Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
53,707 | 13,196,505,956 | IssuesEvent | 2020-08-13 20:47:49 | halide/Halide | https://api.github.com/repos/halide/Halide | closed | CMake/CPack 'distrib' target isn't correct for Windows | build | A while back, someone thoughtfully contributed CPack support to our CMake files, so that `make distrib` worked correctly for CMake as well as make; unfortunately, it needs some extra love for Windows builds in order to match the status quo for Windows:
- Windows requires separate Debug and Release builds; the CMake rules only provide a single build (presumably Release)
- The CMake rules only provide shared-library outputs (bin/Halide.dll), whereas our existing Windows distribs provide Halide.lib as well
This would be really nice to fix, as currently, Windows distribs are produced by custom code in the buildbots, which can (and does) easily fall out of sync with the rest of Halide.
| 1.0 | CMake/CPack 'distrib' target isn't correct for Windows - A while back, someone thoughtfully contributed CPack support to our CMake files, so that `make distrib` worked correctly for CMake as well as make; unfortunately, it needs some extra love for Windows builds in order to match the status quo for Windows:
- Windows requires separate Debug and Release builds; the CMake rules only provide a single build (presumably Release)
- The CMake rules only provide shared-library outputs (bin/Halide.dll), whereas our existing Windows distribs provide Halide.lib as well
This would be really nice to fix, as currently, Windows distribs are produced by custom code in the buildbots, which can (and does) easily fall out of sync with the rest of Halide.
| non_priority | cmake cpack distrib target isn t correct for windows a while back someone thoughtfully contributed cpack support to our cmake files so that make distrib worked correctly for cmake as well as make unfortunately it needs some extra love for windows builds in order to match the status quo for windows windows requires separate debug and release builds the cmake rules only provide a single build presumably release the cmake rules only provide shared library outputs bin halide dll whereas our existing windows distribs provide halide lib as well this would be really nice to fix as currently windows distribs are produced by custom code in the buildbots which can and does easily fall out of sync with the rest of halide | 0 |
331,059 | 28,504,327,977 | IssuesEvent | 2023-04-18 20:00:57 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed | C-bug C-test-failure O-robot O-roachtest GA-blocker T-kv-replication branch-release-23.1.0 | roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on release-23.1.0 @ [ada9871ea92cfc467b045a6a1b8ed2783a243c2f](https://github.com/cockroachdb/cockroach/commits/ada9871ea92cfc467b045a6a1b8ed2783a243c2f):
```
test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1
(monitor.go:127).Wait: monitor failure: monitor task failed: read tcp 172.17.0.3:58046 -> 3.128.192.232:26257: read: connection reset by peer
```
<p>Parameters: <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #100804 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery branch-release-23.1]
- #100341 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [A-kv-replication C-bug C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery T-kv-replication branch-master branch-release-23.1]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-26944 | 2.0 | roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed - roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on release-23.1.0 @ [ada9871ea92cfc467b045a6a1b8ed2783a243c2f](https://github.com/cockroachdb/cockroach/commits/ada9871ea92cfc467b045a6a1b8ed2783a243c2f):
```
test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1
(monitor.go:127).Wait: monitor failure: monitor task failed: read tcp 172.17.0.3:58046 -> 3.128.192.232:26257: read: connection reset by peer
```
<p>Parameters: <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #100804 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery branch-release-23.1]
- #100341 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [A-kv-replication C-bug C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery T-kv-replication branch-master branch-release-23.1]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-26944 | non_priority | roachtest restore tpce aws nodes cpus failed roachtest restore tpce aws nodes cpus with on release test artifacts and logs in artifacts restore tpce aws nodes cpus run monitor go wait monitor failure monitor task failed read tcp read connection reset by peer parameters roachtest cloud aws roachtest cpu roachtest encrypted false roachtest fs roachtest localssd false roachtest ssd help see see same failure on other branches roachtest restore tpce aws nodes cpus failed roachtest restore tpce aws nodes cpus failed cc cockroachdb disaster recovery jira issue crdb | 0 |
72,722 | 19,430,764,592 | IssuesEvent | 2021-12-21 11:40:12 | openego/eGon-data | https://api.github.com/repos/openego/eGon-data | closed | Insert cost assumptions from NEP and technology-data to scenario_parameters | :building_construction: integration | Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39).
Other marginal and capital costs are taken from technology-data repository | 1.0 | Insert cost assumptions from NEP and technology-data to scenario_parameters - Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39).
Other marginal and capital costs are taken from technology-data repository | non_priority | insert cost assumptions from nep and technology data to scenario parameters fuel and prices will be added to the scenario parameters table according to table in nep version entwurf s other marginal and capital costs are taken from technology data repository | 0 |
151,971 | 5,830,952,382 | IssuesEvent | 2017-05-08 18:08:06 | idaholab/raven | https://api.github.com/repos/idaholab/raven | opened | Time Dependent Risk Importance Measures | improvement priority_normal | --------
Issue Description
--------
##### What did you expect to see happen?
This issue refers to the fact that the actual Risk Importance Measures cannot be calculated if time dependent data is provided
##### What did you see instead?
NA
##### Do you have a suggested fix for the development team?
NA
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
Not a defect
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement? improvement
- [ ] 2. Is it tagged with a priority: critical, normal or minor? normal
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? NA
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. NA
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) Yes
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| 1.0 | Time Dependent Risk Importance Measures - --------
Issue Description
--------
##### What did you expect to see happen?
This issue refers to the fact that the actual Risk Importance Measures cannot be calculated if time dependent data is provided
##### What did you see instead?
NA
##### Do you have a suggested fix for the development team?
NA
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
Not a defect
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement? improvement
- [ ] 2. Is it tagged with a priority: critical, normal or minor? normal
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? NA
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. NA
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) Yes
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| priority | time dependent risk importance measures issue description what did you expect to see happen this issue refers to the fact that the actual risk importance measures cannot be calculated if time dependent data is provided what did you see instead na do you have a suggested fix for the development team na please attach the input file s that generate this error the simpler the input the faster we can find the issue not a defect for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement improvement is it tagged with a priority critical normal or minor normal if it will impact requirements or requirements tests is it tagged with requirements na if it is a defect can it cause wrong results for users if so an email needs to be sent to the users na is a rationale provided such as explaining why the improvement is needed or why current code is wrong yes for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided | 1 |
178,061 | 6,598,429,575 | IssuesEvent | 2017-09-16 05:02:53 | minishift/minishift | https://api.github.com/repos/minishift/minishift | closed | Installing CDK fails on Win10 with Hyper-V | kind/bug os/windows priority/major | As noted in #1381 I performed the minimal installation of minishift on my Win10 laptop. Thinking that the CDK might possibly resolve the issue, I downloaded the online installer and ran it. I encountered the following show-stopper error:
```
You have selected Red Hat Container Development Kit, which requires Oracle VirtualBox version 5.1.22 or higher. Please uninstall detected Oracle VirtualBox and restart Installer to continue.
Please enable hardware virtualization support in BIOS for your platform.
```
This error message is nonsense.
* I have enabled virtualization support in the BIOS, and Hyper-V is active.

* Moreover, I have not even installed VirtualBox.

| 1.0 | Installing CDK fails on Win10 with Hyper-V - As noted in #1381 I performed the minimal installation of minishift on my Win10 laptop. Thinking that the CDK might possibly resolve the issue, I downloaded the online installer and ran it. I encountered the following show-stopper error:
```
You have selected Red Hat Container Development Kit, which requires Oracle VirtualBox version 5.1.22 or higher. Please uninstall detected Oracle VirtualBox and restart Installer to continue.
Please enable hardware virtualization support in BIOS for your platform.
```
This error message is nonsense.
* I have enabled virtualization support in the BIOS, and Hyper-V is active.

* Moreover, I have not even installed VirtualBox.

| priority | installing cdk fails on with hyper v as noted in i performed the minimal installation of minishift on my laptop thinking that the cdk might possibly resolve the issue i downloaded the online installer and ran it i encountered the following show stopper error you have selected red hat container development kit which requires oracle virtualbox version or higher please uninstall detected oracle virtualbox and restart installer to continue please enable hardware virtualization support in bios for your platform this error message is nonsense i have enabled virtualization support in the bios and hyper v is active moreover i have not even installed virtualbox | 1 |
279,991 | 8,676,724,382 | IssuesEvent | 2018-11-30 14:55:01 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | opened | User `userId` instead of `username` in collection saga/reducer/api | component: collections priority: p3 | See also: #6609
---
In order to change the collection URLs to use user IDs instead of usernames, we need to change the variable names to reflect this upgrade. In this issue, we should simply rename variables and make sure Flow types and test cases are synchronized.
**For QA:** please make sure collection pages load correctly and that no feature has been altered. | 1.0 | User `userId` instead of `username` in collection saga/reducer/api - See also: #6609
---
In order to change the collection URLs to use user IDs instead of usernames, we need to change the variable names to reflect this upgrade. In this issue, we should simply rename variables and make sure Flow types and test cases are synchronized.
**For QA:** please make sure collection pages load correctly and that no feature has been altered. | priority | user userid instead of username in collection saga reducer api see also in order to change the collection urls to use user ids instead of usernames we need to change the variable names to reflect this upgrade in this issue we should simply rename variables and make sure flow types and test cases are synchronized for qa please make sure collection pages load correctly and that no feature has been altered | 1 |
719,284 | 24,754,464,201 | IssuesEvent | 2022-10-21 16:20:24 | massenergize/frontend-admin | https://api.github.com/repos/massenergize/frontend-admin | closed | DUP of 601 - In HTML Text field, default for links should be to open a new tab | enhancement priority 2 | Currently the default is to use the existing tab. The dialog box says "window",
We want the default to be new tab since that is correct most of the time, and most cadmins may miss this. | 1.0 | DUP of 601 - In HTML Text field, default for links should be to open a new tab - Currently the default is to use the existing tab. The dialog box says "window",
We want the default to be new tab since that is correct most of the time, and most cadmins may miss this. | priority | dup of in html text field default for links should be to open a new tab currently the default is to use the existing tab the dialog box says window we want the default to be new tab since that is correct most of the time and most cadmins may miss this | 1 |
95,036 | 8,528,684,965 | IssuesEvent | 2018-11-03 02:13:01 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: scaledata/jobcoordinator/nodes=3 failed | C-test-failure O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/acd1250b15b7ed3c8938dfd53b8bc53bb53c578c
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stressrace TESTS=scaledata/jobcoordinator/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=999222&tab=buildLog
```
The test failed on master:
test.go:639,cluster.go:1461,scaledata.go:126,scaledata.go:53: signal: interrupt
``` | 1.0 | roachtest: scaledata/jobcoordinator/nodes=3 failed - SHA: https://github.com/cockroachdb/cockroach/commits/acd1250b15b7ed3c8938dfd53b8bc53bb53c578c
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stressrace TESTS=scaledata/jobcoordinator/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=999222&tab=buildLog
```
The test failed on master:
test.go:639,cluster.go:1461,scaledata.go:126,scaledata.go:53: signal: interrupt
``` | non_priority | roachtest scaledata jobcoordinator nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stressrace tests scaledata jobcoordinator nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master test go cluster go scaledata go scaledata go signal interrupt | 0 |
565,735 | 16,768,378,905 | IssuesEvent | 2021-06-14 11:56:10 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.curseforge.com - site is not usable | browser-firefox-ios os-ios priority-normal status-needsinfo | <!-- @browser: Firefox iOS 33.1 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.curseforge.com/minecraft/mc-mods/sophisticated-backpacks
**Browser / Version**: Firefox iOS 33.1
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
When going to the site via a DuckDuckGo search result, a screen appears saying, “Checking your browser before accessing... Redirecting... DDoS protection by Cloudflare...” Then the screen goes blank, i.e., a white screen with no graphics or text.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/25c765a8-03e6-48aa-a913-7af2b26e047d.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.curseforge.com - site is not usable - <!-- @browser: Firefox iOS 33.1 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.curseforge.com/minecraft/mc-mods/sophisticated-backpacks
**Browser / Version**: Firefox iOS 33.1
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
When going to the site via a DuckDuckGo search result, a screen appears saying, “Checking your browser before accessing... Redirecting... DDoS protection by Cloudflare...” Then the screen goes blank, i.e., a white screen with no graphics or text.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/25c765a8-03e6-48aa-a913-7af2b26e047d.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox ios operating system ios tested another browser yes safari problem type site is not usable description page not loading correctly steps to reproduce when going to the site via a duckduckgo search result a screen appears saying “checking your browser before accessing redirecting ddos protection by cloudflare ” then the screen goes blank i e a white screen with no graphics or text view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
74,918 | 14,359,690,965 | IssuesEvent | 2020-11-30 15:56:00 | code4romania/monitorizare-vot-ong | https://api.github.com/repos/code4romania/monitorizare-vot-ong | opened | [BUG] Incorrect polling station info displayed in details screen | angular autumn-2020 bug code4ro-hackdays good first issue help wanted high-priority responses | - Go to answers
- Pick one answers and open the details screen
- Go back to the answers page
- Pick another answer from the list and open the details screen
- Notice that on the top of the details page, the info about the previous polling station in displayed, instead of the info on the currently selected station

| 1.0 | [BUG] Incorrect polling station info displayed in details screen - - Go to answers
- Pick one answers and open the details screen
- Go back to the answers page
- Pick another answer from the list and open the details screen
- Notice that on the top of the details page, the info about the previous polling station in displayed, instead of the info on the currently selected station

| non_priority | incorrect polling station info displayed in details screen go to answers pick one answers and open the details screen go back to the answers page pick another answer from the list and open the details screen notice that on the top of the details page the info about the previous polling station in displayed instead of the info on the currently selected station | 0 |
5,542 | 12,692,902,588 | IssuesEvent | 2020-06-22 01:12:38 | Plant-for-the-Planet-org/treecounter-app | https://api.github.com/repos/Plant-for-the-Planet-org/treecounter-app | opened | Check component/container usage | architecture | Some components do not use associated containers. Check and maybe refactor them. | 1.0 | Check component/container usage - Some components do not use associated containers. Check and maybe refactor them. | non_priority | check component container usage some components do not use associated containers check and maybe refactor them | 0 |
600,300 | 18,293,014,962 | IssuesEvent | 2021-10-05 17:16:19 | FTBTeam/FTB-App | https://api.github.com/repos/FTBTeam/FTB-App | closed | [Feature request] Global Memory Settings Option | enhancement priority/low | **Is your feature request related to a problem? Please describe.**
I download a new mod pack, but the memory settings are set to a fixed default (4gb).
**Describe the solution you'd like**
I'd like to modify the fixed initial default value in a global settings context.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | [Feature request] Global Memory Settings Option - **Is your feature request related to a problem? Please describe.**
I download a new mod pack, but the memory settings are set to a fixed default (4gb).
**Describe the solution you'd like**
I'd like to modify the fixed initial default value in a global settings context.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | global memory settings option is your feature request related to a problem please describe i download a new mod pack but the memory settings are set to a fixed default describe the solution you d like i d like to modify the fixed initial default value in a global settings context describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 1 |
783,059 | 27,517,231,755 | IssuesEvent | 2023-03-06 12:49:58 | NFT-Limited/nf-bugs | https://api.github.com/repos/NFT-Limited/nf-bugs | opened | [Collection/NFT] XDragon Robots | Collection/NFT Priority 1 | **Describe the issue**
Not showing any NFTs from the collection. Reset not helping.
MP4 files.
**Details**
Help us to identify the collection by providing some infos:
REQUIRED
1. Issuer address: rptAoCTeoJg24qgU3WnERE2ino7vzCMXFY
2. Collection Taxon: 523
| 1.0 | [Collection/NFT] XDragon Robots - **Describe the issue**
Not showing any NFTs from the collection. Reset not helping.
MP4 files.
**Details**
Help us to identify the collection by providing some infos:
REQUIRED
1. Issuer address: rptAoCTeoJg24qgU3WnERE2ino7vzCMXFY
2. Collection Taxon: 523
| priority | xdragon robots describe the issue not showing any nfts from the collection reset not helping files details help us to identify the collection by providing some infos required issuer address collection taxon | 1 |
169,000 | 6,393,181,592 | IssuesEvent | 2017-08-04 06:29:33 | CanberraOceanRacingClub/namadgi3 | https://api.github.com/repos/CanberraOceanRacingClub/namadgi3 | closed | Upgrade and fit Gas detector with controller and Solenoid | priority 1: High | Authorised 6/6/17 located on aft of port galley cupboards
who: Quays extra cost to contract | 1.0 | Upgrade and fit Gas detector with controller and Solenoid - Authorised 6/6/17 located on aft of port galley cupboards
who: Quays extra cost to contract | priority | upgrade and fit gas detector with controller and solenoid authorised located on aft of port galley cupboards who quays extra cost to contract | 1 |
308,655 | 26,620,832,715 | IssuesEvent | 2023-01-24 11:05:29 | Kong/kubernetes-ingress-controller | https://api.github.com/repos/Kong/kubernetes-ingress-controller | closed | E2E test failure: data race in `TestDeployAllInOnePostgresWithMultipleReplicas` | bug area/tests | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Data race in tests: https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#logs
```
==================
WARNING: DATA RACE
Read at 0x00c001f1e5b8 by goroutine 309:
bytes.(*Buffer).String()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:65 +0xef
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder.func1()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:333 +0xb9
github.com/stretchr/testify/assert.Eventually.func1()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x39
Previous write at 0x00c001f1e5b8 by goroutine 307:
bytes.(*Buffer).grow()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:145 +0x3c4
bytes.(*Buffer).ReadFrom()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:200 +0x65
io.copyBuffer()
/opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:413 +0x1c5
io.Copy()
/opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:386 +0x64
os/exec.(*Cmd).writerDescriptor.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:407 +0x3d
os/exec.(*Cmd).Start.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:544 +0x35
os/exec.(*Cmd).Start.func2()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:545 +0x47
Goroutine 309 (running) created at:
github.com/stretchr/testify/assert.Eventually()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x3a5
github.com/stretchr/testify/require.Eventually()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/require/require.go:3[61](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:62) +0xb8
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:326 +0x691
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1667
testing.tRunner()
/opt/hostedtoolcache/go/1.19.5/x[64](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:65)/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1493 +0x47
Goroutine 307 (running) created at:
os/exec.(*Cmd).Start()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:543 +0xd74
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:323 +0x450
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1[66](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:67)7
testing.tRunner()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:14[93](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:94) +0x47
==================
```
### Expected Behavior
No data race.
### Steps To Reproduce
_No response_
### Kong Ingress Controller version
_No response_
### Kubernetes version
```shell
1.26
```
### Anything else?
_No response_ | 1.0 | E2E test failure: data race in `TestDeployAllInOnePostgresWithMultipleReplicas` - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Data race in tests: https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#logs
```
==================
WARNING: DATA RACE
Read at 0x00c001f1e5b8 by goroutine 309:
bytes.(*Buffer).String()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:65 +0xef
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder.func1()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:333 +0xb9
github.com/stretchr/testify/assert.Eventually.func1()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x39
Previous write at 0x00c001f1e5b8 by goroutine 307:
bytes.(*Buffer).grow()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:145 +0x3c4
bytes.(*Buffer).ReadFrom()
/opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:200 +0x65
io.copyBuffer()
/opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:413 +0x1c5
io.Copy()
/opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:386 +0x64
os/exec.(*Cmd).writerDescriptor.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:407 +0x3d
os/exec.(*Cmd).Start.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:544 +0x35
os/exec.(*Cmd).Start.func2()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:545 +0x47
Goroutine 309 (running) created at:
github.com/stretchr/testify/assert.Eventually()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x3a5
github.com/stretchr/testify/require.Eventually()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/require/require.go:3[61](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:62) +0xb8
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:326 +0x691
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1667
testing.tRunner()
/opt/hostedtoolcache/go/1.19.5/x[64](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:65)/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1493 +0x47
Goroutine 307 (running) created at:
os/exec.(*Cmd).Start()
/opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:543 +0xd74
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:323 +0x450
github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas()
/home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1[66](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:67)7
testing.tRunner()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:14[93](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:94) +0x47
==================
```
### Expected Behavior
No data race.
### Steps To Reproduce
_No response_
### Kong Ingress Controller version
_No response_
### Kubernetes version
```shell
1.26
```
### Anything else?
_No response_ | non_priority | test failure data race in testdeployallinonepostgreswithmultiplereplicas is there an existing issue for this i have searched the existing issues current behavior data race in tests warning data race read at by goroutine bytes buffer string opt hostedtoolcache go src bytes buffer go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com stretchr testify assert eventually home runner go pkg mod github com stretchr testify assert assertions go previous write at by goroutine bytes buffer grow opt hostedtoolcache go src bytes buffer go bytes buffer readfrom opt hostedtoolcache go src bytes buffer go io copybuffer opt hostedtoolcache go src io io go io copy opt hostedtoolcache go src io io go os exec cmd writerdescriptor opt hostedtoolcache go src os exec exec go os exec cmd start opt hostedtoolcache go src os exec exec go os exec cmd start opt hostedtoolcache go src os exec exec go goroutine running created at github com stretchr testify assert eventually home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify require eventually home runner go pkg mod github com stretchr testify require require go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com kong kubernetes ingress controller test testdeployallinonepostgreswithmultiplereplicas home runner work kubernetes ingress controller kubernetes ingress controller test all in one test go testing trunner opt hostedtoolcache go x testing t run opt hostedtoolcache go src testing testing go goroutine running created at os exec cmd start opt hostedtoolcache go src os exec exec go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com kong kubernetes ingress controller test testdeployallinonepostgreswithmultiplereplicas home runner work kubernetes ingress controller kubernetes ingress controller test all in one test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go expected behavior no data race steps to reproduce no response kong ingress controller version no response kubernetes version shell anything else no response | 0 |
361,461 | 10,709,035,456 | IssuesEvent | 2019-10-24 21:04:03 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate | affected: 2.4 auto-transferred bug category: calib3d priority: normal | Transferred from http://code.opencv.org/issues/2878
```
|| Daniel Danilin on 2013-03-10 16:12
|| Priority: Normal
|| Affected: branch 'master' (2.4.9)
|| Category: calibration, 3d
|| Tracker: Bug
|| Difficulty: None
|| PR:
|| Platform: None / None
```
## "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate
```
I'm getting the "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted." error message if I run stereo_calib example project. Sometimes the variable is 'A'. Here is the stack trace:
<pre>
> opencv_calib3d249d.dll!cvStereoCalibrate(const CvMat * _objectPoints, const CvMat * _imagePoints1, const CvMat * _imagePoints2, const CvMat * _npoints, CvMat * _cameraMatrix1, CvMat * _distCoeffs1, CvMat * _cameraMatrix2, CvMat * _distCoeffs2, CvSize imageSize, CvMat * matR, CvMat * matT, CvMat * matE, CvMat * matF, CvTermCriteria termCrit, int flags) Line 2071 C++
opencv_calib3d249d.dll!cv::stereoCalibrate(const cv::_InputArray & _objectPoints, const cv::_InputArray & _imagePoints1, const cv::_InputArray & _imagePoints2, const cv::_OutputArray & _cameraMatrix1, const cv::_OutputArray & _distCoeffs1, const cv::_OutputArray & _cameraMatrix2, const cv::_OutputArray & _distCoeffs2, cv::Size_<int> imageSize, const cv::_OutputArray & _Rmat, const cv::_OutputArray & _Tmat, const cv::_OutputArray & _Emat, const cv::_OutputArray & _Fmat, cv::TermCriteria criteria, int flags) Line 3305 C++
cpp-example-stereo_calib.exe!StereoCalib(const std::vector<std::basic_string<char,std::char_traits<char>,std::allocator<char> >,std::allocator<std::basic_string<char,std::char_traits<char>,std::allocator<char> > > > & imagelist, cv::Size_<int> boardSize, bool useCalibrated, bool showRectified) Line 183 C++
cpp-example-stereo_calib.exe!main(int argc, char * * argv) Line 404 C++
cpp-example-stereo_calib.exe!__tmainCRTStartup() Line 555 C
cpp-example-stereo_calib.exe!mainCRTStartup() Line 371 C
kernel32.dll!767f8543() Unknown
[Frames below may be incorrect and/or missing, no symbols loaded for kernel32.dll]
ntdll.dll!77eaac69() Unknown
ntdll.dll!77eaac3c() Unknown
</pre>
I'm using origianl stereo_calib.xml and images from repository.
I have Windows 8, CMake 2.8.10.2, Visual Studio 2012, Visual Studio 2010. I have configured the opencv in cmake without cuda. Tried to generate project files for VS 10 and VS 11, but got the same error. The sources code files are from 10.03.2013.
```
## History
| 1.0 | "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate - Transferred from http://code.opencv.org/issues/2878
```
|| Daniel Danilin on 2013-03-10 16:12
|| Priority: Normal
|| Affected: branch 'master' (2.4.9)
|| Category: calibration, 3d
|| Tracker: Bug
|| Difficulty: None
|| PR:
|| Platform: None / None
```
## "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate
```
I'm getting the "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted." error message if I run stereo_calib example project. Sometimes the variable is 'A'. Here is the stack trace:
<pre>
> opencv_calib3d249d.dll!cvStereoCalibrate(const CvMat * _objectPoints, const CvMat * _imagePoints1, const CvMat * _imagePoints2, const CvMat * _npoints, CvMat * _cameraMatrix1, CvMat * _distCoeffs1, CvMat * _cameraMatrix2, CvMat * _distCoeffs2, CvSize imageSize, CvMat * matR, CvMat * matT, CvMat * matE, CvMat * matF, CvTermCriteria termCrit, int flags) Line 2071 C++
opencv_calib3d249d.dll!cv::stereoCalibrate(const cv::_InputArray & _objectPoints, const cv::_InputArray & _imagePoints1, const cv::_InputArray & _imagePoints2, const cv::_OutputArray & _cameraMatrix1, const cv::_OutputArray & _distCoeffs1, const cv::_OutputArray & _cameraMatrix2, const cv::_OutputArray & _distCoeffs2, cv::Size_<int> imageSize, const cv::_OutputArray & _Rmat, const cv::_OutputArray & _Tmat, const cv::_OutputArray & _Emat, const cv::_OutputArray & _Fmat, cv::TermCriteria criteria, int flags) Line 3305 C++
cpp-example-stereo_calib.exe!StereoCalib(const std::vector<std::basic_string<char,std::char_traits<char>,std::allocator<char> >,std::allocator<std::basic_string<char,std::char_traits<char>,std::allocator<char> > > > & imagelist, cv::Size_<int> boardSize, bool useCalibrated, bool showRectified) Line 183 C++
cpp-example-stereo_calib.exe!main(int argc, char * * argv) Line 404 C++
cpp-example-stereo_calib.exe!__tmainCRTStartup() Line 555 C
cpp-example-stereo_calib.exe!mainCRTStartup() Line 371 C
kernel32.dll!767f8543() Unknown
[Frames below may be incorrect and/or missing, no symbols loaded for kernel32.dll]
ntdll.dll!77eaac69() Unknown
ntdll.dll!77eaac3c() Unknown
</pre>
I'm using origianl stereo_calib.xml and images from repository.
I have Windows 8, CMake 2.8.10.2, Visual Studio 2012, Visual Studio 2010. I have configured the opencv in cmake without cuda. Tried to generate project files for VS 10 and VS 11, but got the same error. The sources code files are from 10.03.2013.
```
## History
| priority | run time check failure stack around the variable dk was corrupted in stereocalibrate transferred from daniel danilin on priority normal affected branch master category calibration tracker bug difficulty none pr platform none none run time check failure stack around the variable dk was corrupted in stereocalibrate i m getting the run time check failure stack around the variable dk was corrupted error message if i run stereo calib example project sometimes the variable is a here is the stack trace opencv dll cvstereocalibrate const cvmat objectpoints const cvmat const cvmat const cvmat npoints cvmat cvmat cvmat cvmat cvsize imagesize cvmat matr cvmat matt cvmat mate cvmat matf cvtermcriteria termcrit int flags line c opencv dll cv stereocalibrate const cv inputarray objectpoints const cv inputarray const cv inputarray const cv outputarray const cv outputarray const cv outputarray const cv outputarray cv size imagesize const cv outputarray rmat const cv outputarray tmat const cv outputarray emat const cv outputarray fmat cv termcriteria criteria int flags line c cpp example stereo calib exe stereocalib const std vector std allocator std allocator std allocator imagelist cv size boardsize bool usecalibrated bool showrectified line c cpp example stereo calib exe main int argc char argv line c cpp example stereo calib exe tmaincrtstartup line c cpp example stereo calib exe maincrtstartup line c dll unknown ntdll dll unknown ntdll dll unknown i m using origianl stereo calib xml and images from repository i have windows cmake visual studio visual studio i have configured the opencv in cmake without cuda tried to generate project files for vs and vs but got the same error the sources code files are from history | 1 |
50,833 | 21,420,076,246 | IssuesEvent | 2022-04-22 14:47:51 | microsoft/BotFramework-Composer | https://api.github.com/repos/microsoft/BotFramework-Composer | closed | Bot Feamework Composer - Luis build failed: The model name { boolean } are reserved error when using form dialog feature | Type: Bug customer-reported Bot Services customer-replied-to | <!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
Unable to start bot. Getting Luis build failed: The model name { boolean } are reserved error when trying to use form dialog feature in Bot Framework Composer(v2.1.2)
<!-- Give a clear and concise description of what the bug is. -->
## Version
v2.1.2
<!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). -->
## Screenshots

<!-- If applicable, add screenshots/gif/video to help explain your problem. -->
| 1.0 | Bot Feamework Composer - Luis build failed: The model name { boolean } are reserved error when using form dialog feature - <!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
Unable to start bot. Getting Luis build failed: The model name { boolean } are reserved error when trying to use form dialog feature in Bot Framework Composer(v2.1.2)
<!-- Give a clear and concise description of what the bug is. -->
## Version
v2.1.2
<!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). -->
## Screenshots

<!-- If applicable, add screenshots/gif/video to help explain your problem. -->
| non_priority | bot feamework composer luis build failed the model name boolean are reserved error when using form dialog feature describe the bug unable to start bot getting luis build failed the model name boolean are reserved error when trying to use form dialog feature in bot framework composer version screenshots | 0 |
128,114 | 17,420,887,630 | IssuesEvent | 2021-08-04 01:06:45 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | closed | Failure on initial `pulumi new` doesn't allow for re-use of project name on subsequent `new` | kind/bug resolution/by-design | Initial `pulumi new` failures wont allow for re-using of a project name.
## Expected behavior
Cleanup after initial `new` fails, to force the user to re-run `new` after fixing the issues.
## Current behavior
CLI expects you to run an `up` command
## Steps to reproduce
1. Run `$ pulumi new hello-aws-javascript` on a machine that doesn't have NPM installed
2. Go through onboarding, wait for the failure...
3. Install NPM, and re-run `$ pulumi new hello-aws-javascript` onboarding with the same project name.
Get: `Sorry, '<insert project name>' is not a valid project name. A project with this name already exists.`
| 1.0 | Failure on initial `pulumi new` doesn't allow for re-use of project name on subsequent `new` - Initial `pulumi new` failures wont allow for re-using of a project name.
## Expected behavior
Cleanup after initial `new` fails, to force the user to re-run `new` after fixing the issues.
## Current behavior
CLI expects you to run an `up` command
## Steps to reproduce
1. Run `$ pulumi new hello-aws-javascript` on a machine that doesn't have NPM installed
2. Go through onboarding, wait for the failure...
3. Install NPM, and re-run `$ pulumi new hello-aws-javascript` onboarding with the same project name.
Get: `Sorry, '<insert project name>' is not a valid project name. A project with this name already exists.`
| non_priority | failure on initial pulumi new doesn t allow for re use of project name on subsequent new initial pulumi new failures wont allow for re using of a project name expected behavior cleanup after initial new fails to force the user to re run new after fixing the issues current behavior cli expects you to run an up command steps to reproduce run pulumi new hello aws javascript on a machine that doesn t have npm installed go through onboarding wait for the failure install npm and re run pulumi new hello aws javascript onboarding with the same project name get sorry is not a valid project name a project with this name already exists | 0 |
299,387 | 25,900,992,881 | IssuesEvent | 2022-12-15 05:38:39 | openBackhaul/OperationKeyManagement | https://api.github.com/repos/openBackhaul/OperationKeyManagement | opened | release-number pattern update | testsuite_to_be_changed | Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'.
Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator.
Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder
This scenario "multiple digit in a placeholder" can b:e added to following services:
Service Layer - Acceptance :: Attribute correctness :: release-number checked?
- [ ] /v1/bequeath-your-data-and-die
- [ ] /v1/regard-application
- [ ] /v1/disregard-application
- [ ] /v1/list-applications
- [ ] /v1/regard-updated-link | 1.0 | release-number pattern update - Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'.
Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator.
Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder
This scenario "multiple digit in a placeholder" can b:e added to following services:
Service Layer - Acceptance :: Attribute correctness :: release-number checked?
- [ ] /v1/bequeath-your-data-and-die
- [ ] /v1/regard-application
- [ ] /v1/disregard-application
- [ ] /v1/list-applications
- [ ] /v1/regard-updated-link | non_priority | release number pattern update pattern of release number has been updated to already testcases are available to check for too short release number too long release number letters in release number sign in release number incorrect separator additionally a scenario can be added to test whether in each placeholder for a number only two one or digits are allowed in earlier release number more than two digits are allowed in a placeholder this scenario multiple digit in a placeholder can b e added to following services service layer acceptance attribute correctness release number checked bequeath your data and die regard application disregard application list applications regard updated link | 0 |
349,071 | 24,932,809,719 | IssuesEvent | 2022-10-31 13:03:27 | DuplosFidibuss/academic-time-planner | https://api.github.com/repos/DuplosFidibuss/academic-time-planner | closed | Documentation: describe test projects in methods section | documentation | The test plan project created in #23, the implementation of the loading in #24 and the test Toggl project created in #43 should be described in the methods section. | 1.0 | Documentation: describe test projects in methods section - The test plan project created in #23, the implementation of the loading in #24 and the test Toggl project created in #43 should be described in the methods section. | non_priority | documentation describe test projects in methods section the test plan project created in the implementation of the loading in and the test toggl project created in should be described in the methods section | 0 |
247,813 | 26,735,469,134 | IssuesEvent | 2023-01-30 09:08:45 | MValle21/oathkeeper | https://api.github.com/repos/MValle21/oathkeeper | opened | CVE-2022-25967 (High) detected in eta-1.12.0.tgz | security vulnerability | ## CVE-2022-25967 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eta-1.12.0.tgz</b></p></summary>
<p>Lightweight, fast, and powerful embedded JS template engine</p>
<p>Library home page: <a href="https://registry.npmjs.org/eta/-/eta-1.12.0.tgz">https://registry.npmjs.org/eta/-/eta-1.12.0.tgz</a></p>
<p>Path to dependency file: /docs/package.json</p>
<p>Path to vulnerable library: /docs/node_modules/eta/package.json</p>
<p>
Dependency Hierarchy:
- core-2.0.0-alpha.415a7973f.tgz (Root Library)
- :x: **eta-1.12.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MValle21/oathkeeper/commit/43c00a05bdb772edb5194a57f42ee834b37f3774">43c00a05bdb772edb5194a57f42ee834b37f3774</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package eta before 2.0.0 are vulnerable to Remote Code Execution (RCE) by overwriting template engine configuration variables with view options received from The Express render API. **Note:** This is exploitable only for users who are rendering templates with user-defined data.
<p>Publish Date: 2023-01-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25967>CVE-2022-25967</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-30</p>
<p>Fix Resolution: eta - 2.0.0</p>
</p>
</details>
<p></p>
| True | CVE-2022-25967 (High) detected in eta-1.12.0.tgz - ## CVE-2022-25967 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eta-1.12.0.tgz</b></p></summary>
<p>Lightweight, fast, and powerful embedded JS template engine</p>
<p>Library home page: <a href="https://registry.npmjs.org/eta/-/eta-1.12.0.tgz">https://registry.npmjs.org/eta/-/eta-1.12.0.tgz</a></p>
<p>Path to dependency file: /docs/package.json</p>
<p>Path to vulnerable library: /docs/node_modules/eta/package.json</p>
<p>
Dependency Hierarchy:
- core-2.0.0-alpha.415a7973f.tgz (Root Library)
- :x: **eta-1.12.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MValle21/oathkeeper/commit/43c00a05bdb772edb5194a57f42ee834b37f3774">43c00a05bdb772edb5194a57f42ee834b37f3774</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package eta before 2.0.0 are vulnerable to Remote Code Execution (RCE) by overwriting template engine configuration variables with view options received from The Express render API. **Note:** This is exploitable only for users who are rendering templates with user-defined data.
<p>Publish Date: 2023-01-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25967>CVE-2022-25967</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-30</p>
<p>Fix Resolution: eta - 2.0.0</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in eta tgz cve high severity vulnerability vulnerable library eta tgz lightweight fast and powerful embedded js template engine library home page a href path to dependency file docs package json path to vulnerable library docs node modules eta package json dependency hierarchy core alpha tgz root library x eta tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package eta before are vulnerable to remote code execution rce by overwriting template engine configuration variables with view options received from the express render api note this is exploitable only for users who are rendering templates with user defined data publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution eta | 0 |
302,748 | 9,286,027,591 | IssuesEvent | 2019-03-21 09:15:15 | CSBiology/BioFSharp | https://api.github.com/repos/CSBiology/BioFSharp | closed | [BUG] OboParser neglects first occurences of alt_ids | bug priority-medium project-IO up-for-grabs | ### Description
OboTerms can consists of serveral alt_id items. The OboParser just takes the last occurence of the alt_id keyword and discards previous ones.
### Repro steps
Try to parse the following item (downloaded from http://geneontology.org/page/download-ontology):
[Term]
id: GO:0004748
name: ribonucleoside-diphosphate reductase activity, thioredoxin disulfide as acceptor
namespace: molecular_function
alt_id: GO:0016959
alt_id: GO:0016960
alt_id: GO:0016961
def: "Catalysis…
comment: When thioredoxin…
synonym: "2'-deoxyri…
synonym: "2'-deoxyri…
xref: EC:1.17.4.1
xref: MetaCyc:RIBONUCLEOSIDE-DIP-REDUCTI-RXN
xref: RHEA:23252
is_a: GO:0061731 ! ribonucleoside-diphosphate reductase activity
### Expected behavior
All alt_ids should be stored in the resulting OboTerm
### Actual behavior
Just _GO:0016961_ is stored
| 1.0 | [BUG] OboParser neglects first occurences of alt_ids - ### Description
OboTerms can consists of serveral alt_id items. The OboParser just takes the last occurence of the alt_id keyword and discards previous ones.
### Repro steps
Try to parse the following item (downloaded from http://geneontology.org/page/download-ontology):
[Term]
id: GO:0004748
name: ribonucleoside-diphosphate reductase activity, thioredoxin disulfide as acceptor
namespace: molecular_function
alt_id: GO:0016959
alt_id: GO:0016960
alt_id: GO:0016961
def: "Catalysis…
comment: When thioredoxin…
synonym: "2'-deoxyri…
synonym: "2'-deoxyri…
xref: EC:1.17.4.1
xref: MetaCyc:RIBONUCLEOSIDE-DIP-REDUCTI-RXN
xref: RHEA:23252
is_a: GO:0061731 ! ribonucleoside-diphosphate reductase activity
### Expected behavior
All alt_ids should be stored in the resulting OboTerm
### Actual behavior
Just _GO:0016961_ is stored
| priority | oboparser neglects first occurences of alt ids description oboterms can consists of serveral alt id items the oboparser just takes the last occurence of the alt id keyword and discards previous ones repro steps try to parse the following item downloaded from id go name ribonucleoside diphosphate reductase activity thioredoxin disulfide as acceptor namespace molecular function alt id go alt id go alt id go def catalysis… comment when thioredoxin… synonym deoxyri… synonym deoxyri… xref ec xref metacyc ribonucleoside dip reducti rxn xref rhea is a go ribonucleoside diphosphate reductase activity expected behavior all alt ids should be stored in the resulting oboterm actual behavior just go is stored | 1 |
692,809 | 23,749,569,214 | IssuesEvent | 2022-08-31 19:14:06 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | Unlock my account please | Priority-High (Needed for work) Help wanted | I think I incorrectly entered my password too many times... can you please unlock my account?
Is this the preferred path for fixing such situations? It's my first time...
<img width="413" alt="Screen Shot 2022-08-31 at 10 50 56 AM" src="https://user-images.githubusercontent.com/17605945/187759150-ebb22fc6-0144-41c2-b78c-801a2bca068e.png">
| 1.0 | Unlock my account please - I think I incorrectly entered my password too many times... can you please unlock my account?
Is this the preferred path for fixing such situations? It's my first time...
<img width="413" alt="Screen Shot 2022-08-31 at 10 50 56 AM" src="https://user-images.githubusercontent.com/17605945/187759150-ebb22fc6-0144-41c2-b78c-801a2bca068e.png">
| priority | unlock my account please i think i incorrectly entered my password too many times can you please unlock my account is this the preferred path for fixing such situations it s my first time img width alt screen shot at am src | 1 |
229,436 | 7,574,701,132 | IssuesEvent | 2018-04-23 21:53:31 | kubernetes/kubeadm | https://api.github.com/repos/kubernetes/kubeadm | closed | Nodes registered with FQDN, but labelled with short hostname | kind/bug priority/important-soon | ## What keywords did you search in kubeadm issues before filing this one?
* hostname
## Is this a BUG REPORT or FEATURE REQUEST?
Choose one: BUG REPORT
## Versions
**kubeadm version**: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:40:06Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
**Environment**:
- **Kubernetes version**: 1.9.1
- **Cloud provider or hardware configuration**: AWS
- **OS**: Amazon Linux 2.0 (2017.12) LTS Release Candidate
- **Kernel**: Linux ip-10-67-12-187 4.9.70-2.243.amzn2.x86_64 #1 SMP Thu Jan 4 03:49:00 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
- **Others**:
## What happened?
Build cluster with the following config file:
```yaml
api:
advertiseAddress: 10.67.12.187
bindPort: 6443
apiServerCertSANs:
- master.dev.smartsheet.com
- apiserver.dev.smartsheet.com
authorizationModes:
- Node
- RBAC
certificatesDir: /etc/kubernetes/pki
cloudProvider: aws
etcd:
caFile: ""
certFile: ""
dataDir: /var/lib/etcd
endpoints: null
image: ""
keyFile: ""
imageRepository: gcr.io/google_containers
kubeProxy:
config:
bindAddress: 0.0.0.0
clientConnection:
acceptContentTypes: ""
burst: 10
contentType: application/vnd.kubernetes.protobuf
kubeconfig: /var/lib/kube-proxy/kubeconfig.conf
qps: 5
clusterCIDR: 192.168.0.0/16
configSyncPeriod: 15m0s
conntrack:
max: null
maxPerCore: 32768
min: 131072
tcpCloseWaitTimeout: 1h0m0s
tcpEstablishedTimeout: 24h0m0s
enableProfiling: false
featureGates: ""
healthzBindAddress: 0.0.0.0:10256
hostnameOverride: ""
iptables:
masqueradeAll: false
masqueradeBit: 14
minSyncPeriod: 0s
syncPeriod: 30s
ipvs:
minSyncPeriod: 0s
scheduler: ""
syncPeriod: 30s
metricsBindAddress: 127.0.0.1:10249
mode: ""
oomScoreAdj: -999
portRange: ""
resourceContainer: /kube-proxy
udpTimeoutMilliseconds: 250ms
kubeletConfiguration: {}
kubernetesVersion: v1.9.1
networking:
dnsDomain: cluster.local
podSubnet: 192.168.0.0/16
serviceSubnet: 172.20.0.0/22
nodeName: ip-10-67-12-187.dev.smartsheet.com
token: ""
tokenTTL: 0s
unifiedControlPlaneImage: ""
```
With fully resolvable DNS records for hosts (both forward and reverse records) the nodes register themselves with their FQDN.
```
NAME STATUS ROLES AGE VERSION LABELS
ip-10-67-12-187.dev.smartsheet.com Ready master 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-12-187,node-role.kubernetes.io/master=
ip-10-67-13-11.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-11
ip-10-67-13-178.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-178
ip-10-67-14-40.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-40
ip-10-67-14-66.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-66
ip-10-67-15-129.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-129
ip-10-67-15-22.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-22
```
Some applications target themselves (Rook mon processes) to a specific node using the `kubernetes.io/hostname` label, but use the node name as it is registered formally (i.e. FQDN).
## What you expected to happen?
The `kubernetes.io/hostname` label's value to match the node name.
## How to reproduce it (as minimally and precisely as possible)?
Create a cluster where nodes resolve completely in DNS (both forward and reverse records). Join nodes to the cluster and they should be registered using the node's FQDN. Inspect the `kubernetes.io/hostname` label and its value will be the short hostname instead of the registered node name.
## Anything else we need to know?
I think that the default should be to register the node and set hostname label to the short hostname, but there should be a setting in at least the config file (maybe a CLI switch also) that allows the admin to specify that the FQDN be used in both cases. | 1.0 | Nodes registered with FQDN, but labelled with short hostname - ## What keywords did you search in kubeadm issues before filing this one?
* hostname
## Is this a BUG REPORT or FEATURE REQUEST?
Choose one: BUG REPORT
## Versions
**kubeadm version**: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:40:06Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"}
**Environment**:
- **Kubernetes version**: 1.9.1
- **Cloud provider or hardware configuration**: AWS
- **OS**: Amazon Linux 2.0 (2017.12) LTS Release Candidate
- **Kernel**: Linux ip-10-67-12-187 4.9.70-2.243.amzn2.x86_64 #1 SMP Thu Jan 4 03:49:00 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
- **Others**:
## What happened?
Build cluster with the following config file:
```yaml
api:
advertiseAddress: 10.67.12.187
bindPort: 6443
apiServerCertSANs:
- master.dev.smartsheet.com
- apiserver.dev.smartsheet.com
authorizationModes:
- Node
- RBAC
certificatesDir: /etc/kubernetes/pki
cloudProvider: aws
etcd:
caFile: ""
certFile: ""
dataDir: /var/lib/etcd
endpoints: null
image: ""
keyFile: ""
imageRepository: gcr.io/google_containers
kubeProxy:
config:
bindAddress: 0.0.0.0
clientConnection:
acceptContentTypes: ""
burst: 10
contentType: application/vnd.kubernetes.protobuf
kubeconfig: /var/lib/kube-proxy/kubeconfig.conf
qps: 5
clusterCIDR: 192.168.0.0/16
configSyncPeriod: 15m0s
conntrack:
max: null
maxPerCore: 32768
min: 131072
tcpCloseWaitTimeout: 1h0m0s
tcpEstablishedTimeout: 24h0m0s
enableProfiling: false
featureGates: ""
healthzBindAddress: 0.0.0.0:10256
hostnameOverride: ""
iptables:
masqueradeAll: false
masqueradeBit: 14
minSyncPeriod: 0s
syncPeriod: 30s
ipvs:
minSyncPeriod: 0s
scheduler: ""
syncPeriod: 30s
metricsBindAddress: 127.0.0.1:10249
mode: ""
oomScoreAdj: -999
portRange: ""
resourceContainer: /kube-proxy
udpTimeoutMilliseconds: 250ms
kubeletConfiguration: {}
kubernetesVersion: v1.9.1
networking:
dnsDomain: cluster.local
podSubnet: 192.168.0.0/16
serviceSubnet: 172.20.0.0/22
nodeName: ip-10-67-12-187.dev.smartsheet.com
token: ""
tokenTTL: 0s
unifiedControlPlaneImage: ""
```
With fully resolvable DNS records for hosts (both forward and reverse records) the nodes register themselves with their FQDN.
```
NAME STATUS ROLES AGE VERSION LABELS
ip-10-67-12-187.dev.smartsheet.com Ready master 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-12-187,node-role.kubernetes.io/master=
ip-10-67-13-11.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-11
ip-10-67-13-178.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-178
ip-10-67-14-40.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-40
ip-10-67-14-66.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-66
ip-10-67-15-129.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-129
ip-10-67-15-22.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-22
```
Some applications target themselves (Rook mon processes) to a specific node using the `kubernetes.io/hostname` label, but use the node name as it is registered formally (i.e. FQDN).
## What you expected to happen?
The `kubernetes.io/hostname` label's value to match the node name.
## How to reproduce it (as minimally and precisely as possible)?
Create a cluster where nodes resolve completely in DNS (both forward and reverse records). Join nodes to the cluster and they should be registered using the node's FQDN. Inspect the `kubernetes.io/hostname` label and its value will be the short hostname instead of the registered node name.
## Anything else we need to know?
I think that the default should be to register the node and set hostname label to the short hostname, but there should be a setting in at least the config file (maybe a CLI switch also) that allows the admin to specify that the FQDN be used in both cases. | priority | nodes registered with fqdn but labelled with short hostname what keywords did you search in kubeadm issues before filing this one hostname is this a bug report or feature request choose one bug report versions kubeadm version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux environment kubernetes version cloud provider or hardware configuration aws os amazon linux lts release candidate kernel linux ip smp thu jan utc gnu linux others what happened build cluster with the following config file yaml api advertiseaddress bindport apiservercertsans master dev smartsheet com apiserver dev smartsheet com authorizationmodes node rbac certificatesdir etc kubernetes pki cloudprovider aws etcd cafile certfile datadir var lib etcd endpoints null image keyfile imagerepository gcr io google containers kubeproxy config bindaddress clientconnection acceptcontenttypes burst contenttype application vnd kubernetes protobuf kubeconfig var lib kube proxy kubeconfig conf qps clustercidr configsyncperiod conntrack max null maxpercore min tcpclosewaittimeout tcpestablishedtimeout enableprofiling false featuregates healthzbindaddress hostnameoverride iptables masqueradeall false masqueradebit minsyncperiod syncperiod ipvs minsyncperiod scheduler syncperiod metricsbindaddress mode oomscoreadj portrange resourcecontainer kube proxy udptimeoutmilliseconds kubeletconfiguration kubernetesversion networking dnsdomain cluster local podsubnet servicesubnet nodename ip dev smartsheet com token tokenttl unifiedcontrolplaneimage with fully resolvable dns records for hosts both forward and reverse records the nodes register themselves with their fqdn name status roles age version labels ip dev smartsheet com ready master beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip node role kubernetes io master ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip some applications target themselves rook mon processes to a specific node using the kubernetes io hostname label but use the node name as it is registered formally i e fqdn what you expected to happen the kubernetes io hostname label s value to match the node name how to reproduce it as minimally and precisely as possible create a cluster where nodes resolve completely in dns both forward and reverse records join nodes to the cluster and they should be registered using the node s fqdn inspect the kubernetes io hostname label and its value will be the short hostname instead of the registered node name anything else we need to know i think that the default should be to register the node and set hostname label to the short hostname but there should be a setting in at least the config file maybe a cli switch also that allows the admin to specify that the fqdn be used in both cases | 1 |
341,761 | 30,598,830,235 | IssuesEvent | 2023-07-22 05:12:55 | hajimehoshi/ebiten | https://api.github.com/repos/hajimehoshi/ebiten | closed | .builds, .github/workflows: remove installing development libs for OpenGL | os:linux os:freebsd test | ### Operating System
- [ ] Windows
- [ ] macOS
- [X] Linux
- [X] FreeBSD
- [ ] OpenBSD
- [ ] Android
- [ ] iOS
- [ ] Nintendo Switch
- [ ] Xbox
- [ ] Web Browsers
### What feature would you like to be added?
After https://github.com/hajimehoshi/ebiten/commit/0324db5657e133f85d967ab7b7eab98600935288, we no longer need to install libraries for OpenGL when compiling. Let's simplify yml files.
### Why is this needed?
_No response_ | 1.0 | .builds, .github/workflows: remove installing development libs for OpenGL - ### Operating System
- [ ] Windows
- [ ] macOS
- [X] Linux
- [X] FreeBSD
- [ ] OpenBSD
- [ ] Android
- [ ] iOS
- [ ] Nintendo Switch
- [ ] Xbox
- [ ] Web Browsers
### What feature would you like to be added?
After https://github.com/hajimehoshi/ebiten/commit/0324db5657e133f85d967ab7b7eab98600935288, we no longer need to install libraries for OpenGL when compiling. Let's simplify yml files.
### Why is this needed?
_No response_ | non_priority | builds github workflows remove installing development libs for opengl operating system windows macos linux freebsd openbsd android ios nintendo switch xbox web browsers what feature would you like to be added after we no longer need to install libraries for opengl when compiling let s simplify yml files why is this needed no response | 0 |
230,031 | 25,409,538,999 | IssuesEvent | 2022-11-22 17:45:04 | elastic/beats | https://api.github.com/repos/elastic/beats | opened | [Auditbeat] Prepare System Package to be GA | enhancement Auditbeat Team:Security-External Integrations | **Describe the enhancement:**
We previously had plans to deprecate the `system.package` dataset because in within the Elastic ecosystem we expected that `osquerybeat` could provide this data using various tables like `rpm_package` and `deb_packages`. However, because it cannot provide deltas between previous state and current state the data is not that useful on its own. So we want to improve the Auditbeat system.package dataset such that it can be supported as GA feature and exposed through Elastic Agent.
- [ ] Create a beta Fleet integration for package monitoring that wraps the Auditbeat system.package dataset.
- [ ] Add fsnotify support for triggering updates in near real-time.
- [ ] Migrate to using FlatBuffers and the means of encoding persistent state. This will ensure we have a stable schema for this data between versions and avoid accidental breakages.
- [ ] Update documentation for Auditbeat package dataset.
- [ ] Update documentation related to Auditbeat to Agent migration specifically related to `system.package`. For reference this was added in https://github.com/elastic/observability-docs/pull/2270.
- [ ] Document the Fleet integration as GA using at least version 1.0.0 for the package.
**Describe a specific use case for the enhancement or feature:**
### References
- Replaces https://github.com/elastic/beats/issues/33481
| True | [Auditbeat] Prepare System Package to be GA - **Describe the enhancement:**
We previously had plans to deprecate the `system.package` dataset because in within the Elastic ecosystem we expected that `osquerybeat` could provide this data using various tables like `rpm_package` and `deb_packages`. However, because it cannot provide deltas between previous state and current state the data is not that useful on its own. So we want to improve the Auditbeat system.package dataset such that it can be supported as GA feature and exposed through Elastic Agent.
- [ ] Create a beta Fleet integration for package monitoring that wraps the Auditbeat system.package dataset.
- [ ] Add fsnotify support for triggering updates in near real-time.
- [ ] Migrate to using FlatBuffers and the means of encoding persistent state. This will ensure we have a stable schema for this data between versions and avoid accidental breakages.
- [ ] Update documentation for Auditbeat package dataset.
- [ ] Update documentation related to Auditbeat to Agent migration specifically related to `system.package`. For reference this was added in https://github.com/elastic/observability-docs/pull/2270.
- [ ] Document the Fleet integration as GA using at least version 1.0.0 for the package.
**Describe a specific use case for the enhancement or feature:**
### References
- Replaces https://github.com/elastic/beats/issues/33481
| non_priority | prepare system package to be ga describe the enhancement we previously had plans to deprecate the system package dataset because in within the elastic ecosystem we expected that osquerybeat could provide this data using various tables like rpm package and deb packages however because it cannot provide deltas between previous state and current state the data is not that useful on its own so we want to improve the auditbeat system package dataset such that it can be supported as ga feature and exposed through elastic agent create a beta fleet integration for package monitoring that wraps the auditbeat system package dataset add fsnotify support for triggering updates in near real time migrate to using flatbuffers and the means of encoding persistent state this will ensure we have a stable schema for this data between versions and avoid accidental breakages update documentation for auditbeat package dataset update documentation related to auditbeat to agent migration specifically related to system package for reference this was added in document the fleet integration as ga using at least version for the package describe a specific use case for the enhancement or feature references replaces | 0 |
202,136 | 15,821,453,110 | IssuesEvent | 2021-04-05 20:33:16 | leaf-ai/studio-go-runner | https://api.github.com/repos/leaf-ai/studio-go-runner | closed | Document the Kubernetes Auto-Scaler smoke test for AWS users using eksctl | documentation enhancement | As a cluster operator
I want to be able to manually test the Kubernetes node scaling
In order that during installation baseline ASG can be verified as working
| 1.0 | Document the Kubernetes Auto-Scaler smoke test for AWS users using eksctl - As a cluster operator
I want to be able to manually test the Kubernetes node scaling
In order that during installation baseline ASG can be verified as working
| non_priority | document the kubernetes auto scaler smoke test for aws users using eksctl as a cluster operator i want to be able to manually test the kubernetes node scaling in order that during installation baseline asg can be verified as working | 0 |
413,800 | 27,969,990,244 | IssuesEvent | 2023-03-25 00:25:28 | Webbutvecklare-CL/cl-sektionen.se | https://api.github.com/repos/Webbutvecklare-CL/cl-sektionen.se | closed | Utseende på textsidor | documentation | Just nu har ex VFU och Illabehandling två olika utseende. En ligger på bakgrunden och en ligger som en artikel. Exprimentera och se vilken som fungerar bäst. | 1.0 | Utseende på textsidor - Just nu har ex VFU och Illabehandling två olika utseende. En ligger på bakgrunden och en ligger som en artikel. Exprimentera och se vilken som fungerar bäst. | non_priority | utseende på textsidor just nu har ex vfu och illabehandling två olika utseende en ligger på bakgrunden och en ligger som en artikel exprimentera och se vilken som fungerar bäst | 0 |
617,031 | 19,339,422,120 | IssuesEvent | 2021-12-15 01:29:48 | spaceone-dev/spaceone-design-system | https://api.github.com/repos/spaceone-dev/spaceone-design-system | closed | [Data Table] Update slot props | bug Priority: Low | **Describe the bug**
slot으로 받는 `index` 값이 열 index인지 행 index인지 구분하기가 힘듦.
명확한 구분을 위해 `index`라 되어 있는 값들을 `colIndex`와 `rowIndex`이름으로 추가.
`index`는 이후에 삭제. | 1.0 | [Data Table] Update slot props - **Describe the bug**
slot으로 받는 `index` 값이 열 index인지 행 index인지 구분하기가 힘듦.
명확한 구분을 위해 `index`라 되어 있는 값들을 `colIndex`와 `rowIndex`이름으로 추가.
`index`는 이후에 삭제. | priority | update slot props describe the bug slot으로 받는 index 값이 열 index인지 행 index인지 구분하기가 힘듦 명확한 구분을 위해 index 라 되어 있는 값들을 colindex 와 rowindex 이름으로 추가 index 는 이후에 삭제 | 1 |
395,289 | 11,683,343,289 | IssuesEvent | 2020-03-05 03:04:01 | SnorriDev/thoth | https://api.github.com/repos/SnorriDev/thoth | closed | Move some input checking to a new broadcastBinding event. | low priority wontfix | See Player.java. Some input checking can apparently be moved to a new Event. | 1.0 | Move some input checking to a new broadcastBinding event. - See Player.java. Some input checking can apparently be moved to a new Event. | priority | move some input checking to a new broadcastbinding event see player java some input checking can apparently be moved to a new event | 1 |
32,077 | 6,706,740,103 | IssuesEvent | 2017-10-12 08:23:03 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Issues with Postgres RETURNING with field expressions | C: Functionality P: Medium T: Defect | In the RETURNING clause, Postgres supports "any expression using the table's columns is allowed". jOOQ appears to partially support this, but the behavior varies in surprising ways based on how the library is used. I'd expect this to either be supported and always work or be unsupported and never work, rather than the current behavior where it sometimes works.
I've noticed two issues related to this, so I'm documenting them both here — one involving the generated SQL, and one involving the fields on the resulting Records.
### Expected behavior and actual behavior:
#### Generated SQL Issue
Given the following code:
```java
DSLContext dsl = DSL.using(getConnection(), SQLDialect.POSTGRES);
Table<Record> table = new TableImpl<>(DSL.name("test"));
Field<String> field = DSL.field("field", String.class);
Field<Integer> aliasedFieldAsInt = field.cast(Integer.class).as("int_field");
System.out.println(dsl.insertInto(table).columns(field).values("1").returning(aliasedFieldAsInt).getSQL());
```
I would expect the output to be:
```sql
insert into "test" (field) values (?) returning cast(field as int) as "int_field"
```
However, it is instead:
```sql
insert into "test" (field) values (?) returning "int_field"
```
Note that if I declare the field without an alias, it works as expected:
```java
Field<Integer> fieldAsInt = field.cast(Integer.class);
System.out.println(dsl.insertInto(table).columns(field).values("1").returning(fieldAsInt).getSQL());
```
The above returns the expected value:
```sql
insert into "public"."test" (field) values (?) returning id, cast(field as int)
```
#### asMap() Issue
A second issue I discovered with RETURNING is that calling `fields()` and `intoMap()` on the result works as expected when the table is manually created as above. However, if the table is generated using the schema generation (or the table is manually generated in a manner consistent with the schema generation, namely using the static `createField` method on `AbstractTable`), the `fields()` method returns the list of fields on that table, not the list of fields that were declared in the `returning` method.
For example, given a table named "test" in schema "public" with columns "id" and "field" and a generated `Test` table class, and the following code:
```java
Table<Record> table = Test.TEST;
Field<String> field = DSL.field("field", String.class);
InsertResultStep<Record> returningAsInt = dsl.insertInto(table).columns(field).values("1").returning(field, fieldAsInt);
System.out.println(Arrays.toString(returningAsInt.fetchOne().fields()));
System.out.println(returningAsInt.fetchOne().intoMap());
```
I would expect the output to be analogous to the that of an equivalent select statement:
```
[field, cast(field as integer)]
{field=1, cast=1}
```
But it is instead:
```
["public"."test"."id", "public"."test"."field"]
{id=null, field=1}
``` | 1.0 | Issues with Postgres RETURNING with field expressions - In the RETURNING clause, Postgres supports "any expression using the table's columns is allowed". jOOQ appears to partially support this, but the behavior varies in surprising ways based on how the library is used. I'd expect this to either be supported and always work or be unsupported and never work, rather than the current behavior where it sometimes works.
I've noticed two issues related to this, so I'm documenting them both here — one involving the generated SQL, and one involving the fields on the resulting Records.
### Expected behavior and actual behavior:
#### Generated SQL Issue
Given the following code:
```java
DSLContext dsl = DSL.using(getConnection(), SQLDialect.POSTGRES);
Table<Record> table = new TableImpl<>(DSL.name("test"));
Field<String> field = DSL.field("field", String.class);
Field<Integer> aliasedFieldAsInt = field.cast(Integer.class).as("int_field");
System.out.println(dsl.insertInto(table).columns(field).values("1").returning(aliasedFieldAsInt).getSQL());
```
I would expect the output to be:
```sql
insert into "test" (field) values (?) returning cast(field as int) as "int_field"
```
However, it is instead:
```sql
insert into "test" (field) values (?) returning "int_field"
```
Note that if I declare the field without an alias, it works as expected:
```java
Field<Integer> fieldAsInt = field.cast(Integer.class);
System.out.println(dsl.insertInto(table).columns(field).values("1").returning(fieldAsInt).getSQL());
```
The above returns the expected value:
```sql
insert into "public"."test" (field) values (?) returning id, cast(field as int)
```
#### asMap() Issue
A second issue I discovered with RETURNING is that calling `fields()` and `intoMap()` on the result works as expected when the table is manually created as above. However, if the table is generated using the schema generation (or the table is manually generated in a manner consistent with the schema generation, namely using the static `createField` method on `AbstractTable`), the `fields()` method returns the list of fields on that table, not the list of fields that were declared in the `returning` method.
For example, given a table named "test" in schema "public" with columns "id" and "field" and a generated `Test` table class, and the following code:
```java
Table<Record> table = Test.TEST;
Field<String> field = DSL.field("field", String.class);
InsertResultStep<Record> returningAsInt = dsl.insertInto(table).columns(field).values("1").returning(field, fieldAsInt);
System.out.println(Arrays.toString(returningAsInt.fetchOne().fields()));
System.out.println(returningAsInt.fetchOne().intoMap());
```
I would expect the output to be analogous to the that of an equivalent select statement:
```
[field, cast(field as integer)]
{field=1, cast=1}
```
But it is instead:
```
["public"."test"."id", "public"."test"."field"]
{id=null, field=1}
``` | non_priority | issues with postgres returning with field expressions in the returning clause postgres supports any expression using the table s columns is allowed jooq appears to partially support this but the behavior varies in surprising ways based on how the library is used i d expect this to either be supported and always work or be unsupported and never work rather than the current behavior where it sometimes works i ve noticed two issues related to this so i m documenting them both here — one involving the generated sql and one involving the fields on the resulting records expected behavior and actual behavior generated sql issue given the following code java dslcontext dsl dsl using getconnection sqldialect postgres table table new tableimpl dsl name test field field dsl field field string class field aliasedfieldasint field cast integer class as int field system out println dsl insertinto table columns field values returning aliasedfieldasint getsql i would expect the output to be sql insert into test field values returning cast field as int as int field however it is instead sql insert into test field values returning int field note that if i declare the field without an alias it works as expected java field fieldasint field cast integer class system out println dsl insertinto table columns field values returning fieldasint getsql the above returns the expected value sql insert into public test field values returning id cast field as int asmap issue a second issue i discovered with returning is that calling fields and intomap on the result works as expected when the table is manually created as above however if the table is generated using the schema generation or the table is manually generated in a manner consistent with the schema generation namely using the static createfield method on abstracttable the fields method returns the list of fields on that table not the list of fields that were declared in the returning method for example given a table named test in schema public with columns id and field and a generated test table class and the following code java table table test test field field dsl field field string class insertresultstep returningasint dsl insertinto table columns field values returning field fieldasint system out println arrays tostring returningasint fetchone fields system out println returningasint fetchone intomap i would expect the output to be analogous to the that of an equivalent select statement field cast but it is instead id null field | 0 |
434,248 | 12,515,947,868 | IssuesEvent | 2020-06-03 08:34:58 | eclipse/codewind | https://api.github.com/repos/eclipse/codewind | closed | SVT:Hybrid:Load test request status stuck at "Requested" state | area/portal kind/bug priority/hot | <!-- Please fill out the following form to report a bug. If some fields do not apply to your situation, feel free to skip them.-->
**Codewind version:** 0.12.0
**OS:** CoreOS/RHEL
**Che version:**
**IDE extension version:** 0.12.0
**IDE version:** Eclipse 2019-09
**Kubernetes cluster:** OCP 4.3 / OCP 4.4 GA
**Description:**
I created Appsody open liberty project and tried to submit load test request and it got stuck at "Requested" state and it never changed after that. We have seen this problem during 0.10.0 and 0.11.0 releases and this has been fixed and we could not reproduce during 0.11.0 reelase.
Looks like this problem still exist and not completely addressed.
Here are all the logs from my cluster where I have seen this problem.

[appsodyolhybcw012ocp44ga.txt](https://github.com/eclipse/codewind/files/4609859/appsodyolhybcw012ocp44ga.txt)
[gatekeeper.txt](https://github.com/eclipse/codewind/files/4609860/gatekeeper.txt)
[performance.txt](https://github.com/eclipse/codewind/files/4609861/performance.txt)
[pfe.txt](https://github.com/eclipse/codewind/files/4609862/pfe.txt)
**Steps to reproduce:**
1. Install Codewind 0.12.0 on Hybrid.
2. Create Appsody open liberty project.
3. Request load test after project started.
4. Load test stuck at "Requested" state.
**Workaround:**
<!-- Did you find a way to work around the bug? If so, please describe how you worked around it.-->
Current workaround is to manually refresh your browser page
| 1.0 | SVT:Hybrid:Load test request status stuck at "Requested" state - <!-- Please fill out the following form to report a bug. If some fields do not apply to your situation, feel free to skip them.-->
**Codewind version:** 0.12.0
**OS:** CoreOS/RHEL
**Che version:**
**IDE extension version:** 0.12.0
**IDE version:** Eclipse 2019-09
**Kubernetes cluster:** OCP 4.3 / OCP 4.4 GA
**Description:**
I created Appsody open liberty project and tried to submit load test request and it got stuck at "Requested" state and it never changed after that. We have seen this problem during 0.10.0 and 0.11.0 releases and this has been fixed and we could not reproduce during 0.11.0 reelase.
Looks like this problem still exist and not completely addressed.
Here are all the logs from my cluster where I have seen this problem.

[appsodyolhybcw012ocp44ga.txt](https://github.com/eclipse/codewind/files/4609859/appsodyolhybcw012ocp44ga.txt)
[gatekeeper.txt](https://github.com/eclipse/codewind/files/4609860/gatekeeper.txt)
[performance.txt](https://github.com/eclipse/codewind/files/4609861/performance.txt)
[pfe.txt](https://github.com/eclipse/codewind/files/4609862/pfe.txt)
**Steps to reproduce:**
1. Install Codewind 0.12.0 on Hybrid.
2. Create Appsody open liberty project.
3. Request load test after project started.
4. Load test stuck at "Requested" state.
**Workaround:**
<!-- Did you find a way to work around the bug? If so, please describe how you worked around it.-->
Current workaround is to manually refresh your browser page
| priority | svt hybrid load test request status stuck at requested state codewind version os coreos rhel che version ide extension version ide version eclipse kubernetes cluster ocp ocp ga description i created appsody open liberty project and tried to submit load test request and it got stuck at requested state and it never changed after that we have seen this problem during and releases and this has been fixed and we could not reproduce during reelase looks like this problem still exist and not completely addressed here are all the logs from my cluster where i have seen this problem steps to reproduce install codewind on hybrid create appsody open liberty project request load test after project started load test stuck at requested state workaround current workaround is to manually refresh your browser page | 1 |
56,320 | 31,863,180,325 | IssuesEvent | 2023-09-15 12:27:48 | levibostian/action-bump-major-tag | https://api.github.com/repos/levibostian/action-bump-major-tag | opened | Reduce network calls for getting tags list | performance | # Expected
When getting the repo's git tags, reduce down the number of network requests as little as possible.
# Actual
Today, the action does paging to get all of the repo's git tags.
# Ideas
* Use github's graphql api to specify in more detail what types of tags we want.
* Use caching. | True | Reduce network calls for getting tags list - # Expected
When getting the repo's git tags, reduce down the number of network requests as little as possible.
# Actual
Today, the action does paging to get all of the repo's git tags.
# Ideas
* Use github's graphql api to specify in more detail what types of tags we want.
* Use caching. | non_priority | reduce network calls for getting tags list expected when getting the repo s git tags reduce down the number of network requests as little as possible actual today the action does paging to get all of the repo s git tags ideas use github s graphql api to specify in more detail what types of tags we want use caching | 0 |
76,162 | 7,520,216,392 | IssuesEvent | 2018-04-12 13:56:19 | EOSIO/eos | https://api.github.com/repos/EOSIO/eos | closed | nodeos --config doesn't support path with whitespace on Mac | needs testing - pass | ````
./programs/nodeos/nodeos --config=/Users/antonperkov/Library/Application\ Support/eosio.rel/nodeos/config/config.ini -d /Users/antonperkov/Library/Application\ Support/eosio.rel/ --plugin eosio::wallet_api_plugin --plugin eosio::chain_api_plugin --plugin eosio::account_history_api_plugin --resync
````
fails with
```
140142ms thread-0 main.cpp:97 main ] Throw location unknown (consider using BOOST_THROW_EXCEPTION)
Dynamic exception type: boost::exception_detail::clone_impl<boost::exception_detail::error_info_injector<boost::program_options::invalid_option_value> >
std::exception::what: the argument ('/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini') for option '--config' is invalid
```
stacktrace:
````
(lldb) bt
* thread #1, queue = 'com.apple.main-thread', stop reason = breakpoint 1.1
* frame #0: 0x00007fff7304d1f4 libc++abi.dylib`__cxa_throw
frame #1: 0x00000001000f23e2 nodeos`void boost::throw_exception<boost::bad_lexical_cast>(e=0x00007ffeefbfa068) at throw_exception.hpp:69
frame #2: 0x000000010033e77f nodeos`void boost::conversion::detail::throw_bad_cast<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, boost::filesystem::path>() at bad_lexical_cast.hpp:92
frame #3: 0x000000010033e68a nodeos`boost::filesystem::path boost::lexical_cast<boost::filesystem::path, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > >(arg="/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini") at lexical_cast.hpp:42
frame #4: 0x000000010033e525 nodeos`void boost::program_options::validate<boost::filesystem::path, char>(v=0x00000001076127e8, xs=size=1, (null)=0x0000000000000000, (null)=0) at value_semantic.hpp:92
frame #5: 0x000000010033e04e nodeos`boost::program_options::typed_value<boost::filesystem::path, char>::xparse(this=0x0000000107612620, value_store=0x00000001076127e8, new_tokens=size=1) const at value_semantic.hpp:184
frame #6: 0x00000001018f1119 nodeos`boost::program_options::store(options=0x00007ffeefbfaae0, xm=0x00007ffeefbfab08, utf8=<unavailable>) at variables_map.cpp:81 [opt]
frame #7: 0x00000001000341a8 nodeos`appbase::application::initialize_impl(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980, autostart_plugins=size=4) at application.cpp:107
frame #8: 0x0000000100012243 nodeos`bool appbase::application::initialize<eosio::chain_plugin, eosio::http_plugin, eosio::net_plugin, eosio::producer_plugin>(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980) at application.hpp:62
frame #9: 0x000000010000a4d1 nodeos`main(argc=11, argv=0x00007ffeefbff980) at main.cpp:87
frame #10: 0x00007fff74faa115 libdyld.dylib`start + 1
````
````
34 namespace boost
35 {
36 template <typename Target, typename Source>
37 inline Target lexical_cast(const Source &arg)
38 {
(lldb) l
39 Target result = Target();
40
41 if (!boost::conversion::detail::try_lexical_convert(arg, result)) {
42 boost::conversion::detail::throw_bad_cast<Source, Target>();
43 }
44
45 return result;
46 }
47
48 template <typename Target>
(lldb) p arg
(const std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) $1 = "/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini"
(lldb) p result
(boost::filesystem::path) $0 = (m_pathname = "/Users/antonperkov/Library/Application")
````
| 1.0 | nodeos --config doesn't support path with whitespace on Mac - ````
./programs/nodeos/nodeos --config=/Users/antonperkov/Library/Application\ Support/eosio.rel/nodeos/config/config.ini -d /Users/antonperkov/Library/Application\ Support/eosio.rel/ --plugin eosio::wallet_api_plugin --plugin eosio::chain_api_plugin --plugin eosio::account_history_api_plugin --resync
````
fails with
```
140142ms thread-0 main.cpp:97 main ] Throw location unknown (consider using BOOST_THROW_EXCEPTION)
Dynamic exception type: boost::exception_detail::clone_impl<boost::exception_detail::error_info_injector<boost::program_options::invalid_option_value> >
std::exception::what: the argument ('/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini') for option '--config' is invalid
```
stacktrace:
````
(lldb) bt
* thread #1, queue = 'com.apple.main-thread', stop reason = breakpoint 1.1
* frame #0: 0x00007fff7304d1f4 libc++abi.dylib`__cxa_throw
frame #1: 0x00000001000f23e2 nodeos`void boost::throw_exception<boost::bad_lexical_cast>(e=0x00007ffeefbfa068) at throw_exception.hpp:69
frame #2: 0x000000010033e77f nodeos`void boost::conversion::detail::throw_bad_cast<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, boost::filesystem::path>() at bad_lexical_cast.hpp:92
frame #3: 0x000000010033e68a nodeos`boost::filesystem::path boost::lexical_cast<boost::filesystem::path, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > >(arg="/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini") at lexical_cast.hpp:42
frame #4: 0x000000010033e525 nodeos`void boost::program_options::validate<boost::filesystem::path, char>(v=0x00000001076127e8, xs=size=1, (null)=0x0000000000000000, (null)=0) at value_semantic.hpp:92
frame #5: 0x000000010033e04e nodeos`boost::program_options::typed_value<boost::filesystem::path, char>::xparse(this=0x0000000107612620, value_store=0x00000001076127e8, new_tokens=size=1) const at value_semantic.hpp:184
frame #6: 0x00000001018f1119 nodeos`boost::program_options::store(options=0x00007ffeefbfaae0, xm=0x00007ffeefbfab08, utf8=<unavailable>) at variables_map.cpp:81 [opt]
frame #7: 0x00000001000341a8 nodeos`appbase::application::initialize_impl(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980, autostart_plugins=size=4) at application.cpp:107
frame #8: 0x0000000100012243 nodeos`bool appbase::application::initialize<eosio::chain_plugin, eosio::http_plugin, eosio::net_plugin, eosio::producer_plugin>(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980) at application.hpp:62
frame #9: 0x000000010000a4d1 nodeos`main(argc=11, argv=0x00007ffeefbff980) at main.cpp:87
frame #10: 0x00007fff74faa115 libdyld.dylib`start + 1
````
````
34 namespace boost
35 {
36 template <typename Target, typename Source>
37 inline Target lexical_cast(const Source &arg)
38 {
(lldb) l
39 Target result = Target();
40
41 if (!boost::conversion::detail::try_lexical_convert(arg, result)) {
42 boost::conversion::detail::throw_bad_cast<Source, Target>();
43 }
44
45 return result;
46 }
47
48 template <typename Target>
(lldb) p arg
(const std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) $1 = "/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini"
(lldb) p result
(boost::filesystem::path) $0 = (m_pathname = "/Users/antonperkov/Library/Application")
````
| non_priority | nodeos config doesn t support path with whitespace on mac programs nodeos nodeos config users antonperkov library application support eosio rel nodeos config config ini d users antonperkov library application support eosio rel plugin eosio wallet api plugin plugin eosio chain api plugin plugin eosio account history api plugin resync fails with thread main cpp main throw location unknown consider using boost throw exception dynamic exception type boost exception detail clone impl std exception what the argument users antonperkov library application support eosio rel nodeos config config ini for option config is invalid stacktrace lldb bt thread queue com apple main thread stop reason breakpoint frame libc abi dylib cxa throw frame nodeos void boost throw exception e at throw exception hpp frame nodeos void boost conversion detail throw bad cast std allocator boost filesystem path at bad lexical cast hpp frame nodeos boost filesystem path boost lexical cast std allocator arg users antonperkov library application support eosio rel nodeos config config ini at lexical cast hpp frame nodeos void boost program options validate v xs size null null at value semantic hpp frame nodeos boost program options typed value xparse this value store new tokens size const at value semantic hpp frame nodeos boost program options store options xm at variables map cpp frame nodeos appbase application initialize impl this argc argv autostart plugins size at application cpp frame nodeos bool appbase application initialize this argc argv at application hpp frame nodeos main argc argv at main cpp frame libdyld dylib start namespace boost template inline target lexical cast const source arg lldb l target result target if boost conversion detail try lexical convert arg result boost conversion detail throw bad cast return result template lldb p arg const std basic string std allocator users antonperkov library application support eosio rel nodeos config config ini lldb p result boost filesystem path m pathname users antonperkov library application | 0 |
822,775 | 30,884,415,939 | IssuesEvent | 2023-08-03 20:22:43 | GoogleCloudPlatform/vertex-ai-samples | https://api.github.com/repos/GoogleCloudPlatform/vertex-ai-samples | closed | error: subprocess-exited-with-error | Deploying Iris-detection model using FastAPI and Vertex AI custom container serving | priority: p1 | ## Expected Behavior
No errors when installing packages from the requirements file from the notebook https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb
## Actual Behavior
error: subprocess-exited-with-error
× Preparing metadata (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [794 lines of output]
<string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html
Partial import of sklearn during the build process.
<string>:116: DeprecationWarning:
`numpy.distutils` is deprecated since NumPy 1.23.0, as a result
of the deprecation of `distutils` itself. It will be removed for
Python >= 3.12. For older Python versions it will remain present.
It is recommended to use `setuptools < 60.0` for those Python versions.
For more details, see:
https://numpy.org/devdocs/reference/distutils_status_migration.html
## Steps to Reproduce the Problem
1. Used the Vertex sample, https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb
2. Create the Requirements file:
%%writefile requirements.txt
joblib~=1.0
numpy~=1.20
scikit-learn~=0.24
google-cloud-storage>=1.26.0,<2.0.0dev
2. Run the install:
# Required in Docker serving container
! pip3 install -U -r requirements.txt -q
# For local FastAPI development and running
! pip3 install -U "uvicorn[standard]>=0.12.0,<0.14.0" fastapi~=0.63 -q
# Vertex SDK for Python
! pip3 install -U google-cloud-aiplatform -q
3. Observe the error:
error: subprocess-exited-with-error
× Preparing metadata (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [794 lines of output]
<string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html
Partial import of sklearn during the build process.
<string>:116: DeprecationWarning:
`numpy.distutils` is deprecated since NumPy 1.23.0, as a result
of the deprecation of `distutils` itself. It will be removed for
Python >= 3.12. For older Python versions it will remain present.
It is recommended to use `setuptools < 60.0` for those Python versions.
For more details, see:
https://numpy.org/devdocs/reference/distutils_status_migration.html
<SNIP>
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
× Encountered error while generating package metadata.
╰─> See above for output.
note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
## Specifications
- Version: N/A
- Platform: N/A | 1.0 | error: subprocess-exited-with-error | Deploying Iris-detection model using FastAPI and Vertex AI custom container serving - ## Expected Behavior
No errors when installing packages from the requirements file from the notebook https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb
## Actual Behavior
error: subprocess-exited-with-error
× Preparing metadata (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [794 lines of output]
<string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html
Partial import of sklearn during the build process.
<string>:116: DeprecationWarning:
`numpy.distutils` is deprecated since NumPy 1.23.0, as a result
of the deprecation of `distutils` itself. It will be removed for
Python >= 3.12. For older Python versions it will remain present.
It is recommended to use `setuptools < 60.0` for those Python versions.
For more details, see:
https://numpy.org/devdocs/reference/distutils_status_migration.html
## Steps to Reproduce the Problem
1. Used the Vertex sample, https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb
2. Create the Requirements file:
%%writefile requirements.txt
joblib~=1.0
numpy~=1.20
scikit-learn~=0.24
google-cloud-storage>=1.26.0,<2.0.0dev
2. Run the install:
# Required in Docker serving container
! pip3 install -U -r requirements.txt -q
# For local FastAPI development and running
! pip3 install -U "uvicorn[standard]>=0.12.0,<0.14.0" fastapi~=0.63 -q
# Vertex SDK for Python
! pip3 install -U google-cloud-aiplatform -q
3. Observe the error:
error: subprocess-exited-with-error
× Preparing metadata (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [794 lines of output]
<string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html
Partial import of sklearn during the build process.
<string>:116: DeprecationWarning:
`numpy.distutils` is deprecated since NumPy 1.23.0, as a result
of the deprecation of `distutils` itself. It will be removed for
Python >= 3.12. For older Python versions it will remain present.
It is recommended to use `setuptools < 60.0` for those Python versions.
For more details, see:
https://numpy.org/devdocs/reference/distutils_status_migration.html
<SNIP>
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
× Encountered error while generating package metadata.
╰─> See above for output.
note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
## Specifications
- Version: N/A
- Platform: N/A | priority | error subprocess exited with error deploying iris detection model using fastapi and vertex ai custom container serving expected behavior no errors when installing packages from the requirements file from the notebook actual behavior error subprocess exited with error × preparing metadata pyproject toml did not run successfully │ exit code ╰─ deprecationwarning pkg resources is deprecated as an api see partial import of sklearn during the build process deprecationwarning numpy distutils is deprecated since numpy as a result of the deprecation of distutils itself it will be removed for python for older python versions it will remain present it is recommended to use setuptools for those python versions for more details see steps to reproduce the problem used the vertex sample create the requirements file writefile requirements txt joblib numpy scikit learn google cloud storage run the install required in docker serving container install u r requirements txt q for local fastapi development and running install u uvicorn fastapi q vertex sdk for python install u google cloud aiplatform q observe the error error subprocess exited with error × preparing metadata pyproject toml did not run successfully │ exit code ╰─ deprecationwarning pkg resources is deprecated as an api see partial import of sklearn during the build process deprecationwarning numpy distutils is deprecated since numpy as a result of the deprecation of distutils itself it will be removed for python for older python versions it will remain present it is recommended to use setuptools for those python versions for more details see note this error originates from a subprocess and is likely not a problem with pip error metadata generation failed × encountered error while generating package metadata ╰─ see above for output note this is an issue with the package mentioned above not pip hint see above for details specifications version n a platform n a | 1 |
599,145 | 18,266,407,850 | IssuesEvent | 2021-10-04 08:58:37 | stevenwaterman/Lexoral | https://api.github.com/repos/stevenwaterman/Lexoral | opened | Allow running the functions locally for testing | high priority functions workflows tooling | Currently we can only run them online. This might not be possible with workflows, but because we're generating the workflow yaml with a DSL we could always use that DSL to call the other functions or something weird like that. | 1.0 | Allow running the functions locally for testing - Currently we can only run them online. This might not be possible with workflows, but because we're generating the workflow yaml with a DSL we could always use that DSL to call the other functions or something weird like that. | priority | allow running the functions locally for testing currently we can only run them online this might not be possible with workflows but because we re generating the workflow yaml with a dsl we could always use that dsl to call the other functions or something weird like that | 1 |
516,295 | 14,978,912,162 | IssuesEvent | 2021-01-28 11:29:35 | sButtons/sbuttons | https://api.github.com/repos/sButtons/sbuttons | opened | Fix base icon button class in Documentation Page | Priority: High bug good first issue help wanted up-for-grabs website | In documentation page under [Icon with no background](https://sbuttons.github.io/sbuttons/documentation.html#iconNoBg) change the class of the example button from `base-icon-btn` to `icon-btn`. This should be changed in the button and the code example | 1.0 | Fix base icon button class in Documentation Page - In documentation page under [Icon with no background](https://sbuttons.github.io/sbuttons/documentation.html#iconNoBg) change the class of the example button from `base-icon-btn` to `icon-btn`. This should be changed in the button and the code example | priority | fix base icon button class in documentation page in documentation page under change the class of the example button from base icon btn to icon btn this should be changed in the button and the code example | 1 |
134,959 | 12,643,085,083 | IssuesEvent | 2020-06-16 09:13:57 | durganp1/M2Challenge | https://api.github.com/repos/durganp1/M2Challenge | closed | Create Bio Section | documentation | Desired State:
* personal background
* Hobbies & Interest
* career goals | 1.0 | Create Bio Section - Desired State:
* personal background
* Hobbies & Interest
* career goals | non_priority | create bio section desired state personal background hobbies interest career goals | 0 |
630,779 | 20,117,778,393 | IssuesEvent | 2022-02-07 21:31:50 | arfc/saltproc | https://api.github.com/repos/arfc/saltproc | closed | Creat JSON schema for input | Type:Style Type:Feature Status:5-In Review Priority:2-Normal Difficulty:2-Challenging Comp:Input | Structure the input file based on a validation-ready json input file schema. This schema should be helped by description docstrings where necessary. However, a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears. (declare private or public variable where appropriate). | 1.0 | Creat JSON schema for input - Structure the input file based on a validation-ready json input file schema. This schema should be helped by description docstrings where necessary. However, a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears. (declare private or public variable where appropriate). | priority | creat json schema for input structure the input file based on a validation ready json input file schema this schema should be helped by description docstrings where necessary however a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears declare private or public variable where appropriate | 1 |
722,150 | 24,852,208,806 | IssuesEvent | 2022-10-26 21:10:51 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Make it obvious when a warning is elevated to an error | Priority:2 Type:DCR Area:ErrorHandling Functionality:Restore Style:PackageReference | Customers frequently get confused why restore fails when the message reported doesn't sound like an error, for example NU1603.
I think we should have something in the message that makes it obvious that NuGet considers it a warning, but it was elevated to an error by request. | 1.0 | Make it obvious when a warning is elevated to an error - Customers frequently get confused why restore fails when the message reported doesn't sound like an error, for example NU1603.
I think we should have something in the message that makes it obvious that NuGet considers it a warning, but it was elevated to an error by request. | priority | make it obvious when a warning is elevated to an error customers frequently get confused why restore fails when the message reported doesn t sound like an error for example i think we should have something in the message that makes it obvious that nuget considers it a warning but it was elevated to an error by request | 1 |
792,450 | 27,961,150,062 | IssuesEvent | 2023-03-24 15:44:58 | KDT3-Final-6/final-project-FE | https://api.github.com/repos/KDT3-Final-6/final-project-FE | reopened | Design: Home 마크업 | Status: Available Status: Review Needed Priority: High Type: Feature/Function | ## ✔️ 체크사항
- [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요.
## 💡 개발 목적
- 'Home' page 마크업
## 🌐 세부 내용
- [x] 헤더
- [x] 슬라이더
- [ ] 큐레이션
- [ ] 고객님을 위한 추천여행
- [ ] 그룹별 여행
- [ ] 테마별 인기 여행
- [ ] 여행지소개
- [ ] 실제 고객 여행 후기
- [ ] 이벤트
- [x] 푸터 | 1.0 | Design: Home 마크업 - ## ✔️ 체크사항
- [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요.
## 💡 개발 목적
- 'Home' page 마크업
## 🌐 세부 내용
- [x] 헤더
- [x] 슬라이더
- [ ] 큐레이션
- [ ] 고객님을 위한 추천여행
- [ ] 그룹별 여행
- [ ] 테마별 인기 여행
- [ ] 여행지소개
- [ ] 실제 고객 여행 후기
- [ ] 이벤트
- [x] 푸터 | priority | design home 마크업 ✔️ 체크사항 제목은 개발 페이지 개발 목적 으로 작성해 주세요 💡 개발 목적 home page 마크업 🌐 세부 내용 헤더 슬라이더 큐레이션 고객님을 위한 추천여행 그룹별 여행 테마별 인기 여행 여행지소개 실제 고객 여행 후기 이벤트 푸터 | 1 |
28,322 | 11,610,981,337 | IssuesEvent | 2020-02-26 05:01:48 | istio/istio | https://api.github.com/repos/istio/istio | closed | Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate | area/security lifecycle/needs-triage lifecycle/stale | Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate. The test cases need to cover file mount and workload SDS.
**Affected product area (please put an X in all that apply)**
[ ] Configuration Infrastructure
[ ] Docs
[ ] Installation
[ ] Networking
[ ] Performance and Scalability
[ ] Policies and Telemetry
[X] Security
[ ] Test and Release
[ ] User Experience
[ ] Developer Infrastructure
**Additional context**
https://github.com/istio/istio/issues/13439 | True | Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate - Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate. The test cases need to cover file mount and workload SDS.
**Affected product area (please put an X in all that apply)**
[ ] Configuration Infrastructure
[ ] Docs
[ ] Installation
[ ] Networking
[ ] Performance and Scalability
[ ] Policies and Telemetry
[X] Security
[ ] Test and Release
[ ] User Experience
[ ] Developer Infrastructure
**Additional context**
https://github.com/istio/istio/issues/13439 | non_priority | add integration test case for an internal client calling an internal service and the service is using a wrong certificate add integration test case for an internal client calling an internal service and the service is using a wrong certificate the test cases need to cover file mount and workload sds affected product area please put an x in all that apply configuration infrastructure docs installation networking performance and scalability policies and telemetry security test and release user experience developer infrastructure additional context | 0 |
84,528 | 3,669,167,218 | IssuesEvent | 2016-02-21 02:09:33 | FreezingMoon/AncientBeast | https://api.github.com/repos/FreezingMoon/AncientBeast | opened | plasma indicator vanishes | Coding Priority Visuals | After a player uses plasma points, the plasma indicator (capsule shaped) beneath the Dark Priest doesn't update to show the remaining points, the health indicator (rectangle shape) shows up instead. | 1.0 | plasma indicator vanishes - After a player uses plasma points, the plasma indicator (capsule shaped) beneath the Dark Priest doesn't update to show the remaining points, the health indicator (rectangle shape) shows up instead. | priority | plasma indicator vanishes after a player uses plasma points the plasma indicator capsule shaped beneath the dark priest doesn t update to show the remaining points the health indicator rectangle shape shows up instead | 1 |
611,553 | 18,958,304,732 | IssuesEvent | 2021-11-18 23:30:18 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Door lights do not blink red when access is denied | Type: Bug Priority: 1-Urgent Difficulty: 1-Easy | ## Description
Bug present (approximately) since the recent large sprite merge. | 1.0 | Door lights do not blink red when access is denied - ## Description
Bug present (approximately) since the recent large sprite merge. | priority | door lights do not blink red when access is denied description bug present approximately since the recent large sprite merge | 1 |
436,801 | 12,554,030,518 | IssuesEvent | 2020-06-07 00:22:51 | eclipse-ee4j/glassfish | https://api.github.com/repos/eclipse-ee4j/glassfish | closed | Custom log handler do not work in cluster mode. | Component: logging ERR: Assignee Priority: Minor Stale Type: Bug | 1) I've installed a cluster with 2 instances.
2) Then, I'he created jar-file with my simple custom log handler (just for tests):
package org.company.util;
import java.util.logging.Handler;
import java.util.logging.LogRecord;
import java.util.logging.SimpleFormatter;
import java.io.FileOutputStream;
import java.io.PrintWriter;
/**
*
* @author me
*/
public class LmtLogHandler extends Handler {
FileOutputStream fileOutputStream;
PrintWriter printWriter;
public LmtLogHandler() {
String filename = "MyCustomLogfile.txt";
try
{ fileOutputStream = new FileOutputStream(filename); printWriter = new PrintWriter(fileOutputStream); setFormatter(new SimpleFormatter()); }
catch (Exception e)
{ throw new Error("LogHandler init error!" + e); }
}
@Override
public void publish(LogRecord record) {
if (!isLoggable(record))
{ return; }
printWriter.println("LMT:" + getFormatter().format(record));
printWriter.flush();
}
@Override
public void flush()
{ printWriter.flush(); }
@Override
public void close() throws SecurityException
{ printWriter.close(); }
}
3) I've copied this jar in glassfish lib dirrectory. (/usr/local/glassfish/lib)
4) Then, I've added handler (from administration console) for cluster configuration.
5) Restart server.
There are no files "MyCustomLogfile.txt" in instances
(/usr/local/glassfish/nodeagents/myagent/instance1 or
/usr/local/glassfish/nodeagents/myagent/instance2 or
/usr/local/glassfish/nodeagents/myagent/agent/)
6) Then, I've added handler (from administration console) for server configuration.
7) Restart server.
There is a file MyCustomLogfile.txt in /usr/local/glassfish/domains/domain1/config/
Is is a bug, that custom log handler functionality does not work in cluster mode?
#### Environment
Operating System: Linux
Platform: Linux
#### Affected Versions
[9.1peur2] | 1.0 | Custom log handler do not work in cluster mode. - 1) I've installed a cluster with 2 instances.
2) Then, I'he created jar-file with my simple custom log handler (just for tests):
package org.company.util;
import java.util.logging.Handler;
import java.util.logging.LogRecord;
import java.util.logging.SimpleFormatter;
import java.io.FileOutputStream;
import java.io.PrintWriter;
/**
*
* @author me
*/
public class LmtLogHandler extends Handler {
FileOutputStream fileOutputStream;
PrintWriter printWriter;
public LmtLogHandler() {
String filename = "MyCustomLogfile.txt";
try
{ fileOutputStream = new FileOutputStream(filename); printWriter = new PrintWriter(fileOutputStream); setFormatter(new SimpleFormatter()); }
catch (Exception e)
{ throw new Error("LogHandler init error!" + e); }
}
@Override
public void publish(LogRecord record) {
if (!isLoggable(record))
{ return; }
printWriter.println("LMT:" + getFormatter().format(record));
printWriter.flush();
}
@Override
public void flush()
{ printWriter.flush(); }
@Override
public void close() throws SecurityException
{ printWriter.close(); }
}
3) I've copied this jar in glassfish lib dirrectory. (/usr/local/glassfish/lib)
4) Then, I've added handler (from administration console) for cluster configuration.
5) Restart server.
There are no files "MyCustomLogfile.txt" in instances
(/usr/local/glassfish/nodeagents/myagent/instance1 or
/usr/local/glassfish/nodeagents/myagent/instance2 or
/usr/local/glassfish/nodeagents/myagent/agent/)
6) Then, I've added handler (from administration console) for server configuration.
7) Restart server.
There is a file MyCustomLogfile.txt in /usr/local/glassfish/domains/domain1/config/
Is is a bug, that custom log handler functionality does not work in cluster mode?
#### Environment
Operating System: Linux
Platform: Linux
#### Affected Versions
[9.1peur2] | priority | custom log handler do not work in cluster mode i ve installed a cluster with instances then i he created jar file with my simple custom log handler just for tests package org company util import java util logging handler import java util logging logrecord import java util logging simpleformatter import java io fileoutputstream import java io printwriter author me public class lmtloghandler extends handler fileoutputstream fileoutputstream printwriter printwriter public lmtloghandler string filename mycustomlogfile txt try fileoutputstream new fileoutputstream filename printwriter new printwriter fileoutputstream setformatter new simpleformatter catch exception e throw new error loghandler init error e override public void publish logrecord record if isloggable record return printwriter println lmt getformatter format record printwriter flush override public void flush printwriter flush override public void close throws securityexception printwriter close i ve copied this jar in glassfish lib dirrectory usr local glassfish lib then i ve added handler from administration console for cluster configuration restart server there are no files mycustomlogfile txt in instances usr local glassfish nodeagents myagent or usr local glassfish nodeagents myagent or usr local glassfish nodeagents myagent agent then i ve added handler from administration console for server configuration restart server there is a file mycustomlogfile txt in usr local glassfish domains config is is a bug that custom log handler functionality does not work in cluster mode environment operating system linux platform linux affected versions | 1 |
201,823 | 15,814,370,052 | IssuesEvent | 2021-04-05 09:21:20 | AY2021S2-CS2103T-T12-3/tp | https://api.github.com/repos/AY2021S2-CS2103T-T12-3/tp | closed | [PE-D] Typo for "alias delete" error message | documentation severity.VeryLow | Steps to reproduce:
1. "alias add ls list"
2. "alias delete ls list"
The example given in the error message is a wrong command.
"alias delete ls" as given in the UG deleted the alias successfully.

<!--session: 1617429463103-be450b29-691c-46b2-91ae-62971187af48-->
-------------
Labels: `severity.VeryLow` `type.DocumentationBug`
original: laurenlhy/ped#11 | 1.0 | [PE-D] Typo for "alias delete" error message - Steps to reproduce:
1. "alias add ls list"
2. "alias delete ls list"
The example given in the error message is a wrong command.
"alias delete ls" as given in the UG deleted the alias successfully.

<!--session: 1617429463103-be450b29-691c-46b2-91ae-62971187af48-->
-------------
Labels: `severity.VeryLow` `type.DocumentationBug`
original: laurenlhy/ped#11 | non_priority | typo for alias delete error message steps to reproduce alias add ls list alias delete ls list the example given in the error message is a wrong command alias delete ls as given in the ug deleted the alias successfully labels severity verylow type documentationbug original laurenlhy ped | 0 |
367,849 | 10,862,087,178 | IssuesEvent | 2019-11-14 12:33:26 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.getpostman.com - site is not usable | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:71.0) Gecko/20100101 Firefox/71.0 -->
<!-- @reported_with: -->
**URL**: https://www.getpostman.com/
**Browser / Version**: Firefox 71.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Entire site hangs.
**Steps to Reproduce**:
go to https://www.getpostman.com/ try to click download. Nothing happens.
Scrolling down reveals only white.
Works fine in Chrome
[](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.getpostman.com - site is not usable - <!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:71.0) Gecko/20100101 Firefox/71.0 -->
<!-- @reported_with: -->
**URL**: https://www.getpostman.com/
**Browser / Version**: Firefox 71.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Entire site hangs.
**Steps to Reproduce**:
go to https://www.getpostman.com/ try to click download. Nothing happens.
Scrolling down reveals only white.
Works fine in Chrome
[](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description entire site hangs steps to reproduce go to try to click download nothing happens scrolling down reveals only white works fine in chrome browser configuration none from with ❤️ | 1 |
166,768 | 6,311,178,949 | IssuesEvent | 2017-07-23 17:10:10 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | Update auto-update config of video.js | High Priority in progress Library - Request to Add/Update | As you can see that I manually added the recent few versions, we should exclude `video.cjs.js` & `video.es.js` in most of the cases, so the auto-update config should be updated. | 1.0 | Update auto-update config of video.js - As you can see that I manually added the recent few versions, we should exclude `video.cjs.js` & `video.es.js` in most of the cases, so the auto-update config should be updated. | priority | update auto update config of video js as you can see that i manually added the recent few versions we should exclude video cjs js video es js in most of the cases so the auto update config should be updated | 1 |
370,684 | 25,919,094,213 | IssuesEvent | 2022-12-15 20:05:07 | transparencia-mg/work-stefanini | https://api.github.com/repos/transparencia-mg/work-stefanini | closed | Validação Casos de Testes Upload de Arquivos | bug documentation | @Andrelamor, gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário. Os mesmos podem ser consultados [aqui](https://transparencia-mg.github.io/work-stefanini/dev/testes/01_upload_de_arquivos_recursos_casos_de_teste/). | 1.0 | Validação Casos de Testes Upload de Arquivos - @Andrelamor, gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário. Os mesmos podem ser consultados [aqui](https://transparencia-mg.github.io/work-stefanini/dev/testes/01_upload_de_arquivos_recursos_casos_de_teste/). | non_priority | validação casos de testes upload de arquivos andrelamor gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário os mesmos podem ser consultados | 0 |
761,727 | 26,694,486,456 | IssuesEvent | 2023-01-27 09:09:35 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | CI:frdm_k64f: kernel.common.stack_protection test failure | bug priority: high area: Memory Protection platform: NXP Release Blocker | **Describe the bug**
kernel common stack protection test failure on frdm_k64f
initial found in v3.2.0-3560-g80f87b9480 and still exist in zephyr-v3.2.0-3758-g61f89b7b18a1
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=frdm_k64f
3. make
4. See error
**Expected behavior**
Test PASS
**Impact**
kernel stack protection
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v3.2.0-3758-g61f89b7b18a1 ***
Running TESTSUITE fatal_exception
===================================================================
START - test_fatal
test alt thread 1: generic CPU exception
E: ***** BUS FAULT *****
E: Instruction bus error
E: NXP MPU error, port 3
E: Mode: Supervisor, Instruction Address: 0x20001aa4
E: Type: Read, Master: 0, Regions: 0x8200
E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000014
E: r3/a4: 0x20001aa4 r12/ip: 0x00000000 r14/lr: 0x000007fb
E: xpsr: 0x40000000
E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
E: fpscr: 0x20000388
E: Faulting instruction address (r15/pc): 0x20001aa4
E: >>> ZEPHYR FATAL ERROR 27: Unknown error on CPU 0
E: Current thread: 0x20000020 (unknown)
Caught system error -- reason 27
Wrong crash type got 27 expected 35 or 20
```
**Environment (please complete the following information):**
- OS: (e.g. Linux)
- Toolchain (e.g Zephyr SDK, ...)
- Commit SHA or Version used: v3.2.0-3560-g80f87b9480
| 1.0 | CI:frdm_k64f: kernel.common.stack_protection test failure - **Describe the bug**
kernel common stack protection test failure on frdm_k64f
initial found in v3.2.0-3560-g80f87b9480 and still exist in zephyr-v3.2.0-3758-g61f89b7b18a1
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=frdm_k64f
3. make
4. See error
**Expected behavior**
Test PASS
**Impact**
kernel stack protection
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v3.2.0-3758-g61f89b7b18a1 ***
Running TESTSUITE fatal_exception
===================================================================
START - test_fatal
test alt thread 1: generic CPU exception
E: ***** BUS FAULT *****
E: Instruction bus error
E: NXP MPU error, port 3
E: Mode: Supervisor, Instruction Address: 0x20001aa4
E: Type: Read, Master: 0, Regions: 0x8200
E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000014
E: r3/a4: 0x20001aa4 r12/ip: 0x00000000 r14/lr: 0x000007fb
E: xpsr: 0x40000000
E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
E: fpscr: 0x20000388
E: Faulting instruction address (r15/pc): 0x20001aa4
E: >>> ZEPHYR FATAL ERROR 27: Unknown error on CPU 0
E: Current thread: 0x20000020 (unknown)
Caught system error -- reason 27
Wrong crash type got 27 expected 35 or 20
```
**Environment (please complete the following information):**
- OS: (e.g. Linux)
- Toolchain (e.g Zephyr SDK, ...)
- Commit SHA or Version used: v3.2.0-3560-g80f87b9480
| priority | ci frdm kernel common stack protection test failure describe the bug kernel common stack protection test failure on frdm initial found in and still exist in zephyr to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard frdm make see error expected behavior test pass impact kernel stack protection logs and console output booting zephyr os build zephyr running testsuite fatal exception start test fatal test alt thread generic cpu exception e bus fault e instruction bus error e nxp mpu error port e mode supervisor instruction address e type read master regions e e ip lr e xpsr e s s s s e s s s s e s s s s e s s s s e fpscr e faulting instruction address pc e gt gt gt zephyr fatal error unknown error on cpu e current thread unknown caught system error reason wrong crash type got expected or environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used | 1 |
390,507 | 11,544,463,056 | IssuesEvent | 2020-02-18 11:31:29 | Los-nonos/ZeepCommerce | https://api.github.com/repos/Los-nonos/ZeepCommerce | closed | [Website] Add template material kit for views | frontend priority: high status: DONE | Añadir todo el html y pasarlo a react component el archivo template.html que está en el kit en la homepage | 1.0 | [Website] Add template material kit for views - Añadir todo el html y pasarlo a react component el archivo template.html que está en el kit en la homepage | priority | add template material kit for views añadir todo el html y pasarlo a react component el archivo template html que está en el kit en la homepage | 1 |
50,604 | 7,613,240,119 | IssuesEvent | 2018-05-01 20:29:45 | sonarwhal/sonarwhal | https://api.github.com/repos/sonarwhal/sonarwhal | opened | Add documentation on how to integrate sonarwhal into the development flow | area:documentation difficulty:medium priority:high | There are a few ways to integrate sonarwhal during the development and I expect to have several "articles" ore recipes:
* [ ] Using the local connector while developing
* [ ] In the dev machine running a web server (such as `http-server` if static)
* [ ] In travis/Jenkins.
* This could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production. I have an idea on how to do this with Azure but I'm not sure if we need to go as deep as to show how to do it on each Cloud provider. | 1.0 | Add documentation on how to integrate sonarwhal into the development flow - There are a few ways to integrate sonarwhal during the development and I expect to have several "articles" ore recipes:
* [ ] Using the local connector while developing
* [ ] In the dev machine running a web server (such as `http-server` if static)
* [ ] In travis/Jenkins.
* This could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production. I have an idea on how to do this with Azure but I'm not sure if we need to go as deep as to show how to do it on each Cloud provider. | non_priority | add documentation on how to integrate sonarwhal into the development flow there are a few ways to integrate sonarwhal during the development and i expect to have several articles ore recipes using the local connector while developing in the dev machine running a web server such as http server if static in travis jenkins this could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production i have an idea on how to do this with azure but i m not sure if we need to go as deep as to show how to do it on each cloud provider | 0 |
236,411 | 26,010,598,893 | IssuesEvent | 2022-12-21 01:03:35 | Guillerbr/api-crud-adonisjs | https://api.github.com/repos/Guillerbr/api-crud-adonisjs | opened | CVE-2016-20018 (Medium) detected in knex-0.15.2.tgz | security vulnerability | ## CVE-2016-20018 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>knex-0.15.2.tgz</b></p></summary>
<p>A batteries-included SQL query & schema builder for Postgres, MySQL and SQLite3 and the Browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/knex/-/knex-0.15.2.tgz">https://registry.npmjs.org/knex/-/knex-0.15.2.tgz</a></p>
<p>Path to dependency file: /api-crud-adonisjs/package.json</p>
<p>Path to vulnerable library: /node_modules/knex/package.json</p>
<p>
Dependency Hierarchy:
- lucid-6.1.3.tgz (Root Library)
- :x: **knex-0.15.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Knex Knex.js through 2.3.0 has a limited SQL injection vulnerability that can be exploited to ignore the WHERE clause of a SQL query.
<p>Publish Date: 2022-12-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-20018>CVE-2016-20018</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-20018 (Medium) detected in knex-0.15.2.tgz - ## CVE-2016-20018 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>knex-0.15.2.tgz</b></p></summary>
<p>A batteries-included SQL query & schema builder for Postgres, MySQL and SQLite3 and the Browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/knex/-/knex-0.15.2.tgz">https://registry.npmjs.org/knex/-/knex-0.15.2.tgz</a></p>
<p>Path to dependency file: /api-crud-adonisjs/package.json</p>
<p>Path to vulnerable library: /node_modules/knex/package.json</p>
<p>
Dependency Hierarchy:
- lucid-6.1.3.tgz (Root Library)
- :x: **knex-0.15.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Knex Knex.js through 2.3.0 has a limited SQL injection vulnerability that can be exploited to ignore the WHERE clause of a SQL query.
<p>Publish Date: 2022-12-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-20018>CVE-2016-20018</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in knex tgz cve medium severity vulnerability vulnerable library knex tgz a batteries included sql query schema builder for postgres mysql and and the browser library home page a href path to dependency file api crud adonisjs package json path to vulnerable library node modules knex package json dependency hierarchy lucid tgz root library x knex tgz vulnerable library vulnerability details knex knex js through has a limited sql injection vulnerability that can be exploited to ignore the where clause of a sql query publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
725,222 | 24,954,587,224 | IssuesEvent | 2022-11-01 10:32:26 | thecyberworld/thecyberhub.org | https://api.github.com/repos/thecyberworld/thecyberhub.org | closed | [BUG] Sidebar dropwown not working | ✨ goal: improvement 🛠 goal: fix 🤩 status : Up for Grab 🟥 priority: critical hacktoberfest | ### Describe the bug
The sidebar dropdown not working

when we click on this nav-item nothing happens
### To Reproduce
Click on the Sidebar > learn. dropdown not working
### Expected Behavior
Like this: the drop should be displayed

_No response_
### Screenshot/ Video
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | 1.0 | [BUG] Sidebar dropwown not working - ### Describe the bug
The sidebar dropdown not working

when we click on this nav-item nothing happens
### To Reproduce
Click on the Sidebar > learn. dropdown not working
### Expected Behavior
Like this: the drop should be displayed

_No response_
### Screenshot/ Video
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | priority | sidebar dropwown not working describe the bug the sidebar dropdown not working when we click on this nav item nothing happens to reproduce click on the sidebar learn dropdown not working expected behavior like this the drop should be displayed no response screenshot video no response code of conduct i agree to follow this project s code of conduct | 1 |
175,628 | 21,314,082,367 | IssuesEvent | 2022-04-16 02:03:06 | infrahq/infra | https://api.github.com/repos/infrahq/infra | closed | Issue engines unique keys after initial connection | security authentication status/stale | The engine connection process looks like this:
1. A default API key is generated
2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint
3. The engine presents this API key on subsequent requests
We should migrate to issuing a new token to uniquely identify the engine after initial connection validation:
1. A default API key is generated
2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint
3. The registry creates a new key bound to the engine (with an `issued for` field), scoped to only permissions a registry needs
4. The engine uses this key on subsequent requests | True | Issue engines unique keys after initial connection - The engine connection process looks like this:
1. A default API key is generated
2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint
3. The engine presents this API key on subsequent requests
We should migrate to issuing a new token to uniquely identify the engine after initial connection validation:
1. A default API key is generated
2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint
3. The registry creates a new key bound to the engine (with an `issued for` field), scoped to only permissions a registry needs
4. The engine uses this key on subsequent requests | non_priority | issue engines unique keys after initial connection the engine connection process looks like this a default api key is generated the engine is created and presents this default api key to the registry to connect at the create destination endpoint the engine presents this api key on subsequent requests we should migrate to issuing a new token to uniquely identify the engine after initial connection validation a default api key is generated the engine is created and presents this default api key to the registry to connect at the create destination endpoint the registry creates a new key bound to the engine with an issued for field scoped to only permissions a registry needs the engine uses this key on subsequent requests | 0 |
254,721 | 8,087,309,321 | IssuesEvent | 2018-08-09 00:56:36 | docker/distribution | https://api.github.com/repos/docker/distribution | opened | Repo name validation is misused | bug priority/P1 | Regardless of your registry actual domain name, the Parse method https://github.com/docker/distribution/blob/master/reference/reference.go#L189 is always receiving:
`docker.io/foo/bar` or `docker.io/library/foo` (where `foo/bar` or `foo` is your image).
It seems to be because we are calling https://github.com/docker/distribution/blob/master/reference/normalize.go#L33 on `vars.name` from the context, which is apparently just the repo name (without the domain).
This makes the implementation deviate from the stated intent of the grammar.
If NameTotalLengthMax https://github.com/docker/distribution/blob/master/reference/reference.go#L37 was meant to limit the total size of the image name including the domain, then it must use the actual registry domain (+port).
If NameTotalLengthMax should be applied just on the path component, regardless of the domain, then we should not normalize `vars.name` before validating it (this is what the `WithName` method does, by the way).
A fix for this need to be carefully thought, as in the current situation the actual restriction is that on repo names at 237 characters (with single components) and 245 characters if they contain a slash - introducing proper domain name validation may render existing images invalid.
Let me know if I'm missing something here.
cc @caervs @dmcgowan @tiborvass | 1.0 | Repo name validation is misused - Regardless of your registry actual domain name, the Parse method https://github.com/docker/distribution/blob/master/reference/reference.go#L189 is always receiving:
`docker.io/foo/bar` or `docker.io/library/foo` (where `foo/bar` or `foo` is your image).
It seems to be because we are calling https://github.com/docker/distribution/blob/master/reference/normalize.go#L33 on `vars.name` from the context, which is apparently just the repo name (without the domain).
This makes the implementation deviate from the stated intent of the grammar.
If NameTotalLengthMax https://github.com/docker/distribution/blob/master/reference/reference.go#L37 was meant to limit the total size of the image name including the domain, then it must use the actual registry domain (+port).
If NameTotalLengthMax should be applied just on the path component, regardless of the domain, then we should not normalize `vars.name` before validating it (this is what the `WithName` method does, by the way).
A fix for this need to be carefully thought, as in the current situation the actual restriction is that on repo names at 237 characters (with single components) and 245 characters if they contain a slash - introducing proper domain name validation may render existing images invalid.
Let me know if I'm missing something here.
cc @caervs @dmcgowan @tiborvass | priority | repo name validation is misused regardless of your registry actual domain name the parse method is always receiving docker io foo bar or docker io library foo where foo bar or foo is your image it seems to be because we are calling on vars name from the context which is apparently just the repo name without the domain this makes the implementation deviate from the stated intent of the grammar if nametotallengthmax was meant to limit the total size of the image name including the domain then it must use the actual registry domain port if nametotallengthmax should be applied just on the path component regardless of the domain then we should not normalize vars name before validating it this is what the withname method does by the way a fix for this need to be carefully thought as in the current situation the actual restriction is that on repo names at characters with single components and characters if they contain a slash introducing proper domain name validation may render existing images invalid let me know if i m missing something here cc caervs dmcgowan tiborvass | 1 |
286,080 | 8,783,714,032 | IssuesEvent | 2018-12-20 07:17:52 | syspro-team1/EXTdateManageApp | https://api.github.com/repos/syspro-team1/EXTdateManageApp | opened | Notificationまわりのリファクタリング | enhancement low priority | ProductManagerActivityにごちゃごちゃ通知を設定してるのは
可読性が低い気がする.
クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち. | 1.0 | Notificationまわりのリファクタリング - ProductManagerActivityにごちゃごちゃ通知を設定してるのは
可読性が低い気がする.
クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち. | priority | notificationまわりのリファクタリング productmanageractivityにごちゃごちゃ通知を設定してるのは 可読性が低い気がする. クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち. | 1 |
549,605 | 16,095,718,617 | IssuesEvent | 2021-04-26 23:13:27 | lomcaitlin/cst438_s21_proj03_groupa | https://api.github.com/repos/lomcaitlin/cst438_s21_proj03_groupa | closed | Admin View Users | difficulty 2 priority 3 | As a admin I want to be able to view all users so I can keep track of every user
How will it be tested:
Unit tests for username verification and password verification
Instrumented test to verify extras
| 1.0 | Admin View Users - As a admin I want to be able to view all users so I can keep track of every user
How will it be tested:
Unit tests for username verification and password verification
Instrumented test to verify extras
| priority | admin view users as a admin i want to be able to view all users so i can keep track of every user how will it be tested unit tests for username verification and password verification instrumented test to verify extras | 1 |
323,475 | 9,855,484,347 | IssuesEvent | 2019-06-19 19:31:38 | cloud-custodian/cloud-custodian | https://api.github.com/repos/cloud-custodian/cloud-custodian | closed | tools/mailer - import enum error | area/tools-mailer kind/bug priority/P1 | Hi,
with a 0.8.44, you have add provider(enum) on utils.py, but i can't find this on deploy.py/get_archive() and requirement.txt
(with manual fix, this working again)
with this error, i can't execute update-lambda
could you update the integration of "enum" plz ?
thanks for your help | 1.0 | tools/mailer - import enum error - Hi,
with a 0.8.44, you have add provider(enum) on utils.py, but i can't find this on deploy.py/get_archive() and requirement.txt
(with manual fix, this working again)
with this error, i can't execute update-lambda
could you update the integration of "enum" plz ?
thanks for your help | priority | tools mailer import enum error hi with a you have add provider enum on utils py but i can t find this on deploy py get archive and requirement txt with manual fix this working again with this error i can t execute update lambda could you update the integration of enum plz thanks for your help | 1 |
540,814 | 15,817,518,112 | IssuesEvent | 2021-04-05 14:43:12 | AY2021S2-CS2103T-T12-4/tp | https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp | closed | [PE-D] Formatting of help window | priority.High severity.VeryLow type.Bug | No details provided.

In the help window, the button to copy link is truncated. I can also scroll horizontally even though the information already fits into one screen, and I need to scroll vertically to see all the commands even though they could easily fit in one screen.
<!--session: 1617429943449-e56bf182-40ef-47bf-9213-d664c9ca18a6-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: samuelfangjw/ped#5 | 1.0 | [PE-D] Formatting of help window - No details provided.

In the help window, the button to copy link is truncated. I can also scroll horizontally even though the information already fits into one screen, and I need to scroll vertically to see all the commands even though they could easily fit in one screen.
<!--session: 1617429943449-e56bf182-40ef-47bf-9213-d664c9ca18a6-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: samuelfangjw/ped#5 | priority | formatting of help window no details provided in the help window the button to copy link is truncated i can also scroll horizontally even though the information already fits into one screen and i need to scroll vertically to see all the commands even though they could easily fit in one screen labels severity verylow type functionalitybug original samuelfangjw ped | 1 |
2,400 | 2,607,900,667 | IssuesEvent | 2015-02-26 00:13:24 | chrsmithdemos/zen-coding | https://api.github.com/repos/chrsmithdemos/zen-coding | closed | reference error: use_tab is not defined | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. TYpe ul>li
2. Select 'Expand Abbreviation'
3.
What is the expected output? What do you see instead?
error message: reference error: use_tab is not defined
What version of the product are you using? On what operating system?
0.5.1 on Windows with Aptana
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `freeyl...@gmail.com` on 14 Jan 2010 at 1:40 | 1.0 | reference error: use_tab is not defined - ```
What steps will reproduce the problem?
1. TYpe ul>li
2. Select 'Expand Abbreviation'
3.
What is the expected output? What do you see instead?
error message: reference error: use_tab is not defined
What version of the product are you using? On what operating system?
0.5.1 on Windows with Aptana
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `freeyl...@gmail.com` on 14 Jan 2010 at 1:40 | non_priority | reference error use tab is not defined what steps will reproduce the problem type ul li select expand abbreviation what is the expected output what do you see instead error message reference error use tab is not defined what version of the product are you using on what operating system on windows with aptana please provide any additional information below original issue reported on code google com by freeyl gmail com on jan at | 0 |
518,301 | 15,026,679,509 | IssuesEvent | 2021-02-01 23:08:08 | tysonkaufmann/su-go | https://api.github.com/repos/tysonkaufmann/su-go | opened | [DEV] Add an endpoint to Get Route Information | Medium Priority task | **Related To**
- [View a route](https://github.com/tysonkaufmann/su-go/issues/57)
**Description**
Create an API endpoint `/api/getroute/{ROUTEID}` which will serve `GET` requests:
On Success, the API endpoint should return the information e.g
```
{
"status":"200",
"success":"true",
"data": {
"username":"",
"routeid":"",
"routetitle":"",
"routetype":"",
"routetime":"",
"routedistance":"",
"routedescription":"",
"photos":[],
"mapdata":{}
}
}
```
On Error, the API endpoint should return an error if unsuccessful e.g
```
{
"status":"200",
"success":"false",
"error":"The route does not exist"
}
```
**Development Steps**
- Create integration tests to fully test the API endpoint `/api/getroute/{ROUTEID}` (TDD)
- Route the endpoint in `backend/routes/Routes`
- Create unit tests for any code added in `backend/controllers/Routes` (TDD)
- Add code to the `/api/getroute/{ROUTEID}` to serve the API request
| 1.0 | [DEV] Add an endpoint to Get Route Information - **Related To**
- [View a route](https://github.com/tysonkaufmann/su-go/issues/57)
**Description**
Create an API endpoint `/api/getroute/{ROUTEID}` which will serve `GET` requests:
On Success, the API endpoint should return the information e.g
```
{
"status":"200",
"success":"true",
"data": {
"username":"",
"routeid":"",
"routetitle":"",
"routetype":"",
"routetime":"",
"routedistance":"",
"routedescription":"",
"photos":[],
"mapdata":{}
}
}
```
On Error, the API endpoint should return an error if unsuccessful e.g
```
{
"status":"200",
"success":"false",
"error":"The route does not exist"
}
```
**Development Steps**
- Create integration tests to fully test the API endpoint `/api/getroute/{ROUTEID}` (TDD)
- Route the endpoint in `backend/routes/Routes`
- Create unit tests for any code added in `backend/controllers/Routes` (TDD)
- Add code to the `/api/getroute/{ROUTEID}` to serve the API request
| priority | add an endpoint to get route information related to description create an api endpoint api getroute routeid which will serve get requests on success the api endpoint should return the information e g status success true data username routeid routetitle routetype routetime routedistance routedescription photos mapdata on error the api endpoint should return an error if unsuccessful e g status success false error the route does not exist development steps create integration tests to fully test the api endpoint api getroute routeid tdd route the endpoint in backend routes routes create unit tests for any code added in backend controllers routes tdd add code to the api getroute routeid to serve the api request | 1 |
61,196 | 7,447,608,107 | IssuesEvent | 2018-03-28 13:05:05 | AnSyn/ansyn | https://api.github.com/repos/AnSyn/ansyn | opened | Filters menu - limit section size | Design bug Priority: High Ready for: Development | at the moment each field is showing all of the available values, see "sensor name" in the attached image

As default we should present the top-10 values, and enable the user to "show more" if needed (Expand).
see attached image for inspiration:
<img width="255" alt="screen shot 2018-03-28 at 16 00 03" src="https://user-images.githubusercontent.com/4181572/38030615-c56df46a-32a1-11e8-97b5-47fe03e3baca.png">
| 1.0 | Filters menu - limit section size - at the moment each field is showing all of the available values, see "sensor name" in the attached image

As default we should present the top-10 values, and enable the user to "show more" if needed (Expand).
see attached image for inspiration:
<img width="255" alt="screen shot 2018-03-28 at 16 00 03" src="https://user-images.githubusercontent.com/4181572/38030615-c56df46a-32a1-11e8-97b5-47fe03e3baca.png">
| non_priority | filters menu limit section size at the moment each field is showing all of the available values see sensor name in the attached image as default we should present the top values and enable the user to show more if needed expand see attached image for inspiration img width alt screen shot at src | 0 |
1,150 | 2,870,524,710 | IssuesEvent | 2015-06-07 07:49:51 | pynac/pynac | https://api.github.com/repos/pynac/pynac | opened | cache expairseq::info computations | enhancement performance | The original behaviour is to compute `add`/`mul``::info` every time anew when it's called. | True | cache expairseq::info computations - The original behaviour is to compute `add`/`mul``::info` every time anew when it's called. | non_priority | cache expairseq info computations the original behaviour is to compute add mul info every time anew when it s called | 0 |
755,682 | 26,436,768,450 | IssuesEvent | 2023-01-15 13:42:23 | nrademacher/todo | https://api.github.com/repos/nrademacher/todo | closed | [NRA-117] Client: add pre-release disclaimer | High priority | Add disclaimer on sign-up pages that informs the user that the current version is a pre-release, account data may be lost, etc.
<sub>From [SyncLinear.com](https://synclinear.com) | [NRA-117](https://linear.app/nrademacher/issue/NRA-117/client-add-pre-release-disclaimer)</sub> | 1.0 | [NRA-117] Client: add pre-release disclaimer - Add disclaimer on sign-up pages that informs the user that the current version is a pre-release, account data may be lost, etc.
<sub>From [SyncLinear.com](https://synclinear.com) | [NRA-117](https://linear.app/nrademacher/issue/NRA-117/client-add-pre-release-disclaimer)</sub> | priority | client add pre release disclaimer add disclaimer on sign up pages that informs the user that the current version is a pre release account data may be lost etc from | 1 |
226,495 | 18,022,614,128 | IssuesEvent | 2021-09-16 21:38:19 | tpgxyz/test2 | https://api.github.com/repos/tpgxyz/test2 | opened | Second Open Bug (Bugzilla Bug 2) | bug tests | This issue was created automatically with bugzilla2github
# Bugzilla Bug 2
Date: 2015-05-07 16:21:59 +0200
From: Andriy Berestovskyy <<email@example.org>>
To: Andriy Berestovskyy <<email@example.org>>
CC: email2@example.org
Last updated: 2015-06-26 12:36:40 +0200
## Comment 2
Date: 2015-05-07 16:21:59 +0200
From: Andriy Berestovskyy <<email@example.org>>
Bug description goes here...
| 1.0 | Second Open Bug (Bugzilla Bug 2) - This issue was created automatically with bugzilla2github
# Bugzilla Bug 2
Date: 2015-05-07 16:21:59 +0200
From: Andriy Berestovskyy <<email@example.org>>
To: Andriy Berestovskyy <<email@example.org>>
CC: email2@example.org
Last updated: 2015-06-26 12:36:40 +0200
## Comment 2
Date: 2015-05-07 16:21:59 +0200
From: Andriy Berestovskyy <<email@example.org>>
Bug description goes here...
| non_priority | second open bug bugzilla bug this issue was created automatically with bugzilla bug date from andriy berestovskyy lt gt to andriy berestovskyy lt gt cc example org last updated comment date from andriy berestovskyy lt gt bug description goes here | 0 |
62,861 | 26,191,435,910 | IssuesEvent | 2023-01-03 09:25:44 | azure-deprecation/dashboard | https://api.github.com/repos/azure-deprecation/dashboard | closed | Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 | verified impact:upgrade-required area:feature cloud:public services:functions | Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022
**Deadline:** Dec 03, 2022
**Impacted Services:**
- Azure Functions
**More information:**
- https://azure.microsoft.com/updates/extended-support-for-microsoft-net-core-31-will-end-on-3-december-2022/
- https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version
### Notice
Here's the official report from Microsoft:
> **On 3 December 2022, extended support for Microsoft .NET Core 3.1 will end.** After that date, your applications that are hosted on Functions will continue to run and your applications will not be impacted. However, we'll no longer provide patches or customer service for .NET Core 3.1. Update your Functions applications to [runtime version 4.x, which uses .NET 6](https://docs.microsoft.com/azure/azure-functions/dotnet-isolated-process-guide#supported-versions).
>
> .NET 6 is the latest version with long-term support and provides these enhancements:
>
> - A unified set of base libraries and an SDK that make it easy to share code across any application type.
> - Simplified development with new C# 10 features and minimal APIs.
> - Hot reload that allows you to make code changes without explicit recompiling.
### Timeline
| Phase | Date | Description |
|:------|------|-------------|
|Announcement|Mar 17, 2022|Deprecation was announced|
|Deprecation|Dec 03, 2022|Function Apps will continue to run but no longer receive patches nor support|
### Impact
Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 and upgrade to .NET 6 with Azure Functions v4 is required.
### Required Action
A migration guide to the newer Azure Functions runtime version is available [here](https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version).
Here's the official report from Microsoft:
> To avoid potential service disruptions or security vulnerabilities, [update your Functions applications](https://aka.ms/functions-view-and-update-runtime-version) to runtime version 4.x, which uses .NET 6, before 3 December 2022.
### Contact
You can get in touch through the following options:
- Get answers from Microsoft Q&A ([link](https://aka.ms/functions-c-sharp-qa-link)).
- Contact Azure support ([link](https://portal.azure.com/#blade/Microsoft_Azure_Support/HelpAndSupportBlade/overview)).
| 1.0 | Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 - Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022
**Deadline:** Dec 03, 2022
**Impacted Services:**
- Azure Functions
**More information:**
- https://azure.microsoft.com/updates/extended-support-for-microsoft-net-core-31-will-end-on-3-december-2022/
- https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version
### Notice
Here's the official report from Microsoft:
> **On 3 December 2022, extended support for Microsoft .NET Core 3.1 will end.** After that date, your applications that are hosted on Functions will continue to run and your applications will not be impacted. However, we'll no longer provide patches or customer service for .NET Core 3.1. Update your Functions applications to [runtime version 4.x, which uses .NET 6](https://docs.microsoft.com/azure/azure-functions/dotnet-isolated-process-guide#supported-versions).
>
> .NET 6 is the latest version with long-term support and provides these enhancements:
>
> - A unified set of base libraries and an SDK that make it easy to share code across any application type.
> - Simplified development with new C# 10 features and minimal APIs.
> - Hot reload that allows you to make code changes without explicit recompiling.
### Timeline
| Phase | Date | Description |
|:------|------|-------------|
|Announcement|Mar 17, 2022|Deprecation was announced|
|Deprecation|Dec 03, 2022|Function Apps will continue to run but no longer receive patches nor support|
### Impact
Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 and upgrade to .NET 6 with Azure Functions v4 is required.
### Required Action
A migration guide to the newer Azure Functions runtime version is available [here](https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version).
Here's the official report from Microsoft:
> To avoid potential service disruptions or security vulnerabilities, [update your Functions applications](https://aka.ms/functions-view-and-update-runtime-version) to runtime version 4.x, which uses .NET 6, before 3 December 2022.
### Contact
You can get in touch through the following options:
- Get answers from Microsoft Q&A ([link](https://aka.ms/functions-c-sharp-qa-link)).
- Contact Azure support ([link](https://portal.azure.com/#blade/Microsoft_Azure_Support/HelpAndSupportBlade/overview)).
| non_priority | extended support for net core in azure functions is retiring on december extended support for net core in azure functions is retiring on december deadline dec impacted services azure functions more information notice here s the official report from microsoft on december extended support for microsoft net core will end after that date your applications that are hosted on functions will continue to run and your applications will not be impacted however we ll no longer provide patches or customer service for net core update your functions applications to net is the latest version with long term support and provides these enhancements a unified set of base libraries and an sdk that make it easy to share code across any application type simplified development with new c features and minimal apis hot reload that allows you to make code changes without explicit recompiling timeline phase date description announcement mar deprecation was announced deprecation dec function apps will continue to run but no longer receive patches nor support impact extended support for net core in azure functions is retiring on december and upgrade to net with azure functions is required required action a migration guide to the newer azure functions runtime version is available here s the official report from microsoft to avoid potential service disruptions or security vulnerabilities to runtime version x which uses net before december contact you can get in touch through the following options get answers from microsoft q a contact azure support | 0 |
67,005 | 16,769,019,638 | IssuesEvent | 2021-06-14 12:43:43 | rust-lang/docs.rs | https://api.github.com/repos/rust-lang/docs.rs | reopened | "Failed to parse manifest" on bootloader-0.10.0-alpha-03 | A-builds | Not sure what's going wrong here, `cargo doc` works locally. The crate doesn't show up in the web UI at all.
```
Feb 21 12:36:47 docsrs cratesfyi[28601]: 2021/02/21 12:36:47 [ERROR] docs_rs::build_queue: Failed to build package bootloader-0.10.0-alpha-03 from queue: failed to parse maniest
Feb 21 12:36:47 docsrs cratesfyi[28601]: Backtrace: 0: failure::backtrace::internal::InternalBacktrace::new
Feb 21 12:36:47 docsrs cratesfyi[28601]: 1: <failure::backtrace::Backtrace as core::default::Default>::default
Feb 21 12:36:47 docsrs cratesfyi[28601]: 2: rustwide::build::BuildBuilder::run
Feb 21 12:36:47 docsrs cratesfyi[28601]: 3: docs_rs::docbuilder::rustwide_builder::RustwideBuilder::build_package
Feb 21 12:36:47 docsrs cratesfyi[28601]: 4: docs_rs::build_queue::BuildQueue::process_next_crate
Feb 21 12:36:47 docsrs cratesfyi[28601]: 5: <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: 6: docs_rs::utils::queue_builder::queue_builder
Feb 21 12:36:47 docsrs cratesfyi[28601]: 7: std::sys_common::backtrace::__rust_begin_short_backtrace
Feb 21 12:36:47 docsrs cratesfyi[28601]: 8: core::ops::function::FnOnce::call_once{{vtable.shim}}
Feb 21 12:36:47 docsrs cratesfyi[28601]: 9: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076
Feb 21 12:36:47 docsrs cratesfyi[28601]: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076
Feb 21 12:36:47 docsrs cratesfyi[28601]: std::sys::unix::thread::Thread::new::thread_start
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/libstd/sys/unix/thread.rs:87
Feb 21 12:36:47 docsrs cratesfyi[28601]: 10: start_thread
Feb 21 12:36:47 docsrs cratesfyi[28601]: 11: __clone | 1.0 | "Failed to parse manifest" on bootloader-0.10.0-alpha-03 - Not sure what's going wrong here, `cargo doc` works locally. The crate doesn't show up in the web UI at all.
```
Feb 21 12:36:47 docsrs cratesfyi[28601]: 2021/02/21 12:36:47 [ERROR] docs_rs::build_queue: Failed to build package bootloader-0.10.0-alpha-03 from queue: failed to parse maniest
Feb 21 12:36:47 docsrs cratesfyi[28601]: Backtrace: 0: failure::backtrace::internal::InternalBacktrace::new
Feb 21 12:36:47 docsrs cratesfyi[28601]: 1: <failure::backtrace::Backtrace as core::default::Default>::default
Feb 21 12:36:47 docsrs cratesfyi[28601]: 2: rustwide::build::BuildBuilder::run
Feb 21 12:36:47 docsrs cratesfyi[28601]: 3: docs_rs::docbuilder::rustwide_builder::RustwideBuilder::build_package
Feb 21 12:36:47 docsrs cratesfyi[28601]: 4: docs_rs::build_queue::BuildQueue::process_next_crate
Feb 21 12:36:47 docsrs cratesfyi[28601]: 5: <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: 6: docs_rs::utils::queue_builder::queue_builder
Feb 21 12:36:47 docsrs cratesfyi[28601]: 7: std::sys_common::backtrace::__rust_begin_short_backtrace
Feb 21 12:36:47 docsrs cratesfyi[28601]: 8: core::ops::function::FnOnce::call_once{{vtable.shim}}
Feb 21 12:36:47 docsrs cratesfyi[28601]: 9: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076
Feb 21 12:36:47 docsrs cratesfyi[28601]: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076
Feb 21 12:36:47 docsrs cratesfyi[28601]: std::sys::unix::thread::Thread::new::thread_start
Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/libstd/sys/unix/thread.rs:87
Feb 21 12:36:47 docsrs cratesfyi[28601]: 10: start_thread
Feb 21 12:36:47 docsrs cratesfyi[28601]: 11: __clone | non_priority | failed to parse manifest on bootloader alpha not sure what s going wrong here cargo doc works locally the crate doesn t show up in the web ui at all feb docsrs cratesfyi docs rs build queue failed to build package bootloader alpha from queue failed to parse maniest feb docsrs cratesfyi backtrace failure backtrace internal internalbacktrace new feb docsrs cratesfyi default feb docsrs cratesfyi rustwide build buildbuilder run feb docsrs cratesfyi docs rs docbuilder rustwide builder rustwidebuilder build package feb docsrs cratesfyi docs rs build queue buildqueue process next crate feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi docs rs utils queue builder queue builder feb docsrs cratesfyi std sys common backtrace rust begin short backtrace feb docsrs cratesfyi core ops function fnonce call once vtable shim feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi at rustc src liballoc boxed rs feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi at rustc src liballoc boxed rs feb docsrs cratesfyi std sys unix thread thread new thread start feb docsrs cratesfyi at rustc src libstd sys unix thread rs feb docsrs cratesfyi start thread feb docsrs cratesfyi clone | 0 |
522,797 | 15,168,235,481 | IssuesEvent | 2021-02-12 19:04:18 | FTBTeam/FTB-Quests | https://api.github.com/repos/FTBTeam/FTB-Quests | closed | Completely Broken Quest GUI other than one corner | bug high priority | ftb lib ver:5.4.1.77
ftb quest ver:1.4.1.115
screen shots

: | 1.0 | Completely Broken Quest GUI other than one corner - ftb lib ver:5.4.1.77
ftb quest ver:1.4.1.115
screen shots

: | priority | completely broken quest gui other than one corner ftb lib ver ftb quest ver screen shots | 1 |
270,475 | 23,511,581,166 | IssuesEvent | 2022-08-18 17:04:34 | PowerShell/Microsoft.PowerShell.Archive | https://api.github.com/repos/PowerShell/Microsoft.PowerShell.Archive | closed | Expand-Archive Tests | P0 Size: Large Area - Archive Format Test Scenario Area - Symlinks Area - Error Handling | This is a mega-issue that divides the tests to be added to Expand-Archive. This list is not complete, and more tests will be added over time.
[x] indicates the test has been added to `Expand-Archive.Tests.ps1`.
[ ] indices the test has not yet been added.
All tests will be written in Pester 5. Old tests will be updated.
# Parameter set validation tests
- [x] Validate errors with NULL & EMPTY values for Path, LiteralPath, and DestinationPath
- [x] Throws when invalid path non-existing path is supplied for Path or LiteralPath parameters
- [x] Throws when invalid path non-filesystem path is supplied for Path or LiteralPath parameters
- [x] Throws an error when multiple paths are supplied as input to Path parameter
- [x] Throws an error when multiple paths are supplied as input to LiteralPath parameter
- [x] Throws an error when Path and DestinationPath are the same and -WriteMode Overwrite is specified
- [x] Throws an error when LiteralPath and DestinationPath are the same and WriteMode -Overwrite is specified
- [x] Throws an error when an invalid path is supplied to DestinationPath
# `-DestinationPath` parameter tests
- [x] Throws an error when DestinationPath is an existing file
- [x] Does not throw an error when a directory in the archive has the same destination path as an existing directory
- [x] Writes a non-terminating error when a file in the archive has a destination path that already exists
- [x] Writes a non-terminating error when a file in the archive has a destination path that is an existing directory containing at least 1 item and -WriteMode Overwrite is specified
- [x] Writes a non-terminating error when a file in the archive has a destination path that is the working directory and -WriteMode Overwrite is specified
- [x] Overwrites a file when it is DestinationPath and -WriteMode Overwrite is specified
- [x] Overwrites a file whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified
- [x] Overwrites a directory whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified
- [x] Overwrites a file whose path is the same as the destination path of a directory in the archive when -WriteMode Overwrite is specified
# Relative paths
- [x] Expands an archive when -Path is a relative path
- [x] Expands an archive when -LiteralPath is a relative path
- [x] Expands an archive when -DestinationPath is a relative path
# Special and Wildcard Characters
- [x] Expands an archive when -Path contains wildcard character and resolves to 1 path
- [x] Throws a terminating error when archive when -Path contains wildcard character and resolves to multiple paths
- [x] Expands an archive when -LiteralPath contains [ but no matching ]
- [x] Expands an archive when -DestinationPath contains [ but no matching ]
# Basic functional tests
- [x] Expands an archive when a non-existent directory is specified as -DestinationPath
- [x] Expands an archive when DestinationPath is an existing directory
- [x] Expands an archive to the working directory when it is specified as -DestinationPath
- [x] Expands an archive to a directory with that archive's name when -DestinationPath is not specified
- [x] Expands an archive containing multiple files, non-empty directories, and empty directories
- [x] Expands an archive containing a file whose LastWriteTime is in the past
- [x] Expands an archive containing a directory whose LastWriteTime is in the past
- [x] Throws an error when expanding an archive whose name does not have an extension and -DestinationPath is not specified
# `-PassThru` parameter
- [x] Returns a System.IO.DirectoryInfo object when PassThru is specified
- [x] Does not return an object when PassThru is not specified
- [x] Does not return an object when PassThru is false
# File permissions, locked files, hidden files, symbolic links, etc.
- [x] Expands a read-only archive
- [x] Expands an archive in-use
- [x] Expands an archive containing an entry with non-latin characters
# Pipelining
- [x] Expands an archive when -Path is passed by pipeline
- [x] Expands an archive when -Path is passed by pipeline by name
- [x] Throws an error when multiple paths are passed by pipeline
# Large file tests
- [x] Expands an archive whose size is > 4GB
- [x] Expands an entry whose size is > 4GB
# Archive formats and `-Format` parameter
- [x] Throws an error when an invalid value is supplied to -Format
## Module
- [x] Validate module can be imported when current language is not en-US (1247)
| 1.0 | Expand-Archive Tests - This is a mega-issue that divides the tests to be added to Expand-Archive. This list is not complete, and more tests will be added over time.
[x] indicates the test has been added to `Expand-Archive.Tests.ps1`.
[ ] indices the test has not yet been added.
All tests will be written in Pester 5. Old tests will be updated.
# Parameter set validation tests
- [x] Validate errors with NULL & EMPTY values for Path, LiteralPath, and DestinationPath
- [x] Throws when invalid path non-existing path is supplied for Path or LiteralPath parameters
- [x] Throws when invalid path non-filesystem path is supplied for Path or LiteralPath parameters
- [x] Throws an error when multiple paths are supplied as input to Path parameter
- [x] Throws an error when multiple paths are supplied as input to LiteralPath parameter
- [x] Throws an error when Path and DestinationPath are the same and -WriteMode Overwrite is specified
- [x] Throws an error when LiteralPath and DestinationPath are the same and WriteMode -Overwrite is specified
- [x] Throws an error when an invalid path is supplied to DestinationPath
# `-DestinationPath` parameter tests
- [x] Throws an error when DestinationPath is an existing file
- [x] Does not throw an error when a directory in the archive has the same destination path as an existing directory
- [x] Writes a non-terminating error when a file in the archive has a destination path that already exists
- [x] Writes a non-terminating error when a file in the archive has a destination path that is an existing directory containing at least 1 item and -WriteMode Overwrite is specified
- [x] Writes a non-terminating error when a file in the archive has a destination path that is the working directory and -WriteMode Overwrite is specified
- [x] Overwrites a file when it is DestinationPath and -WriteMode Overwrite is specified
- [x] Overwrites a file whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified
- [x] Overwrites a directory whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified
- [x] Overwrites a file whose path is the same as the destination path of a directory in the archive when -WriteMode Overwrite is specified
# Relative paths
- [x] Expands an archive when -Path is a relative path
- [x] Expands an archive when -LiteralPath is a relative path
- [x] Expands an archive when -DestinationPath is a relative path
# Special and Wildcard Characters
- [x] Expands an archive when -Path contains wildcard character and resolves to 1 path
- [x] Throws a terminating error when archive when -Path contains wildcard character and resolves to multiple paths
- [x] Expands an archive when -LiteralPath contains [ but no matching ]
- [x] Expands an archive when -DestinationPath contains [ but no matching ]
# Basic functional tests
- [x] Expands an archive when a non-existent directory is specified as -DestinationPath
- [x] Expands an archive when DestinationPath is an existing directory
- [x] Expands an archive to the working directory when it is specified as -DestinationPath
- [x] Expands an archive to a directory with that archive's name when -DestinationPath is not specified
- [x] Expands an archive containing multiple files, non-empty directories, and empty directories
- [x] Expands an archive containing a file whose LastWriteTime is in the past
- [x] Expands an archive containing a directory whose LastWriteTime is in the past
- [x] Throws an error when expanding an archive whose name does not have an extension and -DestinationPath is not specified
# `-PassThru` parameter
- [x] Returns a System.IO.DirectoryInfo object when PassThru is specified
- [x] Does not return an object when PassThru is not specified
- [x] Does not return an object when PassThru is false
# File permissions, locked files, hidden files, symbolic links, etc.
- [x] Expands a read-only archive
- [x] Expands an archive in-use
- [x] Expands an archive containing an entry with non-latin characters
# Pipelining
- [x] Expands an archive when -Path is passed by pipeline
- [x] Expands an archive when -Path is passed by pipeline by name
- [x] Throws an error when multiple paths are passed by pipeline
# Large file tests
- [x] Expands an archive whose size is > 4GB
- [x] Expands an entry whose size is > 4GB
# Archive formats and `-Format` parameter
- [x] Throws an error when an invalid value is supplied to -Format
## Module
- [x] Validate module can be imported when current language is not en-US (1247)
| non_priority | expand archive tests this is a mega issue that divides the tests to be added to expand archive this list is not complete and more tests will be added over time indicates the test has been added to expand archive tests indices the test has not yet been added all tests will be written in pester old tests will be updated parameter set validation tests validate errors with null empty values for path literalpath and destinationpath throws when invalid path non existing path is supplied for path or literalpath parameters throws when invalid path non filesystem path is supplied for path or literalpath parameters throws an error when multiple paths are supplied as input to path parameter throws an error when multiple paths are supplied as input to literalpath parameter throws an error when path and destinationpath are the same and writemode overwrite is specified throws an error when literalpath and destinationpath are the same and writemode overwrite is specified throws an error when an invalid path is supplied to destinationpath destinationpath parameter tests throws an error when destinationpath is an existing file does not throw an error when a directory in the archive has the same destination path as an existing directory writes a non terminating error when a file in the archive has a destination path that already exists writes a non terminating error when a file in the archive has a destination path that is an existing directory containing at least item and writemode overwrite is specified writes a non terminating error when a file in the archive has a destination path that is the working directory and writemode overwrite is specified overwrites a file when it is destinationpath and writemode overwrite is specified overwrites a file whose path is the same as the destination path of a file in the archive when writemode overwrite is specified overwrites a directory whose path is the same as the destination path of a file in the archive when writemode overwrite is specified overwrites a file whose path is the same as the destination path of a directory in the archive when writemode overwrite is specified relative paths expands an archive when path is a relative path expands an archive when literalpath is a relative path expands an archive when destinationpath is a relative path special and wildcard characters expands an archive when path contains wildcard character and resolves to path throws a terminating error when archive when path contains wildcard character and resolves to multiple paths expands an archive when literalpath contains expands an archive when destinationpath contains basic functional tests expands an archive when a non existent directory is specified as destinationpath expands an archive when destinationpath is an existing directory expands an archive to the working directory when it is specified as destinationpath expands an archive to a directory with that archive s name when destinationpath is not specified expands an archive containing multiple files non empty directories and empty directories expands an archive containing a file whose lastwritetime is in the past expands an archive containing a directory whose lastwritetime is in the past throws an error when expanding an archive whose name does not have an extension and destinationpath is not specified passthru parameter returns a system io directoryinfo object when passthru is specified does not return an object when passthru is not specified does not return an object when passthru is false file permissions locked files hidden files symbolic links etc expands a read only archive expands an archive in use expands an archive containing an entry with non latin characters pipelining expands an archive when path is passed by pipeline expands an archive when path is passed by pipeline by name throws an error when multiple paths are passed by pipeline large file tests expands an archive whose size is expands an entry whose size is archive formats and format parameter throws an error when an invalid value is supplied to format module validate module can be imported when current language is not en us | 0 |
106,028 | 9,106,411,740 | IssuesEvent | 2019-02-20 23:47:02 | unfoldingWord-dev/translationCore | https://api.github.com/repos/unfoldingWord-dev/translationCore | closed | Don't show any alignment boxes when the verse is blank and there is not Greek text | Kind/Enhancement QA/Retest | 0.8.1-alpha.26-61679a7
 | 1.0 | Don't show any alignment boxes when the verse is blank and there is not Greek text - 0.8.1-alpha.26-61679a7
 | non_priority | don t show any alignment boxes when the verse is blank and there is not greek text alpha | 0 |
86,767 | 10,817,538,252 | IssuesEvent | 2019-11-08 09:57:46 | Marc-AntoineA/A2P | https://api.github.com/repos/Marc-AntoineA/A2P | opened | Updating supervisor welcome view | design supervisor-frontend | The supervisor welcome view was never done and is still in *todo* state. This view should explain the features of the website and give an overview of the opened processes, etc. | 1.0 | Updating supervisor welcome view - The supervisor welcome view was never done and is still in *todo* state. This view should explain the features of the website and give an overview of the opened processes, etc. | non_priority | updating supervisor welcome view the supervisor welcome view was never done and is still in todo state this view should explain the features of the website and give an overview of the opened processes etc | 0 |
101,325 | 8,785,333,096 | IssuesEvent | 2018-12-20 12:40:10 | linkedpipes/dcat-ap-viewer | https://api.github.com/repos/linkedpipes/dcat-ap-viewer | closed | Links in navigation do not work in Firefox (NK-11) | bug test | See [here](https://nkod.nakit.opendata.cz/datov%C3%A9-sady) in FF. Related issue was already resolved in data.gov.cz. | 1.0 | Links in navigation do not work in Firefox (NK-11) - See [here](https://nkod.nakit.opendata.cz/datov%C3%A9-sady) in FF. Related issue was already resolved in data.gov.cz. | non_priority | links in navigation do not work in firefox nk see in ff related issue was already resolved in data gov cz | 0 |
160,942 | 6,105,811,531 | IssuesEvent | 2017-06-21 01:16:47 | ProjectSidewalk/SidewalkWebpage | https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage | closed | After jump, user is always 180 degree turn away from the right direction | potential-intern-assignment Priority: Low pull-request-submitted ui-update | When the user is jumped into a neighborhood, he/she is always facing the wrong way and, hence, has to turn all the way around (u-turn).
Reported by 2 CMSC434 undergrads.

| 1.0 | After jump, user is always 180 degree turn away from the right direction - When the user is jumped into a neighborhood, he/she is always facing the wrong way and, hence, has to turn all the way around (u-turn).
Reported by 2 CMSC434 undergrads.

| priority | after jump user is always degree turn away from the right direction when the user is jumped into a neighborhood he she is always facing the wrong way and hence has to turn all the way around u turn reported by undergrads | 1 |
38,845 | 2,850,497,058 | IssuesEvent | 2015-05-31 16:39:30 | damonkohler/android-scripting | https://api.github.com/repos/damonkohler/android-scripting | closed | WebView resource paths should be relative to script's location, not SL4A's | auto-migrated Priority-Medium Type-Enhancement | ```
When displaying a WebView, relative paths don't work as you'd expect if the
script is in a subdirectory. e.g.:
1. Create a folder in SL4A named test.
2. In that folder, place two files index.html and main.js.
3. In index.html, insert the line: <script src="main.js"></script>. (You might
need head and body too?)
4. In main.is, write: alert("hello")
5. Open index.html. No alert is displayed.
6. Change the include line in index.html to: <script
src="test/main.js"></script>
7. Open index.html again; the alert now appears.
This means that relative paths are based at /mnt/sdcard/sl4a/scripts, rather
than /mnt/sdcard/sl4a/scripts/test. This seems odd to me - a script's working
directory should be the one containing it, so that it can access its resources
without having to know about the directory structure above it. Otherwise the
name of the folder has to be hardcoded into every resource path, which means it
can't be moved, renamed, or copied as a base for another script without a lot
of search-and-replace.
```
Original issue reported on code.google.com by `hyperhac...@gmail.com` on 14 Jul 2012 at 8:20 | 1.0 | WebView resource paths should be relative to script's location, not SL4A's - ```
When displaying a WebView, relative paths don't work as you'd expect if the
script is in a subdirectory. e.g.:
1. Create a folder in SL4A named test.
2. In that folder, place two files index.html and main.js.
3. In index.html, insert the line: <script src="main.js"></script>. (You might
need head and body too?)
4. In main.is, write: alert("hello")
5. Open index.html. No alert is displayed.
6. Change the include line in index.html to: <script
src="test/main.js"></script>
7. Open index.html again; the alert now appears.
This means that relative paths are based at /mnt/sdcard/sl4a/scripts, rather
than /mnt/sdcard/sl4a/scripts/test. This seems odd to me - a script's working
directory should be the one containing it, so that it can access its resources
without having to know about the directory structure above it. Otherwise the
name of the folder has to be hardcoded into every resource path, which means it
can't be moved, renamed, or copied as a base for another script without a lot
of search-and-replace.
```
Original issue reported on code.google.com by `hyperhac...@gmail.com` on 14 Jul 2012 at 8:20 | priority | webview resource paths should be relative to script s location not s when displaying a webview relative paths don t work as you d expect if the script is in a subdirectory e g create a folder in named test in that folder place two files index html and main js in index html insert the line you might need head and body too in main is write alert hello open index html no alert is displayed change the include line in index html to script src test main js open index html again the alert now appears this means that relative paths are based at mnt sdcard scripts rather than mnt sdcard scripts test this seems odd to me a script s working directory should be the one containing it so that it can access its resources without having to know about the directory structure above it otherwise the name of the folder has to be hardcoded into every resource path which means it can t be moved renamed or copied as a base for another script without a lot of search and replace original issue reported on code google com by hyperhac gmail com on jul at | 1 |
358,876 | 10,651,238,205 | IssuesEvent | 2019-10-17 09:56:48 | ntop/ntopng | https://api.github.com/repos/ntop/ntopng | opened | NetworkInterface::periodicStatsUpdate() requires cleanup | enhancement priority ticket | The above method is not just updating statistics as the name says. Instead it also purges entries. Said that its name should change as misleading, on flow-based interfaces the call frequency to this method depends on the flow timeout. While this is acceptable for updates, it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency.
A minimal solution is to call idle at a different frequency. The good solution is to avoid walking the whole hash searching for idle hosts, but rather to accumulate the idle hosts on a vector and thus purge them up without walking. | 1.0 | NetworkInterface::periodicStatsUpdate() requires cleanup - The above method is not just updating statistics as the name says. Instead it also purges entries. Said that its name should change as misleading, on flow-based interfaces the call frequency to this method depends on the flow timeout. While this is acceptable for updates, it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency.
A minimal solution is to call idle at a different frequency. The good solution is to avoid walking the whole hash searching for idle hosts, but rather to accumulate the idle hosts on a vector and thus purge them up without walking. | priority | networkinterface periodicstatsupdate requires cleanup the above method is not just updating statistics as the name says instead it also purges entries said that its name should change as misleading on flow based interfaces the call frequency to this method depends on the flow timeout while this is acceptable for updates it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency a minimal solution is to call idle at a different frequency the good solution is to avoid walking the whole hash searching for idle hosts but rather to accumulate the idle hosts on a vector and thus purge them up without walking | 1 |
421,073 | 12,248,493,906 | IssuesEvent | 2020-05-05 17:35:48 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | cancel call fails to kick CQ under epoll1 | kind/bug lang/c++ priority/P2 | communication for internal issue b/144444259
User reported all threads blocked on C++ CQ dtor waiting for the pending rpc's to finish. The pending rpcs were cancelled previously. The user is using epoll1. | 1.0 | cancel call fails to kick CQ under epoll1 - communication for internal issue b/144444259
User reported all threads blocked on C++ CQ dtor waiting for the pending rpc's to finish. The pending rpcs were cancelled previously. The user is using epoll1. | priority | cancel call fails to kick cq under communication for internal issue b user reported all threads blocked on c cq dtor waiting for the pending rpc s to finish the pending rpcs were cancelled previously the user is using | 1 |
221,227 | 17,314,014,400 | IssuesEvent | 2021-07-27 01:44:59 | logan-giese/whiteboard-course-management | https://api.github.com/repos/logan-giese/whiteboard-course-management | closed | Create unit tests for assignment services | testing | - [x] Create unit tests for assignment fetching (getting all assignments in a course and by ID)
- [x] Create unit tests for assignment creation
- [x] Create unit tests for assignment updates
- [x] Create unit tests for assignment deletion (delete the specific assignment that was created in the creation test) | 1.0 | Create unit tests for assignment services - - [x] Create unit tests for assignment fetching (getting all assignments in a course and by ID)
- [x] Create unit tests for assignment creation
- [x] Create unit tests for assignment updates
- [x] Create unit tests for assignment deletion (delete the specific assignment that was created in the creation test) | non_priority | create unit tests for assignment services create unit tests for assignment fetching getting all assignments in a course and by id create unit tests for assignment creation create unit tests for assignment updates create unit tests for assignment deletion delete the specific assignment that was created in the creation test | 0 |
55,425 | 14,451,298,226 | IssuesEvent | 2020-12-08 10:46:36 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | opened | Severe error calculated density of air is negative | Defect | Issue overview
--------------
user reported on unmet hours that he got the following error:
```
** Severe ** PsyRhoAirFnPbTdbW: RhoAir (Density of Air) is calculated <= 0 [-9.04677].
** ~~~ ** pb =[101100.00], tdb=[-312.09], w=[0.0000000].
** ~~~ ** Routine=CorrectZoneHumRat, During Warmup, Environment=DSDAYTANGIER COOLING 0.4%, at Simulation time=08/21 00:12 - 00:13
** Fatal ** Program terminates due to preceding condition.
```
user supplied the file, which was at E+ 8.7. I updated the file to v9.4 and instead of getting this severe I am getting an actual crash, which isn't something we want to happen, even if it is later found the defect file had problems.
```
Performing Zone Sizing Simulation
...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4%
Warming up
Warming up
Warming up
Performing Zone Sizing Simulation
...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6%
Calculating System sizing
...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4%
Calculating System sizing
...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6%
Adjusting Air System Sizing
Adjusting Standard 62.1 Ventilation Sizing
Initializing Simulation
double free or corruption (out)
Aborted (core dumped)
```
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link: https://unmethours.com/question/49383/severe-error-calculated-density-of-air-is-negative/
Backtrace
```
(lldb) bt
* thread #1, name = 'energyplus', stop reason = hit program assert
frame #0: 0x00007ffff11c7fb7 libc.so.6`__GI_raise(sig=<unavailable>) at raise.c:51
frame #1: 0x00007ffff11c9921 libc.so.6`__GI_abort at abort.c:79
frame #2: 0x00007ffff11b948a libc.so.6`__assert_fail_base(fmt="%s%s%s:%u: %s%sAssertion `%s' failed.\n%n", assertion="contains( i )", file="/home/julien/Software/Others/EnergyPlus/third_party/ObjexxFCL/src/ObjexxFCL/Array1.hh", line=1172, function="T& ObjexxFCL::Array1< <template-parameter-1-1> >::operator()(int) [with T = EnergyPlus::WaterThermalTanks::StratifiedNodeData]") at assert.c:92
frame #3: 0x00007ffff11b9502 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) at assert.c:101
* frame #4: 0x00007ffff4d1a4b1 libenergyplusapi.so.9.4.0`ObjexxFCL::Array1<EnergyPlus::WaterThermalTanks::StratifiedNodeData>::operator(this=0x00005555564ff530, i=0)(int) at Array1.hh:1172
frame #5: 0x00007ffff4cea195 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcWaterThermalTankStratified(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:7218
frame #6: 0x00007ffff4d11882 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcStandardRatings(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:11451
frame #7: 0x00007ffff4ce47d4 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::initialize(this=0x00005555564ff100, state=0x00007fffffffbdf0, FirstHVACIteration=true) at WaterThermalTanks.cc:6183
frame #8: 0x00007ffff4c8c218 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::onInitLoopEquip(this=0x00005555564ff100, state=0x00007fffffffbdf0, calledFromLocation=0x0000555555e45098) at WaterThermalTanks.cc:164
frame #9: 0x00007ffff5ffa702 libenergyplusapi.so.9.4.0`EnergyPlus::DataPlant::CompData::simulate(this=0x0000555555e44f00, state=0x00007fffffffbdf0, FirstHVACIteration=true, InitLoopEquip=0x00007ffff7d6fbdf, GetCompSizFac=true) at Component.cc:119
frame #10: 0x00007ffff5cf1f1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::InitializeLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true) at PlantManager.cc:2204
frame #11: 0x00007ffff5cdbd1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::ManagePlantLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a) at PlantManager.cc:222
frame #12: 0x00007ffff5827b00 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimSelectedEquipment(state=0x00007fffffffbdf0, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a, FirstHVACIteration=0x00007fffffffae2e, LockPlantFlows=false) at HVACManager.cc:1832
frame #13: 0x00007ffff581d054 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:842
frame #14: 0x00007ffff58197f4 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::ManageHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:358
frame #15: 0x00007ffff5a1457a libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::CalcHeatBalanceAir(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:4356
frame #16: 0x00007ffff59b9f7f libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::ManageAirHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:204
frame #17: 0x00007ffff3e451ee libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceSurfaceManager::ManageSurfaceHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceSurfaceManager.cc:281
frame #18: 0x00007ffff5a33c72 libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceManager::ManageHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceManager.cc:363
frame #19: 0x00007ffff47f9269 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::SetupSimulation(state=0x00007fffffffbdf0, ErrorsFound=0x00007fffffffbc23) at SimulationManager.cc:2111
frame #20: 0x00007ffff47ea196 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::ManageSimulation(state=0x00007fffffffbdf0) at SimulationManager.cc:366
frame #21: 0x00007ffff3a08570 libenergyplusapi.so.9.4.0`RunEnergyPlus(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:400
frame #22: 0x00007ffff3a07a17 libenergyplusapi.so.9.4.0`EnergyPlusPgm(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:224
frame #23: 0x000055555578ef66 energyplus`main(argc=6, argv=0x00007fffffffccc8) at main.cc:60
frame #24: 0x00007ffff11aabf7 libc.so.6`__libc_start_main(main=(energyplus`main at main.cc:56), argc=6, argv=0x00007fffffffccc8, init=<unavailable>, fini=<unavailable>, rtld_fini=<unavailable>, stack_end=0x00007fffffffccb8) at libc-start.c:310
frame #25: 0x000055555578ec8a energyplus`_start + 42
```
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | Severe error calculated density of air is negative - Issue overview
--------------
user reported on unmet hours that he got the following error:
```
** Severe ** PsyRhoAirFnPbTdbW: RhoAir (Density of Air) is calculated <= 0 [-9.04677].
** ~~~ ** pb =[101100.00], tdb=[-312.09], w=[0.0000000].
** ~~~ ** Routine=CorrectZoneHumRat, During Warmup, Environment=DSDAYTANGIER COOLING 0.4%, at Simulation time=08/21 00:12 - 00:13
** Fatal ** Program terminates due to preceding condition.
```
user supplied the file, which was at E+ 8.7. I updated the file to v9.4 and instead of getting this severe I am getting an actual crash, which isn't something we want to happen, even if it is later found the defect file had problems.
```
Performing Zone Sizing Simulation
...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4%
Warming up
Warming up
Warming up
Performing Zone Sizing Simulation
...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6%
Calculating System sizing
...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4%
Calculating System sizing
...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6%
Adjusting Air System Sizing
Adjusting Standard 62.1 Ventilation Sizing
Initializing Simulation
double free or corruption (out)
Aborted (core dumped)
```
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link: https://unmethours.com/question/49383/severe-error-calculated-density-of-air-is-negative/
Backtrace
```
(lldb) bt
* thread #1, name = 'energyplus', stop reason = hit program assert
frame #0: 0x00007ffff11c7fb7 libc.so.6`__GI_raise(sig=<unavailable>) at raise.c:51
frame #1: 0x00007ffff11c9921 libc.so.6`__GI_abort at abort.c:79
frame #2: 0x00007ffff11b948a libc.so.6`__assert_fail_base(fmt="%s%s%s:%u: %s%sAssertion `%s' failed.\n%n", assertion="contains( i )", file="/home/julien/Software/Others/EnergyPlus/third_party/ObjexxFCL/src/ObjexxFCL/Array1.hh", line=1172, function="T& ObjexxFCL::Array1< <template-parameter-1-1> >::operator()(int) [with T = EnergyPlus::WaterThermalTanks::StratifiedNodeData]") at assert.c:92
frame #3: 0x00007ffff11b9502 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) at assert.c:101
* frame #4: 0x00007ffff4d1a4b1 libenergyplusapi.so.9.4.0`ObjexxFCL::Array1<EnergyPlus::WaterThermalTanks::StratifiedNodeData>::operator(this=0x00005555564ff530, i=0)(int) at Array1.hh:1172
frame #5: 0x00007ffff4cea195 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcWaterThermalTankStratified(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:7218
frame #6: 0x00007ffff4d11882 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcStandardRatings(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:11451
frame #7: 0x00007ffff4ce47d4 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::initialize(this=0x00005555564ff100, state=0x00007fffffffbdf0, FirstHVACIteration=true) at WaterThermalTanks.cc:6183
frame #8: 0x00007ffff4c8c218 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::onInitLoopEquip(this=0x00005555564ff100, state=0x00007fffffffbdf0, calledFromLocation=0x0000555555e45098) at WaterThermalTanks.cc:164
frame #9: 0x00007ffff5ffa702 libenergyplusapi.so.9.4.0`EnergyPlus::DataPlant::CompData::simulate(this=0x0000555555e44f00, state=0x00007fffffffbdf0, FirstHVACIteration=true, InitLoopEquip=0x00007ffff7d6fbdf, GetCompSizFac=true) at Component.cc:119
frame #10: 0x00007ffff5cf1f1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::InitializeLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true) at PlantManager.cc:2204
frame #11: 0x00007ffff5cdbd1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::ManagePlantLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a) at PlantManager.cc:222
frame #12: 0x00007ffff5827b00 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimSelectedEquipment(state=0x00007fffffffbdf0, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a, FirstHVACIteration=0x00007fffffffae2e, LockPlantFlows=false) at HVACManager.cc:1832
frame #13: 0x00007ffff581d054 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:842
frame #14: 0x00007ffff58197f4 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::ManageHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:358
frame #15: 0x00007ffff5a1457a libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::CalcHeatBalanceAir(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:4356
frame #16: 0x00007ffff59b9f7f libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::ManageAirHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:204
frame #17: 0x00007ffff3e451ee libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceSurfaceManager::ManageSurfaceHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceSurfaceManager.cc:281
frame #18: 0x00007ffff5a33c72 libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceManager::ManageHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceManager.cc:363
frame #19: 0x00007ffff47f9269 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::SetupSimulation(state=0x00007fffffffbdf0, ErrorsFound=0x00007fffffffbc23) at SimulationManager.cc:2111
frame #20: 0x00007ffff47ea196 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::ManageSimulation(state=0x00007fffffffbdf0) at SimulationManager.cc:366
frame #21: 0x00007ffff3a08570 libenergyplusapi.so.9.4.0`RunEnergyPlus(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:400
frame #22: 0x00007ffff3a07a17 libenergyplusapi.so.9.4.0`EnergyPlusPgm(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:224
frame #23: 0x000055555578ef66 energyplus`main(argc=6, argv=0x00007fffffffccc8) at main.cc:60
frame #24: 0x00007ffff11aabf7 libc.so.6`__libc_start_main(main=(energyplus`main at main.cc:56), argc=6, argv=0x00007fffffffccc8, init=<unavailable>, fini=<unavailable>, rtld_fini=<unavailable>, stack_end=0x00007fffffffccb8) at libc-start.c:310
frame #25: 0x000055555578ec8a energyplus`_start + 42
```
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| non_priority | severe error calculated density of air is negative issue overview user reported on unmet hours that he got the following error severe psyrhoairfnpbtdbw rhoair density of air is calculated pb tdb w routine correctzonehumrat during warmup environment dsdaytangier cooling at simulation time fatal program terminates due to preceding condition user supplied the file which was at e i updated the file to and instead of getting this severe i am getting an actual crash which isn t something we want to happen even if it is later found the defect file had problems performing zone sizing simulation for sizing period dsdaytangier cooling warming up warming up warming up performing zone sizing simulation for sizing period dsdaytangier heating calculating system sizing for sizing period dsdaytangier cooling calculating system sizing for sizing period dsdaytangier heating adjusting air system sizing adjusting standard ventilation sizing initializing simulation double free or corruption out aborted core dumped details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link backtrace lldb bt thread name energyplus stop reason hit program assert frame libc so gi raise sig at raise c frame libc so gi abort at abort c frame libc so assert fail base fmt s s s u s sassertion s failed n n assertion contains i file home julien software others energyplus third party objexxfcl src objexxfcl hh line function t objexxfcl operator int at assert c frame libc so gi assert fail assertion file line function at assert c frame libenergyplusapi so objexxfcl operator this i int at hh frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata calcwaterthermaltankstratified this state at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata calcstandardratings this state at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata initialize this state firsthvaciteration true at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata oninitloopequip this state calledfromlocation at waterthermaltanks cc frame libenergyplusapi so energyplus dataplant compdata simulate this state firsthvaciteration true initloopequip getcompsizfac true at component cc frame libenergyplusapi so energyplus plantmanager initializeloops state firsthvaciteration true at plantmanager cc frame libenergyplusapi so energyplus plantmanager manageplantloops state firsthvaciteration true simairloops simzoneequipment simnonzoneequipment simplantloops simeleccircuits at plantmanager cc frame libenergyplusapi so energyplus hvacmanager simselectedequipment state simairloops simzoneequipment simnonzoneequipment simplantloops simeleccircuits firsthvaciteration lockplantflows false at hvacmanager cc frame libenergyplusapi so energyplus hvacmanager simhvac state at hvacmanager cc frame libenergyplusapi so energyplus hvacmanager managehvac state at hvacmanager cc frame libenergyplusapi so energyplus heatbalanceairmanager calcheatbalanceair state at heatbalanceairmanager cc frame libenergyplusapi so energyplus heatbalanceairmanager manageairheatbalance state at heatbalanceairmanager cc frame libenergyplusapi so energyplus heatbalancesurfacemanager managesurfaceheatbalance state at heatbalancesurfacemanager cc frame libenergyplusapi so energyplus heatbalancemanager manageheatbalance state at heatbalancemanager cc frame libenergyplusapi so energyplus simulationmanager setupsimulation state errorsfound at simulationmanager cc frame libenergyplusapi so energyplus simulationmanager managesimulation state at simulationmanager cc frame libenergyplusapi so runenergyplus state filepath xbd xff xff xff at energypluspgm cc frame libenergyplusapi so energypluspgm state filepath xbd xff xff xff at energypluspgm cc frame energyplus main argc argv at main cc frame libc so libc start main main energyplus main at main cc argc argv init fini rtld fini stack end at libc start c frame energyplus start checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 0 |
70,519 | 15,085,829,832 | IssuesEvent | 2021-02-05 19:18:48 | mthbernardes/shaggy-rogers | https://api.github.com/repos/mthbernardes/shaggy-rogers | closed | CVE-2019-20330 (High) detected in jackson-databind-2.9.6.jar - autoclosed | security vulnerability | ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: shaggy-rogers/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- pantomime-2.11.0.jar (Root Library)
- tika-parsers-1.19.1.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-20330 (High) detected in jackson-databind-2.9.6.jar - autoclosed - ## CVE-2019-20330 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: shaggy-rogers/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- pantomime-2.11.0.jar (Root Library)
- tika-parsers-1.19.1.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.
<p>Publish Date: 2020-01-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p>
<p>Release Date: 2020-01-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file shaggy rogers pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy pantomime jar root library tika parsers jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before lacks certain net sf ehcache blocking publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com novemberain pantomime org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before lacks certain net sf ehcache blocking vulnerabilityurl | 0 |
631,931 | 20,165,085,765 | IssuesEvent | 2022-02-10 02:53:16 | ms-club-sliit/msclubwesite-v2 | https://api.github.com/repos/ms-club-sliit/msclubwesite-v2 | closed | Scroll to top alignment issue | bug help wanted high-priority | ## Issue ⚠️
After we add the live chat :speech_balloon: button, the **Scroll to top** goes underneath the chat button. We need to make **Scroll to top** button on top of the chat button.

| 1.0 | Scroll to top alignment issue - ## Issue ⚠️
After we add the live chat :speech_balloon: button, the **Scroll to top** goes underneath the chat button. We need to make **Scroll to top** button on top of the chat button.

| priority | scroll to top alignment issue issue ⚠️ after we add the live chat speech balloon button the scroll to top goes underneath the chat button we need to make scroll to top button on top of the chat button | 1 |
729,100 | 25,109,116,002 | IssuesEvent | 2022-11-08 18:58:57 | deckhouse/deckhouse | https://api.github.com/repos/deckhouse/deckhouse | closed | [monitoring-kubernetes-control-plane] Control plane proxy hangs | area/monitoring type/bug status/no-stale priority/backlog | ```
curl "https://172.16.29.10:10370/metrics"
curl: (7) Failed to connect to 172.16.29.10 port 10370: Connection refused
```
Logs from the pod:
```
kubectl -n d8-monitoring logs control-plane-proxy-425f55b4-wnwwp
Generating RSA private key, 2048 bit long modulus (2 primes)
..........+++++
...........................+++++
e is 65537 (0x010001)
writing RSA key
Signature ok
subject=CN = p-payment-ks-master-0
Getting Private key
nginx: the configuration file /etc/nginx/nginx.conf syntax is ok
nginx: configuration file /etc/nginx/nginx.conf test is successful
2021/08/18 23:40:08 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370"
2021/08/18 23:40:38 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370"
```
A possible solution is to add a liveness probe. | 1.0 | [monitoring-kubernetes-control-plane] Control plane proxy hangs - ```
curl "https://172.16.29.10:10370/metrics"
curl: (7) Failed to connect to 172.16.29.10 port 10370: Connection refused
```
Logs from the pod:
```
kubectl -n d8-monitoring logs control-plane-proxy-425f55b4-wnwwp
Generating RSA private key, 2048 bit long modulus (2 primes)
..........+++++
...........................+++++
e is 65537 (0x010001)
writing RSA key
Signature ok
subject=CN = p-payment-ks-master-0
Getting Private key
nginx: the configuration file /etc/nginx/nginx.conf syntax is ok
nginx: configuration file /etc/nginx/nginx.conf test is successful
2021/08/18 23:40:08 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370"
2021/08/18 23:40:38 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370"
```
A possible solution is to add a liveness probe. | priority | control plane proxy hangs curl curl failed to connect to port connection refused logs from the pod kubectl n monitoring logs control plane proxy wnwwp generating rsa private key bit long modulus primes e is writing rsa key signature ok subject cn p payment ks master getting private key nginx the configuration file etc nginx nginx conf syntax is ok nginx configuration file etc nginx nginx conf test is successful connect failed connection refused while connecting to upstream client server p payment ks master request get metrics http upstream host connect failed connection refused while connecting to upstream client server p payment ks master request get metrics http upstream host a possible solution is to add a liveness probe | 1 |
74,118 | 20,022,434,007 | IssuesEvent | 2022-02-01 17:36:46 | Start9Labs/embassy-os | https://api.github.com/repos/Start9Labs/embassy-os | closed | explore different kernels or base OS's to resolve stability issues | P0 - Blocks Dev Builds | related #901
Currently the raspberry pi has stability issues after being up for several days. Investigation into this has revealed processes are getting stuck in state `R` while being unable to be `SIGKILL`ed. According to https://unix.stackexchange.com/questions/386380/wget-cant-be-killed-with-kill-9, this is either a hardware fault or a kernel bug. Given that there is likely a problem with the kernel, or its compatibility with the pi, we need to upgrade the kernel. We have a few options here:
- Upgrade the kernel in `initialization.sh`
- Use the official Raspberry Pi OS as the base image (https://downloads.raspberrypi.org/raspios_arm64/images/raspios_arm64-2021-05-28/) | 1.0 | explore different kernels or base OS's to resolve stability issues - related #901
Currently the raspberry pi has stability issues after being up for several days. Investigation into this has revealed processes are getting stuck in state `R` while being unable to be `SIGKILL`ed. According to https://unix.stackexchange.com/questions/386380/wget-cant-be-killed-with-kill-9, this is either a hardware fault or a kernel bug. Given that there is likely a problem with the kernel, or its compatibility with the pi, we need to upgrade the kernel. We have a few options here:
- Upgrade the kernel in `initialization.sh`
- Use the official Raspberry Pi OS as the base image (https://downloads.raspberrypi.org/raspios_arm64/images/raspios_arm64-2021-05-28/) | non_priority | explore different kernels or base os s to resolve stability issues related currently the raspberry pi has stability issues after being up for several days investigation into this has revealed processes are getting stuck in state r while being unable to be sigkill ed according to this is either a hardware fault or a kernel bug given that there is likely a problem with the kernel or its compatibility with the pi we need to upgrade the kernel we have a few options here upgrade the kernel in initialization sh use the official raspberry pi os as the base image | 0 |
186,207 | 15,051,035,627 | IssuesEvent | 2021-02-03 13:38:39 | Urban-Analytics/RAMP-UA | https://api.github.com/repos/Urban-Analytics/RAMP-UA | opened | Documentation for the lockdown scenario(s) | documentation | In particular, how do we retrieve the google data for these? | 1.0 | Documentation for the lockdown scenario(s) - In particular, how do we retrieve the google data for these? | non_priority | documentation for the lockdown scenario s in particular how do we retrieve the google data for these | 0 |
9,607 | 3,935,224,451 | IssuesEvent | 2016-04-26 03:38:02 | EmergentOrganization/cell-rpg | https://api.github.com/repos/EmergentOrganization/cell-rpg | closed | weapon collision effect | cat: art cat: code enhancement question / discussion | Some effect is needed to make impact on the ca more apparent. I was thinking about having the impacted squares in the CA switch to a bright color before clearing, but we also have the option of using a particle effect or an explosion sprite. I'm not sure which of these options will look best. | 1.0 | weapon collision effect - Some effect is needed to make impact on the ca more apparent. I was thinking about having the impacted squares in the CA switch to a bright color before clearing, but we also have the option of using a particle effect or an explosion sprite. I'm not sure which of these options will look best. | non_priority | weapon collision effect some effect is needed to make impact on the ca more apparent i was thinking about having the impacted squares in the ca switch to a bright color before clearing but we also have the option of using a particle effect or an explosion sprite i m not sure which of these options will look best | 0 |
105,369 | 4,234,717,617 | IssuesEvent | 2016-07-05 13:02:02 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | An error occurred while building 1.2(openshift_node_dnsmasq) | component/install priority/P2 | Version: oepnshift-origin-1.2.0
Environment: The new system did not install anything(Centos cloud images)
build: 3master+3etcd+3node+1lb[1]+DNS / 1master+2node
command: ansible-playbook ~/openshift-ansible/playbooks/byo/config.yml
Hi,
There is a problem when I build openshift 1.2
{"changed": false, "failed": true, "msg": "Currently, NetworkManager must be installed and enabled prior to installation."}


when i use `ansible-playbook ~/openshift-ansible/playbooks/adhoc/uninstall.yml` and build again,it is ok.but if i want to add a new node,also error,how can i do for this?
thanks. | 1.0 | An error occurred while building 1.2(openshift_node_dnsmasq) - Version: oepnshift-origin-1.2.0
Environment: The new system did not install anything(Centos cloud images)
build: 3master+3etcd+3node+1lb[1]+DNS / 1master+2node
command: ansible-playbook ~/openshift-ansible/playbooks/byo/config.yml
Hi,
There is a problem when I build openshift 1.2
{"changed": false, "failed": true, "msg": "Currently, NetworkManager must be installed and enabled prior to installation."}


when i use `ansible-playbook ~/openshift-ansible/playbooks/adhoc/uninstall.yml` and build again,it is ok.but if i want to add a new node,also error,how can i do for this?
thanks. | priority | an error occurred while building openshift node dnsmasq version oepnshift origin environment the new system did not install anything centos cloud images build dns command ansible playbook openshift ansible playbooks byo config yml hi there is a problem when i build openshift changed false failed true msg currently networkmanager must be installed and enabled prior to installation when i use ansible playbook openshift ansible playbooks adhoc uninstall yml and build again it is ok but if i want to add a new node also error how can i do for this thanks | 1 |
657,600 | 21,797,735,401 | IssuesEvent | 2022-05-15 21:40:17 | kubernetes/website | https://api.github.com/repos/kubernetes/website | closed | kubectl generated command docs do not include 1.24 commands | kind/bug priority/important-soon sig/cli triage/accepted | **This is a Bug Report**
<!-- Thanks for filing an issue! Before submitting, please fill in the following information. -->
<!-- See https://kubernetes.io/docs/contribute/start/ for guidance on writing an actionable issue description. -->
<!--Required Information-->
**Problem:**
https://kubernetes.io/docs/reference/generated/kubectl/kubectl-commands#create does not include commands added in 1.24 (notably `kubectl create token`)
It looks like https://github.com/kubernetes/website/tree/main/static/docs/reference/generated/kubectl is supposed to be auto-generated, but has not been updated for 1.24.
/sig cli | 1.0 | kubectl generated command docs do not include 1.24 commands - **This is a Bug Report**
<!-- Thanks for filing an issue! Before submitting, please fill in the following information. -->
<!-- See https://kubernetes.io/docs/contribute/start/ for guidance on writing an actionable issue description. -->
<!--Required Information-->
**Problem:**
https://kubernetes.io/docs/reference/generated/kubectl/kubectl-commands#create does not include commands added in 1.24 (notably `kubectl create token`)
It looks like https://github.com/kubernetes/website/tree/main/static/docs/reference/generated/kubectl is supposed to be auto-generated, but has not been updated for 1.24.
/sig cli | priority | kubectl generated command docs do not include commands this is a bug report problem does not include commands added in notably kubectl create token it looks like is supposed to be auto generated but has not been updated for sig cli | 1 |
15,389 | 8,873,144,035 | IssuesEvent | 2019-01-11 17:13:41 | snowleopard/hadrian | https://api.github.com/repos/snowleopard/hadrian | closed | Should Hadrian build ghctags and haddock in Stage1? | enhancement performance question | The Make build system needs to build `ghctags` and `haddock` programs in Stage2. For example, see this commit: https://github.com/ghc/ghc/commit/5fb72555f7b7ab67a33583f33ad9160761ca434f.
However, #531 moved `haddock` to Stage1 and it worked fine. I couldn't find an explanation behind the move and reverted this change as part of the cleaning-up-after-531 issue #540. Let's discuss this here.
There is a `Note [No stage2 packages when CrossCompiling or Stage1Only]` in `ghc.mk`, which in particular says:
```
# Here's why:
# - first of all, ghc-stage1 can't use stage0's ghc library (it's too old)
# - neither do we register the ghc library (compiler/stage1) that we build
# with stage0. TODO Why not? We do build it...
# - as a result, we need to a) use ghc-stage2 to build packages that depend on
# the ghc library (e.g. ghctags [4]) and b) exclude those packages when
# ghc-stage2 is not available.
```
But in Hadrian we do register the `ghc` library built in Stage0, so presumably this means we could indeed now build all its dependants in Stage1.
Shall we move `ghctags` and `haddock` to Stage1? This simplifies the build system and also improves the performance, since we don't need to wait for Stage2 GHC to build these utilities. | True | Should Hadrian build ghctags and haddock in Stage1? - The Make build system needs to build `ghctags` and `haddock` programs in Stage2. For example, see this commit: https://github.com/ghc/ghc/commit/5fb72555f7b7ab67a33583f33ad9160761ca434f.
However, #531 moved `haddock` to Stage1 and it worked fine. I couldn't find an explanation behind the move and reverted this change as part of the cleaning-up-after-531 issue #540. Let's discuss this here.
There is a `Note [No stage2 packages when CrossCompiling or Stage1Only]` in `ghc.mk`, which in particular says:
```
# Here's why:
# - first of all, ghc-stage1 can't use stage0's ghc library (it's too old)
# - neither do we register the ghc library (compiler/stage1) that we build
# with stage0. TODO Why not? We do build it...
# - as a result, we need to a) use ghc-stage2 to build packages that depend on
# the ghc library (e.g. ghctags [4]) and b) exclude those packages when
# ghc-stage2 is not available.
```
But in Hadrian we do register the `ghc` library built in Stage0, so presumably this means we could indeed now build all its dependants in Stage1.
Shall we move `ghctags` and `haddock` to Stage1? This simplifies the build system and also improves the performance, since we don't need to wait for Stage2 GHC to build these utilities. | non_priority | should hadrian build ghctags and haddock in the make build system needs to build ghctags and haddock programs in for example see this commit however moved haddock to and it worked fine i couldn t find an explanation behind the move and reverted this change as part of the cleaning up after issue let s discuss this here there is a note in ghc mk which in particular says here s why first of all ghc can t use s ghc library it s too old neither do we register the ghc library compiler that we build with todo why not we do build it as a result we need to a use ghc to build packages that depend on the ghc library e g ghctags and b exclude those packages when ghc is not available but in hadrian we do register the ghc library built in so presumably this means we could indeed now build all its dependants in shall we move ghctags and haddock to this simplifies the build system and also improves the performance since we don t need to wait for ghc to build these utilities | 0 |
127,891 | 10,492,338,984 | IssuesEvent | 2019-09-25 13:06:22 | linkedpipes/etl | https://api.github.com/repos/linkedpipes/etl | closed | Add support for SHACL | enhancement test | Add support for [SHACL](https://www.w3.org/TR/shacl/) validation.
It should be a new component (`q-shacl`) with 2 inputs - Data graph, Shapes graph and 1 output - Validation report.
- There is some [initial work on SHACL support in RDF4J](https://github.com/eclipse/rdf4j-storage/pull/34), however, the support is in form of a Shacl repository, which I am not sure will be useful to our use case
- The [reference implementation](https://github.com/TopQuadrant/shacl) is based on Apache Jena, so I guess such component would have to work with files, shielding the users from the implementation a bit, unless we are willing to add Jena support and all the related transformers.
Related to #631 | 1.0 | Add support for SHACL - Add support for [SHACL](https://www.w3.org/TR/shacl/) validation.
It should be a new component (`q-shacl`) with 2 inputs - Data graph, Shapes graph and 1 output - Validation report.
- There is some [initial work on SHACL support in RDF4J](https://github.com/eclipse/rdf4j-storage/pull/34), however, the support is in form of a Shacl repository, which I am not sure will be useful to our use case
- The [reference implementation](https://github.com/TopQuadrant/shacl) is based on Apache Jena, so I guess such component would have to work with files, shielding the users from the implementation a bit, unless we are willing to add Jena support and all the related transformers.
Related to #631 | non_priority | add support for shacl add support for validation it should be a new component q shacl with inputs data graph shapes graph and output validation report there is some however the support is in form of a shacl repository which i am not sure will be useful to our use case the is based on apache jena so i guess such component would have to work with files shielding the users from the implementation a bit unless we are willing to add jena support and all the related transformers related to | 0 |
253,654 | 27,300,776,629 | IssuesEvent | 2023-02-24 01:37:05 | panasalap/linux-4.19.72_1 | https://api.github.com/repos/panasalap/linux-4.19.72_1 | closed | CVE-2019-19075 (High) detected in linux-yoctov5.4.51 - autoclosed | security vulnerability | ## CVE-2019-19075 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the ca8210_probe() function in drivers/net/ieee802154/ca8210.c in the Linux kernel before 5.3.8 allows attackers to cause a denial of service (memory consumption) by triggering ca8210_get_platform_data() failures, aka CID-6402939ec86e.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19075>CVE-2019-19075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.4-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19075 (High) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2019-19075 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the ca8210_probe() function in drivers/net/ieee802154/ca8210.c in the Linux kernel before 5.3.8 allows attackers to cause a denial of service (memory consumption) by triggering ca8210_get_platform_data() failures, aka CID-6402939ec86e.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19075>CVE-2019-19075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.4-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linux autoclosed cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net c drivers net c vulnerability details a memory leak in the probe function in drivers net c in the linux kernel before allows attackers to cause a denial of service memory consumption by triggering get platform data failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
1,496 | 21,956,078,848 | IssuesEvent | 2022-05-24 12:14:46 | TYPO3-Solr/ext-solr | https://api.github.com/repos/TYPO3-Solr/ext-solr | closed | [BUG] Indexing records with SOLR_RELATION via backend considers hidden relation records | 10LTS BACKPORTABLE | **Describe the bug**
In [Backend context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/backend/Classes/Http/Application.php#L91) TYPO3 will set a VisibilityAspect that allows showing hidden pages and content. When a record is translated and the translation is then hidden, the hidden translation will still get indexed in SOLR_RELATION. This only happens when indexing is done via the TYPO3 backend, so there is a different behaviour from indexing via CLI.
The suggested approach is to reset the VisibilityAspect in `Ext:solr`s Tsfe class to make sure hidden elements are not shown.
Also, I was not able to reproduce this with SOLR_RELATION fields with the PageIndexer. For some reason indexing via CLI is fine even though it is also allows hidden pages and content elements (see [CLI context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/core/Classes/Console/CommandApplication.php#L185)).
---
**To Reproduce**
While this error was found in our internal extension, I tried my best to reproduce this with Ext:news.
This test assumes English as the main language (Index: `core_en`) and German as the second language (Index: `core_de`)
Steps to reproduce the behavior:
1. Create two sys_categories "Category 1" and "Category 2". Create a German translation for both of them ("Kategorie 1" and "Kategorie 2") and disable the translation of "Category 2"
2. Create a Ext:news record and assign both categories to it. Then create a German translation. Make sure both categories are selected there as well.
3. Add News indexer in TypoScript setup:
```typoscript
@import 'EXT:solr/Configuration/TypoScript/Examples/IndexQueueNews/setup.typoscript'
```
3. Index the news record via the TYPO3 backend.
4. Look at the index entries.
In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],`
In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Kategorie 2"],`
**Expected behavior**
In German index, "Kategorie 2" should not show up as it is disabled.
It should behave the same as when indexing via CLI, i.e. it should index the original record.
In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],`
In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Category 2"],`
**Used versions (please complete the following information):**
- TYPO3 Version: 10.4.21
- EXT:solr Version: 11.0.2
- Used Apache Solr Version: 8.5.1, 8.9.0
- PHP Version: 7.4.20 | True | [BUG] Indexing records with SOLR_RELATION via backend considers hidden relation records - **Describe the bug**
In [Backend context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/backend/Classes/Http/Application.php#L91) TYPO3 will set a VisibilityAspect that allows showing hidden pages and content. When a record is translated and the translation is then hidden, the hidden translation will still get indexed in SOLR_RELATION. This only happens when indexing is done via the TYPO3 backend, so there is a different behaviour from indexing via CLI.
The suggested approach is to reset the VisibilityAspect in `Ext:solr`s Tsfe class to make sure hidden elements are not shown.
Also, I was not able to reproduce this with SOLR_RELATION fields with the PageIndexer. For some reason indexing via CLI is fine even though it is also allows hidden pages and content elements (see [CLI context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/core/Classes/Console/CommandApplication.php#L185)).
---
**To Reproduce**
While this error was found in our internal extension, I tried my best to reproduce this with Ext:news.
This test assumes English as the main language (Index: `core_en`) and German as the second language (Index: `core_de`)
Steps to reproduce the behavior:
1. Create two sys_categories "Category 1" and "Category 2". Create a German translation for both of them ("Kategorie 1" and "Kategorie 2") and disable the translation of "Category 2"
2. Create a Ext:news record and assign both categories to it. Then create a German translation. Make sure both categories are selected there as well.
3. Add News indexer in TypoScript setup:
```typoscript
@import 'EXT:solr/Configuration/TypoScript/Examples/IndexQueueNews/setup.typoscript'
```
3. Index the news record via the TYPO3 backend.
4. Look at the index entries.
In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],`
In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Kategorie 2"],`
**Expected behavior**
In German index, "Kategorie 2" should not show up as it is disabled.
It should behave the same as when indexing via CLI, i.e. it should index the original record.
In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],`
In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Category 2"],`
**Used versions (please complete the following information):**
- TYPO3 Version: 10.4.21
- EXT:solr Version: 11.0.2
- Used Apache Solr Version: 8.5.1, 8.9.0
- PHP Version: 7.4.20 | non_priority | indexing records with solr relation via backend considers hidden relation records describe the bug in will set a visibilityaspect that allows showing hidden pages and content when a record is translated and the translation is then hidden the hidden translation will still get indexed in solr relation this only happens when indexing is done via the backend so there is a different behaviour from indexing via cli the suggested approach is to reset the visibilityaspect in ext solr s tsfe class to make sure hidden elements are not shown also i was not able to reproduce this with solr relation fields with the pageindexer for some reason indexing via cli is fine even though it is also allows hidden pages and content elements see to reproduce while this error was found in our internal extension i tried my best to reproduce this with ext news this test assumes english as the main language index core en and german as the second language index core de steps to reproduce the behavior create two sys categories category and category create a german translation for both of them kategorie and kategorie and disable the translation of category create a ext news record and assign both categories to it then create a german translation make sure both categories are selected there as well add news indexer in typoscript setup typoscript import ext solr configuration typoscript examples indexqueuenews setup typoscript index the news record via the backend look at the index entries in core en the entry has category stringm in core de the entry has category stringm expected behavior in german index kategorie should not show up as it is disabled it should behave the same as when indexing via cli i e it should index the original record in core en the entry has category stringm in core de the entry has category stringm used versions please complete the following information version ext solr version used apache solr version php version | 0 |
32,979 | 8,989,513,493 | IssuesEvent | 2019-02-01 00:11:50 | orbeon/orbeon-forms | https://api.github.com/repos/orbeon/orbeon-forms | closed | Incorrect test for form definition versioning | Form Builder Form Runner | We search for the provider on the data side, but we need to search for the provider on the form definition side.
This is usually not a problem, as the same provider is usually configured for form and data.
This has been the case since #1669 in Orbeon Forms 4.6. | 1.0 | Incorrect test for form definition versioning - We search for the provider on the data side, but we need to search for the provider on the form definition side.
This is usually not a problem, as the same provider is usually configured for form and data.
This has been the case since #1669 in Orbeon Forms 4.6. | non_priority | incorrect test for form definition versioning we search for the provider on the data side but we need to search for the provider on the form definition side this is usually not a problem as the same provider is usually configured for form and data this has been the case since in orbeon forms | 0 |
129,024 | 5,082,218,251 | IssuesEvent | 2016-12-29 14:37:02 | vladyslav2/gfwhitelabels | https://api.github.com/repos/vladyslav2/gfwhitelabels | opened | Problems with the DCU Site | Priority | 1. Communication channel not working
DCU site:

GrowthFountain Site:

2. The site does not work properly in mobile
3. I cannot login (when I tried to invest it brought me to this login page which doesn’t work)
4. Hide social sign in on login page

5. When I click “RAISE CAPITAL” on the landing page it prompts SIGN UP but it doesn’t work
6. Hide BLOG––––
| 1.0 | Problems with the DCU Site - 1. Communication channel not working
DCU site:

GrowthFountain Site:

2. The site does not work properly in mobile
3. I cannot login (when I tried to invest it brought me to this login page which doesn’t work)
4. Hide social sign in on login page

5. When I click “RAISE CAPITAL” on the landing page it prompts SIGN UP but it doesn’t work
6. Hide BLOG––––
| priority | problems with the dcu site communication channel not working dcu site growthfountain site the site does not work properly in mobile i cannot login when i tried to invest it brought me to this login page which doesn’t work hide social sign in on login page when i click “raise capital” on the landing page it prompts sign up but it doesn’t work hide blog–––– | 1 |
352,716 | 10,545,323,790 | IssuesEvent | 2019-10-02 18:52:49 | wso2-cellery/sdk | https://api.github.com/repos/wso2-cellery/sdk | closed | Error while executing cellery test | Priority/Highest Severity/Blocker Type/Bug | cellery test wso2cellery/pet-be-cell:latest-dev -n pet-be
✔ Extracting Cell Image wso2cellery/pet-be-cell:latest-dev
✔ Reading Image wso2cellery/pet-be-cell:latest-dev
✔ Starting execution of tests for wso2cellery/pet-be-cell:latest-dev...
? Do you wish to continue with testing above Cell instances (Y/n)?
⠋ Creating telepresence instance
error waiting for telepresence deployment telepresence--telepresence-deployment to be available: exit status 1
Environment: docker-for-mac
| 1.0 | Error while executing cellery test - cellery test wso2cellery/pet-be-cell:latest-dev -n pet-be
✔ Extracting Cell Image wso2cellery/pet-be-cell:latest-dev
✔ Reading Image wso2cellery/pet-be-cell:latest-dev
✔ Starting execution of tests for wso2cellery/pet-be-cell:latest-dev...
? Do you wish to continue with testing above Cell instances (Y/n)?
⠋ Creating telepresence instance
error waiting for telepresence deployment telepresence--telepresence-deployment to be available: exit status 1
Environment: docker-for-mac
| priority | error while executing cellery test cellery test pet be cell latest dev n pet be ✔ extracting cell image pet be cell latest dev ✔ reading image pet be cell latest dev ✔ starting execution of tests for pet be cell latest dev do you wish to continue with testing above cell instances y n ⠋ creating telepresence instance error waiting for telepresence deployment telepresence telepresence deployment to be available exit status environment docker for mac | 1 |
166,383 | 6,303,905,799 | IssuesEvent | 2017-07-21 14:46:03 | BigBrotherTeam/BigBrother | https://api.github.com/repos/BigBrotherTeam/BigBrother | closed | Wrong sound effect is played when PC user open/close door | Category: Gameplay Priority: Low Status: Reproduced | ### Issue description
When any PC user open/close door (include Trapdoor), wrong sound effect (Eye of ender shoot) is played.
### Prerequisite
* PocketMine-MP `g8bb0c53`
* BigBrother master branch 76b03ad
### How to reproduce
* Place any kind of Door block such as Wooden Door, Iron Door or Trap Door
* Right click the placed Door block to open/close
### Server log(crash dump)
no server log related to this issue.
<!--- Thank you for reporting! --> | 1.0 | Wrong sound effect is played when PC user open/close door - ### Issue description
When any PC user open/close door (include Trapdoor), wrong sound effect (Eye of ender shoot) is played.
### Prerequisite
* PocketMine-MP `g8bb0c53`
* BigBrother master branch 76b03ad
### How to reproduce
* Place any kind of Door block such as Wooden Door, Iron Door or Trap Door
* Right click the placed Door block to open/close
### Server log(crash dump)
no server log related to this issue.
<!--- Thank you for reporting! --> | priority | wrong sound effect is played when pc user open close door issue description when any pc user open close door include trapdoor wrong sound effect eye of ender shoot is played prerequisite pocketmine mp bigbrother master branch how to reproduce place any kind of door block such as wooden door iron door or trap door right click the placed door block to open close server log crash dump no server log related to this issue | 1 |
422,853 | 12,287,489,295 | IssuesEvent | 2020-05-09 12:26:52 | googleapis/elixir-google-api | https://api.github.com/repos/googleapis/elixir-google-api | opened | Synthesis failed for Spanner | api: spanner autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate Spanner. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:20:08 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:20:08,292 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-spanner'
2020-05-09 05:20:10 [INFO] Running synthtool
2020-05-09 05:20:10,566 autosynth > Running synthtool
2020-05-09 05:20:10 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--']
2020-05-09 05:20:10,566 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--']
2020-05-09 05:20:10,777 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-spanner
nothing to commit, working tree clean
2020-05-09 05:20:10,853 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:20:11,738 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Spanner
2020-05-09 05:20:15,459 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/spanner/synth.metadata'
2020-05-09 05:20:15 [ERROR] Synthesis failed
2020-05-09 05:20:15,488 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--', 'Spanner']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| 1.0 | Synthesis failed for Spanner - Hello! Autosynth couldn't regenerate Spanner. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:20:08 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:20:08,292 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-spanner'
2020-05-09 05:20:10 [INFO] Running synthtool
2020-05-09 05:20:10,566 autosynth > Running synthtool
2020-05-09 05:20:10 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--']
2020-05-09 05:20:10,566 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--']
2020-05-09 05:20:10,777 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-spanner
nothing to commit, working tree clean
2020-05-09 05:20:10,853 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:20:11,738 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Spanner
2020-05-09 05:20:15,459 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/spanner/synth.metadata'
2020-05-09 05:20:15 [ERROR] Synthesis failed
2020-05-09 05:20:15,488 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--', 'Spanner']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| priority | synthesis failed for spanner hello autosynth couldn t regenerate spanner broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth spanner running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth spanner nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh spanner synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients spanner synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 1 |
26,572 | 26,987,909,588 | IssuesEvent | 2023-02-09 17:27:02 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | opened | `AssetServer` and `gltF` workflows are very prone to failing silently | A-Rendering A-Assets C-Usability | ## What problem does this solve or what need does it fill?
When working with `gltF` files, there are quite a few ways for something to go wrong:
1. The file doesn't exist.
2. The path is wrong.
3. You forgot to specify `#Scene0`.
4. You specified the wrong scene number.
5. Your scale is wildly wrong.
6. Your model is inside out and you have backface culling enabled.
7. Your camera isn't looking at the model.
8. Your material is fully transparent.
9. You're attempting to save the `gltF` scene as something else, possibly a `Mesh` or `StandardMaterial`, rather than a `HandleScene`.
10. You specified an invalid scene path.
Currently, we detect and report 1 and 2. All other cases silently fail, with nothing (apparently) being rendered.
## What solution would you like?
Add error reporting and failures to cases that are flatly wrong: 3, 8, 9, 10.
Warn in cases that are possibly wrong: 4.
Add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model.
## What alternative(s) have you considered?
The categorization of these failure modes is very much up for debate. | True | `AssetServer` and `gltF` workflows are very prone to failing silently - ## What problem does this solve or what need does it fill?
When working with `gltF` files, there are quite a few ways for something to go wrong:
1. The file doesn't exist.
2. The path is wrong.
3. You forgot to specify `#Scene0`.
4. You specified the wrong scene number.
5. Your scale is wildly wrong.
6. Your model is inside out and you have backface culling enabled.
7. Your camera isn't looking at the model.
8. Your material is fully transparent.
9. You're attempting to save the `gltF` scene as something else, possibly a `Mesh` or `StandardMaterial`, rather than a `HandleScene`.
10. You specified an invalid scene path.
Currently, we detect and report 1 and 2. All other cases silently fail, with nothing (apparently) being rendered.
## What solution would you like?
Add error reporting and failures to cases that are flatly wrong: 3, 8, 9, 10.
Warn in cases that are possibly wrong: 4.
Add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model.
## What alternative(s) have you considered?
The categorization of these failure modes is very much up for debate. | non_priority | assetserver and gltf workflows are very prone to failing silently what problem does this solve or what need does it fill when working with gltf files there are quite a few ways for something to go wrong the file doesn t exist the path is wrong you forgot to specify you specified the wrong scene number your scale is wildly wrong your model is inside out and you have backface culling enabled your camera isn t looking at the model your material is fully transparent you re attempting to save the gltf scene as something else possibly a mesh or standardmaterial rather than a handlescene you specified an invalid scene path currently we detect and report and all other cases silently fail with nothing apparently being rendered what solution would you like add error reporting and failures to cases that are flatly wrong warn in cases that are possibly wrong add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model what alternative s have you considered the categorization of these failure modes is very much up for debate | 0 |
123,293 | 12,196,370,325 | IssuesEvent | 2020-04-29 18:57:50 | COVID-19-electronic-health-system/Corona-tracker | https://api.github.com/repos/COVID-19-electronic-health-system/Corona-tracker | closed | [DOCS] Hebrew Translation | documentation |
[CoronaTracker appText.zip](https://github.com/COVID-19-electronic-health-system/Corona-tracker/files/4546409/CoronaTracker.appText.zip)
# ⚠️ IMPORTANT: Please fill out this template to give us as much information as possible to consider/implement this update.
### Summary
<!-- One paragraph explanation of the feature. -->
### Motivation
<!-- Why are we doing this? What use cases does it support? What is the expected outcome? -->
### Possible Alternatives
<!-- A clear and concise description of the alternative solutions you've considered. Be sure to explain why the current documentation isn't suitable for this feature. -->
### Additional Context
<!-- Add any other context or screenshots about the documentation update here. -->
| 1.0 | [DOCS] Hebrew Translation -
[CoronaTracker appText.zip](https://github.com/COVID-19-electronic-health-system/Corona-tracker/files/4546409/CoronaTracker.appText.zip)
# ⚠️ IMPORTANT: Please fill out this template to give us as much information as possible to consider/implement this update.
### Summary
<!-- One paragraph explanation of the feature. -->
### Motivation
<!-- Why are we doing this? What use cases does it support? What is the expected outcome? -->
### Possible Alternatives
<!-- A clear and concise description of the alternative solutions you've considered. Be sure to explain why the current documentation isn't suitable for this feature. -->
### Additional Context
<!-- Add any other context or screenshots about the documentation update here. -->
| non_priority | hebrew translation ⚠️ important please fill out this template to give us as much information as possible to consider implement this update summary motivation possible alternatives additional context | 0 |
163,486 | 20,363,817,295 | IssuesEvent | 2022-02-21 01:31:47 | rgordon95/conFusionAng | https://api.github.com/repos/rgordon95/conFusionAng | opened | CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2022-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /conFusionAng/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- webpack-dev-server-2.11.5.tgz
- sockjs-client-1.1.5.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6.
<p>Publish Date: 2022-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p>
<p>Release Date: 2022-02-14</p>
<p>Fix Resolution: url-parse - 1.5.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz - ## CVE-2022-0512 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /conFusionAng/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- webpack-dev-server-2.11.5.tgz
- sockjs-client-1.1.5.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6.
<p>Publish Date: 2022-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p>
<p>Release Date: 2022-02-14</p>
<p>Fix Resolution: url-parse - 1.5.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file confusionang package json path to vulnerable library node modules url parse package json dependency hierarchy cli tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
294,722 | 9,039,902,515 | IssuesEvent | 2019-02-10 11:42:08 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | drive.google.com - see bug description | browser-firefox-mobile priority-critical | <!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://drive.google.com/drive/mobile/folders/1-0Y9QnZuXMSHPUB-uw_Ed2sgtQ32KOS9?sort=13&direction=a
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: this girl has control of apps, phone carrier my whole phone
**Steps to Reproduce**:
Settings would change by itself everyday. Noticed files n especially assking permission to share my files. Has a camera watching me n audio plus taking my email n name getting Google Pay, Amazon accounts n my phine carrier
[](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190207161357</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Content Security Policy: Ignoring \'report-sample\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[console.timeStamp(CSI/tbsd_) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_tbnd) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ipls) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/jl) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ai) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feis) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feie) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upl_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/as) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/asc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_2) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upr) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivdp) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_1) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/iple) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_files_api_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_partial) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivlc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/pst) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/tbsd_) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/_tbnd) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/first_files_api_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/fll) https://drive.google.com/drive/mobile/:1:147]']
</pre>
</details>
Reported by @kim6571
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | drive.google.com - see bug description - <!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://drive.google.com/drive/mobile/folders/1-0Y9QnZuXMSHPUB-uw_Ed2sgtQ32KOS9?sort=13&direction=a
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: this girl has control of apps, phone carrier my whole phone
**Steps to Reproduce**:
Settings would change by itself everyday. Noticed files n especially assking permission to share my files. Has a camera watching me n audio plus taking my email n name getting Google Pay, Amazon accounts n my phine carrier
[](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190207161357</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Content Security Policy: Ignoring \'report-sample\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[console.timeStamp(CSI/tbsd_) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_tbnd) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ipls) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/jl) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ai) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feis) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feie) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upl_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/as) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/asc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_2) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upr) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivdp) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_1) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/iple) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_files_api_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_partial) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivlc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/pst) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/tbsd_) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/_tbnd) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/first_files_api_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/fll) https://drive.google.com/drive/mobile/:1:147]']
</pre>
</details>
Reported by @kim6571
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | drive google com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description this girl has control of apps phone carrier my whole phone steps to reproduce settings would change by itself everyday noticed files n especially assking permission to share my files has a camera watching me n audio plus taking my email n name getting google pay amazon accounts n my phine carrier browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u reported by from with ❤️ | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.