Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
53,707
13,196,505,956
IssuesEvent
2020-08-13 20:47:49
halide/Halide
https://api.github.com/repos/halide/Halide
closed
CMake/CPack 'distrib' target isn't correct for Windows
build
A while back, someone thoughtfully contributed CPack support to our CMake files, so that `make distrib` worked correctly for CMake as well as make; unfortunately, it needs some extra love for Windows builds in order to match the status quo for Windows: - Windows requires separate Debug and Release builds; the CMake rules only provide a single build (presumably Release) - The CMake rules only provide shared-library outputs (bin/Halide.dll), whereas our existing Windows distribs provide Halide.lib as well This would be really nice to fix, as currently, Windows distribs are produced by custom code in the buildbots, which can (and does) easily fall out of sync with the rest of Halide.
1.0
CMake/CPack 'distrib' target isn't correct for Windows - A while back, someone thoughtfully contributed CPack support to our CMake files, so that `make distrib` worked correctly for CMake as well as make; unfortunately, it needs some extra love for Windows builds in order to match the status quo for Windows: - Windows requires separate Debug and Release builds; the CMake rules only provide a single build (presumably Release) - The CMake rules only provide shared-library outputs (bin/Halide.dll), whereas our existing Windows distribs provide Halide.lib as well This would be really nice to fix, as currently, Windows distribs are produced by custom code in the buildbots, which can (and does) easily fall out of sync with the rest of Halide.
non_priority
cmake cpack distrib target isn t correct for windows a while back someone thoughtfully contributed cpack support to our cmake files so that make distrib worked correctly for cmake as well as make unfortunately it needs some extra love for windows builds in order to match the status quo for windows windows requires separate debug and release builds the cmake rules only provide a single build presumably release the cmake rules only provide shared library outputs bin halide dll whereas our existing windows distribs provide halide lib as well this would be really nice to fix as currently windows distribs are produced by custom code in the buildbots which can and does easily fall out of sync with the rest of halide
0
331,059
28,504,327,977
IssuesEvent
2023-04-18 20:00:57
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed
C-bug C-test-failure O-robot O-roachtest GA-blocker T-kv-replication branch-release-23.1.0
roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on release-23.1.0 @ [ada9871ea92cfc467b045a6a1b8ed2783a243c2f](https://github.com/cockroachdb/cockroach/commits/ada9871ea92cfc467b045a6a1b8ed2783a243c2f): ``` test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1 (monitor.go:127).Wait: monitor failure: monitor task failed: read tcp 172.17.0.3:58046 -> 3.128.192.232:26257: read: connection reset by peer ``` <p>Parameters: <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=8</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #100804 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery branch-release-23.1] - #100341 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [A-kv-replication C-bug C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery T-kv-replication branch-master branch-release-23.1] </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26944
2.0
roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed - roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestFipsNightlyAwsBazel/9578022?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on release-23.1.0 @ [ada9871ea92cfc467b045a6a1b8ed2783a243c2f](https://github.com/cockroachdb/cockroach/commits/ada9871ea92cfc467b045a6a1b8ed2783a243c2f): ``` test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1 (monitor.go:127).Wait: monitor failure: monitor task failed: read tcp 172.17.0.3:58046 -> 3.128.192.232:26257: read: connection reset by peer ``` <p>Parameters: <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=8</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #100804 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery branch-release-23.1] - #100341 roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed [A-kv-replication C-bug C-test-failure GA-blocker O-roachtest O-robot T-disaster-recovery T-kv-replication branch-master branch-release-23.1] </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26944
non_priority
roachtest restore tpce aws nodes cpus failed roachtest restore tpce aws nodes cpus with on release test artifacts and logs in artifacts restore tpce aws nodes cpus run monitor go wait monitor failure monitor task failed read tcp read connection reset by peer parameters roachtest cloud aws roachtest cpu roachtest encrypted false roachtest fs roachtest localssd false roachtest ssd help see see same failure on other branches roachtest restore tpce aws nodes cpus failed roachtest restore tpce aws nodes cpus failed cc cockroachdb disaster recovery jira issue crdb
0
72,722
19,430,764,592
IssuesEvent
2021-12-21 11:40:12
openego/eGon-data
https://api.github.com/repos/openego/eGon-data
closed
Insert cost assumptions from NEP and technology-data to scenario_parameters
:building_construction: integration
Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39). Other marginal and capital costs are taken from technology-data repository
1.0
Insert cost assumptions from NEP and technology-data to scenario_parameters - Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39). Other marginal and capital costs are taken from technology-data repository
non_priority
insert cost assumptions from nep and technology data to scenario parameters fuel and prices will be added to the scenario parameters table according to table in nep version entwurf s other marginal and capital costs are taken from technology data repository
0
151,971
5,830,952,382
IssuesEvent
2017-05-08 18:08:06
idaholab/raven
https://api.github.com/repos/idaholab/raven
opened
Time Dependent Risk Importance Measures
improvement priority_normal
-------- Issue Description -------- ##### What did you expect to see happen? This issue refers to the fact that the actual Risk Importance Measures cannot be calculated if time dependent data is provided ##### What did you see instead? NA ##### Do you have a suggested fix for the development team? NA ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. Not a defect ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? improvement - [ ] 2. Is it tagged with a priority: critical, normal or minor? normal - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? NA - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. NA - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) Yes ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
Time Dependent Risk Importance Measures - -------- Issue Description -------- ##### What did you expect to see happen? This issue refers to the fact that the actual Risk Importance Measures cannot be calculated if time dependent data is provided ##### What did you see instead? NA ##### Do you have a suggested fix for the development team? NA ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. Not a defect ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? improvement - [ ] 2. Is it tagged with a priority: critical, normal or minor? normal - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? NA - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. NA - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) Yes ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
priority
time dependent risk importance measures issue description what did you expect to see happen this issue refers to the fact that the actual risk importance measures cannot be calculated if time dependent data is provided what did you see instead na do you have a suggested fix for the development team na please attach the input file s that generate this error the simpler the input the faster we can find the issue not a defect for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement improvement is it tagged with a priority critical normal or minor normal if it will impact requirements or requirements tests is it tagged with requirements na if it is a defect can it cause wrong results for users if so an email needs to be sent to the users na is a rationale provided such as explaining why the improvement is needed or why current code is wrong yes for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
1
178,061
6,598,429,575
IssuesEvent
2017-09-16 05:02:53
minishift/minishift
https://api.github.com/repos/minishift/minishift
closed
Installing CDK fails on Win10 with Hyper-V
kind/bug os/windows priority/major
As noted in #1381 I performed the minimal installation of minishift on my Win10 laptop. Thinking that the CDK might possibly resolve the issue, I downloaded the online installer and ran it. I encountered the following show-stopper error: ``` You have selected Red Hat Container Development Kit, which requires Oracle VirtualBox version 5.1.22 or higher. Please uninstall detected Oracle VirtualBox and restart Installer to continue. Please enable hardware virtualization support in BIOS for your platform. ``` This error message is nonsense. * I have enabled virtualization support in the BIOS, and Hyper-V is active. ![hyper-v](https://user-images.githubusercontent.com/4177499/30402943-6b38bdb0-98ad-11e7-90f7-d5f41fd96ca4.PNG) * Moreover, I have not even installed VirtualBox. ![no-virtualbox](https://user-images.githubusercontent.com/4177499/30402958-7eb870f6-98ad-11e7-84ef-03bfd8fe3161.PNG)
1.0
Installing CDK fails on Win10 with Hyper-V - As noted in #1381 I performed the minimal installation of minishift on my Win10 laptop. Thinking that the CDK might possibly resolve the issue, I downloaded the online installer and ran it. I encountered the following show-stopper error: ``` You have selected Red Hat Container Development Kit, which requires Oracle VirtualBox version 5.1.22 or higher. Please uninstall detected Oracle VirtualBox and restart Installer to continue. Please enable hardware virtualization support in BIOS for your platform. ``` This error message is nonsense. * I have enabled virtualization support in the BIOS, and Hyper-V is active. ![hyper-v](https://user-images.githubusercontent.com/4177499/30402943-6b38bdb0-98ad-11e7-90f7-d5f41fd96ca4.PNG) * Moreover, I have not even installed VirtualBox. ![no-virtualbox](https://user-images.githubusercontent.com/4177499/30402958-7eb870f6-98ad-11e7-84ef-03bfd8fe3161.PNG)
priority
installing cdk fails on with hyper v as noted in i performed the minimal installation of minishift on my laptop thinking that the cdk might possibly resolve the issue i downloaded the online installer and ran it i encountered the following show stopper error you have selected red hat container development kit which requires oracle virtualbox version or higher please uninstall detected oracle virtualbox and restart installer to continue please enable hardware virtualization support in bios for your platform this error message is nonsense i have enabled virtualization support in the bios and hyper v is active moreover i have not even installed virtualbox
1
279,991
8,676,724,382
IssuesEvent
2018-11-30 14:55:01
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
opened
User `userId` instead of `username` in collection saga/reducer/api
component: collections priority: p3
See also: #6609 --- In order to change the collection URLs to use user IDs instead of usernames, we need to change the variable names to reflect this upgrade. In this issue, we should simply rename variables and make sure Flow types and test cases are synchronized. **For QA:** please make sure collection pages load correctly and that no feature has been altered.
1.0
User `userId` instead of `username` in collection saga/reducer/api - See also: #6609 --- In order to change the collection URLs to use user IDs instead of usernames, we need to change the variable names to reflect this upgrade. In this issue, we should simply rename variables and make sure Flow types and test cases are synchronized. **For QA:** please make sure collection pages load correctly and that no feature has been altered.
priority
user userid instead of username in collection saga reducer api see also in order to change the collection urls to use user ids instead of usernames we need to change the variable names to reflect this upgrade in this issue we should simply rename variables and make sure flow types and test cases are synchronized for qa please make sure collection pages load correctly and that no feature has been altered
1
719,284
24,754,464,201
IssuesEvent
2022-10-21 16:20:24
massenergize/frontend-admin
https://api.github.com/repos/massenergize/frontend-admin
closed
DUP of 601 - In HTML Text field, default for links should be to open a new tab
enhancement priority 2
Currently the default is to use the existing tab. The dialog box says "window", We want the default to be new tab since that is correct most of the time, and most cadmins may miss this.
1.0
DUP of 601 - In HTML Text field, default for links should be to open a new tab - Currently the default is to use the existing tab. The dialog box says "window", We want the default to be new tab since that is correct most of the time, and most cadmins may miss this.
priority
dup of in html text field default for links should be to open a new tab currently the default is to use the existing tab the dialog box says window we want the default to be new tab since that is correct most of the time and most cadmins may miss this
1
95,036
8,528,684,965
IssuesEvent
2018-11-03 02:13:01
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: scaledata/jobcoordinator/nodes=3 failed
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/acd1250b15b7ed3c8938dfd53b8bc53bb53c578c Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stressrace TESTS=scaledata/jobcoordinator/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=999222&tab=buildLog ``` The test failed on master: test.go:639,cluster.go:1461,scaledata.go:126,scaledata.go:53: signal: interrupt ```
1.0
roachtest: scaledata/jobcoordinator/nodes=3 failed - SHA: https://github.com/cockroachdb/cockroach/commits/acd1250b15b7ed3c8938dfd53b8bc53bb53c578c Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stressrace TESTS=scaledata/jobcoordinator/nodes=3 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=999222&tab=buildLog ``` The test failed on master: test.go:639,cluster.go:1461,scaledata.go:126,scaledata.go:53: signal: interrupt ```
non_priority
roachtest scaledata jobcoordinator nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stressrace tests scaledata jobcoordinator nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master test go cluster go scaledata go scaledata go signal interrupt
0
565,735
16,768,378,905
IssuesEvent
2021-06-14 11:56:10
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.curseforge.com - site is not usable
browser-firefox-ios os-ios priority-normal status-needsinfo
<!-- @browser: Firefox iOS 33.1 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.curseforge.com/minecraft/mc-mods/sophisticated-backpacks **Browser / Version**: Firefox iOS 33.1 **Operating System**: iOS 14.4.2 **Tested Another Browser**: Yes Safari **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: When going to the site via a DuckDuckGo search result, a screen appears saying, “Checking your browser before accessing... Redirecting... DDoS protection by Cloudflare...” Then the screen goes blank, i.e., a white screen with no graphics or text. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/25c765a8-03e6-48aa-a913-7af2b26e047d.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.curseforge.com - site is not usable - <!-- @browser: Firefox iOS 33.1 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.curseforge.com/minecraft/mc-mods/sophisticated-backpacks **Browser / Version**: Firefox iOS 33.1 **Operating System**: iOS 14.4.2 **Tested Another Browser**: Yes Safari **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: When going to the site via a DuckDuckGo search result, a screen appears saying, “Checking your browser before accessing... Redirecting... DDoS protection by Cloudflare...” Then the screen goes blank, i.e., a white screen with no graphics or text. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/25c765a8-03e6-48aa-a913-7af2b26e047d.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox ios operating system ios tested another browser yes safari problem type site is not usable description page not loading correctly steps to reproduce when going to the site via a duckduckgo search result a screen appears saying “checking your browser before accessing redirecting ddos protection by cloudflare ” then the screen goes blank i e a white screen with no graphics or text view the screenshot img alt screenshot src browser configuration none from with ❤️
1
74,918
14,359,690,965
IssuesEvent
2020-11-30 15:56:00
code4romania/monitorizare-vot-ong
https://api.github.com/repos/code4romania/monitorizare-vot-ong
opened
[BUG] Incorrect polling station info displayed in details screen
angular autumn-2020 bug code4ro-hackdays good first issue help wanted high-priority responses
- Go to answers - Pick one answers and open the details screen - Go back to the answers page - Pick another answer from the list and open the details screen - Notice that on the top of the details page, the info about the previous polling station in displayed, instead of the info on the currently selected station ![Screenshot from 2020-11-30 17-53-41](https://user-images.githubusercontent.com/1358055/100632200-39be7b00-3335-11eb-971f-22abe6099dbb.png)
1.0
[BUG] Incorrect polling station info displayed in details screen - - Go to answers - Pick one answers and open the details screen - Go back to the answers page - Pick another answer from the list and open the details screen - Notice that on the top of the details page, the info about the previous polling station in displayed, instead of the info on the currently selected station ![Screenshot from 2020-11-30 17-53-41](https://user-images.githubusercontent.com/1358055/100632200-39be7b00-3335-11eb-971f-22abe6099dbb.png)
non_priority
incorrect polling station info displayed in details screen go to answers pick one answers and open the details screen go back to the answers page pick another answer from the list and open the details screen notice that on the top of the details page the info about the previous polling station in displayed instead of the info on the currently selected station
0
5,542
12,692,902,588
IssuesEvent
2020-06-22 01:12:38
Plant-for-the-Planet-org/treecounter-app
https://api.github.com/repos/Plant-for-the-Planet-org/treecounter-app
opened
Check component/container usage
architecture
Some components do not use associated containers. Check and maybe refactor them.
1.0
Check component/container usage - Some components do not use associated containers. Check and maybe refactor them.
non_priority
check component container usage some components do not use associated containers check and maybe refactor them
0
600,300
18,293,014,962
IssuesEvent
2021-10-05 17:16:19
FTBTeam/FTB-App
https://api.github.com/repos/FTBTeam/FTB-App
closed
[Feature request] Global Memory Settings Option
enhancement priority/low
**Is your feature request related to a problem? Please describe.** I download a new mod pack, but the memory settings are set to a fixed default (4gb). **Describe the solution you'd like** I'd like to modify the fixed initial default value in a global settings context. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[Feature request] Global Memory Settings Option - **Is your feature request related to a problem? Please describe.** I download a new mod pack, but the memory settings are set to a fixed default (4gb). **Describe the solution you'd like** I'd like to modify the fixed initial default value in a global settings context. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
priority
global memory settings option is your feature request related to a problem please describe i download a new mod pack but the memory settings are set to a fixed default describe the solution you d like i d like to modify the fixed initial default value in a global settings context describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
783,059
27,517,231,755
IssuesEvent
2023-03-06 12:49:58
NFT-Limited/nf-bugs
https://api.github.com/repos/NFT-Limited/nf-bugs
opened
[Collection/NFT] XDragon Robots
Collection/NFT Priority 1
**Describe the issue** Not showing any NFTs from the collection. Reset not helping. MP4 files. **Details** Help us to identify the collection by providing some infos: REQUIRED 1. Issuer address: rptAoCTeoJg24qgU3WnERE2ino7vzCMXFY 2. Collection Taxon: 523
1.0
[Collection/NFT] XDragon Robots - **Describe the issue** Not showing any NFTs from the collection. Reset not helping. MP4 files. **Details** Help us to identify the collection by providing some infos: REQUIRED 1. Issuer address: rptAoCTeoJg24qgU3WnERE2ino7vzCMXFY 2. Collection Taxon: 523
priority
xdragon robots describe the issue not showing any nfts from the collection reset not helping files details help us to identify the collection by providing some infos required issuer address collection taxon
1
169,000
6,393,181,592
IssuesEvent
2017-08-04 06:29:33
CanberraOceanRacingClub/namadgi3
https://api.github.com/repos/CanberraOceanRacingClub/namadgi3
closed
Upgrade and fit Gas detector with controller and Solenoid
priority 1: High
Authorised 6/6/17 located on aft of port galley cupboards who: Quays extra cost to contract
1.0
Upgrade and fit Gas detector with controller and Solenoid - Authorised 6/6/17 located on aft of port galley cupboards who: Quays extra cost to contract
priority
upgrade and fit gas detector with controller and solenoid authorised located on aft of port galley cupboards who quays extra cost to contract
1
308,655
26,620,832,715
IssuesEvent
2023-01-24 11:05:29
Kong/kubernetes-ingress-controller
https://api.github.com/repos/Kong/kubernetes-ingress-controller
closed
E2E test failure: data race in `TestDeployAllInOnePostgresWithMultipleReplicas`
bug area/tests
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Data race in tests: https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#logs ``` ================== WARNING: DATA RACE Read at 0x00c001f1e5b8 by goroutine 309: bytes.(*Buffer).String() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:65 +0xef github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder.func1() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:333 +0xb9 github.com/stretchr/testify/assert.Eventually.func1() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x39 Previous write at 0x00c001f1e5b8 by goroutine 307: bytes.(*Buffer).grow() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:145 +0x3c4 bytes.(*Buffer).ReadFrom() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:200 +0x65 io.copyBuffer() /opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:413 +0x1c5 io.Copy() /opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:386 +0x64 os/exec.(*Cmd).writerDescriptor.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:407 +0x3d os/exec.(*Cmd).Start.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:544 +0x35 os/exec.(*Cmd).Start.func2() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:545 +0x47 Goroutine 309 (running) created at: github.com/stretchr/testify/assert.Eventually() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x3a5 github.com/stretchr/testify/require.Eventually() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/require/require.go:3[61](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:62) +0xb8 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:326 +0x691 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1667 testing.tRunner() /opt/hostedtoolcache/go/1.19.5/x[64](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:65)/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1493 +0x47 Goroutine 307 (running) created at: os/exec.(*Cmd).Start() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:543 +0xd74 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:323 +0x450 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1[66](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:67)7 testing.tRunner() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:14[93](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:94) +0x47 ================== ``` ### Expected Behavior No data race. ### Steps To Reproduce _No response_ ### Kong Ingress Controller version _No response_ ### Kubernetes version ```shell 1.26 ``` ### Anything else? _No response_
1.0
E2E test failure: data race in `TestDeployAllInOnePostgresWithMultipleReplicas` - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Data race in tests: https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#logs ``` ================== WARNING: DATA RACE Read at 0x00c001f1e5b8 by goroutine 309: bytes.(*Buffer).String() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:65 +0xef github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder.func1() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:333 +0xb9 github.com/stretchr/testify/assert.Eventually.func1() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x39 Previous write at 0x00c001f1e5b8 by goroutine 307: bytes.(*Buffer).grow() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:145 +0x3c4 bytes.(*Buffer).ReadFrom() /opt/hostedtoolcache/go/1.19.5/x64/src/bytes/buffer.go:200 +0x65 io.copyBuffer() /opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:413 +0x1c5 io.Copy() /opt/hostedtoolcache/go/1.19.5/x64/src/io/io.go:386 +0x64 os/exec.(*Cmd).writerDescriptor.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:407 +0x3d os/exec.(*Cmd).Start.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:544 +0x35 os/exec.(*Cmd).Start.func2() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:545 +0x47 Goroutine 309 (running) created at: github.com/stretchr/testify/assert.Eventually() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/assert/assertions.go:1749 +0x3a5 github.com/stretchr/testify/require.Eventually() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.8.1/require/require.go:3[61](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:62) +0xb8 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:326 +0x691 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1667 testing.tRunner() /opt/hostedtoolcache/go/1.19.5/x[64](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:65)/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1493 +0x47 Goroutine 307 (running) created at: os/exec.(*Cmd).Start() /opt/hostedtoolcache/go/1.19.5/x64/src/os/exec/exec.go:543 +0xd74 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.startPortForwarder() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/utils_test.go:323 +0x450 github.com/kong/kubernetes-ingress-controller/v2/test/e2e.TestDeployAllInOnePostgresWithMultipleReplicas() /home/runner/work/kubernetes-ingress-controller/kubernetes-ingress-controller/test/e2e/all_in_one_test.go:273 +0x1[66](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:67)7 testing.tRunner() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.5/x64/src/testing/testing.go:14[93](https://github.com/Kong/kubernetes-ingress-controller/actions/runs/3993055061/jobs/6849453046#step:6:94) +0x47 ================== ``` ### Expected Behavior No data race. ### Steps To Reproduce _No response_ ### Kong Ingress Controller version _No response_ ### Kubernetes version ```shell 1.26 ``` ### Anything else? _No response_
non_priority
test failure data race in testdeployallinonepostgreswithmultiplereplicas is there an existing issue for this i have searched the existing issues current behavior data race in tests warning data race read at by goroutine bytes buffer string opt hostedtoolcache go src bytes buffer go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com stretchr testify assert eventually home runner go pkg mod github com stretchr testify assert assertions go previous write at by goroutine bytes buffer grow opt hostedtoolcache go src bytes buffer go bytes buffer readfrom opt hostedtoolcache go src bytes buffer go io copybuffer opt hostedtoolcache go src io io go io copy opt hostedtoolcache go src io io go os exec cmd writerdescriptor opt hostedtoolcache go src os exec exec go os exec cmd start opt hostedtoolcache go src os exec exec go os exec cmd start opt hostedtoolcache go src os exec exec go goroutine running created at github com stretchr testify assert eventually home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify require eventually home runner go pkg mod github com stretchr testify require require go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com kong kubernetes ingress controller test testdeployallinonepostgreswithmultiplereplicas home runner work kubernetes ingress controller kubernetes ingress controller test all in one test go testing trunner opt hostedtoolcache go x testing t run opt hostedtoolcache go src testing testing go goroutine running created at os exec cmd start opt hostedtoolcache go src os exec exec go github com kong kubernetes ingress controller test startportforwarder home runner work kubernetes ingress controller kubernetes ingress controller test utils test go github com kong kubernetes ingress controller test testdeployallinonepostgreswithmultiplereplicas home runner work kubernetes ingress controller kubernetes ingress controller test all in one test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go expected behavior no data race steps to reproduce no response kong ingress controller version no response kubernetes version shell anything else no response
0
361,461
10,709,035,456
IssuesEvent
2019-10-24 21:04:03
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
"Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate
affected: 2.4 auto-transferred bug category: calib3d priority: normal
Transferred from http://code.opencv.org/issues/2878 ``` || Daniel Danilin on 2013-03-10 16:12 || Priority: Normal || Affected: branch 'master' (2.4.9) || Category: calibration, 3d || Tracker: Bug || Difficulty: None || PR: || Platform: None / None ``` ## "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate ``` I'm getting the "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted." error message if I run stereo_calib example project. Sometimes the variable is 'A'. Here is the stack trace: <pre> > opencv_calib3d249d.dll!cvStereoCalibrate(const CvMat * _objectPoints, const CvMat * _imagePoints1, const CvMat * _imagePoints2, const CvMat * _npoints, CvMat * _cameraMatrix1, CvMat * _distCoeffs1, CvMat * _cameraMatrix2, CvMat * _distCoeffs2, CvSize imageSize, CvMat * matR, CvMat * matT, CvMat * matE, CvMat * matF, CvTermCriteria termCrit, int flags) Line 2071 C++ opencv_calib3d249d.dll!cv::stereoCalibrate(const cv::_InputArray & _objectPoints, const cv::_InputArray & _imagePoints1, const cv::_InputArray & _imagePoints2, const cv::_OutputArray & _cameraMatrix1, const cv::_OutputArray & _distCoeffs1, const cv::_OutputArray & _cameraMatrix2, const cv::_OutputArray & _distCoeffs2, cv::Size_<int> imageSize, const cv::_OutputArray & _Rmat, const cv::_OutputArray & _Tmat, const cv::_OutputArray & _Emat, const cv::_OutputArray & _Fmat, cv::TermCriteria criteria, int flags) Line 3305 C++ cpp-example-stereo_calib.exe!StereoCalib(const std::vector<std::basic_string<char,std::char_traits<char>,std::allocator<char> >,std::allocator<std::basic_string<char,std::char_traits<char>,std::allocator<char> > > > & imagelist, cv::Size_<int> boardSize, bool useCalibrated, bool showRectified) Line 183 C++ cpp-example-stereo_calib.exe!main(int argc, char * * argv) Line 404 C++ cpp-example-stereo_calib.exe!__tmainCRTStartup() Line 555 C cpp-example-stereo_calib.exe!mainCRTStartup() Line 371 C kernel32.dll!767f8543() Unknown [Frames below may be incorrect and/or missing, no symbols loaded for kernel32.dll] ntdll.dll!77eaac69() Unknown ntdll.dll!77eaac3c() Unknown </pre> I'm using origianl stereo_calib.xml and images from repository. I have Windows 8, CMake 2.8.10.2, Visual Studio 2012, Visual Studio 2010. I have configured the opencv in cmake without cuda. Tried to generate project files for VS 10 and VS 11, but got the same error. The sources code files are from 10.03.2013. ``` ## History
1.0
"Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate - Transferred from http://code.opencv.org/issues/2878 ``` || Daniel Danilin on 2013-03-10 16:12 || Priority: Normal || Affected: branch 'master' (2.4.9) || Category: calibration, 3d || Tracker: Bug || Difficulty: None || PR: || Platform: None / None ``` ## "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted" in stereoCalibrate ``` I'm getting the "Run-Time Check Failure #2 - Stack around the variable 'dk' was corrupted." error message if I run stereo_calib example project. Sometimes the variable is 'A'. Here is the stack trace: <pre> > opencv_calib3d249d.dll!cvStereoCalibrate(const CvMat * _objectPoints, const CvMat * _imagePoints1, const CvMat * _imagePoints2, const CvMat * _npoints, CvMat * _cameraMatrix1, CvMat * _distCoeffs1, CvMat * _cameraMatrix2, CvMat * _distCoeffs2, CvSize imageSize, CvMat * matR, CvMat * matT, CvMat * matE, CvMat * matF, CvTermCriteria termCrit, int flags) Line 2071 C++ opencv_calib3d249d.dll!cv::stereoCalibrate(const cv::_InputArray & _objectPoints, const cv::_InputArray & _imagePoints1, const cv::_InputArray & _imagePoints2, const cv::_OutputArray & _cameraMatrix1, const cv::_OutputArray & _distCoeffs1, const cv::_OutputArray & _cameraMatrix2, const cv::_OutputArray & _distCoeffs2, cv::Size_<int> imageSize, const cv::_OutputArray & _Rmat, const cv::_OutputArray & _Tmat, const cv::_OutputArray & _Emat, const cv::_OutputArray & _Fmat, cv::TermCriteria criteria, int flags) Line 3305 C++ cpp-example-stereo_calib.exe!StereoCalib(const std::vector<std::basic_string<char,std::char_traits<char>,std::allocator<char> >,std::allocator<std::basic_string<char,std::char_traits<char>,std::allocator<char> > > > & imagelist, cv::Size_<int> boardSize, bool useCalibrated, bool showRectified) Line 183 C++ cpp-example-stereo_calib.exe!main(int argc, char * * argv) Line 404 C++ cpp-example-stereo_calib.exe!__tmainCRTStartup() Line 555 C cpp-example-stereo_calib.exe!mainCRTStartup() Line 371 C kernel32.dll!767f8543() Unknown [Frames below may be incorrect and/or missing, no symbols loaded for kernel32.dll] ntdll.dll!77eaac69() Unknown ntdll.dll!77eaac3c() Unknown </pre> I'm using origianl stereo_calib.xml and images from repository. I have Windows 8, CMake 2.8.10.2, Visual Studio 2012, Visual Studio 2010. I have configured the opencv in cmake without cuda. Tried to generate project files for VS 10 and VS 11, but got the same error. The sources code files are from 10.03.2013. ``` ## History
priority
run time check failure stack around the variable dk was corrupted in stereocalibrate transferred from daniel danilin on priority normal affected branch master category calibration tracker bug difficulty none pr platform none none run time check failure stack around the variable dk was corrupted in stereocalibrate i m getting the run time check failure stack around the variable dk was corrupted error message if i run stereo calib example project sometimes the variable is a here is the stack trace opencv dll cvstereocalibrate const cvmat objectpoints const cvmat const cvmat const cvmat npoints cvmat cvmat cvmat cvmat cvsize imagesize cvmat matr cvmat matt cvmat mate cvmat matf cvtermcriteria termcrit int flags line c opencv dll cv stereocalibrate const cv inputarray objectpoints const cv inputarray const cv inputarray const cv outputarray const cv outputarray const cv outputarray const cv outputarray cv size imagesize const cv outputarray rmat const cv outputarray tmat const cv outputarray emat const cv outputarray fmat cv termcriteria criteria int flags line c cpp example stereo calib exe stereocalib const std vector std allocator std allocator std allocator imagelist cv size boardsize bool usecalibrated bool showrectified line c cpp example stereo calib exe main int argc char argv line c cpp example stereo calib exe tmaincrtstartup line c cpp example stereo calib exe maincrtstartup line c dll unknown ntdll dll unknown ntdll dll unknown i m using origianl stereo calib xml and images from repository i have windows cmake visual studio visual studio i have configured the opencv in cmake without cuda tried to generate project files for vs and vs but got the same error the sources code files are from history
1
50,833
21,420,076,246
IssuesEvent
2022-04-22 14:47:51
microsoft/BotFramework-Composer
https://api.github.com/repos/microsoft/BotFramework-Composer
closed
Bot Feamework Composer - Luis build failed: The model name { boolean } are reserved error when using form dialog feature
Type: Bug customer-reported Bot Services customer-replied-to
<!-- Please search for your feature request before creating a new one. > <!-- Complete the necessary portions of this template and delete the rest. --> ## Describe the bug Unable to start bot. Getting Luis build failed: The model name { boolean } are reserved error when trying to use form dialog feature in Bot Framework Composer(v2.1.2) <!-- Give a clear and concise description of what the bug is. --> ## Version v2.1.2 <!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). --> ## Screenshots ![image](https://user-images.githubusercontent.com/103985127/163936098-57dc1122-56de-422d-a0f3-e11549d4e484.png) <!-- If applicable, add screenshots/gif/video to help explain your problem. -->
1.0
Bot Feamework Composer - Luis build failed: The model name { boolean } are reserved error when using form dialog feature - <!-- Please search for your feature request before creating a new one. > <!-- Complete the necessary portions of this template and delete the rest. --> ## Describe the bug Unable to start bot. Getting Luis build failed: The model name { boolean } are reserved error when trying to use form dialog feature in Bot Framework Composer(v2.1.2) <!-- Give a clear and concise description of what the bug is. --> ## Version v2.1.2 <!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). --> ## Screenshots ![image](https://user-images.githubusercontent.com/103985127/163936098-57dc1122-56de-422d-a0f3-e11549d4e484.png) <!-- If applicable, add screenshots/gif/video to help explain your problem. -->
non_priority
bot feamework composer luis build failed the model name boolean are reserved error when using form dialog feature describe the bug unable to start bot getting luis build failed the model name boolean are reserved error when trying to use form dialog feature in bot framework composer version screenshots
0
128,114
17,420,887,630
IssuesEvent
2021-08-04 01:06:45
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
closed
Failure on initial `pulumi new` doesn't allow for re-use of project name on subsequent `new`
kind/bug resolution/by-design
Initial `pulumi new` failures wont allow for re-using of a project name. ## Expected behavior Cleanup after initial `new` fails, to force the user to re-run `new` after fixing the issues. ## Current behavior CLI expects you to run an `up` command ## Steps to reproduce 1. Run `$ pulumi new hello-aws-javascript` on a machine that doesn't have NPM installed 2. Go through onboarding, wait for the failure... 3. Install NPM, and re-run `$ pulumi new hello-aws-javascript` onboarding with the same project name. Get: `Sorry, '<insert project name>' is not a valid project name. A project with this name already exists.`
1.0
Failure on initial `pulumi new` doesn't allow for re-use of project name on subsequent `new` - Initial `pulumi new` failures wont allow for re-using of a project name. ## Expected behavior Cleanup after initial `new` fails, to force the user to re-run `new` after fixing the issues. ## Current behavior CLI expects you to run an `up` command ## Steps to reproduce 1. Run `$ pulumi new hello-aws-javascript` on a machine that doesn't have NPM installed 2. Go through onboarding, wait for the failure... 3. Install NPM, and re-run `$ pulumi new hello-aws-javascript` onboarding with the same project name. Get: `Sorry, '<insert project name>' is not a valid project name. A project with this name already exists.`
non_priority
failure on initial pulumi new doesn t allow for re use of project name on subsequent new initial pulumi new failures wont allow for re using of a project name expected behavior cleanup after initial new fails to force the user to re run new after fixing the issues current behavior cli expects you to run an up command steps to reproduce run pulumi new hello aws javascript on a machine that doesn t have npm installed go through onboarding wait for the failure install npm and re run pulumi new hello aws javascript onboarding with the same project name get sorry is not a valid project name a project with this name already exists
0
299,387
25,900,992,881
IssuesEvent
2022-12-15 05:38:39
openBackhaul/OperationKeyManagement
https://api.github.com/repos/openBackhaul/OperationKeyManagement
opened
release-number pattern update
testsuite_to_be_changed
Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'. Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator. Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder This scenario "multiple digit in a placeholder" can b:e added to following services: Service Layer - Acceptance :: Attribute correctness :: release-number checked? - [ ] /v1/bequeath-your-data-and-die - [ ] /v1/regard-application - [ ] /v1/disregard-application - [ ] /v1/list-applications - [ ] /v1/regard-updated-link
1.0
release-number pattern update - Pattern of release-number has been updated to '^([0-9]{1,2}).([0-9]{1,2}).([0-9]{1,2})$'. Already, testcases are available to check for too short release-number, too-long release-number, letters in release-number, sign in release-number , incorrect separator. Additionally a scenario can be added to test whether in each placeholder for a number, only two one or digits are allowed. In earlier release-number more than two digits are allowed in a placeholder This scenario "multiple digit in a placeholder" can b:e added to following services: Service Layer - Acceptance :: Attribute correctness :: release-number checked? - [ ] /v1/bequeath-your-data-and-die - [ ] /v1/regard-application - [ ] /v1/disregard-application - [ ] /v1/list-applications - [ ] /v1/regard-updated-link
non_priority
release number pattern update pattern of release number has been updated to already testcases are available to check for too short release number too long release number letters in release number sign in release number incorrect separator additionally a scenario can be added to test whether in each placeholder for a number only two one or digits are allowed in earlier release number more than two digits are allowed in a placeholder this scenario multiple digit in a placeholder can b e added to following services service layer acceptance attribute correctness release number checked bequeath your data and die regard application disregard application list applications regard updated link
0
349,071
24,932,809,719
IssuesEvent
2022-10-31 13:03:27
DuplosFidibuss/academic-time-planner
https://api.github.com/repos/DuplosFidibuss/academic-time-planner
closed
Documentation: describe test projects in methods section
documentation
The test plan project created in #23, the implementation of the loading in #24 and the test Toggl project created in #43 should be described in the methods section.
1.0
Documentation: describe test projects in methods section - The test plan project created in #23, the implementation of the loading in #24 and the test Toggl project created in #43 should be described in the methods section.
non_priority
documentation describe test projects in methods section the test plan project created in the implementation of the loading in and the test toggl project created in should be described in the methods section
0
247,813
26,735,469,134
IssuesEvent
2023-01-30 09:08:45
MValle21/oathkeeper
https://api.github.com/repos/MValle21/oathkeeper
opened
CVE-2022-25967 (High) detected in eta-1.12.0.tgz
security vulnerability
## CVE-2022-25967 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eta-1.12.0.tgz</b></p></summary> <p>Lightweight, fast, and powerful embedded JS template engine</p> <p>Library home page: <a href="https://registry.npmjs.org/eta/-/eta-1.12.0.tgz">https://registry.npmjs.org/eta/-/eta-1.12.0.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/eta/package.json</p> <p> Dependency Hierarchy: - core-2.0.0-alpha.415a7973f.tgz (Root Library) - :x: **eta-1.12.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MValle21/oathkeeper/commit/43c00a05bdb772edb5194a57f42ee834b37f3774">43c00a05bdb772edb5194a57f42ee834b37f3774</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package eta before 2.0.0 are vulnerable to Remote Code Execution (RCE) by overwriting template engine configuration variables with view options received from The Express render API. **Note:** This is exploitable only for users who are rendering templates with user-defined data. <p>Publish Date: 2023-01-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25967>CVE-2022-25967</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2023-01-30</p> <p>Fix Resolution: eta - 2.0.0</p> </p> </details> <p></p>
True
CVE-2022-25967 (High) detected in eta-1.12.0.tgz - ## CVE-2022-25967 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eta-1.12.0.tgz</b></p></summary> <p>Lightweight, fast, and powerful embedded JS template engine</p> <p>Library home page: <a href="https://registry.npmjs.org/eta/-/eta-1.12.0.tgz">https://registry.npmjs.org/eta/-/eta-1.12.0.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/eta/package.json</p> <p> Dependency Hierarchy: - core-2.0.0-alpha.415a7973f.tgz (Root Library) - :x: **eta-1.12.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MValle21/oathkeeper/commit/43c00a05bdb772edb5194a57f42ee834b37f3774">43c00a05bdb772edb5194a57f42ee834b37f3774</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package eta before 2.0.0 are vulnerable to Remote Code Execution (RCE) by overwriting template engine configuration variables with view options received from The Express render API. **Note:** This is exploitable only for users who are rendering templates with user-defined data. <p>Publish Date: 2023-01-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25967>CVE-2022-25967</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2023-01-30</p> <p>Fix Resolution: eta - 2.0.0</p> </p> </details> <p></p>
non_priority
cve high detected in eta tgz cve high severity vulnerability vulnerable library eta tgz lightweight fast and powerful embedded js template engine library home page a href path to dependency file docs package json path to vulnerable library docs node modules eta package json dependency hierarchy core alpha tgz root library x eta tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package eta before are vulnerable to remote code execution rce by overwriting template engine configuration variables with view options received from the express render api note this is exploitable only for users who are rendering templates with user defined data publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution eta
0
302,748
9,286,027,591
IssuesEvent
2019-03-21 09:15:15
CSBiology/BioFSharp
https://api.github.com/repos/CSBiology/BioFSharp
closed
[BUG] OboParser neglects first occurences of alt_ids
bug priority-medium project-IO up-for-grabs
### Description OboTerms can consists of serveral alt_id items. The OboParser just takes the last occurence of the alt_id keyword and discards previous ones. ### Repro steps Try to parse the following item (downloaded from http://geneontology.org/page/download-ontology): [Term] id: GO:0004748 name: ribonucleoside-diphosphate reductase activity, thioredoxin disulfide as acceptor namespace: molecular_function alt_id: GO:0016959 alt_id: GO:0016960 alt_id: GO:0016961 def: "Catalysis… comment: When thioredoxin… synonym: "2'-deoxyri… synonym: "2'-deoxyri… xref: EC:1.17.4.1 xref: MetaCyc:RIBONUCLEOSIDE-DIP-REDUCTI-RXN xref: RHEA:23252 is_a: GO:0061731 ! ribonucleoside-diphosphate reductase activity ### Expected behavior All alt_ids should be stored in the resulting OboTerm ### Actual behavior Just _GO:0016961_ is stored
1.0
[BUG] OboParser neglects first occurences of alt_ids - ### Description OboTerms can consists of serveral alt_id items. The OboParser just takes the last occurence of the alt_id keyword and discards previous ones. ### Repro steps Try to parse the following item (downloaded from http://geneontology.org/page/download-ontology): [Term] id: GO:0004748 name: ribonucleoside-diphosphate reductase activity, thioredoxin disulfide as acceptor namespace: molecular_function alt_id: GO:0016959 alt_id: GO:0016960 alt_id: GO:0016961 def: "Catalysis… comment: When thioredoxin… synonym: "2'-deoxyri… synonym: "2'-deoxyri… xref: EC:1.17.4.1 xref: MetaCyc:RIBONUCLEOSIDE-DIP-REDUCTI-RXN xref: RHEA:23252 is_a: GO:0061731 ! ribonucleoside-diphosphate reductase activity ### Expected behavior All alt_ids should be stored in the resulting OboTerm ### Actual behavior Just _GO:0016961_ is stored
priority
oboparser neglects first occurences of alt ids description oboterms can consists of serveral alt id items the oboparser just takes the last occurence of the alt id keyword and discards previous ones repro steps try to parse the following item downloaded from id go name ribonucleoside diphosphate reductase activity thioredoxin disulfide as acceptor namespace molecular function alt id go alt id go alt id go def catalysis… comment when thioredoxin… synonym deoxyri… synonym deoxyri… xref ec xref metacyc ribonucleoside dip reducti rxn xref rhea is a go ribonucleoside diphosphate reductase activity expected behavior all alt ids should be stored in the resulting oboterm actual behavior just go is stored
1
692,809
23,749,569,214
IssuesEvent
2022-08-31 19:14:06
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
Unlock my account please
Priority-High (Needed for work) Help wanted
I think I incorrectly entered my password too many times... can you please unlock my account? Is this the preferred path for fixing such situations? It's my first time... <img width="413" alt="Screen Shot 2022-08-31 at 10 50 56 AM" src="https://user-images.githubusercontent.com/17605945/187759150-ebb22fc6-0144-41c2-b78c-801a2bca068e.png">
1.0
Unlock my account please - I think I incorrectly entered my password too many times... can you please unlock my account? Is this the preferred path for fixing such situations? It's my first time... <img width="413" alt="Screen Shot 2022-08-31 at 10 50 56 AM" src="https://user-images.githubusercontent.com/17605945/187759150-ebb22fc6-0144-41c2-b78c-801a2bca068e.png">
priority
unlock my account please i think i incorrectly entered my password too many times can you please unlock my account is this the preferred path for fixing such situations it s my first time img width alt screen shot at am src
1
229,436
7,574,701,132
IssuesEvent
2018-04-23 21:53:31
kubernetes/kubeadm
https://api.github.com/repos/kubernetes/kubeadm
closed
Nodes registered with FQDN, but labelled with short hostname
kind/bug priority/important-soon
## What keywords did you search in kubeadm issues before filing this one? * hostname ## Is this a BUG REPORT or FEATURE REQUEST? Choose one: BUG REPORT ## Versions **kubeadm version**: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:40:06Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} **Environment**: - **Kubernetes version**: 1.9.1 - **Cloud provider or hardware configuration**: AWS - **OS**: Amazon Linux 2.0 (2017.12) LTS Release Candidate - **Kernel**: Linux ip-10-67-12-187 4.9.70-2.243.amzn2.x86_64 #1 SMP Thu Jan 4 03:49:00 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux - **Others**: ## What happened? Build cluster with the following config file: ```yaml api: advertiseAddress: 10.67.12.187 bindPort: 6443 apiServerCertSANs: - master.dev.smartsheet.com - apiserver.dev.smartsheet.com authorizationModes: - Node - RBAC certificatesDir: /etc/kubernetes/pki cloudProvider: aws etcd: caFile: "" certFile: "" dataDir: /var/lib/etcd endpoints: null image: "" keyFile: "" imageRepository: gcr.io/google_containers kubeProxy: config: bindAddress: 0.0.0.0 clientConnection: acceptContentTypes: "" burst: 10 contentType: application/vnd.kubernetes.protobuf kubeconfig: /var/lib/kube-proxy/kubeconfig.conf qps: 5 clusterCIDR: 192.168.0.0/16 configSyncPeriod: 15m0s conntrack: max: null maxPerCore: 32768 min: 131072 tcpCloseWaitTimeout: 1h0m0s tcpEstablishedTimeout: 24h0m0s enableProfiling: false featureGates: "" healthzBindAddress: 0.0.0.0:10256 hostnameOverride: "" iptables: masqueradeAll: false masqueradeBit: 14 minSyncPeriod: 0s syncPeriod: 30s ipvs: minSyncPeriod: 0s scheduler: "" syncPeriod: 30s metricsBindAddress: 127.0.0.1:10249 mode: "" oomScoreAdj: -999 portRange: "" resourceContainer: /kube-proxy udpTimeoutMilliseconds: 250ms kubeletConfiguration: {} kubernetesVersion: v1.9.1 networking: dnsDomain: cluster.local podSubnet: 192.168.0.0/16 serviceSubnet: 172.20.0.0/22 nodeName: ip-10-67-12-187.dev.smartsheet.com token: "" tokenTTL: 0s unifiedControlPlaneImage: "" ``` With fully resolvable DNS records for hosts (both forward and reverse records) the nodes register themselves with their FQDN. ``` NAME STATUS ROLES AGE VERSION LABELS ip-10-67-12-187.dev.smartsheet.com Ready master 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-12-187,node-role.kubernetes.io/master= ip-10-67-13-11.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-11 ip-10-67-13-178.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-178 ip-10-67-14-40.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-40 ip-10-67-14-66.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-66 ip-10-67-15-129.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-129 ip-10-67-15-22.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-22 ``` Some applications target themselves (Rook mon processes) to a specific node using the `kubernetes.io/hostname` label, but use the node name as it is registered formally (i.e. FQDN). ## What you expected to happen? The `kubernetes.io/hostname` label's value to match the node name. ## How to reproduce it (as minimally and precisely as possible)? Create a cluster where nodes resolve completely in DNS (both forward and reverse records). Join nodes to the cluster and they should be registered using the node's FQDN. Inspect the `kubernetes.io/hostname` label and its value will be the short hostname instead of the registered node name. ## Anything else we need to know? I think that the default should be to register the node and set hostname label to the short hostname, but there should be a setting in at least the config file (maybe a CLI switch also) that allows the admin to specify that the FQDN be used in both cases.
1.0
Nodes registered with FQDN, but labelled with short hostname - ## What keywords did you search in kubeadm issues before filing this one? * hostname ## Is this a BUG REPORT or FEATURE REQUEST? Choose one: BUG REPORT ## Versions **kubeadm version**: &version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:40:06Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"linux/amd64"} **Environment**: - **Kubernetes version**: 1.9.1 - **Cloud provider or hardware configuration**: AWS - **OS**: Amazon Linux 2.0 (2017.12) LTS Release Candidate - **Kernel**: Linux ip-10-67-12-187 4.9.70-2.243.amzn2.x86_64 #1 SMP Thu Jan 4 03:49:00 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux - **Others**: ## What happened? Build cluster with the following config file: ```yaml api: advertiseAddress: 10.67.12.187 bindPort: 6443 apiServerCertSANs: - master.dev.smartsheet.com - apiserver.dev.smartsheet.com authorizationModes: - Node - RBAC certificatesDir: /etc/kubernetes/pki cloudProvider: aws etcd: caFile: "" certFile: "" dataDir: /var/lib/etcd endpoints: null image: "" keyFile: "" imageRepository: gcr.io/google_containers kubeProxy: config: bindAddress: 0.0.0.0 clientConnection: acceptContentTypes: "" burst: 10 contentType: application/vnd.kubernetes.protobuf kubeconfig: /var/lib/kube-proxy/kubeconfig.conf qps: 5 clusterCIDR: 192.168.0.0/16 configSyncPeriod: 15m0s conntrack: max: null maxPerCore: 32768 min: 131072 tcpCloseWaitTimeout: 1h0m0s tcpEstablishedTimeout: 24h0m0s enableProfiling: false featureGates: "" healthzBindAddress: 0.0.0.0:10256 hostnameOverride: "" iptables: masqueradeAll: false masqueradeBit: 14 minSyncPeriod: 0s syncPeriod: 30s ipvs: minSyncPeriod: 0s scheduler: "" syncPeriod: 30s metricsBindAddress: 127.0.0.1:10249 mode: "" oomScoreAdj: -999 portRange: "" resourceContainer: /kube-proxy udpTimeoutMilliseconds: 250ms kubeletConfiguration: {} kubernetesVersion: v1.9.1 networking: dnsDomain: cluster.local podSubnet: 192.168.0.0/16 serviceSubnet: 172.20.0.0/22 nodeName: ip-10-67-12-187.dev.smartsheet.com token: "" tokenTTL: 0s unifiedControlPlaneImage: "" ``` With fully resolvable DNS records for hosts (both forward and reverse records) the nodes register themselves with their FQDN. ``` NAME STATUS ROLES AGE VERSION LABELS ip-10-67-12-187.dev.smartsheet.com Ready master 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-12-187,node-role.kubernetes.io/master= ip-10-67-13-11.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-11 ip-10-67-13-178.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2a,kubernetes.io/hostname=ip-10-67-13-178 ip-10-67-14-40.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-40 ip-10-67-14-66.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2b,kubernetes.io/hostname=ip-10-67-14-66 ip-10-67-15-129.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-129 ip-10-67-15-22.dev.smartsheet.com Ready <none> 1h v1.9.1 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=m5.large,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=us-west-2,failure-domain.beta.kubernetes.io/zone=us-west-2c,kubernetes.io/hostname=ip-10-67-15-22 ``` Some applications target themselves (Rook mon processes) to a specific node using the `kubernetes.io/hostname` label, but use the node name as it is registered formally (i.e. FQDN). ## What you expected to happen? The `kubernetes.io/hostname` label's value to match the node name. ## How to reproduce it (as minimally and precisely as possible)? Create a cluster where nodes resolve completely in DNS (both forward and reverse records). Join nodes to the cluster and they should be registered using the node's FQDN. Inspect the `kubernetes.io/hostname` label and its value will be the short hostname instead of the registered node name. ## Anything else we need to know? I think that the default should be to register the node and set hostname label to the short hostname, but there should be a setting in at least the config file (maybe a CLI switch also) that allows the admin to specify that the FQDN be used in both cases.
priority
nodes registered with fqdn but labelled with short hostname what keywords did you search in kubeadm issues before filing this one hostname is this a bug report or feature request choose one bug report versions kubeadm version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux environment kubernetes version cloud provider or hardware configuration aws os amazon linux lts release candidate kernel linux ip smp thu jan utc gnu linux others what happened build cluster with the following config file yaml api advertiseaddress bindport apiservercertsans master dev smartsheet com apiserver dev smartsheet com authorizationmodes node rbac certificatesdir etc kubernetes pki cloudprovider aws etcd cafile certfile datadir var lib etcd endpoints null image keyfile imagerepository gcr io google containers kubeproxy config bindaddress clientconnection acceptcontenttypes burst contenttype application vnd kubernetes protobuf kubeconfig var lib kube proxy kubeconfig conf qps clustercidr configsyncperiod conntrack max null maxpercore min tcpclosewaittimeout tcpestablishedtimeout enableprofiling false featuregates healthzbindaddress hostnameoverride iptables masqueradeall false masqueradebit minsyncperiod syncperiod ipvs minsyncperiod scheduler syncperiod metricsbindaddress mode oomscoreadj portrange resourcecontainer kube proxy udptimeoutmilliseconds kubeletconfiguration kubernetesversion networking dnsdomain cluster local podsubnet servicesubnet nodename ip dev smartsheet com token tokenttl unifiedcontrolplaneimage with fully resolvable dns records for hosts both forward and reverse records the nodes register themselves with their fqdn name status roles age version labels ip dev smartsheet com ready master beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip node role kubernetes io master ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip ip dev smartsheet com ready beta kubernetes io arch beta kubernetes io instance type large beta kubernetes io os linux failure domain beta kubernetes io region us west failure domain beta kubernetes io zone us west kubernetes io hostname ip some applications target themselves rook mon processes to a specific node using the kubernetes io hostname label but use the node name as it is registered formally i e fqdn what you expected to happen the kubernetes io hostname label s value to match the node name how to reproduce it as minimally and precisely as possible create a cluster where nodes resolve completely in dns both forward and reverse records join nodes to the cluster and they should be registered using the node s fqdn inspect the kubernetes io hostname label and its value will be the short hostname instead of the registered node name anything else we need to know i think that the default should be to register the node and set hostname label to the short hostname but there should be a setting in at least the config file maybe a cli switch also that allows the admin to specify that the fqdn be used in both cases
1
341,761
30,598,830,235
IssuesEvent
2023-07-22 05:12:55
hajimehoshi/ebiten
https://api.github.com/repos/hajimehoshi/ebiten
closed
.builds, .github/workflows: remove installing development libs for OpenGL
os:linux os:freebsd test
### Operating System - [ ] Windows - [ ] macOS - [X] Linux - [X] FreeBSD - [ ] OpenBSD - [ ] Android - [ ] iOS - [ ] Nintendo Switch - [ ] Xbox - [ ] Web Browsers ### What feature would you like to be added? After https://github.com/hajimehoshi/ebiten/commit/0324db5657e133f85d967ab7b7eab98600935288, we no longer need to install libraries for OpenGL when compiling. Let's simplify yml files. ### Why is this needed? _No response_
1.0
.builds, .github/workflows: remove installing development libs for OpenGL - ### Operating System - [ ] Windows - [ ] macOS - [X] Linux - [X] FreeBSD - [ ] OpenBSD - [ ] Android - [ ] iOS - [ ] Nintendo Switch - [ ] Xbox - [ ] Web Browsers ### What feature would you like to be added? After https://github.com/hajimehoshi/ebiten/commit/0324db5657e133f85d967ab7b7eab98600935288, we no longer need to install libraries for OpenGL when compiling. Let's simplify yml files. ### Why is this needed? _No response_
non_priority
builds github workflows remove installing development libs for opengl operating system windows macos linux freebsd openbsd android ios nintendo switch xbox web browsers what feature would you like to be added after we no longer need to install libraries for opengl when compiling let s simplify yml files why is this needed no response
0
230,031
25,409,538,999
IssuesEvent
2022-11-22 17:45:04
elastic/beats
https://api.github.com/repos/elastic/beats
opened
[Auditbeat] Prepare System Package to be GA
enhancement Auditbeat Team:Security-External Integrations
**Describe the enhancement:** We previously had plans to deprecate the `system.package` dataset because in within the Elastic ecosystem we expected that `osquerybeat` could provide this data using various tables like `rpm_package` and `deb_packages`. However, because it cannot provide deltas between previous state and current state the data is not that useful on its own. So we want to improve the Auditbeat system.package dataset such that it can be supported as GA feature and exposed through Elastic Agent. - [ ] Create a beta Fleet integration for package monitoring that wraps the Auditbeat system.package dataset. - [ ] Add fsnotify support for triggering updates in near real-time. - [ ] Migrate to using FlatBuffers and the means of encoding persistent state. This will ensure we have a stable schema for this data between versions and avoid accidental breakages. - [ ] Update documentation for Auditbeat package dataset. - [ ] Update documentation related to Auditbeat to Agent migration specifically related to `system.package`. For reference this was added in https://github.com/elastic/observability-docs/pull/2270. - [ ] Document the Fleet integration as GA using at least version 1.0.0 for the package. **Describe a specific use case for the enhancement or feature:** ### References - Replaces https://github.com/elastic/beats/issues/33481
True
[Auditbeat] Prepare System Package to be GA - **Describe the enhancement:** We previously had plans to deprecate the `system.package` dataset because in within the Elastic ecosystem we expected that `osquerybeat` could provide this data using various tables like `rpm_package` and `deb_packages`. However, because it cannot provide deltas between previous state and current state the data is not that useful on its own. So we want to improve the Auditbeat system.package dataset such that it can be supported as GA feature and exposed through Elastic Agent. - [ ] Create a beta Fleet integration for package monitoring that wraps the Auditbeat system.package dataset. - [ ] Add fsnotify support for triggering updates in near real-time. - [ ] Migrate to using FlatBuffers and the means of encoding persistent state. This will ensure we have a stable schema for this data between versions and avoid accidental breakages. - [ ] Update documentation for Auditbeat package dataset. - [ ] Update documentation related to Auditbeat to Agent migration specifically related to `system.package`. For reference this was added in https://github.com/elastic/observability-docs/pull/2270. - [ ] Document the Fleet integration as GA using at least version 1.0.0 for the package. **Describe a specific use case for the enhancement or feature:** ### References - Replaces https://github.com/elastic/beats/issues/33481
non_priority
prepare system package to be ga describe the enhancement we previously had plans to deprecate the system package dataset because in within the elastic ecosystem we expected that osquerybeat could provide this data using various tables like rpm package and deb packages however because it cannot provide deltas between previous state and current state the data is not that useful on its own so we want to improve the auditbeat system package dataset such that it can be supported as ga feature and exposed through elastic agent create a beta fleet integration for package monitoring that wraps the auditbeat system package dataset add fsnotify support for triggering updates in near real time migrate to using flatbuffers and the means of encoding persistent state this will ensure we have a stable schema for this data between versions and avoid accidental breakages update documentation for auditbeat package dataset update documentation related to auditbeat to agent migration specifically related to system package for reference this was added in document the fleet integration as ga using at least version for the package describe a specific use case for the enhancement or feature references replaces
0
202,136
15,821,453,110
IssuesEvent
2021-04-05 20:33:16
leaf-ai/studio-go-runner
https://api.github.com/repos/leaf-ai/studio-go-runner
closed
Document the Kubernetes Auto-Scaler smoke test for AWS users using eksctl
documentation enhancement
As a cluster operator I want to be able to manually test the Kubernetes node scaling In order that during installation baseline ASG can be verified as working
1.0
Document the Kubernetes Auto-Scaler smoke test for AWS users using eksctl - As a cluster operator I want to be able to manually test the Kubernetes node scaling In order that during installation baseline ASG can be verified as working
non_priority
document the kubernetes auto scaler smoke test for aws users using eksctl as a cluster operator i want to be able to manually test the kubernetes node scaling in order that during installation baseline asg can be verified as working
0
413,800
27,969,990,244
IssuesEvent
2023-03-25 00:25:28
Webbutvecklare-CL/cl-sektionen.se
https://api.github.com/repos/Webbutvecklare-CL/cl-sektionen.se
closed
Utseende på textsidor
documentation
Just nu har ex VFU och Illabehandling två olika utseende. En ligger på bakgrunden och en ligger som en artikel. Exprimentera och se vilken som fungerar bäst.
1.0
Utseende på textsidor - Just nu har ex VFU och Illabehandling två olika utseende. En ligger på bakgrunden och en ligger som en artikel. Exprimentera och se vilken som fungerar bäst.
non_priority
utseende på textsidor just nu har ex vfu och illabehandling två olika utseende en ligger på bakgrunden och en ligger som en artikel exprimentera och se vilken som fungerar bäst
0
617,031
19,339,422,120
IssuesEvent
2021-12-15 01:29:48
spaceone-dev/spaceone-design-system
https://api.github.com/repos/spaceone-dev/spaceone-design-system
closed
[Data Table] Update slot props
bug Priority: Low
**Describe the bug** slot으로 받는 `index` 값이 열 index인지 행 index인지 구분하기가 힘듦. 명확한 구분을 위해 `index`라 되어 있는 값들을 `colIndex`와 `rowIndex`이름으로 추가. `index`는 이후에 삭제.
1.0
[Data Table] Update slot props - **Describe the bug** slot으로 받는 `index` 값이 열 index인지 행 index인지 구분하기가 힘듦. 명확한 구분을 위해 `index`라 되어 있는 값들을 `colIndex`와 `rowIndex`이름으로 추가. `index`는 이후에 삭제.
priority
update slot props describe the bug slot으로 받는 index 값이 열 index인지 행 index인지 구분하기가 힘듦 명확한 구분을 위해 index 라 되어 있는 값들을 colindex 와 rowindex 이름으로 추가 index 는 이후에 삭제
1
395,289
11,683,343,289
IssuesEvent
2020-03-05 03:04:01
SnorriDev/thoth
https://api.github.com/repos/SnorriDev/thoth
closed
Move some input checking to a new broadcastBinding event.
low priority wontfix
See Player.java. Some input checking can apparently be moved to a new Event.
1.0
Move some input checking to a new broadcastBinding event. - See Player.java. Some input checking can apparently be moved to a new Event.
priority
move some input checking to a new broadcastbinding event see player java some input checking can apparently be moved to a new event
1
32,077
6,706,740,103
IssuesEvent
2017-10-12 08:23:03
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Issues with Postgres RETURNING with field expressions
C: Functionality P: Medium T: Defect
In the RETURNING clause, Postgres supports "any expression using the table's columns is allowed". jOOQ appears to partially support this, but the behavior varies in surprising ways based on how the library is used. I'd expect this to either be supported and always work or be unsupported and never work, rather than the current behavior where it sometimes works. I've noticed two issues related to this, so I'm documenting them both here — one involving the generated SQL, and one involving the fields on the resulting Records. ### Expected behavior and actual behavior: #### Generated SQL Issue Given the following code: ```java DSLContext dsl = DSL.using(getConnection(), SQLDialect.POSTGRES); Table<Record> table = new TableImpl<>(DSL.name("test")); Field<String> field = DSL.field("field", String.class); Field<Integer> aliasedFieldAsInt = field.cast(Integer.class).as("int_field"); System.out.println(dsl.insertInto(table).columns(field).values("1").returning(aliasedFieldAsInt).getSQL()); ``` I would expect the output to be: ```sql insert into "test" (field) values (?) returning cast(field as int) as "int_field" ``` However, it is instead: ```sql insert into "test" (field) values (?) returning "int_field" ``` Note that if I declare the field without an alias, it works as expected: ```java Field<Integer> fieldAsInt = field.cast(Integer.class); System.out.println(dsl.insertInto(table).columns(field).values("1").returning(fieldAsInt).getSQL()); ``` The above returns the expected value: ```sql insert into "public"."test" (field) values (?) returning id, cast(field as int) ``` #### asMap() Issue A second issue I discovered with RETURNING is that calling `fields()` and `intoMap()` on the result works as expected when the table is manually created as above. However, if the table is generated using the schema generation (or the table is manually generated in a manner consistent with the schema generation, namely using the static `createField` method on `AbstractTable`), the `fields()` method returns the list of fields on that table, not the list of fields that were declared in the `returning` method. For example, given a table named "test" in schema "public" with columns "id" and "field" and a generated `Test` table class, and the following code: ```java Table<Record> table = Test.TEST; Field<String> field = DSL.field("field", String.class); InsertResultStep<Record> returningAsInt = dsl.insertInto(table).columns(field).values("1").returning(field, fieldAsInt); System.out.println(Arrays.toString(returningAsInt.fetchOne().fields())); System.out.println(returningAsInt.fetchOne().intoMap()); ``` I would expect the output to be analogous to the that of an equivalent select statement: ``` [field, cast(field as integer)] {field=1, cast=1} ``` But it is instead: ``` ["public"."test"."id", "public"."test"."field"] {id=null, field=1} ```
1.0
Issues with Postgres RETURNING with field expressions - In the RETURNING clause, Postgres supports "any expression using the table's columns is allowed". jOOQ appears to partially support this, but the behavior varies in surprising ways based on how the library is used. I'd expect this to either be supported and always work or be unsupported and never work, rather than the current behavior where it sometimes works. I've noticed two issues related to this, so I'm documenting them both here — one involving the generated SQL, and one involving the fields on the resulting Records. ### Expected behavior and actual behavior: #### Generated SQL Issue Given the following code: ```java DSLContext dsl = DSL.using(getConnection(), SQLDialect.POSTGRES); Table<Record> table = new TableImpl<>(DSL.name("test")); Field<String> field = DSL.field("field", String.class); Field<Integer> aliasedFieldAsInt = field.cast(Integer.class).as("int_field"); System.out.println(dsl.insertInto(table).columns(field).values("1").returning(aliasedFieldAsInt).getSQL()); ``` I would expect the output to be: ```sql insert into "test" (field) values (?) returning cast(field as int) as "int_field" ``` However, it is instead: ```sql insert into "test" (field) values (?) returning "int_field" ``` Note that if I declare the field without an alias, it works as expected: ```java Field<Integer> fieldAsInt = field.cast(Integer.class); System.out.println(dsl.insertInto(table).columns(field).values("1").returning(fieldAsInt).getSQL()); ``` The above returns the expected value: ```sql insert into "public"."test" (field) values (?) returning id, cast(field as int) ``` #### asMap() Issue A second issue I discovered with RETURNING is that calling `fields()` and `intoMap()` on the result works as expected when the table is manually created as above. However, if the table is generated using the schema generation (or the table is manually generated in a manner consistent with the schema generation, namely using the static `createField` method on `AbstractTable`), the `fields()` method returns the list of fields on that table, not the list of fields that were declared in the `returning` method. For example, given a table named "test" in schema "public" with columns "id" and "field" and a generated `Test` table class, and the following code: ```java Table<Record> table = Test.TEST; Field<String> field = DSL.field("field", String.class); InsertResultStep<Record> returningAsInt = dsl.insertInto(table).columns(field).values("1").returning(field, fieldAsInt); System.out.println(Arrays.toString(returningAsInt.fetchOne().fields())); System.out.println(returningAsInt.fetchOne().intoMap()); ``` I would expect the output to be analogous to the that of an equivalent select statement: ``` [field, cast(field as integer)] {field=1, cast=1} ``` But it is instead: ``` ["public"."test"."id", "public"."test"."field"] {id=null, field=1} ```
non_priority
issues with postgres returning with field expressions in the returning clause postgres supports any expression using the table s columns is allowed jooq appears to partially support this but the behavior varies in surprising ways based on how the library is used i d expect this to either be supported and always work or be unsupported and never work rather than the current behavior where it sometimes works i ve noticed two issues related to this so i m documenting them both here — one involving the generated sql and one involving the fields on the resulting records expected behavior and actual behavior generated sql issue given the following code java dslcontext dsl dsl using getconnection sqldialect postgres table table new tableimpl dsl name test field field dsl field field string class field aliasedfieldasint field cast integer class as int field system out println dsl insertinto table columns field values returning aliasedfieldasint getsql i would expect the output to be sql insert into test field values returning cast field as int as int field however it is instead sql insert into test field values returning int field note that if i declare the field without an alias it works as expected java field fieldasint field cast integer class system out println dsl insertinto table columns field values returning fieldasint getsql the above returns the expected value sql insert into public test field values returning id cast field as int asmap issue a second issue i discovered with returning is that calling fields and intomap on the result works as expected when the table is manually created as above however if the table is generated using the schema generation or the table is manually generated in a manner consistent with the schema generation namely using the static createfield method on abstracttable the fields method returns the list of fields on that table not the list of fields that were declared in the returning method for example given a table named test in schema public with columns id and field and a generated test table class and the following code java table table test test field field dsl field field string class insertresultstep returningasint dsl insertinto table columns field values returning field fieldasint system out println arrays tostring returningasint fetchone fields system out println returningasint fetchone intomap i would expect the output to be analogous to the that of an equivalent select statement field cast but it is instead id null field
0
434,248
12,515,947,868
IssuesEvent
2020-06-03 08:34:58
eclipse/codewind
https://api.github.com/repos/eclipse/codewind
closed
SVT:Hybrid:Load test request status stuck at "Requested" state
area/portal kind/bug priority/hot
<!-- Please fill out the following form to report a bug. If some fields do not apply to your situation, feel free to skip them.--> **Codewind version:** 0.12.0 **OS:** CoreOS/RHEL **Che version:** **IDE extension version:** 0.12.0 **IDE version:** Eclipse 2019-09 **Kubernetes cluster:** OCP 4.3 / OCP 4.4 GA **Description:** I created Appsody open liberty project and tried to submit load test request and it got stuck at "Requested" state and it never changed after that. We have seen this problem during 0.10.0 and 0.11.0 releases and this has been fixed and we could not reproduce during 0.11.0 reelase. Looks like this problem still exist and not completely addressed. Here are all the logs from my cluster where I have seen this problem. ![image](https://user-images.githubusercontent.com/713849/81566892-8591de80-9369-11ea-9f7a-9588e40575c7.png) [appsodyolhybcw012ocp44ga.txt](https://github.com/eclipse/codewind/files/4609859/appsodyolhybcw012ocp44ga.txt) [gatekeeper.txt](https://github.com/eclipse/codewind/files/4609860/gatekeeper.txt) [performance.txt](https://github.com/eclipse/codewind/files/4609861/performance.txt) [pfe.txt](https://github.com/eclipse/codewind/files/4609862/pfe.txt) **Steps to reproduce:** 1. Install Codewind 0.12.0 on Hybrid. 2. Create Appsody open liberty project. 3. Request load test after project started. 4. Load test stuck at "Requested" state. **Workaround:** <!-- Did you find a way to work around the bug? If so, please describe how you worked around it.--> Current workaround is to manually refresh your browser page
1.0
SVT:Hybrid:Load test request status stuck at "Requested" state - <!-- Please fill out the following form to report a bug. If some fields do not apply to your situation, feel free to skip them.--> **Codewind version:** 0.12.0 **OS:** CoreOS/RHEL **Che version:** **IDE extension version:** 0.12.0 **IDE version:** Eclipse 2019-09 **Kubernetes cluster:** OCP 4.3 / OCP 4.4 GA **Description:** I created Appsody open liberty project and tried to submit load test request and it got stuck at "Requested" state and it never changed after that. We have seen this problem during 0.10.0 and 0.11.0 releases and this has been fixed and we could not reproduce during 0.11.0 reelase. Looks like this problem still exist and not completely addressed. Here are all the logs from my cluster where I have seen this problem. ![image](https://user-images.githubusercontent.com/713849/81566892-8591de80-9369-11ea-9f7a-9588e40575c7.png) [appsodyolhybcw012ocp44ga.txt](https://github.com/eclipse/codewind/files/4609859/appsodyolhybcw012ocp44ga.txt) [gatekeeper.txt](https://github.com/eclipse/codewind/files/4609860/gatekeeper.txt) [performance.txt](https://github.com/eclipse/codewind/files/4609861/performance.txt) [pfe.txt](https://github.com/eclipse/codewind/files/4609862/pfe.txt) **Steps to reproduce:** 1. Install Codewind 0.12.0 on Hybrid. 2. Create Appsody open liberty project. 3. Request load test after project started. 4. Load test stuck at "Requested" state. **Workaround:** <!-- Did you find a way to work around the bug? If so, please describe how you worked around it.--> Current workaround is to manually refresh your browser page
priority
svt hybrid load test request status stuck at requested state codewind version os coreos rhel che version ide extension version ide version eclipse kubernetes cluster ocp ocp ga description i created appsody open liberty project and tried to submit load test request and it got stuck at requested state and it never changed after that we have seen this problem during and releases and this has been fixed and we could not reproduce during reelase looks like this problem still exist and not completely addressed here are all the logs from my cluster where i have seen this problem steps to reproduce install codewind on hybrid create appsody open liberty project request load test after project started load test stuck at requested state workaround current workaround is to manually refresh your browser page
1
56,320
31,863,180,325
IssuesEvent
2023-09-15 12:27:48
levibostian/action-bump-major-tag
https://api.github.com/repos/levibostian/action-bump-major-tag
opened
Reduce network calls for getting tags list
performance
# Expected When getting the repo's git tags, reduce down the number of network requests as little as possible. # Actual Today, the action does paging to get all of the repo's git tags. # Ideas * Use github's graphql api to specify in more detail what types of tags we want. * Use caching.
True
Reduce network calls for getting tags list - # Expected When getting the repo's git tags, reduce down the number of network requests as little as possible. # Actual Today, the action does paging to get all of the repo's git tags. # Ideas * Use github's graphql api to specify in more detail what types of tags we want. * Use caching.
non_priority
reduce network calls for getting tags list expected when getting the repo s git tags reduce down the number of network requests as little as possible actual today the action does paging to get all of the repo s git tags ideas use github s graphql api to specify in more detail what types of tags we want use caching
0
76,162
7,520,216,392
IssuesEvent
2018-04-12 13:56:19
EOSIO/eos
https://api.github.com/repos/EOSIO/eos
closed
nodeos --config doesn't support path with whitespace on Mac
needs testing - pass
```` ./programs/nodeos/nodeos --config=/Users/antonperkov/Library/Application\ Support/eosio.rel/nodeos/config/config.ini -d /Users/antonperkov/Library/Application\ Support/eosio.rel/ --plugin eosio::wallet_api_plugin --plugin eosio::chain_api_plugin --plugin eosio::account_history_api_plugin --resync ```` fails with ``` 140142ms thread-0 main.cpp:97 main ] Throw location unknown (consider using BOOST_THROW_EXCEPTION) Dynamic exception type: boost::exception_detail::clone_impl<boost::exception_detail::error_info_injector<boost::program_options::invalid_option_value> > std::exception::what: the argument ('/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini') for option '--config' is invalid ``` stacktrace: ```` (lldb) bt * thread #1, queue = 'com.apple.main-thread', stop reason = breakpoint 1.1 * frame #0: 0x00007fff7304d1f4 libc++abi.dylib`__cxa_throw frame #1: 0x00000001000f23e2 nodeos`void boost::throw_exception<boost::bad_lexical_cast>(e=0x00007ffeefbfa068) at throw_exception.hpp:69 frame #2: 0x000000010033e77f nodeos`void boost::conversion::detail::throw_bad_cast<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, boost::filesystem::path>() at bad_lexical_cast.hpp:92 frame #3: 0x000000010033e68a nodeos`boost::filesystem::path boost::lexical_cast<boost::filesystem::path, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > >(arg="/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini") at lexical_cast.hpp:42 frame #4: 0x000000010033e525 nodeos`void boost::program_options::validate<boost::filesystem::path, char>(v=0x00000001076127e8, xs=size=1, (null)=0x0000000000000000, (null)=0) at value_semantic.hpp:92 frame #5: 0x000000010033e04e nodeos`boost::program_options::typed_value<boost::filesystem::path, char>::xparse(this=0x0000000107612620, value_store=0x00000001076127e8, new_tokens=size=1) const at value_semantic.hpp:184 frame #6: 0x00000001018f1119 nodeos`boost::program_options::store(options=0x00007ffeefbfaae0, xm=0x00007ffeefbfab08, utf8=<unavailable>) at variables_map.cpp:81 [opt] frame #7: 0x00000001000341a8 nodeos`appbase::application::initialize_impl(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980, autostart_plugins=size=4) at application.cpp:107 frame #8: 0x0000000100012243 nodeos`bool appbase::application::initialize<eosio::chain_plugin, eosio::http_plugin, eosio::net_plugin, eosio::producer_plugin>(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980) at application.hpp:62 frame #9: 0x000000010000a4d1 nodeos`main(argc=11, argv=0x00007ffeefbff980) at main.cpp:87 frame #10: 0x00007fff74faa115 libdyld.dylib`start + 1 ```` ```` 34 namespace boost 35 { 36 template <typename Target, typename Source> 37 inline Target lexical_cast(const Source &arg) 38 { (lldb) l 39 Target result = Target(); 40 41 if (!boost::conversion::detail::try_lexical_convert(arg, result)) { 42 boost::conversion::detail::throw_bad_cast<Source, Target>(); 43 } 44 45 return result; 46 } 47 48 template <typename Target> (lldb) p arg (const std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) $1 = "/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini" (lldb) p result (boost::filesystem::path) $0 = (m_pathname = "/Users/antonperkov/Library/Application") ````
1.0
nodeos --config doesn't support path with whitespace on Mac - ```` ./programs/nodeos/nodeos --config=/Users/antonperkov/Library/Application\ Support/eosio.rel/nodeos/config/config.ini -d /Users/antonperkov/Library/Application\ Support/eosio.rel/ --plugin eosio::wallet_api_plugin --plugin eosio::chain_api_plugin --plugin eosio::account_history_api_plugin --resync ```` fails with ``` 140142ms thread-0 main.cpp:97 main ] Throw location unknown (consider using BOOST_THROW_EXCEPTION) Dynamic exception type: boost::exception_detail::clone_impl<boost::exception_detail::error_info_injector<boost::program_options::invalid_option_value> > std::exception::what: the argument ('/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini') for option '--config' is invalid ``` stacktrace: ```` (lldb) bt * thread #1, queue = 'com.apple.main-thread', stop reason = breakpoint 1.1 * frame #0: 0x00007fff7304d1f4 libc++abi.dylib`__cxa_throw frame #1: 0x00000001000f23e2 nodeos`void boost::throw_exception<boost::bad_lexical_cast>(e=0x00007ffeefbfa068) at throw_exception.hpp:69 frame #2: 0x000000010033e77f nodeos`void boost::conversion::detail::throw_bad_cast<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, boost::filesystem::path>() at bad_lexical_cast.hpp:92 frame #3: 0x000000010033e68a nodeos`boost::filesystem::path boost::lexical_cast<boost::filesystem::path, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > >(arg="/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini") at lexical_cast.hpp:42 frame #4: 0x000000010033e525 nodeos`void boost::program_options::validate<boost::filesystem::path, char>(v=0x00000001076127e8, xs=size=1, (null)=0x0000000000000000, (null)=0) at value_semantic.hpp:92 frame #5: 0x000000010033e04e nodeos`boost::program_options::typed_value<boost::filesystem::path, char>::xparse(this=0x0000000107612620, value_store=0x00000001076127e8, new_tokens=size=1) const at value_semantic.hpp:184 frame #6: 0x00000001018f1119 nodeos`boost::program_options::store(options=0x00007ffeefbfaae0, xm=0x00007ffeefbfab08, utf8=<unavailable>) at variables_map.cpp:81 [opt] frame #7: 0x00000001000341a8 nodeos`appbase::application::initialize_impl(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980, autostart_plugins=size=4) at application.cpp:107 frame #8: 0x0000000100012243 nodeos`bool appbase::application::initialize<eosio::chain_plugin, eosio::http_plugin, eosio::net_plugin, eosio::producer_plugin>(this=0x0000000102fb9170, argc=11, argv=0x00007ffeefbff980) at application.hpp:62 frame #9: 0x000000010000a4d1 nodeos`main(argc=11, argv=0x00007ffeefbff980) at main.cpp:87 frame #10: 0x00007fff74faa115 libdyld.dylib`start + 1 ```` ```` 34 namespace boost 35 { 36 template <typename Target, typename Source> 37 inline Target lexical_cast(const Source &arg) 38 { (lldb) l 39 Target result = Target(); 40 41 if (!boost::conversion::detail::try_lexical_convert(arg, result)) { 42 boost::conversion::detail::throw_bad_cast<Source, Target>(); 43 } 44 45 return result; 46 } 47 48 template <typename Target> (lldb) p arg (const std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) $1 = "/Users/antonperkov/Library/Application Support/eosio.rel/nodeos/config/config.ini" (lldb) p result (boost::filesystem::path) $0 = (m_pathname = "/Users/antonperkov/Library/Application") ````
non_priority
nodeos config doesn t support path with whitespace on mac programs nodeos nodeos config users antonperkov library application support eosio rel nodeos config config ini d users antonperkov library application support eosio rel plugin eosio wallet api plugin plugin eosio chain api plugin plugin eosio account history api plugin resync fails with thread main cpp main throw location unknown consider using boost throw exception dynamic exception type boost exception detail clone impl std exception what the argument users antonperkov library application support eosio rel nodeos config config ini for option config is invalid stacktrace lldb bt thread queue com apple main thread stop reason breakpoint frame libc abi dylib cxa throw frame nodeos void boost throw exception e at throw exception hpp frame nodeos void boost conversion detail throw bad cast std allocator boost filesystem path at bad lexical cast hpp frame nodeos boost filesystem path boost lexical cast std allocator arg users antonperkov library application support eosio rel nodeos config config ini at lexical cast hpp frame nodeos void boost program options validate v xs size null null at value semantic hpp frame nodeos boost program options typed value xparse this value store new tokens size const at value semantic hpp frame nodeos boost program options store options xm at variables map cpp frame nodeos appbase application initialize impl this argc argv autostart plugins size at application cpp frame nodeos bool appbase application initialize this argc argv at application hpp frame nodeos main argc argv at main cpp frame libdyld dylib start namespace boost template inline target lexical cast const source arg lldb l target result target if boost conversion detail try lexical convert arg result boost conversion detail throw bad cast return result template lldb p arg const std basic string std allocator users antonperkov library application support eosio rel nodeos config config ini lldb p result boost filesystem path m pathname users antonperkov library application
0
822,775
30,884,415,939
IssuesEvent
2023-08-03 20:22:43
GoogleCloudPlatform/vertex-ai-samples
https://api.github.com/repos/GoogleCloudPlatform/vertex-ai-samples
closed
error: subprocess-exited-with-error | Deploying Iris-detection model using FastAPI and Vertex AI custom container serving
priority: p1
## Expected Behavior No errors when installing packages from the requirements file from the notebook https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb ## Actual Behavior error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [794 lines of output] <string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html Partial import of sklearn during the build process. <string>:116: DeprecationWarning: `numpy.distutils` is deprecated since NumPy 1.23.0, as a result of the deprecation of `distutils` itself. It will be removed for Python >= 3.12. For older Python versions it will remain present. It is recommended to use `setuptools < 60.0` for those Python versions. For more details, see: https://numpy.org/devdocs/reference/distutils_status_migration.html ## Steps to Reproduce the Problem 1. Used the Vertex sample, https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb 2. Create the Requirements file: %%writefile requirements.txt joblib~=1.0 numpy~=1.20 scikit-learn~=0.24 google-cloud-storage>=1.26.0,<2.0.0dev 2. Run the install: # Required in Docker serving container ! pip3 install -U -r requirements.txt -q # For local FastAPI development and running ! pip3 install -U "uvicorn[standard]>=0.12.0,<0.14.0" fastapi~=0.63 -q # Vertex SDK for Python ! pip3 install -U google-cloud-aiplatform -q 3. Observe the error: error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [794 lines of output] <string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html Partial import of sklearn during the build process. <string>:116: DeprecationWarning: `numpy.distutils` is deprecated since NumPy 1.23.0, as a result of the deprecation of `distutils` itself. It will be removed for Python >= 3.12. For older Python versions it will remain present. It is recommended to use `setuptools < 60.0` for those Python versions. For more details, see: https://numpy.org/devdocs/reference/distutils_status_migration.html <SNIP> note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. ## Specifications - Version: N/A - Platform: N/A
1.0
error: subprocess-exited-with-error | Deploying Iris-detection model using FastAPI and Vertex AI custom container serving - ## Expected Behavior No errors when installing packages from the requirements file from the notebook https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb ## Actual Behavior error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [794 lines of output] <string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html Partial import of sklearn during the build process. <string>:116: DeprecationWarning: `numpy.distutils` is deprecated since NumPy 1.23.0, as a result of the deprecation of `distutils` itself. It will be removed for Python >= 3.12. For older Python versions it will remain present. It is recommended to use `setuptools < 60.0` for those Python versions. For more details, see: https://numpy.org/devdocs/reference/distutils_status_migration.html ## Steps to Reproduce the Problem 1. Used the Vertex sample, https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/custom/SDK_Custom_Container_Prediction.ipynb 2. Create the Requirements file: %%writefile requirements.txt joblib~=1.0 numpy~=1.20 scikit-learn~=0.24 google-cloud-storage>=1.26.0,<2.0.0dev 2. Run the install: # Required in Docker serving container ! pip3 install -U -r requirements.txt -q # For local FastAPI development and running ! pip3 install -U "uvicorn[standard]>=0.12.0,<0.14.0" fastapi~=0.63 -q # Vertex SDK for Python ! pip3 install -U google-cloud-aiplatform -q 3. Observe the error: error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [794 lines of output] <string>:17: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html Partial import of sklearn during the build process. <string>:116: DeprecationWarning: `numpy.distutils` is deprecated since NumPy 1.23.0, as a result of the deprecation of `distutils` itself. It will be removed for Python >= 3.12. For older Python versions it will remain present. It is recommended to use `setuptools < 60.0` for those Python versions. For more details, see: https://numpy.org/devdocs/reference/distutils_status_migration.html <SNIP> note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. ## Specifications - Version: N/A - Platform: N/A
priority
error subprocess exited with error deploying iris detection model using fastapi and vertex ai custom container serving expected behavior no errors when installing packages from the requirements file from the notebook actual behavior error subprocess exited with error × preparing metadata pyproject toml did not run successfully │ exit code ╰─ deprecationwarning pkg resources is deprecated as an api see partial import of sklearn during the build process deprecationwarning numpy distutils is deprecated since numpy as a result of the deprecation of distutils itself it will be removed for python for older python versions it will remain present it is recommended to use setuptools for those python versions for more details see steps to reproduce the problem used the vertex sample create the requirements file writefile requirements txt joblib numpy scikit learn google cloud storage run the install required in docker serving container install u r requirements txt q for local fastapi development and running install u uvicorn fastapi q vertex sdk for python install u google cloud aiplatform q observe the error error subprocess exited with error × preparing metadata pyproject toml did not run successfully │ exit code ╰─ deprecationwarning pkg resources is deprecated as an api see partial import of sklearn during the build process deprecationwarning numpy distutils is deprecated since numpy as a result of the deprecation of distutils itself it will be removed for python for older python versions it will remain present it is recommended to use setuptools for those python versions for more details see note this error originates from a subprocess and is likely not a problem with pip error metadata generation failed × encountered error while generating package metadata ╰─ see above for output note this is an issue with the package mentioned above not pip hint see above for details specifications version n a platform n a
1
599,145
18,266,407,850
IssuesEvent
2021-10-04 08:58:37
stevenwaterman/Lexoral
https://api.github.com/repos/stevenwaterman/Lexoral
opened
Allow running the functions locally for testing
high priority functions workflows tooling
Currently we can only run them online. This might not be possible with workflows, but because we're generating the workflow yaml with a DSL we could always use that DSL to call the other functions or something weird like that.
1.0
Allow running the functions locally for testing - Currently we can only run them online. This might not be possible with workflows, but because we're generating the workflow yaml with a DSL we could always use that DSL to call the other functions or something weird like that.
priority
allow running the functions locally for testing currently we can only run them online this might not be possible with workflows but because we re generating the workflow yaml with a dsl we could always use that dsl to call the other functions or something weird like that
1
516,295
14,978,912,162
IssuesEvent
2021-01-28 11:29:35
sButtons/sbuttons
https://api.github.com/repos/sButtons/sbuttons
opened
Fix base icon button class in Documentation Page
Priority: High bug good first issue help wanted up-for-grabs website
In documentation page under [Icon with no background](https://sbuttons.github.io/sbuttons/documentation.html#iconNoBg) change the class of the example button from `base-icon-btn` to `icon-btn`. This should be changed in the button and the code example
1.0
Fix base icon button class in Documentation Page - In documentation page under [Icon with no background](https://sbuttons.github.io/sbuttons/documentation.html#iconNoBg) change the class of the example button from `base-icon-btn` to `icon-btn`. This should be changed in the button and the code example
priority
fix base icon button class in documentation page in documentation page under change the class of the example button from base icon btn to icon btn this should be changed in the button and the code example
1
134,959
12,643,085,083
IssuesEvent
2020-06-16 09:13:57
durganp1/M2Challenge
https://api.github.com/repos/durganp1/M2Challenge
closed
Create Bio Section
documentation
Desired State: * personal background * Hobbies & Interest * career goals
1.0
Create Bio Section - Desired State: * personal background * Hobbies & Interest * career goals
non_priority
create bio section desired state personal background hobbies interest career goals
0
630,779
20,117,778,393
IssuesEvent
2022-02-07 21:31:50
arfc/saltproc
https://api.github.com/repos/arfc/saltproc
closed
Creat JSON schema for input
Type:Style Type:Feature Status:5-In Review Priority:2-Normal Difficulty:2-Challenging Comp:Input
Structure the input file based on a validation-ready json input file schema. This schema should be helped by description docstrings where necessary. However, a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears. (declare private or public variable where appropriate).
1.0
Creat JSON schema for input - Structure the input file based on a validation-ready json input file schema. This schema should be helped by description docstrings where necessary. However, a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears. (declare private or public variable where appropriate).
priority
creat json schema for input structure the input file based on a validation ready json input file schema this schema should be helped by description docstrings where necessary however a description of each variable that each input parameter becomes at runtime should be present in the class where that variable appears declare private or public variable where appropriate
1
722,150
24,852,208,806
IssuesEvent
2022-10-26 21:10:51
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
Make it obvious when a warning is elevated to an error
Priority:2 Type:DCR Area:ErrorHandling Functionality:Restore Style:PackageReference
Customers frequently get confused why restore fails when the message reported doesn't sound like an error, for example NU1603. I think we should have something in the message that makes it obvious that NuGet considers it a warning, but it was elevated to an error by request.
1.0
Make it obvious when a warning is elevated to an error - Customers frequently get confused why restore fails when the message reported doesn't sound like an error, for example NU1603. I think we should have something in the message that makes it obvious that NuGet considers it a warning, but it was elevated to an error by request.
priority
make it obvious when a warning is elevated to an error customers frequently get confused why restore fails when the message reported doesn t sound like an error for example i think we should have something in the message that makes it obvious that nuget considers it a warning but it was elevated to an error by request
1
792,450
27,961,150,062
IssuesEvent
2023-03-24 15:44:58
KDT3-Final-6/final-project-FE
https://api.github.com/repos/KDT3-Final-6/final-project-FE
reopened
Design: Home 마크업
Status: Available Status: Review Needed Priority: High Type: Feature/Function
## ✔️ 체크사항 - [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요. ## 💡 개발 목적 - 'Home' page 마크업 ## 🌐 세부 내용 - [x] 헤더 - [x] 슬라이더 - [ ] 큐레이션 - [ ] 고객님을 위한 추천여행 - [ ] 그룹별 여행 - [ ] 테마별 인기 여행 - [ ] 여행지소개 - [ ] 실제 고객 여행 후기 - [ ] 이벤트 - [x] 푸터
1.0
Design: Home 마크업 - ## ✔️ 체크사항 - [ ] 제목은 `개발 페이지 - 개발 목적` 으로 작성해 주세요. ## 💡 개발 목적 - 'Home' page 마크업 ## 🌐 세부 내용 - [x] 헤더 - [x] 슬라이더 - [ ] 큐레이션 - [ ] 고객님을 위한 추천여행 - [ ] 그룹별 여행 - [ ] 테마별 인기 여행 - [ ] 여행지소개 - [ ] 실제 고객 여행 후기 - [ ] 이벤트 - [x] 푸터
priority
design home 마크업 ✔️ 체크사항 제목은 개발 페이지 개발 목적 으로 작성해 주세요 💡 개발 목적 home page 마크업 🌐 세부 내용 헤더 슬라이더 큐레이션 고객님을 위한 추천여행 그룹별 여행 테마별 인기 여행 여행지소개 실제 고객 여행 후기 이벤트 푸터
1
28,322
11,610,981,337
IssuesEvent
2020-02-26 05:01:48
istio/istio
https://api.github.com/repos/istio/istio
closed
Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate
area/security lifecycle/needs-triage lifecycle/stale
Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate. The test cases need to cover file mount and workload SDS. **Affected product area (please put an X in all that apply)** [ ] Configuration Infrastructure [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [X] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Additional context** https://github.com/istio/istio/issues/13439
True
Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate - Add integration test case for an internal client calling an internal service, and the service is using a wrong certificate. The test cases need to cover file mount and workload SDS. **Affected product area (please put an X in all that apply)** [ ] Configuration Infrastructure [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [X] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Additional context** https://github.com/istio/istio/issues/13439
non_priority
add integration test case for an internal client calling an internal service and the service is using a wrong certificate add integration test case for an internal client calling an internal service and the service is using a wrong certificate the test cases need to cover file mount and workload sds affected product area please put an x in all that apply configuration infrastructure docs installation networking performance and scalability policies and telemetry security test and release user experience developer infrastructure additional context
0
84,528
3,669,167,218
IssuesEvent
2016-02-21 02:09:33
FreezingMoon/AncientBeast
https://api.github.com/repos/FreezingMoon/AncientBeast
opened
plasma indicator vanishes
Coding Priority Visuals
After a player uses plasma points, the plasma indicator (capsule shaped) beneath the Dark Priest doesn't update to show the remaining points, the health indicator (rectangle shape) shows up instead.
1.0
plasma indicator vanishes - After a player uses plasma points, the plasma indicator (capsule shaped) beneath the Dark Priest doesn't update to show the remaining points, the health indicator (rectangle shape) shows up instead.
priority
plasma indicator vanishes after a player uses plasma points the plasma indicator capsule shaped beneath the dark priest doesn t update to show the remaining points the health indicator rectangle shape shows up instead
1
611,553
18,958,304,732
IssuesEvent
2021-11-18 23:30:18
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Door lights do not blink red when access is denied
Type: Bug Priority: 1-Urgent Difficulty: 1-Easy
## Description Bug present (approximately) since the recent large sprite merge.
1.0
Door lights do not blink red when access is denied - ## Description Bug present (approximately) since the recent large sprite merge.
priority
door lights do not blink red when access is denied description bug present approximately since the recent large sprite merge
1
436,801
12,554,030,518
IssuesEvent
2020-06-07 00:22:51
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
Custom log handler do not work in cluster mode.
Component: logging ERR: Assignee Priority: Minor Stale Type: Bug
1) I've installed a cluster with 2 instances. 2) Then, I'he created jar-file with my simple custom log handler (just for tests): package org.company.util; import java.util.logging.Handler; import java.util.logging.LogRecord; import java.util.logging.SimpleFormatter; import java.io.FileOutputStream; import java.io.PrintWriter; /** * * @author me */ public class LmtLogHandler extends Handler { FileOutputStream fileOutputStream; PrintWriter printWriter; public LmtLogHandler() { String filename = "MyCustomLogfile.txt"; try { fileOutputStream = new FileOutputStream(filename); printWriter = new PrintWriter(fileOutputStream); setFormatter(new SimpleFormatter()); } catch (Exception e) { throw new Error("LogHandler init error!" + e); } } @Override public void publish(LogRecord record) { if (!isLoggable(record)) { return; } printWriter.println("LMT:" + getFormatter().format(record)); printWriter.flush(); } @Override public void flush() { printWriter.flush(); } @Override public void close() throws SecurityException { printWriter.close(); } } 3) I've copied this jar in glassfish lib dirrectory. (/usr/local/glassfish/lib) 4) Then, I've added handler (from administration console) for cluster configuration. 5) Restart server. There are no files "MyCustomLogfile.txt" in instances (/usr/local/glassfish/nodeagents/myagent/instance1 or /usr/local/glassfish/nodeagents/myagent/instance2 or /usr/local/glassfish/nodeagents/myagent/agent/) 6) Then, I've added handler (from administration console) for server configuration. 7) Restart server. There is a file MyCustomLogfile.txt in /usr/local/glassfish/domains/domain1/config/ Is is a bug, that custom log handler functionality does not work in cluster mode? #### Environment Operating System: Linux Platform: Linux #### Affected Versions [9.1peur2]
1.0
Custom log handler do not work in cluster mode. - 1) I've installed a cluster with 2 instances. 2) Then, I'he created jar-file with my simple custom log handler (just for tests): package org.company.util; import java.util.logging.Handler; import java.util.logging.LogRecord; import java.util.logging.SimpleFormatter; import java.io.FileOutputStream; import java.io.PrintWriter; /** * * @author me */ public class LmtLogHandler extends Handler { FileOutputStream fileOutputStream; PrintWriter printWriter; public LmtLogHandler() { String filename = "MyCustomLogfile.txt"; try { fileOutputStream = new FileOutputStream(filename); printWriter = new PrintWriter(fileOutputStream); setFormatter(new SimpleFormatter()); } catch (Exception e) { throw new Error("LogHandler init error!" + e); } } @Override public void publish(LogRecord record) { if (!isLoggable(record)) { return; } printWriter.println("LMT:" + getFormatter().format(record)); printWriter.flush(); } @Override public void flush() { printWriter.flush(); } @Override public void close() throws SecurityException { printWriter.close(); } } 3) I've copied this jar in glassfish lib dirrectory. (/usr/local/glassfish/lib) 4) Then, I've added handler (from administration console) for cluster configuration. 5) Restart server. There are no files "MyCustomLogfile.txt" in instances (/usr/local/glassfish/nodeagents/myagent/instance1 or /usr/local/glassfish/nodeagents/myagent/instance2 or /usr/local/glassfish/nodeagents/myagent/agent/) 6) Then, I've added handler (from administration console) for server configuration. 7) Restart server. There is a file MyCustomLogfile.txt in /usr/local/glassfish/domains/domain1/config/ Is is a bug, that custom log handler functionality does not work in cluster mode? #### Environment Operating System: Linux Platform: Linux #### Affected Versions [9.1peur2]
priority
custom log handler do not work in cluster mode i ve installed a cluster with instances then i he created jar file with my simple custom log handler just for tests package org company util import java util logging handler import java util logging logrecord import java util logging simpleformatter import java io fileoutputstream import java io printwriter author me public class lmtloghandler extends handler fileoutputstream fileoutputstream printwriter printwriter public lmtloghandler string filename mycustomlogfile txt try fileoutputstream new fileoutputstream filename printwriter new printwriter fileoutputstream setformatter new simpleformatter catch exception e throw new error loghandler init error e override public void publish logrecord record if isloggable record return printwriter println lmt getformatter format record printwriter flush override public void flush printwriter flush override public void close throws securityexception printwriter close i ve copied this jar in glassfish lib dirrectory usr local glassfish lib then i ve added handler from administration console for cluster configuration restart server there are no files mycustomlogfile txt in instances usr local glassfish nodeagents myagent or usr local glassfish nodeagents myagent or usr local glassfish nodeagents myagent agent then i ve added handler from administration console for server configuration restart server there is a file mycustomlogfile txt in usr local glassfish domains config is is a bug that custom log handler functionality does not work in cluster mode environment operating system linux platform linux affected versions
1
201,823
15,814,370,052
IssuesEvent
2021-04-05 09:21:20
AY2021S2-CS2103T-T12-3/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-3/tp
closed
[PE-D] Typo for "alias delete" error message
documentation severity.VeryLow
Steps to reproduce: 1. "alias add ls list" 2. "alias delete ls list" The example given in the error message is a wrong command. "alias delete ls" as given in the UG deleted the alias successfully. ![image.png](https://raw.githubusercontent.com/laurenlhy/ped/main/files/351e1505-1472-4f9d-820a-292880ae140c.png) <!--session: 1617429463103-be450b29-691c-46b2-91ae-62971187af48--> ------------- Labels: `severity.VeryLow` `type.DocumentationBug` original: laurenlhy/ped#11
1.0
[PE-D] Typo for "alias delete" error message - Steps to reproduce: 1. "alias add ls list" 2. "alias delete ls list" The example given in the error message is a wrong command. "alias delete ls" as given in the UG deleted the alias successfully. ![image.png](https://raw.githubusercontent.com/laurenlhy/ped/main/files/351e1505-1472-4f9d-820a-292880ae140c.png) <!--session: 1617429463103-be450b29-691c-46b2-91ae-62971187af48--> ------------- Labels: `severity.VeryLow` `type.DocumentationBug` original: laurenlhy/ped#11
non_priority
typo for alias delete error message steps to reproduce alias add ls list alias delete ls list the example given in the error message is a wrong command alias delete ls as given in the ug deleted the alias successfully labels severity verylow type documentationbug original laurenlhy ped
0
367,849
10,862,087,178
IssuesEvent
2019-11-14 12:33:26
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.getpostman.com - site is not usable
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 71.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:71.0) Gecko/20100101 Firefox/71.0 --> <!-- @reported_with: --> **URL**: https://www.getpostman.com/ **Browser / Version**: Firefox 71.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Entire site hangs. **Steps to Reproduce**: go to https://www.getpostman.com/ try to click download. Nothing happens. Scrolling down reveals only white. Works fine in Chrome [![Screenshot Description](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb-thumb.jpg)](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.getpostman.com - site is not usable - <!-- @browser: Firefox 71.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:71.0) Gecko/20100101 Firefox/71.0 --> <!-- @reported_with: --> **URL**: https://www.getpostman.com/ **Browser / Version**: Firefox 71.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Entire site hangs. **Steps to Reproduce**: go to https://www.getpostman.com/ try to click download. Nothing happens. Scrolling down reveals only white. Works fine in Chrome [![Screenshot Description](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb-thumb.jpg)](https://webcompat.com/uploads/2019/11/d305a4b1-cc8c-4492-9577-a429178eb6eb.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description entire site hangs steps to reproduce go to try to click download nothing happens scrolling down reveals only white works fine in chrome browser configuration none from with ❤️
1
166,768
6,311,178,949
IssuesEvent
2017-07-23 17:10:10
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
closed
Update auto-update config of video.js
High Priority in progress Library - Request to Add/Update
As you can see that I manually added the recent few versions, we should exclude `video.cjs.js` & `video.es.js` in most of the cases, so the auto-update config should be updated.
1.0
Update auto-update config of video.js - As you can see that I manually added the recent few versions, we should exclude `video.cjs.js` & `video.es.js` in most of the cases, so the auto-update config should be updated.
priority
update auto update config of video js as you can see that i manually added the recent few versions we should exclude video cjs js video es js in most of the cases so the auto update config should be updated
1
370,684
25,919,094,213
IssuesEvent
2022-12-15 20:05:07
transparencia-mg/work-stefanini
https://api.github.com/repos/transparencia-mg/work-stefanini
closed
Validação Casos de Testes Upload de Arquivos
bug documentation
@Andrelamor, gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário. Os mesmos podem ser consultados [aqui](https://transparencia-mg.github.io/work-stefanini/dev/testes/01_upload_de_arquivos_recursos_casos_de_teste/).
1.0
Validação Casos de Testes Upload de Arquivos - @Andrelamor, gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário. Os mesmos podem ser consultados [aqui](https://transparencia-mg.github.io/work-stefanini/dev/testes/01_upload_de_arquivos_recursos_casos_de_teste/).
non_priority
validação casos de testes upload de arquivos andrelamor gostaria de sua ajuda para validar os casos de testes criados para nossa primeira estória de usuário os mesmos podem ser consultados
0
761,727
26,694,486,456
IssuesEvent
2023-01-27 09:09:35
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
CI:frdm_k64f: kernel.common.stack_protection test failure
bug priority: high area: Memory Protection platform: NXP Release Blocker
**Describe the bug** kernel common stack protection test failure on frdm_k64f initial found in v3.2.0-3560-g80f87b9480 and still exist in zephyr-v3.2.0-3758-g61f89b7b18a1 **To Reproduce** Steps to reproduce the behavior: 1. mkdir build; cd build 2. cmake -DBOARD=frdm_k64f 3. make 4. See error **Expected behavior** Test PASS **Impact** kernel stack protection **Logs and console output** ``` *** Booting Zephyr OS build zephyr-v3.2.0-3758-g61f89b7b18a1 *** Running TESTSUITE fatal_exception =================================================================== START - test_fatal test alt thread 1: generic CPU exception E: ***** BUS FAULT ***** E: Instruction bus error E: NXP MPU error, port 3 E: Mode: Supervisor, Instruction Address: 0x20001aa4 E: Type: Read, Master: 0, Regions: 0x8200 E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000014 E: r3/a4: 0x20001aa4 r12/ip: 0x00000000 r14/lr: 0x000007fb E: xpsr: 0x40000000 E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000 E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000 E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000 E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000 E: fpscr: 0x20000388 E: Faulting instruction address (r15/pc): 0x20001aa4 E: &gt;&gt;&gt; ZEPHYR FATAL ERROR 27: Unknown error on CPU 0 E: Current thread: 0x20000020 (unknown) Caught system error -- reason 27 Wrong crash type got 27 expected 35 or 20 ``` **Environment (please complete the following information):** - OS: (e.g. Linux) - Toolchain (e.g Zephyr SDK, ...) - Commit SHA or Version used: v3.2.0-3560-g80f87b9480
1.0
CI:frdm_k64f: kernel.common.stack_protection test failure - **Describe the bug** kernel common stack protection test failure on frdm_k64f initial found in v3.2.0-3560-g80f87b9480 and still exist in zephyr-v3.2.0-3758-g61f89b7b18a1 **To Reproduce** Steps to reproduce the behavior: 1. mkdir build; cd build 2. cmake -DBOARD=frdm_k64f 3. make 4. See error **Expected behavior** Test PASS **Impact** kernel stack protection **Logs and console output** ``` *** Booting Zephyr OS build zephyr-v3.2.0-3758-g61f89b7b18a1 *** Running TESTSUITE fatal_exception =================================================================== START - test_fatal test alt thread 1: generic CPU exception E: ***** BUS FAULT ***** E: Instruction bus error E: NXP MPU error, port 3 E: Mode: Supervisor, Instruction Address: 0x20001aa4 E: Type: Read, Master: 0, Regions: 0x8200 E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000014 E: r3/a4: 0x20001aa4 r12/ip: 0x00000000 r14/lr: 0x000007fb E: xpsr: 0x40000000 E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000 E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000 E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000 E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000 E: fpscr: 0x20000388 E: Faulting instruction address (r15/pc): 0x20001aa4 E: &gt;&gt;&gt; ZEPHYR FATAL ERROR 27: Unknown error on CPU 0 E: Current thread: 0x20000020 (unknown) Caught system error -- reason 27 Wrong crash type got 27 expected 35 or 20 ``` **Environment (please complete the following information):** - OS: (e.g. Linux) - Toolchain (e.g Zephyr SDK, ...) - Commit SHA or Version used: v3.2.0-3560-g80f87b9480
priority
ci frdm kernel common stack protection test failure describe the bug kernel common stack protection test failure on frdm initial found in and still exist in zephyr to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard frdm make see error expected behavior test pass impact kernel stack protection logs and console output booting zephyr os build zephyr running testsuite fatal exception start test fatal test alt thread generic cpu exception e bus fault e instruction bus error e nxp mpu error port e mode supervisor instruction address e type read master regions e e ip lr e xpsr e s s s s e s s s s e s s s s e s s s s e fpscr e faulting instruction address pc e gt gt gt zephyr fatal error unknown error on cpu e current thread unknown caught system error reason wrong crash type got expected or environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used
1
390,507
11,544,463,056
IssuesEvent
2020-02-18 11:31:29
Los-nonos/ZeepCommerce
https://api.github.com/repos/Los-nonos/ZeepCommerce
closed
[Website] Add template material kit for views
frontend priority: high status: DONE
Añadir todo el html y pasarlo a react component el archivo template.html que está en el kit en la homepage
1.0
[Website] Add template material kit for views - Añadir todo el html y pasarlo a react component el archivo template.html que está en el kit en la homepage
priority
add template material kit for views añadir todo el html y pasarlo a react component el archivo template html que está en el kit en la homepage
1
50,604
7,613,240,119
IssuesEvent
2018-05-01 20:29:45
sonarwhal/sonarwhal
https://api.github.com/repos/sonarwhal/sonarwhal
opened
Add documentation on how to integrate sonarwhal into the development flow
area:documentation difficulty:medium priority:high
There are a few ways to integrate sonarwhal during the development and I expect to have several "articles" ore recipes: * [ ] Using the local connector while developing * [ ] In the dev machine running a web server (such as `http-server` if static) * [ ] In travis/Jenkins. * This could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production. I have an idea on how to do this with Azure but I'm not sure if we need to go as deep as to show how to do it on each Cloud provider.
1.0
Add documentation on how to integrate sonarwhal into the development flow - There are a few ways to integrate sonarwhal during the development and I expect to have several "articles" ore recipes: * [ ] Using the local connector while developing * [ ] In the dev machine running a web server (such as `http-server` if static) * [ ] In travis/Jenkins. * This could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production. I have an idea on how to do this with Azure but I'm not sure if we need to go as deep as to show how to do it on each Cloud provider.
non_priority
add documentation on how to integrate sonarwhal into the development flow there are a few ways to integrate sonarwhal during the development and i expect to have several articles ore recipes using the local connector while developing in the dev machine running a web server such as http server if static in travis jenkins this could be similar to the local server or integrate it with a staging environment and if it passes auto publish in production i have an idea on how to do this with azure but i m not sure if we need to go as deep as to show how to do it on each cloud provider
0
236,411
26,010,598,893
IssuesEvent
2022-12-21 01:03:35
Guillerbr/api-crud-adonisjs
https://api.github.com/repos/Guillerbr/api-crud-adonisjs
opened
CVE-2016-20018 (Medium) detected in knex-0.15.2.tgz
security vulnerability
## CVE-2016-20018 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>knex-0.15.2.tgz</b></p></summary> <p>A batteries-included SQL query & schema builder for Postgres, MySQL and SQLite3 and the Browser</p> <p>Library home page: <a href="https://registry.npmjs.org/knex/-/knex-0.15.2.tgz">https://registry.npmjs.org/knex/-/knex-0.15.2.tgz</a></p> <p>Path to dependency file: /api-crud-adonisjs/package.json</p> <p>Path to vulnerable library: /node_modules/knex/package.json</p> <p> Dependency Hierarchy: - lucid-6.1.3.tgz (Root Library) - :x: **knex-0.15.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Knex Knex.js through 2.3.0 has a limited SQL injection vulnerability that can be exploited to ignore the WHERE clause of a SQL query. <p>Publish Date: 2022-12-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-20018>CVE-2016-20018</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-20018 (Medium) detected in knex-0.15.2.tgz - ## CVE-2016-20018 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>knex-0.15.2.tgz</b></p></summary> <p>A batteries-included SQL query & schema builder for Postgres, MySQL and SQLite3 and the Browser</p> <p>Library home page: <a href="https://registry.npmjs.org/knex/-/knex-0.15.2.tgz">https://registry.npmjs.org/knex/-/knex-0.15.2.tgz</a></p> <p>Path to dependency file: /api-crud-adonisjs/package.json</p> <p>Path to vulnerable library: /node_modules/knex/package.json</p> <p> Dependency Hierarchy: - lucid-6.1.3.tgz (Root Library) - :x: **knex-0.15.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Knex Knex.js through 2.3.0 has a limited SQL injection vulnerability that can be exploited to ignore the WHERE clause of a SQL query. <p>Publish Date: 2022-12-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-20018>CVE-2016-20018</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in knex tgz cve medium severity vulnerability vulnerable library knex tgz a batteries included sql query schema builder for postgres mysql and and the browser library home page a href path to dependency file api crud adonisjs package json path to vulnerable library node modules knex package json dependency hierarchy lucid tgz root library x knex tgz vulnerable library vulnerability details knex knex js through has a limited sql injection vulnerability that can be exploited to ignore the where clause of a sql query publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
725,222
24,954,587,224
IssuesEvent
2022-11-01 10:32:26
thecyberworld/thecyberhub.org
https://api.github.com/repos/thecyberworld/thecyberhub.org
closed
[BUG] Sidebar dropwown not working
✨ goal: improvement 🛠 goal: fix 🤩 status : Up for Grab 🟥 priority: critical hacktoberfest
### Describe the bug The sidebar dropdown not working ![image](https://user-images.githubusercontent.com/44284877/198293652-6aa5898c-4165-4a8a-b12b-dbe1f43d3750.png) when we click on this nav-item nothing happens ### To Reproduce Click on the Sidebar > learn. dropdown not working ### Expected Behavior Like this: the drop should be displayed ![image](https://user-images.githubusercontent.com/44284877/198293966-fe1e47ab-9d61-48c4-aadc-4e30b9151f37.png) _No response_ ### Screenshot/ Video _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[BUG] Sidebar dropwown not working - ### Describe the bug The sidebar dropdown not working ![image](https://user-images.githubusercontent.com/44284877/198293652-6aa5898c-4165-4a8a-b12b-dbe1f43d3750.png) when we click on this nav-item nothing happens ### To Reproduce Click on the Sidebar > learn. dropdown not working ### Expected Behavior Like this: the drop should be displayed ![image](https://user-images.githubusercontent.com/44284877/198293966-fe1e47ab-9d61-48c4-aadc-4e30b9151f37.png) _No response_ ### Screenshot/ Video _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
priority
sidebar dropwown not working describe the bug the sidebar dropdown not working when we click on this nav item nothing happens to reproduce click on the sidebar learn dropdown not working expected behavior like this the drop should be displayed no response screenshot video no response code of conduct i agree to follow this project s code of conduct
1
175,628
21,314,082,367
IssuesEvent
2022-04-16 02:03:06
infrahq/infra
https://api.github.com/repos/infrahq/infra
closed
Issue engines unique keys after initial connection
security authentication status/stale
The engine connection process looks like this: 1. A default API key is generated 2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint 3. The engine presents this API key on subsequent requests We should migrate to issuing a new token to uniquely identify the engine after initial connection validation: 1. A default API key is generated 2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint 3. The registry creates a new key bound to the engine (with an `issued for` field), scoped to only permissions a registry needs 4. The engine uses this key on subsequent requests
True
Issue engines unique keys after initial connection - The engine connection process looks like this: 1. A default API key is generated 2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint 3. The engine presents this API key on subsequent requests We should migrate to issuing a new token to uniquely identify the engine after initial connection validation: 1. A default API key is generated 2. The engine is created and presents this default API key to the registry to connect at the create destination endpoint 3. The registry creates a new key bound to the engine (with an `issued for` field), scoped to only permissions a registry needs 4. The engine uses this key on subsequent requests
non_priority
issue engines unique keys after initial connection the engine connection process looks like this a default api key is generated the engine is created and presents this default api key to the registry to connect at the create destination endpoint the engine presents this api key on subsequent requests we should migrate to issuing a new token to uniquely identify the engine after initial connection validation a default api key is generated the engine is created and presents this default api key to the registry to connect at the create destination endpoint the registry creates a new key bound to the engine with an issued for field scoped to only permissions a registry needs the engine uses this key on subsequent requests
0
254,721
8,087,309,321
IssuesEvent
2018-08-09 00:56:36
docker/distribution
https://api.github.com/repos/docker/distribution
opened
Repo name validation is misused
bug priority/P1
Regardless of your registry actual domain name, the Parse method https://github.com/docker/distribution/blob/master/reference/reference.go#L189 is always receiving: `docker.io/foo/bar` or `docker.io/library/foo` (where `foo/bar` or `foo` is your image). It seems to be because we are calling https://github.com/docker/distribution/blob/master/reference/normalize.go#L33 on `vars.name` from the context, which is apparently just the repo name (without the domain). This makes the implementation deviate from the stated intent of the grammar. If NameTotalLengthMax https://github.com/docker/distribution/blob/master/reference/reference.go#L37 was meant to limit the total size of the image name including the domain, then it must use the actual registry domain (+port). If NameTotalLengthMax should be applied just on the path component, regardless of the domain, then we should not normalize `vars.name` before validating it (this is what the `WithName` method does, by the way). A fix for this need to be carefully thought, as in the current situation the actual restriction is that on repo names at 237 characters (with single components) and 245 characters if they contain a slash - introducing proper domain name validation may render existing images invalid. Let me know if I'm missing something here. cc @caervs @dmcgowan @tiborvass
1.0
Repo name validation is misused - Regardless of your registry actual domain name, the Parse method https://github.com/docker/distribution/blob/master/reference/reference.go#L189 is always receiving: `docker.io/foo/bar` or `docker.io/library/foo` (where `foo/bar` or `foo` is your image). It seems to be because we are calling https://github.com/docker/distribution/blob/master/reference/normalize.go#L33 on `vars.name` from the context, which is apparently just the repo name (without the domain). This makes the implementation deviate from the stated intent of the grammar. If NameTotalLengthMax https://github.com/docker/distribution/blob/master/reference/reference.go#L37 was meant to limit the total size of the image name including the domain, then it must use the actual registry domain (+port). If NameTotalLengthMax should be applied just on the path component, regardless of the domain, then we should not normalize `vars.name` before validating it (this is what the `WithName` method does, by the way). A fix for this need to be carefully thought, as in the current situation the actual restriction is that on repo names at 237 characters (with single components) and 245 characters if they contain a slash - introducing proper domain name validation may render existing images invalid. Let me know if I'm missing something here. cc @caervs @dmcgowan @tiborvass
priority
repo name validation is misused regardless of your registry actual domain name the parse method is always receiving docker io foo bar or docker io library foo where foo bar or foo is your image it seems to be because we are calling on vars name from the context which is apparently just the repo name without the domain this makes the implementation deviate from the stated intent of the grammar if nametotallengthmax was meant to limit the total size of the image name including the domain then it must use the actual registry domain port if nametotallengthmax should be applied just on the path component regardless of the domain then we should not normalize vars name before validating it this is what the withname method does by the way a fix for this need to be carefully thought as in the current situation the actual restriction is that on repo names at characters with single components and characters if they contain a slash introducing proper domain name validation may render existing images invalid let me know if i m missing something here cc caervs dmcgowan tiborvass
1
286,080
8,783,714,032
IssuesEvent
2018-12-20 07:17:52
syspro-team1/EXTdateManageApp
https://api.github.com/repos/syspro-team1/EXTdateManageApp
opened
Notificationまわりのリファクタリング
enhancement low priority
ProductManagerActivityにごちゃごちゃ通知を設定してるのは 可読性が低い気がする. クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち.
1.0
Notificationまわりのリファクタリング - ProductManagerActivityにごちゃごちゃ通知を設定してるのは 可読性が低い気がする. クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち.
priority
notificationまわりのリファクタリング productmanageractivityにごちゃごちゃ通知を設定してるのは 可読性が低い気がする. クラスをわけてリスナーを追加するなりして一つのクラスにまとめたい気持ち.
1
549,605
16,095,718,617
IssuesEvent
2021-04-26 23:13:27
lomcaitlin/cst438_s21_proj03_groupa
https://api.github.com/repos/lomcaitlin/cst438_s21_proj03_groupa
closed
Admin View Users
difficulty 2 priority 3
As a admin I want to be able to view all users so I can keep track of every user How will it be tested: Unit tests for username verification and password verification Instrumented test to verify extras
1.0
Admin View Users - As a admin I want to be able to view all users so I can keep track of every user How will it be tested: Unit tests for username verification and password verification Instrumented test to verify extras
priority
admin view users as a admin i want to be able to view all users so i can keep track of every user how will it be tested unit tests for username verification and password verification instrumented test to verify extras
1
323,475
9,855,484,347
IssuesEvent
2019-06-19 19:31:38
cloud-custodian/cloud-custodian
https://api.github.com/repos/cloud-custodian/cloud-custodian
closed
tools/mailer - import enum error
area/tools-mailer kind/bug priority/P1
Hi, with a 0.8.44, you have add provider(enum) on utils.py, but i can't find this on deploy.py/get_archive() and requirement.txt (with manual fix, this working again) with this error, i can't execute update-lambda could you update the integration of "enum" plz ? thanks for your help
1.0
tools/mailer - import enum error - Hi, with a 0.8.44, you have add provider(enum) on utils.py, but i can't find this on deploy.py/get_archive() and requirement.txt (with manual fix, this working again) with this error, i can't execute update-lambda could you update the integration of "enum" plz ? thanks for your help
priority
tools mailer import enum error hi with a you have add provider enum on utils py but i can t find this on deploy py get archive and requirement txt with manual fix this working again with this error i can t execute update lambda could you update the integration of enum plz thanks for your help
1
540,814
15,817,518,112
IssuesEvent
2021-04-05 14:43:12
AY2021S2-CS2103T-T12-4/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp
closed
[PE-D] Formatting of help window
priority.High severity.VeryLow type.Bug
No details provided. ![Screenshot 2021-04-03 at 2.25.53 PM.png](https://raw.githubusercontent.com/samuelfangjw/ped/main/files/e672ffa1-d14f-4766-8e20-e74ba5b5e2cc.png) In the help window, the button to copy link is truncated. I can also scroll horizontally even though the information already fits into one screen, and I need to scroll vertically to see all the commands even though they could easily fit in one screen. <!--session: 1617429943449-e56bf182-40ef-47bf-9213-d664c9ca18a6--> ------------- Labels: `severity.VeryLow` `type.FunctionalityBug` original: samuelfangjw/ped#5
1.0
[PE-D] Formatting of help window - No details provided. ![Screenshot 2021-04-03 at 2.25.53 PM.png](https://raw.githubusercontent.com/samuelfangjw/ped/main/files/e672ffa1-d14f-4766-8e20-e74ba5b5e2cc.png) In the help window, the button to copy link is truncated. I can also scroll horizontally even though the information already fits into one screen, and I need to scroll vertically to see all the commands even though they could easily fit in one screen. <!--session: 1617429943449-e56bf182-40ef-47bf-9213-d664c9ca18a6--> ------------- Labels: `severity.VeryLow` `type.FunctionalityBug` original: samuelfangjw/ped#5
priority
formatting of help window no details provided in the help window the button to copy link is truncated i can also scroll horizontally even though the information already fits into one screen and i need to scroll vertically to see all the commands even though they could easily fit in one screen labels severity verylow type functionalitybug original samuelfangjw ped
1
2,400
2,607,900,667
IssuesEvent
2015-02-26 00:13:24
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
reference error: use_tab is not defined
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. TYpe ul>li 2. Select 'Expand Abbreviation' 3. What is the expected output? What do you see instead? error message: reference error: use_tab is not defined What version of the product are you using? On what operating system? 0.5.1 on Windows with Aptana Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `freeyl...@gmail.com` on 14 Jan 2010 at 1:40
1.0
reference error: use_tab is not defined - ``` What steps will reproduce the problem? 1. TYpe ul>li 2. Select 'Expand Abbreviation' 3. What is the expected output? What do you see instead? error message: reference error: use_tab is not defined What version of the product are you using? On what operating system? 0.5.1 on Windows with Aptana Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `freeyl...@gmail.com` on 14 Jan 2010 at 1:40
non_priority
reference error use tab is not defined what steps will reproduce the problem type ul li select expand abbreviation what is the expected output what do you see instead error message reference error use tab is not defined what version of the product are you using on what operating system on windows with aptana please provide any additional information below original issue reported on code google com by freeyl gmail com on jan at
0
518,301
15,026,679,509
IssuesEvent
2021-02-01 23:08:08
tysonkaufmann/su-go
https://api.github.com/repos/tysonkaufmann/su-go
opened
[DEV] Add an endpoint to Get Route Information
Medium Priority task
**Related To** - [View a route](https://github.com/tysonkaufmann/su-go/issues/57) **Description** Create an API endpoint `/api/getroute/{ROUTEID}` which will serve `GET` requests: On Success, the API endpoint should return the information e.g ``` { "status":"200", "success":"true", "data": { "username":"", "routeid":"", "routetitle":"", "routetype":"", "routetime":"", "routedistance":"", "routedescription":"", "photos":[], "mapdata":{} } } ``` On Error, the API endpoint should return an error if unsuccessful e.g ``` { "status":"200", "success":"false", "error":"The route does not exist" } ``` **Development Steps** - Create integration tests to fully test the API endpoint `/api/getroute/{ROUTEID}` (TDD) - Route the endpoint in `backend/routes/Routes` - Create unit tests for any code added in `backend/controllers/Routes` (TDD) - Add code to the `/api/getroute/{ROUTEID}` to serve the API request
1.0
[DEV] Add an endpoint to Get Route Information - **Related To** - [View a route](https://github.com/tysonkaufmann/su-go/issues/57) **Description** Create an API endpoint `/api/getroute/{ROUTEID}` which will serve `GET` requests: On Success, the API endpoint should return the information e.g ``` { "status":"200", "success":"true", "data": { "username":"", "routeid":"", "routetitle":"", "routetype":"", "routetime":"", "routedistance":"", "routedescription":"", "photos":[], "mapdata":{} } } ``` On Error, the API endpoint should return an error if unsuccessful e.g ``` { "status":"200", "success":"false", "error":"The route does not exist" } ``` **Development Steps** - Create integration tests to fully test the API endpoint `/api/getroute/{ROUTEID}` (TDD) - Route the endpoint in `backend/routes/Routes` - Create unit tests for any code added in `backend/controllers/Routes` (TDD) - Add code to the `/api/getroute/{ROUTEID}` to serve the API request
priority
add an endpoint to get route information related to description create an api endpoint api getroute routeid which will serve get requests on success the api endpoint should return the information e g status success true data username routeid routetitle routetype routetime routedistance routedescription photos mapdata on error the api endpoint should return an error if unsuccessful e g status success false error the route does not exist development steps create integration tests to fully test the api endpoint api getroute routeid tdd route the endpoint in backend routes routes create unit tests for any code added in backend controllers routes tdd add code to the api getroute routeid to serve the api request
1
61,196
7,447,608,107
IssuesEvent
2018-03-28 13:05:05
AnSyn/ansyn
https://api.github.com/repos/AnSyn/ansyn
opened
Filters menu - limit section size
Design bug Priority: High Ready for: Development
at the moment each field is showing all of the available values, see "sensor name" in the attached image ![screen shot 2018-03-28 at 16 01 27](https://user-images.githubusercontent.com/4181572/38030549-9c9f363e-32a1-11e8-8b9a-28a11884314c.png) As default we should present the top-10 values, and enable the user to "show more" if needed (Expand). see attached image for inspiration: <img width="255" alt="screen shot 2018-03-28 at 16 00 03" src="https://user-images.githubusercontent.com/4181572/38030615-c56df46a-32a1-11e8-97b5-47fe03e3baca.png">
1.0
Filters menu - limit section size - at the moment each field is showing all of the available values, see "sensor name" in the attached image ![screen shot 2018-03-28 at 16 01 27](https://user-images.githubusercontent.com/4181572/38030549-9c9f363e-32a1-11e8-8b9a-28a11884314c.png) As default we should present the top-10 values, and enable the user to "show more" if needed (Expand). see attached image for inspiration: <img width="255" alt="screen shot 2018-03-28 at 16 00 03" src="https://user-images.githubusercontent.com/4181572/38030615-c56df46a-32a1-11e8-97b5-47fe03e3baca.png">
non_priority
filters menu limit section size at the moment each field is showing all of the available values see sensor name in the attached image as default we should present the top values and enable the user to show more if needed expand see attached image for inspiration img width alt screen shot at src
0
1,150
2,870,524,710
IssuesEvent
2015-06-07 07:49:51
pynac/pynac
https://api.github.com/repos/pynac/pynac
opened
cache expairseq::info computations
enhancement performance
The original behaviour is to compute `add`/`mul``::info` every time anew when it's called.
True
cache expairseq::info computations - The original behaviour is to compute `add`/`mul``::info` every time anew when it's called.
non_priority
cache expairseq info computations the original behaviour is to compute add mul info every time anew when it s called
0
755,682
26,436,768,450
IssuesEvent
2023-01-15 13:42:23
nrademacher/todo
https://api.github.com/repos/nrademacher/todo
closed
[NRA-117] Client: add pre-release disclaimer
High priority
Add disclaimer on sign-up pages that informs the user that the current version is a pre-release, account data may be lost, etc. <sub>From [SyncLinear.com](https://synclinear.com) | [NRA-117](https://linear.app/nrademacher/issue/NRA-117/client-add-pre-release-disclaimer)</sub>
1.0
[NRA-117] Client: add pre-release disclaimer - Add disclaimer on sign-up pages that informs the user that the current version is a pre-release, account data may be lost, etc. <sub>From [SyncLinear.com](https://synclinear.com) | [NRA-117](https://linear.app/nrademacher/issue/NRA-117/client-add-pre-release-disclaimer)</sub>
priority
client add pre release disclaimer add disclaimer on sign up pages that informs the user that the current version is a pre release account data may be lost etc from
1
226,495
18,022,614,128
IssuesEvent
2021-09-16 21:38:19
tpgxyz/test2
https://api.github.com/repos/tpgxyz/test2
opened
Second Open Bug (Bugzilla Bug 2)
bug tests
This issue was created automatically with bugzilla2github # Bugzilla Bug 2 Date: 2015-05-07 16:21:59 +0200 From: Andriy Berestovskyy &lt;<email@example.org>&gt; To: Andriy Berestovskyy &lt;<email@example.org>&gt; CC: email2@example.org Last updated: 2015-06-26 12:36:40 +0200 ## Comment 2 Date: 2015-05-07 16:21:59 +0200 From: Andriy Berestovskyy &lt;<email@example.org>&gt; Bug description goes here...
1.0
Second Open Bug (Bugzilla Bug 2) - This issue was created automatically with bugzilla2github # Bugzilla Bug 2 Date: 2015-05-07 16:21:59 +0200 From: Andriy Berestovskyy &lt;<email@example.org>&gt; To: Andriy Berestovskyy &lt;<email@example.org>&gt; CC: email2@example.org Last updated: 2015-06-26 12:36:40 +0200 ## Comment 2 Date: 2015-05-07 16:21:59 +0200 From: Andriy Berestovskyy &lt;<email@example.org>&gt; Bug description goes here...
non_priority
second open bug bugzilla bug this issue was created automatically with bugzilla bug date from andriy berestovskyy lt gt to andriy berestovskyy lt gt cc example org last updated comment date from andriy berestovskyy lt gt bug description goes here
0
62,861
26,191,435,910
IssuesEvent
2023-01-03 09:25:44
azure-deprecation/dashboard
https://api.github.com/repos/azure-deprecation/dashboard
closed
Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022
verified impact:upgrade-required area:feature cloud:public services:functions
Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 **Deadline:** Dec 03, 2022 **Impacted Services:** - Azure Functions **More information:** - https://azure.microsoft.com/updates/extended-support-for-microsoft-net-core-31-will-end-on-3-december-2022/ - https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version ### Notice Here's the official report from Microsoft: > **On 3 December 2022, extended support for Microsoft .NET Core 3.1 will end.** After that date, your applications that are hosted on Functions will continue to run and your applications will not be impacted. However, we'll no longer provide patches or customer service for .NET Core 3.1. Update your Functions applications to [runtime version 4.x, which uses .NET 6](https://docs.microsoft.com/azure/azure-functions/dotnet-isolated-process-guide#supported-versions). > > .NET 6 is the latest version with long-term support and provides these enhancements: > > - A unified set of base libraries and an SDK that make it easy to share code across any application type. > - Simplified development with new C# 10 features and minimal APIs. > - Hot reload that allows you to make code changes without explicit recompiling. ### Timeline | Phase | Date | Description | |:------|------|-------------| |Announcement|Mar 17, 2022|Deprecation was announced| |Deprecation|Dec 03, 2022|Function Apps will continue to run but no longer receive patches nor support| ### Impact Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 and upgrade to .NET 6 with Azure Functions v4 is required. ### Required Action A migration guide to the newer Azure Functions runtime version is available [here](https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version). Here's the official report from Microsoft: > To avoid potential service disruptions or security vulnerabilities, [update your Functions applications](https://aka.ms/functions-view-and-update-runtime-version) to runtime version 4.x, which uses .NET 6, before 3 December 2022. ### Contact You can get in touch through the following options: - Get answers from Microsoft Q&A ([link](https://aka.ms/functions-c-sharp-qa-link)). - Contact Azure support ([link](https://portal.azure.com/#blade/Microsoft_Azure_Support/HelpAndSupportBlade/overview)).
1.0
Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 - Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 **Deadline:** Dec 03, 2022 **Impacted Services:** - Azure Functions **More information:** - https://azure.microsoft.com/updates/extended-support-for-microsoft-net-core-31-will-end-on-3-december-2022/ - https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version ### Notice Here's the official report from Microsoft: > **On 3 December 2022, extended support for Microsoft .NET Core 3.1 will end.** After that date, your applications that are hosted on Functions will continue to run and your applications will not be impacted. However, we'll no longer provide patches or customer service for .NET Core 3.1. Update your Functions applications to [runtime version 4.x, which uses .NET 6](https://docs.microsoft.com/azure/azure-functions/dotnet-isolated-process-guide#supported-versions). > > .NET 6 is the latest version with long-term support and provides these enhancements: > > - A unified set of base libraries and an SDK that make it easy to share code across any application type. > - Simplified development with new C# 10 features and minimal APIs. > - Hot reload that allows you to make code changes without explicit recompiling. ### Timeline | Phase | Date | Description | |:------|------|-------------| |Announcement|Mar 17, 2022|Deprecation was announced| |Deprecation|Dec 03, 2022|Function Apps will continue to run but no longer receive patches nor support| ### Impact Extended support for .NET Core 3.1 in Azure Functions is retiring on December 3rd 2022 and upgrade to .NET 6 with Azure Functions v4 is required. ### Required Action A migration guide to the newer Azure Functions runtime version is available [here](https://docs.microsoft.com/azure/azure-functions/set-runtime-version?tabs=portal#view-and-update-the-current-runtime-version). Here's the official report from Microsoft: > To avoid potential service disruptions or security vulnerabilities, [update your Functions applications](https://aka.ms/functions-view-and-update-runtime-version) to runtime version 4.x, which uses .NET 6, before 3 December 2022. ### Contact You can get in touch through the following options: - Get answers from Microsoft Q&A ([link](https://aka.ms/functions-c-sharp-qa-link)). - Contact Azure support ([link](https://portal.azure.com/#blade/Microsoft_Azure_Support/HelpAndSupportBlade/overview)).
non_priority
extended support for net core in azure functions is retiring on december extended support for net core in azure functions is retiring on december deadline dec impacted services azure functions more information notice here s the official report from microsoft on december extended support for microsoft net core will end after that date your applications that are hosted on functions will continue to run and your applications will not be impacted however we ll no longer provide patches or customer service for net core update your functions applications to net is the latest version with long term support and provides these enhancements a unified set of base libraries and an sdk that make it easy to share code across any application type simplified development with new c features and minimal apis hot reload that allows you to make code changes without explicit recompiling timeline phase date description announcement mar deprecation was announced deprecation dec function apps will continue to run but no longer receive patches nor support impact extended support for net core in azure functions is retiring on december and upgrade to net with azure functions is required required action a migration guide to the newer azure functions runtime version is available here s the official report from microsoft to avoid potential service disruptions or security vulnerabilities to runtime version x which uses net before december contact you can get in touch through the following options get answers from microsoft q a contact azure support
0
67,005
16,769,019,638
IssuesEvent
2021-06-14 12:43:43
rust-lang/docs.rs
https://api.github.com/repos/rust-lang/docs.rs
reopened
"Failed to parse manifest" on bootloader-0.10.0-alpha-03
A-builds
Not sure what's going wrong here, `cargo doc` works locally. The crate doesn't show up in the web UI at all. ``` Feb 21 12:36:47 docsrs cratesfyi[28601]: 2021/02/21 12:36:47 [ERROR] docs_rs::build_queue: Failed to build package bootloader-0.10.0-alpha-03 from queue: failed to parse maniest Feb 21 12:36:47 docsrs cratesfyi[28601]: Backtrace: 0: failure::backtrace::internal::InternalBacktrace::new Feb 21 12:36:47 docsrs cratesfyi[28601]: 1: <failure::backtrace::Backtrace as core::default::Default>::default Feb 21 12:36:47 docsrs cratesfyi[28601]: 2: rustwide::build::BuildBuilder::run Feb 21 12:36:47 docsrs cratesfyi[28601]: 3: docs_rs::docbuilder::rustwide_builder::RustwideBuilder::build_package Feb 21 12:36:47 docsrs cratesfyi[28601]: 4: docs_rs::build_queue::BuildQueue::process_next_crate Feb 21 12:36:47 docsrs cratesfyi[28601]: 5: <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: 6: docs_rs::utils::queue_builder::queue_builder Feb 21 12:36:47 docsrs cratesfyi[28601]: 7: std::sys_common::backtrace::__rust_begin_short_backtrace Feb 21 12:36:47 docsrs cratesfyi[28601]: 8: core::ops::function::FnOnce::call_once{{vtable.shim}} Feb 21 12:36:47 docsrs cratesfyi[28601]: 9: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076 Feb 21 12:36:47 docsrs cratesfyi[28601]: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076 Feb 21 12:36:47 docsrs cratesfyi[28601]: std::sys::unix::thread::Thread::new::thread_start Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/libstd/sys/unix/thread.rs:87 Feb 21 12:36:47 docsrs cratesfyi[28601]: 10: start_thread Feb 21 12:36:47 docsrs cratesfyi[28601]: 11: __clone
1.0
"Failed to parse manifest" on bootloader-0.10.0-alpha-03 - Not sure what's going wrong here, `cargo doc` works locally. The crate doesn't show up in the web UI at all. ``` Feb 21 12:36:47 docsrs cratesfyi[28601]: 2021/02/21 12:36:47 [ERROR] docs_rs::build_queue: Failed to build package bootloader-0.10.0-alpha-03 from queue: failed to parse maniest Feb 21 12:36:47 docsrs cratesfyi[28601]: Backtrace: 0: failure::backtrace::internal::InternalBacktrace::new Feb 21 12:36:47 docsrs cratesfyi[28601]: 1: <failure::backtrace::Backtrace as core::default::Default>::default Feb 21 12:36:47 docsrs cratesfyi[28601]: 2: rustwide::build::BuildBuilder::run Feb 21 12:36:47 docsrs cratesfyi[28601]: 3: docs_rs::docbuilder::rustwide_builder::RustwideBuilder::build_package Feb 21 12:36:47 docsrs cratesfyi[28601]: 4: docs_rs::build_queue::BuildQueue::process_next_crate Feb 21 12:36:47 docsrs cratesfyi[28601]: 5: <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: 6: docs_rs::utils::queue_builder::queue_builder Feb 21 12:36:47 docsrs cratesfyi[28601]: 7: std::sys_common::backtrace::__rust_begin_short_backtrace Feb 21 12:36:47 docsrs cratesfyi[28601]: 8: core::ops::function::FnOnce::call_once{{vtable.shim}} Feb 21 12:36:47 docsrs cratesfyi[28601]: 9: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076 Feb 21 12:36:47 docsrs cratesfyi[28601]: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/liballoc/boxed.rs:1076 Feb 21 12:36:47 docsrs cratesfyi[28601]: std::sys::unix::thread::Thread::new::thread_start Feb 21 12:36:47 docsrs cratesfyi[28601]: at /rustc/d3fb005a39e62501b8b0b356166e515ae24e2e54/src/libstd/sys/unix/thread.rs:87 Feb 21 12:36:47 docsrs cratesfyi[28601]: 10: start_thread Feb 21 12:36:47 docsrs cratesfyi[28601]: 11: __clone
non_priority
failed to parse manifest on bootloader alpha not sure what s going wrong here cargo doc works locally the crate doesn t show up in the web ui at all feb docsrs cratesfyi docs rs build queue failed to build package bootloader alpha from queue failed to parse maniest feb docsrs cratesfyi backtrace failure backtrace internal internalbacktrace new feb docsrs cratesfyi default feb docsrs cratesfyi rustwide build buildbuilder run feb docsrs cratesfyi docs rs docbuilder rustwide builder rustwidebuilder build package feb docsrs cratesfyi docs rs build queue buildqueue process next crate feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi docs rs utils queue builder queue builder feb docsrs cratesfyi std sys common backtrace rust begin short backtrace feb docsrs cratesfyi core ops function fnonce call once vtable shim feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi at rustc src liballoc boxed rs feb docsrs cratesfyi as core ops function fnonce call once feb docsrs cratesfyi at rustc src liballoc boxed rs feb docsrs cratesfyi std sys unix thread thread new thread start feb docsrs cratesfyi at rustc src libstd sys unix thread rs feb docsrs cratesfyi start thread feb docsrs cratesfyi clone
0
522,797
15,168,235,481
IssuesEvent
2021-02-12 19:04:18
FTBTeam/FTB-Quests
https://api.github.com/repos/FTBTeam/FTB-Quests
closed
Completely Broken Quest GUI other than one corner
bug high priority
ftb lib ver:5.4.1.77 ftb quest ver:1.4.1.115 screen shots ![](https://user-images.githubusercontent.com/42865598/47655575-ef86d080-db95-11e8-8375-b71358c804f3.png) :
1.0
Completely Broken Quest GUI other than one corner - ftb lib ver:5.4.1.77 ftb quest ver:1.4.1.115 screen shots ![](https://user-images.githubusercontent.com/42865598/47655575-ef86d080-db95-11e8-8375-b71358c804f3.png) :
priority
completely broken quest gui other than one corner ftb lib ver ftb quest ver screen shots
1
270,475
23,511,581,166
IssuesEvent
2022-08-18 17:04:34
PowerShell/Microsoft.PowerShell.Archive
https://api.github.com/repos/PowerShell/Microsoft.PowerShell.Archive
closed
Expand-Archive Tests
P0 Size: Large Area - Archive Format Test Scenario Area - Symlinks Area - Error Handling
This is a mega-issue that divides the tests to be added to Expand-Archive. This list is not complete, and more tests will be added over time. [x] indicates the test has been added to `Expand-Archive.Tests.ps1`. [ ] indices the test has not yet been added. All tests will be written in Pester 5. Old tests will be updated. # Parameter set validation tests - [x] Validate errors with NULL & EMPTY values for Path, LiteralPath, and DestinationPath - [x] Throws when invalid path non-existing path is supplied for Path or LiteralPath parameters - [x] Throws when invalid path non-filesystem path is supplied for Path or LiteralPath parameters - [x] Throws an error when multiple paths are supplied as input to Path parameter - [x] Throws an error when multiple paths are supplied as input to LiteralPath parameter - [x] Throws an error when Path and DestinationPath are the same and -WriteMode Overwrite is specified - [x] Throws an error when LiteralPath and DestinationPath are the same and WriteMode -Overwrite is specified - [x] Throws an error when an invalid path is supplied to DestinationPath # `-DestinationPath` parameter tests - [x] Throws an error when DestinationPath is an existing file - [x] Does not throw an error when a directory in the archive has the same destination path as an existing directory - [x] Writes a non-terminating error when a file in the archive has a destination path that already exists - [x] Writes a non-terminating error when a file in the archive has a destination path that is an existing directory containing at least 1 item and -WriteMode Overwrite is specified - [x] Writes a non-terminating error when a file in the archive has a destination path that is the working directory and -WriteMode Overwrite is specified - [x] Overwrites a file when it is DestinationPath and -WriteMode Overwrite is specified - [x] Overwrites a file whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified - [x] Overwrites a directory whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified - [x] Overwrites a file whose path is the same as the destination path of a directory in the archive when -WriteMode Overwrite is specified # Relative paths - [x] Expands an archive when -Path is a relative path - [x] Expands an archive when -LiteralPath is a relative path - [x] Expands an archive when -DestinationPath is a relative path # Special and Wildcard Characters - [x] Expands an archive when -Path contains wildcard character and resolves to 1 path - [x] Throws a terminating error when archive when -Path contains wildcard character and resolves to multiple paths - [x] Expands an archive when -LiteralPath contains [ but no matching ] - [x] Expands an archive when -DestinationPath contains [ but no matching ] # Basic functional tests - [x] Expands an archive when a non-existent directory is specified as -DestinationPath - [x] Expands an archive when DestinationPath is an existing directory - [x] Expands an archive to the working directory when it is specified as -DestinationPath - [x] Expands an archive to a directory with that archive's name when -DestinationPath is not specified - [x] Expands an archive containing multiple files, non-empty directories, and empty directories - [x] Expands an archive containing a file whose LastWriteTime is in the past - [x] Expands an archive containing a directory whose LastWriteTime is in the past - [x] Throws an error when expanding an archive whose name does not have an extension and -DestinationPath is not specified # `-PassThru` parameter - [x] Returns a System.IO.DirectoryInfo object when PassThru is specified - [x] Does not return an object when PassThru is not specified - [x] Does not return an object when PassThru is false # File permissions, locked files, hidden files, symbolic links, etc. - [x] Expands a read-only archive - [x] Expands an archive in-use - [x] Expands an archive containing an entry with non-latin characters # Pipelining - [x] Expands an archive when -Path is passed by pipeline - [x] Expands an archive when -Path is passed by pipeline by name - [x] Throws an error when multiple paths are passed by pipeline # Large file tests - [x] Expands an archive whose size is > 4GB - [x] Expands an entry whose size is > 4GB # Archive formats and `-Format` parameter - [x] Throws an error when an invalid value is supplied to -Format ## Module - [x] Validate module can be imported when current language is not en-US (1247)
1.0
Expand-Archive Tests - This is a mega-issue that divides the tests to be added to Expand-Archive. This list is not complete, and more tests will be added over time. [x] indicates the test has been added to `Expand-Archive.Tests.ps1`. [ ] indices the test has not yet been added. All tests will be written in Pester 5. Old tests will be updated. # Parameter set validation tests - [x] Validate errors with NULL & EMPTY values for Path, LiteralPath, and DestinationPath - [x] Throws when invalid path non-existing path is supplied for Path or LiteralPath parameters - [x] Throws when invalid path non-filesystem path is supplied for Path or LiteralPath parameters - [x] Throws an error when multiple paths are supplied as input to Path parameter - [x] Throws an error when multiple paths are supplied as input to LiteralPath parameter - [x] Throws an error when Path and DestinationPath are the same and -WriteMode Overwrite is specified - [x] Throws an error when LiteralPath and DestinationPath are the same and WriteMode -Overwrite is specified - [x] Throws an error when an invalid path is supplied to DestinationPath # `-DestinationPath` parameter tests - [x] Throws an error when DestinationPath is an existing file - [x] Does not throw an error when a directory in the archive has the same destination path as an existing directory - [x] Writes a non-terminating error when a file in the archive has a destination path that already exists - [x] Writes a non-terminating error when a file in the archive has a destination path that is an existing directory containing at least 1 item and -WriteMode Overwrite is specified - [x] Writes a non-terminating error when a file in the archive has a destination path that is the working directory and -WriteMode Overwrite is specified - [x] Overwrites a file when it is DestinationPath and -WriteMode Overwrite is specified - [x] Overwrites a file whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified - [x] Overwrites a directory whose path is the same as the destination path of a file in the archive when -WriteMode Overwrite is specified - [x] Overwrites a file whose path is the same as the destination path of a directory in the archive when -WriteMode Overwrite is specified # Relative paths - [x] Expands an archive when -Path is a relative path - [x] Expands an archive when -LiteralPath is a relative path - [x] Expands an archive when -DestinationPath is a relative path # Special and Wildcard Characters - [x] Expands an archive when -Path contains wildcard character and resolves to 1 path - [x] Throws a terminating error when archive when -Path contains wildcard character and resolves to multiple paths - [x] Expands an archive when -LiteralPath contains [ but no matching ] - [x] Expands an archive when -DestinationPath contains [ but no matching ] # Basic functional tests - [x] Expands an archive when a non-existent directory is specified as -DestinationPath - [x] Expands an archive when DestinationPath is an existing directory - [x] Expands an archive to the working directory when it is specified as -DestinationPath - [x] Expands an archive to a directory with that archive's name when -DestinationPath is not specified - [x] Expands an archive containing multiple files, non-empty directories, and empty directories - [x] Expands an archive containing a file whose LastWriteTime is in the past - [x] Expands an archive containing a directory whose LastWriteTime is in the past - [x] Throws an error when expanding an archive whose name does not have an extension and -DestinationPath is not specified # `-PassThru` parameter - [x] Returns a System.IO.DirectoryInfo object when PassThru is specified - [x] Does not return an object when PassThru is not specified - [x] Does not return an object when PassThru is false # File permissions, locked files, hidden files, symbolic links, etc. - [x] Expands a read-only archive - [x] Expands an archive in-use - [x] Expands an archive containing an entry with non-latin characters # Pipelining - [x] Expands an archive when -Path is passed by pipeline - [x] Expands an archive when -Path is passed by pipeline by name - [x] Throws an error when multiple paths are passed by pipeline # Large file tests - [x] Expands an archive whose size is > 4GB - [x] Expands an entry whose size is > 4GB # Archive formats and `-Format` parameter - [x] Throws an error when an invalid value is supplied to -Format ## Module - [x] Validate module can be imported when current language is not en-US (1247)
non_priority
expand archive tests this is a mega issue that divides the tests to be added to expand archive this list is not complete and more tests will be added over time indicates the test has been added to expand archive tests indices the test has not yet been added all tests will be written in pester old tests will be updated parameter set validation tests validate errors with null empty values for path literalpath and destinationpath throws when invalid path non existing path is supplied for path or literalpath parameters throws when invalid path non filesystem path is supplied for path or literalpath parameters throws an error when multiple paths are supplied as input to path parameter throws an error when multiple paths are supplied as input to literalpath parameter throws an error when path and destinationpath are the same and writemode overwrite is specified throws an error when literalpath and destinationpath are the same and writemode overwrite is specified throws an error when an invalid path is supplied to destinationpath destinationpath parameter tests throws an error when destinationpath is an existing file does not throw an error when a directory in the archive has the same destination path as an existing directory writes a non terminating error when a file in the archive has a destination path that already exists writes a non terminating error when a file in the archive has a destination path that is an existing directory containing at least item and writemode overwrite is specified writes a non terminating error when a file in the archive has a destination path that is the working directory and writemode overwrite is specified overwrites a file when it is destinationpath and writemode overwrite is specified overwrites a file whose path is the same as the destination path of a file in the archive when writemode overwrite is specified overwrites a directory whose path is the same as the destination path of a file in the archive when writemode overwrite is specified overwrites a file whose path is the same as the destination path of a directory in the archive when writemode overwrite is specified relative paths expands an archive when path is a relative path expands an archive when literalpath is a relative path expands an archive when destinationpath is a relative path special and wildcard characters expands an archive when path contains wildcard character and resolves to path throws a terminating error when archive when path contains wildcard character and resolves to multiple paths expands an archive when literalpath contains expands an archive when destinationpath contains basic functional tests expands an archive when a non existent directory is specified as destinationpath expands an archive when destinationpath is an existing directory expands an archive to the working directory when it is specified as destinationpath expands an archive to a directory with that archive s name when destinationpath is not specified expands an archive containing multiple files non empty directories and empty directories expands an archive containing a file whose lastwritetime is in the past expands an archive containing a directory whose lastwritetime is in the past throws an error when expanding an archive whose name does not have an extension and destinationpath is not specified passthru parameter returns a system io directoryinfo object when passthru is specified does not return an object when passthru is not specified does not return an object when passthru is false file permissions locked files hidden files symbolic links etc expands a read only archive expands an archive in use expands an archive containing an entry with non latin characters pipelining expands an archive when path is passed by pipeline expands an archive when path is passed by pipeline by name throws an error when multiple paths are passed by pipeline large file tests expands an archive whose size is expands an entry whose size is archive formats and format parameter throws an error when an invalid value is supplied to format module validate module can be imported when current language is not en us
0
106,028
9,106,411,740
IssuesEvent
2019-02-20 23:47:02
unfoldingWord-dev/translationCore
https://api.github.com/repos/unfoldingWord-dev/translationCore
closed
Don't show any alignment boxes when the verse is blank and there is not Greek text
Kind/Enhancement QA/Retest
0.8.1-alpha.26-61679a7 ![image.png](https://images.zenhubusercontent.com/56e1a9316caf56c9d38d8df5/612c2c67-dc2d-4dec-9ee8-76afc0869867)
1.0
Don't show any alignment boxes when the verse is blank and there is not Greek text - 0.8.1-alpha.26-61679a7 ![image.png](https://images.zenhubusercontent.com/56e1a9316caf56c9d38d8df5/612c2c67-dc2d-4dec-9ee8-76afc0869867)
non_priority
don t show any alignment boxes when the verse is blank and there is not greek text alpha
0
86,767
10,817,538,252
IssuesEvent
2019-11-08 09:57:46
Marc-AntoineA/A2P
https://api.github.com/repos/Marc-AntoineA/A2P
opened
Updating supervisor welcome view
design supervisor-frontend
The supervisor welcome view was never done and is still in *todo* state. This view should explain the features of the website and give an overview of the opened processes, etc.
1.0
Updating supervisor welcome view - The supervisor welcome view was never done and is still in *todo* state. This view should explain the features of the website and give an overview of the opened processes, etc.
non_priority
updating supervisor welcome view the supervisor welcome view was never done and is still in todo state this view should explain the features of the website and give an overview of the opened processes etc
0
101,325
8,785,333,096
IssuesEvent
2018-12-20 12:40:10
linkedpipes/dcat-ap-viewer
https://api.github.com/repos/linkedpipes/dcat-ap-viewer
closed
Links in navigation do not work in Firefox (NK-11)
bug test
See [here](https://nkod.nakit.opendata.cz/datov%C3%A9-sady) in FF. Related issue was already resolved in data.gov.cz.
1.0
Links in navigation do not work in Firefox (NK-11) - See [here](https://nkod.nakit.opendata.cz/datov%C3%A9-sady) in FF. Related issue was already resolved in data.gov.cz.
non_priority
links in navigation do not work in firefox nk see in ff related issue was already resolved in data gov cz
0
160,942
6,105,811,531
IssuesEvent
2017-06-21 01:16:47
ProjectSidewalk/SidewalkWebpage
https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage
closed
After jump, user is always 180 degree turn away from the right direction
potential-intern-assignment Priority: Low pull-request-submitted ui-update
When the user is jumped into a neighborhood, he/she is always facing the wrong way and, hence, has to turn all the way around (u-turn). Reported by 2 CMSC434 undergrads. ![jump_wrong_way](https://cloud.githubusercontent.com/assets/2873216/19206162/f47c9d72-8cb4-11e6-9fcd-a09a85cd5b16.png)
1.0
After jump, user is always 180 degree turn away from the right direction - When the user is jumped into a neighborhood, he/she is always facing the wrong way and, hence, has to turn all the way around (u-turn). Reported by 2 CMSC434 undergrads. ![jump_wrong_way](https://cloud.githubusercontent.com/assets/2873216/19206162/f47c9d72-8cb4-11e6-9fcd-a09a85cd5b16.png)
priority
after jump user is always degree turn away from the right direction when the user is jumped into a neighborhood he she is always facing the wrong way and hence has to turn all the way around u turn reported by undergrads
1
38,845
2,850,497,058
IssuesEvent
2015-05-31 16:39:30
damonkohler/android-scripting
https://api.github.com/repos/damonkohler/android-scripting
closed
WebView resource paths should be relative to script's location, not SL4A's
auto-migrated Priority-Medium Type-Enhancement
``` When displaying a WebView, relative paths don't work as you'd expect if the script is in a subdirectory. e.g.: 1. Create a folder in SL4A named test. 2. In that folder, place two files index.html and main.js. 3. In index.html, insert the line: <script src="main.js"></script>. (You might need head and body too?) 4. In main.is, write: alert("hello") 5. Open index.html. No alert is displayed. 6. Change the include line in index.html to: <script src="test/main.js"></script> 7. Open index.html again; the alert now appears. This means that relative paths are based at /mnt/sdcard/sl4a/scripts, rather than /mnt/sdcard/sl4a/scripts/test. This seems odd to me - a script's working directory should be the one containing it, so that it can access its resources without having to know about the directory structure above it. Otherwise the name of the folder has to be hardcoded into every resource path, which means it can't be moved, renamed, or copied as a base for another script without a lot of search-and-replace. ``` Original issue reported on code.google.com by `hyperhac...@gmail.com` on 14 Jul 2012 at 8:20
1.0
WebView resource paths should be relative to script's location, not SL4A's - ``` When displaying a WebView, relative paths don't work as you'd expect if the script is in a subdirectory. e.g.: 1. Create a folder in SL4A named test. 2. In that folder, place two files index.html and main.js. 3. In index.html, insert the line: <script src="main.js"></script>. (You might need head and body too?) 4. In main.is, write: alert("hello") 5. Open index.html. No alert is displayed. 6. Change the include line in index.html to: <script src="test/main.js"></script> 7. Open index.html again; the alert now appears. This means that relative paths are based at /mnt/sdcard/sl4a/scripts, rather than /mnt/sdcard/sl4a/scripts/test. This seems odd to me - a script's working directory should be the one containing it, so that it can access its resources without having to know about the directory structure above it. Otherwise the name of the folder has to be hardcoded into every resource path, which means it can't be moved, renamed, or copied as a base for another script without a lot of search-and-replace. ``` Original issue reported on code.google.com by `hyperhac...@gmail.com` on 14 Jul 2012 at 8:20
priority
webview resource paths should be relative to script s location not s when displaying a webview relative paths don t work as you d expect if the script is in a subdirectory e g create a folder in named test in that folder place two files index html and main js in index html insert the line you might need head and body too in main is write alert hello open index html no alert is displayed change the include line in index html to script src test main js open index html again the alert now appears this means that relative paths are based at mnt sdcard scripts rather than mnt sdcard scripts test this seems odd to me a script s working directory should be the one containing it so that it can access its resources without having to know about the directory structure above it otherwise the name of the folder has to be hardcoded into every resource path which means it can t be moved renamed or copied as a base for another script without a lot of search and replace original issue reported on code google com by hyperhac gmail com on jul at
1
358,876
10,651,238,205
IssuesEvent
2019-10-17 09:56:48
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
opened
NetworkInterface::periodicStatsUpdate() requires cleanup
enhancement priority ticket
The above method is not just updating statistics as the name says. Instead it also purges entries. Said that its name should change as misleading, on flow-based interfaces the call frequency to this method depends on the flow timeout. While this is acceptable for updates, it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency. A minimal solution is to call idle at a different frequency. The good solution is to avoid walking the whole hash searching for idle hosts, but rather to accumulate the idle hosts on a vector and thus purge them up without walking.
1.0
NetworkInterface::periodicStatsUpdate() requires cleanup - The above method is not just updating statistics as the name says. Instead it also purges entries. Said that its name should change as misleading, on flow-based interfaces the call frequency to this method depends on the flow timeout. While this is acceptable for updates, it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency. A minimal solution is to call idle at a different frequency. The good solution is to avoid walking the whole hash searching for idle hosts, but rather to accumulate the idle hosts on a vector and thus purge them up without walking.
priority
networkinterface periodicstatsupdate requires cleanup the above method is not just updating statistics as the name says instead it also purges entries said that its name should change as misleading on flow based interfaces the call frequency to this method depends on the flow timeout while this is acceptable for updates it is wrong for purging and script callbacks as this needs to be done regardless of the update frequency a minimal solution is to call idle at a different frequency the good solution is to avoid walking the whole hash searching for idle hosts but rather to accumulate the idle hosts on a vector and thus purge them up without walking
1
421,073
12,248,493,906
IssuesEvent
2020-05-05 17:35:48
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
cancel call fails to kick CQ under epoll1
kind/bug lang/c++ priority/P2
communication for internal issue b/144444259 User reported all threads blocked on C++ CQ dtor waiting for the pending rpc's to finish. The pending rpcs were cancelled previously. The user is using epoll1.
1.0
cancel call fails to kick CQ under epoll1 - communication for internal issue b/144444259 User reported all threads blocked on C++ CQ dtor waiting for the pending rpc's to finish. The pending rpcs were cancelled previously. The user is using epoll1.
priority
cancel call fails to kick cq under communication for internal issue b user reported all threads blocked on c cq dtor waiting for the pending rpc s to finish the pending rpcs were cancelled previously the user is using
1
221,227
17,314,014,400
IssuesEvent
2021-07-27 01:44:59
logan-giese/whiteboard-course-management
https://api.github.com/repos/logan-giese/whiteboard-course-management
closed
Create unit tests for assignment services
testing
- [x] Create unit tests for assignment fetching (getting all assignments in a course and by ID) - [x] Create unit tests for assignment creation - [x] Create unit tests for assignment updates - [x] Create unit tests for assignment deletion (delete the specific assignment that was created in the creation test)
1.0
Create unit tests for assignment services - - [x] Create unit tests for assignment fetching (getting all assignments in a course and by ID) - [x] Create unit tests for assignment creation - [x] Create unit tests for assignment updates - [x] Create unit tests for assignment deletion (delete the specific assignment that was created in the creation test)
non_priority
create unit tests for assignment services create unit tests for assignment fetching getting all assignments in a course and by id create unit tests for assignment creation create unit tests for assignment updates create unit tests for assignment deletion delete the specific assignment that was created in the creation test
0
55,425
14,451,298,226
IssuesEvent
2020-12-08 10:46:36
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
opened
Severe error calculated density of air is negative
Defect
Issue overview -------------- user reported on unmet hours that he got the following error: ``` ** Severe ** PsyRhoAirFnPbTdbW: RhoAir (Density of Air) is calculated <= 0 [-9.04677]. ** ~~~ ** pb =[101100.00], tdb=[-312.09], w=[0.0000000]. ** ~~~ ** Routine=CorrectZoneHumRat, During Warmup, Environment=DSDAYTANGIER COOLING 0.4%, at Simulation time=08/21 00:12 - 00:13 ** Fatal ** Program terminates due to preceding condition. ``` user supplied the file, which was at E+ 8.7. I updated the file to v9.4 and instead of getting this severe I am getting an actual crash, which isn't something we want to happen, even if it is later found the defect file had problems. ``` Performing Zone Sizing Simulation ...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4% Warming up Warming up Warming up Performing Zone Sizing Simulation ...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6% Calculating System sizing ...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4% Calculating System sizing ...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6% Adjusting Air System Sizing Adjusting Standard 62.1 Ventilation Sizing Initializing Simulation double free or corruption (out) Aborted (core dumped) ``` ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link: https://unmethours.com/question/49383/severe-error-calculated-density-of-air-is-negative/ Backtrace ``` (lldb) bt * thread #1, name = 'energyplus', stop reason = hit program assert frame #0: 0x00007ffff11c7fb7 libc.so.6`__GI_raise(sig=<unavailable>) at raise.c:51 frame #1: 0x00007ffff11c9921 libc.so.6`__GI_abort at abort.c:79 frame #2: 0x00007ffff11b948a libc.so.6`__assert_fail_base(fmt="%s%s%s:%u: %s%sAssertion `%s' failed.\n%n", assertion="contains( i )", file="/home/julien/Software/Others/EnergyPlus/third_party/ObjexxFCL/src/ObjexxFCL/Array1.hh", line=1172, function="T& ObjexxFCL::Array1< <template-parameter-1-1> >::operator()(int) [with T = EnergyPlus::WaterThermalTanks::StratifiedNodeData]") at assert.c:92 frame #3: 0x00007ffff11b9502 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) at assert.c:101 * frame #4: 0x00007ffff4d1a4b1 libenergyplusapi.so.9.4.0`ObjexxFCL::Array1<EnergyPlus::WaterThermalTanks::StratifiedNodeData>::operator(this=0x00005555564ff530, i=0)(int) at Array1.hh:1172 frame #5: 0x00007ffff4cea195 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcWaterThermalTankStratified(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:7218 frame #6: 0x00007ffff4d11882 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcStandardRatings(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:11451 frame #7: 0x00007ffff4ce47d4 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::initialize(this=0x00005555564ff100, state=0x00007fffffffbdf0, FirstHVACIteration=true) at WaterThermalTanks.cc:6183 frame #8: 0x00007ffff4c8c218 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::onInitLoopEquip(this=0x00005555564ff100, state=0x00007fffffffbdf0, calledFromLocation=0x0000555555e45098) at WaterThermalTanks.cc:164 frame #9: 0x00007ffff5ffa702 libenergyplusapi.so.9.4.0`EnergyPlus::DataPlant::CompData::simulate(this=0x0000555555e44f00, state=0x00007fffffffbdf0, FirstHVACIteration=true, InitLoopEquip=0x00007ffff7d6fbdf, GetCompSizFac=true) at Component.cc:119 frame #10: 0x00007ffff5cf1f1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::InitializeLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true) at PlantManager.cc:2204 frame #11: 0x00007ffff5cdbd1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::ManagePlantLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a) at PlantManager.cc:222 frame #12: 0x00007ffff5827b00 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimSelectedEquipment(state=0x00007fffffffbdf0, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a, FirstHVACIteration=0x00007fffffffae2e, LockPlantFlows=false) at HVACManager.cc:1832 frame #13: 0x00007ffff581d054 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:842 frame #14: 0x00007ffff58197f4 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::ManageHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:358 frame #15: 0x00007ffff5a1457a libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::CalcHeatBalanceAir(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:4356 frame #16: 0x00007ffff59b9f7f libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::ManageAirHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:204 frame #17: 0x00007ffff3e451ee libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceSurfaceManager::ManageSurfaceHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceSurfaceManager.cc:281 frame #18: 0x00007ffff5a33c72 libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceManager::ManageHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceManager.cc:363 frame #19: 0x00007ffff47f9269 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::SetupSimulation(state=0x00007fffffffbdf0, ErrorsFound=0x00007fffffffbc23) at SimulationManager.cc:2111 frame #20: 0x00007ffff47ea196 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::ManageSimulation(state=0x00007fffffffbdf0) at SimulationManager.cc:366 frame #21: 0x00007ffff3a08570 libenergyplusapi.so.9.4.0`RunEnergyPlus(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:400 frame #22: 0x00007ffff3a07a17 libenergyplusapi.so.9.4.0`EnergyPlusPgm(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:224 frame #23: 0x000055555578ef66 energyplus`main(argc=6, argv=0x00007fffffffccc8) at main.cc:60 frame #24: 0x00007ffff11aabf7 libc.so.6`__libc_start_main(main=(energyplus`main at main.cc:56), argc=6, argv=0x00007fffffffccc8, init=<unavailable>, fini=<unavailable>, rtld_fini=<unavailable>, stack_end=0x00007fffffffccb8) at libc-start.c:310 frame #25: 0x000055555578ec8a energyplus`_start + 42 ``` ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [x] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Severe error calculated density of air is negative - Issue overview -------------- user reported on unmet hours that he got the following error: ``` ** Severe ** PsyRhoAirFnPbTdbW: RhoAir (Density of Air) is calculated <= 0 [-9.04677]. ** ~~~ ** pb =[101100.00], tdb=[-312.09], w=[0.0000000]. ** ~~~ ** Routine=CorrectZoneHumRat, During Warmup, Environment=DSDAYTANGIER COOLING 0.4%, at Simulation time=08/21 00:12 - 00:13 ** Fatal ** Program terminates due to preceding condition. ``` user supplied the file, which was at E+ 8.7. I updated the file to v9.4 and instead of getting this severe I am getting an actual crash, which isn't something we want to happen, even if it is later found the defect file had problems. ``` Performing Zone Sizing Simulation ...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4% Warming up Warming up Warming up Performing Zone Sizing Simulation ...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6% Calculating System sizing ...for Sizing Period: #1 DSDAYTANGIER COOLING 0.4% Calculating System sizing ...for Sizing Period: #2 DSDAYTANGIER HEATING 99.6% Adjusting Air System Sizing Adjusting Standard 62.1 Ventilation Sizing Initializing Simulation double free or corruption (out) Aborted (core dumped) ``` ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link: https://unmethours.com/question/49383/severe-error-calculated-density-of-air-is-negative/ Backtrace ``` (lldb) bt * thread #1, name = 'energyplus', stop reason = hit program assert frame #0: 0x00007ffff11c7fb7 libc.so.6`__GI_raise(sig=<unavailable>) at raise.c:51 frame #1: 0x00007ffff11c9921 libc.so.6`__GI_abort at abort.c:79 frame #2: 0x00007ffff11b948a libc.so.6`__assert_fail_base(fmt="%s%s%s:%u: %s%sAssertion `%s' failed.\n%n", assertion="contains( i )", file="/home/julien/Software/Others/EnergyPlus/third_party/ObjexxFCL/src/ObjexxFCL/Array1.hh", line=1172, function="T& ObjexxFCL::Array1< <template-parameter-1-1> >::operator()(int) [with T = EnergyPlus::WaterThermalTanks::StratifiedNodeData]") at assert.c:92 frame #3: 0x00007ffff11b9502 libc.so.6`__GI___assert_fail(assertion=<unavailable>, file=<unavailable>, line=<unavailable>, function=<unavailable>) at assert.c:101 * frame #4: 0x00007ffff4d1a4b1 libenergyplusapi.so.9.4.0`ObjexxFCL::Array1<EnergyPlus::WaterThermalTanks::StratifiedNodeData>::operator(this=0x00005555564ff530, i=0)(int) at Array1.hh:1172 frame #5: 0x00007ffff4cea195 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcWaterThermalTankStratified(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:7218 frame #6: 0x00007ffff4d11882 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::CalcStandardRatings(this=0x00005555564ff100, state=0x00007fffffffbdf0) at WaterThermalTanks.cc:11451 frame #7: 0x00007ffff4ce47d4 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::initialize(this=0x00005555564ff100, state=0x00007fffffffbdf0, FirstHVACIteration=true) at WaterThermalTanks.cc:6183 frame #8: 0x00007ffff4c8c218 libenergyplusapi.so.9.4.0`EnergyPlus::WaterThermalTanks::WaterThermalTankData::onInitLoopEquip(this=0x00005555564ff100, state=0x00007fffffffbdf0, calledFromLocation=0x0000555555e45098) at WaterThermalTanks.cc:164 frame #9: 0x00007ffff5ffa702 libenergyplusapi.so.9.4.0`EnergyPlus::DataPlant::CompData::simulate(this=0x0000555555e44f00, state=0x00007fffffffbdf0, FirstHVACIteration=true, InitLoopEquip=0x00007ffff7d6fbdf, GetCompSizFac=true) at Component.cc:119 frame #10: 0x00007ffff5cf1f1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::InitializeLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true) at PlantManager.cc:2204 frame #11: 0x00007ffff5cdbd1a libenergyplusapi.so.9.4.0`EnergyPlus::PlantManager::ManagePlantLoops(state=0x00007fffffffbdf0, FirstHVACIteration=true, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a) at PlantManager.cc:222 frame #12: 0x00007ffff5827b00 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimSelectedEquipment(state=0x00007fffffffbdf0, SimAirLoops=0x00007ffff7d77439, SimZoneEquipment=0x00007ffff7d7743c, SimNonZoneEquipment=0x00007ffff7d7743d, SimPlantLoops=0x00007ffff7d7743b, SimElecCircuits=0x00007ffff7d7743a, FirstHVACIteration=0x00007fffffffae2e, LockPlantFlows=false) at HVACManager.cc:1832 frame #13: 0x00007ffff581d054 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::SimHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:842 frame #14: 0x00007ffff58197f4 libenergyplusapi.so.9.4.0`EnergyPlus::HVACManager::ManageHVAC(state=0x00007fffffffbdf0) at HVACManager.cc:358 frame #15: 0x00007ffff5a1457a libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::CalcHeatBalanceAir(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:4356 frame #16: 0x00007ffff59b9f7f libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceAirManager::ManageAirHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceAirManager.cc:204 frame #17: 0x00007ffff3e451ee libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceSurfaceManager::ManageSurfaceHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceSurfaceManager.cc:281 frame #18: 0x00007ffff5a33c72 libenergyplusapi.so.9.4.0`EnergyPlus::HeatBalanceManager::ManageHeatBalance(state=0x00007fffffffbdf0) at HeatBalanceManager.cc:363 frame #19: 0x00007ffff47f9269 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::SetupSimulation(state=0x00007fffffffbdf0, ErrorsFound=0x00007fffffffbc23) at SimulationManager.cc:2111 frame #20: 0x00007ffff47ea196 libenergyplusapi.so.9.4.0`EnergyPlus::SimulationManager::ManageSimulation(state=0x00007fffffffbdf0) at SimulationManager.cc:366 frame #21: 0x00007ffff3a08570 libenergyplusapi.so.9.4.0`RunEnergyPlus(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:400 frame #22: 0x00007ffff3a07a17 libenergyplusapi.so.9.4.0`EnergyPlusPgm(state=0x00007fffffffbdf0, filepath="\xe0\xbd\xff\xff\xff\U0000007f"...) at EnergyPlusPgm.cc:224 frame #23: 0x000055555578ef66 energyplus`main(argc=6, argv=0x00007fffffffccc8) at main.cc:60 frame #24: 0x00007ffff11aabf7 libc.so.6`__libc_start_main(main=(energyplus`main at main.cc:56), argc=6, argv=0x00007fffffffccc8, init=<unavailable>, fini=<unavailable>, rtld_fini=<unavailable>, stack_end=0x00007fffffffccb8) at libc-start.c:310 frame #25: 0x000055555578ec8a energyplus`_start + 42 ``` ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [x] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
non_priority
severe error calculated density of air is negative issue overview user reported on unmet hours that he got the following error severe psyrhoairfnpbtdbw rhoair density of air is calculated pb tdb w routine correctzonehumrat during warmup environment dsdaytangier cooling at simulation time fatal program terminates due to preceding condition user supplied the file which was at e i updated the file to and instead of getting this severe i am getting an actual crash which isn t something we want to happen even if it is later found the defect file had problems performing zone sizing simulation for sizing period dsdaytangier cooling warming up warming up warming up performing zone sizing simulation for sizing period dsdaytangier heating calculating system sizing for sizing period dsdaytangier cooling calculating system sizing for sizing period dsdaytangier heating adjusting air system sizing adjusting standard ventilation sizing initializing simulation double free or corruption out aborted core dumped details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link backtrace lldb bt thread name energyplus stop reason hit program assert frame libc so gi raise sig at raise c frame libc so gi abort at abort c frame libc so assert fail base fmt s s s u s sassertion s failed n n assertion contains i file home julien software others energyplus third party objexxfcl src objexxfcl hh line function t objexxfcl operator int at assert c frame libc so gi assert fail assertion file line function at assert c frame libenergyplusapi so objexxfcl operator this i int at hh frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata calcwaterthermaltankstratified this state at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata calcstandardratings this state at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata initialize this state firsthvaciteration true at waterthermaltanks cc frame libenergyplusapi so energyplus waterthermaltanks waterthermaltankdata oninitloopequip this state calledfromlocation at waterthermaltanks cc frame libenergyplusapi so energyplus dataplant compdata simulate this state firsthvaciteration true initloopequip getcompsizfac true at component cc frame libenergyplusapi so energyplus plantmanager initializeloops state firsthvaciteration true at plantmanager cc frame libenergyplusapi so energyplus plantmanager manageplantloops state firsthvaciteration true simairloops simzoneequipment simnonzoneequipment simplantloops simeleccircuits at plantmanager cc frame libenergyplusapi so energyplus hvacmanager simselectedequipment state simairloops simzoneequipment simnonzoneequipment simplantloops simeleccircuits firsthvaciteration lockplantflows false at hvacmanager cc frame libenergyplusapi so energyplus hvacmanager simhvac state at hvacmanager cc frame libenergyplusapi so energyplus hvacmanager managehvac state at hvacmanager cc frame libenergyplusapi so energyplus heatbalanceairmanager calcheatbalanceair state at heatbalanceairmanager cc frame libenergyplusapi so energyplus heatbalanceairmanager manageairheatbalance state at heatbalanceairmanager cc frame libenergyplusapi so energyplus heatbalancesurfacemanager managesurfaceheatbalance state at heatbalancesurfacemanager cc frame libenergyplusapi so energyplus heatbalancemanager manageheatbalance state at heatbalancemanager cc frame libenergyplusapi so energyplus simulationmanager setupsimulation state errorsfound at simulationmanager cc frame libenergyplusapi so energyplus simulationmanager managesimulation state at simulationmanager cc frame libenergyplusapi so runenergyplus state filepath xbd xff xff xff at energypluspgm cc frame libenergyplusapi so energypluspgm state filepath xbd xff xff xff at energypluspgm cc frame energyplus main argc argv at main cc frame libc so libc start main main energyplus main at main cc argc argv init fini rtld fini stack end at libc start c frame energyplus start checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
0
70,519
15,085,829,832
IssuesEvent
2021-02-05 19:18:48
mthbernardes/shaggy-rogers
https://api.github.com/repos/mthbernardes/shaggy-rogers
closed
CVE-2019-20330 (High) detected in jackson-databind-2.9.6.jar - autoclosed
security vulnerability
## CVE-2019-20330 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: shaggy-rogers/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - pantomime-2.11.0.jar (Root Library) - tika-parsers-1.19.1.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking. <p>Publish Date: 2020-01-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p> <p>Release Date: 2020-01-03</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-20330 (High) detected in jackson-databind-2.9.6.jar - autoclosed - ## CVE-2019-20330 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: shaggy-rogers/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - pantomime-2.11.0.jar (Root Library) - tika-parsers-1.19.1.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mthbernardes/shaggy-rogers/commit/f72a5cb259e01c0ac208ba3a95eee5232c30fe6c">f72a5cb259e01c0ac208ba3a95eee5232c30fe6c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking. <p>Publish Date: 2020-01-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330>CVE-2019-20330</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2526">https://github.com/FasterXML/jackson-databind/issues/2526</a></p> <p>Release Date: 2020-01-03</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.novemberain:pantomime:2.11.0;org.apache.tika:tika-parsers:1.19.1;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.5,2.9.10.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20330","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.2 lacks certain net.sf.ehcache blocking.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20330","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file shaggy rogers pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy pantomime jar root library tika parsers jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before lacks certain net sf ehcache blocking publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com novemberain pantomime org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before lacks certain net sf ehcache blocking vulnerabilityurl
0
631,931
20,165,085,765
IssuesEvent
2022-02-10 02:53:16
ms-club-sliit/msclubwesite-v2
https://api.github.com/repos/ms-club-sliit/msclubwesite-v2
closed
Scroll to top alignment issue
bug help wanted high-priority
## Issue ⚠️ After we add the live chat :speech_balloon: button, the **Scroll to top** goes underneath the chat button. We need to make **Scroll to top** button on top of the chat button. ![image](https://user-images.githubusercontent.com/57069956/152829461-6b4dc396-ac64-469d-a0e3-76821ebcaea4.png)
1.0
Scroll to top alignment issue - ## Issue ⚠️ After we add the live chat :speech_balloon: button, the **Scroll to top** goes underneath the chat button. We need to make **Scroll to top** button on top of the chat button. ![image](https://user-images.githubusercontent.com/57069956/152829461-6b4dc396-ac64-469d-a0e3-76821ebcaea4.png)
priority
scroll to top alignment issue issue ⚠️ after we add the live chat speech balloon button the scroll to top goes underneath the chat button we need to make scroll to top button on top of the chat button
1
729,100
25,109,116,002
IssuesEvent
2022-11-08 18:58:57
deckhouse/deckhouse
https://api.github.com/repos/deckhouse/deckhouse
closed
[monitoring-kubernetes-control-plane] Control plane proxy hangs
area/monitoring type/bug status/no-stale priority/backlog
``` curl "https://172.16.29.10:10370/metrics" curl: (7) Failed to connect to 172.16.29.10 port 10370: Connection refused ``` Logs from the pod: ``` kubectl -n d8-monitoring logs control-plane-proxy-425f55b4-wnwwp Generating RSA private key, 2048 bit long modulus (2 primes) ..........+++++ ...........................+++++ e is 65537 (0x010001) writing RSA key Signature ok subject=CN = p-payment-ks-master-0 Getting Private key nginx: the configuration file /etc/nginx/nginx.conf syntax is ok nginx: configuration file /etc/nginx/nginx.conf test is successful 2021/08/18 23:40:08 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370" 2021/08/18 23:40:38 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370" ``` A possible solution is to add a liveness probe.
1.0
[monitoring-kubernetes-control-plane] Control plane proxy hangs - ``` curl "https://172.16.29.10:10370/metrics" curl: (7) Failed to connect to 172.16.29.10 port 10370: Connection refused ``` Logs from the pod: ``` kubectl -n d8-monitoring logs control-plane-proxy-425f55b4-wnwwp Generating RSA private key, 2048 bit long modulus (2 primes) ..........+++++ ...........................+++++ e is 65537 (0x010001) writing RSA key Signature ok subject=CN = p-payment-ks-master-0 Getting Private key nginx: the configuration file /etc/nginx/nginx.conf syntax is ok nginx: configuration file /etc/nginx/nginx.conf test is successful 2021/08/18 23:40:08 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370" 2021/08/18 23:40:38 [error] 16#16: *1 connect() failed (111: Connection refused) while connecting to upstream, client: 172.16.29.122, server: p-payment-ks-master-0, request: "GET /metrics HTTP/1.1", upstream: "http://127.0.0.1:2381/metrics", host: "172.16.29.10:10370" ``` A possible solution is to add a liveness probe.
priority
control plane proxy hangs curl curl failed to connect to port connection refused logs from the pod kubectl n monitoring logs control plane proxy wnwwp generating rsa private key bit long modulus primes e is writing rsa key signature ok subject cn p payment ks master getting private key nginx the configuration file etc nginx nginx conf syntax is ok nginx configuration file etc nginx nginx conf test is successful connect failed connection refused while connecting to upstream client server p payment ks master request get metrics http upstream host connect failed connection refused while connecting to upstream client server p payment ks master request get metrics http upstream host a possible solution is to add a liveness probe
1
74,118
20,022,434,007
IssuesEvent
2022-02-01 17:36:46
Start9Labs/embassy-os
https://api.github.com/repos/Start9Labs/embassy-os
closed
explore different kernels or base OS's to resolve stability issues
P0 - Blocks Dev Builds
related #901 Currently the raspberry pi has stability issues after being up for several days. Investigation into this has revealed processes are getting stuck in state `R` while being unable to be `SIGKILL`ed. According to https://unix.stackexchange.com/questions/386380/wget-cant-be-killed-with-kill-9, this is either a hardware fault or a kernel bug. Given that there is likely a problem with the kernel, or its compatibility with the pi, we need to upgrade the kernel. We have a few options here: - Upgrade the kernel in `initialization.sh` - Use the official Raspberry Pi OS as the base image (https://downloads.raspberrypi.org/raspios_arm64/images/raspios_arm64-2021-05-28/)
1.0
explore different kernels or base OS's to resolve stability issues - related #901 Currently the raspberry pi has stability issues after being up for several days. Investigation into this has revealed processes are getting stuck in state `R` while being unable to be `SIGKILL`ed. According to https://unix.stackexchange.com/questions/386380/wget-cant-be-killed-with-kill-9, this is either a hardware fault or a kernel bug. Given that there is likely a problem with the kernel, or its compatibility with the pi, we need to upgrade the kernel. We have a few options here: - Upgrade the kernel in `initialization.sh` - Use the official Raspberry Pi OS as the base image (https://downloads.raspberrypi.org/raspios_arm64/images/raspios_arm64-2021-05-28/)
non_priority
explore different kernels or base os s to resolve stability issues related currently the raspberry pi has stability issues after being up for several days investigation into this has revealed processes are getting stuck in state r while being unable to be sigkill ed according to this is either a hardware fault or a kernel bug given that there is likely a problem with the kernel or its compatibility with the pi we need to upgrade the kernel we have a few options here upgrade the kernel in initialization sh use the official raspberry pi os as the base image
0
186,207
15,051,035,627
IssuesEvent
2021-02-03 13:38:39
Urban-Analytics/RAMP-UA
https://api.github.com/repos/Urban-Analytics/RAMP-UA
opened
Documentation for the lockdown scenario(s)
documentation
In particular, how do we retrieve the google data for these?
1.0
Documentation for the lockdown scenario(s) - In particular, how do we retrieve the google data for these?
non_priority
documentation for the lockdown scenario s in particular how do we retrieve the google data for these
0
9,607
3,935,224,451
IssuesEvent
2016-04-26 03:38:02
EmergentOrganization/cell-rpg
https://api.github.com/repos/EmergentOrganization/cell-rpg
closed
weapon collision effect
cat: art cat: code enhancement question / discussion
Some effect is needed to make impact on the ca more apparent. I was thinking about having the impacted squares in the CA switch to a bright color before clearing, but we also have the option of using a particle effect or an explosion sprite. I'm not sure which of these options will look best.
1.0
weapon collision effect - Some effect is needed to make impact on the ca more apparent. I was thinking about having the impacted squares in the CA switch to a bright color before clearing, but we also have the option of using a particle effect or an explosion sprite. I'm not sure which of these options will look best.
non_priority
weapon collision effect some effect is needed to make impact on the ca more apparent i was thinking about having the impacted squares in the ca switch to a bright color before clearing but we also have the option of using a particle effect or an explosion sprite i m not sure which of these options will look best
0
105,369
4,234,717,617
IssuesEvent
2016-07-05 13:02:02
openshift/origin
https://api.github.com/repos/openshift/origin
closed
An error occurred while building 1.2(openshift_node_dnsmasq)
component/install priority/P2
Version: oepnshift-origin-1.2.0 Environment: The new system did not install anything(Centos cloud images) build: 3master+3etcd+3node+1lb[1]+DNS / 1master+2node command: ansible-playbook ~/openshift-ansible/playbooks/byo/config.yml Hi, There is a problem when I build openshift 1.2 {"changed": false, "failed": true, "msg": "Currently, NetworkManager must be installed and enabled prior to installation."} ![error-1](https://cloud.githubusercontent.com/assets/12937669/16540952/46aaf98e-40a9-11e6-822b-4ca5212418ab.jpg) ![error-2](https://cloud.githubusercontent.com/assets/12937669/16540977/f692bf80-40a9-11e6-8292-f455f69da7e7.jpg) when i use `ansible-playbook ~/openshift-ansible/playbooks/adhoc/uninstall.yml` and build again,it is ok.but if i want to add a new node,also error,how can i do for this? thanks.
1.0
An error occurred while building 1.2(openshift_node_dnsmasq) - Version: oepnshift-origin-1.2.0 Environment: The new system did not install anything(Centos cloud images) build: 3master+3etcd+3node+1lb[1]+DNS / 1master+2node command: ansible-playbook ~/openshift-ansible/playbooks/byo/config.yml Hi, There is a problem when I build openshift 1.2 {"changed": false, "failed": true, "msg": "Currently, NetworkManager must be installed and enabled prior to installation."} ![error-1](https://cloud.githubusercontent.com/assets/12937669/16540952/46aaf98e-40a9-11e6-822b-4ca5212418ab.jpg) ![error-2](https://cloud.githubusercontent.com/assets/12937669/16540977/f692bf80-40a9-11e6-8292-f455f69da7e7.jpg) when i use `ansible-playbook ~/openshift-ansible/playbooks/adhoc/uninstall.yml` and build again,it is ok.but if i want to add a new node,also error,how can i do for this? thanks.
priority
an error occurred while building openshift node dnsmasq version oepnshift origin environment the new system did not install anything centos cloud images build dns command ansible playbook openshift ansible playbooks byo config yml hi there is a problem when i build openshift changed false failed true msg currently networkmanager must be installed and enabled prior to installation when i use ansible playbook openshift ansible playbooks adhoc uninstall yml and build again it is ok but if i want to add a new node also error how can i do for this thanks
1
657,600
21,797,735,401
IssuesEvent
2022-05-15 21:40:17
kubernetes/website
https://api.github.com/repos/kubernetes/website
closed
kubectl generated command docs do not include 1.24 commands
kind/bug priority/important-soon sig/cli triage/accepted
**This is a Bug Report** <!-- Thanks for filing an issue! Before submitting, please fill in the following information. --> <!-- See https://kubernetes.io/docs/contribute/start/ for guidance on writing an actionable issue description. --> <!--Required Information--> **Problem:** https://kubernetes.io/docs/reference/generated/kubectl/kubectl-commands#create does not include commands added in 1.24 (notably `kubectl create token`) It looks like https://github.com/kubernetes/website/tree/main/static/docs/reference/generated/kubectl is supposed to be auto-generated, but has not been updated for 1.24. /sig cli
1.0
kubectl generated command docs do not include 1.24 commands - **This is a Bug Report** <!-- Thanks for filing an issue! Before submitting, please fill in the following information. --> <!-- See https://kubernetes.io/docs/contribute/start/ for guidance on writing an actionable issue description. --> <!--Required Information--> **Problem:** https://kubernetes.io/docs/reference/generated/kubectl/kubectl-commands#create does not include commands added in 1.24 (notably `kubectl create token`) It looks like https://github.com/kubernetes/website/tree/main/static/docs/reference/generated/kubectl is supposed to be auto-generated, but has not been updated for 1.24. /sig cli
priority
kubectl generated command docs do not include commands this is a bug report problem does not include commands added in notably kubectl create token it looks like is supposed to be auto generated but has not been updated for sig cli
1
15,389
8,873,144,035
IssuesEvent
2019-01-11 17:13:41
snowleopard/hadrian
https://api.github.com/repos/snowleopard/hadrian
closed
Should Hadrian build ghctags and haddock in Stage1?
enhancement performance question
The Make build system needs to build `ghctags` and `haddock` programs in Stage2. For example, see this commit: https://github.com/ghc/ghc/commit/5fb72555f7b7ab67a33583f33ad9160761ca434f. However, #531 moved `haddock` to Stage1 and it worked fine. I couldn't find an explanation behind the move and reverted this change as part of the cleaning-up-after-531 issue #540. Let's discuss this here. There is a `Note [No stage2 packages when CrossCompiling or Stage1Only]` in `ghc.mk`, which in particular says: ``` # Here's why: # - first of all, ghc-stage1 can't use stage0's ghc library (it's too old) # - neither do we register the ghc library (compiler/stage1) that we build # with stage0. TODO Why not? We do build it... # - as a result, we need to a) use ghc-stage2 to build packages that depend on # the ghc library (e.g. ghctags [4]) and b) exclude those packages when # ghc-stage2 is not available. ``` But in Hadrian we do register the `ghc` library built in Stage0, so presumably this means we could indeed now build all its dependants in Stage1. Shall we move `ghctags` and `haddock` to Stage1? This simplifies the build system and also improves the performance, since we don't need to wait for Stage2 GHC to build these utilities.
True
Should Hadrian build ghctags and haddock in Stage1? - The Make build system needs to build `ghctags` and `haddock` programs in Stage2. For example, see this commit: https://github.com/ghc/ghc/commit/5fb72555f7b7ab67a33583f33ad9160761ca434f. However, #531 moved `haddock` to Stage1 and it worked fine. I couldn't find an explanation behind the move and reverted this change as part of the cleaning-up-after-531 issue #540. Let's discuss this here. There is a `Note [No stage2 packages when CrossCompiling or Stage1Only]` in `ghc.mk`, which in particular says: ``` # Here's why: # - first of all, ghc-stage1 can't use stage0's ghc library (it's too old) # - neither do we register the ghc library (compiler/stage1) that we build # with stage0. TODO Why not? We do build it... # - as a result, we need to a) use ghc-stage2 to build packages that depend on # the ghc library (e.g. ghctags [4]) and b) exclude those packages when # ghc-stage2 is not available. ``` But in Hadrian we do register the `ghc` library built in Stage0, so presumably this means we could indeed now build all its dependants in Stage1. Shall we move `ghctags` and `haddock` to Stage1? This simplifies the build system and also improves the performance, since we don't need to wait for Stage2 GHC to build these utilities.
non_priority
should hadrian build ghctags and haddock in the make build system needs to build ghctags and haddock programs in for example see this commit however moved haddock to and it worked fine i couldn t find an explanation behind the move and reverted this change as part of the cleaning up after issue let s discuss this here there is a note in ghc mk which in particular says here s why first of all ghc can t use s ghc library it s too old neither do we register the ghc library compiler that we build with todo why not we do build it as a result we need to a use ghc to build packages that depend on the ghc library e g ghctags and b exclude those packages when ghc is not available but in hadrian we do register the ghc library built in so presumably this means we could indeed now build all its dependants in shall we move ghctags and haddock to this simplifies the build system and also improves the performance since we don t need to wait for ghc to build these utilities
0
127,891
10,492,338,984
IssuesEvent
2019-09-25 13:06:22
linkedpipes/etl
https://api.github.com/repos/linkedpipes/etl
closed
Add support for SHACL
enhancement test
Add support for [SHACL](https://www.w3.org/TR/shacl/) validation. It should be a new component (`q-shacl`) with 2 inputs - Data graph, Shapes graph and 1 output - Validation report. - There is some [initial work on SHACL support in RDF4J](https://github.com/eclipse/rdf4j-storage/pull/34), however, the support is in form of a Shacl repository, which I am not sure will be useful to our use case - The [reference implementation](https://github.com/TopQuadrant/shacl) is based on Apache Jena, so I guess such component would have to work with files, shielding the users from the implementation a bit, unless we are willing to add Jena support and all the related transformers. Related to #631
1.0
Add support for SHACL - Add support for [SHACL](https://www.w3.org/TR/shacl/) validation. It should be a new component (`q-shacl`) with 2 inputs - Data graph, Shapes graph and 1 output - Validation report. - There is some [initial work on SHACL support in RDF4J](https://github.com/eclipse/rdf4j-storage/pull/34), however, the support is in form of a Shacl repository, which I am not sure will be useful to our use case - The [reference implementation](https://github.com/TopQuadrant/shacl) is based on Apache Jena, so I guess such component would have to work with files, shielding the users from the implementation a bit, unless we are willing to add Jena support and all the related transformers. Related to #631
non_priority
add support for shacl add support for validation it should be a new component q shacl with inputs data graph shapes graph and output validation report there is some however the support is in form of a shacl repository which i am not sure will be useful to our use case the is based on apache jena so i guess such component would have to work with files shielding the users from the implementation a bit unless we are willing to add jena support and all the related transformers related to
0
253,654
27,300,776,629
IssuesEvent
2023-02-24 01:37:05
panasalap/linux-4.19.72_1
https://api.github.com/repos/panasalap/linux-4.19.72_1
closed
CVE-2019-19075 (High) detected in linux-yoctov5.4.51 - autoclosed
security vulnerability
## CVE-2019-19075 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the ca8210_probe() function in drivers/net/ieee802154/ca8210.c in the Linux kernel before 5.3.8 allows attackers to cause a denial of service (memory consumption) by triggering ca8210_get_platform_data() failures, aka CID-6402939ec86e. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19075>CVE-2019-19075</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19075 (High) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2019-19075 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ieee802154/ca8210.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the ca8210_probe() function in drivers/net/ieee802154/ca8210.c in the Linux kernel before 5.3.8 allows attackers to cause a denial of service (memory consumption) by triggering ca8210_get_platform_data() failures, aka CID-6402939ec86e. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19075>CVE-2019-19075</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19075</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in linux autoclosed cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net c drivers net c vulnerability details a memory leak in the probe function in drivers net c in the linux kernel before allows attackers to cause a denial of service memory consumption by triggering get platform data failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
1,496
21,956,078,848
IssuesEvent
2022-05-24 12:14:46
TYPO3-Solr/ext-solr
https://api.github.com/repos/TYPO3-Solr/ext-solr
closed
[BUG] Indexing records with SOLR_RELATION via backend considers hidden relation records
10LTS BACKPORTABLE
**Describe the bug** In [Backend context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/backend/Classes/Http/Application.php#L91) TYPO3 will set a VisibilityAspect that allows showing hidden pages and content. When a record is translated and the translation is then hidden, the hidden translation will still get indexed in SOLR_RELATION. This only happens when indexing is done via the TYPO3 backend, so there is a different behaviour from indexing via CLI. The suggested approach is to reset the VisibilityAspect in `Ext:solr`s Tsfe class to make sure hidden elements are not shown. Also, I was not able to reproduce this with SOLR_RELATION fields with the PageIndexer. For some reason indexing via CLI is fine even though it is also allows hidden pages and content elements (see [CLI context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/core/Classes/Console/CommandApplication.php#L185)). --- **To Reproduce** While this error was found in our internal extension, I tried my best to reproduce this with Ext:news. This test assumes English as the main language (Index: `core_en`) and German as the second language (Index: `core_de`) Steps to reproduce the behavior: 1. Create two sys_categories "Category 1" and "Category 2". Create a German translation for both of them ("Kategorie 1" and "Kategorie 2") and disable the translation of "Category 2" 2. Create a Ext:news record and assign both categories to it. Then create a German translation. Make sure both categories are selected there as well. 3. Add News indexer in TypoScript setup: ```typoscript @import 'EXT:solr/Configuration/TypoScript/Examples/IndexQueueNews/setup.typoscript' ``` 3. Index the news record via the TYPO3 backend. 4. Look at the index entries. In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],` In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Kategorie 2"],` **Expected behavior** In German index, "Kategorie 2" should not show up as it is disabled. It should behave the same as when indexing via CLI, i.e. it should index the original record. In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],` In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Category 2"],` **Used versions (please complete the following information):** - TYPO3 Version: 10.4.21 - EXT:solr Version: 11.0.2 - Used Apache Solr Version: 8.5.1, 8.9.0 - PHP Version: 7.4.20
True
[BUG] Indexing records with SOLR_RELATION via backend considers hidden relation records - **Describe the bug** In [Backend context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/backend/Classes/Http/Application.php#L91) TYPO3 will set a VisibilityAspect that allows showing hidden pages and content. When a record is translated and the translation is then hidden, the hidden translation will still get indexed in SOLR_RELATION. This only happens when indexing is done via the TYPO3 backend, so there is a different behaviour from indexing via CLI. The suggested approach is to reset the VisibilityAspect in `Ext:solr`s Tsfe class to make sure hidden elements are not shown. Also, I was not able to reproduce this with SOLR_RELATION fields with the PageIndexer. For some reason indexing via CLI is fine even though it is also allows hidden pages and content elements (see [CLI context](https://github.com/TYPO3/typo3/blob/master/typo3/sysext/core/Classes/Console/CommandApplication.php#L185)). --- **To Reproduce** While this error was found in our internal extension, I tried my best to reproduce this with Ext:news. This test assumes English as the main language (Index: `core_en`) and German as the second language (Index: `core_de`) Steps to reproduce the behavior: 1. Create two sys_categories "Category 1" and "Category 2". Create a German translation for both of them ("Kategorie 1" and "Kategorie 2") and disable the translation of "Category 2" 2. Create a Ext:news record and assign both categories to it. Then create a German translation. Make sure both categories are selected there as well. 3. Add News indexer in TypoScript setup: ```typoscript @import 'EXT:solr/Configuration/TypoScript/Examples/IndexQueueNews/setup.typoscript' ``` 3. Index the news record via the TYPO3 backend. 4. Look at the index entries. In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],` In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Kategorie 2"],` **Expected behavior** In German index, "Kategorie 2" should not show up as it is disabled. It should behave the same as when indexing via CLI, i.e. it should index the original record. In `core_en` the entry has: `"category_stringM":["Category 1", "Category 2"],` In `core_de` the entry has: `"category_stringM":["Kategorie 1", "Category 2"],` **Used versions (please complete the following information):** - TYPO3 Version: 10.4.21 - EXT:solr Version: 11.0.2 - Used Apache Solr Version: 8.5.1, 8.9.0 - PHP Version: 7.4.20
non_priority
indexing records with solr relation via backend considers hidden relation records describe the bug in will set a visibilityaspect that allows showing hidden pages and content when a record is translated and the translation is then hidden the hidden translation will still get indexed in solr relation this only happens when indexing is done via the backend so there is a different behaviour from indexing via cli the suggested approach is to reset the visibilityaspect in ext solr s tsfe class to make sure hidden elements are not shown also i was not able to reproduce this with solr relation fields with the pageindexer for some reason indexing via cli is fine even though it is also allows hidden pages and content elements see to reproduce while this error was found in our internal extension i tried my best to reproduce this with ext news this test assumes english as the main language index core en and german as the second language index core de steps to reproduce the behavior create two sys categories category and category create a german translation for both of them kategorie and kategorie and disable the translation of category create a ext news record and assign both categories to it then create a german translation make sure both categories are selected there as well add news indexer in typoscript setup typoscript import ext solr configuration typoscript examples indexqueuenews setup typoscript index the news record via the backend look at the index entries in core en the entry has category stringm in core de the entry has category stringm expected behavior in german index kategorie should not show up as it is disabled it should behave the same as when indexing via cli i e it should index the original record in core en the entry has category stringm in core de the entry has category stringm used versions please complete the following information version ext solr version used apache solr version php version
0
32,979
8,989,513,493
IssuesEvent
2019-02-01 00:11:50
orbeon/orbeon-forms
https://api.github.com/repos/orbeon/orbeon-forms
closed
Incorrect test for form definition versioning
Form Builder Form Runner
We search for the provider on the data side, but we need to search for the provider on the form definition side. This is usually not a problem, as the same provider is usually configured for form and data. This has been the case since #1669 in Orbeon Forms 4.6.
1.0
Incorrect test for form definition versioning - We search for the provider on the data side, but we need to search for the provider on the form definition side. This is usually not a problem, as the same provider is usually configured for form and data. This has been the case since #1669 in Orbeon Forms 4.6.
non_priority
incorrect test for form definition versioning we search for the provider on the data side but we need to search for the provider on the form definition side this is usually not a problem as the same provider is usually configured for form and data this has been the case since in orbeon forms
0
129,024
5,082,218,251
IssuesEvent
2016-12-29 14:37:02
vladyslav2/gfwhitelabels
https://api.github.com/repos/vladyslav2/gfwhitelabels
opened
Problems with the DCU Site
Priority
1. Communication channel not working DCU site: ![image](https://cloud.githubusercontent.com/assets/21956276/21546056/411e8632-cdaa-11e6-8240-a167d6d142f8.png) GrowthFountain Site: ![image](https://cloud.githubusercontent.com/assets/21956276/21546058/44d9725a-cdaa-11e6-9639-fd86eb421afd.png) 2. The site does not work properly in mobile 3. I cannot login (when I tried to invest it brought me to this login page which doesn’t work) 4. Hide social sign in on login page ![image](https://cloud.githubusercontent.com/assets/21956276/21546063/494763d8-cdaa-11e6-95e0-41368fad0ec2.png) 5. When I click “RAISE CAPITAL” on the landing page it prompts SIGN UP but it doesn’t work 6. Hide BLOG––––
1.0
Problems with the DCU Site - 1. Communication channel not working DCU site: ![image](https://cloud.githubusercontent.com/assets/21956276/21546056/411e8632-cdaa-11e6-8240-a167d6d142f8.png) GrowthFountain Site: ![image](https://cloud.githubusercontent.com/assets/21956276/21546058/44d9725a-cdaa-11e6-9639-fd86eb421afd.png) 2. The site does not work properly in mobile 3. I cannot login (when I tried to invest it brought me to this login page which doesn’t work) 4. Hide social sign in on login page ![image](https://cloud.githubusercontent.com/assets/21956276/21546063/494763d8-cdaa-11e6-95e0-41368fad0ec2.png) 5. When I click “RAISE CAPITAL” on the landing page it prompts SIGN UP but it doesn’t work 6. Hide BLOG––––
priority
problems with the dcu site communication channel not working dcu site growthfountain site the site does not work properly in mobile i cannot login when i tried to invest it brought me to this login page which doesn’t work hide social sign in on login page when i click “raise capital” on the landing page it prompts sign up but it doesn’t work hide blog––––
1
352,716
10,545,323,790
IssuesEvent
2019-10-02 18:52:49
wso2-cellery/sdk
https://api.github.com/repos/wso2-cellery/sdk
closed
Error while executing cellery test
Priority/Highest Severity/Blocker Type/Bug
cellery test wso2cellery/pet-be-cell:latest-dev -n pet-be ✔ Extracting Cell Image wso2cellery/pet-be-cell:latest-dev ✔ Reading Image wso2cellery/pet-be-cell:latest-dev ✔ Starting execution of tests for wso2cellery/pet-be-cell:latest-dev... ? Do you wish to continue with testing above Cell instances (Y/n)? ⠋ Creating telepresence instance error waiting for telepresence deployment telepresence--telepresence-deployment to be available: exit status 1 Environment: docker-for-mac
1.0
Error while executing cellery test - cellery test wso2cellery/pet-be-cell:latest-dev -n pet-be ✔ Extracting Cell Image wso2cellery/pet-be-cell:latest-dev ✔ Reading Image wso2cellery/pet-be-cell:latest-dev ✔ Starting execution of tests for wso2cellery/pet-be-cell:latest-dev... ? Do you wish to continue with testing above Cell instances (Y/n)? ⠋ Creating telepresence instance error waiting for telepresence deployment telepresence--telepresence-deployment to be available: exit status 1 Environment: docker-for-mac
priority
error while executing cellery test cellery test pet be cell latest dev n pet be ✔ extracting cell image pet be cell latest dev ✔ reading image pet be cell latest dev ✔ starting execution of tests for pet be cell latest dev do you wish to continue with testing above cell instances y n ⠋ creating telepresence instance error waiting for telepresence deployment telepresence telepresence deployment to be available exit status environment docker for mac
1
166,383
6,303,905,799
IssuesEvent
2017-07-21 14:46:03
BigBrotherTeam/BigBrother
https://api.github.com/repos/BigBrotherTeam/BigBrother
closed
Wrong sound effect is played when PC user open/close door
Category: Gameplay Priority: Low Status: Reproduced
### Issue description When any PC user open/close door (include Trapdoor), wrong sound effect (Eye of ender shoot) is played. ### Prerequisite * PocketMine-MP `g8bb0c53` * BigBrother master branch 76b03ad ### How to reproduce * Place any kind of Door block such as Wooden Door, Iron Door or Trap Door * Right click the placed Door block to open/close ### Server log(crash dump) no server log related to this issue. <!--- Thank you for reporting! -->
1.0
Wrong sound effect is played when PC user open/close door - ### Issue description When any PC user open/close door (include Trapdoor), wrong sound effect (Eye of ender shoot) is played. ### Prerequisite * PocketMine-MP `g8bb0c53` * BigBrother master branch 76b03ad ### How to reproduce * Place any kind of Door block such as Wooden Door, Iron Door or Trap Door * Right click the placed Door block to open/close ### Server log(crash dump) no server log related to this issue. <!--- Thank you for reporting! -->
priority
wrong sound effect is played when pc user open close door issue description when any pc user open close door include trapdoor wrong sound effect eye of ender shoot is played prerequisite pocketmine mp bigbrother master branch how to reproduce place any kind of door block such as wooden door iron door or trap door right click the placed door block to open close server log crash dump no server log related to this issue
1
422,853
12,287,489,295
IssuesEvent
2020-05-09 12:26:52
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for Spanner
api: spanner autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate Spanner. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-09 05:20:08 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-05-09 05:20:08,292 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api Switched to branch 'autosynth-spanner' 2020-05-09 05:20:10 [INFO] Running synthtool 2020-05-09 05:20:10,566 autosynth > Running synthtool 2020-05-09 05:20:10 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--'] 2020-05-09 05:20:10,566 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--'] 2020-05-09 05:20:10,777 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-spanner nothing to commit, working tree clean 2020-05-09 05:20:10,853 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. 2020-05-09 05:20:11,738 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Spanner 2020-05-09 05:20:15,459 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files? Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__ write(self.metadata_file_path) File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write with open(outfile, "w") as fh: FileNotFoundError: [Errno 2] No such file or directory: 'clients/spanner/synth.metadata' 2020-05-09 05:20:15 [ERROR] Synthesis failed 2020-05-09 05:20:15,488 autosynth > Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--', 'Spanner']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
1.0
Synthesis failed for Spanner - Hello! Autosynth couldn't regenerate Spanner. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-09 05:20:08 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-05-09 05:20:08,292 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api Switched to branch 'autosynth-spanner' 2020-05-09 05:20:10 [INFO] Running synthtool 2020-05-09 05:20:10,566 autosynth > Running synthtool 2020-05-09 05:20:10 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--'] 2020-05-09 05:20:10,566 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--'] 2020-05-09 05:20:10,777 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-spanner nothing to commit, working tree clean 2020-05-09 05:20:10,853 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. 2020-05-09 05:20:11,738 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Spanner 2020-05-09 05:20:15,459 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files? Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__ write(self.metadata_file_path) File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write with open(outfile, "w") as fh: FileNotFoundError: [Errno 2] No such file or directory: 'clients/spanner/synth.metadata' 2020-05-09 05:20:15 [ERROR] Synthesis failed 2020-05-09 05:20:15,488 autosynth > Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/spanner/synth.metadata', 'synth.py', '--', 'Spanner']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
priority
synthesis failed for spanner hello autosynth couldn t regenerate spanner broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth spanner running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth spanner nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh spanner synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients spanner synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
1
26,572
26,987,909,588
IssuesEvent
2023-02-09 17:27:02
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
opened
`AssetServer` and `gltF` workflows are very prone to failing silently
A-Rendering A-Assets C-Usability
## What problem does this solve or what need does it fill? When working with `gltF` files, there are quite a few ways for something to go wrong: 1. The file doesn't exist. 2. The path is wrong. 3. You forgot to specify `#Scene0`. 4. You specified the wrong scene number. 5. Your scale is wildly wrong. 6. Your model is inside out and you have backface culling enabled. 7. Your camera isn't looking at the model. 8. Your material is fully transparent. 9. You're attempting to save the `gltF` scene as something else, possibly a `Mesh` or `StandardMaterial`, rather than a `HandleScene`. 10. You specified an invalid scene path. Currently, we detect and report 1 and 2. All other cases silently fail, with nothing (apparently) being rendered. ## What solution would you like? Add error reporting and failures to cases that are flatly wrong: 3, 8, 9, 10. Warn in cases that are possibly wrong: 4. Add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model. ## What alternative(s) have you considered? The categorization of these failure modes is very much up for debate.
True
`AssetServer` and `gltF` workflows are very prone to failing silently - ## What problem does this solve or what need does it fill? When working with `gltF` files, there are quite a few ways for something to go wrong: 1. The file doesn't exist. 2. The path is wrong. 3. You forgot to specify `#Scene0`. 4. You specified the wrong scene number. 5. Your scale is wildly wrong. 6. Your model is inside out and you have backface culling enabled. 7. Your camera isn't looking at the model. 8. Your material is fully transparent. 9. You're attempting to save the `gltF` scene as something else, possibly a `Mesh` or `StandardMaterial`, rather than a `HandleScene`. 10. You specified an invalid scene path. Currently, we detect and report 1 and 2. All other cases silently fail, with nothing (apparently) being rendered. ## What solution would you like? Add error reporting and failures to cases that are flatly wrong: 3, 8, 9, 10. Warn in cases that are possibly wrong: 4. Add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model. ## What alternative(s) have you considered? The categorization of these failure modes is very much up for debate.
non_priority
assetserver and gltf workflows are very prone to failing silently what problem does this solve or what need does it fill when working with gltf files there are quite a few ways for something to go wrong the file doesn t exist the path is wrong you forgot to specify you specified the wrong scene number your scale is wildly wrong your model is inside out and you have backface culling enabled your camera isn t looking at the model your material is fully transparent you re attempting to save the gltf scene as something else possibly a mesh or standardmaterial rather than a handlescene you specified an invalid scene path currently we detect and report and all other cases silently fail with nothing apparently being rendered what solution would you like add error reporting and failures to cases that are flatly wrong warn in cases that are possibly wrong add a prominently documented model debugging tool to look for and report on the remaining cases related to a malformed model what alternative s have you considered the categorization of these failure modes is very much up for debate
0
123,293
12,196,370,325
IssuesEvent
2020-04-29 18:57:50
COVID-19-electronic-health-system/Corona-tracker
https://api.github.com/repos/COVID-19-electronic-health-system/Corona-tracker
closed
[DOCS] Hebrew Translation
documentation
[CoronaTracker appText.zip](https://github.com/COVID-19-electronic-health-system/Corona-tracker/files/4546409/CoronaTracker.appText.zip) # ⚠️ IMPORTANT: Please fill out this template to give us as much information as possible to consider/implement this update. ### Summary <!-- One paragraph explanation of the feature. --> ### Motivation <!-- Why are we doing this? What use cases does it support? What is the expected outcome? --> ### Possible Alternatives <!-- A clear and concise description of the alternative solutions you've considered. Be sure to explain why the current documentation isn't suitable for this feature. --> ### Additional Context <!-- Add any other context or screenshots about the documentation update here. -->
1.0
[DOCS] Hebrew Translation - [CoronaTracker appText.zip](https://github.com/COVID-19-electronic-health-system/Corona-tracker/files/4546409/CoronaTracker.appText.zip) # ⚠️ IMPORTANT: Please fill out this template to give us as much information as possible to consider/implement this update. ### Summary <!-- One paragraph explanation of the feature. --> ### Motivation <!-- Why are we doing this? What use cases does it support? What is the expected outcome? --> ### Possible Alternatives <!-- A clear and concise description of the alternative solutions you've considered. Be sure to explain why the current documentation isn't suitable for this feature. --> ### Additional Context <!-- Add any other context or screenshots about the documentation update here. -->
non_priority
hebrew translation ⚠️ important please fill out this template to give us as much information as possible to consider implement this update summary motivation possible alternatives additional context
0
163,486
20,363,817,295
IssuesEvent
2022-02-21 01:31:47
rgordon95/conFusionAng
https://api.github.com/repos/rgordon95/conFusionAng
opened
CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz
security vulnerability
## CVE-2022-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p>Path to dependency file: /conFusionAng/package.json</p> <p>Path to vulnerable library: /node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - webpack-dev-server-2.11.5.tgz - sockjs-client-1.1.5.tgz - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6. <p>Publish Date: 2022-02-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p> <p>Release Date: 2022-02-14</p> <p>Fix Resolution: url-parse - 1.5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0512 (High) detected in url-parse-1.4.7.tgz - ## CVE-2022-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p>Path to dependency file: /conFusionAng/package.json</p> <p>Path to vulnerable library: /node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - webpack-dev-server-2.11.5.tgz - sockjs-client-1.1.5.tgz - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6. <p>Publish Date: 2022-02-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p> <p>Release Date: 2022-02-14</p> <p>Fix Resolution: url-parse - 1.5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file confusionang package json path to vulnerable library node modules url parse package json dependency hierarchy cli tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource
0
294,722
9,039,902,515
IssuesEvent
2019-02-10 11:42:08
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
drive.google.com - see bug description
browser-firefox-mobile priority-critical
<!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://drive.google.com/drive/mobile/folders/1-0Y9QnZuXMSHPUB-uw_Ed2sgtQ32KOS9?sort=13&direction=a **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: this girl has control of apps, phone carrier my whole phone **Steps to Reproduce**: Settings would change by itself everyday. Noticed files n especially assking permission to share my files. Has a camera watching me n audio plus taking my email n name getting Google Pay, Amazon accounts n my phine carrier [![Screenshot Description](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb-thumb.jpg)](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190207161357</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Content Security Policy: Ignoring \'report-sample\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[console.timeStamp(CSI/tbsd_) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_tbnd) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ipls) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/jl) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ai) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feis) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feie) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upl_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/as) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/asc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_2) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upr) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivdp) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_1) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/iple) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_files_api_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_partial) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivlc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/pst) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/tbsd_) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/_tbnd) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/first_files_api_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/fll) https://drive.google.com/drive/mobile/:1:147]'] </pre> </details> Reported by @kim6571 _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
drive.google.com - see bug description - <!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://drive.google.com/drive/mobile/folders/1-0Y9QnZuXMSHPUB-uw_Ed2sgtQ32KOS9?sort=13&direction=a **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: this girl has control of apps, phone carrier my whole phone **Steps to Reproduce**: Settings would change by itself everyday. Noticed files n especially assking permission to share my files. Has a camera watching me n audio plus taking my email n name getting Google Pay, Amazon accounts n my phine carrier [![Screenshot Description](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb-thumb.jpg)](https://webcompat.com/uploads/2019/2/59bf998e-470f-4530-ae4f-8792bf8296cb.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190207161357</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Content Security Policy: Ignoring \'report-sample\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[console.timeStamp(CSI/tbsd_) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_tbnd) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ipls) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/jl) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ai) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feis) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/feie) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upl_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/as) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/start) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/asc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_2) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/upr) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivdp) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/itemservice_initialize_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/drive_prefetched_1) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/frd_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/iple) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_files_api_request_s) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_partial) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/_first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/ivlc) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/pst) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_account_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/time_to_first_apps_query_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/tbsd_) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/_tbnd) https://apis.google.com/_/scs/apps-static/_/js/k=oz.gapi.en.2EGcoRPZQzg.O/m=client/rt=j/sv=1/d=1/ed=1/am=wQ/rs=AGLTcCNOnpGZaRrnqOmfGcy6pxYZI3ZWCQ/cb=gapi.loaded_0:615:127]', u'[console.timeStamp(CSI/first_files_api_request_e) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/first_item_live_list_success) https://drive.google.com/drive/mobile/:1:147]', u'[console.timeStamp(CSI/fll) https://drive.google.com/drive/mobile/:1:147]'] </pre> </details> Reported by @kim6571 _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
drive google com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description this girl has control of apps phone carrier my whole phone steps to reproduce settings would change by itself everyday noticed files n especially assking permission to share my files has a camera watching me n audio plus taking my email n name getting google pay amazon accounts n my phine carrier browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u reported by from with ❤️
1