Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
17,245
2,986,622,894
IssuesEvent
2015-07-20 05:40:45
jayway/awaitility
https://api.github.com/repos/jayway/awaitility
opened
AssertionError from inside of supplier
auto-migrated Priority-Medium Type-Defect
``` We are using Awaitility inside our tests and some test are throwing AssertionError from inside of Callable supplier. As the result we got ConditionTimeoutException without any cause and it looked like Awaitility just stopped calling our supplier for no reason. I think the problem is that AssertionError is not an Exception, but Awaitility catches only Exceptions in ConditionAwaiter line 57 What is the expected output? What do you see instead? We expect to see stack trace of the AssertionError or it being re-thrown. Instead we see ConditionTimeoutException which doesn't show the real cause What version of the product are you using? On what operating system? 1.6.3 on any operating system ``` Original issue reported on code.google.com by `sho...@gmail.com` on 16 Jun 2015 at 12:32
1.0
AssertionError from inside of supplier - ``` We are using Awaitility inside our tests and some test are throwing AssertionError from inside of Callable supplier. As the result we got ConditionTimeoutException without any cause and it looked like Awaitility just stopped calling our supplier for no reason. I think the problem is that AssertionError is not an Exception, but Awaitility catches only Exceptions in ConditionAwaiter line 57 What is the expected output? What do you see instead? We expect to see stack trace of the AssertionError or it being re-thrown. Instead we see ConditionTimeoutException which doesn't show the real cause What version of the product are you using? On what operating system? 1.6.3 on any operating system ``` Original issue reported on code.google.com by `sho...@gmail.com` on 16 Jun 2015 at 12:32
defect
assertionerror from inside of supplier we are using awaitility inside our tests and some test are throwing assertionerror from inside of callable supplier as the result we got conditiontimeoutexception without any cause and it looked like awaitility just stopped calling our supplier for no reason i think the problem is that assertionerror is not an exception but awaitility catches only exceptions in conditionawaiter line what is the expected output what do you see instead we expect to see stack trace of the assertionerror or it being re thrown instead we see conditiontimeoutexception which doesn t show the real cause what version of the product are you using on what operating system on any operating system original issue reported on code google com by sho gmail com on jun at
1
1,967
2,603,974,195
IssuesEvent
2015-02-24 19:01:05
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳有治疗疱疹的好办法吗
auto-migrated Priority-Medium Type-Defect
``` 沈阳有治疗疱疹的好办法吗〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:10
1.0
沈阳有治疗疱疹的好办法吗 - ``` 沈阳有治疗疱疹的好办法吗〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:10
defect
沈阳有治疗疱疹的好办法吗 沈阳有治疗疱疹的好办法吗〓沈陽軍區政治部醫院性病〓tel�� � 〓 , � ��。 。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 original issue reported on code google com by gmail com on jun at
1
338,538
30,304,245,507
IssuesEvent
2023-07-10 08:24:59
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix layers.test_depthwise_conv2d
Sub Task Ivy Functional API Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5469924771/jobs/9959404555"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix layers.test_depthwise_conv2d - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5469924771/jobs/9959404555"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5505518661"><img src=https://img.shields.io/badge/-success-success></a>
non_defect
fix layers test depthwise jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
0
175,103
27,792,103,333
IssuesEvent
2023-03-17 09:42:47
GoogleChrome/web.dev
https://api.github.com/repos/GoogleChrome/web.dev
closed
Design a page for CSS triggers
P2 design
The team used to maintain csstriggers.com (the site which is currently live there has nothing to do with the old project), which listed all CSS properties and their impact on the browser rendering pipeline. The page looked like this: https://www.lmame-geek.com/css-triggers/ While we don't maintain the page anymore the team is currently looking into recollecting the data and making it available for web.dev to display. Therefore we need a design how this would look on web.dev.
1.0
Design a page for CSS triggers - The team used to maintain csstriggers.com (the site which is currently live there has nothing to do with the old project), which listed all CSS properties and their impact on the browser rendering pipeline. The page looked like this: https://www.lmame-geek.com/css-triggers/ While we don't maintain the page anymore the team is currently looking into recollecting the data and making it available for web.dev to display. Therefore we need a design how this would look on web.dev.
non_defect
design a page for css triggers the team used to maintain csstriggers com the site which is currently live there has nothing to do with the old project which listed all css properties and their impact on the browser rendering pipeline the page looked like this while we don t maintain the page anymore the team is currently looking into recollecting the data and making it available for web dev to display therefore we need a design how this would look on web dev
0
39,646
5,113,712,988
IssuesEvent
2017-01-06 16:10:16
harksys/HawkEye
https://api.github.com/repos/harksys/HawkEye
closed
Enlarge PR/Issue Icons
Design Good for beginners
Enlarging the PR and Issue icons in feed for a bit of a better visual cue on what type of notification you're looking at in the feed would be awesome. The icons that I'm specifically referring to: ![image](https://cloud.githubusercontent.com/assets/502396/21707419/02298a14-d39d-11e6-9a76-1e3d480ff94e.png)
1.0
Enlarge PR/Issue Icons - Enlarging the PR and Issue icons in feed for a bit of a better visual cue on what type of notification you're looking at in the feed would be awesome. The icons that I'm specifically referring to: ![image](https://cloud.githubusercontent.com/assets/502396/21707419/02298a14-d39d-11e6-9a76-1e3d480ff94e.png)
non_defect
enlarge pr issue icons enlarging the pr and issue icons in feed for a bit of a better visual cue on what type of notification you re looking at in the feed would be awesome the icons that i m specifically referring to
0
45,872
11,744,457,031
IssuesEvent
2020-03-12 07:46:10
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
opened
Audit log : Candidate : In the Audit log 'Reject offer letter' should be updated with 'Decline Offer Letter'
Audit Log Defect P2 Release#7 Build#3
When the candidate 'Decline' the offer letter the Audit log is displayed as "Candidate accessed Reject offer letter screen" Similarly also fix this issue when the candidate close the Decline pop or click on X symbol to close the window the audit log should be updated as " Candidate Cancelled Decline offer letter Screen'
1.0
Audit log : Candidate : In the Audit log 'Reject offer letter' should be updated with 'Decline Offer Letter' - When the candidate 'Decline' the offer letter the Audit log is displayed as "Candidate accessed Reject offer letter screen" Similarly also fix this issue when the candidate close the Decline pop or click on X symbol to close the window the audit log should be updated as " Candidate Cancelled Decline offer letter Screen'
non_defect
audit log candidate in the audit log reject offer letter should be updated with decline offer letter when the candidate decline the offer letter the audit log is displayed as candidate accessed reject offer letter screen similarly also fix this issue when the candidate close the decline pop or click on x symbol to close the window the audit log should be updated as candidate cancelled decline offer letter screen
0
68,335
21,639,364,935
IssuesEvent
2022-05-05 17:06:35
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Investigate accessibility tests time out in tugboat
Defect DevOps Platform CMS Team
## Description There are accessibility timeout errors that frequently occur which require manual intervention from the engineers. ### Acceptance Criteria - [ ] Investigate the timeout errors for accessibility tests in PR environments - [ ] Provide a solution to get tests running in a more stable fashion - [ ] Create followup ticket for resolution ![image.png](https://images.zenhubusercontent.com/61671a60c72e6d0337fd8bcd/dd781cbd-3977-499b-9114-e2449bac0fcf) ### CMS Team Please check the team(s) that will do this work. - [ ] `CMS Program` - [x] `Platform CMS Team` - [ ] `Sitewide CMS Team ` (leave Sitewide unchecked and check the specific team instead) - [ ] `⭐️ Content ops` - [ ] `⭐️ CMS experience` - [ ] `⭐️ Offices` - [ ] `⭐️ Product support` - [ ] `⭐️ User support`
1.0
Investigate accessibility tests time out in tugboat - ## Description There are accessibility timeout errors that frequently occur which require manual intervention from the engineers. ### Acceptance Criteria - [ ] Investigate the timeout errors for accessibility tests in PR environments - [ ] Provide a solution to get tests running in a more stable fashion - [ ] Create followup ticket for resolution ![image.png](https://images.zenhubusercontent.com/61671a60c72e6d0337fd8bcd/dd781cbd-3977-499b-9114-e2449bac0fcf) ### CMS Team Please check the team(s) that will do this work. - [ ] `CMS Program` - [x] `Platform CMS Team` - [ ] `Sitewide CMS Team ` (leave Sitewide unchecked and check the specific team instead) - [ ] `⭐️ Content ops` - [ ] `⭐️ CMS experience` - [ ] `⭐️ Offices` - [ ] `⭐️ Product support` - [ ] `⭐️ User support`
defect
investigate accessibility tests time out in tugboat description there are accessibility timeout errors that frequently occur which require manual intervention from the engineers acceptance criteria investigate the timeout errors for accessibility tests in pr environments provide a solution to get tests running in a more stable fashion create followup ticket for resolution cms team please check the team s that will do this work cms program platform cms team sitewide cms team leave sitewide unchecked and check the specific team instead ⭐️ content ops ⭐️ cms experience ⭐️ offices ⭐️ product support ⭐️ user support
1
71,208
23,490,794,347
IssuesEvent
2022-08-17 18:31:05
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Terminal: AutoComplete not working on MyFaces
:lady_beetle: defect
### Describe the bug Originally Reported here: https://github.com/melloware/quarkus-faces/issues/165 **Mojarra:** ![image](https://user-images.githubusercontent.com/4399574/185206133-0dc755a0-fb48-48c4-b3c1-36f315bdfe6f.png) **MyFaces:** ![image](https://user-images.githubusercontent.com/4399574/185206255-3d5311da-a593-480b-b580-47546766d689.png) Looks like the response JSON is being encoded. ### Reproducer Run the showcase locally with ` mvn clean jetty:run -Pnon-ee,myfaces-next-2.3` and navigate to https://www.primefaces.org/showcase/ui/misc/terminal/autocomplete.xhtml and type `git + TAB` ### Expected behavior Terminal Autocomplete works ### PrimeFaces edition Community ### PrimeFaces version 12.0.0-RC2 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.3-next ### Browser(s) _No response_
1.0
Terminal: AutoComplete not working on MyFaces - ### Describe the bug Originally Reported here: https://github.com/melloware/quarkus-faces/issues/165 **Mojarra:** ![image](https://user-images.githubusercontent.com/4399574/185206133-0dc755a0-fb48-48c4-b3c1-36f315bdfe6f.png) **MyFaces:** ![image](https://user-images.githubusercontent.com/4399574/185206255-3d5311da-a593-480b-b580-47546766d689.png) Looks like the response JSON is being encoded. ### Reproducer Run the showcase locally with ` mvn clean jetty:run -Pnon-ee,myfaces-next-2.3` and navigate to https://www.primefaces.org/showcase/ui/misc/terminal/autocomplete.xhtml and type `git + TAB` ### Expected behavior Terminal Autocomplete works ### PrimeFaces edition Community ### PrimeFaces version 12.0.0-RC2 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.3-next ### Browser(s) _No response_
defect
terminal autocomplete not working on myfaces describe the bug originally reported here mojarra myfaces looks like the response json is being encoded reproducer run the showcase locally with mvn clean jetty run pnon ee myfaces next and navigate to and type git tab expected behavior terminal autocomplete works primefaces edition community primefaces version theme no response jsf implementation myfaces jsf version next browser s no response
1
134,886
30,204,940,028
IssuesEvent
2023-07-05 08:48:14
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
[BUG] Analyzers and code fixes not working in C# Dev Kit?
Bug Area-Analyzers VSCode
From vscode-dotnettools created by [DaRosenberg](https://github.com/DaRosenberg): microsoft/vscode-dotnettools#61 ### Describe the Issue Not sure if this is exposing a bug or just a missing feature for now. Can you clarify whether analyzers and code fixes should be working yet or not? ### Steps To Reproduce Works: - Analyzers are run during build, and issues surfaced as problems (errors and warnings) Does NOT work: - Analyzers do not run automatically (neither for solution nor for open documents) - No command to run analyzers imperatively either - Detected issues are highlighted in source (squigglies) but no fixes appear - No settings to enable analyzers or control analysis behavior ### Expected Behavior I expect analyzers functionality to be on par with Visual Studio, or at least on par with the previous OmniSharp-powered C# extension. ### Environment Information - OS: macOS Ventura - VS Code: 1.78.2 - Extension: C# Dev Kit v0.1.8
1.0
[BUG] Analyzers and code fixes not working in C# Dev Kit? - From vscode-dotnettools created by [DaRosenberg](https://github.com/DaRosenberg): microsoft/vscode-dotnettools#61 ### Describe the Issue Not sure if this is exposing a bug or just a missing feature for now. Can you clarify whether analyzers and code fixes should be working yet or not? ### Steps To Reproduce Works: - Analyzers are run during build, and issues surfaced as problems (errors and warnings) Does NOT work: - Analyzers do not run automatically (neither for solution nor for open documents) - No command to run analyzers imperatively either - Detected issues are highlighted in source (squigglies) but no fixes appear - No settings to enable analyzers or control analysis behavior ### Expected Behavior I expect analyzers functionality to be on par with Visual Studio, or at least on par with the previous OmniSharp-powered C# extension. ### Environment Information - OS: macOS Ventura - VS Code: 1.78.2 - Extension: C# Dev Kit v0.1.8
non_defect
analyzers and code fixes not working in c dev kit from vscode dotnettools created by microsoft vscode dotnettools describe the issue not sure if this is exposing a bug or just a missing feature for now can you clarify whether analyzers and code fixes should be working yet or not steps to reproduce works analyzers are run during build and issues surfaced as problems errors and warnings does not work analyzers do not run automatically neither for solution nor for open documents no command to run analyzers imperatively either detected issues are highlighted in source squigglies but no fixes appear no settings to enable analyzers or control analysis behavior expected behavior i expect analyzers functionality to be on par with visual studio or at least on par with the previous omnisharp powered c extension environment information os macos ventura vs code extension c dev kit
0
51,288
13,207,409,152
IssuesEvent
2020-08-14 22:59:42
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
GLshovel File-> Open I3 file on Mac OS 10.5 hangs (Trac #146)
IceTray Incomplete Migration Migrated from Trac defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/146">https://code.icecube.wisc.edu/projects/icecube/ticket/146</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-03-06T01:42:38", "_ts": "1236303758000000", "description": "", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-10-22T16:12:45", "component": "IceTray", "summary": "GLshovel File-> Open I3 file on Mac OS 10.5 hangs", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
GLshovel File-> Open I3 file on Mac OS 10.5 hangs (Trac #146) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/146">https://code.icecube.wisc.edu/projects/icecube/ticket/146</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-03-06T01:42:38", "_ts": "1236303758000000", "description": "", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-10-22T16:12:45", "component": "IceTray", "summary": "GLshovel File-> Open I3 file on Mac OS 10.5 hangs", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
glshovel file open file on mac os hangs trac migrated from json status closed changetime ts description reporter blaufuss cc resolution fixed time component icetray summary glshovel file open file on mac os hangs priority normal keywords milestone owner troy type defect
1
682,094
23,332,271,651
IssuesEvent
2022-08-09 06:49:21
ooni/probe
https://api.github.com/repos/ooni/probe
opened
OONI Probe Mobile includes RiseupVPN option in Settings
bug ooni/probe-mobile priority/medium
**Describe the bug** OONI Probe Mobile includes RiseupVPN in 'Test options' for 'Circumvention' tests while the RiseupVPN test is temporarily disabled. **To Reproduce** 'Settings' > 'Test option' > 'Circumvention' > 'Test RiseupVPN' **Expected behavior** There should be left only two options in the menu ('Test Tor', 'Test Psiphon') **Screenshots** ![IMAGE 2022-08-09 10:47:17](https://user-images.githubusercontent.com/16296135/183582752-9d1e9ac0-c7e8-4c7a-ba54-e19ee8363df5.jpg) **System information (please complete the following information):** - OONI Probe Mobile version: 3.7.0 **Additional context** There is a chance that we might need to return this option to the menu.
1.0
OONI Probe Mobile includes RiseupVPN option in Settings - **Describe the bug** OONI Probe Mobile includes RiseupVPN in 'Test options' for 'Circumvention' tests while the RiseupVPN test is temporarily disabled. **To Reproduce** 'Settings' > 'Test option' > 'Circumvention' > 'Test RiseupVPN' **Expected behavior** There should be left only two options in the menu ('Test Tor', 'Test Psiphon') **Screenshots** ![IMAGE 2022-08-09 10:47:17](https://user-images.githubusercontent.com/16296135/183582752-9d1e9ac0-c7e8-4c7a-ba54-e19ee8363df5.jpg) **System information (please complete the following information):** - OONI Probe Mobile version: 3.7.0 **Additional context** There is a chance that we might need to return this option to the menu.
non_defect
ooni probe mobile includes riseupvpn option in settings describe the bug ooni probe mobile includes riseupvpn in test options for circumvention tests while the riseupvpn test is temporarily disabled to reproduce settings test option circumvention test riseupvpn expected behavior there should be left only two options in the menu test tor test psiphon screenshots system information please complete the following information ooni probe mobile version additional context there is a chance that we might need to return this option to the menu
0
76,594
7,541,080,431
IssuesEvent
2018-04-17 08:44:00
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Update vSphere node template with API fails: "a Config field must be set"
area/host status/resolved status/to-test version/2.0
Trying to update my vSphere node template via API, but results in HTTP response status code 422 `MissingRequired: a Config field must be set` According to: https://github.com/rancher/rancher/blob/master/pkg/api/customization/nodetemplate/validation.go#L12 Seems like a bug, because according to the HTTP payload, the following JSON property is in the JSON: `"driver":"vmwarevsphere"` **Rancher versions:** rancher/server: 2.0.0-beta3 **Docker version: (`docker version`,`docker info` preferred)** ``` root@platania:~# docker version Client: Version: 18.04.0-ce API version: 1.37 Go version: go1.9.4 Git commit: 3d479c0 Built: Tue Apr 10 18:20:32 2018 OS/Arch: linux/amd64 Experimental: false Orchestrator: swarm Server: Engine: Version: 18.04.0-ce API version: 1.37 (minimum version 1.12) Go version: go1.9.4 Git commit: 3d479c0 Built: Tue Apr 10 18:18:40 2018 OS/Arch: linux/amd64 Experimental: false ``` **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** ``` root@platania:~# cat /etc/os-release NAME="Ubuntu" VERSION="16.04.3 LTS (Xenial Xerus)" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 16.04.3 LTS" VERSION_ID="16.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" VERSION_CODENAME=xenial UBUNTU_CODENAME=xenial ``` **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** vSphere (ESXi 6.0) **Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)** **Environment Template: (Cattle/Kubernetes/Swarm/Mesos)** **Steps to Reproduce:** ``` curl --insecure -u "token-ncv2z:jkbwgj78jrnkl87sg75vjbmd9kz7t947xg2dwtjgvbzxv5qzkxqmvd" \ -X PUT \ -H 'Accept: application/json' \ -H 'Content-Type: application/json' \ -d '{"annotations":{}, "created":"2018-04-14T15:27:25Z", "creatorId":"user-hfcpg", "driver":"vmwarevsphere", "engineEnv":{}, "engineInsecureRegistry":[], "engineLabel":{}, "engineOpt":{}, "engineRegistryMirror":[], "id":"user-hfcpg:nodetemplate-bc57f", "labels":{}, "name":"vsphere", "ownerReferences":[], "removed":null, "state":"active", "transitioning":"no", "transitioningMessage":"", "uuid":"55be72ba-3ff8-11e8-a679-0242ac110002", "vmwarevsphereConfig":{"boot2dockerUrl":"https://github.com/boot2docker/boot2docker/releases/download/v17.03.2-ce/boot2docker.iso", "cloudinit":"", "cpuCount":"2", "datacenter":"dc", "datastore":"L09", "diskSize":"10000", "folder":"", "hostsystem":"host", "memorySize":"2048", "network":["dev"], "pool":"", "username":"***", "vcenter":"vcsrv", "vcenterPort":"443"}}' \ 'https://platania/v3/nodeTemplates/user-hfcpg:nodetemplate-bc57f' ``` **Results:** ``` HTTP/1.1 422 Unprocessable Entity Content-Type: application/json Expires: Wed 24 Feb 1982 18:42:00 GMT Set-Cookie: CSRF=51AADB9F99; Path=/ X-Api-Schemas: https://platania/v3/schemas Date: Sat, 14 Apr 2018 15:38:19 GMT Content-Length: 137 {"actions":{},"baseType":"error","code":"MissingRequired","links":{},"message":"a Config field must be set","status":422,"type":"error"} ```
1.0
Update vSphere node template with API fails: "a Config field must be set" - Trying to update my vSphere node template via API, but results in HTTP response status code 422 `MissingRequired: a Config field must be set` According to: https://github.com/rancher/rancher/blob/master/pkg/api/customization/nodetemplate/validation.go#L12 Seems like a bug, because according to the HTTP payload, the following JSON property is in the JSON: `"driver":"vmwarevsphere"` **Rancher versions:** rancher/server: 2.0.0-beta3 **Docker version: (`docker version`,`docker info` preferred)** ``` root@platania:~# docker version Client: Version: 18.04.0-ce API version: 1.37 Go version: go1.9.4 Git commit: 3d479c0 Built: Tue Apr 10 18:20:32 2018 OS/Arch: linux/amd64 Experimental: false Orchestrator: swarm Server: Engine: Version: 18.04.0-ce API version: 1.37 (minimum version 1.12) Go version: go1.9.4 Git commit: 3d479c0 Built: Tue Apr 10 18:18:40 2018 OS/Arch: linux/amd64 Experimental: false ``` **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** ``` root@platania:~# cat /etc/os-release NAME="Ubuntu" VERSION="16.04.3 LTS (Xenial Xerus)" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 16.04.3 LTS" VERSION_ID="16.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" VERSION_CODENAME=xenial UBUNTU_CODENAME=xenial ``` **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** vSphere (ESXi 6.0) **Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)** **Environment Template: (Cattle/Kubernetes/Swarm/Mesos)** **Steps to Reproduce:** ``` curl --insecure -u "token-ncv2z:jkbwgj78jrnkl87sg75vjbmd9kz7t947xg2dwtjgvbzxv5qzkxqmvd" \ -X PUT \ -H 'Accept: application/json' \ -H 'Content-Type: application/json' \ -d '{"annotations":{}, "created":"2018-04-14T15:27:25Z", "creatorId":"user-hfcpg", "driver":"vmwarevsphere", "engineEnv":{}, "engineInsecureRegistry":[], "engineLabel":{}, "engineOpt":{}, "engineRegistryMirror":[], "id":"user-hfcpg:nodetemplate-bc57f", "labels":{}, "name":"vsphere", "ownerReferences":[], "removed":null, "state":"active", "transitioning":"no", "transitioningMessage":"", "uuid":"55be72ba-3ff8-11e8-a679-0242ac110002", "vmwarevsphereConfig":{"boot2dockerUrl":"https://github.com/boot2docker/boot2docker/releases/download/v17.03.2-ce/boot2docker.iso", "cloudinit":"", "cpuCount":"2", "datacenter":"dc", "datastore":"L09", "diskSize":"10000", "folder":"", "hostsystem":"host", "memorySize":"2048", "network":["dev"], "pool":"", "username":"***", "vcenter":"vcsrv", "vcenterPort":"443"}}' \ 'https://platania/v3/nodeTemplates/user-hfcpg:nodetemplate-bc57f' ``` **Results:** ``` HTTP/1.1 422 Unprocessable Entity Content-Type: application/json Expires: Wed 24 Feb 1982 18:42:00 GMT Set-Cookie: CSRF=51AADB9F99; Path=/ X-Api-Schemas: https://platania/v3/schemas Date: Sat, 14 Apr 2018 15:38:19 GMT Content-Length: 137 {"actions":{},"baseType":"error","code":"MissingRequired","links":{},"message":"a Config field must be set","status":422,"type":"error"} ```
non_defect
update vsphere node template with api fails a config field must be set trying to update my vsphere node template via api but results in http response status code missingrequired a config field must be set according to seems like a bug because according to the http payload the following json property is in the json driver vmwarevsphere rancher versions rancher server docker version docker version docker info preferred root platania docker version client version ce api version go version git commit built tue apr os arch linux experimental false orchestrator swarm server engine version ce api version minimum version go version git commit built tue apr os arch linux experimental false operating system and kernel cat etc os release uname r preferred root platania cat etc os release name ubuntu version lts xenial xerus id ubuntu id like debian pretty name ubuntu lts version id home url support url bug report url version codename xenial ubuntu codename xenial type provider of hosts virtualbox bare metal aws gce do vsphere esxi setup details single node rancher vs ha rancher internal db vs external db environment template cattle kubernetes swarm mesos steps to reproduce curl insecure u token x put h accept application json h content type application json d annotations created creatorid user hfcpg driver vmwarevsphere engineenv engineinsecureregistry enginelabel engineopt engineregistrymirror id user hfcpg nodetemplate labels name vsphere ownerreferences removed null state active transitioning no transitioningmessage uuid vmwarevsphereconfig cloudinit cpucount datacenter dc datastore disksize folder hostsystem host memorysize network pool username vcenter vcsrv vcenterport results http unprocessable entity content type application json expires wed feb gmt set cookie csrf path x api schemas date sat apr gmt content length actions basetype error code missingrequired links message a config field must be set status type error
0
65,913
19,786,249,004
IssuesEvent
2022-01-18 07:12:10
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: element.click() does not work when scaling is used
I-defect needs-triaging
### What happened? I use docker-selenium-novnc as a web browser for my phone, so that I can access some website like a computer. Everything works fine when using when the scaling is set to 100%. When I read comics, some picture are slightly too big, so most of the time I set scaling at 90%. ### How can we reproduce the issue? ```shell set scale to 90% then click. ``` ### Relevant log output ```shell the ".click()" waits forever until I scale chrome back to default. ``` ### Operating System docker for win ### Selenium version selenium/hub:4.1.0-20211123 ### What are the browser(s) and version(s) where you see this issue? selenium/node-chrome:4.1.0-20211123 ### What are the browser driver(s) and version(s) where you see this issue? selenium/node-chrome:4.1.0-20211123 ### Are you using Selenium Grid? selenium/hub:4.1.0-20211123
1.0
[🐛 Bug]: element.click() does not work when scaling is used - ### What happened? I use docker-selenium-novnc as a web browser for my phone, so that I can access some website like a computer. Everything works fine when using when the scaling is set to 100%. When I read comics, some picture are slightly too big, so most of the time I set scaling at 90%. ### How can we reproduce the issue? ```shell set scale to 90% then click. ``` ### Relevant log output ```shell the ".click()" waits forever until I scale chrome back to default. ``` ### Operating System docker for win ### Selenium version selenium/hub:4.1.0-20211123 ### What are the browser(s) and version(s) where you see this issue? selenium/node-chrome:4.1.0-20211123 ### What are the browser driver(s) and version(s) where you see this issue? selenium/node-chrome:4.1.0-20211123 ### Are you using Selenium Grid? selenium/hub:4.1.0-20211123
defect
element click does not work when scaling is used what happened i use docker selenium novnc as a web browser for my phone so that i can access some website like a computer everything works fine when using when the scaling is set to when i read comics some picture are slightly too big so most of the time i set scaling at how can we reproduce the issue shell set scale to then click relevant log output shell the click waits forever until i scale chrome back to default operating system docker for win selenium version selenium hub what are the browser s and version s where you see this issue selenium node chrome what are the browser driver s and version s where you see this issue selenium node chrome are you using selenium grid selenium hub
1
257,098
27,561,769,135
IssuesEvent
2023-03-07 22:45:12
samqws-marketing/box_box-ui-elements
https://api.github.com/repos/samqws-marketing/box_box-ui-elements
closed
CVE-2022-24772 (High) detected in node-forge-0.9.0.tgz - autoclosed
security vulnerability
## CVE-2022-24772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.10.1.tgz (Root Library) - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24772>CVE-2022-24772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution (node-forge): 1.3.0</p> <p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2022-24772 (High) detected in node-forge-0.9.0.tgz - autoclosed - ## CVE-2022-24772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.10.1.tgz (Root Library) - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24772>CVE-2022-24772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution (node-forge): 1.3.0</p> <p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_defect
cve high detected in node forge tgz autoclosed cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in head commit a href found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not check for tailing garbage bytes after decoding a digestinfo asn structure this can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution webpack dev server check this box to open an automated fix pr
0
69,906
22,745,457,406
IssuesEvent
2022-07-07 08:47:02
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Crash while cross-signing with my other Element Web device
T-Defect
### Steps to reproduce 1. Initiate a verification request from Element Web on my computer to Element Android on my phone 2. Use the QR Code scanning method to verify the device 3. After scanning the QR code, a spinning process indicator is shown on the phone 4. The app crashes 5. Element Web shows the green checkmark shield ### Outcome #### What did you expect? Element Android to not crash, and for verification to succeed. #### What happened instead? The app crashed. After reopening Element Android, in Security & Privacy it says: Cross-Signing is enabled Keys are trusted. Private keys are not known I'm not sure whether the last line is a cause for concern. ### Your phone model Google Pixel 4a 5G ### Operating system version Android 12 ### Application version and app store Element version 1.4.25 [40104252] (G-b9267), Matrix SDK Version 1.4.25 (1f34d368), olm version 3.2.12 ### Homeserver amorgan.xyz ### Will you send logs? Yes ### Are you willing to provide a PR? No
1.0
Crash while cross-signing with my other Element Web device - ### Steps to reproduce 1. Initiate a verification request from Element Web on my computer to Element Android on my phone 2. Use the QR Code scanning method to verify the device 3. After scanning the QR code, a spinning process indicator is shown on the phone 4. The app crashes 5. Element Web shows the green checkmark shield ### Outcome #### What did you expect? Element Android to not crash, and for verification to succeed. #### What happened instead? The app crashed. After reopening Element Android, in Security & Privacy it says: Cross-Signing is enabled Keys are trusted. Private keys are not known I'm not sure whether the last line is a cause for concern. ### Your phone model Google Pixel 4a 5G ### Operating system version Android 12 ### Application version and app store Element version 1.4.25 [40104252] (G-b9267), Matrix SDK Version 1.4.25 (1f34d368), olm version 3.2.12 ### Homeserver amorgan.xyz ### Will you send logs? Yes ### Are you willing to provide a PR? No
defect
crash while cross signing with my other element web device steps to reproduce initiate a verification request from element web on my computer to element android on my phone use the qr code scanning method to verify the device after scanning the qr code a spinning process indicator is shown on the phone the app crashes element web shows the green checkmark shield outcome what did you expect element android to not crash and for verification to succeed what happened instead the app crashed after reopening element android in security privacy it says cross signing is enabled keys are trusted private keys are not known i m not sure whether the last line is a cause for concern your phone model google pixel operating system version android application version and app store element version g matrix sdk version olm version homeserver amorgan xyz will you send logs yes are you willing to provide a pr no
1
6,610
2,590,202,378
IssuesEvent
2015-02-18 17:28:55
OpenSprites/OpenSprites
https://api.github.com/repos/OpenSprites/OpenSprites
opened
Please - connect to IRC everytime you want to do something!
high priority ongoing
Hello! Everytime we are making something and we are not updating our files actively, someone can do an commit and it mixes things up and its hard to revert everything back. We need to get everyone to join the IRC chat @The-QullzToxic created a while ago. Its really important and we can keep track on whats everyone doing. Even though theres nobody online, stay there! I'm usually online 12 hours a day on the IRC and i would love for someone else to talk there :sweat_smile: https://webchat.freenode.net/ Username: Your GitHub name and connect to #OpenSprites
1.0
Please - connect to IRC everytime you want to do something! - Hello! Everytime we are making something and we are not updating our files actively, someone can do an commit and it mixes things up and its hard to revert everything back. We need to get everyone to join the IRC chat @The-QullzToxic created a while ago. Its really important and we can keep track on whats everyone doing. Even though theres nobody online, stay there! I'm usually online 12 hours a day on the IRC and i would love for someone else to talk there :sweat_smile: https://webchat.freenode.net/ Username: Your GitHub name and connect to #OpenSprites
non_defect
please connect to irc everytime you want to do something hello everytime we are making something and we are not updating our files actively someone can do an commit and it mixes things up and its hard to revert everything back we need to get everyone to join the irc chat the qullztoxic created a while ago its really important and we can keep track on whats everyone doing even though theres nobody online stay there i m usually online hours a day on the irc and i would love for someone else to talk there sweat smile username your github name and connect to opensprites
0
27,607
4,059,765,996
IssuesEvent
2016-05-25 10:45:51
FraserCoyle/sermon-manager-issues
https://api.github.com/repos/FraserCoyle/sermon-manager-issues
opened
[UI] Menu Dropdowns
bug design
The dark highlight area behind 'settings' needs padding left and right. 15px Dropdown box is sitting onto of the menu and has shadow. It needs aligned. ![screen shot 2016-05-25 at 10 18 34](https://cloud.githubusercontent.com/assets/19221166/15537425/3a2177aa-226e-11e6-81df-fd3d65aaeb01.jpg)
1.0
[UI] Menu Dropdowns - The dark highlight area behind 'settings' needs padding left and right. 15px Dropdown box is sitting onto of the menu and has shadow. It needs aligned. ![screen shot 2016-05-25 at 10 18 34](https://cloud.githubusercontent.com/assets/19221166/15537425/3a2177aa-226e-11e6-81df-fd3d65aaeb01.jpg)
non_defect
menu dropdowns the dark highlight area behind settings needs padding left and right dropdown box is sitting onto of the menu and has shadow it needs aligned
0
66,888
8,973,706,759
IssuesEvent
2019-01-29 21:47:25
NervanaSystems/coach
https://api.github.com/repos/NervanaSystems/coach
closed
create a contributor guide
documentation priority/p2
Add expectations and flows for developers and reviewers. Name and place in the repo according to: https://help.github.com/articles/setting-guidelines-for-repository-contributors/
1.0
create a contributor guide - Add expectations and flows for developers and reviewers. Name and place in the repo according to: https://help.github.com/articles/setting-guidelines-for-repository-contributors/
non_defect
create a contributor guide add expectations and flows for developers and reviewers name and place in the repo according to
0
46,047
13,055,844,689
IssuesEvent
2020-07-30 02:54:23
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
'cannot import module' troubleshooting (Trac #506)
Incomplete Migration Migrated from Trac defect documentation
Migrated from https://code.icecube.wisc.edu/ticket/506 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "19:59 < dipo> hi, have a quick question on pybindings, anybody here who can help?\n20:02 <@straszhm> hi\n20:02 <@straszhm> the jebclasses question?\n20:02 < dipo> yes\n20:03 <@straszhm> do you have a file $I3_BUILD/lib/icecube/jebclasses.so?\n20:03 < dipo> yes\n20:04 <@straszhm> hm\n20:05 <@straszhm> no other error messages from your python script?\n20:05 < dipo> no, I'm actually doing things interactively from the prompt\n20:06 <@straszhm> if you just put into a script\n20:06 <@straszhm> 'from icecube import icetray, dataclasses, jebclasses'\n20:06 <@straszhm> (only that)\n20:06 <@straszhm> does it fail?\n20:09 < dipo> sorry it took a while, trying it out, yes, it gave the same error:\n20:10 < dipo> ImportError: cannot import name jebclasses\n20:10 <@straszhm> ok. python has a flag '-v'\n20:10 <@straszhm> that shows verbose information about how it finds imporrts\n20:11 <@straszhm> maybe this will show something. \n20:11 <@straszhm> does it get icetray and dataclasses from the right directory?\n20:11 <@straszhm> (so, run python -v testscript.py)\n20:12 < dipo> yes\n20:12 <@straszhm> anything interesting where the jebclasses search starts?\n20:12 < dipo> import icecube.icetray # dynamically loaded from \n /home/ofadiran/std-processing/build/lib/icecube/icetray.so\n20:12 <@straszhm> right, that looks good\n20:12 < dipo> for icetray, but for jebclasses:\n20:12 < dipo> just the same short error msg that I pasted b4\n20:13 <@straszhm> what is the output of 'file $I3_BUILD/lib/icecube/jebclasses.so'\n20:14 < dipo> /home/ofadiran/std-processing/build/lib/icecube/jebclasses.so: cannot open \n (/home/ofadiran/std-processing/build/lib/icecube/jebclasses.so)\n20:14 <@straszhm> okay, then ls -l $I3_BUILD/lib/icecube\n20:15 <@straszhm> maybe you'll want to rm that jebclasses.so and redo 'make jebclasses-pybindings'\n20:16 <@straszhm> it isn't readable and i bet it is empty or corrupt or something\n20:16 < dipo> o.k, I'll do that ..thanks\n20:16 <@straszhm> obviously the file has to be readable to be loadable into python\n20:16 <@straszhm> what did the ls -l return\n20:16 <@straszhm> (of $I3_BUILD/lib/icecube/jebclasses.so)\n20:17 < dipo> list of .so files, cfirst.so, coordinate_service.so, ..... not including jebclasses\n20:18 <@straszhm> but you said that file existed\n20:18 <@straszhm> but you said that file existed\n20:18 <@straszhm> oops\n20:19 <@straszhm> 20:03 <@straszhm> do you have a file $I3_BUILD/lib/icecube/jebclasses.so?\n20:19 <@straszhm> 20:03 < dipo> yes\n20:19 < dipo> yes, ls -al in $BUILD/lib/ : -rwxrwxr-x 1 ofadiran ofadiran 644258 Dec 19 10:20 \n libjebclasses.so\n20:20 <@straszhm> that's not lib/icecube/jebclasses.so\n20:20 < dipo> I'm sorry, my mistake\n20:20 < dipo> I was looking in $BUILD/lib/\n20:21 < dipo> it does not exist in $BUILD/lib/icecube\n20:21 <@straszhm> ok, well good that what python was telling you makes sense\n20:21 <@straszhm> what happens when you 'make jebclasses-pybindings'\n20:22 < dipo> I'll do that now\n20:25 < dipo> works now, thanks for the help\n20:25 <@straszhm> sure\n", "reporter": "troy", "cc": "", "resolution": "wontfix", "_ts": "1416713877165085", "component": "documentation", "summary": "'cannot import module' troubleshooting", "priority": "normal", "keywords": "", "time": "2008-12-20T01:27:47", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
'cannot import module' troubleshooting (Trac #506) - Migrated from https://code.icecube.wisc.edu/ticket/506 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "19:59 < dipo> hi, have a quick question on pybindings, anybody here who can help?\n20:02 <@straszhm> hi\n20:02 <@straszhm> the jebclasses question?\n20:02 < dipo> yes\n20:03 <@straszhm> do you have a file $I3_BUILD/lib/icecube/jebclasses.so?\n20:03 < dipo> yes\n20:04 <@straszhm> hm\n20:05 <@straszhm> no other error messages from your python script?\n20:05 < dipo> no, I'm actually doing things interactively from the prompt\n20:06 <@straszhm> if you just put into a script\n20:06 <@straszhm> 'from icecube import icetray, dataclasses, jebclasses'\n20:06 <@straszhm> (only that)\n20:06 <@straszhm> does it fail?\n20:09 < dipo> sorry it took a while, trying it out, yes, it gave the same error:\n20:10 < dipo> ImportError: cannot import name jebclasses\n20:10 <@straszhm> ok. python has a flag '-v'\n20:10 <@straszhm> that shows verbose information about how it finds imporrts\n20:11 <@straszhm> maybe this will show something. \n20:11 <@straszhm> does it get icetray and dataclasses from the right directory?\n20:11 <@straszhm> (so, run python -v testscript.py)\n20:12 < dipo> yes\n20:12 <@straszhm> anything interesting where the jebclasses search starts?\n20:12 < dipo> import icecube.icetray # dynamically loaded from \n /home/ofadiran/std-processing/build/lib/icecube/icetray.so\n20:12 <@straszhm> right, that looks good\n20:12 < dipo> for icetray, but for jebclasses:\n20:12 < dipo> just the same short error msg that I pasted b4\n20:13 <@straszhm> what is the output of 'file $I3_BUILD/lib/icecube/jebclasses.so'\n20:14 < dipo> /home/ofadiran/std-processing/build/lib/icecube/jebclasses.so: cannot open \n (/home/ofadiran/std-processing/build/lib/icecube/jebclasses.so)\n20:14 <@straszhm> okay, then ls -l $I3_BUILD/lib/icecube\n20:15 <@straszhm> maybe you'll want to rm that jebclasses.so and redo 'make jebclasses-pybindings'\n20:16 <@straszhm> it isn't readable and i bet it is empty or corrupt or something\n20:16 < dipo> o.k, I'll do that ..thanks\n20:16 <@straszhm> obviously the file has to be readable to be loadable into python\n20:16 <@straszhm> what did the ls -l return\n20:16 <@straszhm> (of $I3_BUILD/lib/icecube/jebclasses.so)\n20:17 < dipo> list of .so files, cfirst.so, coordinate_service.so, ..... not including jebclasses\n20:18 <@straszhm> but you said that file existed\n20:18 <@straszhm> but you said that file existed\n20:18 <@straszhm> oops\n20:19 <@straszhm> 20:03 <@straszhm> do you have a file $I3_BUILD/lib/icecube/jebclasses.so?\n20:19 <@straszhm> 20:03 < dipo> yes\n20:19 < dipo> yes, ls -al in $BUILD/lib/ : -rwxrwxr-x 1 ofadiran ofadiran 644258 Dec 19 10:20 \n libjebclasses.so\n20:20 <@straszhm> that's not lib/icecube/jebclasses.so\n20:20 < dipo> I'm sorry, my mistake\n20:20 < dipo> I was looking in $BUILD/lib/\n20:21 < dipo> it does not exist in $BUILD/lib/icecube\n20:21 <@straszhm> ok, well good that what python was telling you makes sense\n20:21 <@straszhm> what happens when you 'make jebclasses-pybindings'\n20:22 < dipo> I'll do that now\n20:25 < dipo> works now, thanks for the help\n20:25 <@straszhm> sure\n", "reporter": "troy", "cc": "", "resolution": "wontfix", "_ts": "1416713877165085", "component": "documentation", "summary": "'cannot import module' troubleshooting", "priority": "normal", "keywords": "", "time": "2008-12-20T01:27:47", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
cannot import module troubleshooting trac migrated from json status closed changetime description hi have a quick question on pybindings anybody here who can help hi the jebclasses question yes do you have a file build lib icecube jebclasses so yes hm no other error messages from your python script no i m actually doing things interactively from the prompt if you just put into a script from icecube import icetray dataclasses jebclasses only that does it fail sorry it took a while trying it out yes it gave the same error importerror cannot import name jebclasses ok python has a flag v that shows verbose information about how it finds imporrts maybe this will show something does it get icetray and dataclasses from the right directory so run python v testscript py yes anything interesting where the jebclasses search starts import icecube icetray dynamically loaded from n home ofadiran std processing build lib icecube icetray so right that looks good for icetray but for jebclasses just the same short error msg that i pasted what is the output of file build lib icecube jebclasses so home ofadiran std processing build lib icecube jebclasses so cannot open n home ofadiran std processing build lib icecube jebclasses so okay then ls l build lib icecube maybe you ll want to rm that jebclasses so and redo make jebclasses pybindings it isn t readable and i bet it is empty or corrupt or something o k i ll do that thanks obviously the file has to be readable to be loadable into python what did the ls l return of build lib icecube jebclasses so list of so files cfirst so coordinate service so not including jebclasses but you said that file existed but you said that file existed oops do you have a file build lib icecube jebclasses so yes yes ls al in build lib rwxrwxr x ofadiran ofadiran dec n libjebclasses so that s not lib icecube jebclasses so i m sorry my mistake i was looking in build lib it does not exist in build lib icecube ok well good that what python was telling you makes sense what happens when you make jebclasses pybindings i ll do that now works now thanks for the help sure n reporter troy cc resolution wontfix ts component documentation summary cannot import module troubleshooting priority normal keywords time milestone owner nega type defect
1
15,416
2,852,292,100
IssuesEvent
2015-06-01 12:52:21
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[TEST-FAILURE] JCacheClientListenerTest timeout
Team: Client Type: Defect
``` 01:55:30 Running com.hazelcast.client.cache.JCacheClientListenerTest 03:00:16 Build timed out (after 100 minutes). Marking the build as aborted. 03:00:59 FATAL: command execution failed ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK9-Quality-Outreach/40/console
1.0
[TEST-FAILURE] JCacheClientListenerTest timeout - ``` 01:55:30 Running com.hazelcast.client.cache.JCacheClientListenerTest 03:00:16 Build timed out (after 100 minutes). Marking the build as aborted. 03:00:59 FATAL: command execution failed ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK9-Quality-Outreach/40/console
defect
jcacheclientlistenertest timeout running com hazelcast client cache jcacheclientlistenertest build timed out after minutes marking the build as aborted fatal command execution failed
1
55,701
14,640,707,241
IssuesEvent
2020-12-25 03:20:49
unascribed/Fabrication
https://api.github.com/repos/unascribed/Fabrication
closed
Sync Attacker Yaw not working
k: Defect
Sync Attacker Yaw seems to stop working (goes back to default behavior) after player dies Mods: - fabric api - lithium - starlight - fabrication (obviously) - overworld two I am using Java 15 OpenJ9 (but I doubt that would be the problem)
1.0
Sync Attacker Yaw not working - Sync Attacker Yaw seems to stop working (goes back to default behavior) after player dies Mods: - fabric api - lithium - starlight - fabrication (obviously) - overworld two I am using Java 15 OpenJ9 (but I doubt that would be the problem)
defect
sync attacker yaw not working sync attacker yaw seems to stop working goes back to default behavior after player dies mods fabric api lithium starlight fabrication obviously overworld two i am using java but i doubt that would be the problem
1
568,265
16,963,113,932
IssuesEvent
2021-06-29 07:38:30
eventespresso/barista
https://api.github.com/repos/eventespresso/barista
closed
Create GQL Schema and Types for Form Element
C: data systems 🗑 D: EDTR ✏️ P3: med priority 😐 S:1 new 👶🏻 T: feature request 🙏
plz see #757 Form Element (unless we decide on a better name) will be our TS/GQL representation for anything added to a form other than an input such as: - separators - images - headings - any HTML element other than an input Form ELements will utilize the following fields (maybe more): GQL Field | DB Field (*=new) | Type | ??? --- | --- | --- | --- dbId | QST_ID | int | id | UUID* | string | save UUID to DB? content| content| string | text added to element htmlClass | html_class* | string | order | order | int | tag | tag | string | html tag being used [EDIT] possible new field to allow for simpler relations and querying: GQL Field | DB Field (*=new) | Type | ??? --- | --- | --- | --- belongsTo | belongsTo | int/string | ID or UUID of parent form section --- So suggested new schema would be GQL Field | Type | ??? --- | --- | --- id | string | UUID belongsTo | int/string | UUID (or ID) of parent form section content| string | text added to element htmlClass | string | order | int | tag | string | html tag being used
1.0
Create GQL Schema and Types for Form Element - plz see #757 Form Element (unless we decide on a better name) will be our TS/GQL representation for anything added to a form other than an input such as: - separators - images - headings - any HTML element other than an input Form ELements will utilize the following fields (maybe more): GQL Field | DB Field (*=new) | Type | ??? --- | --- | --- | --- dbId | QST_ID | int | id | UUID* | string | save UUID to DB? content| content| string | text added to element htmlClass | html_class* | string | order | order | int | tag | tag | string | html tag being used [EDIT] possible new field to allow for simpler relations and querying: GQL Field | DB Field (*=new) | Type | ??? --- | --- | --- | --- belongsTo | belongsTo | int/string | ID or UUID of parent form section --- So suggested new schema would be GQL Field | Type | ??? --- | --- | --- id | string | UUID belongsTo | int/string | UUID (or ID) of parent form section content| string | text added to element htmlClass | string | order | int | tag | string | html tag being used
non_defect
create gql schema and types for form element plz see form element unless we decide on a better name will be our ts gql representation for anything added to a form other than an input such as separators images headings any html element other than an input form elements will utilize the following fields maybe more gql field db field new type dbid qst id int id uuid string save uuid to db content content string text added to element htmlclass html class string order order int tag tag string html tag being used possible new field to allow for simpler relations and querying gql field db field new type belongsto belongsto int string id or uuid of parent form section so suggested new schema would be gql field type id string uuid belongsto int string uuid or id of parent form section content string text added to element htmlclass string order int tag string html tag being used
0
72,044
23,903,987,076
IssuesEvent
2022-09-08 21:46:48
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
CMS: Vet Center Locations pages do not show mobile on node:view.
Defect Drupal engineering ⭐️ Facilities Vet Center
## Describe the defect When I look at node:view on a Vet Center location page, I can see the "Main and satelite locations" but I can not see "Display additional mobile vet centers" ![image](https://user-images.githubusercontent.com/5752113/189230951-39c770ef-7903-4e83-9ac3-2fe16d497d6e.png) ## To Reproduce Steps to reproduce the behavior: 1. Go to '/node/28131/edit' 2. Notice that it has an entry for a mobile vet center 3. Click on the "View" tab and notice that the mobile vet center does not show. 4. It does show correctly on https://www.va.gov/dundalk-vet-center/locations and the preview server http://preview-prod.vfs.va.gov/preview?nodeId=28131 ## AC / Expected behavior - [ ] When on node view, the field and any entries for it are shown in a similar fashion as "Main and satelite locations" ## Screenshots ![image](https://user-images.githubusercontent.com/5752113/189231634-50c020eb-850e-415c-8489-20586024a7d6.png) ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
CMS: Vet Center Locations pages do not show mobile on node:view. - ## Describe the defect When I look at node:view on a Vet Center location page, I can see the "Main and satelite locations" but I can not see "Display additional mobile vet centers" ![image](https://user-images.githubusercontent.com/5752113/189230951-39c770ef-7903-4e83-9ac3-2fe16d497d6e.png) ## To Reproduce Steps to reproduce the behavior: 1. Go to '/node/28131/edit' 2. Notice that it has an entry for a mobile vet center 3. Click on the "View" tab and notice that the mobile vet center does not show. 4. It does show correctly on https://www.va.gov/dundalk-vet-center/locations and the preview server http://preview-prod.vfs.va.gov/preview?nodeId=28131 ## AC / Expected behavior - [ ] When on node view, the field and any entries for it are shown in a similar fashion as "Main and satelite locations" ## Screenshots ![image](https://user-images.githubusercontent.com/5752113/189231634-50c020eb-850e-415c-8489-20586024a7d6.png) ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
cms vet center locations pages do not show mobile on node view describe the defect when i look at node view on a vet center location page i can see the main and satelite locations but i can not see display additional mobile vet centers to reproduce steps to reproduce the behavior go to node edit notice that it has an entry for a mobile vet center click on the view tab and notice that the mobile vet center does not show it does show correctly on and the preview server ac expected behavior when on node view the field and any entries for it are shown in a similar fashion as main and satelite locations screenshots additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
104,117
8,961,885,927
IssuesEvent
2019-01-28 10:51:18
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
closed
API Test 1 : ApiV1ProjectsIdAutoSuggestionsActivateSuitenameTcnumberGetPathParamTcnumberNullValue
API Test 1 API Test 1
Project : API Test 1 Job : JOB Env : ENV Category : null Tags : null Severity : null Region : AliTest Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/projects/{id}/auto-suggestions/activate/{suiteName}/null Request : Response : Not enough variable values available to expand 'id' Logs : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed]Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed] --- FX Bot ---
2.0
API Test 1 : ApiV1ProjectsIdAutoSuggestionsActivateSuitenameTcnumberGetPathParamTcnumberNullValue - Project : API Test 1 Job : JOB Env : ENV Category : null Tags : null Severity : null Region : AliTest Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/projects/{id}/auto-suggestions/activate/{suiteName}/null Request : Response : Not enough variable values available to expand 'id' Logs : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed]Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed] --- FX Bot ---
non_defect
api test project api test job job env env category null tags null severity null region alitest result fail status code headers endpoint request response not enough variable values available to expand id logs assertion resolved to result assertion resolved to result assertion resolved to result assertion resolved to result fx bot
0
48,838
13,184,754,331
IssuesEvent
2020-08-12 20:01:59
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Problems with PPC in GPU mode on Ubuntu 10.04 (Trac #298)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/298 , reported by icecube and owned by yanezjua@ifh.de_</summary> <p> ```json { "status": "closed", "changetime": "2012-02-02T16:12:13", "description": "First error, compiling:\n/usr/local/cuda/lib/libcudart.so: could not read symbols: File in wrong format\nIt was entering the lib instead of the lib64 directory. Removed the lib directory and created a symlink lib->lib64. Then everything compiled without errors.\n\nThe NVIDIA tests for the card (GPU computing SDK) compile and run with or without the change previously described.\n\nAfter compiling, testing the example script (with the proper path modifications) stops at:\nProcessing files: [filename]\nCUDA error: invalid device function\n\nNo more information.\n\nSystem details:\nUbuntu 10.04 x86_64 2.6.32-33 generic\nNvidia development driver 270.41.19\nCUDA toolkit 4.0.17\nPorts v4\ncmake 2.8.4\nPPC revision V00-00-03\nIcesim 02-05-09 RC\n\nI have tried so far using as well\nPPC r7352, same error\nCUDA toolkit 3.2 (specific for Ubuntu 10.04), same errors\nand the few ideas that the internet could give me. No success.\n\n", "reporter": "icecube", "cc": "", "resolution": "wontfix", "_ts": "1328199133000000", "component": "combo simulation", "summary": "Problems with PPC in GPU mode on Ubuntu 10.04", "priority": "normal", "keywords": "ppc gpu ubuntu cuda", "time": "2011-07-21T16:57:52", "milestone": "", "owner": "yanezjua@ifh.de", "type": "defect" } ``` </p> </details>
1.0
Problems with PPC in GPU mode on Ubuntu 10.04 (Trac #298) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/298 , reported by icecube and owned by yanezjua@ifh.de_</summary> <p> ```json { "status": "closed", "changetime": "2012-02-02T16:12:13", "description": "First error, compiling:\n/usr/local/cuda/lib/libcudart.so: could not read symbols: File in wrong format\nIt was entering the lib instead of the lib64 directory. Removed the lib directory and created a symlink lib->lib64. Then everything compiled without errors.\n\nThe NVIDIA tests for the card (GPU computing SDK) compile and run with or without the change previously described.\n\nAfter compiling, testing the example script (with the proper path modifications) stops at:\nProcessing files: [filename]\nCUDA error: invalid device function\n\nNo more information.\n\nSystem details:\nUbuntu 10.04 x86_64 2.6.32-33 generic\nNvidia development driver 270.41.19\nCUDA toolkit 4.0.17\nPorts v4\ncmake 2.8.4\nPPC revision V00-00-03\nIcesim 02-05-09 RC\n\nI have tried so far using as well\nPPC r7352, same error\nCUDA toolkit 3.2 (specific for Ubuntu 10.04), same errors\nand the few ideas that the internet could give me. No success.\n\n", "reporter": "icecube", "cc": "", "resolution": "wontfix", "_ts": "1328199133000000", "component": "combo simulation", "summary": "Problems with PPC in GPU mode on Ubuntu 10.04", "priority": "normal", "keywords": "ppc gpu ubuntu cuda", "time": "2011-07-21T16:57:52", "milestone": "", "owner": "yanezjua@ifh.de", "type": "defect" } ``` </p> </details>
defect
problems with ppc in gpu mode on ubuntu trac migrated from reported by icecube and owned by yanezjua ifh de json status closed changetime description first error compiling n usr local cuda lib libcudart so could not read symbols file in wrong format nit was entering the lib instead of the directory removed the lib directory and created a symlink lib then everything compiled without errors n nthe nvidia tests for the card gpu computing sdk compile and run with or without the change previously described n nafter compiling testing the example script with the proper path modifications stops at nprocessing files ncuda error invalid device function n nno more information n nsystem details nubuntu generic nnvidia development driver ncuda toolkit nports ncmake nppc revision nicesim rc n ni have tried so far using as well nppc same error ncuda toolkit specific for ubuntu same errors nand the few ideas that the internet could give me no success n n reporter icecube cc resolution wontfix ts component combo simulation summary problems with ppc in gpu mode on ubuntu priority normal keywords ppc gpu ubuntu cuda time milestone owner yanezjua ifh de type defect
1
197,525
6,960,410,951
IssuesEvent
2017-12-08 03:18:01
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
API Issue: user/unlock allows purchase of unavailable items
not yet sectioned priority: minor status: issue: suggestion-discussion
[//]: # (Fill out relevant information - UUID is found in Settings -> API) General Info * UUID: b0413351-405f-416f-8787-947ec1c85199 ### Description [//]: # (Describe bug in detail here. Include pictures if helpful.) Seasonal items that are unavailable at a certain time can still be purchased using user/unlock. For example, Halloween skins
1.0
API Issue: user/unlock allows purchase of unavailable items - [//]: # (Fill out relevant information - UUID is found in Settings -> API) General Info * UUID: b0413351-405f-416f-8787-947ec1c85199 ### Description [//]: # (Describe bug in detail here. Include pictures if helpful.) Seasonal items that are unavailable at a certain time can still be purchased using user/unlock. For example, Halloween skins
non_defect
api issue user unlock allows purchase of unavailable items fill out relevant information uuid is found in settings api general info uuid description describe bug in detail here include pictures if helpful seasonal items that are unavailable at a certain time can still be purchased using user unlock for example halloween skins
0
19,354
3,193,760,727
IssuesEvent
2015-09-30 08:05:40
android-rcs/rcsjta
https://api.github.com/repos/android-rcs/rcsjta
closed
Download File Transfer : SQLiteConstraintException
auto-migrated Priority-Medium Type-Defect
``` Produced only 1 time. The device receives an invitation to transfer a file via HTTP. Upon insertion of the file transfer into the databae provider, an SQLiteConstraintException is thrown telling that "column ft_id is not unique". Content of file transfer provider is not available (no root privilege). Logcat trace is provided. ``` Original issue reported on code.google.com by `lemordan...@gmail.com` on 10 Jun 2015 at 3:10 Attachments: * [SQLiteHttpFT_ConstraintException_column-ft_id_is_not_unique.txt](https://storage.googleapis.com/google-code-attachments/rcsjta/issue-111/comment-0/SQLiteHttpFT_ConstraintException_column-ft_id_is_not_unique.txt)
1.0
Download File Transfer : SQLiteConstraintException - ``` Produced only 1 time. The device receives an invitation to transfer a file via HTTP. Upon insertion of the file transfer into the databae provider, an SQLiteConstraintException is thrown telling that "column ft_id is not unique". Content of file transfer provider is not available (no root privilege). Logcat trace is provided. ``` Original issue reported on code.google.com by `lemordan...@gmail.com` on 10 Jun 2015 at 3:10 Attachments: * [SQLiteHttpFT_ConstraintException_column-ft_id_is_not_unique.txt](https://storage.googleapis.com/google-code-attachments/rcsjta/issue-111/comment-0/SQLiteHttpFT_ConstraintException_column-ft_id_is_not_unique.txt)
defect
download file transfer sqliteconstraintexception produced only time the device receives an invitation to transfer a file via http upon insertion of the file transfer into the databae provider an sqliteconstraintexception is thrown telling that column ft id is not unique content of file transfer provider is not available no root privilege logcat trace is provided original issue reported on code google com by lemordan gmail com on jun at attachments
1
43,977
9,527,822,937
IssuesEvent
2019-04-29 06:16:51
theia-ide/theia
https://api.github.com/repos/theia-ide/theia
closed
[plugin][webview][intermittant] window.postMessageExt is not a function
bug plug-in system vscode
In some of our webviews, we use webview.postMessage() at very early stage to get some data before rendering happens. We get above error intermittently when we open those webviews - mostly first render breaks and others are successful. Going through the source code, I found the place where postMessageExt is injected to window object of iframe at https://github.com/theia-ide/theia/blob/master/packages/plugin-ext/src/main/browser/webview/webview.ts#L99 My initial guess is that it has something to do with timeouts used below there before calling onLoad.
1.0
[plugin][webview][intermittant] window.postMessageExt is not a function - In some of our webviews, we use webview.postMessage() at very early stage to get some data before rendering happens. We get above error intermittently when we open those webviews - mostly first render breaks and others are successful. Going through the source code, I found the place where postMessageExt is injected to window object of iframe at https://github.com/theia-ide/theia/blob/master/packages/plugin-ext/src/main/browser/webview/webview.ts#L99 My initial guess is that it has something to do with timeouts used below there before calling onLoad.
non_defect
window postmessageext is not a function in some of our webviews we use webview postmessage at very early stage to get some data before rendering happens we get above error intermittently when we open those webviews mostly first render breaks and others are successful going through the source code i found the place where postmessageext is injected to window object of iframe at my initial guess is that it has something to do with timeouts used below there before calling onload
0
94,872
27,319,053,741
IssuesEvent
2023-02-24 18:04:48
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
[7.0] Test failure in JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh()
arch-wasm os-linux test-failure Known Build Error
- Affected PR: https://github.com/dotnet/runtime/pull/82554 - Queue: `Build Browser wasm Linux Release AllSubsets_Mono_RuntimeTests` - Job results: https://dev.azure.com/dnceng-public/public/_build/results?buildId=182544&view=logs&j=190ad6c8-5950-568c-cadd-f2dfb7d5a79f&t=c0f6fdc1-ac5d-583c-8ae1-a18de0846552&l=37 - Log file: https://helixre107v0xd1eu3ibi6ka.blob.core.windows.net/dotnet-runtime-refs-pull-82554-merge-7130257ad20d4e91a0/JIT.HardwareIntrinsics.X86.Fma/1/console.08658d22.log?helixlogtype=result ``` Starting: JIT.HardwareIntrinsics.XUnitWrapper (parallel test collections = on, max threads = 2) JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh [FAIL] System.IO.IOException: The system cannot open the device or file specified. : 'NuGet-Migrations' at System.Threading.Mutex.CreateMutexCore(Boolean initiallyOwned, String name, Boolean& createdNew) at System.Threading.Mutex..ctor(Boolean initiallyOwned, String name) at NuGet.Common.Migrations.MigrationRunner.Run() at Microsoft.DotNet.Configurer.DotnetFirstTimeUseConfigurer.Configure() at Microsoft.DotNet.Cli.Program.ConfigureDotNetForFirstTimeUse(IFirstTimeUseNoticeSentinel firstTimeUseNoticeSentinel, IAspNetCertificateSentinel aspNetCertificateSentinel, IFileSentinel toolPathSentinel, Boolean isDotnetBeingInvokedFromNativeInstaller, DotnetFirstRunConfiguration dotnetFirstRunConfiguration, IEnvironmentProvider environmentProvider, Dictionary`2 performanceMeasurements) at Microsoft.DotNet.Cli.Program.ProcessArgs(String[] args, TimeSpan startupTime, ITelemetry telemetryClient) at Microsoft.DotNet.Cli.Program.Main(String[] args) Return code: 1 Raw output file: /datadisks/disk1/work/A5FD0965/w/AD6C0924/uploads/Reports/JIT.HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.output.txt Raw output: BEGIN EXECUTION Test Harness Exitcode is : 1 To run the test: > set CORE_ROOT=/datadisks/disk1/work/A5FD0965/p > /datadisks/disk1/work/A5FD0965/w/AD6C0924/e/JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh Expected: True Actual: False Stack Trace: at JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) Output: System.IO.IOException: The system cannot open the device or file specified. : 'NuGet-Migrations' at System.Threading.Mutex.CreateMutexCore(Boolean initiallyOwned, String name, Boolean& createdNew) at System.Threading.Mutex..ctor(Boolean initiallyOwned, String name) at NuGet.Common.Migrations.MigrationRunner.Run() at Microsoft.DotNet.Configurer.DotnetFirstTimeUseConfigurer.Configure() at Microsoft.DotNet.Cli.Program.ConfigureDotNetForFirstTimeUse(IFirstTimeUseNoticeSentinel firstTimeUseNoticeSentinel, IAspNetCertificateSentinel aspNetCertificateSentinel, IFileSentinel toolPathSentinel, Boolean isDotnetBeingInvokedFromNativeInstaller, DotnetFirstRunConfiguration dotnetFirstRunConfiguration, IEnvironmentProvider environmentProvider, Dictionary`2 performanceMeasurements) at Microsoft.DotNet.Cli.Program.ProcessArgs(String[] args, TimeSpan startupTime, ITelemetry telemetryClient) at Microsoft.DotNet.Cli.Program.Main(String[] args) Return code: 1 Raw output file: /datadisks/disk1/work/A5FD0965/w/AD6C0924/uploads/Reports/JIT.HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.output.txt Raw output: BEGIN EXECUTION Test Harness Exitcode is : 1 To run the test: > set CORE_ROOT=/datadisks/disk1/work/A5FD0965/p > /datadisks/disk1/work/A5FD0965/w/AD6C0924/e/JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh Finished: JIT.HardwareIntrinsics.XUnitWrapper === TEST EXECUTION SUMMARY === JIT.HardwareIntrinsics.XUnitWrapper Total: 4, Errors: 0, Failed: 1, Skipped: 0, Time: 4.876s ``` ```json { "ErrorMessage": "JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh()", "BuildRetry": false, "ErrorPattern": "", "ExcludeConsoleLog": false } ```
1.0
[7.0] Test failure in JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh() - - Affected PR: https://github.com/dotnet/runtime/pull/82554 - Queue: `Build Browser wasm Linux Release AllSubsets_Mono_RuntimeTests` - Job results: https://dev.azure.com/dnceng-public/public/_build/results?buildId=182544&view=logs&j=190ad6c8-5950-568c-cadd-f2dfb7d5a79f&t=c0f6fdc1-ac5d-583c-8ae1-a18de0846552&l=37 - Log file: https://helixre107v0xd1eu3ibi6ka.blob.core.windows.net/dotnet-runtime-refs-pull-82554-merge-7130257ad20d4e91a0/JIT.HardwareIntrinsics.X86.Fma/1/console.08658d22.log?helixlogtype=result ``` Starting: JIT.HardwareIntrinsics.XUnitWrapper (parallel test collections = on, max threads = 2) JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh [FAIL] System.IO.IOException: The system cannot open the device or file specified. : 'NuGet-Migrations' at System.Threading.Mutex.CreateMutexCore(Boolean initiallyOwned, String name, Boolean& createdNew) at System.Threading.Mutex..ctor(Boolean initiallyOwned, String name) at NuGet.Common.Migrations.MigrationRunner.Run() at Microsoft.DotNet.Configurer.DotnetFirstTimeUseConfigurer.Configure() at Microsoft.DotNet.Cli.Program.ConfigureDotNetForFirstTimeUse(IFirstTimeUseNoticeSentinel firstTimeUseNoticeSentinel, IAspNetCertificateSentinel aspNetCertificateSentinel, IFileSentinel toolPathSentinel, Boolean isDotnetBeingInvokedFromNativeInstaller, DotnetFirstRunConfiguration dotnetFirstRunConfiguration, IEnvironmentProvider environmentProvider, Dictionary`2 performanceMeasurements) at Microsoft.DotNet.Cli.Program.ProcessArgs(String[] args, TimeSpan startupTime, ITelemetry telemetryClient) at Microsoft.DotNet.Cli.Program.Main(String[] args) Return code: 1 Raw output file: /datadisks/disk1/work/A5FD0965/w/AD6C0924/uploads/Reports/JIT.HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.output.txt Raw output: BEGIN EXECUTION Test Harness Exitcode is : 1 To run the test: > set CORE_ROOT=/datadisks/disk1/work/A5FD0965/p > /datadisks/disk1/work/A5FD0965/w/AD6C0924/e/JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh Expected: True Actual: False Stack Trace: at JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) Output: System.IO.IOException: The system cannot open the device or file specified. : 'NuGet-Migrations' at System.Threading.Mutex.CreateMutexCore(Boolean initiallyOwned, String name, Boolean& createdNew) at System.Threading.Mutex..ctor(Boolean initiallyOwned, String name) at NuGet.Common.Migrations.MigrationRunner.Run() at Microsoft.DotNet.Configurer.DotnetFirstTimeUseConfigurer.Configure() at Microsoft.DotNet.Cli.Program.ConfigureDotNetForFirstTimeUse(IFirstTimeUseNoticeSentinel firstTimeUseNoticeSentinel, IAspNetCertificateSentinel aspNetCertificateSentinel, IFileSentinel toolPathSentinel, Boolean isDotnetBeingInvokedFromNativeInstaller, DotnetFirstRunConfiguration dotnetFirstRunConfiguration, IEnvironmentProvider environmentProvider, Dictionary`2 performanceMeasurements) at Microsoft.DotNet.Cli.Program.ProcessArgs(String[] args, TimeSpan startupTime, ITelemetry telemetryClient) at Microsoft.DotNet.Cli.Program.Main(String[] args) Return code: 1 Raw output file: /datadisks/disk1/work/A5FD0965/w/AD6C0924/uploads/Reports/JIT.HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.output.txt Raw output: BEGIN EXECUTION Test Harness Exitcode is : 1 To run the test: > set CORE_ROOT=/datadisks/disk1/work/A5FD0965/p > /datadisks/disk1/work/A5FD0965/w/AD6C0924/e/JIT/HardwareIntrinsics/X86/Fma_Vector256/Fma_r/Fma_r.sh Finished: JIT.HardwareIntrinsics.XUnitWrapper === TEST EXECUTION SUMMARY === JIT.HardwareIntrinsics.XUnitWrapper Total: 4, Errors: 0, Failed: 1, Skipped: 0, Time: 4.876s ``` ```json { "ErrorMessage": "JIT_HardwareIntrinsics._X86_Fma_Vector256_Fma_r_Fma_r_._X86_Fma_Vector256_Fma_r_Fma_r_sh()", "BuildRetry": false, "ErrorPattern": "", "ExcludeConsoleLog": false } ```
non_defect
test failure in jit hardwareintrinsics fma fma r fma r fma fma r fma r sh affected pr queue build browser wasm linux release allsubsets mono runtimetests job results log file starting jit hardwareintrinsics xunitwrapper parallel test collections on max threads jit hardwareintrinsics fma fma r fma r sh system io ioexception the system cannot open the device or file specified nuget migrations at system threading mutex createmutexcore boolean initiallyowned string name boolean creatednew at system threading mutex ctor boolean initiallyowned string name at nuget common migrations migrationrunner run at microsoft dotnet configurer dotnetfirsttimeuseconfigurer configure at microsoft dotnet cli program configuredotnetforfirsttimeuse ifirsttimeusenoticesentinel firsttimeusenoticesentinel iaspnetcertificatesentinel aspnetcertificatesentinel ifilesentinel toolpathsentinel boolean isdotnetbeinginvokedfromnativeinstaller dotnetfirstrunconfiguration dotnetfirstrunconfiguration ienvironmentprovider environmentprovider dictionary performancemeasurements at microsoft dotnet cli program processargs string args timespan startuptime itelemetry telemetryclient at microsoft dotnet cli program main string args return code raw output file datadisks work w uploads reports jit hardwareintrinsics fma fma r fma r output txt raw output begin execution test harness exitcode is to run the test set core root datadisks work p datadisks work w e jit hardwareintrinsics fma fma r fma r sh expected true actual false stack trace at jit hardwareintrinsics fma fma r fma r fma fma r fma r sh at system runtimemethodhandle invokemethod object target void arguments signature sig boolean isconstructor at system reflection methodinvoker invoke object obj intptr args bindingflags invokeattr output system io ioexception the system cannot open the device or file specified nuget migrations at system threading mutex createmutexcore boolean initiallyowned string name boolean creatednew at system threading mutex ctor boolean initiallyowned string name at nuget common migrations migrationrunner run at microsoft dotnet configurer dotnetfirsttimeuseconfigurer configure at microsoft dotnet cli program configuredotnetforfirsttimeuse ifirsttimeusenoticesentinel firsttimeusenoticesentinel iaspnetcertificatesentinel aspnetcertificatesentinel ifilesentinel toolpathsentinel boolean isdotnetbeinginvokedfromnativeinstaller dotnetfirstrunconfiguration dotnetfirstrunconfiguration ienvironmentprovider environmentprovider dictionary performancemeasurements at microsoft dotnet cli program processargs string args timespan startuptime itelemetry telemetryclient at microsoft dotnet cli program main string args return code raw output file datadisks work w uploads reports jit hardwareintrinsics fma fma r fma r output txt raw output begin execution test harness exitcode is to run the test set core root datadisks work p datadisks work w e jit hardwareintrinsics fma fma r fma r sh finished jit hardwareintrinsics xunitwrapper test execution summary jit hardwareintrinsics xunitwrapper total errors failed skipped time json errormessage jit hardwareintrinsics fma fma r fma r fma fma r fma r sh buildretry false errorpattern excludeconsolelog false
0
281,605
8,697,397,414
IssuesEvent
2018-12-04 20:09:16
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[engine] Update logging to use slf4j with log4j2
priority: medium task
Update Crafter Engine to use slf4j with log4j2 as the logging framework. Sub-task of https://github.com/craftercms/craftercms/issues/2658.
1.0
[engine] Update logging to use slf4j with log4j2 - Update Crafter Engine to use slf4j with log4j2 as the logging framework. Sub-task of https://github.com/craftercms/craftercms/issues/2658.
non_defect
update logging to use with update crafter engine to use with as the logging framework sub task of
0
7,911
2,611,064,969
IssuesEvent
2015-02-27 00:30:20
alistairreilly/andors-trail
https://api.github.com/repos/alistairreilly/andors-trail
closed
Tavern clutter
auto-migrated Type-Defect
``` Related to issue 77:http://code.google.com/p/andors-trail/issues/detail?id=77 but not the same, so I posted it separately. This is the biggest problem in the tavern in the first village, but it applies elsewhere too: There are a lot of drunks in the first tavern. After grinding in the snake cave (etc), I feel like I have to play dodge-ball to navigate through the crowds of drunks and tables, or else get stopped by the "drink, drink, drink" message for the umpteenth time. playing with a 4-direction D-pad makes it almost impossible, too. Related: in the second town, why do we have the bed farthest from the front? it's just extra hassle for the player, because the bed is a very common destination. ``` Original issue reported on code.google.com by `jch...@email.wm.edu` on 8 Apr 2011 at 4:58 * Merged into: #77
1.0
Tavern clutter - ``` Related to issue 77:http://code.google.com/p/andors-trail/issues/detail?id=77 but not the same, so I posted it separately. This is the biggest problem in the tavern in the first village, but it applies elsewhere too: There are a lot of drunks in the first tavern. After grinding in the snake cave (etc), I feel like I have to play dodge-ball to navigate through the crowds of drunks and tables, or else get stopped by the "drink, drink, drink" message for the umpteenth time. playing with a 4-direction D-pad makes it almost impossible, too. Related: in the second town, why do we have the bed farthest from the front? it's just extra hassle for the player, because the bed is a very common destination. ``` Original issue reported on code.google.com by `jch...@email.wm.edu` on 8 Apr 2011 at 4:58 * Merged into: #77
defect
tavern clutter related to issue but not the same so i posted it separately this is the biggest problem in the tavern in the first village but it applies elsewhere too there are a lot of drunks in the first tavern after grinding in the snake cave etc i feel like i have to play dodge ball to navigate through the crowds of drunks and tables or else get stopped by the drink drink drink message for the umpteenth time playing with a direction d pad makes it almost impossible too related in the second town why do we have the bed farthest from the front it s just extra hassle for the player because the bed is a very common destination original issue reported on code google com by jch email wm edu on apr at merged into
1
360,746
25,308,006,166
IssuesEvent
2022-11-17 15:27:04
mebjas/html5-qrcode
https://api.github.com/repos/mebjas/html5-qrcode
closed
[Feature Request] Move documentation to alternative site / Wiki
enhancement ferature-request documentation
Documentation currently sits on the readme page. - [ ] Move documentation to a different area. - [ ] In the best case it's auto generated using the code. - [ ] Have links to blog articles. - [ ] Migrate mini articles to scanapp articles. - [ ] Add references to readme and cleanup.
1.0
[Feature Request] Move documentation to alternative site / Wiki - Documentation currently sits on the readme page. - [ ] Move documentation to a different area. - [ ] In the best case it's auto generated using the code. - [ ] Have links to blog articles. - [ ] Migrate mini articles to scanapp articles. - [ ] Add references to readme and cleanup.
non_defect
move documentation to alternative site wiki documentation currently sits on the readme page move documentation to a different area in the best case it s auto generated using the code have links to blog articles migrate mini articles to scanapp articles add references to readme and cleanup
0
8,058
5,368,675,282
IssuesEvent
2017-02-22 09:31:16
zaproxy/zaproxy
https://api.github.com/repos/zaproxy/zaproxy
opened
Active Scan API - Allow to start the scans with non-leaf nodes
Component-API enhancement Usability
From https://groups.google.com/d/topic/zaproxy-users/jU2bt0f4b-c/discussion (although that's not the only thread/issue about "URL Not Found"). Steps to reproduce the issue: 1. Start ZAP; 2. Access the targets to be scanned, e.g.: - http://example.com/dir/file1.jsp - http://example.com/dir/file2.jsp 3. Start the active scan, through the ZAP API, with the "directory" that contains the files http://example.com/dir/ 4. Note that it says "URL Not Found in the Scan Tree" (although it exists as non-leaf node). The change will allow to start the active scans through the ZAP API with a directory (non-leaf node), as it's possible through the GUI when selecting the "dir" node.
True
Active Scan API - Allow to start the scans with non-leaf nodes - From https://groups.google.com/d/topic/zaproxy-users/jU2bt0f4b-c/discussion (although that's not the only thread/issue about "URL Not Found"). Steps to reproduce the issue: 1. Start ZAP; 2. Access the targets to be scanned, e.g.: - http://example.com/dir/file1.jsp - http://example.com/dir/file2.jsp 3. Start the active scan, through the ZAP API, with the "directory" that contains the files http://example.com/dir/ 4. Note that it says "URL Not Found in the Scan Tree" (although it exists as non-leaf node). The change will allow to start the active scans through the ZAP API with a directory (non-leaf node), as it's possible through the GUI when selecting the "dir" node.
non_defect
active scan api allow to start the scans with non leaf nodes from although that s not the only thread issue about url not found steps to reproduce the issue start zap access the targets to be scanned e g start the active scan through the zap api with the directory that contains the files note that it says url not found in the scan tree although it exists as non leaf node the change will allow to start the active scans through the zap api with a directory non leaf node as it s possible through the gui when selecting the dir node
0
2,983
2,787,870,697
IssuesEvent
2015-05-08 09:35:11
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
opened
Bug: Column path in tbl j25_menu is rebuilt out of wrong menu
No Code Attached Yet
#### Steps to reproduce the issue Having at least 2 Menus, say menu and menu2 (that ist column `menutype` in tbl `j25_menu`). Attached you find the menu structure at my project with replaced names. - I have marked the entries with id 120 and 239. In the Joomla administration, after editing entry 239 and then save, column `path` is updated wrongly con![screen shot 2015-05-08 at 04 32 48](http://issues.joomla.org/uploads/1/815cbae00747485f425ecf7276c6bb37.png)taining some values of entry 120 (which is not the same menu). #### Expected result I expected entry 239 updated without contents of entry 120. #### Actual result see in description #### System information (as much as possible) Joomla 3.3.6 In ancient times had Joomla 1.5, updated to 2.5 and in turn to 3.x. #### Additional comments I could resolve the problem in `libraries/joomla/table/nested.php` in the 2 methods: * getPath($pk = null, $diagnostic = false) * rebuildPath($pk = null) Both functions select somehow their upper entries - I added `AND p.menutype = n.menutype` to the select queries which resolved the issue for me.
1.0
Bug: Column path in tbl j25_menu is rebuilt out of wrong menu - #### Steps to reproduce the issue Having at least 2 Menus, say menu and menu2 (that ist column `menutype` in tbl `j25_menu`). Attached you find the menu structure at my project with replaced names. - I have marked the entries with id 120 and 239. In the Joomla administration, after editing entry 239 and then save, column `path` is updated wrongly con![screen shot 2015-05-08 at 04 32 48](http://issues.joomla.org/uploads/1/815cbae00747485f425ecf7276c6bb37.png)taining some values of entry 120 (which is not the same menu). #### Expected result I expected entry 239 updated without contents of entry 120. #### Actual result see in description #### System information (as much as possible) Joomla 3.3.6 In ancient times had Joomla 1.5, updated to 2.5 and in turn to 3.x. #### Additional comments I could resolve the problem in `libraries/joomla/table/nested.php` in the 2 methods: * getPath($pk = null, $diagnostic = false) * rebuildPath($pk = null) Both functions select somehow their upper entries - I added `AND p.menutype = n.menutype` to the select queries which resolved the issue for me.
non_defect
bug column path in tbl menu is rebuilt out of wrong menu steps to reproduce the issue having at least menus say menu and that ist column menutype in tbl menu attached you find the menu structure at my project with replaced names i have marked the entries with id and in the joomla administration after editing entry and then save column path is updated wrongly con some values of entry which is not the same menu expected result i expected entry updated without contents of entry actual result see in description system information as much as possible joomla in ancient times had joomla updated to and in turn to x additional comments i could resolve the problem in libraries joomla table nested php in the methods getpath pk null diagnostic false rebuildpath pk null both functions select somehow their upper entries i added and p menutype n menutype to the select queries which resolved the issue for me
0
40,001
20,371,267,575
IssuesEvent
2022-02-21 11:26:55
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Fleet] Improve performance of auto package upgrades
performance Feature:EPM Team:Fleet v8.2.0
We have suspicion that some customers are running into some performance problems on the `/api/fleet/setup` endpoint caused by the "keep policies up-to-date" logic in Fleet setup. This has some opportunity for some optimization improvements which would minimize the overhead of running this logic during Fleet setup. ### Current implementation - Fetch the `id`s for all package policies https://github.com/elastic/kibana/blob/2fa5a87a5f17d8a97b06db7f8d19a30809eb3ade/x-pack/plugins/fleet/server/services/preconfiguration.ts#L357-L367 - For each package policy `id`: - Fetch the full package policy https://github.com/elastic/kibana/blob/2fa5a87a5f17d8a97b06db7f8d19a30809eb3ade/x-pack/plugins/fleet/server/services/managed_package_policies.ts#L39 - Fetch the package info - Fetch the installation - If the package has `keepPoliciesUpToDate == false` or is already up to date, skip this package policy - Run the upgrade dry run for the package policy (which fetches the package policy a 2nd time) - If dry succeeds, run the actual upgrade (which fetches the package policy a 3rd time) ### Proposed implementation Most of the decisions we need to make could be made with much fewer fetches by looking at the packages first and then using that information to filter for the package policies that need to be upgraded. - Fetch the installed packages, filtering for packages with `keepPoliciesUpToDate = true` and `name in [managed package names]` - Maybe we already have these fetched in the previous steps of Fleet setup and could just filter in memory? - For each package that matches the query above: - Fetch all package policies where `packageName = installedPackage.name` and `version != installedPackage.version` - For each package policy that matches query above: - Attempt the dry run using the installed package and package policy that has already been fetched - If dry run succeed, run the actual upgrade using the installed package and package policy that has already been fetched
True
[Fleet] Improve performance of auto package upgrades - We have suspicion that some customers are running into some performance problems on the `/api/fleet/setup` endpoint caused by the "keep policies up-to-date" logic in Fleet setup. This has some opportunity for some optimization improvements which would minimize the overhead of running this logic during Fleet setup. ### Current implementation - Fetch the `id`s for all package policies https://github.com/elastic/kibana/blob/2fa5a87a5f17d8a97b06db7f8d19a30809eb3ade/x-pack/plugins/fleet/server/services/preconfiguration.ts#L357-L367 - For each package policy `id`: - Fetch the full package policy https://github.com/elastic/kibana/blob/2fa5a87a5f17d8a97b06db7f8d19a30809eb3ade/x-pack/plugins/fleet/server/services/managed_package_policies.ts#L39 - Fetch the package info - Fetch the installation - If the package has `keepPoliciesUpToDate == false` or is already up to date, skip this package policy - Run the upgrade dry run for the package policy (which fetches the package policy a 2nd time) - If dry succeeds, run the actual upgrade (which fetches the package policy a 3rd time) ### Proposed implementation Most of the decisions we need to make could be made with much fewer fetches by looking at the packages first and then using that information to filter for the package policies that need to be upgraded. - Fetch the installed packages, filtering for packages with `keepPoliciesUpToDate = true` and `name in [managed package names]` - Maybe we already have these fetched in the previous steps of Fleet setup and could just filter in memory? - For each package that matches the query above: - Fetch all package policies where `packageName = installedPackage.name` and `version != installedPackage.version` - For each package policy that matches query above: - Attempt the dry run using the installed package and package policy that has already been fetched - If dry run succeed, run the actual upgrade using the installed package and package policy that has already been fetched
non_defect
improve performance of auto package upgrades we have suspicion that some customers are running into some performance problems on the api fleet setup endpoint caused by the keep policies up to date logic in fleet setup this has some opportunity for some optimization improvements which would minimize the overhead of running this logic during fleet setup current implementation fetch the id s for all package policies for each package policy id fetch the full package policy fetch the package info fetch the installation if the package has keeppoliciesuptodate false or is already up to date skip this package policy run the upgrade dry run for the package policy which fetches the package policy a time if dry succeeds run the actual upgrade which fetches the package policy a time proposed implementation most of the decisions we need to make could be made with much fewer fetches by looking at the packages first and then using that information to filter for the package policies that need to be upgraded fetch the installed packages filtering for packages with keeppoliciesuptodate true and name in maybe we already have these fetched in the previous steps of fleet setup and could just filter in memory for each package that matches the query above fetch all package policies where packagename installedpackage name and version installedpackage version for each package policy that matches query above attempt the dry run using the installed package and package policy that has already been fetched if dry run succeed run the actual upgrade using the installed package and package policy that has already been fetched
0
73,613
24,718,625,309
IssuesEvent
2022-10-20 08:58:59
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: scipy==1.9.2 causes crash of scikit-learn + joblib
defect
### Describe your issue. Current scipy==1.9.2 on Win (AMD64) causes fail of joblib subprocess using scikit-learn. (Tested in Windows container.) Not replicated on Linux and MacOS even with the same settings. Previous bugfix scipy version works (1.9.1). ### Reproducing Code Example ```python Install at first these packages: python -m pip install -U pip setuptools scipy==1.9.2 joblib sklearn Then run this code (e.g. interactively): from joblib import Parallel, delayed import sklearn def a(): from sklearn.model_selection import cross_val_score return cross_val_score data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) The last line fails (above). To fix, just do (e.g. below) and rerun. python -m pip install -U pip scipy==1.9.1 ``` ### Error message ```shell In [5]: data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) --------------------------------------------------------------------------- _RemoteTraceback Traceback (most recent call last) _RemoteTraceback: """ Traceback (most recent call last): File "C:\Python\lib\site-packages\joblib\externals\loky\process_executor.py", line 428, in _process_worker r = call_item() File "C:\Python\lib\site-packages\joblib\externals\loky\process_executor.py", line 275, in __call__ return self.fn(*self.args, **self.kwargs) File "C:\Python\lib\site-packages\joblib\_parallel_backends.py", line 620, in __call__ return self.func(*args, **kwargs) File "C:\Python\lib\site-packages\joblib\parallel.py", line 288, in __call__ return [func(*args, **kwargs) File "C:\Python\lib\site-packages\joblib\parallel.py", line 288, in <listcomp> return [func(*args, **kwargs) File "<ipython-input-1-bf041901e974>", line 2, in a File "C:\Python\lib\site-packages\sklearn\model_selection\__init__.py", line 23, in <module> from ._validation import cross_val_score File "C:\Python\lib\site-packages\sklearn\model_selection\_validation.py", line 32, in <module> from ..metrics import check_scoring File "C:\Python\lib\site-packages\sklearn\metrics\__init__.py", line 41, in <module> from . import cluster File "C:\Python\lib\site-packages\sklearn\metrics\cluster\__init__.py", line 22, in <module> from ._unsupervised import silhouette_samples File "C:\Python\lib\site-packages\sklearn\metrics\cluster\_unsupervised.py", line 16, in <module> from ..pairwise import pairwise_distances_chunked File "C:\Python\lib\site-packages\sklearn\metrics\pairwise.py", line 33, in <module> from ._pairwise_distances_reduction import PairwiseDistancesArgKmin ImportError: DLL load failed while importing _pairwise_distances_reduction: The specified module could not be found. """ The above exception was the direct cause of the following exception: ImportError Traceback (most recent call last) Cell In [5], line 1 ----> 1 data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) File C:\Python\lib\site-packages\joblib\parallel.py:1098, in Parallel.__call__(self, iterable) 1095 self._iterating = False 1097 with self._backend.retrieval_context(): -> 1098 self.retrieve() 1099 # Make sure that we get a last message telling us we are done 1100 elapsed_time = time.time() - self._start_time File C:\Python\lib\site-packages\joblib\parallel.py:975, in Parallel.retrieve(self) 973 try: 974 if getattr(self._backend, 'supports_timeout', False): --> 975 self._output.extend(job.get(timeout=self.timeout)) 976 else: 977 self._output.extend(job.get()) File C:\Python\lib\site-packages\joblib\_parallel_backends.py:567, in LokyBackend.wrap_future_result(future, timeout) 564 """Wrapper for Future.result to implement the same behaviour as 565 AsyncResults.get from multiprocessing.""" 566 try: --> 567 return future.result(timeout=timeout) 568 except CfTimeoutError as e: 569 raise TimeoutError from e File C:\Python\lib\concurrent\futures\_base.py:458, in Future.result(self, timeout) 456 raise CancelledError() 457 elif self._state == FINISHED: --> 458 return self.__get_result() 459 else: 460 raise TimeoutError() File C:\Python\lib\concurrent\futures\_base.py:403, in Future.__get_result(self) 401 if self._exception: 402 try: --> 403 raise self._exception 404 finally: 405 # Break a reference cycle with the exception in self._exception 406 self = None ImportError: DLL load failed while importing _pairwise_distances_reduction: The specified module could not be found. ``` ### SciPy/NumPy/Python version information 1.9.2 1.23.4 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0)
1.0
BUG: scipy==1.9.2 causes crash of scikit-learn + joblib - ### Describe your issue. Current scipy==1.9.2 on Win (AMD64) causes fail of joblib subprocess using scikit-learn. (Tested in Windows container.) Not replicated on Linux and MacOS even with the same settings. Previous bugfix scipy version works (1.9.1). ### Reproducing Code Example ```python Install at first these packages: python -m pip install -U pip setuptools scipy==1.9.2 joblib sklearn Then run this code (e.g. interactively): from joblib import Parallel, delayed import sklearn def a(): from sklearn.model_selection import cross_val_score return cross_val_score data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) The last line fails (above). To fix, just do (e.g. below) and rerun. python -m pip install -U pip scipy==1.9.1 ``` ### Error message ```shell In [5]: data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) --------------------------------------------------------------------------- _RemoteTraceback Traceback (most recent call last) _RemoteTraceback: """ Traceback (most recent call last): File "C:\Python\lib\site-packages\joblib\externals\loky\process_executor.py", line 428, in _process_worker r = call_item() File "C:\Python\lib\site-packages\joblib\externals\loky\process_executor.py", line 275, in __call__ return self.fn(*self.args, **self.kwargs) File "C:\Python\lib\site-packages\joblib\_parallel_backends.py", line 620, in __call__ return self.func(*args, **kwargs) File "C:\Python\lib\site-packages\joblib\parallel.py", line 288, in __call__ return [func(*args, **kwargs) File "C:\Python\lib\site-packages\joblib\parallel.py", line 288, in <listcomp> return [func(*args, **kwargs) File "<ipython-input-1-bf041901e974>", line 2, in a File "C:\Python\lib\site-packages\sklearn\model_selection\__init__.py", line 23, in <module> from ._validation import cross_val_score File "C:\Python\lib\site-packages\sklearn\model_selection\_validation.py", line 32, in <module> from ..metrics import check_scoring File "C:\Python\lib\site-packages\sklearn\metrics\__init__.py", line 41, in <module> from . import cluster File "C:\Python\lib\site-packages\sklearn\metrics\cluster\__init__.py", line 22, in <module> from ._unsupervised import silhouette_samples File "C:\Python\lib\site-packages\sklearn\metrics\cluster\_unsupervised.py", line 16, in <module> from ..pairwise import pairwise_distances_chunked File "C:\Python\lib\site-packages\sklearn\metrics\pairwise.py", line 33, in <module> from ._pairwise_distances_reduction import PairwiseDistancesArgKmin ImportError: DLL load failed while importing _pairwise_distances_reduction: The specified module could not be found. """ The above exception was the direct cause of the following exception: ImportError Traceback (most recent call last) Cell In [5], line 1 ----> 1 data_results = Parallel(n_jobs=4)(delayed(a)() for i in range(10)) File C:\Python\lib\site-packages\joblib\parallel.py:1098, in Parallel.__call__(self, iterable) 1095 self._iterating = False 1097 with self._backend.retrieval_context(): -> 1098 self.retrieve() 1099 # Make sure that we get a last message telling us we are done 1100 elapsed_time = time.time() - self._start_time File C:\Python\lib\site-packages\joblib\parallel.py:975, in Parallel.retrieve(self) 973 try: 974 if getattr(self._backend, 'supports_timeout', False): --> 975 self._output.extend(job.get(timeout=self.timeout)) 976 else: 977 self._output.extend(job.get()) File C:\Python\lib\site-packages\joblib\_parallel_backends.py:567, in LokyBackend.wrap_future_result(future, timeout) 564 """Wrapper for Future.result to implement the same behaviour as 565 AsyncResults.get from multiprocessing.""" 566 try: --> 567 return future.result(timeout=timeout) 568 except CfTimeoutError as e: 569 raise TimeoutError from e File C:\Python\lib\concurrent\futures\_base.py:458, in Future.result(self, timeout) 456 raise CancelledError() 457 elif self._state == FINISHED: --> 458 return self.__get_result() 459 else: 460 raise TimeoutError() File C:\Python\lib\concurrent\futures\_base.py:403, in Future.__get_result(self) 401 if self._exception: 402 try: --> 403 raise self._exception 404 finally: 405 # Break a reference cycle with the exception in self._exception 406 self = None ImportError: DLL load failed while importing _pairwise_distances_reduction: The specified module could not be found. ``` ### SciPy/NumPy/Python version information 1.9.2 1.23.4 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0)
defect
bug scipy causes crash of scikit learn joblib describe your issue current scipy on win causes fail of joblib subprocess using scikit learn tested in windows container not replicated on linux and macos even with the same settings previous bugfix scipy version works reproducing code example python install at first these packages python m pip install u pip setuptools scipy joblib sklearn then run this code e g interactively from joblib import parallel delayed import sklearn def a from sklearn model selection import cross val score return cross val score data results parallel n jobs delayed a for i in range the last line fails above to fix just do e g below and rerun python m pip install u pip scipy error message shell in data results parallel n jobs delayed a for i in range remotetraceback traceback most recent call last remotetraceback traceback most recent call last file c python lib site packages joblib externals loky process executor py line in process worker r call item file c python lib site packages joblib externals loky process executor py line in call return self fn self args self kwargs file c python lib site packages joblib parallel backends py line in call return self func args kwargs file c python lib site packages joblib parallel py line in call return func args kwargs file c python lib site packages joblib parallel py line in return func args kwargs file line in a file c python lib site packages sklearn model selection init py line in from validation import cross val score file c python lib site packages sklearn model selection validation py line in from metrics import check scoring file c python lib site packages sklearn metrics init py line in from import cluster file c python lib site packages sklearn metrics cluster init py line in from unsupervised import silhouette samples file c python lib site packages sklearn metrics cluster unsupervised py line in from pairwise import pairwise distances chunked file c python lib site packages sklearn metrics pairwise py line in from pairwise distances reduction import pairwisedistancesargkmin importerror dll load failed while importing pairwise distances reduction the specified module could not be found the above exception was the direct cause of the following exception importerror traceback most recent call last cell in line data results parallel n jobs delayed a for i in range file c python lib site packages joblib parallel py in parallel call self iterable self iterating false with self backend retrieval context self retrieve make sure that we get a last message telling us we are done elapsed time time time self start time file c python lib site packages joblib parallel py in parallel retrieve self try if getattr self backend supports timeout false self output extend job get timeout self timeout else self output extend job get file c python lib site packages joblib parallel backends py in lokybackend wrap future result future timeout wrapper for future result to implement the same behaviour as asyncresults get from multiprocessing try return future result timeout timeout except cftimeouterror as e raise timeouterror from e file c python lib concurrent futures base py in future result self timeout raise cancellederror elif self state finished return self get result else raise timeouterror file c python lib concurrent futures base py in future get result self if self exception try raise self exception finally break a reference cycle with the exception in self exception self none importerror dll load failed while importing pairwise distances reduction the specified module could not be found scipy numpy python version information sys version info major minor micro releaselevel final serial
1
19,896
3,273,629,696
IssuesEvent
2015-10-26 04:25:36
npgall/cqengine
https://api.github.com/repos/npgall/cqengine
closed
A small bug in uniqueResult() + resolution
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Try a query that returns a ResultSet with a single item 2. call uniqueResult() on that query 3. an exception is thrown saying that the ResultSet has multiple items What is the expected output? What do you see instead? The single item should be returned What version of the product are you using? On what operating system? latest from svn, XP Please provide any additional information below. in ResultSet.java, line 58 should be: O result = iterator.next(); instead of: O result = iterator().next(); ``` Original issue reported on code.google.com by `gil.assa...@gmail.com` on 29 Oct 2012 at 1:15
1.0
A small bug in uniqueResult() + resolution - ``` What steps will reproduce the problem? 1. Try a query that returns a ResultSet with a single item 2. call uniqueResult() on that query 3. an exception is thrown saying that the ResultSet has multiple items What is the expected output? What do you see instead? The single item should be returned What version of the product are you using? On what operating system? latest from svn, XP Please provide any additional information below. in ResultSet.java, line 58 should be: O result = iterator.next(); instead of: O result = iterator().next(); ``` Original issue reported on code.google.com by `gil.assa...@gmail.com` on 29 Oct 2012 at 1:15
defect
a small bug in uniqueresult resolution what steps will reproduce the problem try a query that returns a resultset with a single item call uniqueresult on that query an exception is thrown saying that the resultset has multiple items what is the expected output what do you see instead the single item should be returned what version of the product are you using on what operating system latest from svn xp please provide any additional information below in resultset java line should be o result iterator next instead of o result iterator next original issue reported on code google com by gil assa gmail com on oct at
1
234,647
18,011,570,030
IssuesEvent
2021-09-16 09:12:14
nvim-telescope/telescope.nvim
https://api.github.com/repos/nvim-telescope/telescope.nvim
closed
Make "ripgrep" a necessary dependency or highlight it's value
documentation enhancement Planned
Hello! I am working as computer vision engineer, and I have multiple datasets with thousands of images in my project directory. I have installed Telescope and was completely frustrated by the speed of searching. Telescope tries to index all directories. Then I started to read relevant issues, documentation, even source code (despite I don't know lua). What I have found in my case: 1) default parameter "file_ignore_patterns" with command like this in init.vim > :lua require('telescope').setup{ defaults = { file_ignore_patterns = {"data/", "datasets/" } } } kind of works, but it just hides these directories from output while continue indexing. 2) "find_files" just ignores .gitignore, but seems like it should use this file. 3) "live_grep" does not work at all. 4) maps like > nnoremap \<leader>ff :lua require("telescope.builtin").find_files{ find_command = { "rg", "--files", "--hidden", "--follow"} }<cr> Shows nothing when I go with \<leader>ff (I have set up leader key properly) So I have found that I need to install ripgrep to use "rg" in find_command (I just did know nothing about ripgrep and what is "rg", I thought it's a some inner-Telescope command), and at the same time all my issues were solved. But it's still unclear for me why it is an optional dependency when it gives such a huge improvement to user experience with Telescope and even allows to use live_grep. My suggestion is to either force users to install ripgrep somehow, or at least emphasize this moment in README or somewhere else. By the way, thanks for awesome tool!
1.0
Make "ripgrep" a necessary dependency or highlight it's value - Hello! I am working as computer vision engineer, and I have multiple datasets with thousands of images in my project directory. I have installed Telescope and was completely frustrated by the speed of searching. Telescope tries to index all directories. Then I started to read relevant issues, documentation, even source code (despite I don't know lua). What I have found in my case: 1) default parameter "file_ignore_patterns" with command like this in init.vim > :lua require('telescope').setup{ defaults = { file_ignore_patterns = {"data/", "datasets/" } } } kind of works, but it just hides these directories from output while continue indexing. 2) "find_files" just ignores .gitignore, but seems like it should use this file. 3) "live_grep" does not work at all. 4) maps like > nnoremap \<leader>ff :lua require("telescope.builtin").find_files{ find_command = { "rg", "--files", "--hidden", "--follow"} }<cr> Shows nothing when I go with \<leader>ff (I have set up leader key properly) So I have found that I need to install ripgrep to use "rg" in find_command (I just did know nothing about ripgrep and what is "rg", I thought it's a some inner-Telescope command), and at the same time all my issues were solved. But it's still unclear for me why it is an optional dependency when it gives such a huge improvement to user experience with Telescope and even allows to use live_grep. My suggestion is to either force users to install ripgrep somehow, or at least emphasize this moment in README or somewhere else. By the way, thanks for awesome tool!
non_defect
make ripgrep a necessary dependency or highlight it s value hello i am working as computer vision engineer and i have multiple datasets with thousands of images in my project directory i have installed telescope and was completely frustrated by the speed of searching telescope tries to index all directories then i started to read relevant issues documentation even source code despite i don t know lua what i have found in my case default parameter file ignore patterns with command like this in init vim lua require telescope setup defaults file ignore patterns data datasets kind of works but it just hides these directories from output while continue indexing find files just ignores gitignore but seems like it should use this file live grep does not work at all maps like nnoremap ff lua require telescope builtin find files find command rg files hidden follow shows nothing when i go with ff i have set up leader key properly so i have found that i need to install ripgrep to use rg in find command i just did know nothing about ripgrep and what is rg i thought it s a some inner telescope command and at the same time all my issues were solved but it s still unclear for me why it is an optional dependency when it gives such a huge improvement to user experience with telescope and even allows to use live grep my suggestion is to either force users to install ripgrep somehow or at least emphasize this moment in readme or somewhere else by the way thanks for awesome tool
0
74,727
25,283,593,411
IssuesEvent
2022-11-16 17:27:55
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
zfs should build without "---without-python or --disable-pyzfs"
Type: Defect
try to build an "ARCH iso" how ever this is zfs-2.1.4 mkdksm conf has been to be patched: ==> dkms install --no-depmod zfs/2.1.4 -k 5.17.5-arch1-1 Deprecated feature: REMAKE_INITRD configure: error: "Python >= 3.4 is required" patch has been changed to (but still no luck): --- scripts/dkms.mkconf.bk 2022-04-30 17:38:29.376581217 +0200 +++ scripts/dkms.mkconf 2022-04-30 17:39:16.753437330 +0200 @@ -28,14 +28,9 @@ PRE_BUILD="configure --prefix=/usr --with-config=kernel - --with-linux=\$( - if [ -e "\${kernel_source_dir/%build/source}" ] - then - echo "\${kernel_source_dir/%build/source}" - else - echo "\${kernel_source_dir}" - fi - ) + --without-python --disable-pyzfs + --disable-pyzfs + --with-linux="\${kernel_source_dir}" --with-linux-obj="\${kernel_source_dir}" \$( [[ -n \"\${ICP_ROOT}\" ]] && \\ @@ -69,7 +63,7 @@ " AUTOINSTALL="yes" REMAKE_INITRD="no" -MAKE[0]="make" +MAKE[0]="make -C module/" STRIP[0]="\$( [[ -r \${PACKAGE_CONFIG} ]] \\ && source \${PACKAGE_CONFIG} \\
1.0
zfs should build without "---without-python or --disable-pyzfs" - try to build an "ARCH iso" how ever this is zfs-2.1.4 mkdksm conf has been to be patched: ==> dkms install --no-depmod zfs/2.1.4 -k 5.17.5-arch1-1 Deprecated feature: REMAKE_INITRD configure: error: "Python >= 3.4 is required" patch has been changed to (but still no luck): --- scripts/dkms.mkconf.bk 2022-04-30 17:38:29.376581217 +0200 +++ scripts/dkms.mkconf 2022-04-30 17:39:16.753437330 +0200 @@ -28,14 +28,9 @@ PRE_BUILD="configure --prefix=/usr --with-config=kernel - --with-linux=\$( - if [ -e "\${kernel_source_dir/%build/source}" ] - then - echo "\${kernel_source_dir/%build/source}" - else - echo "\${kernel_source_dir}" - fi - ) + --without-python --disable-pyzfs + --disable-pyzfs + --with-linux="\${kernel_source_dir}" --with-linux-obj="\${kernel_source_dir}" \$( [[ -n \"\${ICP_ROOT}\" ]] && \\ @@ -69,7 +63,7 @@ " AUTOINSTALL="yes" REMAKE_INITRD="no" -MAKE[0]="make" +MAKE[0]="make -C module/" STRIP[0]="\$( [[ -r \${PACKAGE_CONFIG} ]] \\ && source \${PACKAGE_CONFIG} \\
defect
zfs should build without without python or disable pyzfs try to build an arch iso how ever this is zfs mkdksm conf has been to be patched dkms install no depmod zfs k deprecated feature remake initrd configure error python is required patch has been changed to but still no luck scripts dkms mkconf bk scripts dkms mkconf pre build configure prefix usr with config kernel with linux if then echo kernel source dir build source else echo kernel source dir fi without python disable pyzfs disable pyzfs with linux kernel source dir with linux obj kernel source dir autoinstall yes remake initrd no make make make make c module strip source package config
1
33,952
7,310,556,358
IssuesEvent
2018-02-28 15:19:40
TV-Rename/tvrename
https://api.github.com/repos/TV-Rename/tvrename
closed
Application exit code (command line)
Priority-Medium Retest needed Type-Defect auto-migrated
``` Have the application exit code be != 0 on error. ``` Original issue reported on code.google.com by `tvren...@tvrename.com` on 10 Sep 2009 at 12:59
1.0
Application exit code (command line) - ``` Have the application exit code be != 0 on error. ``` Original issue reported on code.google.com by `tvren...@tvrename.com` on 10 Sep 2009 at 12:59
defect
application exit code command line have the application exit code be on error original issue reported on code google com by tvren tvrename com on sep at
1
51,753
13,211,301,206
IssuesEvent
2020-08-15 22:09:17
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
Coordinate services documentation is Unhelpful (Trac #977)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/977">https://code.icecube.wisc.edu/projects/icecube/ticket/977</a>, reported by icecube</summary> <p> ```json { "status": "closed", "changetime": "2015-06-26T21:13:56", "_ts": "1435353236861155", "description": "Most Astronomers use a coordinate system called J2000 to locate the position of astronomical objects which where the pole is the celestial pole at January 1st 2000. However sometimes a coordinate system called current is used where the pole is wherever the hell the celestial pole it is at the time of observation. Both of these cooridinate systems are refered to as equatorial cooridinates. as the difference is at most 18 arcseconds for time scales on the order of icecubes's lifetime. \n\ncoordinate-services handles these two cooridinate systems in a rather confusing way: to get from equatoraial to local cooridinates it provides two sets of functions:\nLocal2RA(), Local2Dec() and Local2RA_no(), Local2Dec_no()\n\nas well as two sets of functions for the reverse transforms:\nEqua2LocalAzimuth, Equa2LocalZenith() and Equa2LocalAzimuth_inv(), Equal2LocalZenith_inv()\n\nIt is completly unclear weather one should use the _no() functions or not and the _inv() functions or not, the function's doc strings only seem to add to the confusion\n\nIn addition there are the totally confusing Equa2LocalRA and Equal2LocalDec which claim to take an RA and Dec and convert them to another RA and Dec, such a function should not have local in its name.\n\nfurthermore all of these functions have epoch variables which is even more confusing, why would current coordinates need a local function. as well you need to be reading a pretty old paper to find anything in non J2000 coordinates. so epoch should never be anything but 2000 ( it looks like that is the default from the code, but you can't tell from the documentation)\n\nWhat is needed is to figure out which functions are are for J2000 and which ones are for current. and properly document them. Then depreciate the old functions and replace them with functions with the string \"J2000\" and \"Current\". none of which should have an epoch variable. ( If a new epoch becomes popular J2050 for example then new functions can be added )\n\nPython examples should also be provided\n", "reporter": "icecube", "cc": "", "resolution": "wontfix", "time": "2015-05-13T16:30:51", "component": "combo reconstruction", "summary": "Coordinate services documentation is Unhelpful", "priority": "normal", "keywords": "", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
Coordinate services documentation is Unhelpful (Trac #977) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/977">https://code.icecube.wisc.edu/projects/icecube/ticket/977</a>, reported by icecube</summary> <p> ```json { "status": "closed", "changetime": "2015-06-26T21:13:56", "_ts": "1435353236861155", "description": "Most Astronomers use a coordinate system called J2000 to locate the position of astronomical objects which where the pole is the celestial pole at January 1st 2000. However sometimes a coordinate system called current is used where the pole is wherever the hell the celestial pole it is at the time of observation. Both of these cooridinate systems are refered to as equatorial cooridinates. as the difference is at most 18 arcseconds for time scales on the order of icecubes's lifetime. \n\ncoordinate-services handles these two cooridinate systems in a rather confusing way: to get from equatoraial to local cooridinates it provides two sets of functions:\nLocal2RA(), Local2Dec() and Local2RA_no(), Local2Dec_no()\n\nas well as two sets of functions for the reverse transforms:\nEqua2LocalAzimuth, Equa2LocalZenith() and Equa2LocalAzimuth_inv(), Equal2LocalZenith_inv()\n\nIt is completly unclear weather one should use the _no() functions or not and the _inv() functions or not, the function's doc strings only seem to add to the confusion\n\nIn addition there are the totally confusing Equa2LocalRA and Equal2LocalDec which claim to take an RA and Dec and convert them to another RA and Dec, such a function should not have local in its name.\n\nfurthermore all of these functions have epoch variables which is even more confusing, why would current coordinates need a local function. as well you need to be reading a pretty old paper to find anything in non J2000 coordinates. so epoch should never be anything but 2000 ( it looks like that is the default from the code, but you can't tell from the documentation)\n\nWhat is needed is to figure out which functions are are for J2000 and which ones are for current. and properly document them. Then depreciate the old functions and replace them with functions with the string \"J2000\" and \"Current\". none of which should have an epoch variable. ( If a new epoch becomes popular J2050 for example then new functions can be added )\n\nPython examples should also be provided\n", "reporter": "icecube", "cc": "", "resolution": "wontfix", "time": "2015-05-13T16:30:51", "component": "combo reconstruction", "summary": "Coordinate services documentation is Unhelpful", "priority": "normal", "keywords": "", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
defect
coordinate services documentation is unhelpful trac migrated from json status closed changetime ts description most astronomers use a coordinate system called to locate the position of astronomical objects which where the pole is the celestial pole at january however sometimes a coordinate system called current is used where the pole is wherever the hell the celestial pole it is at the time of observation both of these cooridinate systems are refered to as equatorial cooridinates as the difference is at most arcseconds for time scales on the order of icecubes s lifetime n ncoordinate services handles these two cooridinate systems in a rather confusing way to get from equatoraial to local cooridinates it provides two sets of functions and no no n nas well as two sets of functions for the reverse transforms and inv inv n nit is completly unclear weather one should use the no functions or not and the inv functions or not the function s doc strings only seem to add to the confusion n nin addition there are the totally confusing and which claim to take an ra and dec and convert them to another ra and dec such a function should not have local in its name n nfurthermore all of these functions have epoch variables which is even more confusing why would current coordinates need a local function as well you need to be reading a pretty old paper to find anything in non coordinates so epoch should never be anything but it looks like that is the default from the code but you can t tell from the documentation n nwhat is needed is to figure out which functions are are for and which ones are for current and properly document them then depreciate the old functions and replace them with functions with the string and current none of which should have an epoch variable if a new epoch becomes popular for example then new functions can be added n npython examples should also be provided n reporter icecube cc resolution wontfix time component combo reconstruction summary coordinate services documentation is unhelpful priority normal keywords milestone owner type defect
1
461,715
13,234,829,730
IssuesEvent
2020-08-18 16:57:49
Adam-Poppenheimer/Civ-Clone
https://api.github.com/repos/Adam-Poppenheimer/Civ-Clone
opened
Map saving/loading doesn't save the round number
priority: desirable type: bug
I can't see any place where MapComposer serializes the current round , nor can I see a way to set the current round on GameCore manually. But saved games should definitely record this information. Reason for score: It would be fairly trivial to add a CurrentRound field onto SerializableMapData, but if I want to record other types of data not associated with any of the existing subsystems, I might consider adding a CoreComposer class and appropriate calls to it, which'll take a bit longer.
1.0
Map saving/loading doesn't save the round number - I can't see any place where MapComposer serializes the current round , nor can I see a way to set the current round on GameCore manually. But saved games should definitely record this information. Reason for score: It would be fairly trivial to add a CurrentRound field onto SerializableMapData, but if I want to record other types of data not associated with any of the existing subsystems, I might consider adding a CoreComposer class and appropriate calls to it, which'll take a bit longer.
non_defect
map saving loading doesn t save the round number i can t see any place where mapcomposer serializes the current round nor can i see a way to set the current round on gamecore manually but saved games should definitely record this information reason for score it would be fairly trivial to add a currentround field onto serializablemapdata but if i want to record other types of data not associated with any of the existing subsystems i might consider adding a corecomposer class and appropriate calls to it which ll take a bit longer
0
393,750
11,624,224,857
IssuesEvent
2020-02-27 10:22:59
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
Message in alert box isn't clear - disputing form
Needed for V2 launch Priority: Medium
Build: ![image.png](https://images.zenhubusercontent.com/5c07af47035895308911c4c5/6bf1ddf4-8f0b-4512-8e27-2de3491b5e39) 'Tentative Winning outcome has 585.9384 REP already staked for next round. More REP will be needed to make this outcome the Tentative Winner - this' sounds confusing, is 'More REP will be needed to make this outcome the Tentative Winner' referring to the tentative outcome or the selected outcome. The message in the design was: ![image.png](https://images.zenhubusercontent.com/5c07af47035895308911c4c5/ebcc53bd-52f6-44f2-a8a6-c017fcbb9328) ### Filling this bond of 4.0000 REP only completes this current round. 200.0000 additional REP will still be needed to make it the Tentative Outcome. This will require an additional transaction.
1.0
Message in alert box isn't clear - disputing form - Build: ![image.png](https://images.zenhubusercontent.com/5c07af47035895308911c4c5/6bf1ddf4-8f0b-4512-8e27-2de3491b5e39) 'Tentative Winning outcome has 585.9384 REP already staked for next round. More REP will be needed to make this outcome the Tentative Winner - this' sounds confusing, is 'More REP will be needed to make this outcome the Tentative Winner' referring to the tentative outcome or the selected outcome. The message in the design was: ![image.png](https://images.zenhubusercontent.com/5c07af47035895308911c4c5/ebcc53bd-52f6-44f2-a8a6-c017fcbb9328) ### Filling this bond of 4.0000 REP only completes this current round. 200.0000 additional REP will still be needed to make it the Tentative Outcome. This will require an additional transaction.
non_defect
message in alert box isn t clear disputing form build tentative winning outcome has rep already staked for next round more rep will be needed to make this outcome the tentative winner this sounds confusing is more rep will be needed to make this outcome the tentative winner referring to the tentative outcome or the selected outcome the message in the design was filling this bond of rep only completes this current round additional rep will still be needed to make it the tentative outcome this will require an additional transaction
0
50,173
13,187,361,614
IssuesEvent
2020-08-13 03:10:16
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
Nuke TWR (Trac #257)
Migrated from Trac combo core defect
Kill it. Kill it with fire. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/257 , reported by nega and owned by nwhitehorn_</summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:22:56", "description": "Kill it. Kill it with fire.", "reporter": "nega", "cc": "blaufuss, olivas", "resolution": "fixed", "_ts": "1547241776942948", "component": "combo core", "summary": "Nuke TWR", "priority": "normal", "keywords": "twr", "time": "2011-05-11T20:21:18", "milestone": "", "owner": "nwhitehorn", "type": "defect" } ``` </p> </details>
1.0
Nuke TWR (Trac #257) - Kill it. Kill it with fire. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/257 , reported by nega and owned by nwhitehorn_</summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T21:22:56", "description": "Kill it. Kill it with fire.", "reporter": "nega", "cc": "blaufuss, olivas", "resolution": "fixed", "_ts": "1547241776942948", "component": "combo core", "summary": "Nuke TWR", "priority": "normal", "keywords": "twr", "time": "2011-05-11T20:21:18", "milestone": "", "owner": "nwhitehorn", "type": "defect" } ``` </p> </details>
defect
nuke twr trac kill it kill it with fire migrated from reported by nega and owned by nwhitehorn json status closed changetime description kill it kill it with fire reporter nega cc blaufuss olivas resolution fixed ts component combo core summary nuke twr priority normal keywords twr time milestone owner nwhitehorn type defect
1
8,281
2,611,486,073
IssuesEvent
2015-02-27 05:27:23
chrsmith/switchlist
https://api.github.com/repos/chrsmith/switchlist
closed
Hovering for tool tips should work in tables.
auto-migrated Priority-Medium Type-Defect
``` Hovering for tool tips (for details on cargo, etc) should work on both buttons and widgets in the bottom of each window as well as on the various tables. To reproduce: Go to the freight car window, and hover over a cargo name in the table. Nothing happens. Now hover over the cargo button towards the bottom of the screen; you get a helpful tool tip. This would help usability (especially with cargos); the only question is whether an NSTableView allows a tool tip on individual items in the table. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 21 Dec 2011 at 4:12
1.0
Hovering for tool tips should work in tables. - ``` Hovering for tool tips (for details on cargo, etc) should work on both buttons and widgets in the bottom of each window as well as on the various tables. To reproduce: Go to the freight car window, and hover over a cargo name in the table. Nothing happens. Now hover over the cargo button towards the bottom of the screen; you get a helpful tool tip. This would help usability (especially with cargos); the only question is whether an NSTableView allows a tool tip on individual items in the table. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 21 Dec 2011 at 4:12
defect
hovering for tool tips should work in tables hovering for tool tips for details on cargo etc should work on both buttons and widgets in the bottom of each window as well as on the various tables to reproduce go to the freight car window and hover over a cargo name in the table nothing happens now hover over the cargo button towards the bottom of the screen you get a helpful tool tip this would help usability especially with cargos the only question is whether an nstableview allows a tool tip on individual items in the table original issue reported on code google com by rwbowdi gmail com on dec at
1
525,242
15,241,614,878
IssuesEvent
2021-02-19 08:42:47
AY2021S2-CS2103T-W10-1/tp
https://api.github.com/repos/AY2021S2-CS2103T-W10-1/tp
opened
Update Developer Guide
priority.High severity.VeryLow type.Task
Update - target user profile, value proposition, and user stories - use cases - non-functional requirements
1.0
Update Developer Guide - Update - target user profile, value proposition, and user stories - use cases - non-functional requirements
non_defect
update developer guide update target user profile value proposition and user stories use cases non functional requirements
0
608,884
18,850,483,341
IssuesEvent
2021-11-11 20:09:35
salvadorbs/Asuite
https://api.github.com/repos/salvadorbs/Asuite
closed
ClassicMenu issues with left/right click on KDE trayicon
Status: Confirmed Type: Bug Priority: Low LCL
After opening the ClassicMenu with the left click, there are random errors or worse ASuite crashes. The problem lies within the LCL library in QT5 (in KDE distro based). A bug report has been opened in Lazarus Issues https://bugs.freepascal.org/view.php?id=38849
1.0
ClassicMenu issues with left/right click on KDE trayicon - After opening the ClassicMenu with the left click, there are random errors or worse ASuite crashes. The problem lies within the LCL library in QT5 (in KDE distro based). A bug report has been opened in Lazarus Issues https://bugs.freepascal.org/view.php?id=38849
non_defect
classicmenu issues with left right click on kde trayicon after opening the classicmenu with the left click there are random errors or worse asuite crashes the problem lies within the lcl library in in kde distro based a bug report has been opened in lazarus issues
0
305,541
26,391,923,646
IssuesEvent
2023-01-12 16:16:44
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix reduction_ops.test_torch_nanmean
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_nanmean[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-10T03:53:37.0203335Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-10T03:53:37.0204325Z E Falsifying example: test_torch_nanmean( 2023-01-10T03:53:37.0204912Z E dtype_and_x=(['float64'], [array([-1., -1.])], 0), 2023-01-10T03:53:37.0205306Z E keepdims=False, 2023-01-10T03:53:37.0205637Z E with_out=True, 2023-01-10T03:53:37.0205990Z E num_positional_args=0, 2023-01-10T03:53:37.0206359Z E native_array=[False], 2023-01-10T03:53:37.0206723Z E as_variable=[False], 2023-01-10T03:53:37.0207271Z E fn_tree='ivy.functional.frontends.torch.nanmean', 2023-01-10T03:53:37.0207757Z E on_device='cpu', 2023-01-10T03:53:37.0208151Z E frontend='torch', 2023-01-10T03:53:37.0208476Z E ) 2023-01-10T03:53:37.0208766Z E 2023-01-10T03:53:37.0209493Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkAAMoBaYx2AAAmgAH') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_nanmean[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-10T03:53:37.0203335Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-10T03:53:37.0204325Z E Falsifying example: test_torch_nanmean( 2023-01-10T03:53:37.0204912Z E dtype_and_x=(['float64'], [array([-1., -1.])], 0), 2023-01-10T03:53:37.0205306Z E keepdims=False, 2023-01-10T03:53:37.0205637Z E with_out=True, 2023-01-10T03:53:37.0205990Z E num_positional_args=0, 2023-01-10T03:53:37.0206359Z E native_array=[False], 2023-01-10T03:53:37.0206723Z E as_variable=[False], 2023-01-10T03:53:37.0207271Z E fn_tree='ivy.functional.frontends.torch.nanmean', 2023-01-10T03:53:37.0207757Z E on_device='cpu', 2023-01-10T03:53:37.0208151Z E frontend='torch', 2023-01-10T03:53:37.0208476Z E ) 2023-01-10T03:53:37.0208766Z E 2023-01-10T03:53:37.0209493Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkAAMoBaYx2AAAmgAH') as a decorator on your test case </details>
1.0
Fix reduction_ops.test_torch_nanmean - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3879954723/jobs/6617597240" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_nanmean[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-10T03:53:37.0203335Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-10T03:53:37.0204325Z E Falsifying example: test_torch_nanmean( 2023-01-10T03:53:37.0204912Z E dtype_and_x=(['float64'], [array([-1., -1.])], 0), 2023-01-10T03:53:37.0205306Z E keepdims=False, 2023-01-10T03:53:37.0205637Z E with_out=True, 2023-01-10T03:53:37.0205990Z E num_positional_args=0, 2023-01-10T03:53:37.0206359Z E native_array=[False], 2023-01-10T03:53:37.0206723Z E as_variable=[False], 2023-01-10T03:53:37.0207271Z E fn_tree='ivy.functional.frontends.torch.nanmean', 2023-01-10T03:53:37.0207757Z E on_device='cpu', 2023-01-10T03:53:37.0208151Z E frontend='torch', 2023-01-10T03:53:37.0208476Z E ) 2023-01-10T03:53:37.0208766Z E 2023-01-10T03:53:37.0209493Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkAAMoBaYx2AAAmgAH') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_nanmean[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-10T03:53:37.0203335Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-10T03:53:37.0204325Z E Falsifying example: test_torch_nanmean( 2023-01-10T03:53:37.0204912Z E dtype_and_x=(['float64'], [array([-1., -1.])], 0), 2023-01-10T03:53:37.0205306Z E keepdims=False, 2023-01-10T03:53:37.0205637Z E with_out=True, 2023-01-10T03:53:37.0205990Z E num_positional_args=0, 2023-01-10T03:53:37.0206359Z E native_array=[False], 2023-01-10T03:53:37.0206723Z E as_variable=[False], 2023-01-10T03:53:37.0207271Z E fn_tree='ivy.functional.frontends.torch.nanmean', 2023-01-10T03:53:37.0207757Z E on_device='cpu', 2023-01-10T03:53:37.0208151Z E frontend='torch', 2023-01-10T03:53:37.0208476Z E ) 2023-01-10T03:53:37.0208766Z E 2023-01-10T03:53:37.0209493Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkAAMoBaYx2AAAmgAH') as a decorator on your test case </details>
non_defect
fix reduction ops test torch nanmean tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test reduction ops py test torch nanmean e assertionerror the return with a torch backend produced data type of while the return with a numpy backend returned a data type of e falsifying example test torch nanmean e dtype and x e keepdims false e with out true e num positional args e native array e as variable e fn tree ivy functional frontends torch nanmean e on device cpu e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test reduction ops py test torch nanmean e assertionerror the return with a torch backend produced data type of while the return with a numpy backend returned a data type of e falsifying example test torch nanmean e dtype and x e keepdims false e with out true e num positional args e native array e as variable e fn tree ivy functional frontends torch nanmean e on device cpu e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
0
683,642
23,389,787,317
IssuesEvent
2022-08-11 16:40:27
School-Simplified/Timmy-SchoolSimplified
https://api.github.com/repos/School-Simplified/Timmy-SchoolSimplified
closed
Fix `on_command_error` response.
bug Priority: Medium
Make it more user friendly and fix it sending 2 embeds when multiple (stacked) tracebacks occur.
1.0
Fix `on_command_error` response. - Make it more user friendly and fix it sending 2 embeds when multiple (stacked) tracebacks occur.
non_defect
fix on command error response make it more user friendly and fix it sending embeds when multiple stacked tracebacks occur
0
53,432
13,261,601,392
IssuesEvent
2020-08-20 20:11:48
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
segfault when starting steamshovel (Trac #1378)
Migrated from Trac combo core defect
A strange problem: If I start the environment with ./env-shell.sh, then run steamshovel, it segfaults right at the beginning. With "./env-shell.sh steamshovel file.i3" it starts up fine. System: Ubuntu 15.04 icerec version: trunk, Revision: 138186 backtrace: (gdb) run selfveto-results.i3 Starting program: /home/berghaus/events/steamshovel selfveto-results.i3 [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". ERROR (steamshovel): Problem loading icecube python bindings: ImportError: No module named icecube.dataio Are there unbuilt pybindings? (embed.cpp:64 in scripting::PyInterpreter::PyInterpreter(char*)) terminate called after throwing an instance of 'boost::python::error_already_set' Program received signal SIGABRT, Aborted. 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55 55 ../sysdeps/unix/sysv/linux/raise.c: No such file or directory. (gdb) bt https://code.icecube.wisc.edu/ticket/0 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55 #1 0x00007fffeaefeeca in __GI_abort () at abort.c:89 #2 0x00007ffff792706d in __gnu_cxx::__verbose_terminate_handler() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #3 0x00007ffff7924ee6 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007ffff7924f31 in std::terminate() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007ffff7925199 in __cxa_rethrow () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00000000004e23a2 in scripting::PyInterpreter::PyInterpreter(char*) () #7 0x0000000000473fbd in main () <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1378">https://code.icecube.wisc.edu/projects/icecube/ticket/1378</a>, reported by berghausand owned by hdembinski</em></summary> <p> ```json { "status": "closed", "changetime": "2015-10-05T15:28:27", "_ts": "1444058907340211", "description": "A strange problem:\nIf I start the environment with ./env-shell.sh, then run steamshovel, it segfaults right at the beginning. With \"./env-shell.sh steamshovel file.i3\" it starts up fine.\nSystem: Ubuntu 15.04\nicerec version: trunk, Revision: 138186\n\nbacktrace:\n\n(gdb) run selfveto-results.i3\nStarting program: /home/berghaus/events/steamshovel selfveto-results.i3\n[Thread debugging using libthread_db enabled]\nUsing host libthread_db library \"/lib/x86_64-linux-gnu/libthread_db.so.1\".\nERROR (steamshovel): Problem loading icecube python bindings:\nImportError: No module named icecube.dataio\nAre there unbuilt pybindings? (embed.cpp:64 in scripting::PyInterpreter::PyInterpreter(char*))\nterminate called after throwing an instance of 'boost::python::error_already_set'\n\nProgram received signal SIGABRT, Aborted.\n0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55\n55\t../sysdeps/unix/sysv/linux/raise.c: No such file or directory.\n(gdb) bt\n#0 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55\n#1 0x00007fffeaefeeca in __GI_abort () at abort.c:89\n#2 0x00007ffff792706d in __gnu_cxx::__verbose_terminate_handler() ()\n from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#3 0x00007ffff7924ee6 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#4 0x00007ffff7924f31 in std::terminate() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#5 0x00007ffff7925199 in __cxa_rethrow () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#6 0x00000000004e23a2 in scripting::PyInterpreter::PyInterpreter(char*) ()\n#7 0x0000000000473fbd in main ()\n", "reporter": "berghaus", "cc": "", "resolution": "invalid", "time": "2015-10-05T07:29:14", "component": "combo core", "summary": "segfault when starting steamshovel", "priority": "normal", "keywords": "steamshovel", "milestone": "", "owner": "hdembinski", "type": "defect" } ``` </p> </details>
1.0
segfault when starting steamshovel (Trac #1378) - A strange problem: If I start the environment with ./env-shell.sh, then run steamshovel, it segfaults right at the beginning. With "./env-shell.sh steamshovel file.i3" it starts up fine. System: Ubuntu 15.04 icerec version: trunk, Revision: 138186 backtrace: (gdb) run selfveto-results.i3 Starting program: /home/berghaus/events/steamshovel selfveto-results.i3 [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". ERROR (steamshovel): Problem loading icecube python bindings: ImportError: No module named icecube.dataio Are there unbuilt pybindings? (embed.cpp:64 in scripting::PyInterpreter::PyInterpreter(char*)) terminate called after throwing an instance of 'boost::python::error_already_set' Program received signal SIGABRT, Aborted. 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55 55 ../sysdeps/unix/sysv/linux/raise.c: No such file or directory. (gdb) bt https://code.icecube.wisc.edu/ticket/0 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55 #1 0x00007fffeaefeeca in __GI_abort () at abort.c:89 #2 0x00007ffff792706d in __gnu_cxx::__verbose_terminate_handler() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #3 0x00007ffff7924ee6 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007ffff7924f31 in std::terminate() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007ffff7925199 in __cxa_rethrow () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00000000004e23a2 in scripting::PyInterpreter::PyInterpreter(char*) () #7 0x0000000000473fbd in main () <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1378">https://code.icecube.wisc.edu/projects/icecube/ticket/1378</a>, reported by berghausand owned by hdembinski</em></summary> <p> ```json { "status": "closed", "changetime": "2015-10-05T15:28:27", "_ts": "1444058907340211", "description": "A strange problem:\nIf I start the environment with ./env-shell.sh, then run steamshovel, it segfaults right at the beginning. With \"./env-shell.sh steamshovel file.i3\" it starts up fine.\nSystem: Ubuntu 15.04\nicerec version: trunk, Revision: 138186\n\nbacktrace:\n\n(gdb) run selfveto-results.i3\nStarting program: /home/berghaus/events/steamshovel selfveto-results.i3\n[Thread debugging using libthread_db enabled]\nUsing host libthread_db library \"/lib/x86_64-linux-gnu/libthread_db.so.1\".\nERROR (steamshovel): Problem loading icecube python bindings:\nImportError: No module named icecube.dataio\nAre there unbuilt pybindings? (embed.cpp:64 in scripting::PyInterpreter::PyInterpreter(char*))\nterminate called after throwing an instance of 'boost::python::error_already_set'\n\nProgram received signal SIGABRT, Aborted.\n0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55\n55\t../sysdeps/unix/sysv/linux/raise.c: No such file or directory.\n(gdb) bt\n#0 0x00007fffeaefd267 in __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:55\n#1 0x00007fffeaefeeca in __GI_abort () at abort.c:89\n#2 0x00007ffff792706d in __gnu_cxx::__verbose_terminate_handler() ()\n from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#3 0x00007ffff7924ee6 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#4 0x00007ffff7924f31 in std::terminate() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#5 0x00007ffff7925199 in __cxa_rethrow () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6\n#6 0x00000000004e23a2 in scripting::PyInterpreter::PyInterpreter(char*) ()\n#7 0x0000000000473fbd in main ()\n", "reporter": "berghaus", "cc": "", "resolution": "invalid", "time": "2015-10-05T07:29:14", "component": "combo core", "summary": "segfault when starting steamshovel", "priority": "normal", "keywords": "steamshovel", "milestone": "", "owner": "hdembinski", "type": "defect" } ``` </p> </details>
defect
segfault when starting steamshovel trac a strange problem if i start the environment with env shell sh then run steamshovel it segfaults right at the beginning with env shell sh steamshovel file it starts up fine system ubuntu icerec version trunk revision backtrace gdb run selfveto results starting program home berghaus events steamshovel selfveto results using host libthread db library lib linux gnu libthread db so error steamshovel problem loading icecube python bindings importerror no module named icecube dataio are there unbuilt pybindings embed cpp in scripting pyinterpreter pyinterpreter char terminate called after throwing an instance of boost python error already set program received signal sigabrt aborted in gi raise sig sig entry at sysdeps unix sysv linux raise c sysdeps unix sysv linux raise c no such file or directory gdb bt in gi raise sig sig entry at sysdeps unix sysv linux raise c in gi abort at abort c in gnu cxx verbose terminate handler from usr lib linux gnu libstdc so in from usr lib linux gnu libstdc so in std terminate from usr lib linux gnu libstdc so in cxa rethrow from usr lib linux gnu libstdc so in scripting pyinterpreter pyinterpreter char in main migrated from json status closed changetime ts description a strange problem nif i start the environment with env shell sh then run steamshovel it segfaults right at the beginning with env shell sh steamshovel file it starts up fine nsystem ubuntu nicerec version trunk revision n nbacktrace n n gdb run selfveto results nstarting program home berghaus events steamshovel selfveto results n nusing host libthread db library lib linux gnu libthread db so nerror steamshovel problem loading icecube python bindings nimporterror no module named icecube dataio nare there unbuilt pybindings embed cpp in scripting pyinterpreter pyinterpreter char nterminate called after throwing an instance of boost python error already set n nprogram received signal sigabrt aborted in gi raise sig sig entry at sysdeps unix sysv linux raise c t sysdeps unix sysv linux raise c no such file or directory n gdb bt n in gi raise sig sig entry at sysdeps unix sysv linux raise c n in gi abort at abort c n in gnu cxx verbose terminate handler n from usr lib linux gnu libstdc so n in from usr lib linux gnu libstdc so n in std terminate from usr lib linux gnu libstdc so n in cxa rethrow from usr lib linux gnu libstdc so n in scripting pyinterpreter pyinterpreter char n in main n reporter berghaus cc resolution invalid time component combo core summary segfault when starting steamshovel priority normal keywords steamshovel milestone owner hdembinski type defect
1
43,778
11,842,920,302
IssuesEvent
2020-03-24 00:34:30
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
closed
Homepage Articles block: allow multiple instances per post
Blocks [Type] Defect [Type] Enhancement
As reported on newspack-blocks repo (https://github.com/Automattic/newspack-blocks/issues/317), adding multiple Homepage Articles (aka Blog Posts) blocks should be possible. This was done deliberately in https://github.com/Automattic/wp-calypso/issues/38404: https://github.com/Automattic/wp-calypso/blob/b6e7b09b1be3c9b8eb9d790755ec8c9adb1cd23f/apps/full-site-editing/full-site-editing-plugin/blog-posts-block/editor.js#L36 This issue is more of a reminder and a place for further discussion. As @kspilarski mentioned, this surfaced in 2666978-zd
1.0
Homepage Articles block: allow multiple instances per post - As reported on newspack-blocks repo (https://github.com/Automattic/newspack-blocks/issues/317), adding multiple Homepage Articles (aka Blog Posts) blocks should be possible. This was done deliberately in https://github.com/Automattic/wp-calypso/issues/38404: https://github.com/Automattic/wp-calypso/blob/b6e7b09b1be3c9b8eb9d790755ec8c9adb1cd23f/apps/full-site-editing/full-site-editing-plugin/blog-posts-block/editor.js#L36 This issue is more of a reminder and a place for further discussion. As @kspilarski mentioned, this surfaced in 2666978-zd
defect
homepage articles block allow multiple instances per post as reported on newspack blocks repo adding multiple homepage articles aka blog posts blocks should be possible this was done deliberately in this issue is more of a reminder and a place for further discussion as kspilarski mentioned this surfaced in zd
1
91,496
18,504,779,529
IssuesEvent
2021-10-19 17:15:30
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
opened
"New front-end component" issue template
🟨 priority: medium ✨ goal: improvement 💻 aspect: code
We should create a new issue template for front-end component creation. This will allow us to standardize the approach taken when documenting new components. This will make it much easier for contributors to implement new components, as they will know exactly what to expect. ## Things to include - Link to Figma file (obtained by using the `Copy/paste as -> Copy link` feature in Figma <img width="575" alt="CleanShot 2021-10-19 at 13 01 26@2x" src="https://user-images.githubusercontent.com/6351754/137957596-82352aeb-501e-4570-85b9-c1f18ef68afe.png"> - Discussion of needed states (i.e, are focus and hover the same? any loading states?) - List probable props and internal component state - Does the component rely on any global data? Should it _set_ any data, emit events, etc.? - Write some template pseudocode for: - **The component internals**: if a button, what the `<template>` block of `Button.vue` looks like - **The implementation**: if a button, something like `<v-button theme="primary" size="lg" @click="doStuff">Button Text</v-button>` - Follow-up questions for @panchovm if there's any states you see as missing or other issues. ## Things not to include - Screenshots (as they might fall out of date) - Typed out implementation details (i.e, "the surrounding padding should be the "small" value in the design system) as those values might change too.
1.0
"New front-end component" issue template - We should create a new issue template for front-end component creation. This will allow us to standardize the approach taken when documenting new components. This will make it much easier for contributors to implement new components, as they will know exactly what to expect. ## Things to include - Link to Figma file (obtained by using the `Copy/paste as -> Copy link` feature in Figma <img width="575" alt="CleanShot 2021-10-19 at 13 01 26@2x" src="https://user-images.githubusercontent.com/6351754/137957596-82352aeb-501e-4570-85b9-c1f18ef68afe.png"> - Discussion of needed states (i.e, are focus and hover the same? any loading states?) - List probable props and internal component state - Does the component rely on any global data? Should it _set_ any data, emit events, etc.? - Write some template pseudocode for: - **The component internals**: if a button, what the `<template>` block of `Button.vue` looks like - **The implementation**: if a button, something like `<v-button theme="primary" size="lg" @click="doStuff">Button Text</v-button>` - Follow-up questions for @panchovm if there's any states you see as missing or other issues. ## Things not to include - Screenshots (as they might fall out of date) - Typed out implementation details (i.e, "the surrounding padding should be the "small" value in the design system) as those values might change too.
non_defect
new front end component issue template we should create a new issue template for front end component creation this will allow us to standardize the approach taken when documenting new components this will make it much easier for contributors to implement new components as they will know exactly what to expect things to include link to figma file obtained by using the copy paste as copy link feature in figma img width alt cleanshot at src discussion of needed states i e are focus and hover the same any loading states list probable props and internal component state does the component rely on any global data should it set any data emit events etc write some template pseudocode for the component internals if a button what the block of button vue looks like the implementation if a button something like button text follow up questions for panchovm if there s any states you see as missing or other issues things not to include screenshots as they might fall out of date typed out implementation details i e the surrounding padding should be the small value in the design system as those values might change too
0
5,572
8,051,196,019
IssuesEvent
2018-08-01 15:26:02
reactstrap/reactstrap
https://api.github.com/repos/reactstrap/reactstrap
closed
UncontrolledTooltips do not position correctly if the wrapper has an overflow: auto
BS compatibility good first issue help wanted
<!-- The following is part of our issue template, feel free to remove this if it doesn't apply --> ### Issue description - components: `UncontrolledTooltip` - reactstrap version `^6.1.0` - react version `^16.4.0` - bootstrap version `^4.1.1` ### What is happening? The title says it all: UncontrolledTooltips do not position correctly if the wrapper has an overflow: auto. After a lot of searching, [I landed on this popper.js issue](https://github.com/FezVrasta/popper.js/issues/648) which is exactly the problem I am reporting for Reactstrap. ### What should be happening? UncontrolledTooltip should be placed correctly on the target DOM element. ### Steps to reproduce issue 1. Create a div and wrap it in another div with the CSS (or the React style) overflow: auto. 2. Create an UncontrolledTooltip and set as target this div. ### Error message in console No error in console, just UX. ### Code Check the following section. ### A solution I found By looking at the issue [I pointed at previously](https://github.com/FezVrasta/popper.js/issues/648), seems like the problem is the following: > It's the correct behavior, the preventOverflow modifier uses by default, as boundariesElement, the scrollParent, you can change it to viewport or window or anything else supported. So, I resolved the issue by adding the `modifiers` prop to the UncontrolledTooltip: ` <UncontrolledTooltip modifiers={{preventOverflow: {boundariesElement: 'window'}}} autohide={false} delay={{ show: 200, hide: 5 }} placement='left' target={`${id}_${j}`}> ... </UncontrolledTooltip> `
True
UncontrolledTooltips do not position correctly if the wrapper has an overflow: auto - <!-- The following is part of our issue template, feel free to remove this if it doesn't apply --> ### Issue description - components: `UncontrolledTooltip` - reactstrap version `^6.1.0` - react version `^16.4.0` - bootstrap version `^4.1.1` ### What is happening? The title says it all: UncontrolledTooltips do not position correctly if the wrapper has an overflow: auto. After a lot of searching, [I landed on this popper.js issue](https://github.com/FezVrasta/popper.js/issues/648) which is exactly the problem I am reporting for Reactstrap. ### What should be happening? UncontrolledTooltip should be placed correctly on the target DOM element. ### Steps to reproduce issue 1. Create a div and wrap it in another div with the CSS (or the React style) overflow: auto. 2. Create an UncontrolledTooltip and set as target this div. ### Error message in console No error in console, just UX. ### Code Check the following section. ### A solution I found By looking at the issue [I pointed at previously](https://github.com/FezVrasta/popper.js/issues/648), seems like the problem is the following: > It's the correct behavior, the preventOverflow modifier uses by default, as boundariesElement, the scrollParent, you can change it to viewport or window or anything else supported. So, I resolved the issue by adding the `modifiers` prop to the UncontrolledTooltip: ` <UncontrolledTooltip modifiers={{preventOverflow: {boundariesElement: 'window'}}} autohide={false} delay={{ show: 200, hide: 5 }} placement='left' target={`${id}_${j}`}> ... </UncontrolledTooltip> `
non_defect
uncontrolledtooltips do not position correctly if the wrapper has an overflow auto issue description components uncontrolledtooltip reactstrap version react version bootstrap version what is happening the title says it all uncontrolledtooltips do not position correctly if the wrapper has an overflow auto after a lot of searching which is exactly the problem i am reporting for reactstrap what should be happening uncontrolledtooltip should be placed correctly on the target dom element steps to reproduce issue create a div and wrap it in another div with the css or the react style overflow auto create an uncontrolledtooltip and set as target this div error message in console no error in console just ux code check the following section a solution i found by looking at the issue seems like the problem is the following it s the correct behavior the preventoverflow modifier uses by default as boundarieselement the scrollparent you can change it to viewport or window or anything else supported so i resolved the issue by adding the modifiers prop to the uncontrolledtooltip
0
59,856
17,023,267,250
IssuesEvent
2021-07-03 01:08:16
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch 0.9c and Flash 10 under Ubuntu
Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 6.51pm, Thursday, 3rd July 2008]** When using Flash 10 under Ubuntu with Firefox3 i cant press shift to add a point to a way, like i normaly can do. Ok, Flash 10 is still Beta, but maybe someone can keep an eye on this.
1.0
Potlatch 0.9c and Flash 10 under Ubuntu - **[Submitted to the original trac issue database at 6.51pm, Thursday, 3rd July 2008]** When using Flash 10 under Ubuntu with Firefox3 i cant press shift to add a point to a way, like i normaly can do. Ok, Flash 10 is still Beta, but maybe someone can keep an eye on this.
defect
potlatch and flash under ubuntu when using flash under ubuntu with i cant press shift to add a point to a way like i normaly can do ok flash is still beta but maybe someone can keep an eye on this
1
730,961
25,196,423,099
IssuesEvent
2022-11-12 15:23:19
frodeheg/no.sparegris
https://api.github.com/repos/frodeheg/no.sparegris
closed
Saving for the first time returns a very anoying error message
bug High priority
The error message only displays the first time the settings are saved. It was introduced in version 0.18.32 by this commit: https://github.com/frodeheg/no.sparegris/commit/a2b5288ec69556fa637589e7319cb2f48949460d This should be fixed asap so new users do not question the integrity of the app.
1.0
Saving for the first time returns a very anoying error message - The error message only displays the first time the settings are saved. It was introduced in version 0.18.32 by this commit: https://github.com/frodeheg/no.sparegris/commit/a2b5288ec69556fa637589e7319cb2f48949460d This should be fixed asap so new users do not question the integrity of the app.
non_defect
saving for the first time returns a very anoying error message the error message only displays the first time the settings are saved it was introduced in version by this commit this should be fixed asap so new users do not question the integrity of the app
0
17,250
5,358,074,648
IssuesEvent
2017-02-20 20:49:07
numbbo/coco
https://api.github.com/repos/numbbo/coco
closed
Alignment of success probability in tables
Code-LS-Postprocessing Code-MO-Postprocessing Code-Postprocessing Easy Priority-Medium
The many algorithm tables got a strange alignment recently in their last column: ![alignmentmanyalgos](https://cloud.githubusercontent.com/assets/13706996/22593623/c12f33a8-ea1f-11e6-9632-c80de4189c26.png)
3.0
Alignment of success probability in tables - The many algorithm tables got a strange alignment recently in their last column: ![alignmentmanyalgos](https://cloud.githubusercontent.com/assets/13706996/22593623/c12f33a8-ea1f-11e6-9632-c80de4189c26.png)
non_defect
alignment of success probability in tables the many algorithm tables got a strange alignment recently in their last column
0
51,128
13,188,143,650
IssuesEvent
2020-08-13 05:40:21
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
hdfwriter-merge not properly padding rows (Trac #2039)
Migrated from Trac combo core defect
I have a bunch of HDF files produced with I3HDFWriter. Some variables are calculated only for upgoing and some only for downgoing events. As the final level event rate is very low, files may not contain certain tables at all. Merging these files together, the missing tables should be filled with empty rows (i.e. setting exists=0). To illustrate this, I have placed a few files showing this issue in `/data/user/tkintscher/bugs/hdfwriter_merge_missing_rows`. The issue appears in the tables `GFU_BDT_Score_Up` and `GFU_BDT_Score_Down`. '''Example 1 (fails):''' ```text $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT GFU_BDT_Score_Down Dataset {4/Inf} GFU_BDT_Score_Up Dataset {4/Inf} $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 | grep BDT GFU_BDT_Score_Down Dataset {3/Inf} $ hdfwriter-merge -o test_fail.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 ... $ h5ls test_fail.h5 | grep BDT GFU_BDT_Score_Down Dataset {7/Inf} GFU_BDT_Score_Up Dataset {5/Inf} ``` '''Expected outcome:''' As there are seven events combined in both input files, the output tables should have 7 rows. However, the _Up table only has 5. '''Example 2 (works):''' Replacing one of the input files with another one from the same dataset, with the same initial constellation of rows, the issue does not appear: ```text $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT GFU_BDT_Score_Down Dataset {4/Inf} GFU_BDT_Score_Up Dataset {4/Inf} $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 | grep BDT GFU_BDT_Score_Down Dataset {3/Inf} $ hdfwriter-merge -o test_ok.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 ... $ h5ls test_ok.h5 | grep BDT GFU_BDT_Score_Down Dataset {7/Inf} GFU_BDT_Score_Up Dataset {7/Inf} ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2039">https://code.icecube.wisc.edu/ticket/2039</a>, reported by thomas.kintscher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2017-06-28T17:20:04", "description": "I have a bunch of HDF files produced with I3HDFWriter. Some variables are calculated only for upgoing and some only for downgoing events. As the final level event rate is very low, files may not contain certain tables at all. Merging these files together, the missing tables should be filled with empty rows (i.e. setting exists=0).\n\nTo illustrate this, I have placed a few files showing this issue in `/data/user/tkintscher/bugs/hdfwriter_merge_missing_rows`. The issue appears in the tables `GFU_BDT_Score_Up` and `GFU_BDT_Score_Down`.\n\n'''Example 1 (fails):'''\n{{{\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {4/Inf}\nGFU_BDT_Score_Up Dataset {4/Inf}\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {3/Inf}\n$ hdfwriter-merge -o test_fail.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5\n...\n$ h5ls test_fail.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {7/Inf}\nGFU_BDT_Score_Up Dataset {5/Inf}\n}}}\n'''Expected outcome:'''\nAs there are seven events combined in both input files, the output tables should have 7 rows.\nHowever, the _Up table only has 5.\n\n'''Example 2 (works):'''\nReplacing one of the input files with another one from the same dataset, with the same initial constellation of rows, the issue does not appear:\n{{{\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {4/Inf}\nGFU_BDT_Score_Up Dataset {4/Inf}\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {3/Inf}\n$ hdfwriter-merge -o test_ok.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5\n...\n$ h5ls test_ok.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {7/Inf}\nGFU_BDT_Score_Up Dataset {7/Inf}\n}}}", "reporter": "thomas.kintscher", "cc": "", "resolution": "invalid", "_ts": "1498670404950731", "component": "combo core", "summary": "hdfwriter-merge not properly padding rows", "priority": "normal", "keywords": "", "time": "2017-06-13T14:14:10", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
hdfwriter-merge not properly padding rows (Trac #2039) - I have a bunch of HDF files produced with I3HDFWriter. Some variables are calculated only for upgoing and some only for downgoing events. As the final level event rate is very low, files may not contain certain tables at all. Merging these files together, the missing tables should be filled with empty rows (i.e. setting exists=0). To illustrate this, I have placed a few files showing this issue in `/data/user/tkintscher/bugs/hdfwriter_merge_missing_rows`. The issue appears in the tables `GFU_BDT_Score_Up` and `GFU_BDT_Score_Down`. '''Example 1 (fails):''' ```text $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT GFU_BDT_Score_Down Dataset {4/Inf} GFU_BDT_Score_Up Dataset {4/Inf} $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 | grep BDT GFU_BDT_Score_Down Dataset {3/Inf} $ hdfwriter-merge -o test_fail.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 ... $ h5ls test_fail.h5 | grep BDT GFU_BDT_Score_Down Dataset {7/Inf} GFU_BDT_Score_Up Dataset {5/Inf} ``` '''Expected outcome:''' As there are seven events combined in both input files, the output tables should have 7 rows. However, the _Up table only has 5. '''Example 2 (works):''' Replacing one of the input files with another one from the same dataset, with the same initial constellation of rows, the issue does not appear: ```text $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT GFU_BDT_Score_Down Dataset {4/Inf} GFU_BDT_Score_Up Dataset {4/Inf} $ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 | grep BDT GFU_BDT_Score_Down Dataset {3/Inf} $ hdfwriter-merge -o test_ok.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 ... $ h5ls test_ok.h5 | grep BDT GFU_BDT_Score_Down Dataset {7/Inf} GFU_BDT_Score_Up Dataset {7/Inf} ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2039">https://code.icecube.wisc.edu/ticket/2039</a>, reported by thomas.kintscher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2017-06-28T17:20:04", "description": "I have a bunch of HDF files produced with I3HDFWriter. Some variables are calculated only for upgoing and some only for downgoing events. As the final level event rate is very low, files may not contain certain tables at all. Merging these files together, the missing tables should be filled with empty rows (i.e. setting exists=0).\n\nTo illustrate this, I have placed a few files showing this issue in `/data/user/tkintscher/bugs/hdfwriter_merge_missing_rows`. The issue appears in the tables `GFU_BDT_Score_Up` and `GFU_BDT_Score_Down`.\n\n'''Example 1 (fails):'''\n{{{\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {4/Inf}\nGFU_BDT_Score_Up Dataset {4/Inf}\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {3/Inf}\n$ hdfwriter-merge -o test_fail.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091699.h5\n...\n$ h5ls test_fail.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {7/Inf}\nGFU_BDT_Score_Up Dataset {5/Inf}\n}}}\n'''Expected outcome:'''\nAs there are seven events combined in both input files, the output tables should have 7 rows.\nHowever, the _Up table only has 5.\n\n'''Example 2 (works):'''\nReplacing one of the input files with another one from the same dataset, with the same initial constellation of rows, the issue does not appear:\n{{{\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {4/Inf}\nGFU_BDT_Score_Up Dataset {4/Inf}\n$ h5ls /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {3/Inf}\n$ hdfwriter-merge -o test_ok.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091692.h5 /data/user/tkintscher/bugs/hdfwriter_merge_missing_rows/Level2_IC86.2012_corsika.011057.091697.h5\n...\n$ h5ls test_ok.h5 | grep BDT\nGFU_BDT_Score_Down Dataset {7/Inf}\nGFU_BDT_Score_Up Dataset {7/Inf}\n}}}", "reporter": "thomas.kintscher", "cc": "", "resolution": "invalid", "_ts": "1498670404950731", "component": "combo core", "summary": "hdfwriter-merge not properly padding rows", "priority": "normal", "keywords": "", "time": "2017-06-13T14:14:10", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
hdfwriter merge not properly padding rows trac i have a bunch of hdf files produced with some variables are calculated only for upgoing and some only for downgoing events as the final level event rate is very low files may not contain certain tables at all merging these files together the missing tables should be filled with empty rows i e setting exists to illustrate this i have placed a few files showing this issue in data user tkintscher bugs hdfwriter merge missing rows the issue appears in the tables gfu bdt score up and gfu bdt score down example fails text data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt gfu bdt score down dataset inf gfu bdt score up dataset inf data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt gfu bdt score down dataset inf hdfwriter merge o test fail data user tkintscher bugs hdfwriter merge missing rows corsika data user tkintscher bugs hdfwriter merge missing rows corsika test fail grep bdt gfu bdt score down dataset inf gfu bdt score up dataset inf expected outcome as there are seven events combined in both input files the output tables should have rows however the up table only has example works replacing one of the input files with another one from the same dataset with the same initial constellation of rows the issue does not appear text data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt gfu bdt score down dataset inf gfu bdt score up dataset inf data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt gfu bdt score down dataset inf hdfwriter merge o test ok data user tkintscher bugs hdfwriter merge missing rows corsika data user tkintscher bugs hdfwriter merge missing rows corsika test ok grep bdt gfu bdt score down dataset inf gfu bdt score up dataset inf migrated from json status closed changetime description i have a bunch of hdf files produced with some variables are calculated only for upgoing and some only for downgoing events as the final level event rate is very low files may not contain certain tables at all merging these files together the missing tables should be filled with empty rows i e setting exists n nto illustrate this i have placed a few files showing this issue in data user tkintscher bugs hdfwriter merge missing rows the issue appears in the tables gfu bdt score up and gfu bdt score down n n example fails n n data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt ngfu bdt score down dataset inf ngfu bdt score up dataset inf n data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt ngfu bdt score down dataset inf n hdfwriter merge o test fail data user tkintscher bugs hdfwriter merge missing rows corsika data user tkintscher bugs hdfwriter merge missing rows corsika n n test fail grep bdt ngfu bdt score down dataset inf ngfu bdt score up dataset inf n n expected outcome nas there are seven events combined in both input files the output tables should have rows nhowever the up table only has n n example works nreplacing one of the input files with another one from the same dataset with the same initial constellation of rows the issue does not appear n n data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt ngfu bdt score down dataset inf ngfu bdt score up dataset inf n data user tkintscher bugs hdfwriter merge missing rows corsika grep bdt ngfu bdt score down dataset inf n hdfwriter merge o test ok data user tkintscher bugs hdfwriter merge missing rows corsika data user tkintscher bugs hdfwriter merge missing rows corsika n n test ok grep bdt ngfu bdt score down dataset inf ngfu bdt score up dataset inf n reporter thomas kintscher cc resolution invalid ts component combo core summary hdfwriter merge not properly padding rows priority normal keywords time milestone owner jvansanten type defect
1
166,639
12,965,464,712
IssuesEvent
2020-07-20 22:25:08
dapr/cli
https://api.github.com/repos/dapr/cli
opened
[E2E Scenario Tests] for DAPR CLI
P1 area/test
## Describe the proposal Test basic scenarios such as:- - dapr init - dapr init -k - dapr uninstall
1.0
[E2E Scenario Tests] for DAPR CLI - ## Describe the proposal Test basic scenarios such as:- - dapr init - dapr init -k - dapr uninstall
non_defect
for dapr cli describe the proposal test basic scenarios such as dapr init dapr init k dapr uninstall
0
59,647
17,023,190,665
IssuesEvent
2021-07-03 00:47:07
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Junctions in tunnels / on bridges
Component: osmarender Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 3.13pm, Tuesday, 4th December 2007]** The arcades marking the beginning and end of a tunnel should not be shown if all the ways connected at a node have tunnel=yes. For reference, see http://informationfreeway.org/?lat=69.65717098928287&lon=18.970146284989344&zoom=15&layers=B000F000 (or the attached tile from the same location) where there are two roundabouts in the tunnel system.
1.0
Junctions in tunnels / on bridges - **[Submitted to the original trac issue database at 3.13pm, Tuesday, 4th December 2007]** The arcades marking the beginning and end of a tunnel should not be shown if all the ways connected at a node have tunnel=yes. For reference, see http://informationfreeway.org/?lat=69.65717098928287&lon=18.970146284989344&zoom=15&layers=B000F000 (or the attached tile from the same location) where there are two roundabouts in the tunnel system.
defect
junctions in tunnels on bridges the arcades marking the beginning and end of a tunnel should not be shown if all the ways connected at a node have tunnel yes for reference see or the attached tile from the same location where there are two roundabouts in the tunnel system
1
17,767
3,013,014,524
IssuesEvent
2015-07-29 05:23:08
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Decomposition variable not removable
auto-migrated Component-Editor Priority-High Type-Defect
``` What steps will reproduce the problem? 1. load enclosed spec in editor 2. try to remove variable "ClaimsDeadline" in task "Freight Delivery" of the starting net Variable cannot be removed. ``` Original issue reported on code.google.com by `stephan....@googlemail.com` on 9 Mar 2009 at 2:16
1.0
Decomposition variable not removable - ``` What steps will reproduce the problem? 1. load enclosed spec in editor 2. try to remove variable "ClaimsDeadline" in task "Freight Delivery" of the starting net Variable cannot be removed. ``` Original issue reported on code.google.com by `stephan....@googlemail.com` on 9 Mar 2009 at 2:16
defect
decomposition variable not removable what steps will reproduce the problem load enclosed spec in editor try to remove variable claimsdeadline in task freight delivery of the starting net variable cannot be removed original issue reported on code google com by stephan googlemail com on mar at
1
59,544
17,023,157,487
IssuesEvent
2021-07-03 00:37:58
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
GPX upload fails if tags or description fields are empty
Component: rails_port Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 5.05pm, Monday, 7th May 2007]** Rails does not redirect to traces/mine if the description or tag fields are empty when uploading gpx files
1.0
GPX upload fails if tags or description fields are empty - **[Submitted to the original trac issue database at 5.05pm, Monday, 7th May 2007]** Rails does not redirect to traces/mine if the description or tag fields are empty when uploading gpx files
defect
gpx upload fails if tags or description fields are empty rails does not redirect to traces mine if the description or tag fields are empty when uploading gpx files
1
19,237
3,163,418,994
IssuesEvent
2015-09-20 08:19:38
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Methods' optional parameters are not taken into account in generated JavaScript
defect
http://forums.bridge.net/forum/bridge-net-pro/bugs/680 It might be controversial a bit it is a defect or a feature, but in my opinion it is rather a defect. A similar TypeScript-related defect - #292.
1.0
Methods' optional parameters are not taken into account in generated JavaScript - http://forums.bridge.net/forum/bridge-net-pro/bugs/680 It might be controversial a bit it is a defect or a feature, but in my opinion it is rather a defect. A similar TypeScript-related defect - #292.
defect
methods optional parameters are not taken into account in generated javascript it might be controversial a bit it is a defect or a feature but in my opinion it is rather a defect a similar typescript related defect
1
34,566
7,457,299,850
IssuesEvent
2018-03-30 03:17:32
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
KÜ lisakirjelduse lingi RDFi kuvamisel viga
C: AVAR P: highest R: fixed T: defect
**Reported by sven syld on 10 Mar 2014 10:26 UTC** http://dev.raju.teepub/et/description_unit/view/?id=442050 -> lisakirjelduse link FOTIS 390403 -> http://opendata.dev.raju.teepub/photo/11e668e0b3a3 annab vea: ''You have requested a non-existent service "dira.opendata.puri_resolver_factory".''
1.0
KÜ lisakirjelduse lingi RDFi kuvamisel viga - **Reported by sven syld on 10 Mar 2014 10:26 UTC** http://dev.raju.teepub/et/description_unit/view/?id=442050 -> lisakirjelduse link FOTIS 390403 -> http://opendata.dev.raju.teepub/photo/11e668e0b3a3 annab vea: ''You have requested a non-existent service "dira.opendata.puri_resolver_factory".''
defect
kü lisakirjelduse lingi rdfi kuvamisel viga reported by sven syld on mar utc lisakirjelduse link fotis annab vea you have requested a non existent service dira opendata puri resolver factory
1
222,968
7,444,858,187
IssuesEvent
2018-03-28 00:57:48
EvictionLab/eviction-maps
https://api.github.com/repos/EvictionLab/eviction-maps
closed
Choropleth legend not visible on IE 11
bug high priority
The screenshot has the bubble legend visible, but the choropleth legend also isn't visible when the bubble legend isn't displayed. Guessing this is an IE flex bug? <img width="376" alt="screen shot 2018-03-27 at 10 22 47 am" src="https://user-images.githubusercontent.com/8291663/37977179-e2cfee40-31a8-11e8-9892-8e4f2c42d10f.png">
1.0
Choropleth legend not visible on IE 11 - The screenshot has the bubble legend visible, but the choropleth legend also isn't visible when the bubble legend isn't displayed. Guessing this is an IE flex bug? <img width="376" alt="screen shot 2018-03-27 at 10 22 47 am" src="https://user-images.githubusercontent.com/8291663/37977179-e2cfee40-31a8-11e8-9892-8e4f2c42d10f.png">
non_defect
choropleth legend not visible on ie the screenshot has the bubble legend visible but the choropleth legend also isn t visible when the bubble legend isn t displayed guessing this is an ie flex bug img width alt screen shot at am src
0
994
2,594,417,375
IssuesEvent
2015-02-20 03:06:35
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
closed
Erroneous Bonds to Calcium when downloading 2PTC
C: BALL Core P: major R: fixed T: defect
**Reported by dstoeckel on 5 Sep 41842328 21:20 UTC** The Ca2+ ions in 2PTC are covalently bound to the protein. The force fields are unable to parameterize these atoms leading to a runtime error when attempting an energy minimization. Original Report by Oliver: Incorrect bond guessing going on between metal ions and side chain atoms. - download 2PTC - removeWater() - addOptimizedHydrogens() AMBER setup fails because of missing parameters (bond of the Ca2+ with various side-chain and backbone atoms).
1.0
Erroneous Bonds to Calcium when downloading 2PTC - **Reported by dstoeckel on 5 Sep 41842328 21:20 UTC** The Ca2+ ions in 2PTC are covalently bound to the protein. The force fields are unable to parameterize these atoms leading to a runtime error when attempting an energy minimization. Original Report by Oliver: Incorrect bond guessing going on between metal ions and side chain atoms. - download 2PTC - removeWater() - addOptimizedHydrogens() AMBER setup fails because of missing parameters (bond of the Ca2+ with various side-chain and backbone atoms).
defect
erroneous bonds to calcium when downloading reported by dstoeckel on sep utc the ions in are covalently bound to the protein the force fields are unable to parameterize these atoms leading to a runtime error when attempting an energy minimization original report by oliver incorrect bond guessing going on between metal ions and side chain atoms download removewater addoptimizedhydrogens amber setup fails because of missing parameters bond of the with various side chain and backbone atoms
1
173,115
13,387,659,736
IssuesEvent
2020-09-02 16:17:31
astpl1998/Jyoti-Ceramic
https://api.github.com/repos/astpl1998/Jyoti-Ceramic
closed
I&P - DC Challan Print - RDW
17.Testing2_Completed
#43.. Need to do the PFS changes as per the Customer Requirement Regards, Srinivasan Baskaran
1.0
I&P - DC Challan Print - RDW - #43.. Need to do the PFS changes as per the Customer Requirement Regards, Srinivasan Baskaran
non_defect
i p dc challan print rdw need to do the pfs changes as per the customer requirement regards srinivasan baskaran
0
71,089
23,440,250,078
IssuesEvent
2022-08-15 14:14:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
high cpu after recording and play audio message
T-Defect A-Performance S-Major A-Voice Messages O-Occasional
### Steps to reproduce 1. Where are you starting? What can you see? >> chat 2. What do you click? >> record ### Outcome #### What did you expect? normal slow process #### What happened instead? crash after 2 minutes - sound craped after while more and more -- I have to reboot whole #macos ### Operating system macos ### Browser information chrome 104 ### URL for webapp app ### Application version Element version: 1.11.2 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? Yes
1.0
high cpu after recording and play audio message - ### Steps to reproduce 1. Where are you starting? What can you see? >> chat 2. What do you click? >> record ### Outcome #### What did you expect? normal slow process #### What happened instead? crash after 2 minutes - sound craped after while more and more -- I have to reboot whole #macos ### Operating system macos ### Browser information chrome 104 ### URL for webapp app ### Application version Element version: 1.11.2 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? Yes
defect
high cpu after recording and play audio message steps to reproduce where are you starting what can you see chat what do you click record outcome what did you expect normal slow process what happened instead crash after minutes sound craped after while more and more i have to reboot whole macos operating system macos browser information chrome url for webapp app application version element version olm version homeserver matrix org will you send logs yes
1
8,097
3,136,047,246
IssuesEvent
2015-09-10 18:01:16
mozilla/webmaker-core
https://api.github.com/repos/mozilla/webmaker-core
closed
Update documentation to include the use of webmaker-core
documentation
This issue has been migrated from mozilla/webmaker-android#2392. It was originally written by [ryanw-se](https://api.github.com/users/ryanw-se) on Mon Jul 06 2015 and had the following description: > I get a error when running ```npm run build```, it says "webmaker-core isn't installed" but the documentation is rather vague about the process on how to install webmaker-core. On Mon Jul 06 2015, [xmatthewx](https://api.github.com/users/xmatthewx) commented: > https://github.com/mozilla/webmaker-core/ On Mon Jul 06 2015, [ryanw-se](https://api.github.com/users/ryanw-se) commented: > **@****xmatthewx** Thanks, it works now. That documentation should definitely be put in the webmaker-android repository as well, or mentioned to prevent confusion!
1.0
Update documentation to include the use of webmaker-core - This issue has been migrated from mozilla/webmaker-android#2392. It was originally written by [ryanw-se](https://api.github.com/users/ryanw-se) on Mon Jul 06 2015 and had the following description: > I get a error when running ```npm run build```, it says "webmaker-core isn't installed" but the documentation is rather vague about the process on how to install webmaker-core. On Mon Jul 06 2015, [xmatthewx](https://api.github.com/users/xmatthewx) commented: > https://github.com/mozilla/webmaker-core/ On Mon Jul 06 2015, [ryanw-se](https://api.github.com/users/ryanw-se) commented: > **@****xmatthewx** Thanks, it works now. That documentation should definitely be put in the webmaker-android repository as well, or mentioned to prevent confusion!
non_defect
update documentation to include the use of webmaker core this issue has been migrated from mozilla webmaker android it was originally written by on mon jul and had the following description i get a error when running npm run build it says webmaker core isn t installed but the documentation is rather vague about the process on how to install webmaker core on mon jul commented on mon jul commented xmatthewx thanks it works now that documentation should definitely be put in the webmaker android repository as well or mentioned to prevent confusion
0
415,227
12,126,648,854
IssuesEvent
2020-04-22 17:22:37
ml4ai/automates
https://api.github.com/repos/ml4ai/automates
closed
Finish Implementation of TeX2Py
ER Team Mid Priority
# Overview @marcovzla I created a **very rough and undocumented** script to run [SymPy's LaTeX parsing pipeline](https://docs.sympy.org/latest/modules/parsing.html#experimental-latex-parsing). It is located at `scripts/equation_reading/tex2py.py`. What we need to do now is to modify this script to turn it into a callable library routine that takes a tokenized LaTeX equation string as input and outputs a string representation of the equivalent python code. Once we have that I can use Python's `ast` module to turn the mathematic expression code into a parse tree that I will align to a lambda expression extracted from source code. ### Some Important Notes - This implementation expects _chunked_ tokenized LaTeX (i.e. the chunking portion needs to already be accomplished) @marcovzla does this make sense? ### TeX2Py High-level Algorithm At a **very** high-level, here is what the `TeX2Py` algorithm tries to accomplish: ``` (0) Given a string of tokenized LaTeX (call it T) (1) Split T on = into a left-hand side (LHS) and a right-hand side (RHS) (1a) Set aside the LHS, we will return that as-is for now (1b) if there are multiple = (1bi) count every expression other than the LHS as an RHS (2) Remove common LaTeX formatting tokens (e.g. ~, \left, \mathrm, etc) (3) Create a LaTeX variable to simple variable map (call it V) (3a) simple variables will be a single letter (3b) LaTeX variables can use _{}, _, ^, ^{}, _{}^{}, ^{}_{} in their definition (3c) Convert to pythonic form by replacing _{} with _ and ^{} with __ (3d) Create map of pythonic vars to single-letter vars (4)Replace all variables in RHS with one-letter vars in V (5) Perform the translation to python with sympy.parsing.latex.parse_latex (6) Replace one letter vars with the pythonic vars from V (7) Return the results ``` @marcovzla if you have any questions about the above algorithm or any ambiguity associated with this task please @ me in this issue.
1.0
Finish Implementation of TeX2Py - # Overview @marcovzla I created a **very rough and undocumented** script to run [SymPy's LaTeX parsing pipeline](https://docs.sympy.org/latest/modules/parsing.html#experimental-latex-parsing). It is located at `scripts/equation_reading/tex2py.py`. What we need to do now is to modify this script to turn it into a callable library routine that takes a tokenized LaTeX equation string as input and outputs a string representation of the equivalent python code. Once we have that I can use Python's `ast` module to turn the mathematic expression code into a parse tree that I will align to a lambda expression extracted from source code. ### Some Important Notes - This implementation expects _chunked_ tokenized LaTeX (i.e. the chunking portion needs to already be accomplished) @marcovzla does this make sense? ### TeX2Py High-level Algorithm At a **very** high-level, here is what the `TeX2Py` algorithm tries to accomplish: ``` (0) Given a string of tokenized LaTeX (call it T) (1) Split T on = into a left-hand side (LHS) and a right-hand side (RHS) (1a) Set aside the LHS, we will return that as-is for now (1b) if there are multiple = (1bi) count every expression other than the LHS as an RHS (2) Remove common LaTeX formatting tokens (e.g. ~, \left, \mathrm, etc) (3) Create a LaTeX variable to simple variable map (call it V) (3a) simple variables will be a single letter (3b) LaTeX variables can use _{}, _, ^, ^{}, _{}^{}, ^{}_{} in their definition (3c) Convert to pythonic form by replacing _{} with _ and ^{} with __ (3d) Create map of pythonic vars to single-letter vars (4)Replace all variables in RHS with one-letter vars in V (5) Perform the translation to python with sympy.parsing.latex.parse_latex (6) Replace one letter vars with the pythonic vars from V (7) Return the results ``` @marcovzla if you have any questions about the above algorithm or any ambiguity associated with this task please @ me in this issue.
non_defect
finish implementation of overview marcovzla i created a very rough and undocumented script to run it is located at scripts equation reading py what we need to do now is to modify this script to turn it into a callable library routine that takes a tokenized latex equation string as input and outputs a string representation of the equivalent python code once we have that i can use python s ast module to turn the mathematic expression code into a parse tree that i will align to a lambda expression extracted from source code some important notes this implementation expects chunked tokenized latex i e the chunking portion needs to already be accomplished marcovzla does this make sense high level algorithm at a very high level here is what the algorithm tries to accomplish given a string of tokenized latex call it t split t on into a left hand side lhs and a right hand side rhs set aside the lhs we will return that as is for now if there are multiple count every expression other than the lhs as an rhs remove common latex formatting tokens e g left mathrm etc create a latex variable to simple variable map call it v simple variables will be a single letter latex variables can use in their definition convert to pythonic form by replacing with and with create map of pythonic vars to single letter vars replace all variables in rhs with one letter vars in v perform the translation to python with sympy parsing latex parse latex replace one letter vars with the pythonic vars from v return the results marcovzla if you have any questions about the above algorithm or any ambiguity associated with this task please me in this issue
0
2,872
2,607,964,014
IssuesEvent
2015-02-26 00:41:30
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
closed
Support alternate compilers.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. CXX=CC gmake What is the expected output? What do you see instead? CC should be used as the C++ compiler. g++ is instead used. What version of the product are you using? On what operating system? git tip. Solaris. Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `theo.sch...@circonus.com` on 10 Oct 2011 at 9:20
1.0
Support alternate compilers. - ``` What steps will reproduce the problem? 1. CXX=CC gmake What is the expected output? What do you see instead? CC should be used as the C++ compiler. g++ is instead used. What version of the product are you using? On what operating system? git tip. Solaris. Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `theo.sch...@circonus.com` on 10 Oct 2011 at 9:20
defect
support alternate compilers what steps will reproduce the problem cxx cc gmake what is the expected output what do you see instead cc should be used as the c compiler g is instead used what version of the product are you using on what operating system git tip solaris please provide any additional information below original issue reported on code google com by theo sch circonus com on oct at
1
281,995
24,443,001,559
IssuesEvent
2022-10-06 15:45:31
rancher/qa-tasks
https://api.github.com/repos/rancher/qa-tasks
opened
Add support for provisioning hardened clusters
team/area2 area/automation-test
### Issue Description When provisioning clusters, we should have automated tests to harden the clusters. For security purposes, this should be seen as the default, so tests should be added to our framework
1.0
Add support for provisioning hardened clusters - ### Issue Description When provisioning clusters, we should have automated tests to harden the clusters. For security purposes, this should be seen as the default, so tests should be added to our framework
non_defect
add support for provisioning hardened clusters issue description when provisioning clusters we should have automated tests to harden the clusters for security purposes this should be seen as the default so tests should be added to our framework
0
31,501
14,977,928,040
IssuesEvent
2021-01-28 10:08:54
input-output-hk/plutus
https://api.github.com/repos/input-output-hk/plutus
closed
Consider using HashSets in wallet-api
Good first issue Performance Refactoring
At the moment we compare `Script`s using their serialized form. While this is mostly fine, it does mean we're going to these operations on big `ByteString`s. It might be better to use `HashSet`s so we're operating on much smaller hashes instead.
True
Consider using HashSets in wallet-api - At the moment we compare `Script`s using their serialized form. While this is mostly fine, it does mean we're going to these operations on big `ByteString`s. It might be better to use `HashSet`s so we're operating on much smaller hashes instead.
non_defect
consider using hashsets in wallet api at the moment we compare script s using their serialized form while this is mostly fine it does mean we re going to these operations on big bytestring s it might be better to use hashset s so we re operating on much smaller hashes instead
0
57,797
14,219,475,262
IssuesEvent
2020-11-17 13:20:25
whatwg/html
https://api.github.com/repos/whatwg/html
closed
Appcache allows bypassing of Access-Control-Allow-Credentials
security/privacy topic: appcache
Demo: https://github.com/jakearchibald/appcache-credentials Appcache allows you to bypass CORS checks on pages that: - are served with `Access-Control-Allow-Origin: *` - contain sensitive data when fetched with credentials Without appcache, developers can add `Access-Control-Allow-Origin: *` to all resources available on the public internet, even if the data is sensitive with credentials (`Vary: Cookie` must be used if the resource is cachable). If the resource is fetched with credentials the CORS check will fail because `Access-Control-Allow-Credentials` is absent and `Access-Control-Allow-Origin` is not the requesting origin. Appcache caches everything with credentials and `no-cors`, and it serves items from its cache regardless of credentials or mode. This allows an attacker to cache a response requested with credentials, and later retrieve it using a no-credentials fetch, which only requires `Access-Control-Allow-Origin: *` to allow JS to access the content.
True
Appcache allows bypassing of Access-Control-Allow-Credentials - Demo: https://github.com/jakearchibald/appcache-credentials Appcache allows you to bypass CORS checks on pages that: - are served with `Access-Control-Allow-Origin: *` - contain sensitive data when fetched with credentials Without appcache, developers can add `Access-Control-Allow-Origin: *` to all resources available on the public internet, even if the data is sensitive with credentials (`Vary: Cookie` must be used if the resource is cachable). If the resource is fetched with credentials the CORS check will fail because `Access-Control-Allow-Credentials` is absent and `Access-Control-Allow-Origin` is not the requesting origin. Appcache caches everything with credentials and `no-cors`, and it serves items from its cache regardless of credentials or mode. This allows an attacker to cache a response requested with credentials, and later retrieve it using a no-credentials fetch, which only requires `Access-Control-Allow-Origin: *` to allow JS to access the content.
non_defect
appcache allows bypassing of access control allow credentials demo appcache allows you to bypass cors checks on pages that are served with access control allow origin contain sensitive data when fetched with credentials without appcache developers can add access control allow origin to all resources available on the public internet even if the data is sensitive with credentials vary cookie must be used if the resource is cachable if the resource is fetched with credentials the cors check will fail because access control allow credentials is absent and access control allow origin is not the requesting origin appcache caches everything with credentials and no cors and it serves items from its cache regardless of credentials or mode this allows an attacker to cache a response requested with credentials and later retrieve it using a no credentials fetch which only requires access control allow origin to allow js to access the content
0
77,405
26,974,551,657
IssuesEvent
2023-02-09 08:34:34
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
setAllToExcluded also sets fields not specified in insert to excluded when using INSERT .. SET syntax
T: Defect C: Functionality P: Medium E: All Editions
### Expected behavior Only sets fields specified in the insert statement to excluded. ```sql insert into "mm"."test" ("id") values (1) on conflict ("id") do update set "id" = excluded."id" ``` ### Actual behavior Sets all fields of the inserted table to excluded. ```sql insert into "mm"."test" ("id") values (1) on conflict ("id") do update set "id" = excluded."id", "value" = excluded."value" ``` ### Steps to reproduce the problem Example table: ```sql CREATE TABLE "mm"."test" ( id int PRIMARY KEY, value text ); ``` Example insert: ```kotlin insertInto(MM_TEST) .set(MM_TEST.ID, .onDuplicateKeyUpdate() .setAllToExcluded() .execute() ``` ### jOOQ Version 3.17.7 ### Database product and version PostgreSQL 15.1 on x86_64-pc-linux-musl, compiled by gcc (Alpine 12.2.1_git20220924-r4) 12.2.1 20220924, 64-bit ### Java Version openjdk version "17.0.5" 2022-10-18 ### OS Version MacOs Ventura 13.1 (22C65) ### JDBC driver name and version (include name if unofficial driver) org.postgresql:postgresql:42.5.1
1.0
setAllToExcluded also sets fields not specified in insert to excluded when using INSERT .. SET syntax - ### Expected behavior Only sets fields specified in the insert statement to excluded. ```sql insert into "mm"."test" ("id") values (1) on conflict ("id") do update set "id" = excluded."id" ``` ### Actual behavior Sets all fields of the inserted table to excluded. ```sql insert into "mm"."test" ("id") values (1) on conflict ("id") do update set "id" = excluded."id", "value" = excluded."value" ``` ### Steps to reproduce the problem Example table: ```sql CREATE TABLE "mm"."test" ( id int PRIMARY KEY, value text ); ``` Example insert: ```kotlin insertInto(MM_TEST) .set(MM_TEST.ID, .onDuplicateKeyUpdate() .setAllToExcluded() .execute() ``` ### jOOQ Version 3.17.7 ### Database product and version PostgreSQL 15.1 on x86_64-pc-linux-musl, compiled by gcc (Alpine 12.2.1_git20220924-r4) 12.2.1 20220924, 64-bit ### Java Version openjdk version "17.0.5" 2022-10-18 ### OS Version MacOs Ventura 13.1 (22C65) ### JDBC driver name and version (include name if unofficial driver) org.postgresql:postgresql:42.5.1
defect
setalltoexcluded also sets fields not specified in insert to excluded when using insert set syntax expected behavior only sets fields specified in the insert statement to excluded sql insert into mm test id values on conflict id do update set id excluded id actual behavior sets all fields of the inserted table to excluded sql insert into mm test id values on conflict id do update set id excluded id value excluded value steps to reproduce the problem example table sql create table mm test id int primary key value text example insert kotlin insertinto mm test set mm test id onduplicatekeyupdate setalltoexcluded execute jooq version database product and version postgresql on pc linux musl compiled by gcc alpine bit java version openjdk version os version macos ventura jdbc driver name and version include name if unofficial driver org postgresql postgresql
1
141,564
11,426,181,216
IssuesEvent
2020-02-03 21:22:57
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
RestoreCommand_PathTooLongException is disabled for netcore, we need to enable it
Area:PackageSigning Area:Test
RestoreCommand_PathTooLongException is disabled for netcore. We need to enable the test. There's an related issue: https://github.com/NuGet/Home/issues/7588
1.0
RestoreCommand_PathTooLongException is disabled for netcore, we need to enable it - RestoreCommand_PathTooLongException is disabled for netcore. We need to enable the test. There's an related issue: https://github.com/NuGet/Home/issues/7588
non_defect
restorecommand pathtoolongexception is disabled for netcore we need to enable it restorecommand pathtoolongexception is disabled for netcore we need to enable the test there s an related issue
0
100,826
21,524,882,448
IssuesEvent
2022-04-28 17:24:00
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
codeinsights query runner can throttle requests to searcher
team/code-insights
#### Feature request description codeinsights query runner can stop processing insights that exceed a certain number of failures. #### Is your feature request related to a problem? If so, please describe. Per https://github.com/sourcegraph/customer/issues/889, codeinsights query runner is spamming searcher with broken query constantly. We understand codeinsights heavily relies on searcher to generate series. However, on a deployment with limited resources (e.g. docker-compose distro only has one replica of searcher), this would impact other normal unindexed search requests and constantly firsing high CPU and RAM alerts. #### Describe alternatives you've considered. <!-- A description of any alternative solutions or features you've considered. --> #### Additional context <!-- Add any other context or other information you'd like to provide. -->
1.0
codeinsights query runner can throttle requests to searcher - #### Feature request description codeinsights query runner can stop processing insights that exceed a certain number of failures. #### Is your feature request related to a problem? If so, please describe. Per https://github.com/sourcegraph/customer/issues/889, codeinsights query runner is spamming searcher with broken query constantly. We understand codeinsights heavily relies on searcher to generate series. However, on a deployment with limited resources (e.g. docker-compose distro only has one replica of searcher), this would impact other normal unindexed search requests and constantly firsing high CPU and RAM alerts. #### Describe alternatives you've considered. <!-- A description of any alternative solutions or features you've considered. --> #### Additional context <!-- Add any other context or other information you'd like to provide. -->
non_defect
codeinsights query runner can throttle requests to searcher feature request description codeinsights query runner can stop processing insights that exceed a certain number of failures is your feature request related to a problem if so please describe per codeinsights query runner is spamming searcher with broken query constantly we understand codeinsights heavily relies on searcher to generate series however on a deployment with limited resources e g docker compose distro only has one replica of searcher this would impact other normal unindexed search requests and constantly firsing high cpu and ram alerts describe alternatives you ve considered additional context
0
67,870
21,194,944,348
IssuesEvent
2022-04-08 22:38:09
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: Selenium 4.1.3 doesn't Acknowledging proxy config set in ClientConfig
I-defect needs-triaging
### What happened? Selenium 4.1.3 unable to connect to RemoteWebDriver with proxy (no authentication) https://github.com/SeleniumHQ/selenium/commit/897caec5f850adb084d4409d1c5f8f57b6e66d26 I have tried debug. config is always null if (config.credentials() != null) { Credentials credentials = config.credentials(); previous bug: https://github.com/SeleniumHQ/selenium/issues/10231 ### How can we reproduce the issue? ```shell URL browserStackHubURL = new URL("<BS_URL>"); DesiredCapabilities capabilities = new DesiredCapabilities(); capabilities.setCapability("browserName", "chrome"); HashMap<String, Object> browserstackOptions = new HashMap<String, Object>(); browserstackOptions.put("os", "Windows"); browserstackOptions.put("osVersion", "10"); browserstackOptions.put("buildName", "BStack-[Java] Selenium 4 Sample Test"); browserstackOptions.put("sessionName", "Selenium 4 test"); browserstackOptions.put("seleniumVersion", "4.0.0"); capabilities.setCapability("bstack:options", browserstackOptions); InetSocketAddress inetSocketAddress = new InetSocketAddress(Constants.BS_PROXY_HOST, Constants.BS_PROXY_PORT); ClientConfig config = ClientConfig.defaultConfig() .baseUrl(browserStackHubURL) .proxy(new Proxy(HTTP, inetSocketAddress)); WebDriver driver = RemoteWebDriver.builder() .oneOf(capabilities) .config(config) .build(); ``` ### Relevant log output ```shell -Dio.netty.leakDetection.level: simple -Dio.netty.leakDetection.targetRecords: 4 Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@27f8f1cc -Dio.netty.noUnsafe: false Java version: 11 sun.misc.Unsafe.theUnsafe: available sun.misc.Unsafe.copyMemory: available java.nio.Buffer.address: available direct buffer constructor: unavailable: Reflective setAccessible(true) disabled java.nio.Bits.unaligned: available, true jdk.internal.misc.Unsafe.allocateUninitializedArray(int): unavailable: class io.netty.util.internal.PlatformDependent0$6 cannot access class jdk.internal.misc.Unsafe (in module java.base) because module java.base does not export jdk.internal.misc to unnamed module @5c8d685e java.nio.DirectByteBuffer.<init>(long, int): unavailable sun.misc.Unsafe: available maxDirectMemory: 16844324864 bytes (maybe) -Dio.netty.tmpdir: /tmp (java.io.tmpdir) -Dio.netty.bitMode: 64 (sun.arch.data.model) -Dio.netty.maxDirectMemory: -1 bytes -Dio.netty.uninitializedArrayAllocationThreshold: -1 java.nio.ByteBuffer.cleaner(): available -Dio.netty.noPreferDirect: false org.jctools-core.MpscChunkedArrayQueue: available -Dio.netty.threadLocalMap.stringBuilder.initialSize: 1024 -Dio.netty.threadLocalMap.stringBuilder.maxSize: 4096 Default protocols (JDK): [TLSv1.3, TLSv1.2, TLSv1.1, TLSv1] Default cipher suites (JDK): [TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, TLS_RSA_WITH_AES_128_GCM_SHA256, TLS_RSA_WITH_AES_128_CBC_SHA, TLS_RSA_WITH_AES_256_CBC_SHA, TLS_AES_128_GCM_SHA256, TLS_AES_256_GCM_SHA384] -Dio.netty.eventLoopThreads: 16 -Dio.netty.noKeySetOptimization: false -Dio.netty.selectorAutoRebuildThreshold: 512 -Dio.netty.allocator.numHeapArenas: 16 -Dio.netty.allocator.numDirectArenas: 16 -Dio.netty.allocator.pageSize: 8192 -Dio.netty.allocator.maxOrder: 9 -Dio.netty.allocator.chunkSize: 4194304 -Dio.netty.allocator.smallCacheSize: 256 -Dio.netty.allocator.normalCacheSize: 64 -Dio.netty.allocator.maxCachedBufferCapacity: 32768 -Dio.netty.allocator.cacheTrimInterval: 8192 -Dio.netty.allocator.cacheTrimIntervalMillis: 0 -Dio.netty.allocator.useCacheForAllThreads: false -Dio.netty.allocator.maxCachedByteBuffersPerChunk: 1023 -Dio.netty.allocator.type: pooled -Dio.netty.threadLocalDirectBufferSize: 0 -Dio.netty.maxThreadLocalCharBufferSize: 16384 -Dio.netty.processId: 53138 (auto-detected) -Djava.net.preferIPv4Stack: false -Djava.net.preferIPv6Addresses: false Loopback interface: lo (lo, 0:0:0:0:0:0:0:1%lo) /proc/sys/net/core/somaxconn: 128 -Dio.netty.machineId: 00:50:56:ff:fe:8f:65:96 (auto-detected) -Dio.netty.buffer.checkAccessible: true -Dio.netty.buffer.checkBounds: true Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@7a4d3465 connection timed out: hub.browserstack.com/207.254.56.9:443 org.openqa.selenium.SessionNotCreatedException: Could not start a new session. Response code 500. Message: java.net.ConnectException: connection timed out: hub.browserstack.com/207.254.56.9:443 Build info: version: '4.1.3', revision: '7b1ebf28ef' System info: host: 'jnk4stl22', ip: '10.158.23.86', os.name: 'Linux', os.arch: 'amd64', os.version: '3.10.0-1160.49.1.el7.x86_64', java.version: '11.0.1' Driver info: driver.version: unknown at org.openqa.selenium.remote.ProtocolHandshake.createSession(ProtocolHandshake.java:126) at org.openqa.selenium.remote.ProtocolHandshake.createSession(ProtocolHandshake.java:84) at org.openqa.selenium.remote.RemoteWebDriverBuilder.getRemoteDriver(RemoteWebDriverBuilder.java:395) at org.openqa.selenium.remote.RemoteWebDriverBuilder.build(RemoteWebDriverBuilder.java:368) at com.mastercard.commercial.track.testrunners.RunCucumber.beforeClass(RunCucumber.java:78) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.apache.maven.surefire.junitcore.pc.ParallelComputerBuilder$PC$1.run(ParallelComputerBuilder.java:590) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray2(ReflectionUtils.java:206) at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:161) at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:84) at org.apache.maven.plugin.surefire.InPluginVMSurefireStarter.runSuitesInProcess(InPluginVMSurefireStarter.java:87) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1166) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1011) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:857) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:134) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:207) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:307) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:193) at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:106) at org.apache.maven.cli.MavenCli.execute(MavenCli.java:863) at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:288) at org.apache.maven.cli.MavenCli.main(MavenCli.java:199) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:356) ``` ### Operating System MacOs Bigsur ### Selenium version 4.1.3 ### What are the browser(s) and version(s) where you see this issue? N/A ### What are the browser driver(s) and version(s) where you see this issue? N/A ### Are you using Selenium Grid? browser stack
1.0
[🐛 Bug]: Selenium 4.1.3 doesn't Acknowledging proxy config set in ClientConfig - ### What happened? Selenium 4.1.3 unable to connect to RemoteWebDriver with proxy (no authentication) https://github.com/SeleniumHQ/selenium/commit/897caec5f850adb084d4409d1c5f8f57b6e66d26 I have tried debug. config is always null if (config.credentials() != null) { Credentials credentials = config.credentials(); previous bug: https://github.com/SeleniumHQ/selenium/issues/10231 ### How can we reproduce the issue? ```shell URL browserStackHubURL = new URL("<BS_URL>"); DesiredCapabilities capabilities = new DesiredCapabilities(); capabilities.setCapability("browserName", "chrome"); HashMap<String, Object> browserstackOptions = new HashMap<String, Object>(); browserstackOptions.put("os", "Windows"); browserstackOptions.put("osVersion", "10"); browserstackOptions.put("buildName", "BStack-[Java] Selenium 4 Sample Test"); browserstackOptions.put("sessionName", "Selenium 4 test"); browserstackOptions.put("seleniumVersion", "4.0.0"); capabilities.setCapability("bstack:options", browserstackOptions); InetSocketAddress inetSocketAddress = new InetSocketAddress(Constants.BS_PROXY_HOST, Constants.BS_PROXY_PORT); ClientConfig config = ClientConfig.defaultConfig() .baseUrl(browserStackHubURL) .proxy(new Proxy(HTTP, inetSocketAddress)); WebDriver driver = RemoteWebDriver.builder() .oneOf(capabilities) .config(config) .build(); ``` ### Relevant log output ```shell -Dio.netty.leakDetection.level: simple -Dio.netty.leakDetection.targetRecords: 4 Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@27f8f1cc -Dio.netty.noUnsafe: false Java version: 11 sun.misc.Unsafe.theUnsafe: available sun.misc.Unsafe.copyMemory: available java.nio.Buffer.address: available direct buffer constructor: unavailable: Reflective setAccessible(true) disabled java.nio.Bits.unaligned: available, true jdk.internal.misc.Unsafe.allocateUninitializedArray(int): unavailable: class io.netty.util.internal.PlatformDependent0$6 cannot access class jdk.internal.misc.Unsafe (in module java.base) because module java.base does not export jdk.internal.misc to unnamed module @5c8d685e java.nio.DirectByteBuffer.<init>(long, int): unavailable sun.misc.Unsafe: available maxDirectMemory: 16844324864 bytes (maybe) -Dio.netty.tmpdir: /tmp (java.io.tmpdir) -Dio.netty.bitMode: 64 (sun.arch.data.model) -Dio.netty.maxDirectMemory: -1 bytes -Dio.netty.uninitializedArrayAllocationThreshold: -1 java.nio.ByteBuffer.cleaner(): available -Dio.netty.noPreferDirect: false org.jctools-core.MpscChunkedArrayQueue: available -Dio.netty.threadLocalMap.stringBuilder.initialSize: 1024 -Dio.netty.threadLocalMap.stringBuilder.maxSize: 4096 Default protocols (JDK): [TLSv1.3, TLSv1.2, TLSv1.1, TLSv1] Default cipher suites (JDK): [TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, TLS_RSA_WITH_AES_128_GCM_SHA256, TLS_RSA_WITH_AES_128_CBC_SHA, TLS_RSA_WITH_AES_256_CBC_SHA, TLS_AES_128_GCM_SHA256, TLS_AES_256_GCM_SHA384] -Dio.netty.eventLoopThreads: 16 -Dio.netty.noKeySetOptimization: false -Dio.netty.selectorAutoRebuildThreshold: 512 -Dio.netty.allocator.numHeapArenas: 16 -Dio.netty.allocator.numDirectArenas: 16 -Dio.netty.allocator.pageSize: 8192 -Dio.netty.allocator.maxOrder: 9 -Dio.netty.allocator.chunkSize: 4194304 -Dio.netty.allocator.smallCacheSize: 256 -Dio.netty.allocator.normalCacheSize: 64 -Dio.netty.allocator.maxCachedBufferCapacity: 32768 -Dio.netty.allocator.cacheTrimInterval: 8192 -Dio.netty.allocator.cacheTrimIntervalMillis: 0 -Dio.netty.allocator.useCacheForAllThreads: false -Dio.netty.allocator.maxCachedByteBuffersPerChunk: 1023 -Dio.netty.allocator.type: pooled -Dio.netty.threadLocalDirectBufferSize: 0 -Dio.netty.maxThreadLocalCharBufferSize: 16384 -Dio.netty.processId: 53138 (auto-detected) -Djava.net.preferIPv4Stack: false -Djava.net.preferIPv6Addresses: false Loopback interface: lo (lo, 0:0:0:0:0:0:0:1%lo) /proc/sys/net/core/somaxconn: 128 -Dio.netty.machineId: 00:50:56:ff:fe:8f:65:96 (auto-detected) -Dio.netty.buffer.checkAccessible: true -Dio.netty.buffer.checkBounds: true Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@7a4d3465 connection timed out: hub.browserstack.com/207.254.56.9:443 org.openqa.selenium.SessionNotCreatedException: Could not start a new session. Response code 500. Message: java.net.ConnectException: connection timed out: hub.browserstack.com/207.254.56.9:443 Build info: version: '4.1.3', revision: '7b1ebf28ef' System info: host: 'jnk4stl22', ip: '10.158.23.86', os.name: 'Linux', os.arch: 'amd64', os.version: '3.10.0-1160.49.1.el7.x86_64', java.version: '11.0.1' Driver info: driver.version: unknown at org.openqa.selenium.remote.ProtocolHandshake.createSession(ProtocolHandshake.java:126) at org.openqa.selenium.remote.ProtocolHandshake.createSession(ProtocolHandshake.java:84) at org.openqa.selenium.remote.RemoteWebDriverBuilder.getRemoteDriver(RemoteWebDriverBuilder.java:395) at org.openqa.selenium.remote.RemoteWebDriverBuilder.build(RemoteWebDriverBuilder.java:368) at com.mastercard.commercial.track.testrunners.RunCucumber.beforeClass(RunCucumber.java:78) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.apache.maven.surefire.junitcore.pc.ParallelComputerBuilder$PC$1.run(ParallelComputerBuilder.java:590) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray2(ReflectionUtils.java:206) at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:161) at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:84) at org.apache.maven.plugin.surefire.InPluginVMSurefireStarter.runSuitesInProcess(InPluginVMSurefireStarter.java:87) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1166) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1011) at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:857) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:134) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:207) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:307) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:193) at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:106) at org.apache.maven.cli.MavenCli.execute(MavenCli.java:863) at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:288) at org.apache.maven.cli.MavenCli.main(MavenCli.java:199) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:356) ``` ### Operating System MacOs Bigsur ### Selenium version 4.1.3 ### What are the browser(s) and version(s) where you see this issue? N/A ### What are the browser driver(s) and version(s) where you see this issue? N/A ### Are you using Selenium Grid? browser stack
defect
selenium doesn t acknowledging proxy config set in clientconfig what happened selenium unable to connect to remotewebdriver with proxy no authentication i have tried debug config is always null if config credentials null credentials credentials config credentials previous bug how can we reproduce the issue shell url browserstackhuburl new url desiredcapabilities capabilities new desiredcapabilities capabilities setcapability browsername chrome hashmap browserstackoptions new hashmap browserstackoptions put os windows browserstackoptions put osversion browserstackoptions put buildname bstack selenium sample test browserstackoptions put sessionname selenium test browserstackoptions put seleniumversion capabilities setcapability bstack options browserstackoptions inetsocketaddress inetsocketaddress new inetsocketaddress constants bs proxy host constants bs proxy port clientconfig config clientconfig defaultconfig baseurl browserstackhuburl proxy new proxy http inetsocketaddress webdriver driver remotewebdriver builder oneof capabilities config config build relevant log output shell dio netty leakdetection level simple dio netty leakdetection targetrecords loaded default resourceleakdetector io netty util resourceleakdetector dio netty nounsafe false java version sun misc unsafe theunsafe available sun misc unsafe copymemory available java nio buffer address available direct buffer constructor unavailable reflective setaccessible true disabled java nio bits unaligned available true jdk internal misc unsafe allocateuninitializedarray int unavailable class io netty util internal cannot access class jdk internal misc unsafe in module java base because module java base does not export jdk internal misc to unnamed module java nio directbytebuffer long int unavailable sun misc unsafe available maxdirectmemory bytes maybe dio netty tmpdir tmp java io tmpdir dio netty bitmode sun arch data model dio netty maxdirectmemory bytes dio netty uninitializedarrayallocationthreshold java nio bytebuffer cleaner available dio netty nopreferdirect false org jctools core mpscchunkedarrayqueue available dio netty threadlocalmap stringbuilder initialsize dio netty threadlocalmap stringbuilder maxsize default protocols jdk default cipher suites jdk dio netty eventloopthreads dio netty nokeysetoptimization false dio netty selectorautorebuildthreshold dio netty allocator numheaparenas dio netty allocator numdirectarenas dio netty allocator pagesize dio netty allocator maxorder dio netty allocator chunksize dio netty allocator smallcachesize dio netty allocator normalcachesize dio netty allocator maxcachedbuffercapacity dio netty allocator cachetriminterval dio netty allocator cachetrimintervalmillis dio netty allocator usecacheforallthreads false dio netty allocator maxcachedbytebuffersperchunk dio netty allocator type pooled dio netty threadlocaldirectbuffersize dio netty maxthreadlocalcharbuffersize dio netty processid auto detected djava net false djava net false loopback interface lo lo lo proc sys net core somaxconn dio netty machineid ff fe auto detected dio netty buffer checkaccessible true dio netty buffer checkbounds true loaded default resourceleakdetector io netty util resourceleakdetector connection timed out hub browserstack com org openqa selenium sessionnotcreatedexception could not start a new session response code message java net connectexception connection timed out hub browserstack com build info version revision system info host ip os name linux os arch os version java version driver info driver version unknown at org openqa selenium remote protocolhandshake createsession protocolhandshake java at org openqa selenium remote protocolhandshake createsession protocolhandshake java at org openqa selenium remote remotewebdriverbuilder getremotedriver remotewebdriverbuilder java at org openqa selenium remote remotewebdriverbuilder build remotewebdriverbuilder java at com mastercard commercial track testrunners runcucumber beforeclass runcucumber java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc parallelcomputerbuilder pc run parallelcomputerbuilder java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executeeager junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org apache maven surefire util reflectionutils reflectionutils java at org apache maven surefire booter providerfactory providerproxy invoke providerfactory java at org apache maven surefire booter providerfactory invokeprovider providerfactory java at org apache maven plugin surefire inpluginvmsurefirestarter runsuitesinprocess inpluginvmsurefirestarter java at org apache maven plugin surefire abstractsurefiremojo executeprovider abstractsurefiremojo java at org apache maven plugin surefire abstractsurefiremojo executeafterpreconditionschecked abstractsurefiremojo java at org apache maven plugin surefire abstractsurefiremojo execute abstractsurefiremojo java at org apache maven plugin defaultbuildpluginmanager executemojo defaultbuildpluginmanager java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal builder singlethreaded singlethreadedbuilder build singlethreadedbuilder java at org apache maven lifecycle internal lifecyclestarter execute lifecyclestarter java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven execute defaultmaven java at org apache maven cli mavencli execute mavencli java at org apache maven cli mavencli domain mavencli java at org apache maven cli mavencli main mavencli java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org codehaus plexus classworlds launcher launcher launchenhanced launcher java at org codehaus plexus classworlds launcher launcher launch launcher java at org codehaus plexus classworlds launcher launcher mainwithexitcode launcher java at org codehaus plexus classworlds launcher launcher main launcher java operating system macos bigsur selenium version what are the browser s and version s where you see this issue n a what are the browser driver s and version s where you see this issue n a are you using selenium grid browser stack
1
67,969
21,366,829,180
IssuesEvent
2022-04-20 03:14:25
svalinn/DAGMC
https://api.github.com/repos/svalinn/DAGMC
opened
False positive for assignment of vacuum
Type: Defect
@shimwell recently sent me an input with material assignments like "lower_vaccum_vessel", which DAGMC interpreted as a vaccuum assignment. This is caused by the following lines: https://github.com/svalinn/DAGMC/blob/50caabf16b12179f92e71d4dcfb5bf6090b30cb6/src/dagmc/dagmcmetadata.cpp#L209-L212 Prior to https://github.com/svalinn/DAGMC/commit/cc505bb4e4c70c343542de7b0d2eaa211e031d53, this would not have been a problem since only the presence of "Vacuum" (as opposed to "vacuum") in the material assignment would have triggered this behavior. This is still what I would call a bug, the corner case was just rarer. I think we either need to document that "V/vacuum" should not be used in material names or only assign volumes a vacuum condition if the lowercase version of the name matches "mat:vacuum" exactly instead of searching for the presence of "vacuum", which is what happens in the lines linked above. The word vacuum starts to look very strange after typing it so many times at this time of night.
1.0
False positive for assignment of vacuum - @shimwell recently sent me an input with material assignments like "lower_vaccum_vessel", which DAGMC interpreted as a vaccuum assignment. This is caused by the following lines: https://github.com/svalinn/DAGMC/blob/50caabf16b12179f92e71d4dcfb5bf6090b30cb6/src/dagmc/dagmcmetadata.cpp#L209-L212 Prior to https://github.com/svalinn/DAGMC/commit/cc505bb4e4c70c343542de7b0d2eaa211e031d53, this would not have been a problem since only the presence of "Vacuum" (as opposed to "vacuum") in the material assignment would have triggered this behavior. This is still what I would call a bug, the corner case was just rarer. I think we either need to document that "V/vacuum" should not be used in material names or only assign volumes a vacuum condition if the lowercase version of the name matches "mat:vacuum" exactly instead of searching for the presence of "vacuum", which is what happens in the lines linked above. The word vacuum starts to look very strange after typing it so many times at this time of night.
defect
false positive for assignment of vacuum shimwell recently sent me an input with material assignments like lower vaccum vessel which dagmc interpreted as a vaccuum assignment this is caused by the following lines prior to this would not have been a problem since only the presence of vacuum as opposed to vacuum in the material assignment would have triggered this behavior this is still what i would call a bug the corner case was just rarer i think we either need to document that v vacuum should not be used in material names or only assign volumes a vacuum condition if the lowercase version of the name matches mat vacuum exactly instead of searching for the presence of vacuum which is what happens in the lines linked above the word vacuum starts to look very strange after typing it so many times at this time of night
1
8,352
2,611,493,912
IssuesEvent
2015-02-27 05:34:03
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
New Siplified Chinese translation for a262ef09c7e6
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? -- What is the expected output? What do you see instead? -- What version of the product are you using? On what operating system? a262ef09c7e6 Please provide any additional information below. I've only translated hedgewars_zh_CN.ts and zh_CN.txt. No change to the .lua file:) BTW, I love this game! ``` Original issue reported on code.google.com by `cherrot...@gmail.com` on 10 Jan 2012 at 9:16 Attachments: * [ChineseTranslation.tar.bz2](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-345/comment-0/ChineseTranslation.tar.bz2)
1.0
New Siplified Chinese translation for a262ef09c7e6 - ``` What steps will reproduce the problem? -- What is the expected output? What do you see instead? -- What version of the product are you using? On what operating system? a262ef09c7e6 Please provide any additional information below. I've only translated hedgewars_zh_CN.ts and zh_CN.txt. No change to the .lua file:) BTW, I love this game! ``` Original issue reported on code.google.com by `cherrot...@gmail.com` on 10 Jan 2012 at 9:16 Attachments: * [ChineseTranslation.tar.bz2](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-345/comment-0/ChineseTranslation.tar.bz2)
defect
new siplified chinese translation for what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below i ve only translated hedgewars zh cn ts and zh cn txt no change to the lua file btw i love this game original issue reported on code google com by cherrot gmail com on jan at attachments
1
5,332
2,610,185,596
IssuesEvent
2015-02-26 18:58:57
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
解谜怎样有效去色斑
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 化妆品已经普遍到每个人,不管是学生或是年轻人,还是老�� �人,都希望自己的皮肤白嫩细腻。但是皮肤粗糙长斑困扰着� ��多女性。大部分女性都经历过这种事情,白皙的皮肤莫名的 长斑,皮肤变的暗黄粗糙。只是一味的用美白祛斑的化妆品�� �不但皮肤没有得到改善,反而黄褐斑更重,皮肤更粗糙发黑� ��不化妆都不敢出门,全靠涂抹厚厚的化妆品遮盖,长此以往 造成脸上黄褐斑越来越严重,那么去色斑的方法怎么解决呢�� �怎样有效去色斑, 《客户案例》   每个女孩子都希望自己漂亮迷人,更希望有个帅气的白�� �王子陪着身边,可这一切对于我来说都只是个梦想。<br>   我现在是公司的销售主管,能爬到现在的位置,我付出�� �很多。我刚毕业的时候进的现在的公司,那时候我只是个刚� ��校门的学生,什么都不懂,很多公司都是要有工作经验的, 可我什么都不会,别人一听我刚毕业就没下文了,后来经过�� �多失败,我才在现在的公司找到了一个客服的工作,其实客� ��说白了就是为客户服务,我们公司的产品都是因为客服的工 作做的很到位,顾客才会相信才会购买。我在经过了两个星�� �的培训后,走上了工作岗位。在工作中我遇到了形形色色的� ��客,我得去做好沟通,有了投诉,我得去仔细协调,虽然很 累,但是看到自己的业绩我更多的是欣慰。因为我们客服部�� �需要值班的,一般要值班的十二点,就这样过了两年,因为� ��太专注于事业,忽视了自己的情况,连我自己都不知道什么 时候脸上就开始长斑了,等我发觉的时候,他们已经密密麻�� �的长了很多了,脸的两边,眼睛周围,原来还算清秀的脸现� ��简直像黄脸婆了。<br>   我还没有男朋友,现在这样我可怎么嫁人啊,为了把脸�� �的斑去掉,我真是费劲了周折。刚开始我是用的能祛斑的化� ��品,是商场化妆品小姐推荐的,说效果特别好,我信以为真 ,用了很长一段时间当时斑真的淡了很多,我以为没事了,�� �停用又开始反弹了,接着我听说激光祛斑效果很好,做完斑� ��没了,我就又做了,可好景不长,一到夏天就又反弹了,气 死人了,后来有个同事告诉我说「黛芙薇尔精华液」不错,�� �有亲戚在用,效果还挺好。我这回将信将疑了,我都用了那� ��多了都没用,就你一个小精华液就管用了,可我还是经不住 到网上查了查,看到很多人所说这个效果很好,我想既然已�� �用了这么多办法了,也不差这一个了,就试着买了。大约用� ��一个多月的时间斑开始慢慢变淡了,用了有两个多月斑就没 有了,虽然效果挺好的,可我心里还是不放心,毕竟我前几�� �都反弹过,我又慢慢等了又三个月的时间吧,脸上一点都没� ��化,我这个彻底放心了。<br>   现在我已经找到了我的白马王子,我的幸福生活马上就�� �开始了,我的梦想终于实现了。 阅读了怎样有效去色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎样有效去色斑,同时为您分享祛斑小方法 1:醋水:面部生有雀斑者,在洗脸时可在水中加入1~2匙食�� �,有减轻色素沉着作用 2:茄子汁:茄子1个,切片榨汁,磨擦局部,1日3次,十天可� ��效果。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:42
1.0
解谜怎样有效去色斑 - ``` 《摘要》 化妆品已经普遍到每个人,不管是学生或是年轻人,还是老�� �人,都希望自己的皮肤白嫩细腻。但是皮肤粗糙长斑困扰着� ��多女性。大部分女性都经历过这种事情,白皙的皮肤莫名的 长斑,皮肤变的暗黄粗糙。只是一味的用美白祛斑的化妆品�� �不但皮肤没有得到改善,反而黄褐斑更重,皮肤更粗糙发黑� ��不化妆都不敢出门,全靠涂抹厚厚的化妆品遮盖,长此以往 造成脸上黄褐斑越来越严重,那么去色斑的方法怎么解决呢�� �怎样有效去色斑, 《客户案例》   每个女孩子都希望自己漂亮迷人,更希望有个帅气的白�� �王子陪着身边,可这一切对于我来说都只是个梦想。<br>   我现在是公司的销售主管,能爬到现在的位置,我付出�� �很多。我刚毕业的时候进的现在的公司,那时候我只是个刚� ��校门的学生,什么都不懂,很多公司都是要有工作经验的, 可我什么都不会,别人一听我刚毕业就没下文了,后来经过�� �多失败,我才在现在的公司找到了一个客服的工作,其实客� ��说白了就是为客户服务,我们公司的产品都是因为客服的工 作做的很到位,顾客才会相信才会购买。我在经过了两个星�� �的培训后,走上了工作岗位。在工作中我遇到了形形色色的� ��客,我得去做好沟通,有了投诉,我得去仔细协调,虽然很 累,但是看到自己的业绩我更多的是欣慰。因为我们客服部�� �需要值班的,一般要值班的十二点,就这样过了两年,因为� ��太专注于事业,忽视了自己的情况,连我自己都不知道什么 时候脸上就开始长斑了,等我发觉的时候,他们已经密密麻�� �的长了很多了,脸的两边,眼睛周围,原来还算清秀的脸现� ��简直像黄脸婆了。<br>   我还没有男朋友,现在这样我可怎么嫁人啊,为了把脸�� �的斑去掉,我真是费劲了周折。刚开始我是用的能祛斑的化� ��品,是商场化妆品小姐推荐的,说效果特别好,我信以为真 ,用了很长一段时间当时斑真的淡了很多,我以为没事了,�� �停用又开始反弹了,接着我听说激光祛斑效果很好,做完斑� ��没了,我就又做了,可好景不长,一到夏天就又反弹了,气 死人了,后来有个同事告诉我说「黛芙薇尔精华液」不错,�� �有亲戚在用,效果还挺好。我这回将信将疑了,我都用了那� ��多了都没用,就你一个小精华液就管用了,可我还是经不住 到网上查了查,看到很多人所说这个效果很好,我想既然已�� �用了这么多办法了,也不差这一个了,就试着买了。大约用� ��一个多月的时间斑开始慢慢变淡了,用了有两个多月斑就没 有了,虽然效果挺好的,可我心里还是不放心,毕竟我前几�� �都反弹过,我又慢慢等了又三个月的时间吧,脸上一点都没� ��化,我这个彻底放心了。<br>   现在我已经找到了我的白马王子,我的幸福生活马上就�� �开始了,我的梦想终于实现了。 阅读了怎样有效去色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎样有效去色斑,同时为您分享祛斑小方法 1:醋水:面部生有雀斑者,在洗脸时可在水中加入1~2匙食�� �,有减轻色素沉着作用 2:茄子汁:茄子1个,切片榨汁,磨擦局部,1日3次,十天可� ��效果。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:42
defect
解谜怎样有效去色斑 《摘要》 化妆品已经普遍到每个人,不管是学生或是年轻人,还是老�� �人,都希望自己的皮肤白嫩细腻。但是皮肤粗糙长斑困扰着� ��多女性。大部分女性都经历过这种事情,白皙的皮肤莫名的 长斑,皮肤变的暗黄粗糙。只是一味的用美白祛斑的化妆品�� �不但皮肤没有得到改善,反而黄褐斑更重,皮肤更粗糙发黑� ��不化妆都不敢出门,全靠涂抹厚厚的化妆品遮盖,长此以往 造成脸上黄褐斑越来越严重,那么去色斑的方法怎么解决呢�� �怎样有效去色斑, 《客户案例》   每个女孩子都希望自己漂亮迷人,更希望有个帅气的白�� �王子陪着身边,可这一切对于我来说都只是个梦想。   我现在是公司的销售主管,能爬到现在的位置,我付出�� �很多。我刚毕业的时候进的现在的公司,那时候我只是个刚� ��校门的学生,什么都不懂,很多公司都是要有工作经验的, 可我什么都不会,别人一听我刚毕业就没下文了,后来经过�� �多失败,我才在现在的公司找到了一个客服的工作,其实客� ��说白了就是为客户服务,我们公司的产品都是因为客服的工 作做的很到位,顾客才会相信才会购买。我在经过了两个星�� �的培训后,走上了工作岗位。在工作中我遇到了形形色色的� ��客,我得去做好沟通,有了投诉,我得去仔细协调,虽然很 累,但是看到自己的业绩我更多的是欣慰。因为我们客服部�� �需要值班的,一般要值班的十二点,就这样过了两年,因为� ��太专注于事业,忽视了自己的情况,连我自己都不知道什么 时候脸上就开始长斑了,等我发觉的时候,他们已经密密麻�� �的长了很多了,脸的两边,眼睛周围,原来还算清秀的脸现� ��简直像黄脸婆了。   我还没有男朋友,现在这样我可怎么嫁人啊,为了把脸�� �的斑去掉,我真是费劲了周折。刚开始我是用的能祛斑的化� ��品,是商场化妆品小姐推荐的,说效果特别好,我信以为真 ,用了很长一段时间当时斑真的淡了很多,我以为没事了,�� �停用又开始反弹了,接着我听说激光祛斑效果很好,做完斑� ��没了,我就又做了,可好景不长,一到夏天就又反弹了,气 死人了,后来有个同事告诉我说「黛芙薇尔精华液」不错,�� �有亲戚在用,效果还挺好。我这回将信将疑了,我都用了那� ��多了都没用,就你一个小精华液就管用了,可我还是经不住 到网上查了查,看到很多人所说这个效果很好,我想既然已�� �用了这么多办法了,也不差这一个了,就试着买了。大约用� ��一个多月的时间斑开始慢慢变淡了,用了有两个多月斑就没 有了,虽然效果挺好的,可我心里还是不放心,毕竟我前几�� �都反弹过,我又慢慢等了又三个月的时间吧,脸上一点都没� ��化,我这个彻底放心了。   现在我已经找到了我的白马王子,我的幸福生活马上就�� �开始了,我的梦想终于实现了。 阅读了怎样有效去色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 怎样有效去色斑,同时为您分享祛斑小方法 :醋水:面部生有雀斑者, ~ �� �,有减轻色素沉着作用 :茄子汁: ,切片榨汁,磨擦局部, ,十天可� ��效果。 original issue reported on code google com by additive gmail com on jul at
1
52,845
13,225,156,842
IssuesEvent
2020-08-17 20:36:18
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
ipdf does not build (Trac #455)
Migrated from Trac combo core defect
an attempt to build a fresh checkout of analysis trunk ends like this: ```text [ 93%] Building CXX object ipdf/CMakeFiles/ipdf-benchmark.dir/private/benchmark/main.cxx.o Linking CXX shared module ../../../lib/icecube/I3Db.so Linking CXX executable ../bin/ipdf-anim CMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `main': main.cxx:(.text+0x1b0): undefined reference to `I3MediumPropertiesFile::I3MediumPropertiesFile(std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)' main.cxx:(.text+0x22f): undefined reference to `I3MediumService::I3MediumService(I3MediumProperties const&, std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)' CMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const': main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0xff): undefined reference to `I3MediumService::RECO_WAVELENGTH' main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x160): undefined reference to `I3MediumService::Absorptivity(double, double) const' main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x271): undefined reference to `I3MediumService::InvEffScattLength(double, double) const' collect2: ld returned 1 exit status make[2]: *** [bin/ipdf-anim] Error 1 make[1]: *** [ipdf/CMakeFiles/ipdf-anim.dir/all] Error 2 ``` Seems to be related to the latest commits to ipdf. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/455">https://code.icecube.wisc.edu/projects/icecube/ticket/455</a>, reported by hwissingand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-12-07T16:23:37", "_ts": "1354897417000000", "description": "an attempt to build a fresh checkout of analysis trunk ends like this:\n{{{\n[ 93%] Building CXX object ipdf/CMakeFiles/ipdf-benchmark.dir/private/benchmark/main.cxx.o\nLinking CXX shared module ../../../lib/icecube/I3Db.so\nLinking CXX executable ../bin/ipdf-anim\nCMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `main':\nmain.cxx:(.text+0x1b0): undefined reference to `I3MediumPropertiesFile::I3MediumPropertiesFile(std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)'\nmain.cxx:(.text+0x22f): undefined reference to `I3MediumService::I3MediumService(I3MediumProperties const&, std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)'\nCMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const':\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0xff): undefined reference to `I3MediumService::RECO_WAVELENGTH'\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x160): undefined reference to `I3MediumService::Absorptivity(double, double) const'\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x271): undefined reference to `I3MediumService::InvEffScattLength(double, double) const'\ncollect2: ld returned 1 exit status\nmake[2]: *** [bin/ipdf-anim] Error 1\nmake[1]: *** [ipdf/CMakeFiles/ipdf-anim.dir/all] Error 2\n}}}\nSeems to be related to the latest commits to ipdf.\n", "reporter": "hwissing", "cc": "kjmeagher", "resolution": "worksforme", "time": "2012-12-07T00:24:06", "component": "combo core", "summary": "ipdf does not build", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
ipdf does not build (Trac #455) - an attempt to build a fresh checkout of analysis trunk ends like this: ```text [ 93%] Building CXX object ipdf/CMakeFiles/ipdf-benchmark.dir/private/benchmark/main.cxx.o Linking CXX shared module ../../../lib/icecube/I3Db.so Linking CXX executable ../bin/ipdf-anim CMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `main': main.cxx:(.text+0x1b0): undefined reference to `I3MediumPropertiesFile::I3MediumPropertiesFile(std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)' main.cxx:(.text+0x22f): undefined reference to `I3MediumService::I3MediumService(I3MediumProperties const&, std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)' CMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const': main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0xff): undefined reference to `I3MediumService::RECO_WAVELENGTH' main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x160): undefined reference to `I3MediumService::Absorptivity(double, double) const' main.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x271): undefined reference to `I3MediumService::InvEffScattLength(double, double) const' collect2: ld returned 1 exit status make[2]: *** [bin/ipdf-anim] Error 1 make[1]: *** [ipdf/CMakeFiles/ipdf-anim.dir/all] Error 2 ``` Seems to be related to the latest commits to ipdf. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/455">https://code.icecube.wisc.edu/projects/icecube/ticket/455</a>, reported by hwissingand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-12-07T16:23:37", "_ts": "1354897417000000", "description": "an attempt to build a fresh checkout of analysis trunk ends like this:\n{{{\n[ 93%] Building CXX object ipdf/CMakeFiles/ipdf-benchmark.dir/private/benchmark/main.cxx.o\nLinking CXX shared module ../../../lib/icecube/I3Db.so\nLinking CXX executable ../bin/ipdf-anim\nCMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `main':\nmain.cxx:(.text+0x1b0): undefined reference to `I3MediumPropertiesFile::I3MediumPropertiesFile(std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)'\nmain.cxx:(.text+0x22f): undefined reference to `I3MediumService::I3MediumService(I3MediumProperties const&, std::basic_string<char, std::char_traits<char>, std::allocator<char> > const&)'\nCMakeFiles/ipdf-anim.dir/private/anim/main.cxx.o: In function `double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const':\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0xff): undefined reference to `I3MediumService::RECO_WAVELENGTH'\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x160): undefined reference to `I3MediumService::Absorptivity(double, double) const'\nmain.cxx:(.text._ZNK4IPDF6Pandel13GConvolutePEPINS0_10LayeredIceINS0_2H2EEEE9getLogPdfINS_11SimplePEHitENS_12InfiniteMuonEEEdRKT_RKT0_[double IPDF::Pandel::GConvolutePEP<IPDF::Pandel::LayeredIce<IPDF::Pandel::H2> >::getLogPdf<IPDF::SimplePEHit, IPDF::InfiniteMuon>(IPDF::SimplePEHit const&, IPDF::InfiniteMuon const&) const]+0x271): undefined reference to `I3MediumService::InvEffScattLength(double, double) const'\ncollect2: ld returned 1 exit status\nmake[2]: *** [bin/ipdf-anim] Error 1\nmake[1]: *** [ipdf/CMakeFiles/ipdf-anim.dir/all] Error 2\n}}}\nSeems to be related to the latest commits to ipdf.\n", "reporter": "hwissing", "cc": "kjmeagher", "resolution": "worksforme", "time": "2012-12-07T00:24:06", "component": "combo core", "summary": "ipdf does not build", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
ipdf does not build trac an attempt to build a fresh checkout of analysis trunk ends like this text building cxx object ipdf cmakefiles ipdf benchmark dir private benchmark main cxx o linking cxx shared module lib icecube so linking cxx executable bin ipdf anim cmakefiles ipdf anim dir private anim main cxx o in function main main cxx text undefined reference to std basic string std allocator const main cxx text undefined reference to const std basic string std allocator const cmakefiles ipdf anim dir private anim main cxx o in function double ipdf pandel gconvolutepep getlogpdf ipdf simplepehit const ipdf infinitemuon const const main cxx text undefined reference to reco wavelength main cxx text undefined reference to absorptivity double double const main cxx text undefined reference to inveffscattlength double double const ld returned exit status make error make error seems to be related to the latest commits to ipdf migrated from json status closed changetime ts description an attempt to build a fresh checkout of analysis trunk ends like this n n building cxx object ipdf cmakefiles ipdf benchmark dir private benchmark main cxx o nlinking cxx shared module lib icecube so nlinking cxx executable bin ipdf anim ncmakefiles ipdf anim dir private anim main cxx o in function main nmain cxx text undefined reference to std basic string std allocator const nmain cxx text undefined reference to const std basic string std allocator const ncmakefiles ipdf anim dir private anim main cxx o in function double ipdf pandel gconvolutepep getlogpdf ipdf simplepehit const ipdf infinitemuon const const nmain cxx text undefined reference to reco wavelength nmain cxx text undefined reference to absorptivity double double const nmain cxx text undefined reference to inveffscattlength double double const ld returned exit status nmake error nmake error n nseems to be related to the latest commits to ipdf n reporter hwissing cc kjmeagher resolution worksforme time component combo core summary ipdf does not build priority normal keywords milestone owner nega type defect
1
7,076
2,610,325,606
IssuesEvent
2015-02-26 19:44:57
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Increase skirmish space build pad construction time
auto-migrated Priority-Medium Type-Defect
``` Very much needed change as the early game is too dependent on them. ``` ----- Original issue reported on code.google.com by `KillerHurdz@netscape.net` on 8 Jun 2011 at 3:17
1.0
Increase skirmish space build pad construction time - ``` Very much needed change as the early game is too dependent on them. ``` ----- Original issue reported on code.google.com by `KillerHurdz@netscape.net` on 8 Jun 2011 at 3:17
defect
increase skirmish space build pad construction time very much needed change as the early game is too dependent on them original issue reported on code google com by killerhurdz netscape net on jun at
1
15,684
19,847,823,638
IssuesEvent
2022-01-21 08:55:08
ooi-data/CE04OSSM-RID26-07-NUTNRB000-recovered_inst-suna_instrument_recovered
https://api.github.com/repos/ooi-data/CE04OSSM-RID26-07-NUTNRB000-recovered_inst-suna_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:55:07.736867. ## Details Flow name: `CE04OSSM-RID26-07-NUTNRB000-recovered_inst-suna_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:55:07.736867. ## Details Flow name: `CE04OSSM-RID26-07-NUTNRB000-recovered_inst-suna_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
non_defect
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst suna instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
0
796,853
28,129,230,351
IssuesEvent
2023-03-31 20:47:21
slynch8/10x
https://api.github.com/repos/slynch8/10x
reopened
10x notepad perforce integration does not work well
bug Priority 2 current B
When using 10x notepad there is no solution root folder and so P4CLIENT is not used. Because you can be interacting with many different files from different streams, the desired behavior is that the p4 settings would be pulled from the P4CONFIG for whichever file you are interacting with at that time. It seems like 10x notepad should be either connecting to perforce on a per file basis or simply on demand for actions with any given file, using the p4 settings as dictated by the P4CONFIG for that file path.
1.0
10x notepad perforce integration does not work well - When using 10x notepad there is no solution root folder and so P4CLIENT is not used. Because you can be interacting with many different files from different streams, the desired behavior is that the p4 settings would be pulled from the P4CONFIG for whichever file you are interacting with at that time. It seems like 10x notepad should be either connecting to perforce on a per file basis or simply on demand for actions with any given file, using the p4 settings as dictated by the P4CONFIG for that file path.
non_defect
notepad perforce integration does not work well when using notepad there is no solution root folder and so is not used because you can be interacting with many different files from different streams the desired behavior is that the settings would be pulled from the for whichever file you are interacting with at that time it seems like notepad should be either connecting to perforce on a per file basis or simply on demand for actions with any given file using the settings as dictated by the for that file path
0
278,785
30,702,401,939
IssuesEvent
2023-07-27 01:27:03
Nivaskumark/CVE-2020-0074-frameworks_base
https://api.github.com/repos/Nivaskumark/CVE-2020-0074-frameworks_base
reopened
CVE-2020-0227 (High) detected in baseandroid-11.0.0_r39
Mend: dependency security vulnerability
## CVE-2020-0227 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-11.0.0_r39</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0074-frameworks_base/commit/f63c00c11df9fe4c62ee2ed7d5f72e3a7ebec027">f63c00c11df9fe4c62ee2ed7d5f72e3a7ebec027</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/companion/java/com/android/server/companion/CompanionDeviceManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In onCommand of CompanionDeviceManagerService.java, there is a possible permissions bypass due to a missing permission check. This could lead to local escalation of privilege allowing background data usage or launching from the background, with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10Android ID: A-129476618 <p>Publish Date: 2020-07-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0227>CVE-2020-0227</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-07-01">https://source.android.com/security/bulletin/2020-07-01</a></p> <p>Release Date: 2020-07-07</p> <p>Fix Resolution: android-10.0.0_r37,android-9.0.0_r56,android-8.1.0_r76,android-8.0.0_r48</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0227 (High) detected in baseandroid-11.0.0_r39 - ## CVE-2020-0227 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-11.0.0_r39</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0074-frameworks_base/commit/f63c00c11df9fe4c62ee2ed7d5f72e3a7ebec027">f63c00c11df9fe4c62ee2ed7d5f72e3a7ebec027</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/companion/java/com/android/server/companion/CompanionDeviceManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In onCommand of CompanionDeviceManagerService.java, there is a possible permissions bypass due to a missing permission check. This could lead to local escalation of privilege allowing background data usage or launching from the background, with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10Android ID: A-129476618 <p>Publish Date: 2020-07-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0227>CVE-2020-0227</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-07-01">https://source.android.com/security/bulletin/2020-07-01</a></p> <p>Release Date: 2020-07-07</p> <p>Fix Resolution: android-10.0.0_r37,android-9.0.0_r56,android-8.1.0_r76,android-8.0.0_r48</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in baseandroid cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files services companion java com android server companion companiondevicemanagerservice java vulnerability details in oncommand of companiondevicemanagerservice java there is a possible permissions bypass due to a missing permission check this could lead to local escalation of privilege allowing background data usage or launching from the background with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android android android step up your open source security game with mend
0
829,508
31,881,670,623
IssuesEvent
2023-09-16 12:56:18
ubiquity/ubiquibot
https://api.github.com/repos/ubiquity/ubiquibot
opened
Update Request Enhancement
Priority: 1 (Normal) Time: <2 Hours
Occasionally I see that assignees do not take the update request message seriously and a simple fix would be if the bot passes the reply into ChatGPT to understand if it's a valid update message. If not, the bot should: 1. Say something on lines of "this is not a valid update on the task. Please provide a detailed description of any updates not visible in the commit data." 2. continue the original timer for unassignment. > Do you have any updates @seprintour? If you would like to release the bounty back to the DevPool, please comment `/stop` > Last activity time: Mon Sep 11 2023 18:41:07 GMT+0000 (Coordinated Universal Time) 📌 _Originally posted by @seprintour in https://github.com/ubiquity/ubiquibot/issues/743#issuecomment-1722006059_
1.0
Update Request Enhancement - Occasionally I see that assignees do not take the update request message seriously and a simple fix would be if the bot passes the reply into ChatGPT to understand if it's a valid update message. If not, the bot should: 1. Say something on lines of "this is not a valid update on the task. Please provide a detailed description of any updates not visible in the commit data." 2. continue the original timer for unassignment. > Do you have any updates @seprintour? If you would like to release the bounty back to the DevPool, please comment `/stop` > Last activity time: Mon Sep 11 2023 18:41:07 GMT+0000 (Coordinated Universal Time) 📌 _Originally posted by @seprintour in https://github.com/ubiquity/ubiquibot/issues/743#issuecomment-1722006059_
non_defect
update request enhancement occasionally i see that assignees do not take the update request message seriously and a simple fix would be if the bot passes the reply into chatgpt to understand if it s a valid update message if not the bot should say something on lines of this is not a valid update on the task please provide a detailed description of any updates not visible in the commit data continue the original timer for unassignment do you have any updates seprintour if you would like to release the bounty back to the devpool please comment stop last activity time mon sep gmt coordinated universal time 📌 originally posted by seprintour in
0
510,620
14,813,255,226
IssuesEvent
2021-01-14 01:34:44
microsoft/terminal
https://api.github.com/repos/microsoft/terminal
closed
Polish Settings UI: Global Settings polish
Area-Settings UI In-PR Issue-Task Priority-3 Product-Terminal
- [ ] Launch size can be negative - [ ] reorder appearance to always show tabs --> hide title bar --> show terminal title in title bar
1.0
Polish Settings UI: Global Settings polish - - [ ] Launch size can be negative - [ ] reorder appearance to always show tabs --> hide title bar --> show terminal title in title bar
non_defect
polish settings ui global settings polish launch size can be negative reorder appearance to always show tabs hide title bar show terminal title in title bar
0
1,273
2,603,742,539
IssuesEvent
2015-02-24 17:41:35
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
AIModule::onUnitDestroy is not called for units inside a dropship when the dropship gets destroyed
auto-migrated Priority-Low Type-Defect
``` When a dropship is destroyed, all units loaded into it also get destroyed, but onUnitDestroy is only called for the dropship. ``` ----- Original issue reported on code.google.com by `lowerlo...@gmail.com` on 17 Nov 2009 at 5:06
1.0
AIModule::onUnitDestroy is not called for units inside a dropship when the dropship gets destroyed - ``` When a dropship is destroyed, all units loaded into it also get destroyed, but onUnitDestroy is only called for the dropship. ``` ----- Original issue reported on code.google.com by `lowerlo...@gmail.com` on 17 Nov 2009 at 5:06
defect
aimodule onunitdestroy is not called for units inside a dropship when the dropship gets destroyed when a dropship is destroyed all units loaded into it also get destroyed but onunitdestroy is only called for the dropship original issue reported on code google com by lowerlo gmail com on nov at
1
30,215
6,043,332,792
IssuesEvent
2017-06-11 20:55:14
opencaching/opencaching-pl
https://api.github.com/repos/opencaching/opencaching-pl
closed
OKAPI GPX: Some settings are not properly saved/restored
Type_Defect
2017-06-10 16:08 GMT+02:00 Stanisław Pelczar <pelczar.stanislaw@gmail.com>: > Czyżby pojawił się problem z zapamiętywaniem ustawień po spolszczeniu? > Przykładowo: > Nie zapamiętuje "Pokaż moje osobiste notatki bezpośrednio w opisie > skrzynki." > Nie zapamiętuje "W jaki sposób dołączyć obrazki/zdjęcia?" > Nie zapamiętuje "Wpisy do logów:"
1.0
OKAPI GPX: Some settings are not properly saved/restored - 2017-06-10 16:08 GMT+02:00 Stanisław Pelczar <pelczar.stanislaw@gmail.com>: > Czyżby pojawił się problem z zapamiętywaniem ustawień po spolszczeniu? > Przykładowo: > Nie zapamiętuje "Pokaż moje osobiste notatki bezpośrednio w opisie > skrzynki." > Nie zapamiętuje "W jaki sposób dołączyć obrazki/zdjęcia?" > Nie zapamiętuje "Wpisy do logów:"
defect
okapi gpx some settings are not properly saved restored gmt stanisław pelczar czyżby pojawił się problem z zapamiętywaniem ustawień po spolszczeniu przykładowo nie zapamiętuje pokaż moje osobiste notatki bezpośrednio w opisie skrzynki nie zapamiętuje w jaki sposób dołączyć obrazki zdjęcia nie zapamiętuje wpisy do logów
1
8,054
2,611,450,347
IssuesEvent
2015-02-27 04:58:46
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
LAN server broadcast problems
auto-migrated Priority-Medium Type-Defect
``` I hosted a small LAN party with hedgewars 0.9.13 recently, and uncovered some weirdness regarding the ability of people to join games. This might be only one bug with multiple symptoms, or multiple bugs: - once a game has been started, when going back to the lobby, new users cannot find the server (it doesn't show up). It seems to work if you manually enter the IP address of the server, however. - sometimes users can't join the server at all (don't remember if there's an error message), and then the only way to make it work is to have everybody (including the server) exit worms and start over again. ``` Original issue reported on code.google.com by `nekoh...@gmail.com` on 31 Aug 2010 at 4:36
1.0
LAN server broadcast problems - ``` I hosted a small LAN party with hedgewars 0.9.13 recently, and uncovered some weirdness regarding the ability of people to join games. This might be only one bug with multiple symptoms, or multiple bugs: - once a game has been started, when going back to the lobby, new users cannot find the server (it doesn't show up). It seems to work if you manually enter the IP address of the server, however. - sometimes users can't join the server at all (don't remember if there's an error message), and then the only way to make it work is to have everybody (including the server) exit worms and start over again. ``` Original issue reported on code.google.com by `nekoh...@gmail.com` on 31 Aug 2010 at 4:36
defect
lan server broadcast problems i hosted a small lan party with hedgewars recently and uncovered some weirdness regarding the ability of people to join games this might be only one bug with multiple symptoms or multiple bugs once a game has been started when going back to the lobby new users cannot find the server it doesn t show up it seems to work if you manually enter the ip address of the server however sometimes users can t join the server at all don t remember if there s an error message and then the only way to make it work is to have everybody including the server exit worms and start over again original issue reported on code google com by nekoh gmail com on aug at
1
57,945
16,187,366,964
IssuesEvent
2021-05-04 00:07:35
jccastillo0007/equus-ui
https://api.github.com/repos/jccastillo0007/equus-ui
closed
Catálogos en todo lo que es el organigrama - no permitir duplicados
Defecto Fixed
Acabo de capturar registros duplicados y si lo permite. No debe hacerlo. Me refiero al campo nombre. La descripción pues ahí no hay tanto problema.
1.0
Catálogos en todo lo que es el organigrama - no permitir duplicados - Acabo de capturar registros duplicados y si lo permite. No debe hacerlo. Me refiero al campo nombre. La descripción pues ahí no hay tanto problema.
defect
catálogos en todo lo que es el organigrama no permitir duplicados acabo de capturar registros duplicados y si lo permite no debe hacerlo me refiero al campo nombre la descripción pues ahí no hay tanto problema
1
12,032
2,678,027,365
IssuesEvent
2015-03-26 07:12:15
andialbrecht/testmig1
https://api.github.com/repos/andialbrecht/testmig1
closed
Remove "Download" link for each patch
Component-UI Priority-Medium Type-Defect
Original [issue 49](https://code.google.com/p/rietveld/issues/detail?id=49) created by andialbrecht on 2008-09-21T06:40:00.000Z: Chromium's Rietveld's instance is using about 100 MB of storage a week, and I'm looking at reducing that. One thing I'm looking at it is reusing base files when they're the same as earlier patchsets. This doesn't affect Rietveld's UI. One other idea though is to remove the &quot;Download&quot; beside each file. I don't think it's needed, since if someone wants to apply a patch they will most probably want to apply it to all the files. If someone really wants to do it to just one file, then they can edit the full patch themself. Are people ok with this change?
1.0
Remove "Download" link for each patch - Original [issue 49](https://code.google.com/p/rietveld/issues/detail?id=49) created by andialbrecht on 2008-09-21T06:40:00.000Z: Chromium's Rietveld's instance is using about 100 MB of storage a week, and I'm looking at reducing that. One thing I'm looking at it is reusing base files when they're the same as earlier patchsets. This doesn't affect Rietveld's UI. One other idea though is to remove the &quot;Download&quot; beside each file. I don't think it's needed, since if someone wants to apply a patch they will most probably want to apply it to all the files. If someone really wants to do it to just one file, then they can edit the full patch themself. Are people ok with this change?
defect
remove download link for each patch original created by andialbrecht on chromium s rietveld s instance is using about mb of storage a week and i m looking at reducing that one thing i m looking at it is reusing base files when they re the same as earlier patchsets this doesn t affect rietveld s ui one other idea though is to remove the quot download quot beside each file i don t think it s needed since if someone wants to apply a patch they will most probably want to apply it to all the files if someone really wants to do it to just one file then they can edit the full patch themself are people ok with this change
1
23,164
10,852,341,603
IssuesEvent
2019-11-13 12:36:51
ChenLuigi/TestingPOM
https://api.github.com/repos/ChenLuigi/TestingPOM
opened
CVE-2016-0763 (Medium) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar
security vulnerability
## CVE-2016-0763 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p> <details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) </details> <details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p>Core Tomcat implementation</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The setGlobalContext method in org/apache/naming/factory/ResourceLinkFactory.java in Apache Tomcat 7.x before 7.0.68, 8.x before 8.0.31, and 9.x before 9.0.0.M3 does not consider whether ResourceLinkFactory.setGlobalContext callers are authorized, which allows remote authenticated users to bypass intended SecurityManager restrictions and read or write to arbitrary application data, or cause a denial of service (application disruption), via a web application that sets a crafted global context. <p>Publish Date: 2016-02-25 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763>CVE-2016-0763</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763</a></p> <p>Release Date: 2016-02-25</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M3,8.0.32,7.0.68,org.apache.tomcat:tomcat-catalina:9.0.0.M3,8.0.32,7.0.68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-0763 (Medium) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar - ## CVE-2016-0763 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p> <details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) </details> <details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p>Core Tomcat implementation</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The setGlobalContext method in org/apache/naming/factory/ResourceLinkFactory.java in Apache Tomcat 7.x before 7.0.68, 8.x before 8.0.31, and 9.x before 9.0.0.M3 does not consider whether ResourceLinkFactory.setGlobalContext callers are authorized, which allows remote authenticated users to bypass intended SecurityManager restrictions and read or write to arbitrary application data, or cause a denial of service (application disruption), via a web application that sets a crafted global context. <p>Publish Date: 2016-02-25 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763>CVE-2016-0763</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-0763</a></p> <p>Release Date: 2016-02-25</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M3,8.0.32,7.0.68,org.apache.tomcat:tomcat-catalina:9.0.0.M3,8.0.32,7.0.68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in tomcat catalina jar tomcat embed core jar cve medium severity vulnerability vulnerable libraries tomcat catalina jar tomcat embed core jar tomcat catalina jar tomcat servlet engine core classes and standard implementations dependency hierarchy x tomcat catalina jar vulnerable library tomcat embed core jar core tomcat implementation dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href vulnerability details the setglobalcontext method in org apache naming factory resourcelinkfactory java in apache tomcat x before x before and x before does not consider whether resourcelinkfactory setglobalcontext callers are authorized which allows remote authenticated users to bypass intended securitymanager restrictions and read or write to arbitrary application data or cause a denial of service application disruption via a web application that sets a crafted global context publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
132,589
18,268,783,897
IssuesEvent
2021-10-04 11:37:49
artsking/linux-3.0.35
https://api.github.com/repos/artsking/linux-3.0.35
opened
CVE-2019-19046 (Medium) detected in linux-stable-rtv3.8.6
security vulnerability
## CVE-2019-19046 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** A memory leak in the __ipmi_bmc_register() function in drivers/char/ipmi/ipmi_msghandler.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering ida_simple_get() failure, aka CID-4aa7afb0ee20. NOTE: third parties dispute the relevance of this because an attacker cannot realistically control this failure at probe time. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19046>CVE-2019-19046</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/4aa7afb0ee20a97fbf0c5bab3df028d5fb85fdab">https://github.com/torvalds/linux/commit/4aa7afb0ee20a97fbf0c5bab3df028d5fb85fdab</a></p> <p>Release Date: 2019-10-22</p> <p>Fix Resolution: Replace or update the following file: ipmi_msghandler.c</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19046 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2019-19046 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35/commit/5992fa81c6ac1b4e9db13f5408d914525c5b7875">5992fa81c6ac1b4e9db13f5408d914525c5b7875</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_msghandler.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** A memory leak in the __ipmi_bmc_register() function in drivers/char/ipmi/ipmi_msghandler.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering ida_simple_get() failure, aka CID-4aa7afb0ee20. NOTE: third parties dispute the relevance of this because an attacker cannot realistically control this failure at probe time. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19046>CVE-2019-19046</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/4aa7afb0ee20a97fbf0c5bab3df028d5fb85fdab">https://github.com/torvalds/linux/commit/4aa7afb0ee20a97fbf0c5bab3df028d5fb85fdab</a></p> <p>Release Date: 2019-10-22</p> <p>Fix Resolution: Replace or update the following file: ipmi_msghandler.c</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers char ipmi ipmi msghandler c drivers char ipmi ipmi msghandler c drivers char ipmi ipmi msghandler c vulnerability details disputed a memory leak in the ipmi bmc register function in drivers char ipmi ipmi msghandler c in the linux kernel through allows attackers to cause a denial of service memory consumption by triggering ida simple get failure aka cid note third parties dispute the relevance of this because an attacker cannot realistically control this failure at probe time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file ipmi msghandler c step up your open source security game with whitesource
0
58,754
16,744,305,147
IssuesEvent
2021-06-11 13:46:57
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
reopened
xattr=sa is not synced to ZIL
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Distribution Version | Linux Kernel | Architecture | ZFS Version | all SPL Version | <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing xattr=dir will sync to ZIL when run sync, but xattr=sa wil not. ### Describe how to reproduce the problem ``` # zfs create pp/fs0 # zfs create -o xattr=sa pp/fs1 # cd /pp/fs0 # touch $(seq 500) && sync # cd /pp/fs1 # touch $(seq 500) && sync # cd .. # zpool sync pp # for i in $(seq 500); do setfattr -n user.test -v test fs0/$i && setfattr -n user.test -v test fs1/$i; done && sync && sudo sh -c "echo b > /proc/sysrq-trigger" (after reboot) # cd /pp/fs0 # getfattr -d * | wc -l 1500 # cd /pp/fs1 # getfattr -d * | wc -l 0 ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
xattr=sa is not synced to ZIL - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Distribution Version | Linux Kernel | Architecture | ZFS Version | all SPL Version | <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing xattr=dir will sync to ZIL when run sync, but xattr=sa wil not. ### Describe how to reproduce the problem ``` # zfs create pp/fs0 # zfs create -o xattr=sa pp/fs1 # cd /pp/fs0 # touch $(seq 500) && sync # cd /pp/fs1 # touch $(seq 500) && sync # cd .. # zpool sync pp # for i in $(seq 500); do setfattr -n user.test -v test fs0/$i && setfattr -n user.test -v test fs1/$i; done && sync && sudo sh -c "echo b > /proc/sysrq-trigger" (after reboot) # cd /pp/fs0 # getfattr -d * | wc -l 1500 # cd /pp/fs1 # getfattr -d * | wc -l 0 ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
xattr sa is not synced to zil thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information type version name distribution name distribution version linux kernel architecture zfs version all spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing xattr dir will sync to zil when run sync but xattr sa wil not describe how to reproduce the problem zfs create pp zfs create o xattr sa pp cd pp touch seq sync cd pp touch seq sync cd zpool sync pp for i in seq do setfattr n user test v test i setfattr n user test v test i done sync sudo sh c echo b proc sysrq trigger after reboot cd pp getfattr d wc l cd pp getfattr d wc l include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
59,015
17,012,702,274
IssuesEvent
2021-07-02 07:43:08
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Select.fieldsRow() returns derived table fields, not actual select fields
C: Functionality E: All Editions P: Medium T: Defect T: Incompatible change
As can be seen with this simple test case: ```java System.out.println(select(field(name("a", "b"))).fieldsRow()); System.out.println(select(field(name("a", "b"))).getSelect()); ``` The output is: ``` ("alias_119315347"."b") "a"."b" ``` The `fieldsRow()` method (and many others) calls: ```java @Override public final Row fieldsRow() { return asTable().fieldsRow(); } ``` Which produces the wrong result. There is no such `alias_119315347`. Fixing this is likely to have quite the impact. Further analysis will be required.
1.0
Select.fieldsRow() returns derived table fields, not actual select fields - As can be seen with this simple test case: ```java System.out.println(select(field(name("a", "b"))).fieldsRow()); System.out.println(select(field(name("a", "b"))).getSelect()); ``` The output is: ``` ("alias_119315347"."b") "a"."b" ``` The `fieldsRow()` method (and many others) calls: ```java @Override public final Row fieldsRow() { return asTable().fieldsRow(); } ``` Which produces the wrong result. There is no such `alias_119315347`. Fixing this is likely to have quite the impact. Further analysis will be required.
defect
select fieldsrow returns derived table fields not actual select fields as can be seen with this simple test case java system out println select field name a b fieldsrow system out println select field name a b getselect the output is alias b a b the fieldsrow method and many others calls java override public final row fieldsrow return astable fieldsrow which produces the wrong result there is no such alias fixing this is likely to have quite the impact further analysis will be required
1
424,212
12,307,803,941
IssuesEvent
2020-05-12 05:44:04
UTRS2/utrs
https://api.github.com/repos/UTRS2/utrs
opened
Show previous appeals in a table
Priority: Medium enhancement
Previous appeals by user and by ip should be listed on the appeal directly, not including itself. Searching for account appeals only has to worry about appealfor in appeals table. Searching for other IP appeals has to worry about both appealfor and hiddenip in the appeals table. It would also be prudent to search in the logs table under: `{ user = 0, objecttype = appeal, action = create, ip = {ip} }` Can I have you do this @supertassu ? I'm also going to add you to the project so we can add you to assign.
1.0
Show previous appeals in a table - Previous appeals by user and by ip should be listed on the appeal directly, not including itself. Searching for account appeals only has to worry about appealfor in appeals table. Searching for other IP appeals has to worry about both appealfor and hiddenip in the appeals table. It would also be prudent to search in the logs table under: `{ user = 0, objecttype = appeal, action = create, ip = {ip} }` Can I have you do this @supertassu ? I'm also going to add you to the project so we can add you to assign.
non_defect
show previous appeals in a table previous appeals by user and by ip should be listed on the appeal directly not including itself searching for account appeals only has to worry about appealfor in appeals table searching for other ip appeals has to worry about both appealfor and hiddenip in the appeals table it would also be prudent to search in the logs table under user objecttype appeal action create ip ip can i have you do this supertassu i m also going to add you to the project so we can add you to assign
0
64,176
14,657,464,115
IssuesEvent
2020-12-28 15:39:16
fu1771695yongxie/yarn
https://api.github.com/repos/fu1771695yongxie/yarn
opened
CVE-2017-1000048 (High) detected in qs-6.2.3.tgz, nodev7.10.1
security vulnerability
## CVE-2017-1000048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-6.2.3.tgz</b>, <b>nodev7.10.1</b></p></summary> <p> <details><summary><b>qs-6.2.3.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.3.tgz">https://registry.npmjs.org/qs/-/qs-6.2.3.tgz</a></p> <p>Path to dependency file: yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/package.json</p> <p>Path to vulnerable library: yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/loggly/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - karma-2.0.5.tgz (Root Library) - log4js-2.11.0.tgz - loggly-1.1.1.tgz - request-2.75.0.tgz - :x: **qs-6.2.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash. <p>Publish Date: 2017-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000048">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000048</a></p> <p>Release Date: 2017-07-17</p> <p>Fix Resolution: 6.0.4,6.1.2,6.2.3,6.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-1000048 (High) detected in qs-6.2.3.tgz, nodev7.10.1 - ## CVE-2017-1000048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-6.2.3.tgz</b>, <b>nodev7.10.1</b></p></summary> <p> <details><summary><b>qs-6.2.3.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.3.tgz">https://registry.npmjs.org/qs/-/qs-6.2.3.tgz</a></p> <p>Path to dependency file: yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/package.json</p> <p>Path to vulnerable library: yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/loggly/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - karma-2.0.5.tgz (Root Library) - log4js-2.11.0.tgz - loggly-1.1.1.tgz - request-2.75.0.tgz - :x: **qs-6.2.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash. <p>Publish Date: 2017-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000048">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000048</a></p> <p>Release Date: 2017-07-17</p> <p>Fix Resolution: 6.0.4,6.1.2,6.2.3,6.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in qs tgz cve high severity vulnerability vulnerable libraries qs tgz qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file yarn tests fixtures install install update auth no offline integrity offline mirror abab package package json path to vulnerable library yarn tests fixtures install install update auth no offline integrity offline mirror abab package node modules loggly node modules qs package json dependency hierarchy karma tgz root library tgz loggly tgz request tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
80,286
23,162,805,598
IssuesEvent
2022-07-29 19:50:50
Turing-Complete-Game/Suggestions-and-Issues
https://api.github.com/repos/Turing-Complete-Game/Suggestions-and-Issues
closed
[Bug]: programs changing in levels
bug Category: UI Severity: Moderate Build: Development
### What happened? programs seem to change around entering the sandbox ![image](https://user-images.githubusercontent.com/61679450/142742674-b7b9c0ba-b19f-46e8-9d40-53c76ed3c4ed.png) ### Version Latest ### Operating System win 10 ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[Bug]: programs changing in levels - ### What happened? programs seem to change around entering the sandbox ![image](https://user-images.githubusercontent.com/61679450/142742674-b7b9c0ba-b19f-46e8-9d40-53c76ed3c4ed.png) ### Version Latest ### Operating System win 10 ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_defect
programs changing in levels what happened programs seem to change around entering the sandbox version latest operating system win code of conduct i agree to follow this project s code of conduct
0
53,681
13,262,088,695
IssuesEvent
2020-08-20 21:04:58
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[bootcamp] Test ticket (Trac #1865)
Migrated from Trac cmake defect
This is just for illustration purposes. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1865">https://code.icecube.wisc.edu/projects/icecube/ticket/1865</a>, reported by olivasand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-10-01T08:53:31", "_ts": "1475312011603176", "description": "This is just for illustration purposes.", "reporter": "olivas", "cc": "", "resolution": "invalid", "time": "2016-10-01T08:53:05", "component": "cmake", "summary": "[bootcamp] Test ticket", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[bootcamp] Test ticket (Trac #1865) - This is just for illustration purposes. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1865">https://code.icecube.wisc.edu/projects/icecube/ticket/1865</a>, reported by olivasand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-10-01T08:53:31", "_ts": "1475312011603176", "description": "This is just for illustration purposes.", "reporter": "olivas", "cc": "", "resolution": "invalid", "time": "2016-10-01T08:53:05", "component": "cmake", "summary": "[bootcamp] Test ticket", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
test ticket trac this is just for illustration purposes migrated from json status closed changetime ts description this is just for illustration purposes reporter olivas cc resolution invalid time component cmake summary test ticket priority normal keywords milestone owner olivas type defect
1