Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
107,215
11,520,772,044
IssuesEvent
2020-02-14 15:25:10
jethrokuan/org-roam
https://api.github.com/repos/jethrokuan/org-roam
closed
Require org-roam fails with cannot find 'f file
documentation priority: high
When I try to load org-roam I get the following error Debugger entered--Lisp error: (file-missing "Cannot open load file" "No such file or directory" "f") require(f) eval-buffer(#<buffer *load*> nil "/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" nil t) ; Reading at buffer position 229 load-with-code-conversion("/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" "/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" nil t) require(org-roam) eval((require (quote org-roam)) nil) elisp--eval-last-sexp(nil) eval-last-sexp(nil) funcall-interactively(eval-last-sexp nil) call-interactively(eval-last-sexp nil nil) command-execute(eval-last-sexp) https://github.com/jethrokuan/org-roam/blob/32865d3d0cc5c647904211fb59f121580fc6e68a/org-roam.el#L12
1.0
Require org-roam fails with cannot find 'f file - When I try to load org-roam I get the following error Debugger entered--Lisp error: (file-missing "Cannot open load file" "No such file or directory" "f") require(f) eval-buffer(#<buffer *load*> nil "/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" nil t) ; Reading at buffer position 229 load-with-code-conversion("/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" "/home/ANT.AMAZON.COM/renukr/.emacs.d/lisp/org-roam/org-roam.el" nil t) require(org-roam) eval((require (quote org-roam)) nil) elisp--eval-last-sexp(nil) eval-last-sexp(nil) funcall-interactively(eval-last-sexp nil) call-interactively(eval-last-sexp nil nil) command-execute(eval-last-sexp) https://github.com/jethrokuan/org-roam/blob/32865d3d0cc5c647904211fb59f121580fc6e68a/org-roam.el#L12
non_defect
require org roam fails with cannot find f file when i try to load org roam i get the following error debugger entered lisp error file missing cannot open load file no such file or directory f require f eval buffer nil home ant amazon com renukr emacs d lisp org roam org roam el nil t reading at buffer position load with code conversion home ant amazon com renukr emacs d lisp org roam org roam el home ant amazon com renukr emacs d lisp org roam org roam el nil t require org roam eval require quote org roam nil elisp eval last sexp nil eval last sexp nil funcall interactively eval last sexp nil call interactively eval last sexp nil nil command execute eval last sexp
0
25,730
4,426,483,653
IssuesEvent
2016-08-16 18:27:56
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
opened
Bridge Compiler adds mscorlib reference
defect
### Steps To Reproduce - Create a Class Library project; - Install **Bridge** package; - Rebuild the project; - Open the output dll (with JetBrains dotPeek, for example): the dll contains reference to **mscorlib**.
1.0
Bridge Compiler adds mscorlib reference - ### Steps To Reproduce - Create a Class Library project; - Install **Bridge** package; - Rebuild the project; - Open the output dll (with JetBrains dotPeek, for example): the dll contains reference to **mscorlib**.
defect
bridge compiler adds mscorlib reference steps to reproduce create a class library project install bridge package rebuild the project open the output dll with jetbrains dotpeek for example the dll contains reference to mscorlib
1
3,835
2,610,069,876
IssuesEvent
2015-02-26 18:20:30
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
台州割包茎哪里正规
auto-migrated Priority-Medium Type-Defect
``` 台州割包茎哪里正规【台州五洲生殖医院】24小时健康咨询热 线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒 江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118� ��198及椒江一金清公交车直达枫南小区,乘坐107、105、109、112 、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:50
1.0
台州割包茎哪里正规 - ``` 台州割包茎哪里正规【台州五洲生殖医院】24小时健康咨询热 线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒 江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118� ��198及椒江一金清公交车直达枫南小区,乘坐107、105、109、112 、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:50
defect
台州割包茎哪里正规 台州割包茎哪里正规【台州五洲生殖医院】 线 微信号tzwzszyy 医院地址 台州市椒 (枫南大转盘旁)乘车线路 、 、 � �� , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
40,671
10,113,125,656
IssuesEvent
2019-07-30 16:00:57
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
Condominios - CxC - Cartera general
defect
Debe mostrar el nombre del propietario (ahora mismo muestra el nombre del inquilino), así como el id casa. Eliminar el checkbox de 60, 90 y 120, ya que aquí no aplica. <img width="336" alt="carteraGeneral" src="https://user-images.githubusercontent.com/2912775/62145554-49e5c680-b2b9-11e9-9b0e-29496912061a.PNG">
1.0
Condominios - CxC - Cartera general - Debe mostrar el nombre del propietario (ahora mismo muestra el nombre del inquilino), así como el id casa. Eliminar el checkbox de 60, 90 y 120, ya que aquí no aplica. <img width="336" alt="carteraGeneral" src="https://user-images.githubusercontent.com/2912775/62145554-49e5c680-b2b9-11e9-9b0e-29496912061a.PNG">
defect
condominios cxc cartera general debe mostrar el nombre del propietario ahora mismo muestra el nombre del inquilino así como el id casa eliminar el checkbox de y ya que aquí no aplica img width alt carterageneral src
1
24,013
3,898,756,625
IssuesEvent
2016-04-17 09:28:46
OSMBuildings/OSMBuildings
https://api.github.com/repos/OSMBuildings/OSMBuildings
opened
Fix Pointer events
DEFECT FEATURE
Pointer enter/leave is not handled properly. I.e. when starting a drag on map and move outside. Also 3-finger gesture for tilt could be added.
1.0
Fix Pointer events - Pointer enter/leave is not handled properly. I.e. when starting a drag on map and move outside. Also 3-finger gesture for tilt could be added.
defect
fix pointer events pointer enter leave is not handled properly i e when starting a drag on map and move outside also finger gesture for tilt could be added
1
71,114
23,453,047,105
IssuesEvent
2022-08-16 06:06:14
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
"Something went wrong" while joining a room from a Space page
T-Defect
### Steps to reproduce 1. Go to "Matrix Community" space page. 2. Search for "maubot" rooms. 3. Click "Join" buttons next to "Bots" subspace and "Maubot" room. 4. Wait for the joins to finish. ### Outcome #### What did you expect? I am joined to the "Bots" subspace and also to the "Maubot" room. #### What happened instead? At some point, I saw that the word "joined" appeared next to the "Bots" subspace, while the "Maubot" join was still in progress (rotating circle indicator). When I checked on the Element tab later, there was the "Something went wrong" screen. ### Operating system Gentoo Linux ### Browser information Firefox 91.11.0esr (64-bit) ### URL for webapp https://app.element.io ### Application version Element version: 1.11.2 Olm version: 3.2.12 ### Homeserver Dendrite 0.9.1 ### Will you send logs? Yes
1.0
"Something went wrong" while joining a room from a Space page - ### Steps to reproduce 1. Go to "Matrix Community" space page. 2. Search for "maubot" rooms. 3. Click "Join" buttons next to "Bots" subspace and "Maubot" room. 4. Wait for the joins to finish. ### Outcome #### What did you expect? I am joined to the "Bots" subspace and also to the "Maubot" room. #### What happened instead? At some point, I saw that the word "joined" appeared next to the "Bots" subspace, while the "Maubot" join was still in progress (rotating circle indicator). When I checked on the Element tab later, there was the "Something went wrong" screen. ### Operating system Gentoo Linux ### Browser information Firefox 91.11.0esr (64-bit) ### URL for webapp https://app.element.io ### Application version Element version: 1.11.2 Olm version: 3.2.12 ### Homeserver Dendrite 0.9.1 ### Will you send logs? Yes
defect
something went wrong while joining a room from a space page steps to reproduce go to matrix community space page search for maubot rooms click join buttons next to bots subspace and maubot room wait for the joins to finish outcome what did you expect i am joined to the bots subspace and also to the maubot room what happened instead at some point i saw that the word joined appeared next to the bots subspace while the maubot join was still in progress rotating circle indicator when i checked on the element tab later there was the something went wrong screen operating system gentoo linux browser information firefox bit url for webapp application version element version olm version homeserver dendrite will you send logs yes
1
246,650
7,895,585,455
IssuesEvent
2018-06-29 04:20:00
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Bug in depth cueing.
Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 4 - Crash / Wrong Results Support Group: Any bug version: 2.6.0
From an e-mail from Jean Favre: hello one of my users has brought to my attention a bad behavior in depth cueing. After inspecting the source code, I have corrected a typographic error, and confirmed that the depth cueing now works correctly. the typos are in avt/Plotter/avtSoftwareShader.C line 611,612 the indices used in initialization of start[3] and end[3] are 0,2,2 and 0,2,2 and should instead be 0,1,2 and 0,1,2 I trust someone can correct that on the repo. ----------------- Jean/CSCS -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 12/18/2012 07:17 pm Original update: 12/18/2012 07:41 pm Ticket number: 1285
1.0
Bug in depth cueing. - From an e-mail from Jean Favre: hello one of my users has brought to my attention a bad behavior in depth cueing. After inspecting the source code, I have corrected a typographic error, and confirmed that the depth cueing now works correctly. the typos are in avt/Plotter/avtSoftwareShader.C line 611,612 the indices used in initialization of start[3] and end[3] are 0,2,2 and 0,2,2 and should instead be 0,1,2 and 0,1,2 I trust someone can correct that on the repo. ----------------- Jean/CSCS -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 12/18/2012 07:17 pm Original update: 12/18/2012 07:41 pm Ticket number: 1285
non_defect
bug in depth cueing from an e mail from jean favre hello one of my users has brought to my attention a bad behavior in depth cueing after inspecting the source code i have corrected a typographic error and confirmed that the depth cueing now works correctly the typos are in avt plotter avtsoftwareshader c line the indices used in initialization of start and end are and and should instead be and i trust someone can correct that on the repo jean cscs redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation pm original update pm ticket number
0
422,892
12,287,799,822
IssuesEvent
2020-05-09 13:56:13
gravitational/gravity
https://api.github.com/repos/gravitational/gravity
opened
[BUG] Adding node in dashboard shows incorrect steps number
kind/bug priority/2 ui
**Describe the bug** When a node is being added the dashboard provides the steps progress. The number of steps reached is higher then the total "out of" given which is 10. The image attached shows getting to 11 out of 10 though getting to 13 was observed. ![steps adding node](https://user-images.githubusercontent.com/60704961/81475406-fa361300-91d9-11ea-816d-e29e4b1182d2.png) **To Reproduce** Have a existing gravity installation. Add a node though a join and observe the add in the dashboard as it installs. Steps followed: 1. Installed gravity:7.0.4 image and established a user admin 2. Copied the image file to another node 3. Ran sudo ./gravity join <ip> --token=<token id> --role=node See logs below **Expected behavior** The " out of #" would match to the total number of steps completed. **Environment (please complete the following information):** - OS [e.g. Redhat 7.4]: Ubuntu 16.04.04 - Gravity [e.g. 5.5.4]: 7.0.4 - Platform [e.g. Vmware, AWS]: GCP **Logs** [logsAddNode.txt](https://github.com/gravitational/gravity/files/4603545/logsAddNode.txt)
1.0
[BUG] Adding node in dashboard shows incorrect steps number - **Describe the bug** When a node is being added the dashboard provides the steps progress. The number of steps reached is higher then the total "out of" given which is 10. The image attached shows getting to 11 out of 10 though getting to 13 was observed. ![steps adding node](https://user-images.githubusercontent.com/60704961/81475406-fa361300-91d9-11ea-816d-e29e4b1182d2.png) **To Reproduce** Have a existing gravity installation. Add a node though a join and observe the add in the dashboard as it installs. Steps followed: 1. Installed gravity:7.0.4 image and established a user admin 2. Copied the image file to another node 3. Ran sudo ./gravity join <ip> --token=<token id> --role=node See logs below **Expected behavior** The " out of #" would match to the total number of steps completed. **Environment (please complete the following information):** - OS [e.g. Redhat 7.4]: Ubuntu 16.04.04 - Gravity [e.g. 5.5.4]: 7.0.4 - Platform [e.g. Vmware, AWS]: GCP **Logs** [logsAddNode.txt](https://github.com/gravitational/gravity/files/4603545/logsAddNode.txt)
non_defect
adding node in dashboard shows incorrect steps number describe the bug when a node is being added the dashboard provides the steps progress the number of steps reached is higher then the total out of given which is the image attached shows getting to out of though getting to was observed to reproduce have a existing gravity installation add a node though a join and observe the add in the dashboard as it installs steps followed installed gravity image and established a user admin copied the image file to another node ran sudo gravity join token role node see logs below expected behavior the out of would match to the total number of steps completed environment please complete the following information os ubuntu gravity platform gcp logs
0
447,776
31,722,316,632
IssuesEvent
2023-09-10 14:52:03
platers/obsidian-linter
https://api.github.com/repos/platers/obsidian-linter
opened
Doc: Compatability with obsidian properties
documentation
This is somewhere between an issue, discussion and documentation request, but with the new obsidian properties (which I really like so far!), I've been facing issues with linting. I haven't had time to investigate the problems fully, but wondering what others have experienced and what others have noticed are good settings. I've noticed that my linter settings disagree with obsidians preferred rules and they seem to be at a war with each other linting back and forth :D. This is mostly with arrays and sorting, but haven't quite nailed it down yet. ## What I Would Like Documented - what are the most compatible settings with linter yaml-rules and the new obsidian property rules - should these be set as default for the plugin? ## Possible issues - I've had cases where the frontmatter completely scrambles on saving
1.0
Doc: Compatability with obsidian properties - This is somewhere between an issue, discussion and documentation request, but with the new obsidian properties (which I really like so far!), I've been facing issues with linting. I haven't had time to investigate the problems fully, but wondering what others have experienced and what others have noticed are good settings. I've noticed that my linter settings disagree with obsidians preferred rules and they seem to be at a war with each other linting back and forth :D. This is mostly with arrays and sorting, but haven't quite nailed it down yet. ## What I Would Like Documented - what are the most compatible settings with linter yaml-rules and the new obsidian property rules - should these be set as default for the plugin? ## Possible issues - I've had cases where the frontmatter completely scrambles on saving
non_defect
doc compatability with obsidian properties this is somewhere between an issue discussion and documentation request but with the new obsidian properties which i really like so far i ve been facing issues with linting i haven t had time to investigate the problems fully but wondering what others have experienced and what others have noticed are good settings i ve noticed that my linter settings disagree with obsidians preferred rules and they seem to be at a war with each other linting back and forth d this is mostly with arrays and sorting but haven t quite nailed it down yet what i would like documented what are the most compatible settings with linter yaml rules and the new obsidian property rules should these be set as default for the plugin possible issues i ve had cases where the frontmatter completely scrambles on saving
0
3,232
2,537,518,802
IssuesEvent
2015-01-26 21:08:52
web2py/web2py
https://api.github.com/repos/web2py/web2py
opened
Python Crashes When Trying To Get The Scheduler To Run Alongside The Web Server.
2–5 stars bug imported Priority-Medium
_From [gheit..._at_gmail.com](https://code.google.com/u/110741554702776491750/) on June 19, 2014 22:52:40_ What steps will reproduce the problem? 1. nohup python web2py.py -K asdf12 -X --nogui -p 21000 -i 127.0.0.1 -a '' & What is the expected output? What do you see instead? What is expected is to have my web server and queued scheduled task to run. The web server runs, but my queued scheduled task never get's assigned a worker. What version of the product are you using? On what operating system? Web2py 2.9.5-stable+timestamp.2014.03.16.02.35.39 Mac OS X 10.9.3 (Mavericks) Python Version 2.7.5 Please provide any additional information below. I'm trying to get the scheduler to run alongside the web server. The OS alerts me that that Python quit unexpectedly and dumps the details to the screen. The web server will run none-the-less just fine. It's the worker that never gets started. I'm not sure what the -X is doing, but I think it has something to do with it. Main thing that stands out in the OS dump is the following below: *** multi-threaded process forked *** crashed on child side of fork pre-exec Thread 0 Crashed Everything works fine if I startup the web server and scheduler separately though. _Original issue: http://code.google.com/p/web2py/issues/detail?id=1945_
1.0
Python Crashes When Trying To Get The Scheduler To Run Alongside The Web Server. - _From [gheit..._at_gmail.com](https://code.google.com/u/110741554702776491750/) on June 19, 2014 22:52:40_ What steps will reproduce the problem? 1. nohup python web2py.py -K asdf12 -X --nogui -p 21000 -i 127.0.0.1 -a '' & What is the expected output? What do you see instead? What is expected is to have my web server and queued scheduled task to run. The web server runs, but my queued scheduled task never get's assigned a worker. What version of the product are you using? On what operating system? Web2py 2.9.5-stable+timestamp.2014.03.16.02.35.39 Mac OS X 10.9.3 (Mavericks) Python Version 2.7.5 Please provide any additional information below. I'm trying to get the scheduler to run alongside the web server. The OS alerts me that that Python quit unexpectedly and dumps the details to the screen. The web server will run none-the-less just fine. It's the worker that never gets started. I'm not sure what the -X is doing, but I think it has something to do with it. Main thing that stands out in the OS dump is the following below: *** multi-threaded process forked *** crashed on child side of fork pre-exec Thread 0 Crashed Everything works fine if I startup the web server and scheduler separately though. _Original issue: http://code.google.com/p/web2py/issues/detail?id=1945_
non_defect
python crashes when trying to get the scheduler to run alongside the web server from on june what steps will reproduce the problem nohup python py k x nogui p i a what is the expected output what do you see instead what is expected is to have my web server and queued scheduled task to run the web server runs but my queued scheduled task never get s assigned a worker what version of the product are you using on what operating system stable timestamp mac os x mavericks python version please provide any additional information below i m trying to get the scheduler to run alongside the web server the os alerts me that that python quit unexpectedly and dumps the details to the screen the web server will run none the less just fine it s the worker that never gets started i m not sure what the x is doing but i think it has something to do with it main thing that stands out in the os dump is the following below multi threaded process forked crashed on child side of fork pre exec thread crashed everything works fine if i startup the web server and scheduler separately though original issue
0
205,607
15,649,729,615
IssuesEvent
2021-03-23 08:01:25
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
reopened
`thttpclient_ssl.nim`, `thttpclient_ssl_remotenetwork.nim` fails on linux: genericDeepCopyAux SIGSEGV
Medium Priority OS/Arch specific Stdlib Test suite
tests/untestable/thttpclient_ssl.nim fails on linux i386 (EDIT: also amd64) genericDeepCopyAux SIGSEGV ### Example nim r -d:ssl --threads:on tests/untestable/thttpclient_ssl.nim ### Current Output in CI on linux i386 (found while working on https://github.com/nim-lang/Nim/pull/16221): ``` 2020-12-12T08:08:45.1952425Z Output: 2020-12-12T08:08:45.1952654Z 2020-12-12T08:08:45.1953144Z [Suite] SSL certificate check - disabled 2020-12-12T08:08:45.1953433Z 2020-12-12T08:08:45.1953924Z [Suite] SSL certificate check - httpclient 2020-12-12T08:08:45.1954835Z no-common-name (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1956590Z no-subject (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1957540Z incomplete-chain (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1957928Z 2020-12-12T08:08:45.1958318Z [Suite] SSL certificate check - httpclient - threaded 2020-12-12T08:08:45.1959068Z no-common-name (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1959968Z no-subject (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1960898Z incomplete-chain (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1961382Z Traceback (most recent call last) 2020-12-12T08:08:45.1961633Z thttpclient_ssl.nim(173) thttpclient_ssl 2020-12-12T08:08:45.1961887Z threadpool.nim(263) ^ 2020-12-12T08:08:45.1962125Z deepcopy.nim(190) genericDeepCopy 2020-12-12T08:08:45.1962396Z deepcopy.nim(152) genericDeepCopyAux 2020-12-12T08:08:45.1962715Z SIGSEGV: Illegal storage access. (Attempt to read from nil?) ``` ### Expected Output works ### Additional Information devel 1.5.1 0b73106ccf983565d34654ee8b1167827b53b01a note that before https://github.com/nim-lang/Nim/pull/16221, this test wasn't tested on i386: ``` matrix: os: [ubuntu-18.04, macos-10.15, windows-2019] cpu: [amd64] ``` it's either very flaky or always fails, not sure ## links [flaky test tests/untestable/thttpclient_ssl.nim: genericDeepCopyAux SIGSEGV · Issue #410 · timotheecour/Nim](https://github.com/timotheecour/Nim/issues/410) [thttpclient_ssl flaky test on Linux_i386 · Issue #448 · timotheecour/Nim](https://github.com/timotheecour/Nim/issues/448)
1.0
`thttpclient_ssl.nim`, `thttpclient_ssl_remotenetwork.nim` fails on linux: genericDeepCopyAux SIGSEGV - tests/untestable/thttpclient_ssl.nim fails on linux i386 (EDIT: also amd64) genericDeepCopyAux SIGSEGV ### Example nim r -d:ssl --threads:on tests/untestable/thttpclient_ssl.nim ### Current Output in CI on linux i386 (found while working on https://github.com/nim-lang/Nim/pull/16221): ``` 2020-12-12T08:08:45.1952425Z Output: 2020-12-12T08:08:45.1952654Z 2020-12-12T08:08:45.1953144Z [Suite] SSL certificate check - disabled 2020-12-12T08:08:45.1953433Z 2020-12-12T08:08:45.1953924Z [Suite] SSL certificate check - httpclient 2020-12-12T08:08:45.1954835Z no-common-name (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1956590Z no-subject (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1957540Z incomplete-chain (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1957928Z 2020-12-12T08:08:45.1958318Z [Suite] SSL certificate check - httpclient - threaded 2020-12-12T08:08:45.1959068Z no-common-name (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1959968Z no-subject (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1960898Z incomplete-chain (dubious_broken) raised: error:14090086:SSL routines:ssl3_get_server_certificate:certificate verify failed 2020-12-12T08:08:45.1961382Z Traceback (most recent call last) 2020-12-12T08:08:45.1961633Z thttpclient_ssl.nim(173) thttpclient_ssl 2020-12-12T08:08:45.1961887Z threadpool.nim(263) ^ 2020-12-12T08:08:45.1962125Z deepcopy.nim(190) genericDeepCopy 2020-12-12T08:08:45.1962396Z deepcopy.nim(152) genericDeepCopyAux 2020-12-12T08:08:45.1962715Z SIGSEGV: Illegal storage access. (Attempt to read from nil?) ``` ### Expected Output works ### Additional Information devel 1.5.1 0b73106ccf983565d34654ee8b1167827b53b01a note that before https://github.com/nim-lang/Nim/pull/16221, this test wasn't tested on i386: ``` matrix: os: [ubuntu-18.04, macos-10.15, windows-2019] cpu: [amd64] ``` it's either very flaky or always fails, not sure ## links [flaky test tests/untestable/thttpclient_ssl.nim: genericDeepCopyAux SIGSEGV · Issue #410 · timotheecour/Nim](https://github.com/timotheecour/Nim/issues/410) [thttpclient_ssl flaky test on Linux_i386 · Issue #448 · timotheecour/Nim](https://github.com/timotheecour/Nim/issues/448)
non_defect
thttpclient ssl nim thttpclient ssl remotenetwork nim fails on linux genericdeepcopyaux sigsegv tests untestable thttpclient ssl nim fails on linux edit also genericdeepcopyaux sigsegv example nim r d ssl threads on tests untestable thttpclient ssl nim current output in ci on linux found while working on output ssl certificate check disabled ssl certificate check httpclient no common name dubious broken raised error ssl routines get server certificate certificate verify failed no subject dubious broken raised error ssl routines get server certificate certificate verify failed incomplete chain dubious broken raised error ssl routines get server certificate certificate verify failed ssl certificate check httpclient threaded no common name dubious broken raised error ssl routines get server certificate certificate verify failed no subject dubious broken raised error ssl routines get server certificate certificate verify failed incomplete chain dubious broken raised error ssl routines get server certificate certificate verify failed traceback most recent call last thttpclient ssl nim thttpclient ssl threadpool nim deepcopy nim genericdeepcopy deepcopy nim genericdeepcopyaux sigsegv illegal storage access attempt to read from nil expected output works additional information devel note that before this test wasn t tested on matrix os cpu it s either very flaky or always fails not sure links
0
41,146
2,868,980,042
IssuesEvent
2015-06-05 22:20:55
dart-lang/pub
https://api.github.com/repos/dart-lang/pub
closed
Support all package sources in "pub global activate"
enhancement Fixed Priority-Medium
<a href="https://github.com/seaneagan"><img src="https://avatars.githubusercontent.com/u/444270?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [seaneagan](https://github.com/seaneagan)** _Originally opened as dart-lang/sdk#19902_ ---- Currently we have: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate &lt;package&gt; [&lt;version&gt;] which is limited to binaries from &quot;hosted&quot; packages. &quot;path&quot; and &quot;git&quot; packages are needed as well: &nbsp;&nbsp;&nbsp;&nbsp;https://www.dartlang.org/tools/pub/dependencies.html#dependency-sources Seems like the most extensible way to do this would be to mimic how they are specified in pubspec.yaml: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate &lt;package&gt; [&lt;yaml snippet&gt;] Examples: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;&gt;=1.0 &lt;2.0&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;path: /path/to/foo&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;git: {url: 'git://...', ref: '...'}&quot; Suggesting yaml since there is no standard way to specify deep object literals via command line args. Alternatively just add dedicated args for each package source type: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --path /path/to/foo &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --git.url git://... --git.ref ... &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --version &quot;&gt;=1.0 &lt;2.0&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --hosted.url http://... --hosted.name foo --version &quot;&gt;=1.0 &lt;2.0&quot; Presumably the same command-line interface would be used for issue dart-lang/pub#982.
1.0
Support all package sources in "pub global activate" - <a href="https://github.com/seaneagan"><img src="https://avatars.githubusercontent.com/u/444270?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [seaneagan](https://github.com/seaneagan)** _Originally opened as dart-lang/sdk#19902_ ---- Currently we have: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate &lt;package&gt; [&lt;version&gt;] which is limited to binaries from &quot;hosted&quot; packages. &quot;path&quot; and &quot;git&quot; packages are needed as well: &nbsp;&nbsp;&nbsp;&nbsp;https://www.dartlang.org/tools/pub/dependencies.html#dependency-sources Seems like the most extensible way to do this would be to mimic how they are specified in pubspec.yaml: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate &lt;package&gt; [&lt;yaml snippet&gt;] Examples: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;&gt;=1.0 &lt;2.0&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;path: /path/to/foo&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo &quot;git: {url: 'git://...', ref: '...'}&quot; Suggesting yaml since there is no standard way to specify deep object literals via command line args. Alternatively just add dedicated args for each package source type: &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --path /path/to/foo &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --git.url git://... --git.ref ... &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --version &quot;&gt;=1.0 &lt;2.0&quot; &nbsp;&nbsp;&nbsp;&nbsp;pub global activate foo --hosted.url http://... --hosted.name foo --version &quot;&gt;=1.0 &lt;2.0&quot; Presumably the same command-line interface would be used for issue dart-lang/pub#982.
non_defect
support all package sources in pub global activate issue by originally opened as dart lang sdk currently we have nbsp nbsp nbsp nbsp pub global activate lt package gt which is limited to binaries from quot hosted quot packages quot path quot and quot git quot packages are needed as well nbsp nbsp nbsp nbsp seems like the most extensible way to do this would be to mimic how they are specified in pubspec yaml nbsp nbsp nbsp nbsp pub global activate lt package gt examples nbsp nbsp nbsp nbsp pub global activate foo quot gt lt quot nbsp nbsp nbsp nbsp pub global activate foo quot path path to foo quot nbsp nbsp nbsp nbsp pub global activate foo quot git url git ref quot suggesting yaml since there is no standard way to specify deep object literals via command line args alternatively just add dedicated args for each package source type nbsp nbsp nbsp nbsp pub global activate foo path path to foo nbsp nbsp nbsp nbsp pub global activate foo git url git git ref nbsp nbsp nbsp nbsp pub global activate foo version quot gt lt quot nbsp nbsp nbsp nbsp pub global activate foo hosted url hosted name foo version quot gt lt quot presumably the same command line interface would be used for issue dart lang pub
0
284,221
8,736,703,661
IssuesEvent
2018-12-11 20:18:14
WTobor/DevAdventCalendar
https://api.github.com/repos/WTobor/DevAdventCalendar
closed
Add count of windows opened by user
enhancement high priority
After log in, above calendar, show information about windows opened by user (correct answered) / all opened windows - ex: "Twoje statystyki wyglądają następująco: 5/10 otwartych okienek".
1.0
Add count of windows opened by user - After log in, above calendar, show information about windows opened by user (correct answered) / all opened windows - ex: "Twoje statystyki wyglądają następująco: 5/10 otwartych okienek".
non_defect
add count of windows opened by user after log in above calendar show information about windows opened by user correct answered all opened windows ex twoje statystyki wyglądają następująco otwartych okienek
0
476,440
13,744,988,489
IssuesEvent
2020-10-06 01:32:40
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
[UX] Add new API - No validations in select API types: SOAP endpoint and Existing API
2.1.0 Priority/Low Type/Improvement Type/UX
**Description:** ***Not fulfilling [checklist items](https://docs.google.com/spreadsheets/d/1l6YKXSbmtykvvn_NvX6uJbXSsZvpT8jn72Qoi_FoJq8/edit#gid=1221574205):*** Error prevention - Is validations provided ***Related task:*** Create API ***Issue:*** Add new API Validations are not provided for SOAP endpoint and Existing API Swagger File Swagger URL. API Design - No validations in "import API definition" for Swagger File Swagger URL. ***Proposed Solution:*** Add file extensions validations for API Swagger File and URL validations for Swagger URL and SOAP endpoint URL **Suggested Labels:** UX, Improvement, 2.1.0 **Affected Product Version:** 2.1.0
1.0
[UX] Add new API - No validations in select API types: SOAP endpoint and Existing API - **Description:** ***Not fulfilling [checklist items](https://docs.google.com/spreadsheets/d/1l6YKXSbmtykvvn_NvX6uJbXSsZvpT8jn72Qoi_FoJq8/edit#gid=1221574205):*** Error prevention - Is validations provided ***Related task:*** Create API ***Issue:*** Add new API Validations are not provided for SOAP endpoint and Existing API Swagger File Swagger URL. API Design - No validations in "import API definition" for Swagger File Swagger URL. ***Proposed Solution:*** Add file extensions validations for API Swagger File and URL validations for Swagger URL and SOAP endpoint URL **Suggested Labels:** UX, Improvement, 2.1.0 **Affected Product Version:** 2.1.0
non_defect
add new api no validations in select api types soap endpoint and existing api description not fulfilling error prevention is validations provided related task create api issue add new api validations are not provided for soap endpoint and existing api swagger file swagger url api design no validations in import api definition for swagger file swagger url proposed solution add file extensions validations for api swagger file and url validations for swagger url and soap endpoint url suggested labels ux improvement affected product version
0
162,435
6,153,208,281
IssuesEvent
2017-06-28 09:23:25
architecture-building-systems/CEAforArcGIS
https://api.github.com/repos/architecture-building-systems/CEAforArcGIS
opened
Import Error: no module named arcpy with CEA Planning V. 2.2.a4
bug Priority 1
As described in #697, the radiation script seems to not run with an installation of CEA Planning V 2.2.a4, yielding the following error: ![image](https://user-images.githubusercontent.com/2969564/27629860-f6b125d6-5bf3-11e7-9edf-10a882edf62c.png) ![image](https://user-images.githubusercontent.com/2969564/27629872-fbc2dfa6-5bf3-11e7-82bd-57cdb5b46cb1.png) @kerembeygo, I created this as a separate issue so that I can close your original issue.
1.0
Import Error: no module named arcpy with CEA Planning V. 2.2.a4 - As described in #697, the radiation script seems to not run with an installation of CEA Planning V 2.2.a4, yielding the following error: ![image](https://user-images.githubusercontent.com/2969564/27629860-f6b125d6-5bf3-11e7-9edf-10a882edf62c.png) ![image](https://user-images.githubusercontent.com/2969564/27629872-fbc2dfa6-5bf3-11e7-82bd-57cdb5b46cb1.png) @kerembeygo, I created this as a separate issue so that I can close your original issue.
non_defect
import error no module named arcpy with cea planning v as described in the radiation script seems to not run with an installation of cea planning v yielding the following error kerembeygo i created this as a separate issue so that i can close your original issue
0
47,330
13,056,124,335
IssuesEvent
2020-07-30 03:43:53
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
pan along z-axis in glshovel (Trac #356)
Migrated from Trac defect glshovel
It would be very, very helpful if holding <SHIFT> and moving the mouse vertically would make glshovel pan along the z-axis. I am looking at low-energy events in DeepCore (see a [http://icecube.wisc.edu/~mdunkman/sample_i3files/Level2a_nugen_numu_IC79.006600.0000xx.000001.geo.processed.i3 sample event in its own i3 file]). Based on my selection criteria, there are only 20-30 DOMs hit near the center of the detector. In order to see which strings are hit, I need to zoom in quite a bit but when I do so, the lower region of the detector gets cropped out. I am able to pan in x,y using <SHIFT> and dragging the mouse left-to-right, but moving the mouse up and down just zooms in and out more. Currently, I need to contort the camera angle to be highly tilted in order to see any of the events that I care about, and it is very time-consuming and challenging to get a good angle to show what is happening. Thanks! -Matt Migrated from https://code.icecube.wisc.edu/ticket/356 ```json { "status": "closed", "changetime": "2013-08-23T16:40:48", "description": "It would be very, very helpful if holding <SHIFT> and moving the mouse vertically would make glshovel pan along the z-axis.\n\nI am looking at low-energy events in DeepCore (see a [http://icecube.wisc.edu/~mdunkman/sample_i3files/Level2a_nugen_numu_IC79.006600.0000xx.000001.geo.processed.i3 sample event in its own i3 file]). Based on my selection criteria, there are only 20-30 DOMs hit near the center of the detector. In order to see which strings are hit, I need to zoom in quite a bit but when I do so, the lower region of the detector gets cropped out.\n\nI am able to pan in x,y using <SHIFT> and dragging the mouse left-to-right, but moving the mouse up and down just zooms in and out more.\n\nCurrently, I need to contort the camera angle to be highly tilted in order to see any of the events that I care about, and it is very time-consuming and challenging to get a good angle to show what is happening.\n\nThanks! -Matt", "reporter": "dunkman", "cc": "mdunkman@gmail.com", "resolution": "fixed", "_ts": "1377276048000000", "component": "glshovel", "summary": "pan along z-axis in glshovel", "priority": "normal", "keywords": "", "time": "2012-02-23T21:16:20", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
pan along z-axis in glshovel (Trac #356) - It would be very, very helpful if holding <SHIFT> and moving the mouse vertically would make glshovel pan along the z-axis. I am looking at low-energy events in DeepCore (see a [http://icecube.wisc.edu/~mdunkman/sample_i3files/Level2a_nugen_numu_IC79.006600.0000xx.000001.geo.processed.i3 sample event in its own i3 file]). Based on my selection criteria, there are only 20-30 DOMs hit near the center of the detector. In order to see which strings are hit, I need to zoom in quite a bit but when I do so, the lower region of the detector gets cropped out. I am able to pan in x,y using <SHIFT> and dragging the mouse left-to-right, but moving the mouse up and down just zooms in and out more. Currently, I need to contort the camera angle to be highly tilted in order to see any of the events that I care about, and it is very time-consuming and challenging to get a good angle to show what is happening. Thanks! -Matt Migrated from https://code.icecube.wisc.edu/ticket/356 ```json { "status": "closed", "changetime": "2013-08-23T16:40:48", "description": "It would be very, very helpful if holding <SHIFT> and moving the mouse vertically would make glshovel pan along the z-axis.\n\nI am looking at low-energy events in DeepCore (see a [http://icecube.wisc.edu/~mdunkman/sample_i3files/Level2a_nugen_numu_IC79.006600.0000xx.000001.geo.processed.i3 sample event in its own i3 file]). Based on my selection criteria, there are only 20-30 DOMs hit near the center of the detector. In order to see which strings are hit, I need to zoom in quite a bit but when I do so, the lower region of the detector gets cropped out.\n\nI am able to pan in x,y using <SHIFT> and dragging the mouse left-to-right, but moving the mouse up and down just zooms in and out more.\n\nCurrently, I need to contort the camera angle to be highly tilted in order to see any of the events that I care about, and it is very time-consuming and challenging to get a good angle to show what is happening.\n\nThanks! -Matt", "reporter": "dunkman", "cc": "mdunkman@gmail.com", "resolution": "fixed", "_ts": "1377276048000000", "component": "glshovel", "summary": "pan along z-axis in glshovel", "priority": "normal", "keywords": "", "time": "2012-02-23T21:16:20", "milestone": "", "owner": "olivas", "type": "defect" } ```
defect
pan along z axis in glshovel trac it would be very very helpful if holding and moving the mouse vertically would make glshovel pan along the z axis i am looking at low energy events in deepcore see a based on my selection criteria there are only doms hit near the center of the detector in order to see which strings are hit i need to zoom in quite a bit but when i do so the lower region of the detector gets cropped out i am able to pan in x y using and dragging the mouse left to right but moving the mouse up and down just zooms in and out more currently i need to contort the camera angle to be highly tilted in order to see any of the events that i care about and it is very time consuming and challenging to get a good angle to show what is happening thanks matt migrated from json status closed changetime description it would be very very helpful if holding and moving the mouse vertically would make glshovel pan along the z axis n ni am looking at low energy events in deepcore see a based on my selection criteria there are only doms hit near the center of the detector in order to see which strings are hit i need to zoom in quite a bit but when i do so the lower region of the detector gets cropped out n ni am able to pan in x y using and dragging the mouse left to right but moving the mouse up and down just zooms in and out more n ncurrently i need to contort the camera angle to be highly tilted in order to see any of the events that i care about and it is very time consuming and challenging to get a good angle to show what is happening n nthanks matt reporter dunkman cc mdunkman gmail com resolution fixed ts component glshovel summary pan along z axis in glshovel priority normal keywords time milestone owner olivas type defect
1
8,302
2,611,486,532
IssuesEvent
2015-02-27 05:27:39
chrsmith/switchlist
https://api.github.com/repos/chrsmith/switchlist
closed
receive/ship indication on suggest cargos shouldn't be editable
auto-migrated Priority-Medium Type-Defect
``` Receive/ship indicator shouldn't be editable, or at least should be changeable. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 29 Sep 2012 at 7:20 * Merged into: #48
1.0
receive/ship indication on suggest cargos shouldn't be editable - ``` Receive/ship indicator shouldn't be editable, or at least should be changeable. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 29 Sep 2012 at 7:20 * Merged into: #48
defect
receive ship indication on suggest cargos shouldn t be editable receive ship indicator shouldn t be editable or at least should be changeable original issue reported on code google com by rwbowdi gmail com on sep at merged into
1
64,498
18,717,377,058
IssuesEvent
2021-11-03 07:36:46
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
TS2769: No overload matches this call.   Overload 1 of 2, '(props: InputTextProps | Readonly<InputTextProps>): InputText', gave the following error.
defect
### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours **I'm submitting a ...** (check one with "x") ``` [X] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Codesandbox Case (Bug Reports)** Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed. https://codesandbox.io/s/misty-cache-02ngq?file=/src/App.tsx **Current behavior** <!-- Describe how the bug manifests. --> TS2769: No overload matches this call.   Overload 1 of 2, '(props: InputTextProps | Readonly<InputTextProps>): InputText', gave the following error.     Type 'RefObject<InputText>' is not assignable to type 'string | (string & ((instance: HTMLInputElement | null) => void)) | (string & RefObject<HTMLInputElement>) | (RefObject<InputText> & string) | ... 6 more ... | undefined'.       Type 'RefObject<InputText>' is not assignable to type '((instance: InputText | null) => void) & RefObject<HTMLInputElement>'.         Type 'RefObject<InputText>' is not assignable to type '(instance: InputText | null) => void'.           Type 'RefObject<InputText>' provides no match for the signature '(instance: InputText | null): void'.   Overload 2 of 2, '(props: InputTextProps, context: any): InputText', gave the following error.     Type 'RefObject<InputText>' is not assignable to type 'string | (string & ((instance: HTMLInputElement | null) => void)) | (string & RefObject<HTMLInputElement>) | (RefObject<InputText> & string) | ... 6 more ... | undefined'.       Type 'RefObject<InputText>' is not assignable to type '((instance: InputText | null) => void) & RefObject<HTMLInputElement>'.         Type 'RefObject<InputText>' is not assignable to type '(instance: InputText | null) => void'. **Expected behavior** <!-- Describe what the behavior would be without the bug. --> No Type Error **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4). --> **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> Window, VS CODE, yarn, local * **React version:** <!-- Check whether this is still an issue in the most recent React version --> 16.12.0 * **PrimeReact version:** <!-- Check whether this is still an issue in the most recent PrimeReact version --> 6.5.1 * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> chrome all * **Language:** [all | TypeScript X.X | ES6/7 | ES5] * TypeScript 4.1.3
1.0
TS2769: No overload matches this call.   Overload 1 of 2, '(props: InputTextProps | Readonly<InputTextProps>): InputText', gave the following error. - ### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours **I'm submitting a ...** (check one with "x") ``` [X] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Codesandbox Case (Bug Reports)** Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed. https://codesandbox.io/s/misty-cache-02ngq?file=/src/App.tsx **Current behavior** <!-- Describe how the bug manifests. --> TS2769: No overload matches this call.   Overload 1 of 2, '(props: InputTextProps | Readonly<InputTextProps>): InputText', gave the following error.     Type 'RefObject<InputText>' is not assignable to type 'string | (string & ((instance: HTMLInputElement | null) => void)) | (string & RefObject<HTMLInputElement>) | (RefObject<InputText> & string) | ... 6 more ... | undefined'.       Type 'RefObject<InputText>' is not assignable to type '((instance: InputText | null) => void) & RefObject<HTMLInputElement>'.         Type 'RefObject<InputText>' is not assignable to type '(instance: InputText | null) => void'.           Type 'RefObject<InputText>' provides no match for the signature '(instance: InputText | null): void'.   Overload 2 of 2, '(props: InputTextProps, context: any): InputText', gave the following error.     Type 'RefObject<InputText>' is not assignable to type 'string | (string & ((instance: HTMLInputElement | null) => void)) | (string & RefObject<HTMLInputElement>) | (RefObject<InputText> & string) | ... 6 more ... | undefined'.       Type 'RefObject<InputText>' is not assignable to type '((instance: InputText | null) => void) & RefObject<HTMLInputElement>'.         Type 'RefObject<InputText>' is not assignable to type '(instance: InputText | null) => void'. **Expected behavior** <!-- Describe what the behavior would be without the bug. --> No Type Error **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4). --> **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> Window, VS CODE, yarn, local * **React version:** <!-- Check whether this is still an issue in the most recent React version --> 16.12.0 * **PrimeReact version:** <!-- Check whether this is still an issue in the most recent PrimeReact version --> 6.5.1 * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> chrome all * **Language:** [all | TypeScript X.X | ES6/7 | ES5] * TypeScript 4.1.3
defect
no overload matches this call   overload of props inputtextprops readonly inputtext gave the following error there is no guarantee in receiving an immediate response in github issue tracker if you d like to secure our response you may consider primereact pro support where support is provided within business hours i m submitting a check one with x bug report feature request support request please do not submit support request here instead see codesandbox case bug reports please fork the codesandbox below and create a case demonstrating your bug report issues without a codesandbox have much less possibility to be reviewed current behavior no overload matches this call   overload of props inputtextprops readonly inputtext gave the following error     type refobject is not assignable to type string string instance htmlinputelement null void string refobject refobject string more undefined       type refobject is not assignable to type instance inputtext null void refobject         type refobject is not assignable to type instance inputtext null void           type refobject provides no match for the signature instance inputtext null void   overload of props inputtextprops context any inputtext gave the following error     type refobject is not assignable to type string string instance htmlinputelement null void string refobject refobject string more undefined       type refobject is not assignable to type instance inputtext null void refobject         type refobject is not assignable to type instance inputtext null void expected behavior no type error minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point please tell us about your environment window vs code yarn local react version primereact version browser chrome all language typescript
1
55,891
14,762,420,536
IssuesEvent
2021-01-09 03:34:07
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Cannot import pool following VM crash (indirect-0 ONLINE)
Status: Triage Needed Type: Defect
Distribution Name | Centos Distribution Version | 7 Linux Kernel | 3.10.0-1127.10.1.el7 Architecture | x86_64 ZFS Version | kmod-zfs-0.8.4-1.el7.x86_64 SPL Version | 2.0.0-1 Following an abrupt vm reboot , the pool does not import (with -f or -F either). `` [pryzbyj@database ~]$ sudo zpool import pool: zfs id: 500727213717931933 state: FAULTED status: One or more devices contains corrupted data. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the '-f' flag. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-5E config: zfs FAULTED corrupted data indirect-0 ONLINE vdf FAULTED corrupted data vdd ONLINE vdf ONLINE vdg ONLINE vdh ONLINE indirect-6 ONLINE indirect-7 ONLINE vdi ONLINE [pryzbyj@database ~]$ sudo zpool import zfs cannot import 'zfs': one or more devices is currently unavailable ``
1.0
Cannot import pool following VM crash (indirect-0 ONLINE) - Distribution Name | Centos Distribution Version | 7 Linux Kernel | 3.10.0-1127.10.1.el7 Architecture | x86_64 ZFS Version | kmod-zfs-0.8.4-1.el7.x86_64 SPL Version | 2.0.0-1 Following an abrupt vm reboot , the pool does not import (with -f or -F either). `` [pryzbyj@database ~]$ sudo zpool import pool: zfs id: 500727213717931933 state: FAULTED status: One or more devices contains corrupted data. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the '-f' flag. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-5E config: zfs FAULTED corrupted data indirect-0 ONLINE vdf FAULTED corrupted data vdd ONLINE vdf ONLINE vdg ONLINE vdh ONLINE indirect-6 ONLINE indirect-7 ONLINE vdi ONLINE [pryzbyj@database ~]$ sudo zpool import zfs cannot import 'zfs': one or more devices is currently unavailable ``
defect
cannot import pool following vm crash indirect online distribution name centos distribution version linux kernel architecture zfs version kmod zfs spl version following an abrupt vm reboot the pool does not import with f or f either sudo zpool import pool zfs id state faulted status one or more devices contains corrupted data action the pool cannot be imported due to damaged devices or data the pool may be active on another system but can be imported using the f flag see config zfs faulted corrupted data indirect online vdf faulted corrupted data vdd online vdf online vdg online vdh online indirect online indirect online vdi online sudo zpool import zfs cannot import zfs one or more devices is currently unavailable
1
34,798
16,684,320,179
IssuesEvent
2021-06-08 06:05:18
tailscale/tailscale
https://api.github.com/repos/tailscale/tailscale
closed
MacOS BigSur 11.1 M1 Chip: Excessive CPU usage even when Tailscale is in Disconnected Mode
L2 More users OS-macos P1 Nuisance T3 Performance/Debugging
OS: MacOS/11.1 CPU: M1 Device: Mac Mini M1 16GB TailScale Version: 1.2.4 Problem: Tailscale uses ~4% of CPU in Intel Rosetta mode even when it is in disconnected mode. Expect: Tailscale should use close to 0% cpu amortized over time. To put in perspective, Android M1 (alpha) Android Emulator uses ~5% CPU idle Attached is the screenshot of above situation. ![tailscal_mac_min_m1](https://user-images.githubusercontent.com/417764/102428271-345b7300-404d-11eb-88ff-68f54c4a5782.png) .
True
MacOS BigSur 11.1 M1 Chip: Excessive CPU usage even when Tailscale is in Disconnected Mode - OS: MacOS/11.1 CPU: M1 Device: Mac Mini M1 16GB TailScale Version: 1.2.4 Problem: Tailscale uses ~4% of CPU in Intel Rosetta mode even when it is in disconnected mode. Expect: Tailscale should use close to 0% cpu amortized over time. To put in perspective, Android M1 (alpha) Android Emulator uses ~5% CPU idle Attached is the screenshot of above situation. ![tailscal_mac_min_m1](https://user-images.githubusercontent.com/417764/102428271-345b7300-404d-11eb-88ff-68f54c4a5782.png) .
non_defect
macos bigsur chip excessive cpu usage even when tailscale is in disconnected mode os macos cpu device mac mini tailscale version problem tailscale uses of cpu in intel rosetta mode even when it is in disconnected mode expect tailscale should use close to cpu amortized over time to put in perspective android alpha android emulator uses cpu idle attached is the screenshot of above situation
0
129,147
18,071,064,604
IssuesEvent
2021-09-21 03:01:32
Dima2022/JS-Demo
https://api.github.com/repos/Dima2022/JS-Demo
opened
CVE-2019-16776 (High) detected in npm-3.10.10.tgz
security vulnerability
## CVE-2019-16776 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-3.10.10.tgz</b></p></summary> <p>a package manager for JavaScript</p> <p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-3.10.10.tgz">https://registry.npmjs.org/npm/-/npm-3.10.10.tgz</a></p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - :x: **npm-3.10.10.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/d2b50a157c9dcc579fb01370d66876e9f4472962">d2b50a157c9dcc579fb01370d66876e9f4472962</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option. <p>Publish Date: 2019-12-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p> <p>Release Date: 2019-12-13</p> <p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-16776","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-16776 (High) detected in npm-3.10.10.tgz - ## CVE-2019-16776 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-3.10.10.tgz</b></p></summary> <p>a package manager for JavaScript</p> <p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-3.10.10.tgz">https://registry.npmjs.org/npm/-/npm-3.10.10.tgz</a></p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - :x: **npm-3.10.10.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/d2b50a157c9dcc579fb01370d66876e9f4472962">d2b50a157c9dcc579fb01370d66876e9f4472962</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option. <p>Publish Date: 2019-12-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p> <p>Release Date: 2019-12-13</p> <p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"3.10.10","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"npm - 6.13.3;yarn - 1.21.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-16776","vulnerabilityDetails":"Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user\u0027s system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in npm tgz cve high severity vulnerability vulnerable library npm tgz a package manager for javascript library home page a href dependency hierarchy grunt npm install tgz root library x npm tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user s system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution npm yarn isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt npm install npm isminimumfixversionavailable true minimumfixversion npm yarn basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option vulnerabilityurl
0
330,315
24,255,980,006
IssuesEvent
2022-09-27 17:48:30
icflorescu/mantine-datatable
https://api.github.com/repos/icflorescu/mantine-datatable
closed
Docs website navigation regression introduced when adding the CHANGELOG
bug documentation
Docs website navigation regression introduced when adding the CHANGELOG: ![image](https://user-images.githubusercontent.com/581999/192599502-56250995-597e-4260-8fe5-0ccfdb4c544f.png)
1.0
Docs website navigation regression introduced when adding the CHANGELOG - Docs website navigation regression introduced when adding the CHANGELOG: ![image](https://user-images.githubusercontent.com/581999/192599502-56250995-597e-4260-8fe5-0ccfdb4c544f.png)
non_defect
docs website navigation regression introduced when adding the changelog docs website navigation regression introduced when adding the changelog
0
22,807
3,701,251,267
IssuesEvent
2016-02-29 12:23:33
extnet/Ext.NET
https://api.github.com/repos/extnet/Ext.NET
closed
Ext.net.Image does not properly handle ImageUrl= setting
4.x defect
A ```xml <ext:Image runat="server" ImageUrl="http://speed.ext.net/identity/extnet-logo-large-vertical.svg" /> ``` Is turning into: ```xml <img id="ext-element-4" style="border: currentColor; border-image: none; width: 1586px;" src=""> ``` So, no image is displayed at all. Reported on: [Image can not be displayed by Image field](http://forums.ext.net/showthread.php?60667).
1.0
Ext.net.Image does not properly handle ImageUrl= setting - A ```xml <ext:Image runat="server" ImageUrl="http://speed.ext.net/identity/extnet-logo-large-vertical.svg" /> ``` Is turning into: ```xml <img id="ext-element-4" style="border: currentColor; border-image: none; width: 1586px;" src=""> ``` So, no image is displayed at all. Reported on: [Image can not be displayed by Image field](http://forums.ext.net/showthread.php?60667).
defect
ext net image does not properly handle imageurl setting a xml is turning into xml so no image is displayed at all reported on
1
73,830
24,815,645,537
IssuesEvent
2022-10-25 12:58:31
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
closed
Build with XCode 14: error: Signing for "DTCoreText-Resources" requires a development team
T-Defect A-Developer-Experience S-Major O-Uncommon os:iOS16
When building with XCode 14.0.1 I get the error ``` ios-client/Pods/Pods.xcodeproj: error: Signing for "DTCoreText-Resources" requires a development team. Select a development team in the Signing & Capabilities editor. (in target 'DTCoreText-Resources' from project 'Pods') ``` To build the IPA I use the command `bundle exec fastlane app_store build_number:<your_build_number>` as mentioned in the Readme. What I tried: To set the development team ID I inserted the following part after the method `cocoapods(repo_update: true)` (after this method because it removes the team id) in the lane `build_release` ``` update_project_team( teamid: "ZWBT3HXRYW", path: "./Pods/Pods.xcodeproj", targets: "DTCoreText-Resources" ) ``` After I did that I got this error (although the method `disable_automatic_code_signing` still was used): ``` ios-client/Pods/Pods.xcodeproj: error: DTCoreText-Resources has conflicting provisioning settings. DTCoreText-Resources is automatically signed, but code signing identity Apple Distribution: Intevation GmbH (ZWBT3HXRYW) has been manually specified. Set the code signing identity value to "Apple Development" in the build settings editor, or switch to manual signing in the Signing & Capabilities editor. (in target 'DTCoreText-Resources' from project 'Pods') ``` So I disabled automatic code signing explicitely for DTCoreText-Resources with: ``` update_code_signing_settings( use_automatic_signing: false, path: "./Pods/Pods.xcodeproj", team_id: "ZWBT3HXRYW", targets: "DTCoreText-Resources" ) ``` This lead to the next error: `[!] Seems to be a very old project file format - please open your project file in a more recent version of Xcode` In [this issue](https://github.com/fastlane/fastlane/issues/9270) I read that in some cases after using `update_code_signing_settings` the property `TargetAttributes` was missing. Do you have an idea what I can do to solve this problem?
1.0
Build with XCode 14: error: Signing for "DTCoreText-Resources" requires a development team - When building with XCode 14.0.1 I get the error ``` ios-client/Pods/Pods.xcodeproj: error: Signing for "DTCoreText-Resources" requires a development team. Select a development team in the Signing & Capabilities editor. (in target 'DTCoreText-Resources' from project 'Pods') ``` To build the IPA I use the command `bundle exec fastlane app_store build_number:<your_build_number>` as mentioned in the Readme. What I tried: To set the development team ID I inserted the following part after the method `cocoapods(repo_update: true)` (after this method because it removes the team id) in the lane `build_release` ``` update_project_team( teamid: "ZWBT3HXRYW", path: "./Pods/Pods.xcodeproj", targets: "DTCoreText-Resources" ) ``` After I did that I got this error (although the method `disable_automatic_code_signing` still was used): ``` ios-client/Pods/Pods.xcodeproj: error: DTCoreText-Resources has conflicting provisioning settings. DTCoreText-Resources is automatically signed, but code signing identity Apple Distribution: Intevation GmbH (ZWBT3HXRYW) has been manually specified. Set the code signing identity value to "Apple Development" in the build settings editor, or switch to manual signing in the Signing & Capabilities editor. (in target 'DTCoreText-Resources' from project 'Pods') ``` So I disabled automatic code signing explicitely for DTCoreText-Resources with: ``` update_code_signing_settings( use_automatic_signing: false, path: "./Pods/Pods.xcodeproj", team_id: "ZWBT3HXRYW", targets: "DTCoreText-Resources" ) ``` This lead to the next error: `[!] Seems to be a very old project file format - please open your project file in a more recent version of Xcode` In [this issue](https://github.com/fastlane/fastlane/issues/9270) I read that in some cases after using `update_code_signing_settings` the property `TargetAttributes` was missing. Do you have an idea what I can do to solve this problem?
defect
build with xcode error signing for dtcoretext resources requires a development team when building with xcode i get the error ios client pods pods xcodeproj error signing for dtcoretext resources requires a development team select a development team in the signing capabilities editor in target dtcoretext resources from project pods to build the ipa i use the command bundle exec fastlane app store build number as mentioned in the readme what i tried to set the development team id i inserted the following part after the method cocoapods repo update true after this method because it removes the team id in the lane build release update project team teamid path pods pods xcodeproj targets dtcoretext resources after i did that i got this error although the method disable automatic code signing still was used ios client pods pods xcodeproj error dtcoretext resources has conflicting provisioning settings dtcoretext resources is automatically signed but code signing identity apple distribution intevation gmbh has been manually specified set the code signing identity value to apple development in the build settings editor or switch to manual signing in the signing capabilities editor in target dtcoretext resources from project pods so i disabled automatic code signing explicitely for dtcoretext resources with update code signing settings use automatic signing false path pods pods xcodeproj team id targets dtcoretext resources this lead to the next error seems to be a very old project file format please open your project file in a more recent version of xcode in i read that in some cases after using update code signing settings the property targetattributes was missing do you have an idea what i can do to solve this problem
1
22,645
3,671,464,801
IssuesEvent
2016-02-22 07:36:02
jetti777Ltd/mochiweb
https://api.github.com/repos/jetti777Ltd/mochiweb
closed
Suggestion for mochiweb_cookies:cookie
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. mochiweb_cookies:cookie("sessionid", "123345", [{path, "/"}]). What is the expected output? What do you see instead? Expected: {"Set-Cookie", "session_id=123345; Path=/"} Instead: {"Set-Cookie", "session_id=\"123345\"; Path=\"/\""} Consequence: The browser does not send the cookie correctly What version of the product are you using? On what operating system? I am using the current Subversion trunk Please provide any additional information below. I have attached a suggestion (I did not make a subversion patch file out of it, but just attached the replacement cookie()) for a modification to the cookie() function, containing the following changes: - split function into base_cookie(), cookie_expires(), cookie_secure(), cookie_domain() and cookie_path() - does not use quote() function With these changes, the cookie handling seems to work in my environment. Regards, Wei-ju Wu ``` Original issue reported on code.google.com by `we...@web.de` on 2 Jan 2008 at 8:25 Attachments: * [cookie_patch.erl](https://storage.googleapis.com/google-code-attachments/mochiweb/issue-4/comment-0/cookie_patch.erl)
1.0
Suggestion for mochiweb_cookies:cookie - ``` What steps will reproduce the problem? 1. mochiweb_cookies:cookie("sessionid", "123345", [{path, "/"}]). What is the expected output? What do you see instead? Expected: {"Set-Cookie", "session_id=123345; Path=/"} Instead: {"Set-Cookie", "session_id=\"123345\"; Path=\"/\""} Consequence: The browser does not send the cookie correctly What version of the product are you using? On what operating system? I am using the current Subversion trunk Please provide any additional information below. I have attached a suggestion (I did not make a subversion patch file out of it, but just attached the replacement cookie()) for a modification to the cookie() function, containing the following changes: - split function into base_cookie(), cookie_expires(), cookie_secure(), cookie_domain() and cookie_path() - does not use quote() function With these changes, the cookie handling seems to work in my environment. Regards, Wei-ju Wu ``` Original issue reported on code.google.com by `we...@web.de` on 2 Jan 2008 at 8:25 Attachments: * [cookie_patch.erl](https://storage.googleapis.com/google-code-attachments/mochiweb/issue-4/comment-0/cookie_patch.erl)
defect
suggestion for mochiweb cookies cookie what steps will reproduce the problem mochiweb cookies cookie sessionid what is the expected output what do you see instead expected set cookie session id path instead set cookie session id path consequence the browser does not send the cookie correctly what version of the product are you using on what operating system i am using the current subversion trunk please provide any additional information below i have attached a suggestion i did not make a subversion patch file out of it but just attached the replacement cookie for a modification to the cookie function containing the following changes split function into base cookie cookie expires cookie secure cookie domain and cookie path does not use quote function with these changes the cookie handling seems to work in my environment regards wei ju wu original issue reported on code google com by we web de on jan at attachments
1
72,397
24,103,144,678
IssuesEvent
2022-09-20 04:05:44
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Crash `Recvmsg error: Connection reset by peer (104) Trace/breakpoint trap (core dumped)`
T-Defect
### Steps to reproduce 1. Open element ``` $ element-desktop /home/user/.config/Element exists: yes /home/user/.config/Riot exists: no No update_base_url is defined: auto update is disabled Fetching translation json for locale: en_EN Changing application language to en-us,en Fetching translation json for locale: en-us Fetching translation json for locale: en Resetting the UI components after locale change Resetting the UI components after locale change Changing application language to en-us Fetching translation json for locale: en-us Resetting the UI components after locale change [141732:0919/210047.337243:ERROR:broker_posix.cc(40)] Recvmsg error: Connection reset by peer (104) Trace/breakpoint trap (core dumped) ``` 2. Leave the client opened ### Outcome #### What did you expect? No crash. #### What happened instead? Crash ### Operating system Arch Linux ### Application version `element-desktop 1.11.4-1` ### How did you install the app? aur ### Homeserver frogg.ie on docker running `matrixdotorg/synapse:latest`. Not sure how to get version from the container. ### Will you send logs? Yes
1.0
Crash `Recvmsg error: Connection reset by peer (104) Trace/breakpoint trap (core dumped)` - ### Steps to reproduce 1. Open element ``` $ element-desktop /home/user/.config/Element exists: yes /home/user/.config/Riot exists: no No update_base_url is defined: auto update is disabled Fetching translation json for locale: en_EN Changing application language to en-us,en Fetching translation json for locale: en-us Fetching translation json for locale: en Resetting the UI components after locale change Resetting the UI components after locale change Changing application language to en-us Fetching translation json for locale: en-us Resetting the UI components after locale change [141732:0919/210047.337243:ERROR:broker_posix.cc(40)] Recvmsg error: Connection reset by peer (104) Trace/breakpoint trap (core dumped) ``` 2. Leave the client opened ### Outcome #### What did you expect? No crash. #### What happened instead? Crash ### Operating system Arch Linux ### Application version `element-desktop 1.11.4-1` ### How did you install the app? aur ### Homeserver frogg.ie on docker running `matrixdotorg/synapse:latest`. Not sure how to get version from the container. ### Will you send logs? Yes
defect
crash recvmsg error connection reset by peer trace breakpoint trap core dumped steps to reproduce open element element desktop home user config element exists yes home user config riot exists no no update base url is defined auto update is disabled fetching translation json for locale en en changing application language to en us en fetching translation json for locale en us fetching translation json for locale en resetting the ui components after locale change resetting the ui components after locale change changing application language to en us fetching translation json for locale en us resetting the ui components after locale change recvmsg error connection reset by peer trace breakpoint trap core dumped leave the client opened outcome what did you expect no crash what happened instead crash operating system arch linux application version element desktop how did you install the app aur homeserver frogg ie on docker running matrixdotorg synapse latest not sure how to get version from the container will you send logs yes
1
52,316
13,224,648,773
IssuesEvent
2020-08-17 19:33:41
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[cscd-llh] remove minuit dependency (Trac #2009)
Incomplete Migration Migrated from Trac analysis defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2009">https://code.icecube.wisc.edu/projects/icecube/ticket/2009</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "_ts": "1550067295757382", "description": "nobody appears to be using this functionality, all scripts use \"Powel\"", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2017-05-09T17:17:41", "component": "analysis", "summary": "[cscd-llh] remove minuit dependency", "priority": "minor", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
[cscd-llh] remove minuit dependency (Trac #2009) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2009">https://code.icecube.wisc.edu/projects/icecube/ticket/2009</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "_ts": "1550067295757382", "description": "nobody appears to be using this functionality, all scripts use \"Powel\"", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2017-05-09T17:17:41", "component": "analysis", "summary": "[cscd-llh] remove minuit dependency", "priority": "minor", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
remove minuit dependency trac migrated from json status closed changetime ts description nobody appears to be using this functionality all scripts use powel reporter kjmeagher cc resolution fixed time component analysis summary remove minuit dependency priority minor keywords milestone owner kjmeagher type defect
1
75,448
9,855,140,394
IssuesEvent
2019-06-19 18:37:31
utopia-airlines/UtopiaSearchService
https://api.github.com/repos/utopia-airlines/UtopiaSearchService
opened
Add documentation of query parameters
documentation enhancement
If this were a more traditional REST API, I would be able to figure out how to interact with this by reading the route-creation lines in the controller code, but in this service there's only one route, for `/`. So please document what query parameters are recognized, and how to use them; `README.md` would be a good place for this, but a comment above the route in question would also work.
1.0
Add documentation of query parameters - If this were a more traditional REST API, I would be able to figure out how to interact with this by reading the route-creation lines in the controller code, but in this service there's only one route, for `/`. So please document what query parameters are recognized, and how to use them; `README.md` would be a good place for this, but a comment above the route in question would also work.
non_defect
add documentation of query parameters if this were a more traditional rest api i would be able to figure out how to interact with this by reading the route creation lines in the controller code but in this service there s only one route for so please document what query parameters are recognized and how to use them readme md would be a good place for this but a comment above the route in question would also work
0
37,696
8,474,811,262
IssuesEvent
2018-10-24 17:09:44
brainvisa/testbidon
https://api.github.com/repos/brainvisa/testbidon
closed
Windows: soma openslide is not able to read images
Category: soma-io Component: Resolution Priority: Normal Status: Closed Tracker: Defect
--- Author Name: **Souedet, Nicolas** (Souedet, Nicolas) Original Redmine Issue: 17635, https://bioproj.extra.cea.fr/redmine/issues/17635 Original Date: 2017-11-28 Original Assignee: Souedet, Nicolas --- This is the consequence of a wrong use of type int64_t in header field resolutions_dimension
1.0
Windows: soma openslide is not able to read images - --- Author Name: **Souedet, Nicolas** (Souedet, Nicolas) Original Redmine Issue: 17635, https://bioproj.extra.cea.fr/redmine/issues/17635 Original Date: 2017-11-28 Original Assignee: Souedet, Nicolas --- This is the consequence of a wrong use of type int64_t in header field resolutions_dimension
defect
windows soma openslide is not able to read images author name souedet nicolas souedet nicolas original redmine issue original date original assignee souedet nicolas this is the consequence of a wrong use of type t in header field resolutions dimension
1
35,270
7,674,154,349
IssuesEvent
2018-05-15 02:12:11
cshc/cshc-web
https://api.github.com/repos/cshc/cshc-web
closed
Restrict max-width of Match Report and Blog images
defect
e.g. '/matches/2625/' Need some appropriate CSS (and probably a useful class name so we can make the CSS as generic as possible)
1.0
Restrict max-width of Match Report and Blog images - e.g. '/matches/2625/' Need some appropriate CSS (and probably a useful class name so we can make the CSS as generic as possible)
defect
restrict max width of match report and blog images e g matches need some appropriate css and probably a useful class name so we can make the css as generic as possible
1
39,929
8,705,410,516
IssuesEvent
2018-12-05 22:19:43
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
Auto-populate Git Co-Authors with list of Live Share guests
area: source control community feedback requested feature-request scenario: pair programming vs vscode
Pair programming is one of the most common use cases for Live Share, and in order to make it easier to attribute developers you’ve paired with, we should explore how to auto-populate the host’s commit messages with the list of guests they collaborated with in a Live Share session (as appropriate). Now that GitHub supports a defacto [commit convention](https://help.github.com/articles/creating-a-commit-with-multiple-authors), we could simply detect the name and emails of developers who were in a Live Share session with you, and generate the respective “co-authored-by” trailers. Functionally, this would behave similarly to [this extension](https://marketplace.visualstudio.com/items?itemName=drrouman.git-coauthors), but wouldn’t require you to manually manage your list of collaborators, or explicitly append the trailer to your commit message (which you could easily forget to do!). In general, this kind of experience could really help “close the loop” for pair programming, ensure folks on the team feel correctly attributed regardless who’s doing the commit, and keep your project history as semanticallly rich as possible. From a scoping perspective, I believe that at least initially, we would focus on enhancing the in-tool SCM experience (e.g. the “Source Control” tab in VS Code), and recommend existing solutions for developers that use the Git CLI for performing their commits (e.g. [git-mob](https://github.com/findmypast-oss/git-mob#workflow--usage), [git-duet](https://github.com/git-duet/git-duet))
1.0
Auto-populate Git Co-Authors with list of Live Share guests - Pair programming is one of the most common use cases for Live Share, and in order to make it easier to attribute developers you’ve paired with, we should explore how to auto-populate the host’s commit messages with the list of guests they collaborated with in a Live Share session (as appropriate). Now that GitHub supports a defacto [commit convention](https://help.github.com/articles/creating-a-commit-with-multiple-authors), we could simply detect the name and emails of developers who were in a Live Share session with you, and generate the respective “co-authored-by” trailers. Functionally, this would behave similarly to [this extension](https://marketplace.visualstudio.com/items?itemName=drrouman.git-coauthors), but wouldn’t require you to manually manage your list of collaborators, or explicitly append the trailer to your commit message (which you could easily forget to do!). In general, this kind of experience could really help “close the loop” for pair programming, ensure folks on the team feel correctly attributed regardless who’s doing the commit, and keep your project history as semanticallly rich as possible. From a scoping perspective, I believe that at least initially, we would focus on enhancing the in-tool SCM experience (e.g. the “Source Control” tab in VS Code), and recommend existing solutions for developers that use the Git CLI for performing their commits (e.g. [git-mob](https://github.com/findmypast-oss/git-mob#workflow--usage), [git-duet](https://github.com/git-duet/git-duet))
non_defect
auto populate git co authors with list of live share guests pair programming is one of the most common use cases for live share and in order to make it easier to attribute developers you’ve paired with we should explore how to auto populate the host’s commit messages with the list of guests they collaborated with in a live share session as appropriate now that github supports a defacto we could simply detect the name and emails of developers who were in a live share session with you and generate the respective “co authored by” trailers functionally this would behave similarly to but wouldn’t require you to manually manage your list of collaborators or explicitly append the trailer to your commit message which you could easily forget to do in general this kind of experience could really help “close the loop” for pair programming ensure folks on the team feel correctly attributed regardless who’s doing the commit and keep your project history as semanticallly rich as possible from a scoping perspective i believe that at least initially we would focus on enhancing the in tool scm experience e g the “source control” tab in vs code and recommend existing solutions for developers that use the git cli for performing their commits e g
0
4,739
2,610,154,035
IssuesEvent
2015-02-26 18:49:02
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Space Layers
auto-migrated Priority-Medium Type-Defect
``` Adjust layers for Providence and Bulwark ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:09
1.0
Space Layers - ``` Adjust layers for Providence and Bulwark ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:09
defect
space layers adjust layers for providence and bulwark original issue reported on code google com by gmail com on jan at
1
74,261
25,030,802,565
IssuesEvent
2022-11-04 12:12:21
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Pop out jitsi call does not work when openidtoken-jwt is enabled
T-Defect
### Steps to reproduce 1. enable openidtoken-jwt authentication for jitsi 2. add jitsi widget to room 3. click "Pop out" 4. "Joining" Window (jitsi.html) opens ### Outcome #### What did you expect? I can join conference #### What happened instead? When clicking on "Join Conference" it says "Failed to load Jitsi widget" ### Operating system - ### Browser information - ### URL for webapp _No response_ ### Application version 1.11.13 ### Homeserver synapse 1.70.0 ### Will you send logs? No
1.0
Pop out jitsi call does not work when openidtoken-jwt is enabled - ### Steps to reproduce 1. enable openidtoken-jwt authentication for jitsi 2. add jitsi widget to room 3. click "Pop out" 4. "Joining" Window (jitsi.html) opens ### Outcome #### What did you expect? I can join conference #### What happened instead? When clicking on "Join Conference" it says "Failed to load Jitsi widget" ### Operating system - ### Browser information - ### URL for webapp _No response_ ### Application version 1.11.13 ### Homeserver synapse 1.70.0 ### Will you send logs? No
defect
pop out jitsi call does not work when openidtoken jwt is enabled steps to reproduce enable openidtoken jwt authentication for jitsi add jitsi widget to room click pop out joining window jitsi html opens outcome what did you expect i can join conference what happened instead when clicking on join conference it says failed to load jitsi widget operating system browser information url for webapp no response application version homeserver synapse will you send logs no
1
744,042
25,925,735,274
IssuesEvent
2022-12-16 04:11:39
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
[Core] Spammable CC
Spell Core Priority: High Status: Needs Confirmation Bug Report from Discord
spammable CC such as polymorph or cyclone or entangling roots Jimmy17 OP — 12/11/2022 11:34 PM All these spells should have an error message saying ''a more power spell is already active'' if you use it on a target that is already struck by any of those CC and the active CC's duration is greater than the one that is being casted. for example i cast roots which is 8 sec duration > if i try to cast root when my 1st roo at 6 sec it shouldn't work and it sohuld say ''a more power spell is already active'' but if the 1st root is at 3 secs then my root should be casted cz the 2nd casted root is 4 sec ( of of the DR) which is higher than 3 secs.
1.0
[Core] Spammable CC - spammable CC such as polymorph or cyclone or entangling roots Jimmy17 OP — 12/11/2022 11:34 PM All these spells should have an error message saying ''a more power spell is already active'' if you use it on a target that is already struck by any of those CC and the active CC's duration is greater than the one that is being casted. for example i cast roots which is 8 sec duration > if i try to cast root when my 1st roo at 6 sec it shouldn't work and it sohuld say ''a more power spell is already active'' but if the 1st root is at 3 secs then my root should be casted cz the 2nd casted root is 4 sec ( of of the DR) which is higher than 3 secs.
non_defect
spammable cc spammable cc such as polymorph or cyclone or entangling roots op — pm all these spells should have an error message saying a more power spell is already active if you use it on a target that is already struck by any of those cc and the active cc s duration is greater than the one that is being casted for example i cast roots which is sec duration if i try to cast root when my roo at sec it shouldn t work and it sohuld say a more power spell is already active but if the root is at secs then my root should be casted cz the casted root is sec of of the dr which is higher than secs
0
68,973
22,037,314,196
IssuesEvent
2022-05-28 19:59:11
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
Faulty Elements Generated by FancyExtruderGenerator
T: defect P: normal C: Meshing
## Bug Description The current algorithm used in `FancyExtruderGenerator` may generate faulty 3D elements during extrusion. To be specify, the generated 3D elements (i.e., PRISM6, PRISM18, HEX8, and HEX27) do not follow the node indexing convention. As a result, these 3D elements have negative volume and lead to negative Jacobian. ## Steps to Reproduce This issue usually happens in two cases: 1. If a 2D element (mesh) can be extruded in (x y z) (e.g., (0 0 1)) direction without any issue, simply use (-x, -y, -z) (e.g., (0 0 -1)) direction to extrude it will lead to a faulty 3D element (mesh). 2. For a 2D mesh in XY-plane, if we use `TransformGenerator` to rotate it (0 180 0) to achieve an effective "mirroring" operation, and then stitch it to another original 2D mesh in XY-plane. The result 2D mesh cannot be extruded correctly in any direction because it contains two types of 2D elements: one can be extruded in (x y z) direction, the other can be extruded in (-x -y -z) direction. ## Impact `FancyExtruderGenerator` may generate faulty elements that cannot be used in FEM simulation without warning/error messages.
1.0
Faulty Elements Generated by FancyExtruderGenerator - ## Bug Description The current algorithm used in `FancyExtruderGenerator` may generate faulty 3D elements during extrusion. To be specify, the generated 3D elements (i.e., PRISM6, PRISM18, HEX8, and HEX27) do not follow the node indexing convention. As a result, these 3D elements have negative volume and lead to negative Jacobian. ## Steps to Reproduce This issue usually happens in two cases: 1. If a 2D element (mesh) can be extruded in (x y z) (e.g., (0 0 1)) direction without any issue, simply use (-x, -y, -z) (e.g., (0 0 -1)) direction to extrude it will lead to a faulty 3D element (mesh). 2. For a 2D mesh in XY-plane, if we use `TransformGenerator` to rotate it (0 180 0) to achieve an effective "mirroring" operation, and then stitch it to another original 2D mesh in XY-plane. The result 2D mesh cannot be extruded correctly in any direction because it contains two types of 2D elements: one can be extruded in (x y z) direction, the other can be extruded in (-x -y -z) direction. ## Impact `FancyExtruderGenerator` may generate faulty elements that cannot be used in FEM simulation without warning/error messages.
defect
faulty elements generated by fancyextrudergenerator bug description the current algorithm used in fancyextrudergenerator may generate faulty elements during extrusion to be specify the generated elements i e and do not follow the node indexing convention as a result these elements have negative volume and lead to negative jacobian steps to reproduce this issue usually happens in two cases if a element mesh can be extruded in x y z e g direction without any issue simply use x y z e g direction to extrude it will lead to a faulty element mesh for a mesh in xy plane if we use transformgenerator to rotate it to achieve an effective mirroring operation and then stitch it to another original mesh in xy plane the result mesh cannot be extruded correctly in any direction because it contains two types of elements one can be extruded in x y z direction the other can be extruded in x y z direction impact fancyextrudergenerator may generate faulty elements that cannot be used in fem simulation without warning error messages
1
241,999
7,836,479,600
IssuesEvent
2018-06-17 20:01:13
Pugabyte/BearNation
https://api.github.com/repos/Pugabyte/BearNation
closed
BlockHunt
minigames priority
LibsDisguises supports disguising as a falling block Maybe use something similar to freeze code to make players "sit"
1.0
BlockHunt - LibsDisguises supports disguising as a falling block Maybe use something similar to freeze code to make players "sit"
non_defect
blockhunt libsdisguises supports disguising as a falling block maybe use something similar to freeze code to make players sit
0
32,653
6,885,821,639
IssuesEvent
2017-11-21 17:15:04
pymc-devs/pymc3
https://api.github.com/repos/pymc-devs/pymc3
closed
sample_ppc uses a single chain only
beginner_friendly defects
Judging both by the behavior of `sample_ppc` and its implementation code I see that it samples from a single chain of a `MultiTrace`. I guess that's because indexing into such trace returns samples from a single chain only and its `len()` is equal to the number of samples per chain. Not sure if that's on purpose, but it's certainly unexpected. ## Versions and main components * PyMC3 Version: 3.1 * Theano Version: 0.9.0 * Python Version: 3.6 * Operating system: Linux * How did you install PyMC3: conda
1.0
sample_ppc uses a single chain only - Judging both by the behavior of `sample_ppc` and its implementation code I see that it samples from a single chain of a `MultiTrace`. I guess that's because indexing into such trace returns samples from a single chain only and its `len()` is equal to the number of samples per chain. Not sure if that's on purpose, but it's certainly unexpected. ## Versions and main components * PyMC3 Version: 3.1 * Theano Version: 0.9.0 * Python Version: 3.6 * Operating system: Linux * How did you install PyMC3: conda
defect
sample ppc uses a single chain only judging both by the behavior of sample ppc and its implementation code i see that it samples from a single chain of a multitrace i guess that s because indexing into such trace returns samples from a single chain only and its len is equal to the number of samples per chain not sure if that s on purpose but it s certainly unexpected versions and main components version theano version python version operating system linux how did you install conda
1
159,118
13,756,214,029
IssuesEvent
2020-10-06 19:34:47
glotzerlab/coxeter
https://api.github.com/repos/glotzerlab/coxeter
closed
Family523 S Documentation Corrections
documentation
Seems that there are some discrepancies with the documentation of the [inverse/ golden ratio](https://coxeter.readthedocs.io/en/latest/package-shape-families.html#coxeter.shape_families.Family523.s). The text is taken for S while the value and attribute name are s. And only one of the attributes is shown. It might also be good to mention in the general class doc string that these Ss are based on the golden ratio. The current doc string is correct, but does rely on recognition of that form of the well known irrational.
1.0
Family523 S Documentation Corrections - Seems that there are some discrepancies with the documentation of the [inverse/ golden ratio](https://coxeter.readthedocs.io/en/latest/package-shape-families.html#coxeter.shape_families.Family523.s). The text is taken for S while the value and attribute name are s. And only one of the attributes is shown. It might also be good to mention in the general class doc string that these Ss are based on the golden ratio. The current doc string is correct, but does rely on recognition of that form of the well known irrational.
non_defect
s documentation corrections seems that there are some discrepancies with the documentation of the the text is taken for s while the value and attribute name are s and only one of the attributes is shown it might also be good to mention in the general class doc string that these ss are based on the golden ratio the current doc string is correct but does rely on recognition of that form of the well known irrational
0
17,408
3,004,216,653
IssuesEvent
2015-07-25 18:12:53
canbican/wordpress-java
https://api.github.com/repos/canbican/wordpress-java
closed
Issues on Java 1.6 MAC OS X Snow Leopard
auto-migrated Priority-Medium Type-Defect
``` Shortly, I wasn't able to use the JAR from here, so i've compiled it on Java 1.6. Now it works, so I'm attaching it in case anybody else needs it. ``` Original issue reported on code.google.com by `vazdauta...@gmail.com` on 18 Jul 2013 at 3:48 Attachments: * [wordpress-java-1.0-SNAPSHOT.jar](https://storage.googleapis.com/google-code-attachments/wordpress-java/issue-43/comment-0/wordpress-java-1.0-SNAPSHOT.jar)
1.0
Issues on Java 1.6 MAC OS X Snow Leopard - ``` Shortly, I wasn't able to use the JAR from here, so i've compiled it on Java 1.6. Now it works, so I'm attaching it in case anybody else needs it. ``` Original issue reported on code.google.com by `vazdauta...@gmail.com` on 18 Jul 2013 at 3:48 Attachments: * [wordpress-java-1.0-SNAPSHOT.jar](https://storage.googleapis.com/google-code-attachments/wordpress-java/issue-43/comment-0/wordpress-java-1.0-SNAPSHOT.jar)
defect
issues on java mac os x snow leopard shortly i wasn t able to use the jar from here so i ve compiled it on java now it works so i m attaching it in case anybody else needs it original issue reported on code google com by vazdauta gmail com on jul at attachments
1
197,425
22,594,724,580
IssuesEvent
2022-06-29 00:47:43
interserver/mailbaby-mail-api
https://api.github.com/repos/interserver/mailbaby-mail-api
closed
CVE-2022-0639 (Medium) detected in url-parse-1.5.1.tgz - autoclosed
security vulnerability
## CVE-2022-0639 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /public/elements/package.json</p> <p>Path to vulnerable library: /public/elements/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - @stoplight/elements-demo-7.0.0-beta.1.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7. <p>Publish Date: 2022-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p> <p>Release Date: 2022-02-17</p> <p>Fix Resolution: url-parse - 1.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0639 (Medium) detected in url-parse-1.5.1.tgz - autoclosed - ## CVE-2022-0639 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /public/elements/package.json</p> <p>Path to vulnerable library: /public/elements/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - @stoplight/elements-demo-7.0.0-beta.1.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/interserver/mailbaby-mail-api/commit/f27f03f1ab403aa46166f9c03a997576599a7c43">f27f03f1ab403aa46166f9c03a997576599a7c43</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7. <p>Publish Date: 2022-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p> <p>Release Date: 2022-02-17</p> <p>Fix Resolution: url-parse - 1.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in url parse tgz autoclosed cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file public elements package json path to vulnerable library public elements node modules url parse package json dependency hierarchy stoplight elements demo beta tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with mend
0
3,919
2,610,084,062
IssuesEvent
2015-02-26 18:25:38
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳彩光嫩肤能祛痘吗
auto-migrated Priority-Medium Type-Defect
``` 深圳彩光嫩肤能祛痘吗【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:54
1.0
深圳彩光嫩肤能祛痘吗 - ``` 深圳彩光嫩肤能祛痘吗【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:54
defect
深圳彩光嫩肤能祛痘吗 深圳彩光嫩肤能祛痘吗【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 original issue reported on code google com by szft com on may at
1
295,104
9,081,606,188
IssuesEvent
2019-02-17 03:14:36
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
sanitycheck failure: [`nocache`] build failures with sdk-ng-0.10.0
bug priority: medium
``` FAILED: sam_e70_xplained/samples/net/dhcpv4_client/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/dns_resolve/test_llmnr has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/dns_resolve/test_mdns has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/gptp/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/nats/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/big_http_download/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/dumb_http_server/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo_client/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo_server/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/http_get/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/stats/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/syslog_net/test_with_timefuncs has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/syslog_net/test_without_timefuncs has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/telnet/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/vlan/test has unrecognized binary sections: ['nocache'] ```
1.0
sanitycheck failure: [`nocache`] build failures with sdk-ng-0.10.0 - ``` FAILED: sam_e70_xplained/samples/net/dhcpv4_client/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/dns_resolve/test_llmnr has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/dns_resolve/test_mdns has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/gptp/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/nats/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/big_http_download/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/dumb_http_server/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo_client/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/echo_server/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/sockets/http_get/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/stats/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/syslog_net/test_with_timefuncs has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/syslog_net/test_without_timefuncs has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/telnet/test has unrecognized binary sections: ['nocache'] FAILED: sam_e70_xplained/samples/net/vlan/test has unrecognized binary sections: ['nocache'] ```
non_defect
sanitycheck failure build failures with sdk ng failed sam xplained samples net client test has unrecognized binary sections failed sam xplained samples net dns resolve test llmnr has unrecognized binary sections failed sam xplained samples net dns resolve test mdns has unrecognized binary sections failed sam xplained samples net gptp test has unrecognized binary sections failed sam xplained samples net nats test has unrecognized binary sections failed sam xplained samples net sockets big http download test has unrecognized binary sections failed sam xplained samples net sockets dumb http server test has unrecognized binary sections failed sam xplained samples net sockets echo test has unrecognized binary sections failed sam xplained samples net sockets echo client test has unrecognized binary sections failed sam xplained samples net sockets echo server test has unrecognized binary sections failed sam xplained samples net sockets http get test has unrecognized binary sections failed sam xplained samples net stats test has unrecognized binary sections failed sam xplained samples net syslog net test with timefuncs has unrecognized binary sections failed sam xplained samples net syslog net test without timefuncs has unrecognized binary sections failed sam xplained samples net telnet test has unrecognized binary sections failed sam xplained samples net vlan test has unrecognized binary sections
0
407,315
27,610,498,766
IssuesEvent
2023-03-09 15:40:59
CarsOk/castillo_del_rey
https://api.github.com/repos/CarsOk/castillo_del_rey
opened
Actas y documentos
documentation
# Yo como Aprendiz sena # Quiero llevar a cabo la realización de las Actas y documentos de mi proyecto formativo # Para que de esta forma la información este ordenada y lista para funcionar Criterios de aceptación -[] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente. -[] la información debe ser verídica y completa.
1.0
Actas y documentos - # Yo como Aprendiz sena # Quiero llevar a cabo la realización de las Actas y documentos de mi proyecto formativo # Para que de esta forma la información este ordenada y lista para funcionar Criterios de aceptación -[] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente. -[] la información debe ser verídica y completa.
non_defect
actas y documentos yo como aprendiz sena quiero llevar a cabo la realización de las actas y documentos de mi proyecto formativo para que de esta forma la información este ordenada y lista para funcionar criterios de aceptación debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente la información debe ser verídica y completa
0
17,867
3,013,547,535
IssuesEvent
2015-07-29 09:39:15
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Nested Exception Handling
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? An example is specified through a Test case which is attached. Steps are explained there in detail. What is the expected output? What do you see instead? expected: The execution order of tasks should be: C, B and A. see: C -> B -> the A task will be halted What version of the product are you using? On what operating system? 2.1 Please provide any additional information below. Attached please find a detailed description of the problem in a microsoft Word Document. ``` Original issue reported on code.google.com by `Jalali.A...@gmail.com` on 30 Mar 2011 at 2:37 Attachments: * [3. Nested Exception handling-1.docx](https://storage.googleapis.com/google-code-attachments/yawl/issue-430/comment-0/3. Nested Exception handling-1.docx)
1.0
Nested Exception Handling - ``` What steps will reproduce the problem? An example is specified through a Test case which is attached. Steps are explained there in detail. What is the expected output? What do you see instead? expected: The execution order of tasks should be: C, B and A. see: C -> B -> the A task will be halted What version of the product are you using? On what operating system? 2.1 Please provide any additional information below. Attached please find a detailed description of the problem in a microsoft Word Document. ``` Original issue reported on code.google.com by `Jalali.A...@gmail.com` on 30 Mar 2011 at 2:37 Attachments: * [3. Nested Exception handling-1.docx](https://storage.googleapis.com/google-code-attachments/yawl/issue-430/comment-0/3. Nested Exception handling-1.docx)
defect
nested exception handling what steps will reproduce the problem an example is specified through a test case which is attached steps are explained there in detail what is the expected output what do you see instead expected the execution order of tasks should be c b and a see c b the a task will be halted what version of the product are you using on what operating system please provide any additional information below attached please find a detailed description of the problem in a microsoft word document original issue reported on code google com by jalali a gmail com on mar at attachments nested exception handling docx
1
61,925
17,023,810,900
IssuesEvent
2021-07-03 03:58:57
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Relation members won't download
Component: api Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 2.14am, Friday, 27th July 2012]** If I try to download members of relation 1593692 in JOSM I get an internal server error. It's not the size (significantly larger relations download just fine). I'm not sure if this is a JOSM or API bug, but the error message seems to indicate the latter.
1.0
Relation members won't download - **[Submitted to the original trac issue database at 2.14am, Friday, 27th July 2012]** If I try to download members of relation 1593692 in JOSM I get an internal server error. It's not the size (significantly larger relations download just fine). I'm not sure if this is a JOSM or API bug, but the error message seems to indicate the latter.
defect
relation members won t download if i try to download members of relation in josm i get an internal server error it s not the size significantly larger relations download just fine i m not sure if this is a josm or api bug but the error message seems to indicate the latter
1
142,661
13,037,299,505
IssuesEvent
2020-07-28 13:37:46
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Include a hint where to find documentation for DevTools in docs overview
status: forward-port type: documentation
Forward port of issue #22622 to 2.4.x.
1.0
Include a hint where to find documentation for DevTools in docs overview - Forward port of issue #22622 to 2.4.x.
non_defect
include a hint where to find documentation for devtools in docs overview forward port of issue to x
0
46,023
13,055,839,995
IssuesEvent
2020-07-30 02:53:31
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
trunk offline-software (cmake) testing framework does not support multiple test executables per project (Trac #453)
Incomplete Migration Migrated from Trac combo core defect
Migrated from https://code.icecube.wisc.edu/ticket/453 ```json { "status": "closed", "changetime": "2013-01-27T15:04:22", "description": "mmc-icetray and neutrino-generator use multiple \"i3_test_executable\" statements in their respective CMakeLists.txt files. This fails during the cmake stage (rebuild_cache) with current offline-software/trunk (as of r95785) with:\n{{{\nCMake Error at cmake/project.cmake:456 (add_test):\n add_test given test NAME \"mmc-icetray::unit_tests\" which already exists in\n this directory.\nCall Stack (most recent call first):\n mmc-icetray/CMakeLists.txt:19 (i3_test_executable)\n}}}\nand \n{{{\nCMake Error at cmake/project.cmake:456 (add_test):\n add_test given test NAME \"neutrino-generator::unit_tests\" which already\n exists in this directory.\nCall Stack (most recent call first):\n neutrino-generator/CMakeLists.txt:20 (i3_test_executable)\n}}}\n\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "_ts": "1359299062000000", "component": "combo core", "summary": "trunk offline-software (cmake) testing framework does not support multiple test executables per project", "priority": "normal", "keywords": "", "time": "2012-11-26T23:04:21", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
trunk offline-software (cmake) testing framework does not support multiple test executables per project (Trac #453) - Migrated from https://code.icecube.wisc.edu/ticket/453 ```json { "status": "closed", "changetime": "2013-01-27T15:04:22", "description": "mmc-icetray and neutrino-generator use multiple \"i3_test_executable\" statements in their respective CMakeLists.txt files. This fails during the cmake stage (rebuild_cache) with current offline-software/trunk (as of r95785) with:\n{{{\nCMake Error at cmake/project.cmake:456 (add_test):\n add_test given test NAME \"mmc-icetray::unit_tests\" which already exists in\n this directory.\nCall Stack (most recent call first):\n mmc-icetray/CMakeLists.txt:19 (i3_test_executable)\n}}}\nand \n{{{\nCMake Error at cmake/project.cmake:456 (add_test):\n add_test given test NAME \"neutrino-generator::unit_tests\" which already\n exists in this directory.\nCall Stack (most recent call first):\n neutrino-generator/CMakeLists.txt:20 (i3_test_executable)\n}}}\n\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "_ts": "1359299062000000", "component": "combo core", "summary": "trunk offline-software (cmake) testing framework does not support multiple test executables per project", "priority": "normal", "keywords": "", "time": "2012-11-26T23:04:21", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
trunk offline software cmake testing framework does not support multiple test executables per project trac migrated from json status closed changetime description mmc icetray and neutrino generator use multiple test executable statements in their respective cmakelists txt files this fails during the cmake stage rebuild cache with current offline software trunk as of with n ncmake error at cmake project cmake add test n add test given test name mmc icetray unit tests which already exists in n this directory ncall stack most recent call first n mmc icetray cmakelists txt test executable n nand n ncmake error at cmake project cmake add test n add test given test name neutrino generator unit tests which already n exists in this directory ncall stack most recent call first n neutrino generator cmakelists txt test executable n n n reporter claudio kopper cc resolution fixed ts component combo core summary trunk offline software cmake testing framework does not support multiple test executables per project priority normal keywords time milestone owner nega type defect
1
57,415
15,772,096,175
IssuesEvent
2021-03-31 21:21:12
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
sitewide, design system, CMS — 508-defect-2 [AXE-CORE]: Heading levels should increase by one - footer issue
508-defect-2 508-issue-headings 508/Accessibility cms components design system frontend sitewide triage vsa
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) ```diff ! Team & project: vsa-public-websites, AMA follow-on ``` **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Headings on a web page should progress, increasing by one, in sections of content. The headings in the footer are currently tagged as `<h4>`. On pages that don't have an `<h3>`, this results in an accessibility issue. Moderate Template Issue | Heading levels should only increase by one `<h4 class="va-footer-linkgroup-title>Veteran programs and services</h4>` 1. After you request a decision review — https://preview-prod.vfs.va.gov/preview?nodeId=3008 1. Supplemental Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3011 1. Higher-Level Reviews — https://preview-prod.vfs.va.gov/preview?nodeId=3003 1. Board Appeals — https://preview-prod.vfs.va.gov/preview?nodeId=3029 1. Contested Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3030 1. Insurance claims — https://preview-prod.vfs.va.gov/preview?nodeId=3025 1. FAQs — https://preview-prod.vfs.va.gov/preview?nodeId=3004 1. Get help with review request — https://preview-prod.vfs.va.gov/preview?nodeId=3009 ### Why it Matters The underlying purpose of headers is to convey the structure of the page. For sighted users, the same purpose is achieved using different sizes of text. Text size, however, is not helpful for users of screen readers, because a screen reader identifies a header only if it is properly marked-up. When heading elements are applied correctly, the page becomes much easier to navigate for screen reader users and sighted users alike. In the same way that sighted users can glance at a page and get a sense of its contents, users of screen readers can do the same by navigating through headings. Well written and properly ordered headings can save users, especially those who use screen readers, a lot of time and frustration. The purpose of headings is to describe the structure of the webpage, not just highlight important text. They should be brief, clear, unique, and marked with h1 through h6 elements applied in hierarchical order. All of these qualities make headings valuable tools for screen reader users. Similar to the way sighted users can glance at a page and get a sense of its contents, screen reader users can navigate through headings. Well written and properly ordered headings can save screen reader time and frustration. In addition to making the page more accessible, headings have other benefits since search engines use headings when filtering, ordering, and displaying results. Improving the accessibility of your site can also have the effect of making your page more findable. ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to navigate the hierarchy of the page content using heading levels to save time and frustration. ## Environment * Operating System: all * Browser: all * Screenreading device: any * Server destination: staging & production ## Steps to Recreate 1. Enter any of the above urls in browser, in the VA environment 1. Have developer tools open, and the axe browser extension loaded 1. Enter Urgent care for facility type and Community urgent care providers for service type 1. Run an axe audit 1. Verify that there is an error of "Heading levels should only increase by one" ## Possible Fixes (optional) Change the structure of the footer info lists to use a `<dl>` or change the `<h4>` to an `<h2>`. ## WCAG or Vendor Guidance (optional) * [axe-core 3.4 - Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/3.4/heading-order) * [MDN dialog element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dialog) * [MDN ARIA: alert role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/Alert_Role) * [MDN Description List element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dl)
1.0
sitewide, design system, CMS — 508-defect-2 [AXE-CORE]: Heading levels should increase by one - footer issue - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) ```diff ! Team & project: vsa-public-websites, AMA follow-on ``` **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Headings on a web page should progress, increasing by one, in sections of content. The headings in the footer are currently tagged as `<h4>`. On pages that don't have an `<h3>`, this results in an accessibility issue. Moderate Template Issue | Heading levels should only increase by one `<h4 class="va-footer-linkgroup-title>Veteran programs and services</h4>` 1. After you request a decision review — https://preview-prod.vfs.va.gov/preview?nodeId=3008 1. Supplemental Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3011 1. Higher-Level Reviews — https://preview-prod.vfs.va.gov/preview?nodeId=3003 1. Board Appeals — https://preview-prod.vfs.va.gov/preview?nodeId=3029 1. Contested Claims — https://preview-prod.vfs.va.gov/preview?nodeId=3030 1. Insurance claims — https://preview-prod.vfs.va.gov/preview?nodeId=3025 1. FAQs — https://preview-prod.vfs.va.gov/preview?nodeId=3004 1. Get help with review request — https://preview-prod.vfs.va.gov/preview?nodeId=3009 ### Why it Matters The underlying purpose of headers is to convey the structure of the page. For sighted users, the same purpose is achieved using different sizes of text. Text size, however, is not helpful for users of screen readers, because a screen reader identifies a header only if it is properly marked-up. When heading elements are applied correctly, the page becomes much easier to navigate for screen reader users and sighted users alike. In the same way that sighted users can glance at a page and get a sense of its contents, users of screen readers can do the same by navigating through headings. Well written and properly ordered headings can save users, especially those who use screen readers, a lot of time and frustration. The purpose of headings is to describe the structure of the webpage, not just highlight important text. They should be brief, clear, unique, and marked with h1 through h6 elements applied in hierarchical order. All of these qualities make headings valuable tools for screen reader users. Similar to the way sighted users can glance at a page and get a sense of its contents, screen reader users can navigate through headings. Well written and properly ordered headings can save screen reader time and frustration. In addition to making the page more accessible, headings have other benefits since search engines use headings when filtering, ordering, and displaying results. Improving the accessibility of your site can also have the effect of making your page more findable. ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to navigate the hierarchy of the page content using heading levels to save time and frustration. ## Environment * Operating System: all * Browser: all * Screenreading device: any * Server destination: staging & production ## Steps to Recreate 1. Enter any of the above urls in browser, in the VA environment 1. Have developer tools open, and the axe browser extension loaded 1. Enter Urgent care for facility type and Community urgent care providers for service type 1. Run an axe audit 1. Verify that there is an error of "Heading levels should only increase by one" ## Possible Fixes (optional) Change the structure of the footer info lists to use a `<dl>` or change the `<h4>` to an `<h2>`. ## WCAG or Vendor Guidance (optional) * [axe-core 3.4 - Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/3.4/heading-order) * [MDN dialog element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dialog) * [MDN ARIA: alert role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/Alert_Role) * [MDN Description List element](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/dl)
defect
sitewide design system cms — defect heading levels should increase by one footer issue diff team project vsa public websites ama follow on feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description headings on a web page should progress increasing by one in sections of content the headings in the footer are currently tagged as on pages that don t have an this results in an accessibility issue moderate template issue heading levels should only increase by one veteran programs and services after you request a decision review — supplemental claims — higher level reviews — board appeals — contested claims — insurance claims — faqs — get help with review request — why it matters the underlying purpose of headers is to convey the structure of the page for sighted users the same purpose is achieved using different sizes of text text size however is not helpful for users of screen readers because a screen reader identifies a header only if it is properly marked up when heading elements are applied correctly the page becomes much easier to navigate for screen reader users and sighted users alike in the same way that sighted users can glance at a page and get a sense of its contents users of screen readers can do the same by navigating through headings well written and properly ordered headings can save users especially those who use screen readers a lot of time and frustration the purpose of headings is to describe the structure of the webpage not just highlight important text they should be brief clear unique and marked with through elements applied in hierarchical order all of these qualities make headings valuable tools for screen reader users similar to the way sighted users can glance at a page and get a sense of its contents screen reader users can navigate through headings well written and properly ordered headings can save screen reader time and frustration in addition to making the page more accessible headings have other benefits since search engines use headings when filtering ordering and displaying results improving the accessibility of your site can also have the effect of making your page more findable point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to navigate the hierarchy of the page content using heading levels to save time and frustration environment operating system all browser all screenreading device any server destination staging production steps to recreate enter any of the above urls in browser in the va environment have developer tools open and the axe browser extension loaded enter urgent care for facility type and community urgent care providers for service type run an axe audit verify that there is an error of heading levels should only increase by one possible fixes optional change the structure of the footer info lists to use a or change the to an wcag or vendor guidance optional
1
341,721
24,709,843,809
IssuesEvent
2022-10-19 22:59:47
cogeotiff/rio-tiler
https://api.github.com/repos/cogeotiff/rio-tiler
closed
COGReader constructor doesn't actually take a crs parameter
resolved-by-pr documentation
Tested this in 3.1.6 and 3.1.0. The COGReader constructor says it takes a "crs" parameter for the dataset crs: > crs (rasterio.crs.CRS): Dataset CRS. However, actually passing this parameter results in an immediate error. ``` from rio_tiler.io import COGReader import rasterio with COGReader("s3://hrrrzarr/sfc/20210101/20210101_00z_anl.zarr/surface/TMP/surface/TMP", crs=rasterio.CRS.from_dict(proj_params)) as image: print(image.dataset) ``` Error: > TypeError: __init__() got an unexpected keyword argument 'crs' Without the crs parameter, COGReader can read the zarr dataset in that example, but the dataset itself doesn't contain the CRS info so I can't do much with the resulting image. Please let me know if there's a workaround to set the dataset crs.
1.0
COGReader constructor doesn't actually take a crs parameter - Tested this in 3.1.6 and 3.1.0. The COGReader constructor says it takes a "crs" parameter for the dataset crs: > crs (rasterio.crs.CRS): Dataset CRS. However, actually passing this parameter results in an immediate error. ``` from rio_tiler.io import COGReader import rasterio with COGReader("s3://hrrrzarr/sfc/20210101/20210101_00z_anl.zarr/surface/TMP/surface/TMP", crs=rasterio.CRS.from_dict(proj_params)) as image: print(image.dataset) ``` Error: > TypeError: __init__() got an unexpected keyword argument 'crs' Without the crs parameter, COGReader can read the zarr dataset in that example, but the dataset itself doesn't contain the CRS info so I can't do much with the resulting image. Please let me know if there's a workaround to set the dataset crs.
non_defect
cogreader constructor doesn t actually take a crs parameter tested this in and the cogreader constructor says it takes a crs parameter for the dataset crs crs rasterio crs crs dataset crs however actually passing this parameter results in an immediate error from rio tiler io import cogreader import rasterio with cogreader hrrrzarr sfc anl zarr surface tmp surface tmp crs rasterio crs from dict proj params as image print image dataset error typeerror init got an unexpected keyword argument crs without the crs parameter cogreader can read the zarr dataset in that example but the dataset itself doesn t contain the crs info so i can t do much with the resulting image please let me know if there s a workaround to set the dataset crs
0
634,516
20,364,239,676
IssuesEvent
2022-02-21 02:23:44
latteart-org/latteart
https://api.github.com/repos/latteart-org/latteart
opened
リモート接続した状態でインポートするとローカルにインポートされてしまう
Type: Bug Priority: Must
**Describe the bug** リモート接続した状態でインポートするとローカルにインポートされてしまう **Expected behavior** インポートのzipファイル取得はローカル固定で、インポート先は接続しているリポジトリとなる。
1.0
リモート接続した状態でインポートするとローカルにインポートされてしまう - **Describe the bug** リモート接続した状態でインポートするとローカルにインポートされてしまう **Expected behavior** インポートのzipファイル取得はローカル固定で、インポート先は接続しているリポジトリとなる。
non_defect
リモート接続した状態でインポートするとローカルにインポートされてしまう describe the bug リモート接続した状態でインポートするとローカルにインポートされてしまう expected behavior インポートのzipファイル取得はローカル固定で、インポート先は接続しているリポジトリとなる。
0
364,117
10,759,132,863
IssuesEvent
2019-10-31 16:06:13
ZTLARTCC/ZTL_website
https://api.github.com/repos/ZTLARTCC/ZTL_website
closed
Visiting Requests. Stop OBS from visiting requests
enhancement good first issue high priority
Make sure OBS ratings can’t submit requests, this should block out the noobs.
1.0
Visiting Requests. Stop OBS from visiting requests - Make sure OBS ratings can’t submit requests, this should block out the noobs.
non_defect
visiting requests stop obs from visiting requests make sure obs ratings can’t submit requests this should block out the noobs
0
34,176
9,303,105,736
IssuesEvent
2019-03-24 15:05:53
eclipse/sumo
https://api.github.com/repos/eclipse/sumo
closed
source release does not allow building from source
c:build sumo-user
- missing files (FileHelpersTest.cpp) - version.py runs and assumes existing git repository
1.0
source release does not allow building from source - - missing files (FileHelpersTest.cpp) - version.py runs and assumes existing git repository
non_defect
source release does not allow building from source missing files filehelperstest cpp version py runs and assumes existing git repository
0
5,501
2,610,188,834
IssuesEvent
2015-02-26 18:59:49
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
指南脸上长色斑如何去掉
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 黄褐斑是每个人女人的痛,如何淡化黄褐斑是有斑点女人最�� �做的事情。其实生活中是有很多的方法可以淡化脸上黄褐斑� ��。我们来看看有哪些有效的淡化黄褐斑的方法吧~脸上长色�� �如何去掉, 《客户案例》   谢天谢地!我终于摆脱斑的烦恼了!呵呵,简直太高兴了!�� �起以前,还真有点说不出口,因为家族遗传,从小脸上就有� ��,随着年龄的增长,斑也是如影随形,真的讨厌死了,满脸 都是小星星啊,害的我小侄女老师抓我的脸,感情啊是把斑�� �成我脸上的脏东西了呢,那个尴尬啊,是没法说,为了俺的� ��丽,可是狠下心要去斑了,说起祛斑,可是走了不少的弯路 呢,祛斑霜,用了不下三种,效果确实一般般,幸好我买的�� �好点的,对皮肤刺激不怎么大,至少没有什么副作用。遮瑕� ��、粉底,天天没断过,自己都烦了呢,后来为了省事,直接 做了个激光祛斑,可一去看,人家医生说我是疤痕体质,不�� �做激光,唉。你说怎么这么倒霉呢。还好,我以前一个同学� ��我介绍了一种产品,「黛芙薇尔精华液」,说是精华的,效 果一流滴,你别说,我这个同学脸上的斑还真是没了呢,心�� �不如行到,我也赶快去他们网站上订购了2个周期的产品,收 到后就迫不及待 使用了,呵呵,我是天天照镜子啊,希望奇迹发生,呵呵,�� �天还是挺仁慈的,使用了一个星期以后,感觉斑淡了点,有� ��希望,继续用啊,不知不觉的,两个月过去了,我可美的不 得了啦,斑已经没了呢,现在我一抱我小侄女,她就亲我的�� �呢,,呵呵,这个小色女!以前怎么不亲我哩! 阅读了脸上长色斑如何去掉,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑如何去掉,同时为您分享祛斑小方法 1、把干净的黄豆放入醋瓶中,盖上盖子,一个月之后,就可� ��吃了。 2、量的多少自己看着办,黄豆泡了一段时间后会涨大,一定� ��让醋浸过黄豆 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:34
1.0
指南脸上长色斑如何去掉 - ``` 《摘要》 黄褐斑是每个人女人的痛,如何淡化黄褐斑是有斑点女人最�� �做的事情。其实生活中是有很多的方法可以淡化脸上黄褐斑� ��。我们来看看有哪些有效的淡化黄褐斑的方法吧~脸上长色�� �如何去掉, 《客户案例》   谢天谢地!我终于摆脱斑的烦恼了!呵呵,简直太高兴了!�� �起以前,还真有点说不出口,因为家族遗传,从小脸上就有� ��,随着年龄的增长,斑也是如影随形,真的讨厌死了,满脸 都是小星星啊,害的我小侄女老师抓我的脸,感情啊是把斑�� �成我脸上的脏东西了呢,那个尴尬啊,是没法说,为了俺的� ��丽,可是狠下心要去斑了,说起祛斑,可是走了不少的弯路 呢,祛斑霜,用了不下三种,效果确实一般般,幸好我买的�� �好点的,对皮肤刺激不怎么大,至少没有什么副作用。遮瑕� ��、粉底,天天没断过,自己都烦了呢,后来为了省事,直接 做了个激光祛斑,可一去看,人家医生说我是疤痕体质,不�� �做激光,唉。你说怎么这么倒霉呢。还好,我以前一个同学� ��我介绍了一种产品,「黛芙薇尔精华液」,说是精华的,效 果一流滴,你别说,我这个同学脸上的斑还真是没了呢,心�� �不如行到,我也赶快去他们网站上订购了2个周期的产品,收 到后就迫不及待 使用了,呵呵,我是天天照镜子啊,希望奇迹发生,呵呵,�� �天还是挺仁慈的,使用了一个星期以后,感觉斑淡了点,有� ��希望,继续用啊,不知不觉的,两个月过去了,我可美的不 得了啦,斑已经没了呢,现在我一抱我小侄女,她就亲我的�� �呢,,呵呵,这个小色女!以前怎么不亲我哩! 阅读了脸上长色斑如何去掉,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑如何去掉,同时为您分享祛斑小方法 1、把干净的黄豆放入醋瓶中,盖上盖子,一个月之后,就可� ��吃了。 2、量的多少自己看着办,黄豆泡了一段时间后会涨大,一定� ��让醋浸过黄豆 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:34
defect
指南脸上长色斑如何去掉 《摘要》 黄褐斑是每个人女人的痛,如何淡化黄褐斑是有斑点女人最�� �做的事情。其实生活中是有很多的方法可以淡化脸上黄褐斑� ��。我们来看看有哪些有效的淡化黄褐斑的方法吧 脸上长色�� �如何去掉, 《客户案例》   谢天谢地 我终于摆脱斑的烦恼了 呵呵,简直太高兴了 �� �起以前,还真有点说不出口,因为家族遗传,从小脸上就有� ��,随着年龄的增长,斑也是如影随形,真的讨厌死了,满脸 都是小星星啊,害的我小侄女老师抓我的脸,感情啊是把斑�� �成我脸上的脏东西了呢,那个尴尬啊,是没法说,为了俺的� ��丽,可是狠下心要去斑了,说起祛斑,可是走了不少的弯路 呢,祛斑霜,用了不下三种,效果确实一般般,幸好我买的�� �好点的,对皮肤刺激不怎么大,至少没有什么副作用。遮瑕� ��、粉底,天天没断过,自己都烦了呢,后来为了省事,直接 做了个激光祛斑,可一去看,人家医生说我是疤痕体质,不�� �做激光,唉。你说怎么这么倒霉呢。还好,我以前一个同学� ��我介绍了一种产品,「黛芙薇尔精华液」,说是精华的,效 果一流滴,你别说,我这个同学脸上的斑还真是没了呢,心�� �不如行到, ,收 到后就迫不及待 使用了,呵呵,我是天天照镜子啊,希望奇迹发生,呵呵,�� �天还是挺仁慈的,使用了一个星期以后,感觉斑淡了点,有� ��希望,继续用啊,不知不觉的,两个月过去了,我可美的不 得了啦,斑已经没了呢,现在我一抱我小侄女,她就亲我的�� �呢,,呵呵,这个小色女 以前怎么不亲我哩 阅读了脸上长色斑如何去掉,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑如何去掉,同时为您分享祛斑小方法 、把干净的黄豆放入醋瓶中,盖上盖子,一个月之后,就可� ��吃了。 、量的多少自己看着办,黄豆泡了一段时间后会涨大,一定� ��让醋浸过黄豆 original issue reported on code google com by additive gmail com on jul at
1
77,798
27,167,453,285
IssuesEvent
2023-02-17 16:24:56
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
508-defect-2 [SCREENREADER]: "On behalf of" string on the sign as representative field should programmatically describe the field
508/Accessibility 1010-team 508-defect-2 508-issue-screenreader legal-rep 1010-cg va-text-input sitewide CAIA sitewide accessibility
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** Josh ## User Story or Problem Statement As a screen reader in forms mode, I expect descriptive, non-interactive text between interactive form elements (in this case, a field and a checkbox) to be programmatically linked to the objects they are describing so I don't miss them. ## Details The string "On behalf of [Veteran name]" will most likely be skipped by most screen readers as it is non-interactive text that is sandwiched between two interactive form elements. Because it describes the field "Enter your name to sign as the Veteran's representative," it should be programmatically linked to it using `aria-describedby`. ## Acceptance Criteria - [ ] Upon tabbing to the field "Enter your name to sign as the Veteran's representative," the screen reader user also hears "on behalf of [Veteran's name]". ## Steps to Recreate Using a screen reader, tab to the field "Enter your name to sign as the Veteran's representative" and confirm that "on behalf of..." is not announced. ## Proposed Solution (if known) Use `aria-describedby` to link the "on behalf of..." string to the form field. ## Screenshots or Trace Logs <img width="1199" alt="Screen Shot 2021-05-10 at 10 49 33 PM" src="https://user-images.githubusercontent.com/14154792/117756432-266ce100-b1ec-11eb-92f7-c9e4cb7b69ca.png"> <img width="1576" alt="Screen Shot 2021-05-11 at 12 06 13 AM" src="https://user-images.githubusercontent.com/14154792/117756801-d80c1200-b1ec-11eb-9c86-ef4926517f82.png">
1.0
508-defect-2 [SCREENREADER]: "On behalf of" string on the sign as representative field should programmatically describe the field - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** Josh ## User Story or Problem Statement As a screen reader in forms mode, I expect descriptive, non-interactive text between interactive form elements (in this case, a field and a checkbox) to be programmatically linked to the objects they are describing so I don't miss them. ## Details The string "On behalf of [Veteran name]" will most likely be skipped by most screen readers as it is non-interactive text that is sandwiched between two interactive form elements. Because it describes the field "Enter your name to sign as the Veteran's representative," it should be programmatically linked to it using `aria-describedby`. ## Acceptance Criteria - [ ] Upon tabbing to the field "Enter your name to sign as the Veteran's representative," the screen reader user also hears "on behalf of [Veteran's name]". ## Steps to Recreate Using a screen reader, tab to the field "Enter your name to sign as the Veteran's representative" and confirm that "on behalf of..." is not announced. ## Proposed Solution (if known) Use `aria-describedby` to link the "on behalf of..." string to the form field. ## Screenshots or Trace Logs <img width="1199" alt="Screen Shot 2021-05-10 at 10 49 33 PM" src="https://user-images.githubusercontent.com/14154792/117756432-266ce100-b1ec-11eb-92f7-c9e4cb7b69ca.png"> <img width="1576" alt="Screen Shot 2021-05-11 at 12 06 13 AM" src="https://user-images.githubusercontent.com/14154792/117756801-d80c1200-b1ec-11eb-9c86-ef4926517f82.png">
defect
defect on behalf of string on the sign as representative field should programmatically describe the field enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact josh user story or problem statement as a screen reader in forms mode i expect descriptive non interactive text between interactive form elements in this case a field and a checkbox to be programmatically linked to the objects they are describing so i don t miss them details the string on behalf of will most likely be skipped by most screen readers as it is non interactive text that is sandwiched between two interactive form elements because it describes the field enter your name to sign as the veteran s representative it should be programmatically linked to it using aria describedby acceptance criteria upon tabbing to the field enter your name to sign as the veteran s representative the screen reader user also hears on behalf of steps to recreate using a screen reader tab to the field enter your name to sign as the veteran s representative and confirm that on behalf of is not announced proposed solution if known use aria describedby to link the on behalf of string to the form field screenshots or trace logs img width alt screen shot at pm src img width alt screen shot at am src
1
600,709
18,352,532,124
IssuesEvent
2021-10-08 14:08:33
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
[3.2.0] Errors in Application Owner Change
Type/Bug Priority/Normal
### Description: The Application Owner Change functionality is not working as expected in the APIM v3.2.0 (1619587854304). After updating the Application Owner from the Admin portal, when trying to update the specific Application using the new owner, the following error-traces are captured in the Carbon logs. Please find the reproducing steps below ### Steps to reproduce: - Created a user as foo in the Super tenant with Subscriber Role - Logged into the Devportal with foo and admin user and created an Application with Foo user named: FooApplication and generated the Keys for Production - Logged to the Admin portal and changed the owner of the FooApplication from Foo user to Admin user - Logged to the Devportal using the Admin user and able to see the FooApplication listed, but when trying to update the Production OAuth App (example: change the expiry), we saw the following error log in the carbon logs ```log org.wso2.carbon.identity.oauth.dcr.exception.DCRMServerException: Error occurred while updating service provider foo_FooApplication_PRODUCTION at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:1.8.0_201] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[?:1.8.0_201] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:1.8.0_201] at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[?:1.8.0_201] at org.wso2.carbon.identity.base.IdentityException.error(IdentityException.java:103) ~[org.wso2.carbon.identity.base_5.17.5.jar:?] at org.wso2.carbon.identity.oauth.dcr.util.DCRMUtils.generateServerException(DCRMUtils.java:123) ~[org.wso2.carbon.identity.oauth.dcr_6.4.2.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:622) ~[classes/:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateApplication(DCRMService.java:115) ~[classes/:?] at org.wso2.is.key.manager.operations.endpoint.impl.DcrApiServiceImpl.updateApplication(DcrApiServiceImpl.java:130) [classes/:?] at org.wso2.is.key.manager.operations.endpoint.DcrApi.updateApplication(DcrApi.java:118) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_201] ... at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201] Caused by: org.wso2.carbon.identity.application.common.IdentityApplicationManagementException: Error occurred while updating the application: admin_FooApplication_PRODUCTION. Illegal Access! User: foo does not have access to update the application: 'admin_FooApplication_PRODUCTION' in tenantDomain: carbon.super at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.updateApplication(ApplicationManagementServiceImpl.java:599) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:620) ~[classes/:?] ... 61 more Caused by: org.wso2.carbon.identity.application.common.IdentityApplicationManagementClientException: Illegal Access! User: foo does not have access to update the application: 'admin_FooApplication_PRODUCTION' in tenantDomain: carbon.super at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.buildClientException(ApplicationManagementServiceImpl.java:2436) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.validateAuthorization(ApplicationManagementServiceImpl.java:2299) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.doPreUpdateChecks(ApplicationManagementServiceImpl.java:2238) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.updateApplication(ApplicationManagementServiceImpl.java:587) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:620) ~[classes/:?] ... 61 more ``` ### Affected Product Version: - API Manager v3.2.0 (1619587854304)
1.0
[3.2.0] Errors in Application Owner Change - ### Description: The Application Owner Change functionality is not working as expected in the APIM v3.2.0 (1619587854304). After updating the Application Owner from the Admin portal, when trying to update the specific Application using the new owner, the following error-traces are captured in the Carbon logs. Please find the reproducing steps below ### Steps to reproduce: - Created a user as foo in the Super tenant with Subscriber Role - Logged into the Devportal with foo and admin user and created an Application with Foo user named: FooApplication and generated the Keys for Production - Logged to the Admin portal and changed the owner of the FooApplication from Foo user to Admin user - Logged to the Devportal using the Admin user and able to see the FooApplication listed, but when trying to update the Production OAuth App (example: change the expiry), we saw the following error log in the carbon logs ```log org.wso2.carbon.identity.oauth.dcr.exception.DCRMServerException: Error occurred while updating service provider foo_FooApplication_PRODUCTION at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:1.8.0_201] at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[?:1.8.0_201] at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:1.8.0_201] at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[?:1.8.0_201] at org.wso2.carbon.identity.base.IdentityException.error(IdentityException.java:103) ~[org.wso2.carbon.identity.base_5.17.5.jar:?] at org.wso2.carbon.identity.oauth.dcr.util.DCRMUtils.generateServerException(DCRMUtils.java:123) ~[org.wso2.carbon.identity.oauth.dcr_6.4.2.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:622) ~[classes/:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateApplication(DCRMService.java:115) ~[classes/:?] at org.wso2.is.key.manager.operations.endpoint.impl.DcrApiServiceImpl.updateApplication(DcrApiServiceImpl.java:130) [classes/:?] at org.wso2.is.key.manager.operations.endpoint.DcrApi.updateApplication(DcrApi.java:118) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_201] ... at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201] Caused by: org.wso2.carbon.identity.application.common.IdentityApplicationManagementException: Error occurred while updating the application: admin_FooApplication_PRODUCTION. Illegal Access! User: foo does not have access to update the application: 'admin_FooApplication_PRODUCTION' in tenantDomain: carbon.super at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.updateApplication(ApplicationManagementServiceImpl.java:599) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:620) ~[classes/:?] ... 61 more Caused by: org.wso2.carbon.identity.application.common.IdentityApplicationManagementClientException: Illegal Access! User: foo does not have access to update the application: 'admin_FooApplication_PRODUCTION' in tenantDomain: carbon.super at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.buildClientException(ApplicationManagementServiceImpl.java:2436) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.validateAuthorization(ApplicationManagementServiceImpl.java:2299) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.doPreUpdateChecks(ApplicationManagementServiceImpl.java:2238) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.carbon.identity.application.mgt.ApplicationManagementServiceImpl.updateApplication(ApplicationManagementServiceImpl.java:587) ~[org.wso2.carbon.identity.application.mgt_5.17.5.jar:?] at org.wso2.is.key.manager.operations.endpoint.dcr.service.DCRMService.updateServiceProvider(DCRMService.java:620) ~[classes/:?] ... 61 more ``` ### Affected Product Version: - API Manager v3.2.0 (1619587854304)
non_defect
errors in application owner change description the application owner change functionality is not working as expected in the apim after updating the application owner from the admin portal when trying to update the specific application using the new owner the following error traces are captured in the carbon logs please find the reproducing steps below steps to reproduce created a user as foo in the super tenant with subscriber role logged into the devportal with foo and admin user and created an application with foo user named fooapplication and generated the keys for production logged to the admin portal and changed the owner of the fooapplication from foo user to admin user logged to the devportal using the admin user and able to see the fooapplication listed but when trying to update the production oauth app example change the expiry we saw the following error log in the carbon logs log org carbon identity oauth dcr exception dcrmserverexception error occurred while updating service provider foo fooapplication production at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at org carbon identity base identityexception error identityexception java at org carbon identity oauth dcr util dcrmutils generateserverexception dcrmutils java at org is key manager operations endpoint dcr service dcrmservice updateserviceprovider dcrmservice java at org is key manager operations endpoint dcr service dcrmservice updateapplication dcrmservice java at org is key manager operations endpoint impl dcrapiserviceimpl updateapplication dcrapiserviceimpl java at org is key manager operations endpoint dcrapi updateapplication dcrapi java at sun reflect nativemethodaccessorimpl native method at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by org carbon identity application common identityapplicationmanagementexception error occurred while updating the application admin fooapplication production illegal access user foo does not have access to update the application admin fooapplication production in tenantdomain carbon super at org carbon identity application mgt applicationmanagementserviceimpl updateapplication applicationmanagementserviceimpl java at org is key manager operations endpoint dcr service dcrmservice updateserviceprovider dcrmservice java more caused by org carbon identity application common identityapplicationmanagementclientexception illegal access user foo does not have access to update the application admin fooapplication production in tenantdomain carbon super at org carbon identity application mgt applicationmanagementserviceimpl buildclientexception applicationmanagementserviceimpl java at org carbon identity application mgt applicationmanagementserviceimpl validateauthorization applicationmanagementserviceimpl java at org carbon identity application mgt applicationmanagementserviceimpl dopreupdatechecks applicationmanagementserviceimpl java at org carbon identity application mgt applicationmanagementserviceimpl updateapplication applicationmanagementserviceimpl java at org is key manager operations endpoint dcr service dcrmservice updateserviceprovider dcrmservice java more affected product version api manager
0
45,561
12,867,966,248
IssuesEvent
2020-07-10 07:59:20
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Out of thin air exceptions on the client
Source: Internal Team: Client Type: Defect
After the fix https://github.com/hazelcast/hazelcast/pull/15441 we have removed exception stack trace rewrite on the server. Since the client proxies do not use correct method (which seems to be `invocationFuture.joinInternal`), the exceptions that come from the remote are thrown with the stack trace on the remote. This makes it hard to follow the stack trace. The solution is already available on the member side. This issue is to address the problems on the client-side.
1.0
Out of thin air exceptions on the client - After the fix https://github.com/hazelcast/hazelcast/pull/15441 we have removed exception stack trace rewrite on the server. Since the client proxies do not use correct method (which seems to be `invocationFuture.joinInternal`), the exceptions that come from the remote are thrown with the stack trace on the remote. This makes it hard to follow the stack trace. The solution is already available on the member side. This issue is to address the problems on the client-side.
defect
out of thin air exceptions on the client after the fix we have removed exception stack trace rewrite on the server since the client proxies do not use correct method which seems to be invocationfuture joininternal the exceptions that come from the remote are thrown with the stack trace on the remote this makes it hard to follow the stack trace the solution is already available on the member side this issue is to address the problems on the client side
1
49,501
20,772,866,765
IssuesEvent
2022-03-16 07:31:49
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Reporting] CSV Export URL generation adds an absolute time range filter
bug Feature:Discover Feature:Reporting loe:hours Team:AppServicesSv impact:medium Team:DataDiscovery
**Kibana version:** 7.16.1 **Elasticsearch version:** Aligned **Describe the bug:** CSV Export URL contains an unexpected time filter, which is not present at the time of the visualization. **Steps to reproduce:** 1. Create an index pattern which contains data during the last N days. I've used `.monitoring-es*` 2. Choose `Last 24h` in the timepicker (note the URL contains only `(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-24h%2Fh,to:now))&_a=(columns:!(),filters:!(),index:'10881820-3b02-11eb-a646-7340c405390d',interval:auto,query:(language:kuery,query:''),sort:!(!(timestamp,desc)))` 3. Save the Saved Search - The URL still shows `(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-24h%2Fh,to:now))&_a=(columns:!(),filters:!(),grid:(),hideChart:!f,index:'10881820-3b02-11eb-a646-7340c405390d',interval:auto,query:(language:kuery,query:''),sort:!(!(timestamp,desc)))` 4. Go on `Share / CSV Export / Copy POST URL` 5. You'll get `.../api/reporting/generate/csv_searchsource?jobParams=%28browserTimezone%3AEurope%2FDublin%2Ccolumns%3A%21%28%29%2CobjectType%3Asearch%2CsearchSource%3A%28fields%3A%21%28%28field%3A%27%2A%27%2Cinclude_unmapped%3Atrue%29%29%2Cfilter%3A%21%28%28meta%3A%28field%3Atimestamp%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparams%3A%28%29%29%2Cquery%3A%28range%3A%28timestamp%3A%28format%3Astrict_date_optional_time%2Cgte%3Anow-24h%2Fh%2Clte%3Anow%29%29%29%29%2C%28meta%3A%28field%3Atimestamp%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparams%3A%28%29%29%2Cquery%3A%28range%3A%28timestamp%3A%28format%3Astrict_date_optional_time%2Cgte%3A%272022-01-09T18%3A00%3A00.000Z%27%2Clte%3A%272022-01-10T18%3A14%3A50.694Z%27%29%29%29%29%29%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparent%3A%28filter%3A%21%28%29%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cquery%3A%28language%3Akuery%2Cquery%3A%27%27%29%29%2Csort%3A%21%28%28timestamp%3Adesc%29%29%2CtrackTotalHits%3A%21t%29%2Ctitle%3Ademo-csv-export%2Cversion%3A%277.16.1%27%29` URL Decoded params: ``` (browserTimezone:Europe/Dublin,columns:!(),objectType:search,searchSource:(fields:!((field:'*',include_unmapped:true)),filter:!((meta:(field:timestamp,index:'10881820-3b02-11eb-a646-7340c405390d',params:()),query:(range:(timestamp:(format:strict_date_optional_time,gte:now-24h/h,lte:now)))),(meta:(field:timestamp,index:'10881820-3b02-11eb-a646-7340c405390d',params:()),query:(range:(timestamp:(format:strict_date_optional_time,gte:'2022-01-09T18:00:00.000Z',lte:'2022-01-10T18:14:50.694Z'))))),index:'10881820-3b02-11eb-a646-7340c405390d',parent:(filter:!(),index:'10881820-3b02-11eb-a646-7340c405390d',query:(language:kuery,query:'')),sort:!((timestamp:desc)),trackTotalHits:!t),title:demo-csv-export,version:'7.16.1') ``` We can see there's an additional query which passes the "current" time, meaning this URL cannot be used for automated reports except manually trimming away the second `range` query (`query:(range:(timestamp:(format:strict_date_optional_time,gte:'2022-01-09T18:00:00.000Z',lte:'2022-01-10T18:14:50.694Z')))`). Note that reloading the saved search, the additional time `range` filter is not present. Only the relative filter is present. **Expected behavior:** Generate a CSV Export URL which only includes the relative timepicker.
1.0
[Reporting] CSV Export URL generation adds an absolute time range filter - **Kibana version:** 7.16.1 **Elasticsearch version:** Aligned **Describe the bug:** CSV Export URL contains an unexpected time filter, which is not present at the time of the visualization. **Steps to reproduce:** 1. Create an index pattern which contains data during the last N days. I've used `.monitoring-es*` 2. Choose `Last 24h` in the timepicker (note the URL contains only `(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-24h%2Fh,to:now))&_a=(columns:!(),filters:!(),index:'10881820-3b02-11eb-a646-7340c405390d',interval:auto,query:(language:kuery,query:''),sort:!(!(timestamp,desc)))` 3. Save the Saved Search - The URL still shows `(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-24h%2Fh,to:now))&_a=(columns:!(),filters:!(),grid:(),hideChart:!f,index:'10881820-3b02-11eb-a646-7340c405390d',interval:auto,query:(language:kuery,query:''),sort:!(!(timestamp,desc)))` 4. Go on `Share / CSV Export / Copy POST URL` 5. You'll get `.../api/reporting/generate/csv_searchsource?jobParams=%28browserTimezone%3AEurope%2FDublin%2Ccolumns%3A%21%28%29%2CobjectType%3Asearch%2CsearchSource%3A%28fields%3A%21%28%28field%3A%27%2A%27%2Cinclude_unmapped%3Atrue%29%29%2Cfilter%3A%21%28%28meta%3A%28field%3Atimestamp%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparams%3A%28%29%29%2Cquery%3A%28range%3A%28timestamp%3A%28format%3Astrict_date_optional_time%2Cgte%3Anow-24h%2Fh%2Clte%3Anow%29%29%29%29%2C%28meta%3A%28field%3Atimestamp%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparams%3A%28%29%29%2Cquery%3A%28range%3A%28timestamp%3A%28format%3Astrict_date_optional_time%2Cgte%3A%272022-01-09T18%3A00%3A00.000Z%27%2Clte%3A%272022-01-10T18%3A14%3A50.694Z%27%29%29%29%29%29%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cparent%3A%28filter%3A%21%28%29%2Cindex%3A%2710881820-3b02-11eb-a646-7340c405390d%27%2Cquery%3A%28language%3Akuery%2Cquery%3A%27%27%29%29%2Csort%3A%21%28%28timestamp%3Adesc%29%29%2CtrackTotalHits%3A%21t%29%2Ctitle%3Ademo-csv-export%2Cversion%3A%277.16.1%27%29` URL Decoded params: ``` (browserTimezone:Europe/Dublin,columns:!(),objectType:search,searchSource:(fields:!((field:'*',include_unmapped:true)),filter:!((meta:(field:timestamp,index:'10881820-3b02-11eb-a646-7340c405390d',params:()),query:(range:(timestamp:(format:strict_date_optional_time,gte:now-24h/h,lte:now)))),(meta:(field:timestamp,index:'10881820-3b02-11eb-a646-7340c405390d',params:()),query:(range:(timestamp:(format:strict_date_optional_time,gte:'2022-01-09T18:00:00.000Z',lte:'2022-01-10T18:14:50.694Z'))))),index:'10881820-3b02-11eb-a646-7340c405390d',parent:(filter:!(),index:'10881820-3b02-11eb-a646-7340c405390d',query:(language:kuery,query:'')),sort:!((timestamp:desc)),trackTotalHits:!t),title:demo-csv-export,version:'7.16.1') ``` We can see there's an additional query which passes the "current" time, meaning this URL cannot be used for automated reports except manually trimming away the second `range` query (`query:(range:(timestamp:(format:strict_date_optional_time,gte:'2022-01-09T18:00:00.000Z',lte:'2022-01-10T18:14:50.694Z')))`). Note that reloading the saved search, the additional time `range` filter is not present. Only the relative filter is present. **Expected behavior:** Generate a CSV Export URL which only includes the relative timepicker.
non_defect
csv export url generation adds an absolute time range filter kibana version elasticsearch version aligned describe the bug csv export url contains an unexpected time filter which is not present at the time of the visualization steps to reproduce create an index pattern which contains data during the last n days i ve used monitoring es choose last in the timepicker note the url contains only filters refreshinterval pause t value time from now to now a columns filters index interval auto query language kuery query sort timestamp desc save the saved search the url still shows filters refreshinterval pause t value time from now to now a columns filters grid hidechart f index interval auto query language kuery query sort timestamp desc go on share csv export copy post url you ll get api reporting generate csv searchsource jobparams unmapped date optional time date optional time csv export url decoded params browsertimezone europe dublin columns objecttype search searchsource fields field include unmapped true filter meta field timestamp index params query range timestamp format strict date optional time gte now h lte now meta field timestamp index params query range timestamp format strict date optional time gte lte index parent filter index query language kuery query sort timestamp desc tracktotalhits t title demo csv export version we can see there s an additional query which passes the current time meaning this url cannot be used for automated reports except manually trimming away the second range query query range timestamp format strict date optional time gte lte note that reloading the saved search the additional time range filter is not present only the relative filter is present expected behavior generate a csv export url which only includes the relative timepicker
0
333,621
24,382,965,264
IssuesEvent
2022-10-04 09:22:39
100mslive/100ms-flutter
https://api.github.com/repos/100mslive/100ms-flutter
closed
Add description of working of Example apps in ReadMe
documentation
Add descriptions for Example apps in - - Provider - MobX - GetX - BloC - Riverpod - To be added in ReadMe files of these apps. - Add details on how onJoin, onPeerUpdate & onTrackUpdates are used. - Creation & usage of PeerTrackNodes - Usage of State Management libraries - Use of different UI Components - Sorting & Changing Layout functions
1.0
Add description of working of Example apps in ReadMe - Add descriptions for Example apps in - - Provider - MobX - GetX - BloC - Riverpod - To be added in ReadMe files of these apps. - Add details on how onJoin, onPeerUpdate & onTrackUpdates are used. - Creation & usage of PeerTrackNodes - Usage of State Management libraries - Use of different UI Components - Sorting & Changing Layout functions
non_defect
add description of working of example apps in readme add descriptions for example apps in provider mobx getx bloc riverpod to be added in readme files of these apps add details on how onjoin onpeerupdate ontrackupdates are used creation usage of peertracknodes usage of state management libraries use of different ui components sorting changing layout functions
0
35,372
12,322,277,550
IssuesEvent
2020-05-13 10:04:15
fufunoyu/example-npm-circle
https://api.github.com/repos/fufunoyu/example-npm-circle
opened
CVE-2020-7608 (High) detected in yargs-parser-11.1.1.tgz
security vulnerability
## CVE-2020-7608 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-11.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/example-npm-circle/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/example-npm-circle/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - protractor-5.4.4.tgz (Root Library) - yargs-12.0.5.tgz - :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-npm-circle/commit/4b9d35337365451db5cf7aad10d58ff1a3dc8acd">4b9d35337365451db5cf7aad10d58ff1a3dc8acd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"11.1.1","isTransitiveDependency":true,"dependencyTree":"protractor:5.4.4;yargs:12.0.5;yargs-parser:11.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7608 (High) detected in yargs-parser-11.1.1.tgz - ## CVE-2020-7608 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-11.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/example-npm-circle/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/example-npm-circle/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - protractor-5.4.4.tgz (Root Library) - yargs-12.0.5.tgz - :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-npm-circle/commit/4b9d35337365451db5cf7aad10d58ff1a3dc8acd">4b9d35337365451db5cf7aad10d58ff1a3dc8acd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"11.1.1","isTransitiveDependency":true,"dependencyTree":"protractor:5.4.4;yargs:12.0.5;yargs-parser:11.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in yargs parser tgz cve high severity vulnerability vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file tmp ws scm example npm circle package json path to vulnerable library tmp ws scm example npm circle node modules yargs parser package json dependency hierarchy protractor tgz root library yargs tgz x yargs parser tgz vulnerable library found in head commit a href vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
0
43,351
11,634,009,337
IssuesEvent
2020-02-28 09:30:47
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
rec: Boost context is no longer detected with boost >= 1.72.0
defect rec
- Program: Recursor - Issue type: Bug report ### Short description On Arch with boost 1.72.0: ``` checking for the Boost context library... no checking for getcontext... yes checking for makecontext... yes checking for swapcontext... yes checking what context library to use for MTasker... System V ucontexts ``` ``` configure:14335: checking for the Boost context library configure:14401: g++ -c -Wall -Wextra -Wshadow -Wno-unused-parameter -g -O2 -pthread conftest.cpp >&5 conftest.cpp:43:10: fatal error: boost/context/execution_context_v2.hpp: No such file or directory 43 | #include <boost/context/execution_context_v2.hpp> | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ``` This happens because our m4 detection tries to use `execution_context_v2.hpp` when the boost version is greater than or equal to 1.61, but unfortunately this API was removed in 1.72.0 beta 1: https://github.com/boostorg/context/commit/0c754f90c53ddb8d041c2ccbc78da5cb484c5f1f#diff-7750a754757adf2b59d3dd0cad775071
1.0
rec: Boost context is no longer detected with boost >= 1.72.0 - - Program: Recursor - Issue type: Bug report ### Short description On Arch with boost 1.72.0: ``` checking for the Boost context library... no checking for getcontext... yes checking for makecontext... yes checking for swapcontext... yes checking what context library to use for MTasker... System V ucontexts ``` ``` configure:14335: checking for the Boost context library configure:14401: g++ -c -Wall -Wextra -Wshadow -Wno-unused-parameter -g -O2 -pthread conftest.cpp >&5 conftest.cpp:43:10: fatal error: boost/context/execution_context_v2.hpp: No such file or directory 43 | #include <boost/context/execution_context_v2.hpp> | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ``` This happens because our m4 detection tries to use `execution_context_v2.hpp` when the boost version is greater than or equal to 1.61, but unfortunately this API was removed in 1.72.0 beta 1: https://github.com/boostorg/context/commit/0c754f90c53ddb8d041c2ccbc78da5cb484c5f1f#diff-7750a754757adf2b59d3dd0cad775071
defect
rec boost context is no longer detected with boost program recursor issue type bug report short description on arch with boost checking for the boost context library no checking for getcontext yes checking for makecontext yes checking for swapcontext yes checking what context library to use for mtasker system v ucontexts configure checking for the boost context library configure g c wall wextra wshadow wno unused parameter g pthread conftest cpp conftest cpp fatal error boost context execution context hpp no such file or directory include this happens because our detection tries to use execution context hpp when the boost version is greater than or equal to but unfortunately this api was removed in beta
1
40,433
12,793,904,668
IssuesEvent
2020-07-02 05:31:30
TIBCOSoftware/TCSTK-case-manager-app
https://api.github.com/repos/TIBCOSoftware/TCSTK-case-manager-app
closed
CVE-2020-7608 (Medium) detected in yargs-parser-11.1.1.tgz, yargs-parser-13.1.1.tgz
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-11.1.1.tgz</b>, <b>yargs-parser-13.1.1.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/TCSTK-case-manager-app/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/TCSTK-case-manager-app/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.803.21.tgz (Root Library) - webpack-dev-server-3.9.0.tgz - yargs-12.0.5.tgz - :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/TCSTK-case-manager-app/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/TCSTK-case-manager-app/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-8.2.14.tgz (Root Library) - yargs-13.1.0.tgz - :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"11.1.1","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.803.21;webpack-dev-server:3.9.0;yargs:12.0.5;yargs-parser:11.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"13.1.1","isTransitiveDependency":true,"dependencyTree":"@angular/compiler-cli:8.2.14;yargs:13.1.0;yargs-parser:13.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7608 (Medium) detected in yargs-parser-11.1.1.tgz, yargs-parser-13.1.1.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-11.1.1.tgz</b>, <b>yargs-parser-13.1.1.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/TCSTK-case-manager-app/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/TCSTK-case-manager-app/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - build-angular-0.803.21.tgz (Root Library) - webpack-dev-server-3.9.0.tgz - yargs-12.0.5.tgz - :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/TCSTK-case-manager-app/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/TCSTK-case-manager-app/node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - compiler-cli-8.2.14.tgz (Root Library) - yargs-13.1.0.tgz - :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"11.1.1","isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.803.21;webpack-dev-server:3.9.0;yargs:12.0.5;yargs-parser:11.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"13.1.1","isTransitiveDependency":true,"dependencyTree":"@angular/compiler-cli:8.2.14;yargs:13.1.0;yargs-parser:13.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in yargs parser tgz yargs parser tgz cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file tmp ws scm tcstk case manager app package json path to vulnerable library tmp ws scm tcstk case manager app node modules yargs parser package json dependency hierarchy build angular tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file tmp ws scm tcstk case manager app package json path to vulnerable library tmp ws scm tcstk case manager app node modules angular compiler cli node modules yargs parser package json dependency hierarchy compiler cli tgz root library yargs tgz x yargs parser tgz vulnerable library vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
0
29,965
5,965,561,226
IssuesEvent
2017-05-30 11:57:40
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Several DB2 fixes
C: DB: DB2 P: Medium R: Fixed T: Defect
- [x] #6117 DB2 doesn't support qualified column references in FOR UPDATE OF - [x] #6118 RESTRICT is a mandatory keyword in DB2's DROP SCHEMA command - [ ] ~~#6120 Add support for DB2 RENAME INDEX statement~~ - [ ] ~~#6121 Add support for DB2 RENAME TABLE statement~~ - [ ] ~~#6122 Fix DDL Support annotations for DB2~~ - [x] #6127 DELETE RETURNING must be emulated on DB2 with OLD TABLE() not FINAL TABLE()
1.0
Several DB2 fixes - - [x] #6117 DB2 doesn't support qualified column references in FOR UPDATE OF - [x] #6118 RESTRICT is a mandatory keyword in DB2's DROP SCHEMA command - [ ] ~~#6120 Add support for DB2 RENAME INDEX statement~~ - [ ] ~~#6121 Add support for DB2 RENAME TABLE statement~~ - [ ] ~~#6122 Fix DDL Support annotations for DB2~~ - [x] #6127 DELETE RETURNING must be emulated on DB2 with OLD TABLE() not FINAL TABLE()
defect
several fixes doesn t support qualified column references in for update of restrict is a mandatory keyword in s drop schema command add support for rename index statement add support for rename table statement fix ddl support annotations for delete returning must be emulated on with old table not final table
1
294,850
9,049,446,202
IssuesEvent
2019-02-12 04:42:15
adamgibbons/ics
https://api.github.com/repos/adamgibbons/ics
closed
Getting validation error when 'recurrenceRule' value set
bug help wanted high priority
Getting following error when we try to set `recurrenceRule` field value `error: ValidationError: "recurrenceRule" is not allowed`
1.0
Getting validation error when 'recurrenceRule' value set - Getting following error when we try to set `recurrenceRule` field value `error: ValidationError: "recurrenceRule" is not allowed`
non_defect
getting validation error when recurrencerule value set getting following error when we try to set recurrencerule field value error validationerror recurrencerule is not allowed
0
127,576
5,036,844,976
IssuesEvent
2016-12-17 09:33:05
ludo237/claudioludovico
https://api.github.com/repos/ludo237/claudioludovico
closed
404 Page not working
Bug Priority: Medium Status: Pending
When a page does not exists the application will redirect to the standard error page instead of a dedicated custom 404 page.
1.0
404 Page not working - When a page does not exists the application will redirect to the standard error page instead of a dedicated custom 404 page.
non_defect
page not working when a page does not exists the application will redirect to the standard error page instead of a dedicated custom page
0
2,350
2,607,897,230
IssuesEvent
2015-02-26 00:11:56
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
Hotkey for "Wrap wih abbreviation" in Windows
auto-migrated Priority-Medium Type-Defect
``` В новой версии Zen Coding for Aptana v0.5 появилась отличная функция "Wrap wih abbreviation", ей соответствует комбинация Command+E Какая комбинация соответствует этой команде в ОС Windows? ``` ----- Original issue reported on code.google.com by `metaller...@gmail.com` on 10 Nov 2009 at 8:11
1.0
Hotkey for "Wrap wih abbreviation" in Windows - ``` В новой версии Zen Coding for Aptana v0.5 появилась отличная функция "Wrap wih abbreviation", ей соответствует комбинация Command+E Какая комбинация соответствует этой команде в ОС Windows? ``` ----- Original issue reported on code.google.com by `metaller...@gmail.com` on 10 Nov 2009 at 8:11
defect
hotkey for wrap wih abbreviation in windows в новой версии zen coding for aptana появилась отличная функция wrap wih abbreviation ей соответствует комбинация command e какая комбинация соответствует этой команде в ос windows original issue reported on code google com by metaller gmail com on nov at
1
42,933
11,385,681,406
IssuesEvent
2020-01-29 11:38:22
line/armeria
https://api.github.com/repos/line/armeria
closed
Dcom.linecorp.armeria.dumpOpenSslInfo=true flag throws server into infinite loop
defect
Version: Armeria 0.95 ``` 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.543 WARN 6 --- [ main] s.c.a.AnnotationConfigApplicationContext : Exception encountered during context initialization - cancelling refresh attempt: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'armeriaServer' defined in com.linecorp.armeria.spring.ArmeriaAutoConfiguration: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError 2020-01-23 19:24:36.577 INFO 6 --- [ main] ConditionEvaluationReportLoggingListener : Error starting ApplicationContext. To display the conditions report re-run your application with 'debug' enabled. 2020-01-23 19:24:36.595 ERROR 6 --- [ main] o.s.b.SpringApplication : Application run failed org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'armeriaServer' defined in com.linecorp.armeria.spring.ArmeriaAutoConfiguration: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:645) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:625) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1338) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1177) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:557) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:517) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:323) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:321) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:202) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:879) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:878) ~[spring-context-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:550) ~[spring-context-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:747) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:397) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.SpringApplication.run(SpringApplication.java:315) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:140) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at zipkin.server.ZipkinServer.main(ZipkinServer.java:54) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_232] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_232] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_232] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_232] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48) [zipkin-server/:?] at org.springframework.boot.loader.Launcher.launch(Launcher.java:87) [zipkin-server/:?] at org.springframework.boot.loader.Launcher.launch(Launcher.java:51) [zipkin-server/:?] at org.springframework.boot.loader.PropertiesLauncher.main(PropertiesLauncher.java:597) [zipkin-server/:?] Caused by: org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:185) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:640) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] ... 25 more Caused by: java.lang.StackOverflowError at org.apache.logging.log4j.core.pattern.DatePatternConverter.getThreadLocalFormatter(DatePatternConverter.java:299) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.formatWithoutAllocation(DatePatternConverter.java:295) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.format(DatePatternConverter.java:288) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.format(DatePatternConverter.java:265) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.PatternFormatter.format(PatternFormatter.java:38) ~[log4j-core-2.12.1.jar:2.12.1] at org.springframework.boot.logging.log4j2.ColorConverter.format(ColorConverter.java:121) ~[spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.apache.logging.log4j.core.pattern.PatternFormatter.format(PatternFormatter.java:38) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout$PatternSerializer.toSerializable(PatternLayout.java:341) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.toText(PatternLayout.java:240) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.encode(PatternLayout.java:225) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.encode(PatternLayout.java:59) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.directEncodeEvent(AbstractOutputStreamAppender.java:197) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.tryAppend(AbstractOutputStreamAppender.java:190) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.append(AbstractOutputStreamAppender.java:181) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.tryCallAppender(AppenderControl.java:156) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppender0(AppenderControl.java:129) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppenderPreventRecursion(AppenderControl.java:120) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppender(AppenderControl.java:84) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.callAppenders(LoggerConfig.java:543) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.processLogEvent(LoggerConfig.java:502) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:485) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:460) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AwaitCompletionReliabilityStrategy.log(AwaitCompletionReliabilityStrategy.java:82) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.Logger.log(Logger.java:162) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2190) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2144) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2127) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:2026) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1899) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.slf4j.Log4jLogger.info(Log4jLogger.java:189) ~[log4j-slf4j-impl-2.12.1.jar:2.12.1] at com.linecorp.armeria.common.Flags.getNormalized(Flags.java:931) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.getBoolean(Flags.java:875) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.getBoolean(Flags.java:871) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:413) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:426) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:426) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] ```
1.0
Dcom.linecorp.armeria.dumpOpenSslInfo=true flag throws server into infinite loop - Version: Armeria 0.95 ``` 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : com.linecorp.armeria.useOpenSsl: true (default) 2020-01-23 19:24:36.539 INFO 6 --- [ main] c.l.a.c.Flags : Using OpenSSL: BoringSSL, 0x1010007f 2020-01-23 19:24:36.543 WARN 6 --- [ main] s.c.a.AnnotationConfigApplicationContext : Exception encountered during context initialization - cancelling refresh attempt: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'armeriaServer' defined in com.linecorp.armeria.spring.ArmeriaAutoConfiguration: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError 2020-01-23 19:24:36.577 INFO 6 --- [ main] ConditionEvaluationReportLoggingListener : Error starting ApplicationContext. To display the conditions report re-run your application with 'debug' enabled. 2020-01-23 19:24:36.595 ERROR 6 --- [ main] o.s.b.SpringApplication : Application run failed org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'armeriaServer' defined in com.linecorp.armeria.spring.ArmeriaAutoConfiguration: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:645) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:625) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1338) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1177) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:557) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:517) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:323) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:321) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:202) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:879) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:878) ~[spring-context-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:550) ~[spring-context-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:747) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:397) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.SpringApplication.run(SpringApplication.java:315) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:140) [spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at zipkin.server.ZipkinServer.main(ZipkinServer.java:54) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_232] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_232] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_232] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_232] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48) [zipkin-server/:?] at org.springframework.boot.loader.Launcher.launch(Launcher.java:87) [zipkin-server/:?] at org.springframework.boot.loader.Launcher.launch(Launcher.java:51) [zipkin-server/:?] at org.springframework.boot.loader.PropertiesLauncher.main(PropertiesLauncher.java:597) [zipkin-server/:?] Caused by: org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.linecorp.armeria.server.Server]: Factory method 'armeriaServer' threw exception; nested exception is java.lang.StackOverflowError at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:185) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:640) ~[spring-beans-5.2.0.RELEASE.jar:5.2.0.RELEASE] ... 25 more Caused by: java.lang.StackOverflowError at org.apache.logging.log4j.core.pattern.DatePatternConverter.getThreadLocalFormatter(DatePatternConverter.java:299) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.formatWithoutAllocation(DatePatternConverter.java:295) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.format(DatePatternConverter.java:288) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.DatePatternConverter.format(DatePatternConverter.java:265) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.pattern.PatternFormatter.format(PatternFormatter.java:38) ~[log4j-core-2.12.1.jar:2.12.1] at org.springframework.boot.logging.log4j2.ColorConverter.format(ColorConverter.java:121) ~[spring-boot-2.2.0.RELEASE.jar:2.2.0.RELEASE] at org.apache.logging.log4j.core.pattern.PatternFormatter.format(PatternFormatter.java:38) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout$PatternSerializer.toSerializable(PatternLayout.java:341) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.toText(PatternLayout.java:240) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.encode(PatternLayout.java:225) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.layout.PatternLayout.encode(PatternLayout.java:59) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.directEncodeEvent(AbstractOutputStreamAppender.java:197) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.tryAppend(AbstractOutputStreamAppender.java:190) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.append(AbstractOutputStreamAppender.java:181) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.tryCallAppender(AppenderControl.java:156) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppender0(AppenderControl.java:129) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppenderPreventRecursion(AppenderControl.java:120) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AppenderControl.callAppender(AppenderControl.java:84) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.callAppenders(LoggerConfig.java:543) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.processLogEvent(LoggerConfig.java:502) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:485) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:460) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.config.AwaitCompletionReliabilityStrategy.log(AwaitCompletionReliabilityStrategy.java:82) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.core.Logger.log(Logger.java:162) ~[log4j-core-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2190) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2144) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2127) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:2026) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1899) ~[log4j-api-2.12.1.jar:2.12.1] at org.apache.logging.slf4j.Log4jLogger.info(Log4jLogger.java:189) ~[log4j-slf4j-impl-2.12.1.jar:2.12.1] at com.linecorp.armeria.common.Flags.getNormalized(Flags.java:931) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.getBoolean(Flags.java:875) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.getBoolean(Flags.java:871) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:413) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:426) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.common.Flags.useOpenSsl(Flags.java:426) ~[armeria-0.95.0.jar:?] at com.linecorp.armeria.internal.SslContextUtil.createSslContext(SslContextUtil.java:90) ~[armeria-0.95.0.jar:?] ```
defect
dcom linecorp armeria dumpopensslinfo true flag throws server into infinite loop version armeria info c l a c flags using openssl boringssl info c l a c flags com linecorp armeria useopenssl true default info c l a c flags using openssl boringssl info c l a c flags com linecorp armeria useopenssl true default info c l a c flags using openssl boringssl info c l a c flags com linecorp armeria useopenssl true default info c l a c flags using openssl boringssl warn s c a annotationconfigapplicationcontext exception encountered during context initialization cancelling refresh attempt org springframework beans factory beancreationexception error creating bean with name armeriaserver defined in com linecorp armeria spring armeriaautoconfiguration bean instantiation via factory method failed nested exception is org springframework beans beaninstantiationexception failed to instantiate factory method armeriaserver threw exception nested exception is java lang stackoverflowerror info conditionevaluationreportlogginglistener error starting applicationcontext to display the conditions report re run your application with debug enabled error o s b springapplication application run failed org springframework beans factory beancreationexception error creating bean with name armeriaserver defined in com linecorp armeria spring armeriaautoconfiguration bean instantiation via factory method failed nested exception is org springframework beans beaninstantiationexception failed to instantiate factory method armeriaserver threw exception nested exception is java lang stackoverflowerror at org springframework beans factory support constructorresolver instantiate constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory support defaultlistablebeanfactory preinstantiatesingletons defaultlistablebeanfactory java at org springframework context support abstractapplicationcontext finishbeanfactoryinitialization abstractapplicationcontext java at org springframework context support abstractapplicationcontext refresh abstractapplicationcontext java at org springframework boot springapplication refresh springapplication java at org springframework boot springapplication refreshcontext springapplication java at org springframework boot springapplication run springapplication java at org springframework boot builder springapplicationbuilder run springapplicationbuilder java at zipkin server zipkinserver main zipkinserver java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org springframework boot loader mainmethodrunner run mainmethodrunner java at org springframework boot loader launcher launch launcher java at org springframework boot loader launcher launch launcher java at org springframework boot loader propertieslauncher main propertieslauncher java caused by org springframework beans beaninstantiationexception failed to instantiate factory method armeriaserver threw exception nested exception is java lang stackoverflowerror at org springframework beans factory support simpleinstantiationstrategy instantiate simpleinstantiationstrategy java at org springframework beans factory support constructorresolver instantiate constructorresolver java more caused by java lang stackoverflowerror at org apache logging core pattern datepatternconverter getthreadlocalformatter datepatternconverter java at org apache logging core pattern datepatternconverter formatwithoutallocation datepatternconverter java at org apache logging core pattern datepatternconverter format datepatternconverter java at org apache logging core pattern datepatternconverter format datepatternconverter java at org apache logging core pattern patternformatter format patternformatter java at org springframework boot logging colorconverter format colorconverter java at org apache logging core pattern patternformatter format patternformatter java at org apache logging core layout patternlayout patternserializer toserializable patternlayout java at org apache logging core layout patternlayout totext patternlayout java at org apache logging core layout patternlayout encode patternlayout java at org apache logging core layout patternlayout encode patternlayout java at org apache logging core appender abstractoutputstreamappender directencodeevent abstractoutputstreamappender java at org apache logging core appender abstractoutputstreamappender tryappend abstractoutputstreamappender java at org apache logging core appender abstractoutputstreamappender append abstractoutputstreamappender java at org apache logging core config appendercontrol trycallappender appendercontrol java at org apache logging core config appendercontrol appendercontrol java at org apache logging core config appendercontrol callappenderpreventrecursion appendercontrol java at org apache logging core config appendercontrol callappender appendercontrol java at org apache logging core config loggerconfig callappenders loggerconfig java at org apache logging core config loggerconfig processlogevent loggerconfig java at org apache logging core config loggerconfig log loggerconfig java at org apache logging core config loggerconfig log loggerconfig java at org apache logging core config awaitcompletionreliabilitystrategy log awaitcompletionreliabilitystrategy java at org apache logging core logger log logger java at org apache logging spi abstractlogger trylogmessage abstractlogger java at org apache logging spi abstractlogger logmessagetrackrecursion abstractlogger java at org apache logging spi abstractlogger logmessagesafely abstractlogger java at org apache logging spi abstractlogger logmessage abstractlogger java at org apache logging spi abstractlogger logifenabled abstractlogger java at org apache logging info java at com linecorp armeria common flags getnormalized flags java at com linecorp armeria common flags getboolean flags java at com linecorp armeria common flags getboolean flags java at com linecorp armeria common flags useopenssl flags java at com linecorp armeria internal sslcontextutil createsslcontext sslcontextutil java at com linecorp armeria common flags useopenssl flags java at com linecorp armeria internal sslcontextutil createsslcontext sslcontextutil java at com linecorp armeria common flags useopenssl flags java at com linecorp armeria internal sslcontextutil createsslcontext sslcontextutil java
1
45,004
12,516,505,373
IssuesEvent
2020-06-03 09:29:50
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
DataTable: column/columns/row support passthrough attributes
defect
Column renderer will generate td/th elements, but passThroughAttributes Map items gets ignored for td elements, they are only rendered for th elements for dynamic columns component, the passthrough attributes are missing for th and td. for Row component either used in columnGroup or in a panelGrid, the passthrough attributes are missing in the client-side in the tr element. I verified that the attributes where set correctly in the passThroughAttributes Map of these components through the JSF PassThroughAttributeHandler **Environment:** - PF Version: _7.0_ - JSF + version: _Mojarra 2.3.9.SP06_
1.0
DataTable: column/columns/row support passthrough attributes - Column renderer will generate td/th elements, but passThroughAttributes Map items gets ignored for td elements, they are only rendered for th elements for dynamic columns component, the passthrough attributes are missing for th and td. for Row component either used in columnGroup or in a panelGrid, the passthrough attributes are missing in the client-side in the tr element. I verified that the attributes where set correctly in the passThroughAttributes Map of these components through the JSF PassThroughAttributeHandler **Environment:** - PF Version: _7.0_ - JSF + version: _Mojarra 2.3.9.SP06_
defect
datatable column columns row support passthrough attributes column renderer will generate td th elements but passthroughattributes map items gets ignored for td elements they are only rendered for th elements for dynamic columns component the passthrough attributes are missing for th and td for row component either used in columngroup or in a panelgrid the passthrough attributes are missing in the client side in the tr element i verified that the attributes where set correctly in the passthroughattributes map of these components through the jsf passthroughattributehandler environment pf version jsf version mojarra
1
66,381
20,164,473,461
IssuesEvent
2022-02-10 01:55:16
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Defect: facilities flagged 'removal complete' do not show on /admin/content/flagged
Defect vsa-facilities ⭐️ Product support Needs refining Platform CMS Team Sitewide CMS Team Lighthouse Facility API
## Describe the defect The View at /admin/content/flagged is not showing flagged items if they are only flagged "Removal complete". regardless of the filter setting for "Show "Removal complete" ## To Reproduce Steps to reproduce the behavior: 1. Flag a facility "Removal complete" 2. Go to /admin/content/flagged 3. Change the "Show "Removal complete" to either "-any-' or 'True' 4. The facility you just flagged should show, but it does not. ## Expected behavior 1. Flag a facility "Removal complete" 2. Go to /admin/content/flagged 3. Change the "Show "Removal complete" to either "-any-' or 'True' and click Filter 4. The facility you just flagged should show. 5. Change the "Show "Removal complete" to either "False' and click Filter 4. The facility you just flagged should NOT show. ## Screenshots If applicable, add screenshots to help explain your problem. ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please leave only the team that will do this work selected. - [x] `Platform CMS Team` - [x] `Sitewide CMS Team` - [ ] `⭐️ Content ops` - [ ] `⭐️ CMS experience` - [ ] `⭐️ Offices` - [x] `⭐️ Product support` - [ ] `⭐️ User support`
1.0
Defect: facilities flagged 'removal complete' do not show on /admin/content/flagged - ## Describe the defect The View at /admin/content/flagged is not showing flagged items if they are only flagged "Removal complete". regardless of the filter setting for "Show "Removal complete" ## To Reproduce Steps to reproduce the behavior: 1. Flag a facility "Removal complete" 2. Go to /admin/content/flagged 3. Change the "Show "Removal complete" to either "-any-' or 'True' 4. The facility you just flagged should show, but it does not. ## Expected behavior 1. Flag a facility "Removal complete" 2. Go to /admin/content/flagged 3. Change the "Show "Removal complete" to either "-any-' or 'True' and click Filter 4. The facility you just flagged should show. 5. Change the "Show "Removal complete" to either "False' and click Filter 4. The facility you just flagged should NOT show. ## Screenshots If applicable, add screenshots to help explain your problem. ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please leave only the team that will do this work selected. - [x] `Platform CMS Team` - [x] `Sitewide CMS Team` - [ ] `⭐️ Content ops` - [ ] `⭐️ CMS experience` - [ ] `⭐️ Offices` - [x] `⭐️ Product support` - [ ] `⭐️ User support`
defect
defect facilities flagged removal complete do not show on admin content flagged describe the defect the view at admin content flagged is not showing flagged items if they are only flagged removal complete regardless of the filter setting for show removal complete to reproduce steps to reproduce the behavior flag a facility removal complete go to admin content flagged change the show removal complete to either any or true the facility you just flagged should show but it does not expected behavior flag a facility removal complete go to admin content flagged change the show removal complete to either any or true and click filter the facility you just flagged should show change the show removal complete to either false and click filter the facility you just flagged should not show screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please leave only the team that will do this work selected platform cms team sitewide cms team ⭐️ content ops ⭐️ cms experience ⭐️ offices ⭐️ product support ⭐️ user support
1
11,315
2,649,110,646
IssuesEvent
2015-03-14 16:13:25
Paradoxianer/projectconceptor
https://api.github.com/repos/Paradoxianer/projectconceptor
closed
Implement Find Command
auto-migrated Priority-Medium Type-Defect
``` Find Command is unimplemented at the Moment First implement simple text search (Option Attribut or just Node search) Second implement "structure" search... like find all Nodes with 3 Conncetions ``` Original issue reported on code.google.com by `two4...@gmail.com` on 29 Dec 2013 at 10:42
1.0
Implement Find Command - ``` Find Command is unimplemented at the Moment First implement simple text search (Option Attribut or just Node search) Second implement "structure" search... like find all Nodes with 3 Conncetions ``` Original issue reported on code.google.com by `two4...@gmail.com` on 29 Dec 2013 at 10:42
defect
implement find command find command is unimplemented at the moment first implement simple text search option attribut or just node search second implement structure search like find all nodes with conncetions original issue reported on code google com by gmail com on dec at
1
17,893
3,013,568,576
IssuesEvent
2015-07-29 09:46:42
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Invalid XML returned by YEngine.getStateForCase(YIdentifier caseID)
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Create a YEngine object 2. Load a specification 3. Start a case 4. Call YEngine.getStateForCase(caseID) with the case's ID What is the expected output? What do you see instead? Expected (only concerns this one element): <nextElementRef id="foobar" documentation=""/> Received: <nextElementRef id="foobar" documentation=""> Please use labels and text to provide additional information. The slash '/' closing an empty XML tag is missing. It needs to be added in YEngine.class:999. The problem exists in trunk/release2/src/org/yawlfoundation/yawl/engine/YEngine.java revision 1798. Don't know about other revisions. ``` Original issue reported on code.google.com by `florian....@gmail.com` on 5 Dec 2011 at 3:13
1.0
Invalid XML returned by YEngine.getStateForCase(YIdentifier caseID) - ``` What steps will reproduce the problem? 1. Create a YEngine object 2. Load a specification 3. Start a case 4. Call YEngine.getStateForCase(caseID) with the case's ID What is the expected output? What do you see instead? Expected (only concerns this one element): <nextElementRef id="foobar" documentation=""/> Received: <nextElementRef id="foobar" documentation=""> Please use labels and text to provide additional information. The slash '/' closing an empty XML tag is missing. It needs to be added in YEngine.class:999. The problem exists in trunk/release2/src/org/yawlfoundation/yawl/engine/YEngine.java revision 1798. Don't know about other revisions. ``` Original issue reported on code.google.com by `florian....@gmail.com` on 5 Dec 2011 at 3:13
defect
invalid xml returned by yengine getstateforcase yidentifier caseid what steps will reproduce the problem create a yengine object load a specification start a case call yengine getstateforcase caseid with the case s id what is the expected output what do you see instead expected only concerns this one element nextelementref id foobar documentation received please use labels and text to provide additional information the slash closing an empty xml tag is missing it needs to be added in yengine class the problem exists in trunk src org yawlfoundation yawl engine yengine java revision don t know about other revisions original issue reported on code google com by florian gmail com on dec at
1
290,993
8,915,908,255
IssuesEvent
2019-01-19 11:18:19
ictsc/ictsc-score-server
https://api.github.com/repos/ictsc/ictsc-score-server
opened
スポンサー部屋用ページ
UI priority: high
スポンサー部屋のディスプレイで表示する画面が欲しい * スコアと問題の解き具合を表示する * スポンサー部屋にこの一画面を表示すれば情報が分かるようにする * 一画面で表示できるサイズに収めるのは厳しいかもしれない 要相談 * participantとno_loginから見えないようにする * そもそも他チームが取れる情報はAPI側で制限されてるので問題ない?
1.0
スポンサー部屋用ページ - スポンサー部屋のディスプレイで表示する画面が欲しい * スコアと問題の解き具合を表示する * スポンサー部屋にこの一画面を表示すれば情報が分かるようにする * 一画面で表示できるサイズに収めるのは厳しいかもしれない 要相談 * participantとno_loginから見えないようにする * そもそも他チームが取れる情報はAPI側で制限されてるので問題ない?
non_defect
スポンサー部屋用ページ スポンサー部屋のディスプレイで表示する画面が欲しい スコアと問題の解き具合を表示する スポンサー部屋にこの一画面を表示すれば情報が分かるようにする 一画面で表示できるサイズに収めるのは厳しいかもしれない 要相談 participantとno loginから見えないようにする そもそも他チームが取れる情報はapi側で制限されてるので問題ない
0
275,282
20,915,900,802
IssuesEvent
2022-03-24 13:25:16
dhowe/AdNauseam
https://api.github.com/repos/dhowe/AdNauseam
opened
Documentation Changes about Strict Blocking
Documentation
### Original post we currently have: ----------- #### What is ‘Strict Blocking’, and when should I use it? By default AdNauseam only blocks requests that do not interfere with the collection of ads. However, when ‘Strict Blocking’ mode is enabled, AdNauseam will instead block all requests that match a blocking rule from an enabled third-party list. This means that LESS ads will be collected, placed in the vault, and later clicked, and that AdNauseam will be LESS effective in its primary function (for example, the majority of Google ads won’t be collected or clicked). Thus for most users ‘Strict Blocking’ mode is NOT RECOMMENDED. However, if you are primarily interested in blocking ads, you _may_ see better performance with ‘Strict Blocking’ enabled (depending on your settings and the specific sites you visit). You can read more about blocking and hiding rules in AdNauseam [here](https://github.com/dhowe/AdNauseam/wiki/Developer-FAQ#what-is-the-relationship-between-blocking-and-hiding-rules-in-ublock-and-adn). Generally speaking this mode is for advanced users with specific use-cases (e.g., testing), so please be sure to understand the ramifications before enabling it. A BETTER OPTION is often to use ‘Strict Blocking’ in combination with dynamic filtering rules. You can follow [uBlock's dynamic filtering rule syntax](https://github.com/gorhill/uBlock/wiki/Dynamic-filtering:-rule-syntax) to create your own strict blocking rules to block requests based on sites, 3rd-party domains, and request types. To compose strict-blocking rules, you will need to use the syntax `strictBlock` for the "action" component. Here are a few examples: * To strict-block all the requests for facebook.com, you can use the following rule: `facebook.com * * strictBlock` * To strict-block all the requests coming from doubleclick.net for a certain site, you can use the following rule (this will stop google ads from rendering and prevent them from being collected and clicked by AdNauseam): `mysite.com doubleclick.net * strictBlock` * To strict-block all 3rd-party scripts for a certain site, you can use the following rule: `facebook.com * 3p-script strictBlock` * Please note the difference between `strictBlock` and `block`. The `block` action blocks requests according to the dynamic filtering rules, while `strictBlock` only blocks a request if it triggers a blocking rule in one of the filter lists.
1.0
Documentation Changes about Strict Blocking - ### Original post we currently have: ----------- #### What is ‘Strict Blocking’, and when should I use it? By default AdNauseam only blocks requests that do not interfere with the collection of ads. However, when ‘Strict Blocking’ mode is enabled, AdNauseam will instead block all requests that match a blocking rule from an enabled third-party list. This means that LESS ads will be collected, placed in the vault, and later clicked, and that AdNauseam will be LESS effective in its primary function (for example, the majority of Google ads won’t be collected or clicked). Thus for most users ‘Strict Blocking’ mode is NOT RECOMMENDED. However, if you are primarily interested in blocking ads, you _may_ see better performance with ‘Strict Blocking’ enabled (depending on your settings and the specific sites you visit). You can read more about blocking and hiding rules in AdNauseam [here](https://github.com/dhowe/AdNauseam/wiki/Developer-FAQ#what-is-the-relationship-between-blocking-and-hiding-rules-in-ublock-and-adn). Generally speaking this mode is for advanced users with specific use-cases (e.g., testing), so please be sure to understand the ramifications before enabling it. A BETTER OPTION is often to use ‘Strict Blocking’ in combination with dynamic filtering rules. You can follow [uBlock's dynamic filtering rule syntax](https://github.com/gorhill/uBlock/wiki/Dynamic-filtering:-rule-syntax) to create your own strict blocking rules to block requests based on sites, 3rd-party domains, and request types. To compose strict-blocking rules, you will need to use the syntax `strictBlock` for the "action" component. Here are a few examples: * To strict-block all the requests for facebook.com, you can use the following rule: `facebook.com * * strictBlock` * To strict-block all the requests coming from doubleclick.net for a certain site, you can use the following rule (this will stop google ads from rendering and prevent them from being collected and clicked by AdNauseam): `mysite.com doubleclick.net * strictBlock` * To strict-block all 3rd-party scripts for a certain site, you can use the following rule: `facebook.com * 3p-script strictBlock` * Please note the difference between `strictBlock` and `block`. The `block` action blocks requests according to the dynamic filtering rules, while `strictBlock` only blocks a request if it triggers a blocking rule in one of the filter lists.
non_defect
documentation changes about strict blocking original post we currently have what is ‘strict blocking’ and when should i use it by default adnauseam only blocks requests that do not interfere with the collection of ads however when ‘strict blocking’ mode is enabled adnauseam will instead block all requests that match a blocking rule from an enabled third party list this means that less ads will be collected placed in the vault and later clicked and that adnauseam will be less effective in its primary function for example the majority of google ads won’t be collected or clicked thus for most users ‘strict blocking’ mode is not recommended however if you are primarily interested in blocking ads you may see better performance with ‘strict blocking’ enabled depending on your settings and the specific sites you visit you can read more about blocking and hiding rules in adnauseam generally speaking this mode is for advanced users with specific use cases e g testing so please be sure to understand the ramifications before enabling it a better option is often to use ‘strict blocking’ in combination with dynamic filtering rules you can follow to create your own strict blocking rules to block requests based on sites party domains and request types to compose strict blocking rules you will need to use the syntax strictblock for the action component here are a few examples to strict block all the requests for facebook com you can use the following rule facebook com strictblock to strict block all the requests coming from doubleclick net for a certain site you can use the following rule this will stop google ads from rendering and prevent them from being collected and clicked by adnauseam mysite com doubleclick net strictblock to strict block all party scripts for a certain site you can use the following rule facebook com script strictblock please note the difference between strictblock and block the block action blocks requests according to the dynamic filtering rules while strictblock only blocks a request if it triggers a blocking rule in one of the filter lists
0
536,269
15,707,018,403
IssuesEvent
2021-03-26 18:15:04
sopra-fs21-group-03/Client
https://api.github.com/repos/sopra-fs21-group-03/Client
opened
Every user can see what the users before have done so far
medium priority task
Time estimate: 0.7h "This task is part of user story #13"
1.0
Every user can see what the users before have done so far - Time estimate: 0.7h "This task is part of user story #13"
non_defect
every user can see what the users before have done so far time estimate this task is part of user story
0
8,865
2,612,919,059
IssuesEvent
2015-02-27 17:30:42
chrsmith/windows-package-manager
https://api.github.com/repos/chrsmith/windows-package-manager
closed
Remove default repo not work
auto-migrated Milestone-1.18 Type-Defect
``` 1. npackdcl remove-repo -u https://windows-package-manager.googlecode.com/hg/repository/Rep.xml The repository was removed successfully 2. npackdcl remove-repo -u https://windows-package-manager.googlecode.com/hg/repository/Rep.xml The repository was removed successfully Npackd says that the repository is removed but this one is not removed! Windows 7 | NpackdCL-1.17.9 ``` Original issue reported on code.google.com by `gregory....@gmail.com` on 1 Mar 2013 at 2:56
1.0
Remove default repo not work - ``` 1. npackdcl remove-repo -u https://windows-package-manager.googlecode.com/hg/repository/Rep.xml The repository was removed successfully 2. npackdcl remove-repo -u https://windows-package-manager.googlecode.com/hg/repository/Rep.xml The repository was removed successfully Npackd says that the repository is removed but this one is not removed! Windows 7 | NpackdCL-1.17.9 ``` Original issue reported on code.google.com by `gregory....@gmail.com` on 1 Mar 2013 at 2:56
defect
remove default repo not work npackdcl remove repo u the repository was removed successfully npackdcl remove repo u the repository was removed successfully npackd says that the repository is removed but this one is not removed windows npackdcl original issue reported on code google com by gregory gmail com on mar at
1
32,053
6,694,168,235
IssuesEvent
2017-10-10 00:03:19
ericpromislow/tabhunter
https://api.github.com/repos/ericpromislow/tabhunter
closed
updated firefox, now tabhunter no longer moves to correct location in tab bar when used
auto-migrated Priority-Medium Type-Defect
``` I just upgraded firefox from 14.0.1, where tabhunter worked perfectly, to 25.0.1. Now tabhunter properly shifts the window focus to the selected tab, but it no longer moves to the right place in the tabbar. So the selected tab may not even show in the tabbar when one has many tabs open. What steps will reproduce the problem? 1. Select a tab to shift to in tabhunter, and double click 2. 3. What is the expected output? What do you see instead? expect to not only shift the focus of the main window to the selected tab, but also for it to move the tab bar to the location with that selected tab. Instead, the focus is properly shifted, but the tab bar isn't shifted such that the selected tab can be seen. I have to scroll through the entire tab bar to find the tab I used tab hunter to go to. What version of the product are you using? On what operating system? v. 1.0.2, in firefox 25.0.1, on Windows 7 home premium Please provide any additional information below. ``` Original issue reported on code.google.com by `Eclec...@gmail.com` on 27 Nov 2013 at 6:47
1.0
updated firefox, now tabhunter no longer moves to correct location in tab bar when used - ``` I just upgraded firefox from 14.0.1, where tabhunter worked perfectly, to 25.0.1. Now tabhunter properly shifts the window focus to the selected tab, but it no longer moves to the right place in the tabbar. So the selected tab may not even show in the tabbar when one has many tabs open. What steps will reproduce the problem? 1. Select a tab to shift to in tabhunter, and double click 2. 3. What is the expected output? What do you see instead? expect to not only shift the focus of the main window to the selected tab, but also for it to move the tab bar to the location with that selected tab. Instead, the focus is properly shifted, but the tab bar isn't shifted such that the selected tab can be seen. I have to scroll through the entire tab bar to find the tab I used tab hunter to go to. What version of the product are you using? On what operating system? v. 1.0.2, in firefox 25.0.1, on Windows 7 home premium Please provide any additional information below. ``` Original issue reported on code.google.com by `Eclec...@gmail.com` on 27 Nov 2013 at 6:47
defect
updated firefox now tabhunter no longer moves to correct location in tab bar when used i just upgraded firefox from where tabhunter worked perfectly to now tabhunter properly shifts the window focus to the selected tab but it no longer moves to the right place in the tabbar so the selected tab may not even show in the tabbar when one has many tabs open what steps will reproduce the problem select a tab to shift to in tabhunter and double click what is the expected output what do you see instead expect to not only shift the focus of the main window to the selected tab but also for it to move the tab bar to the location with that selected tab instead the focus is properly shifted but the tab bar isn t shifted such that the selected tab can be seen i have to scroll through the entire tab bar to find the tab i used tab hunter to go to what version of the product are you using on what operating system v in firefox on windows home premium please provide any additional information below original issue reported on code google com by eclec gmail com on nov at
1
344,925
24,835,262,435
IssuesEvent
2022-10-26 08:21:01
keptn/community
https://api.github.com/repos/keptn/community
opened
Create a community/hackathons page
good first issue community documentation
We want to have a online hackathon/Hacktoiberfest contributing guide, similar to https://www.jenkins.io/events/hacktoberfest/ that would describe everything for newcomer episodic contributors
1.0
Create a community/hackathons page - We want to have a online hackathon/Hacktoiberfest contributing guide, similar to https://www.jenkins.io/events/hacktoberfest/ that would describe everything for newcomer episodic contributors
non_defect
create a community hackathons page we want to have a online hackathon hacktoiberfest contributing guide similar to that would describe everything for newcomer episodic contributors
0
53,898
29,045,495,377
IssuesEvent
2023-05-13 13:59:23
IntelPython/dpnp
https://api.github.com/repos/IntelPython/dpnp
opened
L2 norm: dpnp on CPU is 5 times slower than NumPy
performance
Results from dpbench of L2 norm workload (running on a laptop with 11th Gen processor and Iris Xe graphics): | workload | numpy | dpnp CPU | dpnp GPU | size | | :--- | :---: | :---: | :---: | :---: | | L2 norm | 76 ms | 366 ms | 267 ms | 4194304 | demonstrates dpnp execution time on CPU which is 5 times greater than one of NumPy.
True
L2 norm: dpnp on CPU is 5 times slower than NumPy - Results from dpbench of L2 norm workload (running on a laptop with 11th Gen processor and Iris Xe graphics): | workload | numpy | dpnp CPU | dpnp GPU | size | | :--- | :---: | :---: | :---: | :---: | | L2 norm | 76 ms | 366 ms | 267 ms | 4194304 | demonstrates dpnp execution time on CPU which is 5 times greater than one of NumPy.
non_defect
norm dpnp on cpu is times slower than numpy results from dpbench of norm workload running on a laptop with gen processor and iris xe graphics workload numpy dpnp cpu dpnp gpu size norm ms ms ms demonstrates dpnp execution time on cpu which is times greater than one of numpy
0
29,824
5,909,086,393
IssuesEvent
2017-05-19 22:27:55
sukona/Grapevine
https://api.github.com/repos/sukona/Grapevine
closed
Mono lacks support for ExtendedProtectionSelector
defect Server
This lack of support leads to such problems as these, recently logged on StackOverflow: [Running Grapevine REST server in Mono Docker container](http://stackoverflow.com/questions/41270527/running-grapevine-rest-server-in-mono-docker-container) ``` Error: System.TypeLoadException: Could not load type 'Grapevine.Interfaces.Server.HttpListener' from assembly 'Grapevine, Version=4.0.0.195, Culture=neutral, PublicKeyToken=null'. at Boerse.BoersenApplication.Main (System.String[] args) [0x0002f] in <407ced228a394aa7b9fc2fa883a239a9>:0 TypeRef ResolutionScope not yet handled (57) for .ExtendedProtectionSelector in image /boerse/Grapevine.dll Could not load signature of Grapevine.Interfaces.Server.HttpListener:get_ExtendedProtectionSelectorDelegate due to: Could not resolve type with token 0100003a assembly: type:ExtendedProtectionSelector member:<none> TypeRef ResolutionScope not yet handled (57) for .ExtendedProtectionSelector in image /boerse/Grapevine.dll Could not load signature of Grapevine.Interfaces.Server.IHttpListener:get_ExtendedProtectionSelectorDelegate due to: Could not resolve type with token 0100003a assembly: type:ExtendedProtectionSelector member:<none> ``` [Cannot load type HttpListener in simple Grapevine server example](http://stackoverflow.com/questions/41087072/cannot-load-type-httplistener-in-simple-grapevine-server-example) ``` System.TypeLoadException: Could not load type 'Grapevine.Interfaces.Server.HttpListener' from assembly 'Grapevine, Version=4.0.0.195, Culture=neutral, PublicKeyToken=null'. at Grapevine.Server.RestServer..ctor () [0x00006] in <5da3c1fcf3364795b3df98bfc8b714aa>:0 at TestServer.MainClass.Main (System.String[] args) [0x0000b] in /Users/blah/Projects/Test/TestServer/Program.cs:12 ``` Both of which *might* be resolved once there is a Grapevine version that targets .NET 4.5 or 4.6.
1.0
Mono lacks support for ExtendedProtectionSelector - This lack of support leads to such problems as these, recently logged on StackOverflow: [Running Grapevine REST server in Mono Docker container](http://stackoverflow.com/questions/41270527/running-grapevine-rest-server-in-mono-docker-container) ``` Error: System.TypeLoadException: Could not load type 'Grapevine.Interfaces.Server.HttpListener' from assembly 'Grapevine, Version=4.0.0.195, Culture=neutral, PublicKeyToken=null'. at Boerse.BoersenApplication.Main (System.String[] args) [0x0002f] in <407ced228a394aa7b9fc2fa883a239a9>:0 TypeRef ResolutionScope not yet handled (57) for .ExtendedProtectionSelector in image /boerse/Grapevine.dll Could not load signature of Grapevine.Interfaces.Server.HttpListener:get_ExtendedProtectionSelectorDelegate due to: Could not resolve type with token 0100003a assembly: type:ExtendedProtectionSelector member:<none> TypeRef ResolutionScope not yet handled (57) for .ExtendedProtectionSelector in image /boerse/Grapevine.dll Could not load signature of Grapevine.Interfaces.Server.IHttpListener:get_ExtendedProtectionSelectorDelegate due to: Could not resolve type with token 0100003a assembly: type:ExtendedProtectionSelector member:<none> ``` [Cannot load type HttpListener in simple Grapevine server example](http://stackoverflow.com/questions/41087072/cannot-load-type-httplistener-in-simple-grapevine-server-example) ``` System.TypeLoadException: Could not load type 'Grapevine.Interfaces.Server.HttpListener' from assembly 'Grapevine, Version=4.0.0.195, Culture=neutral, PublicKeyToken=null'. at Grapevine.Server.RestServer..ctor () [0x00006] in <5da3c1fcf3364795b3df98bfc8b714aa>:0 at TestServer.MainClass.Main (System.String[] args) [0x0000b] in /Users/blah/Projects/Test/TestServer/Program.cs:12 ``` Both of which *might* be resolved once there is a Grapevine version that targets .NET 4.5 or 4.6.
defect
mono lacks support for extendedprotectionselector this lack of support leads to such problems as these recently logged on stackoverflow error system typeloadexception could not load type grapevine interfaces server httplistener from assembly grapevine version culture neutral publickeytoken null at boerse boersenapplication main system string args in typeref resolutionscope not yet handled for extendedprotectionselector in image boerse grapevine dll could not load signature of grapevine interfaces server httplistener get extendedprotectionselectordelegate due to could not resolve type with token assembly type extendedprotectionselector member typeref resolutionscope not yet handled for extendedprotectionselector in image boerse grapevine dll could not load signature of grapevine interfaces server ihttplistener get extendedprotectionselectordelegate due to could not resolve type with token assembly type extendedprotectionselector member system typeloadexception could not load type grapevine interfaces server httplistener from assembly grapevine version culture neutral publickeytoken null at grapevine server restserver ctor in at testserver mainclass main system string args in users blah projects test testserver program cs both of which might be resolved once there is a grapevine version that targets net or
1
82,068
31,897,610,573
IssuesEvent
2023-09-18 04:22:14
zed-industries/community
https://api.github.com/repos/zed-industries/community
closed
cannot close the inline assist API key pane
defect triage admin read
### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it Press the inline assist button. An pane occupying the right half of the editor is shown asking for the OpenAI API key. There's no button or command to close it. ### Environment Zed: v0.103.1 (stable) OS: macOS 13.5.2 Memory: 32 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
1.0
cannot close the inline assist API key pane - ### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it Press the inline assist button. An pane occupying the right half of the editor is shown asking for the OpenAI API key. There's no button or command to close it. ### Environment Zed: v0.103.1 (stable) OS: macOS 13.5.2 Memory: 32 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
defect
cannot close the inline assist api key pane check for existing issues completed describe the bug provide steps to reproduce it press the inline assist button an pane occupying the right half of the editor is shown asking for the openai api key there s no button or command to close it environment zed stable os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature no response if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response
1
308,840
23,269,846,570
IssuesEvent
2022-08-04 21:27:50
LLNL/Abmarl
https://api.github.com/repos/LLNL/Abmarl
opened
Restructure docs for updated workflow
documentation
Having added the OpenSpielWrapper and our own Training prototype, Abmarl is becoming less dependent on RLlib and more of an all-purpose tool. After we finish #275 and #236, we'll be even less dependent on RLlib. Our documentation still has RLlib at the center, and we'll need to restructure them. The docs should be restructured so that it matches our project structure better. We'll have a page on the RL-loop and reinforcement learning in general. We'll have a page on workflow with sub pages for Abmarl's trainers, RLlib integration, OpenSpiel integration, etc. We'll have a page on GSF. We'll have the other pages as well. We may even want to add a "quick start" page that highlights all the features at once. We should also add new projects (here in GitHub) for integration with RLlib, OpenSpiel, etc. Should be done after #275 and #236
1.0
Restructure docs for updated workflow - Having added the OpenSpielWrapper and our own Training prototype, Abmarl is becoming less dependent on RLlib and more of an all-purpose tool. After we finish #275 and #236, we'll be even less dependent on RLlib. Our documentation still has RLlib at the center, and we'll need to restructure them. The docs should be restructured so that it matches our project structure better. We'll have a page on the RL-loop and reinforcement learning in general. We'll have a page on workflow with sub pages for Abmarl's trainers, RLlib integration, OpenSpiel integration, etc. We'll have a page on GSF. We'll have the other pages as well. We may even want to add a "quick start" page that highlights all the features at once. We should also add new projects (here in GitHub) for integration with RLlib, OpenSpiel, etc. Should be done after #275 and #236
non_defect
restructure docs for updated workflow having added the openspielwrapper and our own training prototype abmarl is becoming less dependent on rllib and more of an all purpose tool after we finish and we ll be even less dependent on rllib our documentation still has rllib at the center and we ll need to restructure them the docs should be restructured so that it matches our project structure better we ll have a page on the rl loop and reinforcement learning in general we ll have a page on workflow with sub pages for abmarl s trainers rllib integration openspiel integration etc we ll have a page on gsf we ll have the other pages as well we may even want to add a quick start page that highlights all the features at once we should also add new projects here in github for integration with rllib openspiel etc should be done after and
0
289,922
32,008,580,244
IssuesEvent
2023-09-21 16:20:59
quarkusio/quarkus
https://api.github.com/repos/quarkusio/quarkus
closed
Renamed configuration property to enable CORS filter not recognized
kind/bug area/config area/security
### Describe the bug The [Quarkus 3.4 migration guide](https://github.com/quarkusio/quarkus/wiki/Migration-Guide-3.4) states that the configuration property `quarkus.http.cors` has been renamed to `quarkus.http.cors.enabled`. When renaming this configuration property accordingly after upgrading to Qkuarkus `3.4.1` our application doesn't seem to recognize the configuration property `quarkus.http.cors.enabled`. ``` > ./gradlew quarkusDev ... 2023-09-21 16:09:01,377 WARN [io.qua.config] (Quarkus Main Thread) Unrecognized configuration key "quarkus.http.cors.enabled" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ... ``` When running tests, the error message is slightly different: ``` > ./gradlew check ... Unrecognized configuration key "%s" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ... ``` The "old" configuration property `quarkus.http.cors` works as expected. Also, the [HTTP reference guide](https://quarkus.io/guides/http-reference#cors-filter) doesn't mention the renamed property at all. ### Expected behavior Quarkus correctly recognizes the renamed configuration property `quarkus.http.cors.enabled`. ### Actual behavior Using a CORS configuration like ```properties quarkus.http.cors.enabled = true quarkus.http.cors.origins = /.*/ quarkus.http.cors.methods = GET,PUT,POST,DELETE quarkus.http.cors.headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.exposed-headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.access-control-max-age = 24H quarkus.http.cors.access-control-allow-credentials = true ``` results in the following error message at application startup: ``` Unrecognized configuration key "quarkus.http.cors.enabled" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ``` ### How to Reproduce? Create a fresh Quarkus project and add a corresponding CORS filter configuration to application configuration: 1. Create a fresh Quarkus project: ```sh quarkus create app my-groupId:my-artifactId \ --extension=resteasy-reactive,resteasy-reactive-jackson \ --gradle-kotlin-dsl ``` 1. Add CORS configuration filter with renamed configuration property to `application.properties`: ```properties quarkus.http.cors.enabled = true quarkus.http.cors.origins = /.*/ quarkus.http.cors.methods = GET,PUT,POST,DELETE quarkus.http.cors.headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.exposed-headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.access-control-max-age = 24H quarkus.http.cors.access-control-allow-credentials = true ``` 1. Start application in development mode: ```sh ./gradlew quarkusDev ``` 1. Run tests: ```sh ./gradlew check ``` ### Output of `uname -a` or `ver` _No response_ ### Output of `java -version` openjdk version "17.0.7" 2023-04-18 LTS ### GraalVM version (if different from Java) _No response_ ### Quarkus version or git rev 3.4.1 ### Build tool (ie. output of `mvnw --version` or `gradlew --version`) Gradle 8.3 ### Additional information _No response_
True
Renamed configuration property to enable CORS filter not recognized - ### Describe the bug The [Quarkus 3.4 migration guide](https://github.com/quarkusio/quarkus/wiki/Migration-Guide-3.4) states that the configuration property `quarkus.http.cors` has been renamed to `quarkus.http.cors.enabled`. When renaming this configuration property accordingly after upgrading to Qkuarkus `3.4.1` our application doesn't seem to recognize the configuration property `quarkus.http.cors.enabled`. ``` > ./gradlew quarkusDev ... 2023-09-21 16:09:01,377 WARN [io.qua.config] (Quarkus Main Thread) Unrecognized configuration key "quarkus.http.cors.enabled" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ... ``` When running tests, the error message is slightly different: ``` > ./gradlew check ... Unrecognized configuration key "%s" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ... ``` The "old" configuration property `quarkus.http.cors` works as expected. Also, the [HTTP reference guide](https://quarkus.io/guides/http-reference#cors-filter) doesn't mention the renamed property at all. ### Expected behavior Quarkus correctly recognizes the renamed configuration property `quarkus.http.cors.enabled`. ### Actual behavior Using a CORS configuration like ```properties quarkus.http.cors.enabled = true quarkus.http.cors.origins = /.*/ quarkus.http.cors.methods = GET,PUT,POST,DELETE quarkus.http.cors.headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.exposed-headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.access-control-max-age = 24H quarkus.http.cors.access-control-allow-credentials = true ``` results in the following error message at application startup: ``` Unrecognized configuration key "quarkus.http.cors.enabled" was provided; it will be ignored; verify that the dependency extension for this configuration is set or that you did not make a typo ``` ### How to Reproduce? Create a fresh Quarkus project and add a corresponding CORS filter configuration to application configuration: 1. Create a fresh Quarkus project: ```sh quarkus create app my-groupId:my-artifactId \ --extension=resteasy-reactive,resteasy-reactive-jackson \ --gradle-kotlin-dsl ``` 1. Add CORS configuration filter with renamed configuration property to `application.properties`: ```properties quarkus.http.cors.enabled = true quarkus.http.cors.origins = /.*/ quarkus.http.cors.methods = GET,PUT,POST,DELETE quarkus.http.cors.headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.exposed-headers = Content-Type,Content-Length,Authorization,X-Requested-With,Accept,Origin quarkus.http.cors.access-control-max-age = 24H quarkus.http.cors.access-control-allow-credentials = true ``` 1. Start application in development mode: ```sh ./gradlew quarkusDev ``` 1. Run tests: ```sh ./gradlew check ``` ### Output of `uname -a` or `ver` _No response_ ### Output of `java -version` openjdk version "17.0.7" 2023-04-18 LTS ### GraalVM version (if different from Java) _No response_ ### Quarkus version or git rev 3.4.1 ### Build tool (ie. output of `mvnw --version` or `gradlew --version`) Gradle 8.3 ### Additional information _No response_
non_defect
renamed configuration property to enable cors filter not recognized describe the bug the states that the configuration property quarkus http cors has been renamed to quarkus http cors enabled when renaming this configuration property accordingly after upgrading to qkuarkus our application doesn t seem to recognize the configuration property quarkus http cors enabled gradlew quarkusdev warn quarkus main thread unrecognized configuration key quarkus http cors enabled was provided it will be ignored verify that the dependency extension for this configuration is set or that you did not make a typo when running tests the error message is slightly different gradlew check unrecognized configuration key s was provided it will be ignored verify that the dependency extension for this configuration is set or that you did not make a typo the old configuration property quarkus http cors works as expected also the doesn t mention the renamed property at all expected behavior quarkus correctly recognizes the renamed configuration property quarkus http cors enabled actual behavior using a cors configuration like properties quarkus http cors enabled true quarkus http cors origins quarkus http cors methods get put post delete quarkus http cors headers content type content length authorization x requested with accept origin quarkus http cors exposed headers content type content length authorization x requested with accept origin quarkus http cors access control max age quarkus http cors access control allow credentials true results in the following error message at application startup unrecognized configuration key quarkus http cors enabled was provided it will be ignored verify that the dependency extension for this configuration is set or that you did not make a typo how to reproduce create a fresh quarkus project and add a corresponding cors filter configuration to application configuration create a fresh quarkus project sh quarkus create app my groupid my artifactid extension resteasy reactive resteasy reactive jackson gradle kotlin dsl add cors configuration filter with renamed configuration property to application properties properties quarkus http cors enabled true quarkus http cors origins quarkus http cors methods get put post delete quarkus http cors headers content type content length authorization x requested with accept origin quarkus http cors exposed headers content type content length authorization x requested with accept origin quarkus http cors access control max age quarkus http cors access control allow credentials true start application in development mode sh gradlew quarkusdev run tests sh gradlew check output of uname a or ver no response output of java version openjdk version lts graalvm version if different from java no response quarkus version or git rev build tool ie output of mvnw version or gradlew version gradle additional information no response
0
16,679
21,781,964,371
IssuesEvent
2022-05-13 20:03:09
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
[FR] Lens shading map (DNG GainMap) support
feature: enhancement scope: camera support scope: image processing
Some smartphone RAW DNGs contain a lens shading map, which is like a generalization of vignetting correction. Unlike the existing vignetting correction in lensfun, it can be an arbitrary shape rather than radially symmetric, and a separate map is applied to each raw color channel separately to correct spatially varying color casts. Also, it's generated for each individual raw image rather than being constant for a specific camera+lens module. I'm not sure how widely used this feature is across various phones and camera apps. The RAW DNGs from the Google Camera app on my Pixel 4a have it and the vignetting is pretty noticeable if the correction is not done. The [Android Camera2 API documentation](https://developer.android.com/reference/android/hardware/camera2/CaptureResult#STATISTICS_LENS_SHADING_CORRECTION_MAP) has a detailed description of how the correction works. The [Adobe DNG spec](https://wwwimages.adobe.com/content/dam/Adobe/en/products/photoshop/pdfs/dng_spec_1.5.0.0.pdf) describes how the lens shading map is encoded as GainMap opcodes within the OpcodeList2 exif tag. There is a [working implementation in ART](https://bitbucket.org/agriggio/art/src/master/rtengine/gainmap.cc) - it is enabled by the "Flat-Field" module in the Raw category when "Embedded in metadata" is checked. I have been trying to figure out how this could be implemented in Darktable. According to the DNG spec, as a stage opcode (in OpcodeList2) the gain map should be applied to linear raw data after black level subtraction but before demosaicing. For the typical case of a RGGB Bayer sensor where the GainMap for both of the G channels is identical (like from Pixel 4a) it would probably end up with the same result if the GainMaps were applied to the demosaiced RGB image prior to input color profile, but it probably makes more sense to apply it to the raw data according to the spec. One possibility is to do the correction within RawSpeed. It already has the ability to parse DNG opcodes (other than GainMap), but currently this is only done for lossy DNGs. It could be modified to also apply black level subtraction and GainMaps to RAW DNGs that have the gain map, and then it would just output 0 for the black levels so that rawprepare does not perform any further black level subtraction. I'm not sure how the UI would work if there was a need to make this correction optional, since it wouldn't be part of any module. Another possibility is to do the correction within the pipeline. I think this would be more complicated to implement. It could be its own module that comes after rawprepare, or maybe it could be added to rawprepare. Similarly to #7092 it would need to access the additional info from the exif which is not currently stored in dt_image_t or the image database. It could be awkward to add there because it can be quite large (20kb from the Pixel 4a) and there is not an explicit upper bound on the size defined by the file format. Maybe it could be managed similarly to dt_image_t.profile - it's not stored in the structure or the sql database, there is only a pointer to it. The colorin module loads it from the image file, allocates memory for it dynamically, and it's freed when it is removed from the image cache. The module implementing the gainmap could do something similar - if it hasn't been loaded into dt_image_t already, read it out of the exif, allocate memory, and store a pointer to it in dt_image_t.
1.0
[FR] Lens shading map (DNG GainMap) support - Some smartphone RAW DNGs contain a lens shading map, which is like a generalization of vignetting correction. Unlike the existing vignetting correction in lensfun, it can be an arbitrary shape rather than radially symmetric, and a separate map is applied to each raw color channel separately to correct spatially varying color casts. Also, it's generated for each individual raw image rather than being constant for a specific camera+lens module. I'm not sure how widely used this feature is across various phones and camera apps. The RAW DNGs from the Google Camera app on my Pixel 4a have it and the vignetting is pretty noticeable if the correction is not done. The [Android Camera2 API documentation](https://developer.android.com/reference/android/hardware/camera2/CaptureResult#STATISTICS_LENS_SHADING_CORRECTION_MAP) has a detailed description of how the correction works. The [Adobe DNG spec](https://wwwimages.adobe.com/content/dam/Adobe/en/products/photoshop/pdfs/dng_spec_1.5.0.0.pdf) describes how the lens shading map is encoded as GainMap opcodes within the OpcodeList2 exif tag. There is a [working implementation in ART](https://bitbucket.org/agriggio/art/src/master/rtengine/gainmap.cc) - it is enabled by the "Flat-Field" module in the Raw category when "Embedded in metadata" is checked. I have been trying to figure out how this could be implemented in Darktable. According to the DNG spec, as a stage opcode (in OpcodeList2) the gain map should be applied to linear raw data after black level subtraction but before demosaicing. For the typical case of a RGGB Bayer sensor where the GainMap for both of the G channels is identical (like from Pixel 4a) it would probably end up with the same result if the GainMaps were applied to the demosaiced RGB image prior to input color profile, but it probably makes more sense to apply it to the raw data according to the spec. One possibility is to do the correction within RawSpeed. It already has the ability to parse DNG opcodes (other than GainMap), but currently this is only done for lossy DNGs. It could be modified to also apply black level subtraction and GainMaps to RAW DNGs that have the gain map, and then it would just output 0 for the black levels so that rawprepare does not perform any further black level subtraction. I'm not sure how the UI would work if there was a need to make this correction optional, since it wouldn't be part of any module. Another possibility is to do the correction within the pipeline. I think this would be more complicated to implement. It could be its own module that comes after rawprepare, or maybe it could be added to rawprepare. Similarly to #7092 it would need to access the additional info from the exif which is not currently stored in dt_image_t or the image database. It could be awkward to add there because it can be quite large (20kb from the Pixel 4a) and there is not an explicit upper bound on the size defined by the file format. Maybe it could be managed similarly to dt_image_t.profile - it's not stored in the structure or the sql database, there is only a pointer to it. The colorin module loads it from the image file, allocates memory for it dynamically, and it's freed when it is removed from the image cache. The module implementing the gainmap could do something similar - if it hasn't been loaded into dt_image_t already, read it out of the exif, allocate memory, and store a pointer to it in dt_image_t.
non_defect
lens shading map dng gainmap support some smartphone raw dngs contain a lens shading map which is like a generalization of vignetting correction unlike the existing vignetting correction in lensfun it can be an arbitrary shape rather than radially symmetric and a separate map is applied to each raw color channel separately to correct spatially varying color casts also it s generated for each individual raw image rather than being constant for a specific camera lens module i m not sure how widely used this feature is across various phones and camera apps the raw dngs from the google camera app on my pixel have it and the vignetting is pretty noticeable if the correction is not done the has a detailed description of how the correction works the describes how the lens shading map is encoded as gainmap opcodes within the exif tag there is a it is enabled by the flat field module in the raw category when embedded in metadata is checked i have been trying to figure out how this could be implemented in darktable according to the dng spec as a stage opcode in the gain map should be applied to linear raw data after black level subtraction but before demosaicing for the typical case of a rggb bayer sensor where the gainmap for both of the g channels is identical like from pixel it would probably end up with the same result if the gainmaps were applied to the demosaiced rgb image prior to input color profile but it probably makes more sense to apply it to the raw data according to the spec one possibility is to do the correction within rawspeed it already has the ability to parse dng opcodes other than gainmap but currently this is only done for lossy dngs it could be modified to also apply black level subtraction and gainmaps to raw dngs that have the gain map and then it would just output for the black levels so that rawprepare does not perform any further black level subtraction i m not sure how the ui would work if there was a need to make this correction optional since it wouldn t be part of any module another possibility is to do the correction within the pipeline i think this would be more complicated to implement it could be its own module that comes after rawprepare or maybe it could be added to rawprepare similarly to it would need to access the additional info from the exif which is not currently stored in dt image t or the image database it could be awkward to add there because it can be quite large from the pixel and there is not an explicit upper bound on the size defined by the file format maybe it could be managed similarly to dt image t profile it s not stored in the structure or the sql database there is only a pointer to it the colorin module loads it from the image file allocates memory for it dynamically and it s freed when it is removed from the image cache the module implementing the gainmap could do something similar if it hasn t been loaded into dt image t already read it out of the exif allocate memory and store a pointer to it in dt image t
0
179,516
13,884,956,290
IssuesEvent
2020-10-18 18:07:25
daniel-norris/neu_ui
https://api.github.com/repos/daniel-norris/neu_ui
closed
Create a simple test for the CardHeader component
good first issue hacktoberfest help wanted tests
**Is your feature request related to a problem? Please describe.** Test coverage across the application is low. We need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future. **Describe the solution you'd like** We need to implement better test coverage across the library. Ideally each component should be accompanied by a test case written using Jest. At a minimum the test should check whether the component successfully shows any child `props`. We should have a test case covering the CardHeader component implemented using Jest. More info on Jest can be found [here](https://jestjs.io/). For examples of how this is done, take a look at existing test cases in the library. An example would be the typography or button components. This is part of epic #19.
1.0
Create a simple test for the CardHeader component - **Is your feature request related to a problem? Please describe.** Test coverage across the application is low. We need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future. **Describe the solution you'd like** We need to implement better test coverage across the library. Ideally each component should be accompanied by a test case written using Jest. At a minimum the test should check whether the component successfully shows any child `props`. We should have a test case covering the CardHeader component implemented using Jest. More info on Jest can be found [here](https://jestjs.io/). For examples of how this is done, take a look at existing test cases in the library. An example would be the typography or button components. This is part of epic #19.
non_defect
create a simple test for the cardheader component is your feature request related to a problem please describe test coverage across the application is low we need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future describe the solution you d like we need to implement better test coverage across the library ideally each component should be accompanied by a test case written using jest at a minimum the test should check whether the component successfully shows any child props we should have a test case covering the cardheader component implemented using jest more info on jest can be found for examples of how this is done take a look at existing test cases in the library an example would be the typography or button components this is part of epic
0
91,242
11,488,909,766
IssuesEvent
2020-02-11 14:41:20
ibm-openbmc/dev
https://api.github.com/repos/ibm-openbmc/dev
closed
GUI : Design : Creating symbols library for sketch
GUI UI Design
## SMEs **BMC**: Ryan Arnell Jandra ## Use Case Creating symbols library for sketch based on new community approved design. Once the symbols are created, the library will be very easy to use and the design will be consistent across the board ## InVision Prototype TBD ## Design Issue (phosphor-webui) TBD ## References/Resources - [eBMC Feature Item: <add item number>](https://ibm.box.com/s/j15ux3yfjycy4or4azbqyhqq11lbya0r) - feature discovery folder: - user research notes*: - user research synthesis: * This folder is restricted in accordance with GDPR guidelines.
1.0
GUI : Design : Creating symbols library for sketch - ## SMEs **BMC**: Ryan Arnell Jandra ## Use Case Creating symbols library for sketch based on new community approved design. Once the symbols are created, the library will be very easy to use and the design will be consistent across the board ## InVision Prototype TBD ## Design Issue (phosphor-webui) TBD ## References/Resources - [eBMC Feature Item: <add item number>](https://ibm.box.com/s/j15ux3yfjycy4or4azbqyhqq11lbya0r) - feature discovery folder: - user research notes*: - user research synthesis: * This folder is restricted in accordance with GDPR guidelines.
non_defect
gui design creating symbols library for sketch smes bmc ryan arnell jandra use case creating symbols library for sketch based on new community approved design once the symbols are created the library will be very easy to use and the design will be consistent across the board invision prototype tbd design issue phosphor webui tbd references resources feature discovery folder user research notes user research synthesis this folder is restricted in accordance with gdpr guidelines
0
27,454
5,025,293,032
IssuesEvent
2016-12-15 08:38:30
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
opened
e-Wallet_Mobile App (Loyalty-My Vouchers Used/Expired Page) 15122016
APAC Venture Defect - Medium (Sev-3) Dev Environment
Defect Description : Clicking on the used or expired voucher does nothing, unable to check the voucher details Device : Both android & IOS UserID : TNG001 ![wrong tagging for used vouchers](https://cloud.githubusercontent.com/assets/23113211/21217006/dba46f08-c2e4-11e6-9ca7-45b9eb9bfb51.jpg) Kindly refer attached POT for your perusal.
1.0
e-Wallet_Mobile App (Loyalty-My Vouchers Used/Expired Page) 15122016 - Defect Description : Clicking on the used or expired voucher does nothing, unable to check the voucher details Device : Both android & IOS UserID : TNG001 ![wrong tagging for used vouchers](https://cloud.githubusercontent.com/assets/23113211/21217006/dba46f08-c2e4-11e6-9ca7-45b9eb9bfb51.jpg) Kindly refer attached POT for your perusal.
defect
e wallet mobile app loyalty my vouchers used expired page defect description clicking on the used or expired voucher does nothing unable to check the voucher details device both android ios userid kindly refer attached pot for your perusal
1
52,660
13,224,886,538
IssuesEvent
2020-08-17 20:02:52
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
omkey inherits from i3Frame object (Trac #117)
IceTray Migrated from Trac defect
where did this come from? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/117">https://code.icecube.wisc.edu/projects/icecube/ticket/117</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "_ts": "1416713876900096", "description": "where did this come from?", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "time": "2008-08-30T16:06:14", "component": "IceTray", "summary": "omkey inherits from i3Frame object", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
omkey inherits from i3Frame object (Trac #117) - where did this come from? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/117">https://code.icecube.wisc.edu/projects/icecube/ticket/117</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "_ts": "1416713876900096", "description": "where did this come from?", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "time": "2008-08-30T16:06:14", "component": "IceTray", "summary": "omkey inherits from i3Frame object", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
omkey inherits from object trac where did this come from migrated from json status closed changetime ts description where did this come from reporter troy cc resolution wont or cant fix time component icetray summary omkey inherits from object priority major keywords milestone owner troy type defect
1
67,640
21,038,684,661
IssuesEvent
2022-03-31 10:13:08
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
MLocationBody 'Expand map' z-index stacks over message composer
T-Defect
### Steps to reproduce <img width="454" alt="Screenshot 2022-03-31 at 12 12 03" src="https://user-images.githubusercontent.com/3055605/161032140-0b855b7c-bd43-4a9d-9964-118295eb1c1a.png"> ### Outcome #### What did you expect? #### What happened instead? ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
MLocationBody 'Expand map' z-index stacks over message composer - ### Steps to reproduce <img width="454" alt="Screenshot 2022-03-31 at 12 12 03" src="https://user-images.githubusercontent.com/3055605/161032140-0b855b7c-bd43-4a9d-9964-118295eb1c1a.png"> ### Outcome #### What did you expect? #### What happened instead? ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
mlocationbody expand map z index stacks over message composer steps to reproduce img width alt screenshot at src outcome what did you expect what happened instead operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
1
75,718
26,012,107,869
IssuesEvent
2022-12-21 03:28:11
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
zfs destroy can't destroy zvol that are detected with permanent errors
Type: Defect Status: Stale
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | ProxmoxVE Distribution Version | 6.4-13 Kernel Version | 5.4.128-1-pve Architecture | amd64 OpenZFS Version | zfs-kmod-2.0.5-pve1~bpo10+1 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing I had some issues with bad cables that caused my zpool can't import some days ago. Then I changed these cables and used `zpool import -f -F -T <tgx> poolname` to revert it to a previous state (the steps are: https://github.com/openzfs/zfs/issues/6497#issuecomment-917718107). Now there are still some permanent errors with some zvols when I check with `zpool status -v`. I want to destroy these zvols to clean these errors, but I can't destroy them (except one of them), it shows `cannot open 'xxx': I/O error`. How to get rid of them? ``` root@pve52:~# zfs mount -a cannot iterate filesystems: I/O error root@pve52:~# zpool status -v pool: zfs52 state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A scan: scrub repaired 8K in 1 days 00:48:01 with 8 errors on Mon Sep 13 02:48:13 2021 config: NAME STATE READ WRITE CKSUM zfs52 DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 ata-WDC_WD120EMFZ-11A6JA0_QGG3AB2T DEGRADED 0 0 32 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGDS91T DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGE07HT DEGRADED 0 0 27 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGL1ZVT DEGRADED 0 0 28 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGH5V04T DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G502KL DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G6H9LL DEGRADED 0 0 16 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G9TYHL DEGRADED 0 0 19 too many errors ata-WDC_WD120EMFZ-11A6JA0_XHG0J1MD DEGRADED 0 0 28 too many errors errors: Permanent errors have been detected in the following files: zfs52/enc/dir:<0xf0583> zfs52/enc/vol/vm-115-disk-0:<0x0> zfs52/enc/vol/vm-112-disk-0:<0x0> zfs52/enc/vol/subvol-103-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-2:<0x0> zfs52/enc/vol/vm-113-disk-0:<0x0> zfs52/enc/vol/vm-116-disk-0:<0x0> root@pve52:~# zfs destroy zfs52/enc/vol/vm-115-disk-0 cannot open 'zfs52/enc/vol/vm-115-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-112-disk-0 cannot open 'zfs52/enc/vol/vm-112-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/subvol-103-disk-0 //<---- note: this is the only one that can be destroyed root@pve52:~# zfs destroy zfs52/enc/vol/vm-151-disk-0 cannot open 'zfs52/enc/vol/vm-151-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-151-disk-2 cannot open 'zfs52/enc/vol/vm-151-disk-2': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-113-disk-0 cannot open 'zfs52/enc/vol/vm-113-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-116-disk-0 cannot open 'zfs52/enc/vol/vm-116-disk-0': I/O error ``` This is what it looks like after destroyed `zfs52/enc/vol/subvol-103-disk-0`. All others are still there... ``` root@pve52:~# zpool status -v pool: zfs52 state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A scan: scrub repaired 8K in 1 days 00:48:01 with 8 errors on Mon Sep 13 02:48:13 2021 config: NAME STATE READ WRITE CKSUM zfs52 DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 ata-WDC_WD120EMFZ-11A6JA0_QGG3AB2T DEGRADED 0 0 36 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGDS91T DEGRADED 0 0 24 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGE07HT DEGRADED 0 0 32 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGL1ZVT DEGRADED 0 0 34 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGH5V04T DEGRADED 0 0 26 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G502KL DEGRADED 0 0 26 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G6H9LL DEGRADED 0 0 18 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G9TYHL DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_XHG0J1MD DEGRADED 0 0 32 too many errors errors: Permanent errors have been detected in the following files: zfs52/enc/dir:<0xf0583> zfs52/enc/vol/vm-115-disk-0:<0x0> zfs52/enc/vol/vm-112-disk-0:<0x0> <0x4d>:<0x0> zfs52/enc/vol/vm-151-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-2:<0x0> zfs52/enc/vol/vm-113-disk-0:<0x0> zfs52/enc/vol/vm-116-disk-0:<0x0> <0xffffffffffffffff>:<0x0> ``` ### Describe how to reproduce the problem Not sure how to reproduce because it's initially caused by bad cables. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> dmesg log appears to be normal. The only warning are these (unknown if it's related with zfs or not) ``` [ 31.518469] L1TF CPU bug present and SMT on, data leak possible. See CVE-2018-3646 and https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/l1tf.html for details. [ 35.987493] WARNING: can't open objset 2068, error 5 [ 36.002540] WARNING: can't open objset 1720, error 5 [ 36.003232] WARNING: can't open objset 269, error 5 [ 36.032697] WARNING: can't open objset 2949, error 5 [ 36.037329] WARNING: can't open objset 612, error 5 [ 36.046358] WARNING: can't open objset 626, error 5 [ 36.203316] WARNING: can't open objset for 269, error 5 [ 36.203424] WARNING: can't open objset for 626, error 5 [ 36.203732] WARNING: can't open objset for 2068, error 5 [ 36.203780] WARNING: can't open objset for 2949, error 5 [ 36.203905] WARNING: can't open objset for 612, error 5 [ 36.203943] WARNING: can't open objset for 1720, error 5 [ 44.920979] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 44.921010] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 74.809359] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 74.809395] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 105.028711] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 105.028789] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 105.028797] mei_me 0000:00:16.0: reset: reached maximal consecutive resets: disabling the device [ 3939.672762] zd0: p1 p2 [ 3940.133015] zd16: p1 p2 [ 3940.700128] zd32: p1 p2 p3 [ 3941.185524] zd48: p1 p2 p3 [ 3941.650778] zd64: p1 [ 3942.101795] zd80: p1 [14547.446785] device tap118i0 entered promiscuous mode [14547.476417] fwbr118i0: port 1(fwln118i0) entered blocking state [14547.476418] fwbr118i0: port 1(fwln118i0) entered disabled state ```
1.0
zfs destroy can't destroy zvol that are detected with permanent errors - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | ProxmoxVE Distribution Version | 6.4-13 Kernel Version | 5.4.128-1-pve Architecture | amd64 OpenZFS Version | zfs-kmod-2.0.5-pve1~bpo10+1 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing I had some issues with bad cables that caused my zpool can't import some days ago. Then I changed these cables and used `zpool import -f -F -T <tgx> poolname` to revert it to a previous state (the steps are: https://github.com/openzfs/zfs/issues/6497#issuecomment-917718107). Now there are still some permanent errors with some zvols when I check with `zpool status -v`. I want to destroy these zvols to clean these errors, but I can't destroy them (except one of them), it shows `cannot open 'xxx': I/O error`. How to get rid of them? ``` root@pve52:~# zfs mount -a cannot iterate filesystems: I/O error root@pve52:~# zpool status -v pool: zfs52 state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A scan: scrub repaired 8K in 1 days 00:48:01 with 8 errors on Mon Sep 13 02:48:13 2021 config: NAME STATE READ WRITE CKSUM zfs52 DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 ata-WDC_WD120EMFZ-11A6JA0_QGG3AB2T DEGRADED 0 0 32 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGDS91T DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGE07HT DEGRADED 0 0 27 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGL1ZVT DEGRADED 0 0 28 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGH5V04T DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G502KL DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G6H9LL DEGRADED 0 0 16 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G9TYHL DEGRADED 0 0 19 too many errors ata-WDC_WD120EMFZ-11A6JA0_XHG0J1MD DEGRADED 0 0 28 too many errors errors: Permanent errors have been detected in the following files: zfs52/enc/dir:<0xf0583> zfs52/enc/vol/vm-115-disk-0:<0x0> zfs52/enc/vol/vm-112-disk-0:<0x0> zfs52/enc/vol/subvol-103-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-2:<0x0> zfs52/enc/vol/vm-113-disk-0:<0x0> zfs52/enc/vol/vm-116-disk-0:<0x0> root@pve52:~# zfs destroy zfs52/enc/vol/vm-115-disk-0 cannot open 'zfs52/enc/vol/vm-115-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-112-disk-0 cannot open 'zfs52/enc/vol/vm-112-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/subvol-103-disk-0 //<---- note: this is the only one that can be destroyed root@pve52:~# zfs destroy zfs52/enc/vol/vm-151-disk-0 cannot open 'zfs52/enc/vol/vm-151-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-151-disk-2 cannot open 'zfs52/enc/vol/vm-151-disk-2': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-113-disk-0 cannot open 'zfs52/enc/vol/vm-113-disk-0': I/O error root@pve52:~# zfs destroy zfs52/enc/vol/vm-116-disk-0 cannot open 'zfs52/enc/vol/vm-116-disk-0': I/O error ``` This is what it looks like after destroyed `zfs52/enc/vol/subvol-103-disk-0`. All others are still there... ``` root@pve52:~# zpool status -v pool: zfs52 state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A scan: scrub repaired 8K in 1 days 00:48:01 with 8 errors on Mon Sep 13 02:48:13 2021 config: NAME STATE READ WRITE CKSUM zfs52 DEGRADED 0 0 0 raidz1-0 DEGRADED 0 0 0 ata-WDC_WD120EMFZ-11A6JA0_QGG3AB2T DEGRADED 0 0 36 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGDS91T DEGRADED 0 0 24 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGE07HT DEGRADED 0 0 32 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGGL1ZVT DEGRADED 0 0 34 too many errors ata-WDC_WD120EMFZ-11A6JA0_QGH5V04T DEGRADED 0 0 26 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G502KL DEGRADED 0 0 26 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G6H9LL DEGRADED 0 0 18 too many errors ata-WDC_WD120EMFZ-11A6JA0_X1G9TYHL DEGRADED 0 0 22 too many errors ata-WDC_WD120EMFZ-11A6JA0_XHG0J1MD DEGRADED 0 0 32 too many errors errors: Permanent errors have been detected in the following files: zfs52/enc/dir:<0xf0583> zfs52/enc/vol/vm-115-disk-0:<0x0> zfs52/enc/vol/vm-112-disk-0:<0x0> <0x4d>:<0x0> zfs52/enc/vol/vm-151-disk-0:<0x0> zfs52/enc/vol/vm-151-disk-2:<0x0> zfs52/enc/vol/vm-113-disk-0:<0x0> zfs52/enc/vol/vm-116-disk-0:<0x0> <0xffffffffffffffff>:<0x0> ``` ### Describe how to reproduce the problem Not sure how to reproduce because it's initially caused by bad cables. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> dmesg log appears to be normal. The only warning are these (unknown if it's related with zfs or not) ``` [ 31.518469] L1TF CPU bug present and SMT on, data leak possible. See CVE-2018-3646 and https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/l1tf.html for details. [ 35.987493] WARNING: can't open objset 2068, error 5 [ 36.002540] WARNING: can't open objset 1720, error 5 [ 36.003232] WARNING: can't open objset 269, error 5 [ 36.032697] WARNING: can't open objset 2949, error 5 [ 36.037329] WARNING: can't open objset 612, error 5 [ 36.046358] WARNING: can't open objset 626, error 5 [ 36.203316] WARNING: can't open objset for 269, error 5 [ 36.203424] WARNING: can't open objset for 626, error 5 [ 36.203732] WARNING: can't open objset for 2068, error 5 [ 36.203780] WARNING: can't open objset for 2949, error 5 [ 36.203905] WARNING: can't open objset for 612, error 5 [ 36.203943] WARNING: can't open objset for 1720, error 5 [ 44.920979] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 44.921010] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 74.809359] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 74.809395] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 105.028711] mei_me 0000:00:16.0: timer: init clients timeout hbm_state = 2. [ 105.028789] mei_me 0000:00:16.0: unexpected reset: dev_state = INIT_CLIENTS fw status = 001F0252 348A0E26 00000000 00084000 00000000 00000000 [ 105.028797] mei_me 0000:00:16.0: reset: reached maximal consecutive resets: disabling the device [ 3939.672762] zd0: p1 p2 [ 3940.133015] zd16: p1 p2 [ 3940.700128] zd32: p1 p2 p3 [ 3941.185524] zd48: p1 p2 p3 [ 3941.650778] zd64: p1 [ 3942.101795] zd80: p1 [14547.446785] device tap118i0 entered promiscuous mode [14547.476417] fwbr118i0: port 1(fwln118i0) entered blocking state [14547.476418] fwbr118i0: port 1(fwln118i0) entered disabled state ```
defect
zfs destroy can t destroy zvol that are detected with permanent errors thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name proxmoxve distribution version kernel version pve architecture openzfs version zfs kmod command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing i had some issues with bad cables that caused my zpool can t import some days ago then i changed these cables and used zpool import f f t poolname to revert it to a previous state the steps are now there are still some permanent errors with some zvols when i check with zpool status v i want to destroy these zvols to clean these errors but i can t destroy them except one of them it shows cannot open xxx i o error how to get rid of them root zfs mount a cannot iterate filesystems i o error root zpool status v pool state degraded status one or more devices has experienced an error resulting in data corruption applications may be affected action restore the file in question if possible otherwise restore the entire pool from backup see scan scrub repaired in days with errors on mon sep config name state read write cksum degraded degraded ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors errors permanent errors have been detected in the following files enc dir enc vol vm disk enc vol vm disk enc vol subvol disk enc vol vm disk enc vol vm disk enc vol vm disk enc vol vm disk root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error root zfs destroy enc vol subvol disk note this is the only one that can be destroyed root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error root zfs destroy enc vol vm disk cannot open enc vol vm disk i o error this is what it looks like after destroyed enc vol subvol disk all others are still there root zpool status v pool state degraded status one or more devices has experienced an error resulting in data corruption applications may be affected action restore the file in question if possible otherwise restore the entire pool from backup see scan scrub repaired in days with errors on mon sep config name state read write cksum degraded degraded ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors ata wdc degraded too many errors errors permanent errors have been detected in the following files enc dir enc vol vm disk enc vol vm disk enc vol vm disk enc vol vm disk enc vol vm disk enc vol vm disk describe how to reproduce the problem not sure how to reproduce because it s initially caused by bad cables include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with dmesg log appears to be normal the only warning are these unknown if it s related with zfs or not cpu bug present and smt on data leak possible see cve and for details warning can t open objset error warning can t open objset error warning can t open objset error warning can t open objset error warning can t open objset error warning can t open objset error warning can t open objset for error warning can t open objset for error warning can t open objset for error warning can t open objset for error warning can t open objset for error warning can t open objset for error mei me timer init clients timeout hbm state mei me unexpected reset dev state init clients fw status mei me timer init clients timeout hbm state mei me unexpected reset dev state init clients fw status mei me timer init clients timeout hbm state mei me unexpected reset dev state init clients fw status mei me reset reached maximal consecutive resets disabling the device device entered promiscuous mode port entered blocking state port entered disabled state
1
343,388
10,329,142,358
IssuesEvent
2019-09-02 11:23:54
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
timesofindia.indiatimes.com - site is not usable
browser-firefox-mobile engine-gecko priority-important
<!-- @browser: Firefox Mobile 69.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 --> <!-- @reported_with: web --> **URL**: https://timesofindia.indiatimes.com/city/chandigarh/fake-insurance-claim-scam-lawyer-among-3-booked-in-hisar/articleshow/70812138.cms?from=mdr **Browser / Version**: Firefox Mobile 69.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Page loads and the site is visible for a flash and then only white page **Steps to Reproduce**: Load the url <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
timesofindia.indiatimes.com - site is not usable - <!-- @browser: Firefox Mobile 69.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 --> <!-- @reported_with: web --> **URL**: https://timesofindia.indiatimes.com/city/chandigarh/fake-insurance-claim-scam-lawyer-among-3-booked-in-hisar/articleshow/70812138.cms?from=mdr **Browser / Version**: Firefox Mobile 69.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Page loads and the site is visible for a flash and then only white page **Steps to Reproduce**: Load the url <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
timesofindia indiatimes com site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description page loads and the site is visible for a flash and then only white page steps to reproduce load the url browser configuration none from with ❤️
0
79,186
7,698,027,752
IssuesEvent
2018-05-18 21:05:51
Princeton-CDH/ppa-django
https://api.github.com/repos/Princeton-CDH/ppa-django
closed
Histogram dates are unordered
awaiting testing
my local (correct): <img width="446" alt="screen shot 2018-05-18 at 11 35 05 am" src="https://user-images.githubusercontent.com/4924494/40243947-8e743a3a-5a8f-11e8-9432-14a5beff64e6.png"> qa current: <img width="693" alt="screen shot 2018-05-18 at 11 35 40 am" src="https://user-images.githubusercontent.com/4924494/40243966-9e3cab1e-5a8f-11e8-98f4-c7295abc0996.png"> the `<dt>` and `<dd>` elements are appearing out of order (not arranged chronologically) which leads me to believe that results in `facet_ranges.pub_date` in the view context are unordered in qa for some reason.
1.0
Histogram dates are unordered - my local (correct): <img width="446" alt="screen shot 2018-05-18 at 11 35 05 am" src="https://user-images.githubusercontent.com/4924494/40243947-8e743a3a-5a8f-11e8-9432-14a5beff64e6.png"> qa current: <img width="693" alt="screen shot 2018-05-18 at 11 35 40 am" src="https://user-images.githubusercontent.com/4924494/40243966-9e3cab1e-5a8f-11e8-98f4-c7295abc0996.png"> the `<dt>` and `<dd>` elements are appearing out of order (not arranged chronologically) which leads me to believe that results in `facet_ranges.pub_date` in the view context are unordered in qa for some reason.
non_defect
histogram dates are unordered my local correct img width alt screen shot at am src qa current img width alt screen shot at am src the and elements are appearing out of order not arranged chronologically which leads me to believe that results in facet ranges pub date in the view context are unordered in qa for some reason
0
33,551
7,160,969,659
IssuesEvent
2018-01-28 08:22:25
stevespringett/dependency-track
https://api.github.com/repos/stevespringett/dependency-track
closed
cve downloads only as far as 2011/dependency-check analysis issue
defect in progress
Hi, I'm using the master 3.0 dev branch & have uploaded a few test projects scanned by dependency check on Jenkins; I can publish the results to dependency-track but see some strange things in the logs that also exist in the UI. 1. the UI only has CVEs up to 2011, thus the projects are not fully scanned: 2017-10-18 07:00:32,575 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://static.nvd.nist.gov/feeds/json/cve/1.0/nvdcve-1.0-2010.json.gz 2017-10-18 07:00:34,075 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.0-2010.json.gz 2017-10-18 07:00:34,386 [] INFO [org.owasp.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.0-2010.json 2017-10-18 07:00:41,686 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:41,687 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://nvd.nist.gov/download/nvdcve-2011.xml.gz 2017-10-18 07:00:42,886 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-2011.xml.gz 2017-10-18 07:00:43,069 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:43,070 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://nvd.nist.gov/feeds/xml/cve/nvdcve-2.0-2011.xml.gz 2017-10-18 07:00:45,454 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-2.0-2011.xml.gz 2017-10-18 07:00:45,976 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:45,977 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://static.nvd.nist.gov/feeds/json/cve/1.0/nvdcve-1.0-2011.json.gz 2017-10-18 07:00:46,760 [] INFO [alpine.AlpineServlet] Starting Dependency-Track 2017-10-18 07:00:49,364 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.0-2011.json.gz 2017-10-18 07:00:50,258 [] INFO [org.owasp.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.0-2011.json 2. none of the CVEs found in dependency check are uploaded to the UI, looking at the logs: 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: data.directory='/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,929 [] INFO [org.owasp.dependencycheck.agent.DependencyCheckScanAgent] Populating settings 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: autoupdate='false' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.central.enabled='true' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.nexus.enabled='true' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.nexus.proxy='true' 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Archive Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer File Name Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Jar Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Hint Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer CPE Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer False Positive Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Cpe Suppression Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Dependency Bundling Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Dependency Merging Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer NVD CVE Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Vulnerability Suppression Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.CentralAnalyzer] Enabling the Central analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Central Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.NexusAnalyzer] Nexus analyzer disabled, using Central instead 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Nexus Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Nuspec Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Assembly Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer OpenSSL Source Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Node Security Platform Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Ruby Bundle Audit Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Version Filter Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Settings.getDataFile() - file: '/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Connection String: 'jdbc:h2:file:/root/.dependency-track/dependency-check/data/dc;MV_STORE=FALSE;AUTOCOMMIT=ON;LOCK_MODE=0;FILE_LOCK=NO' 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Settings.getDataFile() - file: '/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,954 [] DEBUG [org.owasp.dependencycheck.data.nvdcve.CveDB] Database product: h2 2017-10-18 06:59:17,956 [] ERROR [org.owasp.dependencycheck.Engine] No documents exist Unable to continue dependency-check analysis. 2017-10-18 06:59:17,956 [] DEBUG [org.owasp.dependencycheck.Engine] org.owasp.dependencycheck.exception.NoDataException: No documents exist at org.owasp.dependencycheck.Engine.ensureDataExists(Engine.java:897) at org.owasp.dependencycheck.Engine.analyzeDependencies(Engine.java:604) at org.owasp.dependencycheck.agent.DependencyCheckScanAgent.executeDependencyCheck(DependencyCheckScanAgent.java:887) at org.owasp.dependencycheck.agent.DependencyCheckScanAgent.execute(DependencyCheckScanAgent.java:972) at org.owasp.dependencytrack.tasks.DependencyCheckTask.analyze(DependencyCheckTask.java:149) at org.owasp.dependencytrack.tasks.DependencyCheckTask.performAnalysis(DependencyCheckTask.java:92) at org.owasp.dependencytrack.tasks.DependencyCheckTask.inform(DependencyCheckTask.java:64) at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:86) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) I have attached some screenshots of what I see in terminal & UI. Thanks, AD ![nistdir](https://user-images.githubusercontent.com/4255262/31716116-52464dec-b3fe-11e7-82d9-875ec563a918.PNG) ![vulnerabilitesui](https://user-images.githubusercontent.com/4255262/31716122-5a6520a2-b3fe-11e7-9fbc-a76391660269.PNG) ![projectsview](https://user-images.githubusercontent.com/4255262/31716127-5f41580c-b3fe-11e7-90d2-537702b9621e.PNG)
1.0
cve downloads only as far as 2011/dependency-check analysis issue - Hi, I'm using the master 3.0 dev branch & have uploaded a few test projects scanned by dependency check on Jenkins; I can publish the results to dependency-track but see some strange things in the logs that also exist in the UI. 1. the UI only has CVEs up to 2011, thus the projects are not fully scanned: 2017-10-18 07:00:32,575 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://static.nvd.nist.gov/feeds/json/cve/1.0/nvdcve-1.0-2010.json.gz 2017-10-18 07:00:34,075 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.0-2010.json.gz 2017-10-18 07:00:34,386 [] INFO [org.owasp.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.0-2010.json 2017-10-18 07:00:41,686 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:41,687 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://nvd.nist.gov/download/nvdcve-2011.xml.gz 2017-10-18 07:00:42,886 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-2011.xml.gz 2017-10-18 07:00:43,069 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:43,070 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://nvd.nist.gov/feeds/xml/cve/nvdcve-2.0-2011.xml.gz 2017-10-18 07:00:45,454 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-2.0-2011.xml.gz 2017-10-18 07:00:45,976 [] ERROR [org.owasp.dependencytrack.tasks.NistMirrorTask] Failed to determine content length 2017-10-18 07:00:45,977 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Downloading https://static.nvd.nist.gov/feeds/json/cve/1.0/nvdcve-1.0-2011.json.gz 2017-10-18 07:00:46,760 [] INFO [alpine.AlpineServlet] Starting Dependency-Track 2017-10-18 07:00:49,364 [] INFO [org.owasp.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.0-2011.json.gz 2017-10-18 07:00:50,258 [] INFO [org.owasp.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.0-2011.json 2. none of the CVEs found in dependency check are uploaded to the UI, looking at the logs: 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: data.directory='/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,929 [] INFO [org.owasp.dependencycheck.agent.DependencyCheckScanAgent] Populating settings 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: autoupdate='false' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.central.enabled='true' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.nexus.enabled='true' 2017-10-18 06:59:17,929 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Setting: analyzer.nexus.proxy='true' 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Archive Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer File Name Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Jar Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Hint Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer CPE Analyzer 2017-10-18 06:59:17,930 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer False Positive Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Cpe Suppression Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Dependency Bundling Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Dependency Merging Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer NVD CVE Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Vulnerability Suppression Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.CentralAnalyzer] Enabling the Central analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Central Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.NexusAnalyzer] Nexus analyzer disabled, using Central instead 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Nexus Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Nuspec Analyzer 2017-10-18 06:59:17,931 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Assembly Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer OpenSSL Source Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Node Security Platform Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Ruby Bundle Audit Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.analyzer.AnalyzerService] Loaded Analyzer Version Filter Analyzer 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Settings.getDataFile() - file: '/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Connection String: 'jdbc:h2:file:/root/.dependency-track/dependency-check/data/dc;MV_STORE=FALSE;AUTOCOMMIT=ON;LOCK_MODE=0;FILE_LOCK=NO' 2017-10-18 06:59:17,932 [] DEBUG [org.owasp.dependencycheck.utils.Settings] Settings.getDataFile() - file: '/root/.dependency-track/dependency-check/data' 2017-10-18 06:59:17,954 [] DEBUG [org.owasp.dependencycheck.data.nvdcve.CveDB] Database product: h2 2017-10-18 06:59:17,956 [] ERROR [org.owasp.dependencycheck.Engine] No documents exist Unable to continue dependency-check analysis. 2017-10-18 06:59:17,956 [] DEBUG [org.owasp.dependencycheck.Engine] org.owasp.dependencycheck.exception.NoDataException: No documents exist at org.owasp.dependencycheck.Engine.ensureDataExists(Engine.java:897) at org.owasp.dependencycheck.Engine.analyzeDependencies(Engine.java:604) at org.owasp.dependencycheck.agent.DependencyCheckScanAgent.executeDependencyCheck(DependencyCheckScanAgent.java:887) at org.owasp.dependencycheck.agent.DependencyCheckScanAgent.execute(DependencyCheckScanAgent.java:972) at org.owasp.dependencytrack.tasks.DependencyCheckTask.analyze(DependencyCheckTask.java:149) at org.owasp.dependencytrack.tasks.DependencyCheckTask.performAnalysis(DependencyCheckTask.java:92) at org.owasp.dependencytrack.tasks.DependencyCheckTask.inform(DependencyCheckTask.java:64) at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:86) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) I have attached some screenshots of what I see in terminal & UI. Thanks, AD ![nistdir](https://user-images.githubusercontent.com/4255262/31716116-52464dec-b3fe-11e7-82d9-875ec563a918.PNG) ![vulnerabilitesui](https://user-images.githubusercontent.com/4255262/31716122-5a6520a2-b3fe-11e7-9fbc-a76391660269.PNG) ![projectsview](https://user-images.githubusercontent.com/4255262/31716127-5f41580c-b3fe-11e7-90d2-537702b9621e.PNG)
defect
cve downloads only as far as dependency check analysis issue hi i m using the master dev branch have uploaded a few test projects scanned by dependency check on jenkins i can publish the results to dependency track but see some strange things in the logs that also exist in the ui the ui only has cves up to thus the projects are not fully scanned info downloading info uncompressing nvdcve json gz info parsing nvdcve json error failed to determine content length info downloading info uncompressing nvdcve xml gz error failed to determine content length info downloading info uncompressing nvdcve xml gz error failed to determine content length info downloading info starting dependency track info uncompressing nvdcve json gz info parsing nvdcve json none of the cves found in dependency check are uploaded to the ui looking at the logs debug setting data directory root dependency track dependency check data info populating settings debug setting autoupdate false debug setting analyzer central enabled true debug setting analyzer nexus enabled true debug setting analyzer nexus proxy true debug loaded analyzer archive analyzer debug loaded analyzer file name analyzer debug loaded analyzer jar analyzer debug loaded analyzer hint analyzer debug loaded analyzer cpe analyzer debug loaded analyzer false positive analyzer debug loaded analyzer cpe suppression analyzer debug loaded analyzer dependency bundling analyzer debug loaded analyzer dependency merging analyzer debug loaded analyzer nvd cve analyzer debug loaded analyzer vulnerability suppression analyzer debug enabling the central analyzer debug loaded analyzer central analyzer debug nexus analyzer disabled using central instead debug loaded analyzer nexus analyzer debug loaded analyzer nuspec analyzer debug loaded analyzer assembly analyzer debug loaded analyzer openssl source analyzer debug loaded analyzer node security platform analyzer debug loaded analyzer ruby bundle audit analyzer debug loaded analyzer version filter analyzer debug settings getdatafile file root dependency track dependency check data debug connection string jdbc file root dependency track dependency check data dc mv store false autocommit on lock mode file lock no debug settings getdatafile file root dependency track dependency check data debug database product error no documents exist unable to continue dependency check analysis debug org owasp dependencycheck exception nodataexception no documents exist at org owasp dependencycheck engine ensuredataexists engine java at org owasp dependencycheck engine analyzedependencies engine java at org owasp dependencycheck agent dependencycheckscanagent executedependencycheck dependencycheckscanagent java at org owasp dependencycheck agent dependencycheckscanagent execute dependencycheckscanagent java at org owasp dependencytrack tasks dependencychecktask analyze dependencychecktask java at org owasp dependencytrack tasks dependencychecktask performanalysis dependencychecktask java at org owasp dependencytrack tasks dependencychecktask inform dependencychecktask java at alpine event framework baseeventservice lambda publish baseeventservice java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java i have attached some screenshots of what i see in terminal ui thanks ad
1
63,300
17,572,960,844
IssuesEvent
2021-08-15 03:55:48
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Mac: Text replacement bubble gets 'stuck'
T-Defect
### Steps to reproduce 1. Open System Preferences/Keyboard and go to the Text tab. For this bug report I'm assuming that the default entry of "omw" mapped to "On my way!" is present. 2. Go to Element and type "omw" (without accepting the text replacement) and press Return to send the message. ### What happened? The message ("omw") gets sent as expected, but the "On my way!" suggestion bubble persists and needs to be manually closed. ### What did you expect? The suggestion bubble should disappear automatically once the message has been sent. ### Operating system macOS 11.4 ### Application version Element 1.7.34 ### How did you install the app? From https://element.io/get-started
1.0
Mac: Text replacement bubble gets 'stuck' - ### Steps to reproduce 1. Open System Preferences/Keyboard and go to the Text tab. For this bug report I'm assuming that the default entry of "omw" mapped to "On my way!" is present. 2. Go to Element and type "omw" (without accepting the text replacement) and press Return to send the message. ### What happened? The message ("omw") gets sent as expected, but the "On my way!" suggestion bubble persists and needs to be manually closed. ### What did you expect? The suggestion bubble should disappear automatically once the message has been sent. ### Operating system macOS 11.4 ### Application version Element 1.7.34 ### How did you install the app? From https://element.io/get-started
defect
mac text replacement bubble gets stuck steps to reproduce open system preferences keyboard and go to the text tab for this bug report i m assuming that the default entry of omw mapped to on my way is present go to element and type omw without accepting the text replacement and press return to send the message what happened the message omw gets sent as expected but the on my way suggestion bubble persists and needs to be manually closed what did you expect the suggestion bubble should disappear automatically once the message has been sent operating system macos application version element how did you install the app from
1
52,861
13,225,175,419
IssuesEvent
2020-08-17 20:38:28
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
Photospline error with new OSX tools (Trac #484)
Migrated from Trac combo reconstruction defect
I just updated all my laptop software to celebrate the collaboration meeting being over. This included: IceRec V04-06-00 Apple Xcode 5.1 Apple Command Line Developer Tools for mavericks 5.1.0.0 clang Apple LLVM version 5.1 (clang-503.0.38) (based on LLVM 3.4svn) When I tried to compile, photospline gave the error below. Claudio tells me that the "unrecognized argument" has always been unrecognized, but it used to be just a warning, and now it's a full-blown error. Error message below. thanks, Laura [ 15%] Building C object photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o clang: error: unknown argument: '-fpeel-loops' [-Wunused-command-line-argument-hard-error-in-future] clang: note: this will be a hard error (cannot be downgraded to a warning) in the future make[2]: *** [photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o] Error 1 make[1]: *** [photospline/CMakeFiles/photospline.dir/all] Error 2 make: *** [all] Error 2 <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/484">https://code.icecube.wisc.edu/projects/icecube/ticket/484</a>, reported by gladstoneand owned by meike.dewith</em></summary> <p> ```json { "status": "closed", "changetime": "2014-04-08T10:13:02", "_ts": "1396951982000000", "description": "I just updated all my laptop software to celebrate the collaboration meeting being over. This included: \n IceRec V04-06-00[[BR]]\n Apple Xcode 5.1[[BR]]\n Apple Command Line Developer Tools for mavericks 5.1.0.0[[BR]]\n clang Apple LLVM version 5.1 (clang-503.0.38) (based on LLVM 3.4svn)\n\nWhen I tried to compile, photospline gave the error below. Claudio tells me that the \"unrecognized argument\" has always been unrecognized, but it used to be just a warning, and now it's a full-blown error. \n\nError message below. [[BR]]\nthanks, [[BR]]\nLaura\n\n[ 15%] Building C object photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o[[BR]]\nclang: error: unknown argument: '-fpeel-loops' [-Wunused-command-line-argument-hard-error-in-future][[BR]]\nclang: note: this will be a hard error (cannot be downgraded to a warning) in the future[[BR]]\nmake[2]: *** [photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o] Error 1[[BR]]\nmake[1]: *** [photospline/CMakeFiles/photospline.dir/all] Error 2[[BR]]\nmake: *** [all] Error 2", "reporter": "gladstone", "cc": "", "resolution": "fixed", "time": "2014-03-11T23:34:28", "component": "combo reconstruction", "summary": "Photospline error with new OSX tools", "priority": "normal", "keywords": "", "milestone": "", "owner": "meike.dewith", "type": "defect" } ``` </p> </details>
1.0
Photospline error with new OSX tools (Trac #484) - I just updated all my laptop software to celebrate the collaboration meeting being over. This included: IceRec V04-06-00 Apple Xcode 5.1 Apple Command Line Developer Tools for mavericks 5.1.0.0 clang Apple LLVM version 5.1 (clang-503.0.38) (based on LLVM 3.4svn) When I tried to compile, photospline gave the error below. Claudio tells me that the "unrecognized argument" has always been unrecognized, but it used to be just a warning, and now it's a full-blown error. Error message below. thanks, Laura [ 15%] Building C object photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o clang: error: unknown argument: '-fpeel-loops' [-Wunused-command-line-argument-hard-error-in-future] clang: note: this will be a hard error (cannot be downgraded to a warning) in the future make[2]: *** [photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o] Error 1 make[1]: *** [photospline/CMakeFiles/photospline.dir/all] Error 2 make: *** [all] Error 2 <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/484">https://code.icecube.wisc.edu/projects/icecube/ticket/484</a>, reported by gladstoneand owned by meike.dewith</em></summary> <p> ```json { "status": "closed", "changetime": "2014-04-08T10:13:02", "_ts": "1396951982000000", "description": "I just updated all my laptop software to celebrate the collaboration meeting being over. This included: \n IceRec V04-06-00[[BR]]\n Apple Xcode 5.1[[BR]]\n Apple Command Line Developer Tools for mavericks 5.1.0.0[[BR]]\n clang Apple LLVM version 5.1 (clang-503.0.38) (based on LLVM 3.4svn)\n\nWhen I tried to compile, photospline gave the error below. Claudio tells me that the \"unrecognized argument\" has always been unrecognized, but it used to be just a warning, and now it's a full-blown error. \n\nError message below. [[BR]]\nthanks, [[BR]]\nLaura\n\n[ 15%] Building C object photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o[[BR]]\nclang: error: unknown argument: '-fpeel-loops' [-Wunused-command-line-argument-hard-error-in-future][[BR]]\nclang: note: this will be a hard error (cannot be downgraded to a warning) in the future[[BR]]\nmake[2]: *** [photospline/CMakeFiles/photospline.dir/private/lib/bspline.c.o] Error 1[[BR]]\nmake[1]: *** [photospline/CMakeFiles/photospline.dir/all] Error 2[[BR]]\nmake: *** [all] Error 2", "reporter": "gladstone", "cc": "", "resolution": "fixed", "time": "2014-03-11T23:34:28", "component": "combo reconstruction", "summary": "Photospline error with new OSX tools", "priority": "normal", "keywords": "", "milestone": "", "owner": "meike.dewith", "type": "defect" } ``` </p> </details>
defect
photospline error with new osx tools trac i just updated all my laptop software to celebrate the collaboration meeting being over this included icerec apple xcode apple command line developer tools for mavericks clang apple llvm version clang based on llvm when i tried to compile photospline gave the error below claudio tells me that the unrecognized argument has always been unrecognized but it used to be just a warning and now it s a full blown error error message below thanks laura building c object photospline cmakefiles photospline dir private lib bspline c o clang error unknown argument fpeel loops clang note this will be a hard error cannot be downgraded to a warning in the future make error make error make error migrated from json status closed changetime ts description i just updated all my laptop software to celebrate the collaboration meeting being over this included n icerec n apple xcode n apple command line developer tools for mavericks n clang apple llvm version clang based on llvm n nwhen i tried to compile photospline gave the error below claudio tells me that the unrecognized argument has always been unrecognized but it used to be just a warning and now it s a full blown error n nerror message below nthanks nlaura n n building c object photospline cmakefiles photospline dir private lib bspline c o nclang error unknown argument fpeel loops nclang note this will be a hard error cannot be downgraded to a warning in the future nmake error nmake error nmake error reporter gladstone cc resolution fixed time component combo reconstruction summary photospline error with new osx tools priority normal keywords milestone owner meike dewith type defect
1
9,273
2,615,142,526
IssuesEvent
2015-03-01 06:17:34
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
closed
WARNING: Receive timeout occurred
auto-migrated Priority-Triage Type-Defect
``` Backtrack 5R1 and a AWUS036 (Realtek RTL8187L), using r82, AP is a D-Link DSL 2680 and power is between -20 and -45 so very good (I'm right next to it basically). It won't associate so I'm using aireplay-ng mon0 --fakeauth 600 -e TALKTALK-4FFXXX -a XX:XX:XX:XX:XX:XX -h YY:YY:YY:YY:YY:YY, then reaver -i mon0 -b XX:XX:XX:XX:XX:XX -vv -A. Reaver Output: [+] Associated with XX:XX:XX:XX:XX:XX (ESSID: TALKTALK-4FFXXX) [+] Trying pin 12345670 [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request ...which occurs over and over. Capture attached. ``` Original issue reported on code.google.com by `bdee...@gmail.com` on 9 Jan 2012 at 10:37
1.0
WARNING: Receive timeout occurred - ``` Backtrack 5R1 and a AWUS036 (Realtek RTL8187L), using r82, AP is a D-Link DSL 2680 and power is between -20 and -45 so very good (I'm right next to it basically). It won't associate so I'm using aireplay-ng mon0 --fakeauth 600 -e TALKTALK-4FFXXX -a XX:XX:XX:XX:XX:XX -h YY:YY:YY:YY:YY:YY, then reaver -i mon0 -b XX:XX:XX:XX:XX:XX -vv -A. Reaver Output: [+] Associated with XX:XX:XX:XX:XX:XX (ESSID: TALKTALK-4FFXXX) [+] Trying pin 12345670 [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request ...which occurs over and over. Capture attached. ``` Original issue reported on code.google.com by `bdee...@gmail.com` on 9 Jan 2012 at 10:37
defect
warning receive timeout occurred backtrack and a realtek using ap is a d link dsl and power is between and so very good i m right next to it basically it won t associate so i m using aireplay ng fakeauth e talktalk a xx xx xx xx xx xx h yy yy yy yy yy yy then reaver i b xx xx xx xx xx xx vv a reaver output associated with xx xx xx xx xx xx essid talktalk trying pin sending eapol start request warning receive timeout occurred sending eapol start request warning receive timeout occurred sending eapol start request which occurs over and over capture attached original issue reported on code google com by bdee gmail com on jan at
1
77,716
14,910,639,871
IssuesEvent
2021-01-22 09:52:48
firecracker-microvm/firecracker
https://api.github.com/repos/firecracker-microvm/firecracker
closed
[Code improvement] deduplicate literal HTTP responses in tests
Codebase: Refactoring Contribute: Good First Issue Contribute: Help Wanted
There are many tests with literal hardcoded HTTP responses that bloat the code. Some of them even have data embedded in them, making those tests hard to maintain. Example possible deduplication: in https://github.com/firecracker-microvm/firecracker/blob/e8200f3c3eaba014220e447e8d426c8cf8607eec/src/api_server/src/parsed_request.rs#L511 : ```diff - let expected_response = format!( - "HTTP/1.1 200 \r\n\ - Server: Firecracker API\r\n\ - Connection: keep-alive\r\n\ - Content-Type: application/json\r\n\ - Content-Length: 122\r\n\r\n{}", - VmConfig::default().to_string() - ); + let body = VmConfig::default().to_string(); + let expected_response = http_response_ok(&body); ); ``` where `http_response_ok()` could be reused in all tests. Example `http_response_ok()` definition: ```rust fn http_response_ok(body: &str) { format!( "HTTP/1.1 200 \r\n\ Server: Firecracker API\r\n\ Connection: keep-alive\r\n\ Content-Type: application/json\r\n\ Content-Length: {}\r\n\r\n{}", status_code, body.len(), body, ) } ```
1.0
[Code improvement] deduplicate literal HTTP responses in tests - There are many tests with literal hardcoded HTTP responses that bloat the code. Some of them even have data embedded in them, making those tests hard to maintain. Example possible deduplication: in https://github.com/firecracker-microvm/firecracker/blob/e8200f3c3eaba014220e447e8d426c8cf8607eec/src/api_server/src/parsed_request.rs#L511 : ```diff - let expected_response = format!( - "HTTP/1.1 200 \r\n\ - Server: Firecracker API\r\n\ - Connection: keep-alive\r\n\ - Content-Type: application/json\r\n\ - Content-Length: 122\r\n\r\n{}", - VmConfig::default().to_string() - ); + let body = VmConfig::default().to_string(); + let expected_response = http_response_ok(&body); ); ``` where `http_response_ok()` could be reused in all tests. Example `http_response_ok()` definition: ```rust fn http_response_ok(body: &str) { format!( "HTTP/1.1 200 \r\n\ Server: Firecracker API\r\n\ Connection: keep-alive\r\n\ Content-Type: application/json\r\n\ Content-Length: {}\r\n\r\n{}", status_code, body.len(), body, ) } ```
non_defect
deduplicate literal http responses in tests there are many tests with literal hardcoded http responses that bloat the code some of them even have data embedded in them making those tests hard to maintain example possible deduplication in diff let expected response format http r n server firecracker api r n connection keep alive r n content type application json r n content length r n r n vmconfig default to string let body vmconfig default to string let expected response http response ok body where http response ok could be reused in all tests example http response ok definition rust fn http response ok body str format http r n server firecracker api r n connection keep alive r n content type application json r n content length r n r n status code body len body
0
49,381
13,186,650,015
IssuesEvent
2020-08-13 00:52:16
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[tableio] test_subevents is broken (Trac #1256)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1256">https://code.icecube.wisc.edu/ticket/1256</a>, reported by kkrings and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:13:59", "description": "`test_subvents.py` is not listed under `i3_test_scripts` but it is definitely broken. I think it should be removed or fixed.", "reporter": "kkrings", "cc": "", "resolution": "fixed", "_ts": "1458335639558230", "component": "combo core", "summary": "[tableio] test_subevents is broken", "priority": "blocker", "keywords": "", "time": "2015-08-20T19:48:05", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[tableio] test_subevents is broken (Trac #1256) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1256">https://code.icecube.wisc.edu/ticket/1256</a>, reported by kkrings and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:13:59", "description": "`test_subvents.py` is not listed under `i3_test_scripts` but it is definitely broken. I think it should be removed or fixed.", "reporter": "kkrings", "cc": "", "resolution": "fixed", "_ts": "1458335639558230", "component": "combo core", "summary": "[tableio] test_subevents is broken", "priority": "blocker", "keywords": "", "time": "2015-08-20T19:48:05", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
test subevents is broken trac migrated from json status closed changetime description test subvents py is not listed under test scripts but it is definitely broken i think it should be removed or fixed reporter kkrings cc resolution fixed ts component combo core summary test subevents is broken priority blocker keywords time milestone owner jvansanten type defect
1
740,471
25,753,085,887
IssuesEvent
2022-12-08 14:34:59
threefoldtech/freeflow_twin_main
https://api.github.com/repos/threefoldtech/freeflow_twin_main
closed
[DASHBOARD] Like icon moves around if you hover comment
priority_minor layout
Hovering over a comment moves the like icon around. It also shows the icon to be clickable but it isn't. ![image](https://user-images.githubusercontent.com/36887810/195327595-4334b999-228a-4c46-af78-38e1251d3a35.png)
1.0
[DASHBOARD] Like icon moves around if you hover comment - Hovering over a comment moves the like icon around. It also shows the icon to be clickable but it isn't. ![image](https://user-images.githubusercontent.com/36887810/195327595-4334b999-228a-4c46-af78-38e1251d3a35.png)
non_defect
like icon moves around if you hover comment hovering over a comment moves the like icon around it also shows the icon to be clickable but it isn t
0
20,614
10,542,555,665
IssuesEvent
2019-10-02 13:25:02
TIBCOSoftware/js-workshops
https://api.github.com/repos/TIBCOSoftware/js-workshops
opened
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/chai-spies/test/browser/index.html</p> <p>Path to vulnerable library: /js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/chai-spies/test/browser/index.html,/js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/js-workshops/commit/25c0018beaa00ae9261f3d3c00f032aa925f92b1">25c0018beaa00ae9261f3d3c00f032aa925f92b1</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p>
True
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/chai-spies/test/browser/index.html</p> <p>Path to vulnerable library: /js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/chai-spies/test/browser/index.html,/js-workshops/UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/js-workshops/commit/25c0018beaa00ae9261f3d3c00f032aa925f92b1">25c0018beaa00ae9261f3d3c00f032aa925f92b1</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p>
non_defect
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm js workshops urbanroast tutorial resources funnel master node modules chai spies test browser index html path to vulnerable library js workshops urbanroast tutorial resources funnel master node modules chai spies test browser index html js workshops urbanroast tutorial resources funnel master node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
76,578
26,493,994,557
IssuesEvent
2023-01-18 02:43:43
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG: CDF and PMF of binomial function not same with extreme values
defect
### Describe your issue. I wanted to calculate the chance that I inhale at least one molecule of Ceasars words (see [here](https://www.theguardian.com/books/2017/jul/16/caesars-last-breath-sam-kean-review-decoding-the-secrets-of-the-air-around-us)). I thought to calculate the chance of inhaling zero molecules. ### Reproducing Code Example ```python from scipy.stats import binom def calculate(n, p, r): print (f"{n=} {p=} {r=}") print (f"PMF The chance that you inhale {r} molecules {binom.pmf(r, n, p)}") print (f"CDF The chance that you inhale {r} molecules {binom.cdf(r, n, p)}") n = 25.0*10**21 p = 1.0*10**-21 r = 0 calculate(n, p, r) # My output is # PMF The chance that you inhale 0 molecules 1.0 # CDF The chance that you inhale 0 molecules 1.388794386496407e-11 # In R both results are 1.388794386496407e-11 ``` ### Error message ```shell - ``` ### SciPy/NumPy/Python version information 1.8.0 1.22.3 sys.version_info(major=3, minor=10, micro=3, releaselevel='final', serial=0)
1.0
BUG: CDF and PMF of binomial function not same with extreme values - ### Describe your issue. I wanted to calculate the chance that I inhale at least one molecule of Ceasars words (see [here](https://www.theguardian.com/books/2017/jul/16/caesars-last-breath-sam-kean-review-decoding-the-secrets-of-the-air-around-us)). I thought to calculate the chance of inhaling zero molecules. ### Reproducing Code Example ```python from scipy.stats import binom def calculate(n, p, r): print (f"{n=} {p=} {r=}") print (f"PMF The chance that you inhale {r} molecules {binom.pmf(r, n, p)}") print (f"CDF The chance that you inhale {r} molecules {binom.cdf(r, n, p)}") n = 25.0*10**21 p = 1.0*10**-21 r = 0 calculate(n, p, r) # My output is # PMF The chance that you inhale 0 molecules 1.0 # CDF The chance that you inhale 0 molecules 1.388794386496407e-11 # In R both results are 1.388794386496407e-11 ``` ### Error message ```shell - ``` ### SciPy/NumPy/Python version information 1.8.0 1.22.3 sys.version_info(major=3, minor=10, micro=3, releaselevel='final', serial=0)
defect
bug cdf and pmf of binomial function not same with extreme values describe your issue i wanted to calculate the chance that i inhale at least one molecule of ceasars words see i thought to calculate the chance of inhaling zero molecules reproducing code example python from scipy stats import binom def calculate n p r print f n p r print f pmf the chance that you inhale r molecules binom pmf r n p print f cdf the chance that you inhale r molecules binom cdf r n p n p r calculate n p r my output is pmf the chance that you inhale molecules cdf the chance that you inhale molecules in r both results are error message shell scipy numpy python version information sys version info major minor micro releaselevel final serial
1
16,598
2,920,381,489
IssuesEvent
2015-06-24 18:39:25
kronometrix/recording
https://api.github.com/repos/kronometrix/recording
closed
hdwrec on CloudLinux errors
defect-high
hdwrec reports errors regarding release operating system: [root@cl564 log]# cat /etc/redhat-release CloudLinux Server release 5.11 (Vladislav Volkov) [root@cl564 log]# cat hdwrec.log Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. relos is not found. This seems to be a problem with core package: Linux::Distribution which does not know about CloudLinux.
1.0
hdwrec on CloudLinux errors - hdwrec reports errors regarding release operating system: [root@cl564 log]# cat /etc/redhat-release CloudLinux Server release 5.11 (Vladislav Volkov) [root@cl564 log]# cat hdwrec.log Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. Use of uninitialized value $relos in printf at /opt/kronometrix/bin/hdwrec line 147. relos is not found. This seems to be a problem with core package: Linux::Distribution which does not know about CloudLinux.
defect
hdwrec on cloudlinux errors hdwrec reports errors regarding release operating system cat etc redhat release cloudlinux server release vladislav volkov cat hdwrec log use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line use of uninitialized value relos in printf at opt kronometrix bin hdwrec line relos is not found this seems to be a problem with core package linux distribution which does not know about cloudlinux
1