Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
108,420
4,344,652,587
IssuesEvent
2016-07-29 09:16:16
zeit/hyperterm
https://api.github.com/repos/zeit/hyperterm
opened
Schema for plugins and extensions
Priority: High Status: Community feedback wanted Type: Enhancement
Since most (if not all) of the projects within the node ecosystem are using this techniques, I think we should adapt one of them in favour of making the experience consistent for all users. ## Technique 1 - **Plugins:** `hyperterm-plugin-<name>` (example: "hyperterm-plugin-power") - **Themes:** `hyperterm-theme-<name>` (example: "hyperterm-theme-yellow") ## Technique 2 - **Plugins:** `hyperterm-<plugin-name>` (example: "hyperterm-power") - **Themes:** `hyperterm-<theme-name>` (example: "hyperterm-yellow") - I'm for technique 1 because it's much more explicit and forces developers to use (existing) plugins for extending their theme with complex functionality. What I'd like to do is create a "theme" property as well as keep the existing "plugins" array within the config file. This way, the user only needs to enter the name of the theme/plugin and the string will automatically get prefixed behind the curtains and downloaded from npm: ```js plugins: [ 'links', 'blink', 'cwd' ] ``` Results in these packages getting installed: - hyperterm-plugin-links - hyperterm-plugin-blink - hyperterm-plugin-cwd AND ```js theme: 'yellow' ``` Results in the "hyperterm-theme-yellow" theme getting installed.
1.0
Schema for plugins and extensions - Since most (if not all) of the projects within the node ecosystem are using this techniques, I think we should adapt one of them in favour of making the experience consistent for all users. ## Technique 1 - **Plugins:** `hyperterm-plugin-<name>` (example: "hyperterm-plugin-power") - **Themes:** `hyperterm-theme-<name>` (example: "hyperterm-theme-yellow") ## Technique 2 - **Plugins:** `hyperterm-<plugin-name>` (example: "hyperterm-power") - **Themes:** `hyperterm-<theme-name>` (example: "hyperterm-yellow") - I'm for technique 1 because it's much more explicit and forces developers to use (existing) plugins for extending their theme with complex functionality. What I'd like to do is create a "theme" property as well as keep the existing "plugins" array within the config file. This way, the user only needs to enter the name of the theme/plugin and the string will automatically get prefixed behind the curtains and downloaded from npm: ```js plugins: [ 'links', 'blink', 'cwd' ] ``` Results in these packages getting installed: - hyperterm-plugin-links - hyperterm-plugin-blink - hyperterm-plugin-cwd AND ```js theme: 'yellow' ``` Results in the "hyperterm-theme-yellow" theme getting installed.
non_defect
schema for plugins and extensions since most if not all of the projects within the node ecosystem are using this techniques i think we should adapt one of them in favour of making the experience consistent for all users technique plugins hyperterm plugin example hyperterm plugin power themes hyperterm theme example hyperterm theme yellow technique plugins hyperterm example hyperterm power themes hyperterm example hyperterm yellow i m for technique because it s much more explicit and forces developers to use existing plugins for extending their theme with complex functionality what i d like to do is create a theme property as well as keep the existing plugins array within the config file this way the user only needs to enter the name of the theme plugin and the string will automatically get prefixed behind the curtains and downloaded from npm js plugins links blink cwd results in these packages getting installed hyperterm plugin links hyperterm plugin blink hyperterm plugin cwd and js theme yellow results in the hyperterm theme yellow theme getting installed
0
384,747
11,402,661,334
IssuesEvent
2020-01-31 04:12:04
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
ImportError: cannot import name 'proto_pb2' from 'grpc' in python
disposition/stale kind/question lang/Python priority/P2
Hi everyone : D **I create a proto.proto file and then create proto_pb2.py and proto_pb2_grpc,py with protoc :** `>> python3 -m grpc_tools.protoc -I. grpc/proto.proto --python_out . --grpc_python_out . ` but when run the server.py it's **bug** is shown : ``` Traceback (most recent call last): File "server.py", line 3, in <module> import proto_pb2_grpc as proto_pb2_grpc File "/home/mehrdad/Downloads/profile-dgraph/grpc/proto_pb2_grpc.py", line 4, in <module> from grpc import proto_pb2 as grpc_dot_proto__pb2 ImportError: cannot import name 'proto_pb2' from 'grpc' (/home/mehrdad/anaconda3/lib/python3.7/site-packages/grpc/__init__.py) ``` **the server.py include :** ``` import grpc import proto_pb2 import proto_pb2_grpc and ... ``` Can you help me ?!
1.0
ImportError: cannot import name 'proto_pb2' from 'grpc' in python - Hi everyone : D **I create a proto.proto file and then create proto_pb2.py and proto_pb2_grpc,py with protoc :** `>> python3 -m grpc_tools.protoc -I. grpc/proto.proto --python_out . --grpc_python_out . ` but when run the server.py it's **bug** is shown : ``` Traceback (most recent call last): File "server.py", line 3, in <module> import proto_pb2_grpc as proto_pb2_grpc File "/home/mehrdad/Downloads/profile-dgraph/grpc/proto_pb2_grpc.py", line 4, in <module> from grpc import proto_pb2 as grpc_dot_proto__pb2 ImportError: cannot import name 'proto_pb2' from 'grpc' (/home/mehrdad/anaconda3/lib/python3.7/site-packages/grpc/__init__.py) ``` **the server.py include :** ``` import grpc import proto_pb2 import proto_pb2_grpc and ... ``` Can you help me ?!
non_defect
importerror cannot import name proto from grpc in python hi everyone d i create a proto proto file and then create proto py and proto grpc py with protoc m grpc tools protoc i grpc proto proto python out grpc python out but when run the server py it s bug is shown traceback most recent call last file server py line in import proto grpc as proto grpc file home mehrdad downloads profile dgraph grpc proto grpc py line in from grpc import proto as grpc dot proto importerror cannot import name proto from grpc home mehrdad lib site packages grpc init py the server py include import grpc import proto import proto grpc and can you help me
0
190,635
6,820,848,265
IssuesEvent
2017-11-07 15:09:27
DecipherNow/gm-fabric-dashboard
https://api.github.com/repos/DecipherNow/gm-fabric-dashboard
closed
Refactor GMServiceView into a scene under `src/Main/scenes`
in progress priority-3 refactor
Currently the service dashboard is structured as a subcomponent of the Fabric View. In reality, this should be a distinct scene that's a sibling of Fabric or Instance.
1.0
Refactor GMServiceView into a scene under `src/Main/scenes` - Currently the service dashboard is structured as a subcomponent of the Fabric View. In reality, this should be a distinct scene that's a sibling of Fabric or Instance.
non_defect
refactor gmserviceview into a scene under src main scenes currently the service dashboard is structured as a subcomponent of the fabric view in reality this should be a distinct scene that s a sibling of fabric or instance
0
308,413
9,438,950,783
IssuesEvent
2019-04-14 05:45:15
CS2113-AY1819S2-M11-2/main
https://api.github.com/repos/CS2113-AY1819S2-M11-2/main
closed
As a user, I must be able to delete notes that I no longer want
priority.High type.Story v1.4
So that I can view relevant tasks easily
1.0
As a user, I must be able to delete notes that I no longer want - So that I can view relevant tasks easily
non_defect
as a user i must be able to delete notes that i no longer want so that i can view relevant tasks easily
0
203,677
15,886,929,665
IssuesEvent
2021-04-10 00:06:16
GigiaJ/JondTextTranslator
https://api.github.com/repos/GigiaJ/JondTextTranslator
closed
Finish adding comments to methods in need
documentation enhancement
Most things are commented fairly well, but some things could still use comments to explain what they do better.
1.0
Finish adding comments to methods in need - Most things are commented fairly well, but some things could still use comments to explain what they do better.
non_defect
finish adding comments to methods in need most things are commented fairly well but some things could still use comments to explain what they do better
0
341,440
10,297,558,571
IssuesEvent
2019-08-28 08:51:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.instagram.com - design is broken
browser-fenix engine-gecko priority-critical
<!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Instagram stories doesn't appear well **Steps to Reproduce**: When you touch some story it shows in full screen horizontal and not in vertical [![Screenshot Description](https://webcompat.com/uploads/2019/8/10b03147-c755-4409-a670-1312c32d3f5e-thumb.jpeg)](https://webcompat.com/uploads/2019/8/10b03147-c755-4409-a670-1312c32d3f5e.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.instagram.com - design is broken - <!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Instagram stories doesn't appear well **Steps to Reproduce**: When you touch some story it shows in full screen horizontal and not in vertical [![Screenshot Description](https://webcompat.com/uploads/2019/8/10b03147-c755-4409-a670-1312c32d3f5e-thumb.jpeg)](https://webcompat.com/uploads/2019/8/10b03147-c755-4409-a670-1312c32d3f5e.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
design is broken url browser version firefox mobile operating system android tested another browser yes problem type design is broken description instagram stories doesn t appear well steps to reproduce when you touch some story it shows in full screen horizontal and not in vertical browser configuration none from with ❤️
0
56,091
14,926,943,799
IssuesEvent
2021-01-24 13:37:32
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
DataTable: csv export does not work with field
defect
@melloware @Rapster The datatable exporter seems not work with the new column field attribute notation if i have not missed any additionally required adaptations. Just add the following column to /showcase/ui/data/dataexporter/basic.xhtml to verify: ``` <p:column field="code" headerText="CodeNew"/> ``` The exporter delivers empty fields for such a column.
1.0
DataTable: csv export does not work with field - @melloware @Rapster The datatable exporter seems not work with the new column field attribute notation if i have not missed any additionally required adaptations. Just add the following column to /showcase/ui/data/dataexporter/basic.xhtml to verify: ``` <p:column field="code" headerText="CodeNew"/> ``` The exporter delivers empty fields for such a column.
defect
datatable csv export does not work with field melloware rapster the datatable exporter seems not work with the new column field attribute notation if i have not missed any additionally required adaptations just add the following column to showcase ui data dataexporter basic xhtml to verify p column field code headertext codenew the exporter delivers empty fields for such a column
1
75,631
25,961,736,047
IssuesEvent
2022-12-19 00:21:00
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
Message text is truncated without warning
T-Defect
### Steps to reproduce End of message is truncated on iOS element (and visable on web/elsewhere). I sent said message into encrypted chat. I haven't seen this issue since the iOS beta. ![truncated](https://user-images.githubusercontent.com/10473201/208327231-0fb1310f-107f-4ebf-9f14-b6036e946210.jpeg) <img width="371" alt="non-truncated" src="https://user-images.githubusercontent.com/10473201/208327234-f5d5f133-d098-45b9-9354-d5e9ab8be2c7.png"> ### Outcome #### What did you expect? Message to be displayed intact. #### What happened instead? Message is truncated. ### Your phone model iPhone 14 Pro ### Operating system version iOS 16.0.2 ### Application version Element iOS 1.9.13 ### Homeserver Synapse via matrix docker ansible deploy (not sure how to check version) ### Will you send logs? Yes
1.0
Message text is truncated without warning - ### Steps to reproduce End of message is truncated on iOS element (and visable on web/elsewhere). I sent said message into encrypted chat. I haven't seen this issue since the iOS beta. ![truncated](https://user-images.githubusercontent.com/10473201/208327231-0fb1310f-107f-4ebf-9f14-b6036e946210.jpeg) <img width="371" alt="non-truncated" src="https://user-images.githubusercontent.com/10473201/208327234-f5d5f133-d098-45b9-9354-d5e9ab8be2c7.png"> ### Outcome #### What did you expect? Message to be displayed intact. #### What happened instead? Message is truncated. ### Your phone model iPhone 14 Pro ### Operating system version iOS 16.0.2 ### Application version Element iOS 1.9.13 ### Homeserver Synapse via matrix docker ansible deploy (not sure how to check version) ### Will you send logs? Yes
defect
message text is truncated without warning steps to reproduce end of message is truncated on ios element and visable on web elsewhere i sent said message into encrypted chat i haven t seen this issue since the ios beta img width alt non truncated src outcome what did you expect message to be displayed intact what happened instead message is truncated your phone model iphone pro operating system version ios application version element ios homeserver synapse via matrix docker ansible deploy not sure how to check version will you send logs yes
1
33,571
7,166,893,008
IssuesEvent
2018-01-29 18:43:20
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Paginator numbers options first and last don't work like expected
Defect pagination
This is a (multiple allowed): * [x] bug * CakePHP Version: 3.5.10 * Platform and Target: Ubuntu 17.10 - Nginx 1.12.1 - PHP 7.1.11 ### What you did <?= $this->Paginator->numbers(['modulus' => 4, 'prev' => '<', 'next' => '>', 'first' => 1, 'last' => 1]) ?> ### What happened This is the generated code for the last param `<li class="page-item"><a href="/catalogue?page=247" class="page-link">1</a></li>` ### What you expected to happen The page number should be the last aka 247, not 1 like the doc say about it, Whether you want first links generated, set to an integer to define the number of ‘first’ links to generate. `<li class="page-item"><a href="/catalogue?page=247" class="page-link">247</a></li>` ### Possible fix `if (is_int($last) && $params['page'] <= $lower) {` [https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1090](https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1090) `if (is_int($options['last']) && $params['page'] <= $lower) {` The same goes for the first function `if (is_int($first) && $params['page'] >= $first) {` [https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1029](https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1029) `if (is_int($options['first']) && $params['page'] >= $first) {`
1.0
Paginator numbers options first and last don't work like expected - This is a (multiple allowed): * [x] bug * CakePHP Version: 3.5.10 * Platform and Target: Ubuntu 17.10 - Nginx 1.12.1 - PHP 7.1.11 ### What you did <?= $this->Paginator->numbers(['modulus' => 4, 'prev' => '<', 'next' => '>', 'first' => 1, 'last' => 1]) ?> ### What happened This is the generated code for the last param `<li class="page-item"><a href="/catalogue?page=247" class="page-link">1</a></li>` ### What you expected to happen The page number should be the last aka 247, not 1 like the doc say about it, Whether you want first links generated, set to an integer to define the number of ‘first’ links to generate. `<li class="page-item"><a href="/catalogue?page=247" class="page-link">247</a></li>` ### Possible fix `if (is_int($last) && $params['page'] <= $lower) {` [https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1090](https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1090) `if (is_int($options['last']) && $params['page'] <= $lower) {` The same goes for the first function `if (is_int($first) && $params['page'] >= $first) {` [https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1029](https://github.com/cakephp/cakephp/blob/master/src/View/Helper/PaginatorHelper.php#L1029) `if (is_int($options['first']) && $params['page'] >= $first) {`
defect
paginator numbers options first and last don t work like expected this is a multiple allowed bug cakephp version platform and target ubuntu nginx php what you did paginator numbers what happened this is the generated code for the last param what you expected to happen the page number should be the last aka not like the doc say about it whether you want first links generated set to an integer to define the number of ‘first’ links to generate possible fix if is int last params lower if is int options params lower the same goes for the first function if is int first params first if is int options params first
1
1,856
2,603,972,526
IssuesEvent
2015-02-24 19:00:37
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳男性生殖器疣
auto-migrated Priority-Medium Type-Defect
``` 沈阳男性生殖器疣〓沈陽軍區政治部醫院性病〓TEL:024-3102330 8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于� ��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
1.0
沈阳男性生殖器疣 - ``` 沈阳男性生殖器疣〓沈陽軍區政治部醫院性病〓TEL:024-3102330 8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于� ��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
defect
沈阳男性生殖器疣 沈阳男性生殖器疣〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位于� �� 。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 original issue reported on code google com by gmail com on jun at
1
66,386
20,169,425,762
IssuesEvent
2022-02-10 09:03:45
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Notifications right panel: Something went wrong!
T-Defect X-Regression
### Steps to reproduce 1. Not sure, this just appeared. ### Outcome > Something went wrong! > Please [create a new issue](https://github.com/vector-im/element-web/issues/new/choose) on GitHub so that we can investigate this bug. ### Operating system mac ### Browser information chrome ### URL for webapp develop.element.io ### Application version 762fc53c6150-react-254dbeeccb26-js-41bf8c2d5f4a ### Homeserver element.io ### Will you send logs? Yes
1.0
Notifications right panel: Something went wrong! - ### Steps to reproduce 1. Not sure, this just appeared. ### Outcome > Something went wrong! > Please [create a new issue](https://github.com/vector-im/element-web/issues/new/choose) on GitHub so that we can investigate this bug. ### Operating system mac ### Browser information chrome ### URL for webapp develop.element.io ### Application version 762fc53c6150-react-254dbeeccb26-js-41bf8c2d5f4a ### Homeserver element.io ### Will you send logs? Yes
defect
notifications right panel something went wrong steps to reproduce not sure this just appeared outcome something went wrong please on github so that we can investigate this bug operating system mac browser information chrome url for webapp develop element io application version react js homeserver element io will you send logs yes
1
15,720
2,869,002,190
IssuesEvent
2015-06-05 22:30:06
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Allow transformers to generate/modify code in packages as a result of whole-code evaluation
Area-Pkg Pkg-Barback Priority-Medium Triaged Type-Defect
Right now transformers can only modify code in packages when that package is being transformed, but there are a number of scenarios where the transformation on the package needs to be done with the results of some whole-program analysis. The specific scenario I'm running into is generating code for Angular in a lazy-loading environment. In a non-lazy environment the generated caches are all injected into the application entry point, but in the lazy-loaded scenario the code needs to be injected into the root of the lazy library. The issue is that the calculation of what needs to be cached needs to take the entire app into account. One approach discussed is the 'global transformers' which execute after all package transformers and can modify every package contents. Marking as critical as it is required for Angular's transformers once lazy-loading is supported.
1.0
Allow transformers to generate/modify code in packages as a result of whole-code evaluation - Right now transformers can only modify code in packages when that package is being transformed, but there are a number of scenarios where the transformation on the package needs to be done with the results of some whole-program analysis. The specific scenario I'm running into is generating code for Angular in a lazy-loading environment. In a non-lazy environment the generated caches are all injected into the application entry point, but in the lazy-loaded scenario the code needs to be injected into the root of the lazy library. The issue is that the calculation of what needs to be cached needs to take the entire app into account. One approach discussed is the 'global transformers' which execute after all package transformers and can modify every package contents. Marking as critical as it is required for Angular's transformers once lazy-loading is supported.
defect
allow transformers to generate modify code in packages as a result of whole code evaluation right now transformers can only modify code in packages when that package is being transformed but there are a number of scenarios where the transformation on the package needs to be done with the results of some whole program analysis the specific scenario i m running into is generating code for angular in a lazy loading environment in a non lazy environment the generated caches are all injected into the application entry point but in the lazy loaded scenario the code needs to be injected into the root of the lazy library the issue is that the calculation of what needs to be cached needs to take the entire app into account one approach discussed is the global transformers which execute after all package transformers and can modify every package contents marking as critical as it is required for angular s transformers once lazy loading is supported
1
464,648
13,337,046,015
IssuesEvent
2020-08-28 08:35:54
onvif/specs
https://api.github.com/repos/onvif/specs
opened
Clarify RecordingJobStateChange data
Priority minor Recording Control
#2701 Current Problem The DTT tests that the payload does not contain Sources information which absolutely makes sense to limit the size of the event. The specification specifies that other information shall be provided . Proposal Add to 5.25.1 The device shall omit the Sources parameter when emitting the event. Attachments (0) Oldest first Newest first Comments only Change History (1) Changed 4 months ago by hans.busch@de.bosch.com Correction, our sample forgot to repeat the State information in the ElementItem. The DTT was just creating wrong output. My conclusion is that the whole element item should be made optional since the only mandatory items state and token repeat the simple item parameters. Note that this CR would probably require a relaxation of the DTT.
1.0
Clarify RecordingJobStateChange data - #2701 Current Problem The DTT tests that the payload does not contain Sources information which absolutely makes sense to limit the size of the event. The specification specifies that other information shall be provided . Proposal Add to 5.25.1 The device shall omit the Sources parameter when emitting the event. Attachments (0) Oldest first Newest first Comments only Change History (1) Changed 4 months ago by hans.busch@de.bosch.com Correction, our sample forgot to repeat the State information in the ElementItem. The DTT was just creating wrong output. My conclusion is that the whole element item should be made optional since the only mandatory items state and token repeat the simple item parameters. Note that this CR would probably require a relaxation of the DTT.
non_defect
clarify recordingjobstatechange data current problem the dtt tests that the payload does not contain sources information which absolutely makes sense to limit the size of the event the specification specifies that other information shall be provided proposal add to the device shall omit the sources parameter when emitting the event attachments oldest first newest first comments only change history changed months ago by hans busch de bosch com correction our sample forgot to repeat the state information in the elementitem the dtt was just creating wrong output my conclusion is that the whole element item should be made optional since the only mandatory items state and token repeat the simple item parameters note that this cr would probably require a relaxation of the dtt
0
38,703
8,952,343,428
IssuesEvent
2019-01-25 16:17:30
svigerske/ipopt-donotuse
https://api.github.com/repos/svigerske/ipopt-donotuse
closed
wget not found when building IPOPT on MinGW/MSYS
Ipopt defect
Issue created by migration from Trac. Original creator: jdpipe Original creation time: 2010-05-06 07:38:33 Assignee: ipopt-team Version: 3.8 I use the 'wget' from the GnuWin32 project, which by default installs in c:\Program Files\GnuWin32\bin. I then add that folder to my PATH in MSYS and all GnuWin32 commands are then available straight away. These are much easier to install than the MSYS/MinGW stuff, and work fine. Problem is that the various 'get.PACKAGENAME' scripts in the ThirdParty section of IPOPT fail to detect my 'wget', because the call 'which wget | wc -w' returns a number more than 1, but a value of 1 is the only allowable. The solution to this bug, I believe, is a to simply change 'wc -w' to 'wc -l' for all of these 'wget' calls. Alternatively, I wonder if the return value of the 'which' command can be checked? For ease of building, maybe a 'get.ThirdParty' script could be provided that automatically downloaded all possible third-party libraries (except HSL which can't be automated...easily).
1.0
wget not found when building IPOPT on MinGW/MSYS - Issue created by migration from Trac. Original creator: jdpipe Original creation time: 2010-05-06 07:38:33 Assignee: ipopt-team Version: 3.8 I use the 'wget' from the GnuWin32 project, which by default installs in c:\Program Files\GnuWin32\bin. I then add that folder to my PATH in MSYS and all GnuWin32 commands are then available straight away. These are much easier to install than the MSYS/MinGW stuff, and work fine. Problem is that the various 'get.PACKAGENAME' scripts in the ThirdParty section of IPOPT fail to detect my 'wget', because the call 'which wget | wc -w' returns a number more than 1, but a value of 1 is the only allowable. The solution to this bug, I believe, is a to simply change 'wc -w' to 'wc -l' for all of these 'wget' calls. Alternatively, I wonder if the return value of the 'which' command can be checked? For ease of building, maybe a 'get.ThirdParty' script could be provided that automatically downloaded all possible third-party libraries (except HSL which can't be automated...easily).
defect
wget not found when building ipopt on mingw msys issue created by migration from trac original creator jdpipe original creation time assignee ipopt team version i use the wget from the project which by default installs in c program files bin i then add that folder to my path in msys and all commands are then available straight away these are much easier to install than the msys mingw stuff and work fine problem is that the various get packagename scripts in the thirdparty section of ipopt fail to detect my wget because the call which wget wc w returns a number more than but a value of is the only allowable the solution to this bug i believe is a to simply change wc w to wc l for all of these wget calls alternatively i wonder if the return value of the which command can be checked for ease of building maybe a get thirdparty script could be provided that automatically downloaded all possible third party libraries except hsl which can t be automated easily
1
8,706
7,374,082,286
IssuesEvent
2018-03-13 19:10:30
SpiderLabs/ModSecurity
https://api.github.com/repos/SpiderLabs/ModSecurity
closed
LibModSecurity + Nginx + ClamAV
3.x Pending feedback Type - Usage libmodsecurity
Hi all Using latest version of ModSecuity and nginx connector. However even in debug level 9 it still wont show me what issue is. I have a rule below: `SecRule FILES_TMPNAMES "@inspectFile /usr/local/nginx/sbin/test.lua" "phase:2,t:none,log,deny,msg:'Malicous File Attachment Identified.',id:99999"` And in modsec settings I have enabled secresponsebody and following: `SecTmpSaveUploadedFiles On SecUploadDir /usr/local/nginx/tmp SecUploadKeepFiles On SecUploadFileMode 0777 ` I have a test php page to upload a file to nginx server but even with test virus it does not pick anything up. Here is debug log: https://pastebin.com/GjdXM3yF and here is the lua script https://gist.github.com/angeloxx/97714f9108b3642460564acdcd37b34a thanks
True
LibModSecurity + Nginx + ClamAV - Hi all Using latest version of ModSecuity and nginx connector. However even in debug level 9 it still wont show me what issue is. I have a rule below: `SecRule FILES_TMPNAMES "@inspectFile /usr/local/nginx/sbin/test.lua" "phase:2,t:none,log,deny,msg:'Malicous File Attachment Identified.',id:99999"` And in modsec settings I have enabled secresponsebody and following: `SecTmpSaveUploadedFiles On SecUploadDir /usr/local/nginx/tmp SecUploadKeepFiles On SecUploadFileMode 0777 ` I have a test php page to upload a file to nginx server but even with test virus it does not pick anything up. Here is debug log: https://pastebin.com/GjdXM3yF and here is the lua script https://gist.github.com/angeloxx/97714f9108b3642460564acdcd37b34a thanks
non_defect
libmodsecurity nginx clamav hi all using latest version of modsecuity and nginx connector however even in debug level it still wont show me what issue is i have a rule below secrule files tmpnames inspectfile usr local nginx sbin test lua phase t none log deny msg malicous file attachment identified id and in modsec settings i have enabled secresponsebody and following sectmpsaveuploadedfiles on secuploaddir usr local nginx tmp secuploadkeepfiles on secuploadfilemode i have a test php page to upload a file to nginx server but even with test virus it does not pick anything up here is debug log and here is the lua script thanks
0
50,120
13,187,333,606
IssuesEvent
2020-08-13 03:04:47
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
ROOTSYS in `make tarball` tarball (Trac #149)
Migrated from Trac cmake defect
root under some circumstances rummages through various directories under ROOTSYS. What are they, does the 'make tarball' step need to copy more stuff under I3_BUILD/CMAKE_INSTALL_PREFIX/lib/? <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/149 , reported by troy and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2011-04-14T19:21:48", "description": "root under some circumstances rummages through various directories under ROOTSYS. What are they, does the 'make tarball' step need to copy more stuff under I3_BUILD/CMAKE_INSTALL_PREFIX/lib/?\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1302808908000000", "component": "cmake", "summary": "ROOTSYS in `make tarball` tarball", "priority": "normal", "keywords": "", "time": "2008-11-18T14:04:08", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
ROOTSYS in `make tarball` tarball (Trac #149) - root under some circumstances rummages through various directories under ROOTSYS. What are they, does the 'make tarball' step need to copy more stuff under I3_BUILD/CMAKE_INSTALL_PREFIX/lib/? <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/149 , reported by troy and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2011-04-14T19:21:48", "description": "root under some circumstances rummages through various directories under ROOTSYS. What are they, does the 'make tarball' step need to copy more stuff under I3_BUILD/CMAKE_INSTALL_PREFIX/lib/?\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1302808908000000", "component": "cmake", "summary": "ROOTSYS in `make tarball` tarball", "priority": "normal", "keywords": "", "time": "2008-11-18T14:04:08", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
rootsys in make tarball tarball trac root under some circumstances rummages through various directories under rootsys what are they does the make tarball step need to copy more stuff under build cmake install prefix lib migrated from reported by troy and owned by troy json status closed changetime description root under some circumstances rummages through various directories under rootsys what are they does the make tarball step need to copy more stuff under build cmake install prefix lib n n reporter troy cc resolution fixed ts component cmake summary rootsys in make tarball tarball priority normal keywords time milestone owner troy type defect
1
26,580
4,768,870,588
IssuesEvent
2016-10-26 10:33:27
gbif/ipt
https://api.github.com/repos/gbif/ipt
closed
Project data are not saved
bug Metadata MetadataEditor Priority-High Type-Defect
Odd that I never noticed, but **project data has been removed** for all datasets where we populated this. ## Example: http://dataset.inbo.be/bird-tracking-gull-occurrences * We definitely added project data to this dataset. We have a record of it in the [dataset metadata we maintain on GitHub](https://github.com/inbo/data-publication/blob/master/datasets/bird-tracking-gull-occurrences/metadata.md#project-data) * All fields for project data are empty in the metadata editor * No project information appears on the dataset page on [IPT](http://dataset.inbo.be/bird-tracking-gull-occurrences) nor [GBIF](http://www.gbif.org/dataset/83e20573-f7dd-4852-9159-21566e1e691e) Oddly though, if I re-add project metadata in the editor (title & funding), that information is saved, but: * Does not appear on the dataset preview page (http://data.inbo.be/ipt/resource/preview?r=bird-tracking-gull-occurrences) * Is apparently lost on publication (i.e. all project data fields are empty again). Does this bug only affects us? If not, since when do republication empty project information? Has project information ever appeared on GBIF?
1.0
Project data are not saved - Odd that I never noticed, but **project data has been removed** for all datasets where we populated this. ## Example: http://dataset.inbo.be/bird-tracking-gull-occurrences * We definitely added project data to this dataset. We have a record of it in the [dataset metadata we maintain on GitHub](https://github.com/inbo/data-publication/blob/master/datasets/bird-tracking-gull-occurrences/metadata.md#project-data) * All fields for project data are empty in the metadata editor * No project information appears on the dataset page on [IPT](http://dataset.inbo.be/bird-tracking-gull-occurrences) nor [GBIF](http://www.gbif.org/dataset/83e20573-f7dd-4852-9159-21566e1e691e) Oddly though, if I re-add project metadata in the editor (title & funding), that information is saved, but: * Does not appear on the dataset preview page (http://data.inbo.be/ipt/resource/preview?r=bird-tracking-gull-occurrences) * Is apparently lost on publication (i.e. all project data fields are empty again). Does this bug only affects us? If not, since when do republication empty project information? Has project information ever appeared on GBIF?
defect
project data are not saved odd that i never noticed but project data has been removed for all datasets where we populated this example we definitely added project data to this dataset we have a record of it in the all fields for project data are empty in the metadata editor no project information appears on the dataset page on nor oddly though if i re add project metadata in the editor title funding that information is saved but does not appear on the dataset preview page is apparently lost on publication i e all project data fields are empty again does this bug only affects us if not since when do republication empty project information has project information ever appeared on gbif
1
62,204
17,023,872,188
IssuesEvent
2021-07-03 04:17:40
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
New zoom slower
Component: website Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 2.25pm, Monday, 5th August 2013]** Back when there was no animation during zooming, you could instantly zoom into a much higher magnification with a long scroll of the mouse wheel. Now, every magnification is treated as an event during which no new commands are read. That is, you have to wait a short time after each zooming event before you can zoom in again, making zooming a slower operation. Now with the new map controls and the disappearance of the zoom scale, zooming is made further inconvenient. Are there any benefits of this functionality? If not, can the old zooming behaviour be brought back?
1.0
New zoom slower - **[Submitted to the original trac issue database at 2.25pm, Monday, 5th August 2013]** Back when there was no animation during zooming, you could instantly zoom into a much higher magnification with a long scroll of the mouse wheel. Now, every magnification is treated as an event during which no new commands are read. That is, you have to wait a short time after each zooming event before you can zoom in again, making zooming a slower operation. Now with the new map controls and the disappearance of the zoom scale, zooming is made further inconvenient. Are there any benefits of this functionality? If not, can the old zooming behaviour be brought back?
defect
new zoom slower back when there was no animation during zooming you could instantly zoom into a much higher magnification with a long scroll of the mouse wheel now every magnification is treated as an event during which no new commands are read that is you have to wait a short time after each zooming event before you can zoom in again making zooming a slower operation now with the new map controls and the disappearance of the zoom scale zooming is made further inconvenient are there any benefits of this functionality if not can the old zooming behaviour be brought back
1
11,532
5,028,663,002
IssuesEvent
2016-12-15 18:54:13
quicklisp/quicklisp-projects
https://api.github.com/repos/quicklisp/quicklisp-projects
closed
Please add stl
canbuild
Reads triangle data from binary STL files used in 3D graphics and for 3D printing. Can be found here: https://github.com/jl2/stl
1.0
Please add stl - Reads triangle data from binary STL files used in 3D graphics and for 3D printing. Can be found here: https://github.com/jl2/stl
non_defect
please add stl reads triangle data from binary stl files used in graphics and for printing can be found here
0
61,652
17,023,749,043
IssuesEvent
2021-07-03 03:38:15
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
osmarender vs Mapnik formats result in different latitudes for north edge of map
Component: admin Priority: minor Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 6.37am, Friday, 30th September 2011]** Greetings,[[BR]] This is my first post, please bear with me.[[BR]] I've noticed that when generating map data from [http://openstreetmap.org], that there appears to be some inconsistency with the latitude when exporting maps in different formats.[[BR]] For example, if I export a map of Northern Ontario, Canada with manually entered coordinates:[[BR]] 57[[BR]] -95.2 -79.5[[BR]] 52[[BR]] [[BR]] a. first in Mapnik format (png, 3050000)[[BR]] b. then in Osmarender format (png, 7)[[BR]] the two maps obviously differ in latitude: the osmarender map is further south. The differences are I believe mathematically significant. Map file outputs are attached.[[BR]] I personally am most concerned about the coordinate accuracy of the Osmarender generation. Generating different Osmarender images at different Zooms gives a consistent approach to latitude, which is a good thing! 8-) [[BR]] 1. Ubuntu linux 10.04, Firefox 3.6.23[[BR]] 2. [http://openstreetmap.org] Export feature[[BR]] 3. step by step[[BR]] 3a open [http://openstreetmap.org] in firefox[[BR]] 3b click on Export button[[BR]] 3c click on Export button again to see N W E S text entry boxes[[BR]] 3d enter values above into respective coordinate text boxes[[BR]] 3e click on MapnikImage[[BR]] 3f enter maximum scale (2350000)[[BR]] 3g click Export, save image to disk (osm_nnontario_mapnik_2350000.png)[[BR]] 3h click OsmarenderImage[[BR]] 3i select Zoom of 7[[BR]] 3j click Export, save image to disk (osm_nnontario_osmarender_7b.png)[[BR]] 3k open both images and compare 4 Expected both images to cover the **same** coordinates.[[BR]] 5 Osmarender map is further south than Mapnik. Unsure which is at correct latitude.[[BR]] CU,[[BR]] james
1.0
osmarender vs Mapnik formats result in different latitudes for north edge of map - **[Submitted to the original trac issue database at 6.37am, Friday, 30th September 2011]** Greetings,[[BR]] This is my first post, please bear with me.[[BR]] I've noticed that when generating map data from [http://openstreetmap.org], that there appears to be some inconsistency with the latitude when exporting maps in different formats.[[BR]] For example, if I export a map of Northern Ontario, Canada with manually entered coordinates:[[BR]] 57[[BR]] -95.2 -79.5[[BR]] 52[[BR]] [[BR]] a. first in Mapnik format (png, 3050000)[[BR]] b. then in Osmarender format (png, 7)[[BR]] the two maps obviously differ in latitude: the osmarender map is further south. The differences are I believe mathematically significant. Map file outputs are attached.[[BR]] I personally am most concerned about the coordinate accuracy of the Osmarender generation. Generating different Osmarender images at different Zooms gives a consistent approach to latitude, which is a good thing! 8-) [[BR]] 1. Ubuntu linux 10.04, Firefox 3.6.23[[BR]] 2. [http://openstreetmap.org] Export feature[[BR]] 3. step by step[[BR]] 3a open [http://openstreetmap.org] in firefox[[BR]] 3b click on Export button[[BR]] 3c click on Export button again to see N W E S text entry boxes[[BR]] 3d enter values above into respective coordinate text boxes[[BR]] 3e click on MapnikImage[[BR]] 3f enter maximum scale (2350000)[[BR]] 3g click Export, save image to disk (osm_nnontario_mapnik_2350000.png)[[BR]] 3h click OsmarenderImage[[BR]] 3i select Zoom of 7[[BR]] 3j click Export, save image to disk (osm_nnontario_osmarender_7b.png)[[BR]] 3k open both images and compare 4 Expected both images to cover the **same** coordinates.[[BR]] 5 Osmarender map is further south than Mapnik. Unsure which is at correct latitude.[[BR]] CU,[[BR]] james
defect
osmarender vs mapnik formats result in different latitudes for north edge of map greetings this is my first post please bear with me i ve noticed that when generating map data from that there appears to be some inconsistency with the latitude when exporting maps in different formats for example if i export a map of northern ontario canada with manually entered coordinates a first in mapnik format png b then in osmarender format png the two maps obviously differ in latitude the osmarender map is further south the differences are i believe mathematically significant map file outputs are attached i personally am most concerned about the coordinate accuracy of the osmarender generation generating different osmarender images at different zooms gives a consistent approach to latitude which is a good thing ubuntu linux firefox export feature step by step open in firefox click on export button click on export button again to see n w e s text entry boxes enter values above into respective coordinate text boxes click on mapnikimage enter maximum scale click export save image to disk osm nnontario mapnik png click osmarenderimage select zoom of click export save image to disk osm nnontario osmarender png open both images and compare expected both images to cover the same coordinates osmarender map is further south than mapnik unsure which is at correct latitude cu james
1
29,457
5,693,473,673
IssuesEvent
2017-04-15 01:52:32
hugotacito/django-pagination
https://api.github.com/repos/hugotacito/django-pagination
closed
InfinitePaginator should re-define start_index
auto-migrated Priority-Medium Type-Defect
``` The default django Page class defines start_index to have a special case when self.paginator.count == 0. This will fail in the InfinitePaginator, which needs to redefine start_index without that special case. See trivial patch attached. ``` Original issue reported on code.google.com by `andrew.a...@gmail.com` on 14 Sep 2010 at 6:38 Attachments: - [paginator.patch](https://storage.googleapis.com/google-code-attachments/django-pagination/issue-73/comment-0/paginator.patch)
1.0
InfinitePaginator should re-define start_index - ``` The default django Page class defines start_index to have a special case when self.paginator.count == 0. This will fail in the InfinitePaginator, which needs to redefine start_index without that special case. See trivial patch attached. ``` Original issue reported on code.google.com by `andrew.a...@gmail.com` on 14 Sep 2010 at 6:38 Attachments: - [paginator.patch](https://storage.googleapis.com/google-code-attachments/django-pagination/issue-73/comment-0/paginator.patch)
defect
infinitepaginator should re define start index the default django page class defines start index to have a special case when self paginator count this will fail in the infinitepaginator which needs to redefine start index without that special case see trivial patch attached original issue reported on code google com by andrew a gmail com on sep at attachments
1
193,379
6,884,457,625
IssuesEvent
2017-11-21 13:07:15
citusdata/citus
https://api.github.com/repos/citusdata/citus
closed
Feature regression on subquery pushdowns with reference tables from Citus 7 to 7.1
priority:high
@saicitus just reported that three queries that used to work in Citus 7, but that #no longer do on Citus 7.1. All queries are related to subquery pushdowns in reference tables in Citus. @saicitus will share more details in the upcoming days. <pre> SELECT DBE.id, DBE.event, DBE.userid, EXTRACT(EPOCH FROM FSE.timestamp) AS my_time, DBE.pointid, NULL AS info FROM database_event AS DBE INNER JOIN ( SELECT DISTINCT test.id FROM testpoint AS test WHERE test.unique_id IN ('00cfdd') ) AS TEST_NAMES ON TEST_NAMES.id = DBE.pointid ORDER BY DBE.my_time DESC LIMIT 20 OFFSET 0; </pre> The `database_event` table is distributed on `pointid`. The `testpoint` table is a reference table. In Citus 7, this query executes nicely. In Citus 7.1, we see the following error message: <pre> psql:q_distinct.sql:31: ERROR: cannot push down this subquery DETAIL: Distinct on columns without partition column is currently unsupported </pre> Removing the `DISTINCT` clause also doesn't help. We then get this error message in Citus 7.1 <pre> psql:q.sql:31: ERROR: cannot push down this subquery DETAIL: Group by list without partition column is currently unsupported </pre>
1.0
Feature regression on subquery pushdowns with reference tables from Citus 7 to 7.1 - @saicitus just reported that three queries that used to work in Citus 7, but that #no longer do on Citus 7.1. All queries are related to subquery pushdowns in reference tables in Citus. @saicitus will share more details in the upcoming days. <pre> SELECT DBE.id, DBE.event, DBE.userid, EXTRACT(EPOCH FROM FSE.timestamp) AS my_time, DBE.pointid, NULL AS info FROM database_event AS DBE INNER JOIN ( SELECT DISTINCT test.id FROM testpoint AS test WHERE test.unique_id IN ('00cfdd') ) AS TEST_NAMES ON TEST_NAMES.id = DBE.pointid ORDER BY DBE.my_time DESC LIMIT 20 OFFSET 0; </pre> The `database_event` table is distributed on `pointid`. The `testpoint` table is a reference table. In Citus 7, this query executes nicely. In Citus 7.1, we see the following error message: <pre> psql:q_distinct.sql:31: ERROR: cannot push down this subquery DETAIL: Distinct on columns without partition column is currently unsupported </pre> Removing the `DISTINCT` clause also doesn't help. We then get this error message in Citus 7.1 <pre> psql:q.sql:31: ERROR: cannot push down this subquery DETAIL: Group by list without partition column is currently unsupported </pre>
non_defect
feature regression on subquery pushdowns with reference tables from citus to saicitus just reported that three queries that used to work in citus but that no longer do on citus all queries are related to subquery pushdowns in reference tables in citus saicitus will share more details in the upcoming days select dbe id dbe event dbe userid extract epoch from fse timestamp as my time dbe pointid null as info from database event as dbe inner join select distinct test id from testpoint as test where test unique id in as test names on test names id dbe pointid order by dbe my time desc limit offset the database event table is distributed on pointid the testpoint table is a reference table in citus this query executes nicely in citus we see the following error message psql q distinct sql error cannot push down this subquery detail distinct on columns without partition column is currently unsupported removing the distinct clause also doesn t help we then get this error message in citus psql q sql error cannot push down this subquery detail group by list without partition column is currently unsupported
0
405,134
11,868,234,931
IssuesEvent
2020-03-26 08:48:31
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
AMP setting panel is not showing when ""WPML for AMP" pluign is active.
[Priority: HIGH] bug
The user is not able to see AMP Setting Panel(https://prnt.sc/rmbysh). getting blank page in amp settings. when we deactivate the "WPML for AMP" plugin then the AMP setting panel works properly. Note: this issue is not occurring in the localhost. Ref: https://secure.helpscout.net/conversation/1114266259/117604?folderId=3257665
1.0
AMP setting panel is not showing when ""WPML for AMP" pluign is active. - The user is not able to see AMP Setting Panel(https://prnt.sc/rmbysh). getting blank page in amp settings. when we deactivate the "WPML for AMP" plugin then the AMP setting panel works properly. Note: this issue is not occurring in the localhost. Ref: https://secure.helpscout.net/conversation/1114266259/117604?folderId=3257665
non_defect
amp setting panel is not showing when wpml for amp pluign is active the user is not able to see amp setting panel getting blank page in amp settings when we deactivate the wpml for amp plugin then the amp setting panel works properly note this issue is not occurring in the localhost ref
0
180,565
14,787,500,219
IssuesEvent
2021-01-12 07:43:32
PaxJaromeMalues/arma3_cram
https://api.github.com/repos/PaxJaromeMalues/arma3_cram
opened
hit chance
documentation enhancement
Currently the turret is set to fire for a random amount of time onto the targeted projectile. The length of those salvos is stored in `_shots`. For testing and development this provides enough variation in hit chances to be performant. Later on a real chance to hit might need to be implemented.
1.0
hit chance - Currently the turret is set to fire for a random amount of time onto the targeted projectile. The length of those salvos is stored in `_shots`. For testing and development this provides enough variation in hit chances to be performant. Later on a real chance to hit might need to be implemented.
non_defect
hit chance currently the turret is set to fire for a random amount of time onto the targeted projectile the length of those salvos is stored in shots for testing and development this provides enough variation in hit chances to be performant later on a real chance to hit might need to be implemented
0
42,098
17,029,352,332
IssuesEvent
2021-07-04 08:26:00
MOONGIFT/moongift.dev
https://api.github.com/repos/MOONGIFT/moongift.dev
opened
Napkin – Backend in the Browser
Backend Web API Web service
https://www.napkin.io/ Write, run, and deploy API endpoints, all from the browser. No infra. No boilerplate. No devops. Just code. バックエンドサービス。
1.0
Napkin – Backend in the Browser - https://www.napkin.io/ Write, run, and deploy API endpoints, all from the browser. No infra. No boilerplate. No devops. Just code. バックエンドサービス。
non_defect
napkin – backend in the browser write run and deploy api endpoints all from the browser no infra no boilerplate no devops just code バックエンドサービス。
0
77,294
26,902,772,390
IssuesEvent
2023-02-06 16:44:51
dotCMS/core
https://api.github.com/repos/dotCMS/core
opened
Error when update zero size file through Webdav.
Type : Defect Triage
### Problem Statement When try to update a zero size file through webdav an error appears, this error is related with a `fk_contentlet_version_info_working` value. ### Steps to Reproduce ## Windows commands: Create zero size file `type nul > zero-size.file` Create non-zero size file `echo not empty > non-zero-size.file` Check the file sizes `dir *.file` Upload non-zero size file with webdav using curl `curl --verbose --user "admin@dotcms.com:admin" -T ".\non-zero-size.file" https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` Upload zero size file with webdav using curl `curl --verbose --user "admin@dotcms.com:admin" -T ".\zero-size.file" https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` ## Linux commands: Create zero size file `touch zero-size.file` Create non-zero size file `echo not empty > non-zero-size.file` Check the file sizes `ls *.file -al` Upload non-zero size file with webdav using curl `curl --verbose --user 'admin@dotcms.com:admin' -T './non-zero-size.file' https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` Upload zero size file with webdav using curl `curl --verbose --user 'admin@dotcms.com:admin' -T './zero-size.file' https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` ### Following error message appears when upload/update zero size file: ``` [06/02/23 16:37:51:138 UTC] WARN webdav.HostResourceImpl: An error occurred while creating new file: test-file.vtl in this path: /webdav/live/1/default/ ERROR: update or delete on table "contentlet" violates foreign key constraint "fk_contentlet_version_info_working" on table "contentlet_version_info" Detail: Key (inode)=(3a5ebb93-7dd8-4898-bb30-3fb959d5b222) is still referenced from table "contentlet_version_info".{ "SQL": ["delete from contentlet where inode=?"], "maxRows": [-1], "offest": [0], "params": "3a5ebb93-7dd8-4898-bb30-3fb959d5b222" } com.dotmarketing.exception.DotDataException: ERROR: update or delete on table "contentlet" violates foreign key constraint "fk_contentlet_version_info_working" on table "contentlet_version_info" Detail: Key (inode)=(3a5ebb93-7dd8-4898-bb30-3fb959d5b222) is still referenced from table "contentlet_version_info".{ "SQL": ["delete from contentlet where inode=?"], "maxRows": [-1], "offest": [0], "params": "3a5ebb93-7dd8-4898-bb30-3fb959d5b222" } at com.dotmarketing.common.db.DotConnect.loadResult(DotConnect.java:273) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentFactoryImpl.delete(ESContentFactoryImpl.java:603) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentFactoryImpl.deleteVersion(ESContentFactoryImpl.java:766) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl.deleteVersion_aroundBody88(ESContentletAPIImpl.java:2870) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl$AjcClosure89.run(ESContentletAPIImpl.java:1) ~[dotcms_22.03.2_999999.jar:?] at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.9.2.jar:?] at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.db.LocalTransaction.wrapReturnWithListeners(LocalTransaction.java:119) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.aspects.interceptors.WrapInTransactionMethodInterceptor.invoke(WrapInTransactionMethodInterceptor.java:30) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.aspects.aspectj.WrapInTransactionAspect.invoke(WrapInTransactionAspect.java:41) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl.deleteVersion(ESContentletAPIImpl.java:2856) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.portlets.contentlet.business.ContentletAPIInterceptor.deleteVersion(ContentletAPIInterceptor.java:2362) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.webdav.DotWebdavHelper.setResourceContent(DotWebdavHelper.java:849) ~[dotcms_22.03.2_999999.jar:?] ``` ![Screenshot 2023-02-06 at 8 36 37 AM](https://user-images.githubusercontent.com/12899960/217030103-0468ea83-5e1c-401f-b9c5-5fbeff6f9c3a.png) ![Screenshot 2023-02-06 at 8 37 09 AM](https://user-images.githubusercontent.com/12899960/217030213-016e78ea-bacc-4046-8eda-2a019c3 ![Screenshot 2023-02-06 at 8 37 27 AM](https://user-images.githubusercontent.com/12899960/217030275-ea613166-4171-4cbb-a154-93b73c7e3b4c.png) f894b.png) ### Acceptance Criteria In both cases (zero file size or non-zero file size) the file can be updated/upload and it's showing good results in backend dashboard, but those error messages still in the system logs. ### dotCMS Version This was tested locally with 22.03.2 version and demo site 23.01. ### Proposed Objective Application Performance ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. https://dotcms.zendesk.com/agent/tickets/109728 ### Assumptions & Initiation Needs _No response_ ### Sub-Tasks & Estimates _No response_
1.0
Error when update zero size file through Webdav. - ### Problem Statement When try to update a zero size file through webdav an error appears, this error is related with a `fk_contentlet_version_info_working` value. ### Steps to Reproduce ## Windows commands: Create zero size file `type nul > zero-size.file` Create non-zero size file `echo not empty > non-zero-size.file` Check the file sizes `dir *.file` Upload non-zero size file with webdav using curl `curl --verbose --user "admin@dotcms.com:admin" -T ".\non-zero-size.file" https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` Upload zero size file with webdav using curl `curl --verbose --user "admin@dotcms.com:admin" -T ".\zero-size.file" https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` ## Linux commands: Create zero size file `touch zero-size.file` Create non-zero size file `echo not empty > non-zero-size.file` Check the file sizes `ls *.file -al` Upload non-zero size file with webdav using curl `curl --verbose --user 'admin@dotcms.com:admin' -T './non-zero-size.file' https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` Upload zero size file with webdav using curl `curl --verbose --user 'admin@dotcms.com:admin' -T './zero-size.file' https://demo.dotcms.com/webdav/live/1/demo.dotcms.com/test-file.vtl` ### Following error message appears when upload/update zero size file: ``` [06/02/23 16:37:51:138 UTC] WARN webdav.HostResourceImpl: An error occurred while creating new file: test-file.vtl in this path: /webdav/live/1/default/ ERROR: update or delete on table "contentlet" violates foreign key constraint "fk_contentlet_version_info_working" on table "contentlet_version_info" Detail: Key (inode)=(3a5ebb93-7dd8-4898-bb30-3fb959d5b222) is still referenced from table "contentlet_version_info".{ "SQL": ["delete from contentlet where inode=?"], "maxRows": [-1], "offest": [0], "params": "3a5ebb93-7dd8-4898-bb30-3fb959d5b222" } com.dotmarketing.exception.DotDataException: ERROR: update or delete on table "contentlet" violates foreign key constraint "fk_contentlet_version_info_working" on table "contentlet_version_info" Detail: Key (inode)=(3a5ebb93-7dd8-4898-bb30-3fb959d5b222) is still referenced from table "contentlet_version_info".{ "SQL": ["delete from contentlet where inode=?"], "maxRows": [-1], "offest": [0], "params": "3a5ebb93-7dd8-4898-bb30-3fb959d5b222" } at com.dotmarketing.common.db.DotConnect.loadResult(DotConnect.java:273) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentFactoryImpl.delete(ESContentFactoryImpl.java:603) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentFactoryImpl.deleteVersion(ESContentFactoryImpl.java:766) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl.deleteVersion_aroundBody88(ESContentletAPIImpl.java:2870) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl$AjcClosure89.run(ESContentletAPIImpl.java:1) ~[dotcms_22.03.2_999999.jar:?] at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.9.2.jar:?] at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.db.LocalTransaction.wrapReturnWithListeners(LocalTransaction.java:119) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.aspects.interceptors.WrapInTransactionMethodInterceptor.invoke(WrapInTransactionMethodInterceptor.java:30) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.aspects.aspectj.WrapInTransactionAspect.invoke(WrapInTransactionAspect.java:41) ~[dotcms_22.03.2_999999.jar:?] at com.dotcms.content.elasticsearch.business.ESContentletAPIImpl.deleteVersion(ESContentletAPIImpl.java:2856) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.portlets.contentlet.business.ContentletAPIInterceptor.deleteVersion(ContentletAPIInterceptor.java:2362) ~[dotcms_22.03.2_999999.jar:?] at com.dotmarketing.webdav.DotWebdavHelper.setResourceContent(DotWebdavHelper.java:849) ~[dotcms_22.03.2_999999.jar:?] ``` ![Screenshot 2023-02-06 at 8 36 37 AM](https://user-images.githubusercontent.com/12899960/217030103-0468ea83-5e1c-401f-b9c5-5fbeff6f9c3a.png) ![Screenshot 2023-02-06 at 8 37 09 AM](https://user-images.githubusercontent.com/12899960/217030213-016e78ea-bacc-4046-8eda-2a019c3 ![Screenshot 2023-02-06 at 8 37 27 AM](https://user-images.githubusercontent.com/12899960/217030275-ea613166-4171-4cbb-a154-93b73c7e3b4c.png) f894b.png) ### Acceptance Criteria In both cases (zero file size or non-zero file size) the file can be updated/upload and it's showing good results in backend dashboard, but those error messages still in the system logs. ### dotCMS Version This was tested locally with 22.03.2 version and demo site 23.01. ### Proposed Objective Application Performance ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. https://dotcms.zendesk.com/agent/tickets/109728 ### Assumptions & Initiation Needs _No response_ ### Sub-Tasks & Estimates _No response_
defect
error when update zero size file through webdav problem statement when try to update a zero size file through webdav an error appears this error is related with a fk contentlet version info working value steps to reproduce windows commands create zero size file type nul zero size file create non zero size file echo not empty non zero size file check the file sizes dir file upload non zero size file with webdav using curl curl verbose user admin dotcms com admin t non zero size file upload zero size file with webdav using curl curl verbose user admin dotcms com admin t zero size file linux commands create zero size file touch zero size file create non zero size file echo not empty non zero size file check the file sizes ls file al upload non zero size file with webdav using curl curl verbose user admin dotcms com admin t non zero size file upload zero size file with webdav using curl curl verbose user admin dotcms com admin t zero size file following error message appears when upload update zero size file warn webdav hostresourceimpl an error occurred while creating new file test file vtl in this path webdav live default error update or delete on table contentlet violates foreign key constraint fk contentlet version info working on table contentlet version info detail key inode is still referenced from table contentlet version info sql maxrows offest params com dotmarketing exception dotdataexception error update or delete on table contentlet violates foreign key constraint fk contentlet version info working on table contentlet version info detail key inode is still referenced from table contentlet version info sql maxrows offest params at com dotmarketing common db dotconnect loadresult dotconnect java at com dotcms content elasticsearch business escontentfactoryimpl delete escontentfactoryimpl java at com dotcms content elasticsearch business escontentfactoryimpl deleteversion escontentfactoryimpl java at com dotcms content elasticsearch business escontentletapiimpl deleteversion escontentletapiimpl java at com dotcms content elasticsearch business escontentletapiimpl run escontentletapiimpl java at org aspectj runtime reflect joinpointimpl proceed joinpointimpl java at com dotcms aspects aspectj aspectjdelegatemethodinvocation proceed aspectjdelegatemethodinvocation java at com dotmarketing db localtransaction wrapreturnwithlisteners localtransaction java at com dotcms aspects interceptors wrapintransactionmethodinterceptor invoke wrapintransactionmethodinterceptor java at com dotcms aspects aspectj wrapintransactionaspect invoke wrapintransactionaspect java at com dotcms content elasticsearch business escontentletapiimpl deleteversion escontentletapiimpl java at com dotmarketing portlets contentlet business contentletapiinterceptor deleteversion contentletapiinterceptor java at com dotmarketing webdav dotwebdavhelper setresourcecontent dotwebdavhelper java png acceptance criteria in both cases zero file size or non zero file size the file can be updated upload and it s showing good results in backend dashboard but those error messages still in the system logs dotcms version this was tested locally with version and demo site proposed objective application performance proposed priority priority average external links slack conversations support tickets figma designs etc assumptions initiation needs no response sub tasks estimates no response
1
14,676
2,831,388,436
IssuesEvent
2015-05-24 15:53:22
nobodyguy/dslrdashboard
https://api.github.com/repos/nobodyguy/dslrdashboard
closed
D600 Live View not working
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Using Nikon D-600 camera with WU-1B wireless device and Samgsung Galaxy S4 phone 2. 3. What is the expected output? What do you see instead? Live View will not function - get message to use Scene or User Mode. All other functions of app work. What version of the product are you using? On what operating system? Latest version of app (downloaded 3/28) and Android 4.3 Please provide any additional information below. ``` Original issue reported on code.google.com by `stevean...@gmail.com` on 29 Mar 2014 at 12:53
1.0
D600 Live View not working - ``` What steps will reproduce the problem? 1. Using Nikon D-600 camera with WU-1B wireless device and Samgsung Galaxy S4 phone 2. 3. What is the expected output? What do you see instead? Live View will not function - get message to use Scene or User Mode. All other functions of app work. What version of the product are you using? On what operating system? Latest version of app (downloaded 3/28) and Android 4.3 Please provide any additional information below. ``` Original issue reported on code.google.com by `stevean...@gmail.com` on 29 Mar 2014 at 12:53
defect
live view not working what steps will reproduce the problem using nikon d camera with wu wireless device and samgsung galaxy phone what is the expected output what do you see instead live view will not function get message to use scene or user mode all other functions of app work what version of the product are you using on what operating system latest version of app downloaded and android please provide any additional information below original issue reported on code google com by stevean gmail com on mar at
1
161,046
20,120,388,715
IssuesEvent
2022-02-08 01:14:11
Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
https://api.github.com/repos/Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
opened
CVE-2022-23573 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl
security vulnerability
## CVE-2022-23573 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Tensorflow is an Open Source Machine Learning Framework. The implementation of `AssignOp` can result in copying uninitialized data to a new tensor. This later results in undefined behavior. The implementation has a check that the left hand side of the assignment is initialized (to minimize number of allocations), but does not check that the right hand side is also initialized. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range. <p>Publish Date: 2022-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23573>CVE-2022-23573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-q85f-69q7-55h2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-q85f-69q7-55h2</a></p> <p>Release Date: 2022-02-04</p> <p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-cpu - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-gpu - 2.5.3,2.6.3,2.7.1,2.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-23573 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2022-23573 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Tensorflow is an Open Source Machine Learning Framework. The implementation of `AssignOp` can result in copying uninitialized data to a new tensor. This later results in undefined behavior. The implementation has a check that the left hand side of the assignment is initialized (to minimize number of allocations), but does not check that the right hand side is also initialized. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range. <p>Publish Date: 2022-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23573>CVE-2022-23573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-q85f-69q7-55h2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-q85f-69q7-55h2</a></p> <p>Release Date: 2022-02-04</p> <p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-cpu - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-gpu - 2.5.3,2.6.3,2.7.1,2.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file requirements txt path to vulnerable library raindrop removal with light field image using image inpainting requirements txt dependency hierarchy x tensorflow whl vulnerable library found in base branch main vulnerability details tensorflow is an open source machine learning framework the implementation of assignop can result in copying uninitialized data to a new tensor this later results in undefined behavior the implementation has a check that the left hand side of the assignment is initialized to minimize number of allocations but does not check that the right hand side is also initialized the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
50,615
13,187,624,433
IssuesEvent
2020-08-13 04:01:42
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051)
Migrated from Trac combo simulation defect
I don't have astropy. So get this error for the current trunk when cmake configures itself: Traceback (most recent call last): File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module> import astropy.io.fits as pyfits ImportError: No module named astropy.io.fits Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-15T07:42:17", "description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.", "reporter": "hdembinski", "cc": "", "resolution": "fixed", "_ts": "1436946137322111", "component": "combo simulation", "summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present", "priority": "critical", "keywords": "photospline astropy", "time": "2015-07-15T00:12:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051) - I don't have astropy. So get this error for the current trunk when cmake configures itself: Traceback (most recent call last): File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module> import astropy.io.fits as pyfits ImportError: No module named astropy.io.fits Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-15T07:42:17", "description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.", "reporter": "hdembinski", "cc": "", "resolution": "fixed", "_ts": "1436946137322111", "component": "combo simulation", "summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present", "priority": "critical", "keywords": "photospline astropy", "time": "2015-07-15T00:12:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
defect
photospline pybindings fails during cmake s configure step if astropy is not present trac i don t have astropy so get this error for the current trunk when cmake configures itself traceback most recent call last file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in import astropy io fits as pyfits importerror no module named astropy io fits maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user migrated from json status closed changetime description i don t have astropy so get this error for the current trunk when cmake configures itself n ntraceback most recent call last n file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in n import astropy io fits as pyfits nimporterror no module named astropy io fits n nmaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user reporter hdembinski cc resolution fixed ts component combo simulation summary photospline pybindings fails during cmake s configure step if astropy is not present priority critical keywords photospline astropy time milestone owner type defect
1
9,815
8,713,022,856
IssuesEvent
2018-12-07 00:35:33
ssube/isolex
https://api.github.com/repos/ssube/isolex
closed
move service lifecycle to module
service/bot status/progress type/feature
### Summary The `Bot` is currently acting as a service locator and manager for itself and every other service. This should be implemented in a DI module that tracks (starts and stops) services on behalf of the bot. This module will have some helpers to create/get services within the bot's managed scope, much like the prometheus registry for metrics. ### Scope - [x] add service locator module - [x] manage service lifecycle from the bot, through the module - [x] move `createService` and `getService` methods - [x] add kind/name search to `getService` (along with id) ### Use Case The bot is getting too big (#37).
1.0
move service lifecycle to module - ### Summary The `Bot` is currently acting as a service locator and manager for itself and every other service. This should be implemented in a DI module that tracks (starts and stops) services on behalf of the bot. This module will have some helpers to create/get services within the bot's managed scope, much like the prometheus registry for metrics. ### Scope - [x] add service locator module - [x] manage service lifecycle from the bot, through the module - [x] move `createService` and `getService` methods - [x] add kind/name search to `getService` (along with id) ### Use Case The bot is getting too big (#37).
non_defect
move service lifecycle to module summary the bot is currently acting as a service locator and manager for itself and every other service this should be implemented in a di module that tracks starts and stops services on behalf of the bot this module will have some helpers to create get services within the bot s managed scope much like the prometheus registry for metrics scope add service locator module manage service lifecycle from the bot through the module move createservice and getservice methods add kind name search to getservice along with id use case the bot is getting too big
0
154,913
13,595,099,237
IssuesEvent
2020-09-22 02:00:46
UnBArqDsw/2020.1_G2_TCLDL
https://api.github.com/repos/UnBArqDsw/2020.1_G2_TCLDL
closed
User case diagram
documentation dynamic-2
**_Issue_ type** DOC X - _Issue_ title [Documentation] **Description** Make the user case diagram **Screenshots** If you think that it would be necessary, add screenshots here to help you explain your _issue_. **Tasks** - [x] reference the documentation - [x] make all the diagrams cited in the description of the issue - [x] link or reference the user stories - [x] make the artifacts available here tracked **Acceptance Criteria** - [ ] criteria 1 - [ ] ...
1.0
User case diagram - **_Issue_ type** DOC X - _Issue_ title [Documentation] **Description** Make the user case diagram **Screenshots** If you think that it would be necessary, add screenshots here to help you explain your _issue_. **Tasks** - [x] reference the documentation - [x] make all the diagrams cited in the description of the issue - [x] link or reference the user stories - [x] make the artifacts available here tracked **Acceptance Criteria** - [ ] criteria 1 - [ ] ...
non_defect
user case diagram issue type doc x issue title description make the user case diagram screenshots if you think that it would be necessary add screenshots here to help you explain your issue tasks reference the documentation make all the diagrams cited in the description of the issue link or reference the user stories make the artifacts available here tracked acceptance criteria criteria
0
56,231
6,507,198,532
IssuesEvent
2017-08-24 12:20:24
chauncy-crib/tagprobot
https://api.github.com/repos/chauncy-crib/tagprobot
opened
Change `setupTiles` in test/tiles.spec.js to take a team color as a parameter
unit-testing
Right now it takes in a boolean `myColorIsBlue`
1.0
Change `setupTiles` in test/tiles.spec.js to take a team color as a parameter - Right now it takes in a boolean `myColorIsBlue`
non_defect
change setuptiles in test tiles spec js to take a team color as a parameter right now it takes in a boolean mycolorisblue
0
243,537
20,423,758,767
IssuesEvent
2022-02-24 00:06:49
VOREStation/VOREStation
https://api.github.com/repos/VOREStation/VOREStation
closed
Non-emagged holographic carp are digesting people
Sev: 2-High Type: Vore-Related Task: Needs Testing Status: No Response
#### Brief description of the issue After the updates fixing vore bellies and whatnot, the holodeck carp are eating and digesting people (with real, non-halloss damage that persists and everything) even with the safety features on. Their normal attacks still don't do any damage though, so it's just restricted to vore. They DO spit you out before you die, but with how much damage you accrue it seems you usually just die of suffocation. Also you cannot attack them from the inside, which is very unusual. There's also reports of them just leaving the holodeck and running rampant and breaking doors even with the safety features on, but I haven't witnessed it myself and I'm still testing things to see if that happens. If it's true, I'll make a separate issue for it. Update: Haven't seen it happen, must've just been people getting confused by emagged carp. #### What you expected to happen They'd either just endo me or deal halloss digestion damage. #### What actually happened They were immediately in digest mode and dealt very real damage. #### Steps to reproduce - Go to holodeck - Switch it to space - Wait for carp to spawn and smack it - Perish #### Code Revision - release - 98f5770ddf8ce271bf306cd5d85002e20d75f756 - master - 2018-03-25 e904582d94bb07df2bdcd5360e989c3bd843abe7 (Server I did the testing on, but these issues have been observed on the release version as well)
1.0
Non-emagged holographic carp are digesting people - #### Brief description of the issue After the updates fixing vore bellies and whatnot, the holodeck carp are eating and digesting people (with real, non-halloss damage that persists and everything) even with the safety features on. Their normal attacks still don't do any damage though, so it's just restricted to vore. They DO spit you out before you die, but with how much damage you accrue it seems you usually just die of suffocation. Also you cannot attack them from the inside, which is very unusual. There's also reports of them just leaving the holodeck and running rampant and breaking doors even with the safety features on, but I haven't witnessed it myself and I'm still testing things to see if that happens. If it's true, I'll make a separate issue for it. Update: Haven't seen it happen, must've just been people getting confused by emagged carp. #### What you expected to happen They'd either just endo me or deal halloss digestion damage. #### What actually happened They were immediately in digest mode and dealt very real damage. #### Steps to reproduce - Go to holodeck - Switch it to space - Wait for carp to spawn and smack it - Perish #### Code Revision - release - 98f5770ddf8ce271bf306cd5d85002e20d75f756 - master - 2018-03-25 e904582d94bb07df2bdcd5360e989c3bd843abe7 (Server I did the testing on, but these issues have been observed on the release version as well)
non_defect
non emagged holographic carp are digesting people brief description of the issue after the updates fixing vore bellies and whatnot the holodeck carp are eating and digesting people with real non halloss damage that persists and everything even with the safety features on their normal attacks still don t do any damage though so it s just restricted to vore they do spit you out before you die but with how much damage you accrue it seems you usually just die of suffocation also you cannot attack them from the inside which is very unusual there s also reports of them just leaving the holodeck and running rampant and breaking doors even with the safety features on but i haven t witnessed it myself and i m still testing things to see if that happens if it s true i ll make a separate issue for it update haven t seen it happen must ve just been people getting confused by emagged carp what you expected to happen they d either just endo me or deal halloss digestion damage what actually happened they were immediately in digest mode and dealt very real damage steps to reproduce go to holodeck switch it to space wait for carp to spawn and smack it perish code revision release master server i did the testing on but these issues have been observed on the release version as well
0
43,733
11,812,153,155
IssuesEvent
2020-03-19 19:35:12
carbon-design-system/ibm-security
https://api.github.com/repos/carbon-design-system/ibm-security
opened
`LoadingMessage` missing storybook knobs and props documentation
Defect severity 4
## Bug - `LoadingMessage` missing storybook knobs and props documentation **Expected behavior -** The `LoadingMessage` storybook demo should have functional knobs and also props documentation. **Actual behavior -** The `LoadingMessage` storybook demo is missing these features: https://ibm-security.carbondesignsystem.com/?path=/story/components-loading--with-message Only some props are documented in the "Show info" panel. ### Steps for reproducing See: https://ibm-security.carbondesignsystem.com/?path=/story/components-loading--with-message
1.0
`LoadingMessage` missing storybook knobs and props documentation - ## Bug - `LoadingMessage` missing storybook knobs and props documentation **Expected behavior -** The `LoadingMessage` storybook demo should have functional knobs and also props documentation. **Actual behavior -** The `LoadingMessage` storybook demo is missing these features: https://ibm-security.carbondesignsystem.com/?path=/story/components-loading--with-message Only some props are documented in the "Show info" panel. ### Steps for reproducing See: https://ibm-security.carbondesignsystem.com/?path=/story/components-loading--with-message
defect
loadingmessage missing storybook knobs and props documentation bug loadingmessage missing storybook knobs and props documentation expected behavior the loadingmessage storybook demo should have functional knobs and also props documentation actual behavior the loadingmessage storybook demo is missing these features only some props are documented in the show info panel steps for reproducing see
1
306,979
23,177,156,612
IssuesEvent
2022-07-31 15:37:36
PoCInnovation/SpaceVector
https://api.github.com/repos/PoCInnovation/SpaceVector
closed
[Feature]: Make a schema for the application
documentation
### Description You have to build the first step of the application : The vector database. Since the application is complex, you should draw a schema of the application. ### Expected behaviour The team have to: - [x] [Draw](https://excalidraw.com/) a schema - [ ] Put the schema in [README.MD](https://github.com/PoCInnovation/SpaceVector/blob/main/README.md) - [ ] Explain the schema in [README.MD](https://github.com/PoCInnovation/SpaceVector/blob/main/README.md) ### Environment / ### Proposed solution /
1.0
[Feature]: Make a schema for the application - ### Description You have to build the first step of the application : The vector database. Since the application is complex, you should draw a schema of the application. ### Expected behaviour The team have to: - [x] [Draw](https://excalidraw.com/) a schema - [ ] Put the schema in [README.MD](https://github.com/PoCInnovation/SpaceVector/blob/main/README.md) - [ ] Explain the schema in [README.MD](https://github.com/PoCInnovation/SpaceVector/blob/main/README.md) ### Environment / ### Proposed solution /
non_defect
make a schema for the application description you have to build the first step of the application the vector database since the application is complex you should draw a schema of the application expected behaviour the team have to a schema put the schema in explain the schema in environment proposed solution
0
73,206
24,507,492,472
IssuesEvent
2022-10-10 17:47:50
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
CMS: Vet Center Locations pages do not show mobile on node:view.
Defect Drupal engineering ⭐️ Facilities Vet Center
## Describe the defect When I look at node:view on a Vet Center location page, I can see the "Main and satelite locations" but I can not see "Display additional mobile vet centers" ![image](https://user-images.githubusercontent.com/5752113/189230951-39c770ef-7903-4e83-9ac3-2fe16d497d6e.png) Surfaced in [this slack thread](https://dsva.slack.com/archives/C02730UEZPS/p1662667671380849) ## To Reproduce Steps to reproduce the behavior: 1. Go to '/node/28131/edit' 2. Notice that it has an entry for a mobile vet center 3. Click on the "View" tab and notice that the mobile vet center does not show. 4. It does show correctly on https://www.va.gov/dundalk-vet-center/locations and the preview server http://preview-prod.vfs.va.gov/preview?nodeId=28131 ## AC / Expected behavior - [ ] When on node view, the field and any entries for it are shown in a similar fashion as "Main and satelite locations" ## Screenshots ![image](https://user-images.githubusercontent.com/5752113/189231634-50c020eb-850e-415c-8489-20586024a7d6.png) ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
CMS: Vet Center Locations pages do not show mobile on node:view. - ## Describe the defect When I look at node:view on a Vet Center location page, I can see the "Main and satelite locations" but I can not see "Display additional mobile vet centers" ![image](https://user-images.githubusercontent.com/5752113/189230951-39c770ef-7903-4e83-9ac3-2fe16d497d6e.png) Surfaced in [this slack thread](https://dsva.slack.com/archives/C02730UEZPS/p1662667671380849) ## To Reproduce Steps to reproduce the behavior: 1. Go to '/node/28131/edit' 2. Notice that it has an entry for a mobile vet center 3. Click on the "View" tab and notice that the mobile vet center does not show. 4. It does show correctly on https://www.va.gov/dundalk-vet-center/locations and the preview server http://preview-prod.vfs.va.gov/preview?nodeId=28131 ## AC / Expected behavior - [ ] When on node view, the field and any entries for it are shown in a similar fashion as "Main and satelite locations" ## Screenshots ![image](https://user-images.githubusercontent.com/5752113/189231634-50c020eb-850e-415c-8489-20586024a7d6.png) ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
cms vet center locations pages do not show mobile on node view describe the defect when i look at node view on a vet center location page i can see the main and satelite locations but i can not see display additional mobile vet centers surfaced in to reproduce steps to reproduce the behavior go to node edit notice that it has an entry for a mobile vet center click on the view tab and notice that the mobile vet center does not show it does show correctly on and the preview server ac expected behavior when on node view the field and any entries for it are shown in a similar fashion as main and satelite locations screenshots additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
23,989
3,883,369,775
IssuesEvent
2016-04-13 13:40:21
svnpenn/mp4v2
https://api.github.com/repos/svnpenn/mp4v2
closed
Extremely Poor Performace with Native Win32 - Std Fileprovider
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Open file for modification 2. copy samples to single file and close 3. Repeat this 20 to 50 times (or more) I upgraded my libmp4v2 library from a trunk version dated 2008-12-18. The modifications for cross-platform abstraction for file I/O and the switch to Native windows function (CreateFileA ) creates serious file read write performance degradation. Reading about 40 audio files (MP4Modify, Read/write samples, MP4Close ... do again for next file and so on) and combining them in a single m4a file takes 50 secs using the old file routines. The new routines take over 5 minutes and nearly max out /hang the cpu. I reverted all the file routines back to before the R296 change. Performance was restored. What is the expected output? What do you see instead? reasonable read/write performance for multiple open/close of files. What version of the product are you using? On what operating system? Version 1.9, 1.9.1 and trunk R355 all exhibit the problem. Tried on WinXP SP3 and Win7 Please provide any additional information below. ``` Original issue reported on code.google.com by `jeffreyl...@gmail.com` on 11 Jan 2010 at 10:01
1.0
Extremely Poor Performace with Native Win32 - Std Fileprovider - ``` What steps will reproduce the problem? 1. Open file for modification 2. copy samples to single file and close 3. Repeat this 20 to 50 times (or more) I upgraded my libmp4v2 library from a trunk version dated 2008-12-18. The modifications for cross-platform abstraction for file I/O and the switch to Native windows function (CreateFileA ) creates serious file read write performance degradation. Reading about 40 audio files (MP4Modify, Read/write samples, MP4Close ... do again for next file and so on) and combining them in a single m4a file takes 50 secs using the old file routines. The new routines take over 5 minutes and nearly max out /hang the cpu. I reverted all the file routines back to before the R296 change. Performance was restored. What is the expected output? What do you see instead? reasonable read/write performance for multiple open/close of files. What version of the product are you using? On what operating system? Version 1.9, 1.9.1 and trunk R355 all exhibit the problem. Tried on WinXP SP3 and Win7 Please provide any additional information below. ``` Original issue reported on code.google.com by `jeffreyl...@gmail.com` on 11 Jan 2010 at 10:01
defect
extremely poor performace with native std fileprovider what steps will reproduce the problem open file for modification copy samples to single file and close repeat this to times or more i upgraded my library from a trunk version dated the modifications for cross platform abstraction for file i o and the switch to native windows function createfilea creates serious file read write performance degradation reading about audio files read write samples do again for next file and so on and combining them in a single file takes secs using the old file routines the new routines take over minutes and nearly max out hang the cpu i reverted all the file routines back to before the change performance was restored what is the expected output what do you see instead reasonable read write performance for multiple open close of files what version of the product are you using on what operating system version and trunk all exhibit the problem tried on winxp and please provide any additional information below original issue reported on code google com by jeffreyl gmail com on jan at
1
61,291
17,023,659,018
IssuesEvent
2021-07-03 03:09:06
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Should dump.wiki.osm.org work?
Component: wiki Priority: trivial Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 1.11pm, Monday, 6th December 2010]** dump.wiki.osm.org domain returns 403 Forbidden. http://dump.wiki.osm.org/ I'm betting it should go to: :-) http://dump.wiki.openstreetmap.org/
1.0
Should dump.wiki.osm.org work? - **[Submitted to the original trac issue database at 1.11pm, Monday, 6th December 2010]** dump.wiki.osm.org domain returns 403 Forbidden. http://dump.wiki.osm.org/ I'm betting it should go to: :-) http://dump.wiki.openstreetmap.org/
defect
should dump wiki osm org work dump wiki osm org domain returns forbidden i m betting it should go to
1
74,399
25,104,528,814
IssuesEvent
2022-11-08 15:44:54
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
GenericConstantRankTwoTensor.C does not instantiate the templated classes
T: defect P: normal
## Bug Description Those lines missing can create problems when inheriting from GenericConstantRankTwoTensor in an application (seems to be compiler dependent). ## Steps to Reproduce I've attached a Dockerfile that reproduces this issue when installing an application with a subclass of that class. [moose-issue.tar.gz](https://github.com/idaholab/moose/files/9959536/moose-issue.tar.gz) ## Impact I'll survive without but this can be fixed easily
1.0
GenericConstantRankTwoTensor.C does not instantiate the templated classes - ## Bug Description Those lines missing can create problems when inheriting from GenericConstantRankTwoTensor in an application (seems to be compiler dependent). ## Steps to Reproduce I've attached a Dockerfile that reproduces this issue when installing an application with a subclass of that class. [moose-issue.tar.gz](https://github.com/idaholab/moose/files/9959536/moose-issue.tar.gz) ## Impact I'll survive without but this can be fixed easily
defect
genericconstantranktwotensor c does not instantiate the templated classes bug description those lines missing can create problems when inheriting from genericconstantranktwotensor in an application seems to be compiler dependent steps to reproduce i ve attached a dockerfile that reproduces this issue when installing an application with a subclass of that class impact i ll survive without but this can be fixed easily
1
759,683
26,605,910,555
IssuesEvent
2023-01-23 19:19:11
meisnate12/Plex-Meta-Manager
https://api.github.com/repos/meisnate12/Plex-Meta-Manager
closed
Unraid Docker could not download icon pmm.png
bug priority:low
### Version Number 1.17.3 ### What branch are you on? master ### Describe the Bug UnRaid Docker GUI can't download the icon - each time loading the docker page it says _"Plex-Meta-Manager: Could not download icon https://metamanager.wiki/en/latest/_static/pmm.png"._ The icon file appears to have been moved from that page. ### Relevant Collection/Overlay/Playlist Definition _No response_ ### Logs https://gist.github.com/jakabo27/a8a9bfd2e70d01716ab8767450ca35be
1.0
Unraid Docker could not download icon pmm.png - ### Version Number 1.17.3 ### What branch are you on? master ### Describe the Bug UnRaid Docker GUI can't download the icon - each time loading the docker page it says _"Plex-Meta-Manager: Could not download icon https://metamanager.wiki/en/latest/_static/pmm.png"._ The icon file appears to have been moved from that page. ### Relevant Collection/Overlay/Playlist Definition _No response_ ### Logs https://gist.github.com/jakabo27/a8a9bfd2e70d01716ab8767450ca35be
non_defect
unraid docker could not download icon pmm png version number what branch are you on master describe the bug unraid docker gui can t download the icon each time loading the docker page it says plex meta manager could not download icon the icon file appears to have been moved from that page relevant collection overlay playlist definition no response logs
0
322,695
27,625,383,706
IssuesEvent
2023-03-10 06:07:44
italia/design-angular-kit
https://api.github.com/repos/italia/design-angular-kit
closed
Complete Header Component
wontfix docs new component need for tests
## Description Complete and add docs and tests for component `Header` following the vanilla JS implementation present in [Bootstrap Italia documentation](https://italia.github.io/bootstrap-italia/docs/menu-di-navigazione/header/). ## Checklist - [ ] Verify and update, markup and classes of the resulting DOM template, compared to Bootstrap Italia 2 - [ ] Check the rendering of the component (CSS application including spacing, dimensions, typography, ...), compared to Bootstrap Italia and the new UI kit - [ ] Check the behavior of the component (JavaScript, user interaction, states, keyboard interaction for accessibility, ...), compared to Bootstrap Italia 2 - [ ] Verify the accessibility of the component, including automatic tests and manual evaluations by a11y experts, if possible - [ ] Evaluate the need to supplement documentation with more detailed information. - [ ] Write tests for this component - [ ] Write documentation for this component <!-- If you need help: Developers Italia Slack (https://developersitalia.slack.com/messages/C7VPAUVB3)! -->
1.0
Complete Header Component - ## Description Complete and add docs and tests for component `Header` following the vanilla JS implementation present in [Bootstrap Italia documentation](https://italia.github.io/bootstrap-italia/docs/menu-di-navigazione/header/). ## Checklist - [ ] Verify and update, markup and classes of the resulting DOM template, compared to Bootstrap Italia 2 - [ ] Check the rendering of the component (CSS application including spacing, dimensions, typography, ...), compared to Bootstrap Italia and the new UI kit - [ ] Check the behavior of the component (JavaScript, user interaction, states, keyboard interaction for accessibility, ...), compared to Bootstrap Italia 2 - [ ] Verify the accessibility of the component, including automatic tests and manual evaluations by a11y experts, if possible - [ ] Evaluate the need to supplement documentation with more detailed information. - [ ] Write tests for this component - [ ] Write documentation for this component <!-- If you need help: Developers Italia Slack (https://developersitalia.slack.com/messages/C7VPAUVB3)! -->
non_defect
complete header component description complete and add docs and tests for component header following the vanilla js implementation present in checklist verify and update markup and classes of the resulting dom template compared to bootstrap italia check the rendering of the component css application including spacing dimensions typography compared to bootstrap italia and the new ui kit check the behavior of the component javascript user interaction states keyboard interaction for accessibility compared to bootstrap italia verify the accessibility of the component including automatic tests and manual evaluations by experts if possible evaluate the need to supplement documentation with more detailed information write tests for this component write documentation for this component
0
294,733
22,161,653,541
IssuesEvent
2022-06-04 15:38:30
Sheweny/framework
https://api.github.com/repos/Sheweny/framework
opened
Minimal changelog of V4
documentation meta V4 semver major
# Minimal changelog of V4 - ## **For breaking changes, please see #92** ### Client - [ShewenyClientOptions]: Add `disableCooldownForAdmins` option ([3fbd807](https://github.com/Sheweny/framework/commit/3fbd80743d3a96be218af5365ec97aaca5f41c5b)) - [ClientUtil]: `getX` return an array (Array.from(interator)) instedof an iterator. ([0cab451](https://github.com/Sheweny/framework/commit/0cab45103be8f7862fee80d8ebe17824e6fd04d6)) - [ClientUtil]: Add `resolveCommand` function ([88d93cd](https://github.com/Sheweny/framework/commit/88d93cd41125eec1498984372024eda763b42d8a)). ### Managers - #90 - [CommandsManager] : Remove `loadAndRegisterAll()` => now in `loadAll()` ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [EventsManager] : Remove `loadAndRegisterAll()` => now in `loadAll()` ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [Managers]: Remove `loadAll` option ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [Managers]: Removed auto-saving of structures in `client.collection` (it remains automatic through the client options) ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [CommandsManager]: Add `description` to the default object ([7c10846](https://github.com/Sheweny/framework/commit/7c10846b17bf7652986ed21c1770d832adc89b1d)) - Modals support ([d084556](https://github.com/Sheweny/framework/commit/d08455697cac51f32305d0f6ed24577878c9f8a0)) - [ModalsManager]: Create this manager ### Structures - Modals support ([d084556](https://github.com/Sheweny/framework/commit/d08455697cac51f32305d0f6ed24577878c9f8a0)) - [Modal]: Create this structure - [Command]: Add "COMMAND" type in message command ([88d93cd](https://github.com/Sheweny/framework/commit/88d93cd41125eec1498984372024eda763b42d8a)). - [Command]: Add "GLOBAL" type in channel option ([2933afd](https://github.com/Sheweny/framework/commit/2933afd56abb7c0523ba0d4ee2313ce1e2a801db)). - [Structures]: Allow regular expressions and strings in same array ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)) ### Events - [CommandsManager]: emit invalidChannel event when command is in bad channel ([b39ae2c](https://github.com/Sheweny/framework/commit/b39ae2c96cdf0fa751d3c789af382a373ff61766)) ## Cooldowns - [Cooldowns]: Refactor cooldown system (access in client.cooldowns) #87 ([1224753](https://github.com/Sheweny/framework/commit/12247534465494d6afba59b70f41a291d4e26925)) ### Loader V2 - See: #88 - Remove loadFiles function - Remove readDirAndPush function
1.0
Minimal changelog of V4 - # Minimal changelog of V4 - ## **For breaking changes, please see #92** ### Client - [ShewenyClientOptions]: Add `disableCooldownForAdmins` option ([3fbd807](https://github.com/Sheweny/framework/commit/3fbd80743d3a96be218af5365ec97aaca5f41c5b)) - [ClientUtil]: `getX` return an array (Array.from(interator)) instedof an iterator. ([0cab451](https://github.com/Sheweny/framework/commit/0cab45103be8f7862fee80d8ebe17824e6fd04d6)) - [ClientUtil]: Add `resolveCommand` function ([88d93cd](https://github.com/Sheweny/framework/commit/88d93cd41125eec1498984372024eda763b42d8a)). ### Managers - #90 - [CommandsManager] : Remove `loadAndRegisterAll()` => now in `loadAll()` ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [EventsManager] : Remove `loadAndRegisterAll()` => now in `loadAll()` ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [Managers]: Remove `loadAll` option ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [Managers]: Removed auto-saving of structures in `client.collection` (it remains automatic through the client options) ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)). - [CommandsManager]: Add `description` to the default object ([7c10846](https://github.com/Sheweny/framework/commit/7c10846b17bf7652986ed21c1770d832adc89b1d)) - Modals support ([d084556](https://github.com/Sheweny/framework/commit/d08455697cac51f32305d0f6ed24577878c9f8a0)) - [ModalsManager]: Create this manager ### Structures - Modals support ([d084556](https://github.com/Sheweny/framework/commit/d08455697cac51f32305d0f6ed24577878c9f8a0)) - [Modal]: Create this structure - [Command]: Add "COMMAND" type in message command ([88d93cd](https://github.com/Sheweny/framework/commit/88d93cd41125eec1498984372024eda763b42d8a)). - [Command]: Add "GLOBAL" type in channel option ([2933afd](https://github.com/Sheweny/framework/commit/2933afd56abb7c0523ba0d4ee2313ce1e2a801db)). - [Structures]: Allow regular expressions and strings in same array ([c7f41b1](https://github.com/Sheweny/framework/commit/c7f41b156514c6787fab6c2c437c5daa74c7fec7)) ### Events - [CommandsManager]: emit invalidChannel event when command is in bad channel ([b39ae2c](https://github.com/Sheweny/framework/commit/b39ae2c96cdf0fa751d3c789af382a373ff61766)) ## Cooldowns - [Cooldowns]: Refactor cooldown system (access in client.cooldowns) #87 ([1224753](https://github.com/Sheweny/framework/commit/12247534465494d6afba59b70f41a291d4e26925)) ### Loader V2 - See: #88 - Remove loadFiles function - Remove readDirAndPush function
non_defect
minimal changelog of minimal changelog of for breaking changes please see client add disablecooldownforadmins option getx return an array array from interator instedof an iterator add resolvecommand function managers remove loadandregisterall now in loadall remove loadandregisterall now in loadall remove loadall option removed auto saving of structures in client collection it remains automatic through the client options add description to the default object modals support create this manager structures modals support create this structure add command type in message command add global type in channel option allow regular expressions and strings in same array events emit invalidchannel event when command is in bad channel cooldowns refactor cooldown system access in client cooldowns loader see remove loadfiles function remove readdirandpush function
0
651,715
21,485,866,399
IssuesEvent
2022-04-26 23:18:07
canonical-web-and-design/maas-ui
https://api.github.com/repos/canonical-web-and-design/maas-ui
closed
Reuse hostname validation for all node forms
Priority: Low
We should add this to `base/validation.ts` and reuse it for other node types _Originally posted by @Caleb-Ellis in https://github.com/canonical-web-and-design/maas-ui/pull/3805#discussion_r848979094_
1.0
Reuse hostname validation for all node forms - We should add this to `base/validation.ts` and reuse it for other node types _Originally posted by @Caleb-Ellis in https://github.com/canonical-web-and-design/maas-ui/pull/3805#discussion_r848979094_
non_defect
reuse hostname validation for all node forms we should add this to base validation ts and reuse it for other node types originally posted by caleb ellis in
0
171,097
13,218,398,613
IssuesEvent
2020-08-17 08:40:23
kubernetes-sigs/azuredisk-csi-driver
https://api.github.com/repos/kubernetes-sigs/azuredisk-csi-driver
closed
add disk detach e2e test
help wanted kind/test size/S
**Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like in detail** <!-- A clear and concise description of what you want to happen. --> We only have disk attach e2e test, also need to make sure after pod deleted, disk is in unattached state **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
add disk detach e2e test - **Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like in detail** <!-- A clear and concise description of what you want to happen. --> We only have disk attach e2e test, also need to make sure after pod deleted, disk is in unattached state **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
non_defect
add disk detach test is your feature request related to a problem please describe describe the solution you d like in detail we only have disk attach test also need to make sure after pod deleted disk is in unattached state describe alternatives you ve considered additional context
0
19,051
3,130,399,760
IssuesEvent
2015-09-09 09:12:10
Slyum/slyum
https://api.github.com/repos/Slyum/slyum
closed
Rechargement de fichiers
auto-migrated Priority-Medium Type-Defect
``` Lorsque le diagramme devient grand, après l'enregistrement des modifications, si le focus de la fenêtre est perdu, le logiciel demande s'il faut recharger le contenu du fichier qui aurait été modifié extérieurement (ce qui n'est pas le cas). ``` Original issue reported on code.google.com by `davidmis...@gmail.com` on 29 Sep 2014 at 2:26
1.0
Rechargement de fichiers - ``` Lorsque le diagramme devient grand, après l'enregistrement des modifications, si le focus de la fenêtre est perdu, le logiciel demande s'il faut recharger le contenu du fichier qui aurait été modifié extérieurement (ce qui n'est pas le cas). ``` Original issue reported on code.google.com by `davidmis...@gmail.com` on 29 Sep 2014 at 2:26
defect
rechargement de fichiers lorsque le diagramme devient grand après l enregistrement des modifications si le focus de la fenêtre est perdu le logiciel demande s il faut recharger le contenu du fichier qui aurait été modifié extérieurement ce qui n est pas le cas original issue reported on code google com by davidmis gmail com on sep at
1
25,403
11,170,683,969
IssuesEvent
2019-12-28 14:48:24
theWhiteFox/react-tic-tac-toe
https://api.github.com/repos/theWhiteFox/react-tic-tac-toe
opened
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/react-tic-tac-toe/node_modules/sockjs/examples/hapi/html/index.html</p> <p>Path to vulnerable library: /react-tic-tac-toe/node_modules/sockjs/examples/hapi/html/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/express-3.x/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/echo/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/multiplex/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/react-tic-tac-toe/commit/bdca72865016047a3c8d585fce7299b0b9087c7f">bdca72865016047a3c8d585fce7299b0b9087c7f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/react-tic-tac-toe/node_modules/sockjs/examples/hapi/html/index.html</p> <p>Path to vulnerable library: /react-tic-tac-toe/node_modules/sockjs/examples/hapi/html/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/express-3.x/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/echo/index.html,/react-tic-tac-toe/node_modules/sockjs/examples/multiplex/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/react-tic-tac-toe/commit/bdca72865016047a3c8d585fce7299b0b9087c7f">bdca72865016047a3c8d585fce7299b0b9087c7f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm react tic tac toe node modules sockjs examples hapi html index html path to vulnerable library react tic tac toe node modules sockjs examples hapi html index html react tic tac toe node modules sockjs examples express x index html react tic tac toe node modules sockjs examples echo index html react tic tac toe node modules sockjs examples multiplex index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
256,667
22,090,619,809
IssuesEvent
2022-06-01 05:28:37
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
closed
Breadcrumb: Convert Protractor Tests to Puppeteer
type: unit testing for: dev
**Is your feature request related to a problem or use case? Please describe.** Convert breadcrumb component protractor tests to puppeteer. **Describe the solution you'd like** Convert all tests scenarios in protractor to puppeteer. For the naming can be, `{component}.puppeteer-spec.js` **Describe alternatives you've considered** N/A
1.0
Breadcrumb: Convert Protractor Tests to Puppeteer - **Is your feature request related to a problem or use case? Please describe.** Convert breadcrumb component protractor tests to puppeteer. **Describe the solution you'd like** Convert all tests scenarios in protractor to puppeteer. For the naming can be, `{component}.puppeteer-spec.js` **Describe alternatives you've considered** N/A
non_defect
breadcrumb convert protractor tests to puppeteer is your feature request related to a problem or use case please describe convert breadcrumb component protractor tests to puppeteer describe the solution you d like convert all tests scenarios in protractor to puppeteer for the naming can be component puppeteer spec js describe alternatives you ve considered n a
0
16,555
2,917,640,974
IssuesEvent
2015-06-24 00:00:15
dart-lang/logging
https://api.github.com/repos/dart-lang/logging
opened
Add "error" and "info"-method to Logger
Area-Pkg Pkg-Logging Priority-Unassigned Triaged Type-Defect
_From @DartBot on July 16, 2013 13:12_ *This issue was originally filed by off...&#064;mikemitterer.at* _____ **What steps will reproduce the problem?** final Logger logger = new Logger(&quot;rest.DefeaultRESTErrorProcessor&quot;); logger.error(&quot;Does not work&quot;); **Please provide any additional information below.** Most Logging-Frameworks use &quot;error&quot; to print out error-messages. You use &quot;severe&quot;, this is OK but give us &quot;error&quot; to. The same with &quot;info&quot; - I don't know who is using fine, finer, finest - means nothing. fine, finer, finest???? - please provide us with &quot;info&quot; in addition to &quot;fine&quot; _Copied from original issue: dart-lang/sdk#11845_
1.0
Add "error" and "info"-method to Logger - _From @DartBot on July 16, 2013 13:12_ *This issue was originally filed by off...&#064;mikemitterer.at* _____ **What steps will reproduce the problem?** final Logger logger = new Logger(&quot;rest.DefeaultRESTErrorProcessor&quot;); logger.error(&quot;Does not work&quot;); **Please provide any additional information below.** Most Logging-Frameworks use &quot;error&quot; to print out error-messages. You use &quot;severe&quot;, this is OK but give us &quot;error&quot; to. The same with &quot;info&quot; - I don't know who is using fine, finer, finest - means nothing. fine, finer, finest???? - please provide us with &quot;info&quot; in addition to &quot;fine&quot; _Copied from original issue: dart-lang/sdk#11845_
defect
add error and info method to logger from dartbot on july this issue was originally filed by off mikemitterer at what steps will reproduce the problem final logger logger new logger quot rest defeaultresterrorprocessor quot logger error quot does not work quot please provide any additional information below most logging frameworks use quot error quot to print out error messages you use quot severe quot this is ok but give us quot error quot to the same with quot info quot i don t know who is using fine finer finest means nothing fine finer finest please provide us with quot info quot in addition to quot fine quot copied from original issue dart lang sdk
1
64,468
18,684,709,668
IssuesEvent
2021-11-01 10:54:44
obophenotype/cell-ontology
https://api.github.com/repos/obophenotype/cell-ontology
closed
CL:0002339 'prostate stem cell' vs CL:0002341 'basal cell of prostate epithelium'
Priority-Medium Type-Defect auto-migrated autoclosed-unfixed
``` If we define 'basal cell of prostate epithelium' as its name suggests it leads to an equivalence axiom: EquivalentClasses(CL:0002339 'prostate stem cell' CL:0002341 'basal cell of prostate epithelium' ) ``` Original issue reported on code.google.com by `cmung...@gmail.com` on 20 Feb 2013 at 6:23
1.0
CL:0002339 'prostate stem cell' vs CL:0002341 'basal cell of prostate epithelium' - ``` If we define 'basal cell of prostate epithelium' as its name suggests it leads to an equivalence axiom: EquivalentClasses(CL:0002339 'prostate stem cell' CL:0002341 'basal cell of prostate epithelium' ) ``` Original issue reported on code.google.com by `cmung...@gmail.com` on 20 Feb 2013 at 6:23
defect
cl prostate stem cell vs cl basal cell of prostate epithelium if we define basal cell of prostate epithelium as its name suggests it leads to an equivalence axiom equivalentclasses cl prostate stem cell cl basal cell of prostate epithelium original issue reported on code google com by cmung gmail com on feb at
1
311,197
23,375,107,876
IssuesEvent
2022-08-11 01:29:32
missive/emoji-mart
https://api.github.com/repos/missive/emoji-mart
closed
getEmojiDataFromNative equivalent in v5 ?
documentation
Hi there, I'm using this function in v3 to convert native emoji bytes to emoji-mart data. Is there an equivalent in v5 or is there any plan to bring back this feature? Thank you Eric
1.0
getEmojiDataFromNative equivalent in v5 ? - Hi there, I'm using this function in v3 to convert native emoji bytes to emoji-mart data. Is there an equivalent in v5 or is there any plan to bring back this feature? Thank you Eric
non_defect
getemojidatafromnative equivalent in hi there i m using this function in to convert native emoji bytes to emoji mart data is there an equivalent in or is there any plan to bring back this feature thank you eric
0
4,870
2,610,159,072
IssuesEvent
2015-02-26 18:50:29
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Gameplay Error
auto-migrated Priority-Medium Type-Defect
``` Missing Hutt fighter... Code???? ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 3:53
1.0
Gameplay Error - ``` Missing Hutt fighter... Code???? ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 3:53
defect
gameplay error missing hutt fighter code original issue reported on code google com by gmail com on jan at
1
81,700
31,390,914,141
IssuesEvent
2023-08-26 10:25:51
nats-io/nats-server
https://api.github.com/repos/nats-io/nats-server
opened
HTTP headers are not set for the monitoring server when the status is not 200 OK
defect
### What version were you using? nats-server: v2.9.21 ### What environment was the server running in? Linux amd64 ### Is this defect reproducible? All the monitoring server endpoints return `http.StatusOK` except the `/healthz` endpoint when the health check fails. Requests made to the `/healthz` endpoint when the health check fails will receive responses with the default HTTP headers, for example the `Content-Type` header will be set to the value `text/plain; charset=utf-8` instead of `application/json` and `application/javascript` for JSONP. This is happening because setting the headers on an `http.ResponseWriter` has no effect after setting the status code with `ResponseWriter.WriteHeader`. The function [`ResponseHandler()`](https://github.com/mdawar/nats-server/blob/5b18e80d424926eca48091c6eb2e2a56fdffdc5d/server/monitor.go#L2305) sets the headers, but when the health check fails, the function [`HandleHealthz()`](https://github.com/mdawar/nats-server/blob/5b18e80d424926eca48091c6eb2e2a56fdffdc5d/server/monitor.go#L3080) calls `WriteHeader()` before calling `ResponseHandler()` which causes this issue. I will add tests that reproduce this issue in a pull request. ### Given the capability you are leveraging, describe your expectation? The HTTP headers for the `/healthz` endpoint should be set properly when the health check fails. ### Given the expectation, what is the defect you are observing? Setting the headers has no effect after calling `ResponseWriter.WriteHeader()`.
1.0
HTTP headers are not set for the monitoring server when the status is not 200 OK - ### What version were you using? nats-server: v2.9.21 ### What environment was the server running in? Linux amd64 ### Is this defect reproducible? All the monitoring server endpoints return `http.StatusOK` except the `/healthz` endpoint when the health check fails. Requests made to the `/healthz` endpoint when the health check fails will receive responses with the default HTTP headers, for example the `Content-Type` header will be set to the value `text/plain; charset=utf-8` instead of `application/json` and `application/javascript` for JSONP. This is happening because setting the headers on an `http.ResponseWriter` has no effect after setting the status code with `ResponseWriter.WriteHeader`. The function [`ResponseHandler()`](https://github.com/mdawar/nats-server/blob/5b18e80d424926eca48091c6eb2e2a56fdffdc5d/server/monitor.go#L2305) sets the headers, but when the health check fails, the function [`HandleHealthz()`](https://github.com/mdawar/nats-server/blob/5b18e80d424926eca48091c6eb2e2a56fdffdc5d/server/monitor.go#L3080) calls `WriteHeader()` before calling `ResponseHandler()` which causes this issue. I will add tests that reproduce this issue in a pull request. ### Given the capability you are leveraging, describe your expectation? The HTTP headers for the `/healthz` endpoint should be set properly when the health check fails. ### Given the expectation, what is the defect you are observing? Setting the headers has no effect after calling `ResponseWriter.WriteHeader()`.
defect
http headers are not set for the monitoring server when the status is not ok what version were you using nats server what environment was the server running in linux is this defect reproducible all the monitoring server endpoints return http statusok except the healthz endpoint when the health check fails requests made to the healthz endpoint when the health check fails will receive responses with the default http headers for example the content type header will be set to the value text plain charset utf instead of application json and application javascript for jsonp this is happening because setting the headers on an http responsewriter has no effect after setting the status code with responsewriter writeheader the function sets the headers but when the health check fails the function calls writeheader before calling responsehandler which causes this issue i will add tests that reproduce this issue in a pull request given the capability you are leveraging describe your expectation the http headers for the healthz endpoint should be set properly when the health check fails given the expectation what is the defect you are observing setting the headers has no effect after calling responsewriter writeheader
1
404,917
27,499,562,538
IssuesEvent
2023-03-05 14:42:26
felangel/bloc
https://api.github.com/repos/felangel/bloc
opened
docs: fluttercountertutorial wrong example
documentation
**Description** Hi, There is an error on this documentation: https://bloclibrary.dev/#/fluttercountertutorial This code is wrong: `export 'cubit/counter_cubit.dart'; export 'view/view.dart'` and should be replaced by `export 'cubit/counter_cubit.dart'; export 'view/counter_view.dart'`
1.0
docs: fluttercountertutorial wrong example - **Description** Hi, There is an error on this documentation: https://bloclibrary.dev/#/fluttercountertutorial This code is wrong: `export 'cubit/counter_cubit.dart'; export 'view/view.dart'` and should be replaced by `export 'cubit/counter_cubit.dart'; export 'view/counter_view.dart'`
non_defect
docs fluttercountertutorial wrong example description hi there is an error on this documentation this code is wrong export cubit counter cubit dart export view view dart and should be replaced by export cubit counter cubit dart export view counter view dart
0
408,138
27,645,648,820
IssuesEvent
2023-03-10 22:40:06
wagtail/wagtail
https://api.github.com/repos/wagtail/wagtail
closed
Documentation search delivers 404 result
Documentation
The following result delivers a 404 result. The 4th item https://docs.wagtail.org/en/latest/search.html?q=search https://docs.wagtail.org/en/latest/reference/contrib/postgres_search.html?highlight=search
1.0
Documentation search delivers 404 result - The following result delivers a 404 result. The 4th item https://docs.wagtail.org/en/latest/search.html?q=search https://docs.wagtail.org/en/latest/reference/contrib/postgres_search.html?highlight=search
non_defect
documentation search delivers result the following result delivers a result the item
0
345,150
30,792,523,116
IssuesEvent
2023-07-31 17:14:44
B-ki/ft_irc
https://api.github.com/repos/B-ki/ft_irc
closed
Make tests for al lots of different commands
feature test
Make tests for those commands: - JOIN - TOPIC - INVITE - KICK - PART
1.0
Make tests for al lots of different commands - Make tests for those commands: - JOIN - TOPIC - INVITE - KICK - PART
non_defect
make tests for al lots of different commands make tests for those commands join topic invite kick part
0
266,740
28,432,844,084
IssuesEvent
2023-04-15 01:19:20
mgh3326/querydsl
https://api.github.com/repos/mgh3326/querydsl
opened
CVE-2023-20863 (Medium) detected in spring-expression-5.2.3.RELEASE.jar
Mend: dependency security vulnerability
## CVE-2023-20863 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.2.3.RELEASE.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.2.3.RELEASE/d0c6bb10758805b2153c589686b8045554bfac2d/spring-expression-5.2.3.RELEASE.jar,/root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.2.3.RELEASE/d0c6bb10758805b2153c589686b8045554bfac2d/spring-expression-5.2.3.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-actuator-2.2.3.RELEASE.jar (Root Library) - spring-boot-starter-2.2.3.RELEASE.jar - spring-boot-2.2.3.RELEASE.jar - spring-context-5.2.3.RELEASE.jar - :x: **spring-expression-5.2.3.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In spring framework versions prior to 5.2.24 release+ ,5.3.27+ and 6.0.8+ , it is possible for a user to provide a specially crafted SpEL expression that may cause a denial-of-service (DoS) condition. <p>Publish Date: 2023-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20863>CVE-2023-20863</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2023-20863">https://nvd.nist.gov/vuln/detail/CVE-2023-20863</a></p> <p>Release Date: 2023-04-13</p> <p>Fix Resolution (org.springframework:spring-expression): 5.2.24.RELEASE</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-20863 (Medium) detected in spring-expression-5.2.3.RELEASE.jar - ## CVE-2023-20863 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.2.3.RELEASE.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.2.3.RELEASE/d0c6bb10758805b2153c589686b8045554bfac2d/spring-expression-5.2.3.RELEASE.jar,/root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.2.3.RELEASE/d0c6bb10758805b2153c589686b8045554bfac2d/spring-expression-5.2.3.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-actuator-2.2.3.RELEASE.jar (Root Library) - spring-boot-starter-2.2.3.RELEASE.jar - spring-boot-2.2.3.RELEASE.jar - spring-context-5.2.3.RELEASE.jar - :x: **spring-expression-5.2.3.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In spring framework versions prior to 5.2.24 release+ ,5.3.27+ and 6.0.8+ , it is possible for a user to provide a specially crafted SpEL expression that may cause a denial-of-service (DoS) condition. <p>Publish Date: 2023-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20863>CVE-2023-20863</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2023-20863">https://nvd.nist.gov/vuln/detail/CVE-2023-20863</a></p> <p>Release Date: 2023-04-13</p> <p>Fix Resolution (org.springframework:spring-expression): 5.2.24.RELEASE</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-actuator): 2.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in spring expression release jar cve medium severity vulnerability vulnerable library spring expression release jar spring expression language spel library home page a href path to dependency file build gradle path to vulnerable library root gradle caches modules files org springframework spring expression release spring expression release jar root gradle caches modules files org springframework spring expression release spring expression release jar dependency hierarchy spring boot starter actuator release jar root library spring boot starter release jar spring boot release jar spring context release jar x spring expression release jar vulnerable library vulnerability details in spring framework versions prior to release and it is possible for a user to provide a specially crafted spel expression that may cause a denial of service dos condition publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring expression release direct dependency fix resolution org springframework boot spring boot starter actuator step up your open source security game with mend
0
122,656
4,838,665,636
IssuesEvent
2016-11-09 05:09:41
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
opened
[FVT]: diskless provision is in netbooting status after the privision is finished
priority:high type:bug
xCAT 2.12.4 both on rhels7.3 and sles11.4 Steps to reproduce 1. Install xCAT 2.12.4 2. Define CN 3. Do diskless provision on CN 4. After the provision, check CN's status ``` [root@c910f02c08p05 autotest]# lsxcatd -v Version 2.12.4 (git commit 1b5160257787eb027ecc5311ec9a3778f903657b, built Mon Nov 7 06:16:10 EST 2016) [root@c910f02c08p05 autotest]# lsdef -l c910f02c08p07 Object name: c910f02c08p07 arch=ppc64 cons=hmc currchain=boot currstate=netboot rhels7.3-ppc64-compute groups=all hcp=c910hmc01 hidden=0 hwtype=lpar id=7 initrd=xcat/osimage/rhels7.3-ppc64-netboot-compute/initrd-stateless.gz kcmdline=imgurl=http://c910f02c08p05:80//install/netboot/rhels7.3/ppc64/compute/rootimg.cpio.gz XCAT=c910f02c08p05:3001 NODE=c910f02c08p07 FC=0 BOOTIF=4a:c8:fd:4a:c0:03 kernel=xcat/osimage/rhels7.3-ppc64-netboot-compute/kernel mac=4a:c8:fd:4a:c0:03 mgt=hmc monserver=c910f02c08p05 netboot=grub2 nfsserver=c910f02c08p05 nodetype=ppc os=rhels7.3 parent=c910f02fsp08 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles pprofile=c910f02c08p07 profile=compute provmethod=rhels7.3-ppc64-netboot-compute status=netbooting <== here status should be booted statustime=11-08-2016 21:40:35 tftpserver=c910f02c08p05 updatestatus=synced updatestatustime=11-07-2016 14:26:05 xcatmaster=c910f02c08p05 ``` The node is in problem state, please debug on this node. Thanks!
1.0
[FVT]: diskless provision is in netbooting status after the privision is finished - xCAT 2.12.4 both on rhels7.3 and sles11.4 Steps to reproduce 1. Install xCAT 2.12.4 2. Define CN 3. Do diskless provision on CN 4. After the provision, check CN's status ``` [root@c910f02c08p05 autotest]# lsxcatd -v Version 2.12.4 (git commit 1b5160257787eb027ecc5311ec9a3778f903657b, built Mon Nov 7 06:16:10 EST 2016) [root@c910f02c08p05 autotest]# lsdef -l c910f02c08p07 Object name: c910f02c08p07 arch=ppc64 cons=hmc currchain=boot currstate=netboot rhels7.3-ppc64-compute groups=all hcp=c910hmc01 hidden=0 hwtype=lpar id=7 initrd=xcat/osimage/rhels7.3-ppc64-netboot-compute/initrd-stateless.gz kcmdline=imgurl=http://c910f02c08p05:80//install/netboot/rhels7.3/ppc64/compute/rootimg.cpio.gz XCAT=c910f02c08p05:3001 NODE=c910f02c08p07 FC=0 BOOTIF=4a:c8:fd:4a:c0:03 kernel=xcat/osimage/rhels7.3-ppc64-netboot-compute/kernel mac=4a:c8:fd:4a:c0:03 mgt=hmc monserver=c910f02c08p05 netboot=grub2 nfsserver=c910f02c08p05 nodetype=ppc os=rhels7.3 parent=c910f02fsp08 postbootscripts=otherpkgs postscripts=syslog,remoteshell,syncfiles pprofile=c910f02c08p07 profile=compute provmethod=rhels7.3-ppc64-netboot-compute status=netbooting <== here status should be booted statustime=11-08-2016 21:40:35 tftpserver=c910f02c08p05 updatestatus=synced updatestatustime=11-07-2016 14:26:05 xcatmaster=c910f02c08p05 ``` The node is in problem state, please debug on this node. Thanks!
non_defect
diskless provision is in netbooting status after the privision is finished xcat both on and steps to reproduce install xcat define cn do diskless provision on cn after the provision check cn s status lsxcatd v version git commit built mon nov est lsdef l object name arch cons hmc currchain boot currstate netboot compute groups all hcp hidden hwtype lpar id initrd xcat osimage netboot compute initrd stateless gz kcmdline imgurl xcat node fc bootif fd kernel xcat osimage netboot compute kernel mac fd mgt hmc monserver netboot nfsserver nodetype ppc os parent postbootscripts otherpkgs postscripts syslog remoteshell syncfiles pprofile profile compute provmethod netboot compute status netbooting here status should be booted statustime tftpserver updatestatus synced updatestatustime xcatmaster the node is in problem state please debug on this node thanks
0
18,553
3,696,626,510
IssuesEvent
2016-02-27 03:45:10
softlayer/sl-ember-components
https://api.github.com/repos/softlayer/sl-ember-components
closed
Unit | Component | sl pagination: gotoPage() sends the changePage action
ready sl-pagination tests
``` not ok 551 PhantomJS 1.9 - Unit | Component | sl pagination: gotoPage() sends the changePage action --- actual: > null message: > Died on test #1 at test (http://localhost:7357/assets/test-support.js:3025) at testWrapper (http://localhost:7357/assets/test-support.js:6192) at test (http://localhost:7357/assets/test-support.js:6205) at http://localhost:7357/assets/tests.js:25674 at http://localhost:7357/assets/vendor.js:152 at tryFinally (http://localhost:7357/assets/vendor.js:33) at http://localhost:7357/assets/vendor.js:158 at http://localhost:7357/assets/test-loader.js:60 at http://localhost:7357/assets/test-loader.js:51 at http://localhost:7357/assets/test-loader.js:82 at http://localhost:7357/assets/test-support.js:6024: 'undefined' is not a function (evaluating 'pages.find(function (page) { return page.active; })') Log: | ... ```
1.0
Unit | Component | sl pagination: gotoPage() sends the changePage action - ``` not ok 551 PhantomJS 1.9 - Unit | Component | sl pagination: gotoPage() sends the changePage action --- actual: > null message: > Died on test #1 at test (http://localhost:7357/assets/test-support.js:3025) at testWrapper (http://localhost:7357/assets/test-support.js:6192) at test (http://localhost:7357/assets/test-support.js:6205) at http://localhost:7357/assets/tests.js:25674 at http://localhost:7357/assets/vendor.js:152 at tryFinally (http://localhost:7357/assets/vendor.js:33) at http://localhost:7357/assets/vendor.js:158 at http://localhost:7357/assets/test-loader.js:60 at http://localhost:7357/assets/test-loader.js:51 at http://localhost:7357/assets/test-loader.js:82 at http://localhost:7357/assets/test-support.js:6024: 'undefined' is not a function (evaluating 'pages.find(function (page) { return page.active; })') Log: | ... ```
non_defect
unit component sl pagination gotopage sends the changepage action not ok phantomjs unit component sl pagination gotopage sends the changepage action actual null message died on test at test at testwrapper at test at at at tryfinally at at at at at undefined is not a function evaluating pages find function page return page active log
0
495,925
14,290,254,994
IssuesEvent
2020-11-23 20:35:27
LorittaBot/Loritta
https://api.github.com/repos/LorittaBot/Loritta
closed
Fix user retribution on actions being incorrect [5 BRL]
Bounty 🤑 Module: Loritta (Discord) 🎀 Priority: Medium Type: Bug 🐞
![https://cdn.discordapp.com/attachments/664431430159302674/780086898525470760/unknown.png](https://cdn.discordapp.com/attachments/664431430159302674/780086898525470760/unknown.png) When you click on the reaction, Loritta is not showing the correct users. In the second example, it should be "Kiodos" and "I'm Miguel", not "I'm Miguel" and "Kiodos"
1.0
Fix user retribution on actions being incorrect [5 BRL] - ![https://cdn.discordapp.com/attachments/664431430159302674/780086898525470760/unknown.png](https://cdn.discordapp.com/attachments/664431430159302674/780086898525470760/unknown.png) When you click on the reaction, Loritta is not showing the correct users. In the second example, it should be "Kiodos" and "I'm Miguel", not "I'm Miguel" and "Kiodos"
non_defect
fix user retribution on actions being incorrect when you click on the reaction loritta is not showing the correct users in the second example it should be kiodos and i m miguel not i m miguel and kiodos
0
24,889
4,118,947,276
IssuesEvent
2016-06-08 13:27:03
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Extra semicolons added sometimes and taken away
defect
**Note:** These are just the lines with the ; problems not all of them, see [Live Bridge](http://live.bridge.net/#2628d95670c80d4cab285b3ac6b27010). ### Expected ```javascript var c = Bridge.fn.combine(a, b); var d = Bridge.fn.remove(c, a); Bridge.ns("Program", $_); ``` ### Actual ```javascript var c = Bridge.fn.combine(a, b);; var d = Bridge.fn.remove(c, a);; Bridge.ns("Program", $_) ``` ### Steps To Reproduce [Bridge Live](http://live.bridge.net/#2628d95670c80d4cab285b3ac6b27010) ```csharp using System; public static class Program { public static void Main (string[] args) { Action<int> a = (i) => Console.WriteLine("a({0})", i); Action<int> b = (i) => Console.WriteLine("b({0})", i); Action<int> c = (Action<int>)Delegate.Combine(a, b); a(1); b(2); c(3); var d = (Action<int>)Delegate.Remove(c, a); d(4); } public static void PrintNumber (int x) { Console.WriteLine("PrintNumber({0})", x); } } ```
1.0
Extra semicolons added sometimes and taken away - **Note:** These are just the lines with the ; problems not all of them, see [Live Bridge](http://live.bridge.net/#2628d95670c80d4cab285b3ac6b27010). ### Expected ```javascript var c = Bridge.fn.combine(a, b); var d = Bridge.fn.remove(c, a); Bridge.ns("Program", $_); ``` ### Actual ```javascript var c = Bridge.fn.combine(a, b);; var d = Bridge.fn.remove(c, a);; Bridge.ns("Program", $_) ``` ### Steps To Reproduce [Bridge Live](http://live.bridge.net/#2628d95670c80d4cab285b3ac6b27010) ```csharp using System; public static class Program { public static void Main (string[] args) { Action<int> a = (i) => Console.WriteLine("a({0})", i); Action<int> b = (i) => Console.WriteLine("b({0})", i); Action<int> c = (Action<int>)Delegate.Combine(a, b); a(1); b(2); c(3); var d = (Action<int>)Delegate.Remove(c, a); d(4); } public static void PrintNumber (int x) { Console.WriteLine("PrintNumber({0})", x); } } ```
defect
extra semicolons added sometimes and taken away note these are just the lines with the problems not all of them see expected javascript var c bridge fn combine a b var d bridge fn remove c a bridge ns program actual javascript var c bridge fn combine a b var d bridge fn remove c a bridge ns program steps to reproduce csharp using system public static class program public static void main string args action a i console writeline a i action b i console writeline b i action c action delegate combine a b a b c var d action delegate remove c a d public static void printnumber int x console writeline printnumber x
1
488,327
14,076,009,679
IssuesEvent
2020-11-04 09:52:18
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
opened
Exceptions logged when shutting down with downloads running
bug: exception component: downloads priority: 2 - low
When downloading a couple of files and closing the window while the downloads are still running (on ebed435a0c0202e24c31cf0de7c76f4c05a5b607 with Qt 5.15.1), I get: ``` 10:50:25 ERROR: Uncaught exception Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1057, in _on_begin_remove_row self.beginRemoveRows(QModelIndex(), idx, idx) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: Error while recovering pages: wrapped C/C++ object of type TabWidget has been deleted Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 213, in _get_exception_info pages = self._recover_pages(forgiving=True) File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 140, in _recover_pages for tab in tabbed_browser.widgets(): File "/home/florian/proj/qutebrowser/git/qutebrowser/mainwindow/tabbedbrowser.py", line 286, in widgets for i in range(self.widget.count()): RuntimeError: wrapped C/C++ object of type TabWidget has been deleted Error in sys.excepthook: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 288, in exception_hook self._app.lastWindowClosed.disconnect( RuntimeError: wrapped C/C++ object of type Application has been deleted Original exception was: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1057, in _on_begin_remove_row self.beginRemoveRows(QModelIndex(), idx, idx) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: ARGH, there was an exception while the crash dialog is already shown: RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: ARGH, there was an exception while the crash dialog is already shown: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1069, in _on_data_changed start_index = self.index(0, 0) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted ```
1.0
Exceptions logged when shutting down with downloads running - When downloading a couple of files and closing the window while the downloads are still running (on ebed435a0c0202e24c31cf0de7c76f4c05a5b607 with Qt 5.15.1), I get: ``` 10:50:25 ERROR: Uncaught exception Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1057, in _on_begin_remove_row self.beginRemoveRows(QModelIndex(), idx, idx) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: Error while recovering pages: wrapped C/C++ object of type TabWidget has been deleted Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 213, in _get_exception_info pages = self._recover_pages(forgiving=True) File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 140, in _recover_pages for tab in tabbed_browser.widgets(): File "/home/florian/proj/qutebrowser/git/qutebrowser/mainwindow/tabbedbrowser.py", line 286, in widgets for i in range(self.widget.count()): RuntimeError: wrapped C/C++ object of type TabWidget has been deleted Error in sys.excepthook: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/misc/crashsignal.py", line 288, in exception_hook self._app.lastWindowClosed.disconnect( RuntimeError: wrapped C/C++ object of type Application has been deleted Original exception was: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1057, in _on_begin_remove_row self.beginRemoveRows(QModelIndex(), idx, idx) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: ARGH, there was an exception while the crash dialog is already shown: RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted 10:50:25 ERROR: ARGH, there was an exception while the crash dialog is already shown: Traceback (most recent call last): File "/home/florian/proj/qutebrowser/git/qutebrowser/browser/downloads.py", line 1069, in _on_data_changed start_index = self.index(0, 0) RuntimeError: wrapped C/C++ object of type DownloadModel has been deleted ```
non_defect
exceptions logged when shutting down with downloads running when downloading a couple of files and closing the window while the downloads are still running on with qt i get error uncaught exception traceback most recent call last file home florian proj qutebrowser git qutebrowser browser downloads py line in on begin remove row self beginremoverows qmodelindex idx idx runtimeerror wrapped c c object of type downloadmodel has been deleted error error while recovering pages wrapped c c object of type tabwidget has been deleted traceback most recent call last file home florian proj qutebrowser git qutebrowser misc crashsignal py line in get exception info pages self recover pages forgiving true file home florian proj qutebrowser git qutebrowser misc crashsignal py line in recover pages for tab in tabbed browser widgets file home florian proj qutebrowser git qutebrowser mainwindow tabbedbrowser py line in widgets for i in range self widget count runtimeerror wrapped c c object of type tabwidget has been deleted error in sys excepthook traceback most recent call last file home florian proj qutebrowser git qutebrowser misc crashsignal py line in exception hook self app lastwindowclosed disconnect runtimeerror wrapped c c object of type application has been deleted original exception was traceback most recent call last file home florian proj qutebrowser git qutebrowser browser downloads py line in on begin remove row self beginremoverows qmodelindex idx idx runtimeerror wrapped c c object of type downloadmodel has been deleted error argh there was an exception while the crash dialog is already shown runtimeerror wrapped c c object of type downloadmodel has been deleted error argh there was an exception while the crash dialog is already shown traceback most recent call last file home florian proj qutebrowser git qutebrowser browser downloads py line in on data changed start index self index runtimeerror wrapped c c object of type downloadmodel has been deleted
0
5,261
2,610,184,295
IssuesEvent
2015-02-26 18:58:35
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
分享怎么调理色斑
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。怎么调理色斑, 《客户案例》   陆小姐 22岁<br>   自从长了满脸黄褐斑后。很多时候我都在想:如果老天�� �我一个愿望,我决不会考虑太多,就是单纯的希望脸上白白� ��净的。或许很多姐妹都不能想象也不能体会脸上很多黄褐斑 对一个女孩子来说是多大的痛苦。特别是当你很年轻,年轻�� �还没有足够成熟的心态来面对同学、伙伴、长辈或无心或故� ��的嘲笑时,那样的感觉是言语无法形容和表达的。<br>   我的黄褐斑是遗传性的,大概在小学4年纪左右开始蔓延� ��而且一直有越长越多的趋势。大概长斑的人皮肤都很白,所 以即使脸上很多斑,但见到我的人还是会说:你好白啊。有�� �白?大家可以想象一下,一张布满咖啡色小点的白纸和一张静 面的白纸相比,前者肯定会显得暗一些,而我就是前者,我�� �了斑还是比普通人白,晕啊,这样就显得斑点更明显了。<br>   在后来的祛斑路途中,是「黛芙薇尔精华液」把我从多�� �的痛苦中解救出来。当在我为脸上的黄褐斑迷惘的时候,是� ��母为我送来了冬天的一缕阳光,帮我带来了「黛芙薇尔精华 液」。使用一段时间后,我就发现脸上黄褐斑没有了以往的�� �张气焰了,有慢慢消退的迹象,看到一点效果后的我非常高� ��,于是继续每天坚持使用。一个周期下来,我惊奇的发现脸 上的黄褐斑居然消灭了不少,这个效果对我来说真的是太振�� �人心了。高兴之余妈妈又帮我订购了两个周期的产品进行巩� ��治疗。<br>   没想到,仅用了三个周期的时间,就把我的遗传性黄褐�� �彻底的祛除了,世界上再也没有比这更让我高兴的事了,看� ��自己现在洁白无暇的脸蛋,真的好开心,心情也变得开朗了 ,我阴暗的天空终于放晴了! 阅读了怎么调理色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么调理色斑,同时为您分享祛斑小方法 防治雀斑的办法 1、避免阳光和紫外线照射。如果长时间在太阳下活动,应戴� ��阔边帽和墨镜,以避免阳光的直射,并且在脸上涂上防晒霜 可避免皮肤晒黑。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:18
1.0
分享怎么调理色斑 - ``` 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。怎么调理色斑, 《客户案例》   陆小姐 22岁<br>   自从长了满脸黄褐斑后。很多时候我都在想:如果老天�� �我一个愿望,我决不会考虑太多,就是单纯的希望脸上白白� ��净的。或许很多姐妹都不能想象也不能体会脸上很多黄褐斑 对一个女孩子来说是多大的痛苦。特别是当你很年轻,年轻�� �还没有足够成熟的心态来面对同学、伙伴、长辈或无心或故� ��的嘲笑时,那样的感觉是言语无法形容和表达的。<br>   我的黄褐斑是遗传性的,大概在小学4年纪左右开始蔓延� ��而且一直有越长越多的趋势。大概长斑的人皮肤都很白,所 以即使脸上很多斑,但见到我的人还是会说:你好白啊。有�� �白?大家可以想象一下,一张布满咖啡色小点的白纸和一张静 面的白纸相比,前者肯定会显得暗一些,而我就是前者,我�� �了斑还是比普通人白,晕啊,这样就显得斑点更明显了。<br>   在后来的祛斑路途中,是「黛芙薇尔精华液」把我从多�� �的痛苦中解救出来。当在我为脸上的黄褐斑迷惘的时候,是� ��母为我送来了冬天的一缕阳光,帮我带来了「黛芙薇尔精华 液」。使用一段时间后,我就发现脸上黄褐斑没有了以往的�� �张气焰了,有慢慢消退的迹象,看到一点效果后的我非常高� ��,于是继续每天坚持使用。一个周期下来,我惊奇的发现脸 上的黄褐斑居然消灭了不少,这个效果对我来说真的是太振�� �人心了。高兴之余妈妈又帮我订购了两个周期的产品进行巩� ��治疗。<br>   没想到,仅用了三个周期的时间,就把我的遗传性黄褐�� �彻底的祛除了,世界上再也没有比这更让我高兴的事了,看� ��自己现在洁白无暇的脸蛋,真的好开心,心情也变得开朗了 ,我阴暗的天空终于放晴了! 阅读了怎么调理色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么调理色斑,同时为您分享祛斑小方法 防治雀斑的办法 1、避免阳光和紫外线照射。如果长时间在太阳下活动,应戴� ��阔边帽和墨镜,以避免阳光的直射,并且在脸上涂上防晒霜 可避免皮肤晒黑。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:18
defect
分享怎么调理色斑 《摘要》 哭的时候,我会闭上眼睛不让它流泪;孤独寂寞的时候,我�� �静静的想着某人;伤心的时候,我会找个地方静静的发呆,� ��后告诉自己,还是要面对坚持下去;难过的时候,我会伪装 自己,对别人说:我很好、我很开心;失落的时候,我会笑�� �对自己说,没事的,一切总会过去。黄褐斑最主要的就是不� ��等长了黄褐斑才想要祛斑,那是得不偿失的,劳神费才的事 情。同时在去除黄褐斑的时候你要知道面部黄褐斑是怎么来�� �,这才能有效去除黄褐斑。怎么调理色斑, 《客户案例》   陆小姐   自从长了满脸黄褐斑后。很多时候我都在想:如果老天�� �我一个愿望,我决不会考虑太多,就是单纯的希望脸上白白� ��净的。或许很多姐妹都不能想象也不能体会脸上很多黄褐斑 对一个女孩子来说是多大的痛苦。特别是当你很年轻,年轻�� �还没有足够成熟的心态来面对同学、伙伴、长辈或无心或故� ��的嘲笑时,那样的感觉是言语无法形容和表达的。   我的黄褐斑是遗传性的, � ��而且一直有越长越多的趋势。大概长斑的人皮肤都很白,所 以即使脸上很多斑,但见到我的人还是会说:你好白啊。有�� �白 大家可以想象一下,一张布满咖啡色小点的白纸和一张静 面的白纸相比,前者肯定会显得暗一些,而我就是前者,我�� �了斑还是比普通人白,晕啊,这样就显得斑点更明显了。   在后来的祛斑路途中,是「黛芙薇尔精华液」把我从多�� �的痛苦中解救出来。当在我为脸上的黄褐斑迷惘的时候,是� ��母为我送来了冬天的一缕阳光,帮我带来了「黛芙薇尔精华 液」。使用一段时间后,我就发现脸上黄褐斑没有了以往的�� �张气焰了,有慢慢消退的迹象,看到一点效果后的我非常高� ��,于是继续每天坚持使用。一个周期下来,我惊奇的发现脸 上的黄褐斑居然消灭了不少,这个效果对我来说真的是太振�� �人心了。高兴之余妈妈又帮我订购了两个周期的产品进行巩� ��治疗。   没想到,仅用了三个周期的时间,就把我的遗传性黄褐�� �彻底的祛除了,世界上再也没有比这更让我高兴的事了,看� ��自己现在洁白无暇的脸蛋,真的好开心,心情也变得开朗了 ,我阴暗的天空终于放晴了 阅读了怎么调理色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么调理色斑,同时为您分享祛斑小方法 防治雀斑的办法 、避免阳光和紫外线照射。如果长时间在太阳下活动,应戴� ��阔边帽和墨镜,以避免阳光的直射,并且在脸上涂上防晒霜 可避免皮肤晒黑。 original issue reported on code google com by additive gmail com on jul at
1
54,902
14,012,234,051
IssuesEvent
2020-10-29 08:44:25
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
zfs sawtooth performance
Status: Triage Needed Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Ubuntu Linux Distribution Version | 20.04 Linux Kernel | 5.4.0-51-generic Architecture | x86 ZFS Version | zfs-0.8.3-1ubuntu12.4 SPL Version | zfs-0.8.3-1ubuntu12.4 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> System Configuration | Value --- | --- Ram | 128GB CPU | Intel Xeon Silver 4208 CPU @ 2.10 GHz No of Cores | 16 Disk Type | SSD ### Describe the problem you're observing #### ZVOL There is a very heavy fluctuations in IOPS during the elapsed time of 300 seconds ![image](https://user-images.githubusercontent.com/33537695/97541853-da92e100-19eb-11eb-963c-4f3b54c1d9e9.png) <p align="center">IOPS vs Time </p> #### Raw SSD Much less fluctuation in IOPS during the elapsed time of 300 seconds ![image](https://user-images.githubusercontent.com/33537695/97544858-32334b80-19f0-11eb-8941-1be80e3704c3.png) <p align="center">IOPS vs Time </p> ### Describe how to reproduce the problem 1. Zpool was created with one phsical disk (SSD) `zpool create -o ashift=12 tank <device1>` 2. Zvol was created using this command `zfs create -o dedup=on -o compression=lz4 -o sync=always -o volblocksize=128K -s -V 300G tank/vol1` 3. Primarycache was set to metadata on pool `zfs set primarycache=metadata tank` 4. Install vdbench 5. Fill the disk using this synthetic workload ` dedupratio=4 dedupunit=4k compratio=4 messagescan=no sd=default,openflags=directio,align=4k sd=sd2,lun=/dev/zd0 wd=wd_default,sd=* wd=fill,sd=sd*,xfersize=(4k,3,8k,8,16k,11,32k,12,64k,19,128k,21,256k,26),rdpct=0,seekpct=eof rd=default rd=fill,wd=fill,iorate=max,interval=30,elapsed=9000,forthreads=16 ` 6. after first workload is finished , run the second workload ` dedupratio=4 dedupunit=4k compratio=4 sd=default,openflags=directio,align=4k sd=sd1,lun=/dev/zd0 wd=wd_vsi_read,rdpct=100,xfersize=(4k,25.68,8k,26.31,16k,6.4,32k,7.52,60k,10.52,128k,9.82,252k,7.31,504k,6.19,984k,0.23,1032k,0.02),seekpct=80,range=(0,100),sd=sd* wd=wd_vsi_write,rdpct=0,xfersize=(4k,59.62,8k,14.81,16k,8.97,32k,6.13,60k,5.35,100k,1.63,212k,1.39,432k,1.95,920k,0.14,1032k,0.01),seekpct=80,range=(50,100),sd=sd* rd=rd_vsi_ramp,wd=wd_vsi*,forrdpct=57.40,iorate=max,warmup=60,interval=1,elapsed=300,forthreads=8 ` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
zfs sawtooth performance - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Ubuntu Linux Distribution Version | 20.04 Linux Kernel | 5.4.0-51-generic Architecture | x86 ZFS Version | zfs-0.8.3-1ubuntu12.4 SPL Version | zfs-0.8.3-1ubuntu12.4 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> System Configuration | Value --- | --- Ram | 128GB CPU | Intel Xeon Silver 4208 CPU @ 2.10 GHz No of Cores | 16 Disk Type | SSD ### Describe the problem you're observing #### ZVOL There is a very heavy fluctuations in IOPS during the elapsed time of 300 seconds ![image](https://user-images.githubusercontent.com/33537695/97541853-da92e100-19eb-11eb-963c-4f3b54c1d9e9.png) <p align="center">IOPS vs Time </p> #### Raw SSD Much less fluctuation in IOPS during the elapsed time of 300 seconds ![image](https://user-images.githubusercontent.com/33537695/97544858-32334b80-19f0-11eb-8941-1be80e3704c3.png) <p align="center">IOPS vs Time </p> ### Describe how to reproduce the problem 1. Zpool was created with one phsical disk (SSD) `zpool create -o ashift=12 tank <device1>` 2. Zvol was created using this command `zfs create -o dedup=on -o compression=lz4 -o sync=always -o volblocksize=128K -s -V 300G tank/vol1` 3. Primarycache was set to metadata on pool `zfs set primarycache=metadata tank` 4. Install vdbench 5. Fill the disk using this synthetic workload ` dedupratio=4 dedupunit=4k compratio=4 messagescan=no sd=default,openflags=directio,align=4k sd=sd2,lun=/dev/zd0 wd=wd_default,sd=* wd=fill,sd=sd*,xfersize=(4k,3,8k,8,16k,11,32k,12,64k,19,128k,21,256k,26),rdpct=0,seekpct=eof rd=default rd=fill,wd=fill,iorate=max,interval=30,elapsed=9000,forthreads=16 ` 6. after first workload is finished , run the second workload ` dedupratio=4 dedupunit=4k compratio=4 sd=default,openflags=directio,align=4k sd=sd1,lun=/dev/zd0 wd=wd_vsi_read,rdpct=100,xfersize=(4k,25.68,8k,26.31,16k,6.4,32k,7.52,60k,10.52,128k,9.82,252k,7.31,504k,6.19,984k,0.23,1032k,0.02),seekpct=80,range=(0,100),sd=sd* wd=wd_vsi_write,rdpct=0,xfersize=(4k,59.62,8k,14.81,16k,8.97,32k,6.13,60k,5.35,100k,1.63,212k,1.39,432k,1.95,920k,0.14,1032k,0.01),seekpct=80,range=(50,100),sd=sd* rd=rd_vsi_ramp,wd=wd_vsi*,forrdpct=57.40,iorate=max,warmup=60,interval=1,elapsed=300,forthreads=8 ` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
zfs sawtooth performance thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu linux distribution version linux kernel generic architecture zfs version zfs spl version zfs commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version system configuration value ram cpu intel xeon silver cpu ghz no of cores disk type ssd describe the problem you re observing zvol there is a very heavy fluctuations in iops during the elapsed time of seconds iops vs time raw ssd much less fluctuation in iops during the elapsed time of seconds iops vs time describe how to reproduce the problem zpool was created with one phsical disk ssd zpool create o ashift tank zvol was created using this command zfs create o dedup on o compression o sync always o volblocksize s v tank primarycache was set to metadata on pool zfs set primarycache metadata tank install vdbench fill the disk using this synthetic workload dedupratio dedupunit compratio messagescan no sd default openflags directio align sd lun dev wd wd default sd wd fill sd sd xfersize rdpct seekpct eof rd default rd fill wd fill iorate max interval elapsed forthreads after first workload is finished run the second workload dedupratio dedupunit compratio sd default openflags directio align sd lun dev wd wd vsi read rdpct xfersize seekpct range sd sd wd wd vsi write rdpct xfersize seekpct range sd sd rd rd vsi ramp wd wd vsi forrdpct iorate max warmup interval elapsed forthreads include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
79,192
28,035,923,404
IssuesEvent
2023-03-28 15:04:22
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
[Markup and meta data] HTML markup isn't valid. (09.01.1)
508/Accessibility vaos 508-defect-2 collab-cycle-feedback Staging CCIssue09.01 CC-Dashboard appointment-list
### General Information #### VFS team name VA Online Scheduling (VAOS) #### VFS product name VA Online Scheduling #### VFS feature name Appointment List #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details The first column in each row has an `aria-label` providing screen-reader-friendly date-time information. `aria-label` isn't a valid attribute for `div` elements. The actual impact of this is limited, since the most-used screen readers (JAWS, NVDA, VoiceOver) will still be able to parse this as intended. But I think there's a good chance that some screen readers outside the big three will have trouble with it. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A ### Platform Recommendation Using a `<span class="sr-only">` instead of an `aria-label` should accomplish the same goal without running into any invalid HTML. --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
1.0
[Markup and meta data] HTML markup isn't valid. (09.01.1) - ### General Information #### VFS team name VA Online Scheduling (VAOS) #### VFS product name VA Online Scheduling #### VFS feature name Appointment List #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details The first column in each row has an `aria-label` providing screen-reader-friendly date-time information. `aria-label` isn't a valid attribute for `div` elements. The actual impact of this is limited, since the most-used screen readers (JAWS, NVDA, VoiceOver) will still be able to parse this as intended. But I think there's a good chance that some screen readers outside the big three will have trouble with it. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A ### Platform Recommendation Using a `<span class="sr-only">` instead of an `aria-label` should accomplish the same goal without running into any invalid HTML. --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
defect
html markup isn t valid general information vfs team name va online scheduling vaos vfs product name va online scheduling vfs feature name appointment list point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue html markup isn t valid issue details the first column in each row has an aria label providing screen reader friendly date time information aria label isn t a valid attribute for div elements the actual impact of this is limited since the most used screen readers jaws nvda voiceover will still be able to parse this as intended but i think there s a good chance that some screen readers outside the big three will have trouble with it link screenshot or steps to recreate va gov experience standard other references wcag sc a platform recommendation using a instead of an aria label should accomplish the same goal without running into any invalid html vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up
1
172,147
14,350,788,920
IssuesEvent
2020-11-29 22:26:12
guuppy/guuppy-application
https://api.github.com/repos/guuppy/guuppy-application
opened
[TODO] change readme
documentation
## needs - [ ] remove flutter readme template - [ ] append components list - [ ] append page lists (with components) - [ ] introduce guuppy application
1.0
[TODO] change readme - ## needs - [ ] remove flutter readme template - [ ] append components list - [ ] append page lists (with components) - [ ] introduce guuppy application
non_defect
change readme needs remove flutter readme template append components list append page lists with components introduce guuppy application
0
333,788
10,131,317,150
IssuesEvent
2019-08-01 19:13:05
desktop/desktop
https://api.github.com/repos/desktop/desktop
opened
Binary file conflict resolution error message
bug priority-3
When resolving a conflict for a binary file, I noticed two things not consistent with the normal flow of a conflict when dealing with non-binary files. We surface a message stating the conflict must be resolved via the command line, yet when you go to the conflict modal, you can select the file from source A or B and avoid the command line. Recommended text: "This file is in conflict and can be resolved via the command line or by returning to the conflict modal". Are there other binary file conflicts that cannot be resolved in the modal and that this message applies too? ![bug](https://user-images.githubusercontent.com/14828183/62320461-f18b0200-b43b-11e9-8fad-717bfb6b3e2e.png) 2.1.1-beta3, mac
1.0
Binary file conflict resolution error message - When resolving a conflict for a binary file, I noticed two things not consistent with the normal flow of a conflict when dealing with non-binary files. We surface a message stating the conflict must be resolved via the command line, yet when you go to the conflict modal, you can select the file from source A or B and avoid the command line. Recommended text: "This file is in conflict and can be resolved via the command line or by returning to the conflict modal". Are there other binary file conflicts that cannot be resolved in the modal and that this message applies too? ![bug](https://user-images.githubusercontent.com/14828183/62320461-f18b0200-b43b-11e9-8fad-717bfb6b3e2e.png) 2.1.1-beta3, mac
non_defect
binary file conflict resolution error message when resolving a conflict for a binary file i noticed two things not consistent with the normal flow of a conflict when dealing with non binary files we surface a message stating the conflict must be resolved via the command line yet when you go to the conflict modal you can select the file from source a or b and avoid the command line recommended text this file is in conflict and can be resolved via the command line or by returning to the conflict modal are there other binary file conflicts that cannot be resolved in the modal and that this message applies too mac
0
19,430
3,202,749,759
IssuesEvent
2015-10-02 15:29:24
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Near cache on client size for replicated map does not gets invalidated after replicated map changes
Team: Client Type: Defect
This issue was initially described in https://groups.google.com/forum/#!topic/hazelcast/QkSlH2UWhsI I'am trying to add NearCache to ReplicatedMap and configure it to invalidate values on change. ```java ClientConfig clientConfig = new ClientConfig(); NearCacheConfig nnc = new NearCacheConfig(); nnc.setInvalidateOnChange(true); nnc.setInMemoryFormat(InMemoryFormat.OBJECT); Map<String, NearCacheConfig> cachedMaps = new HashMap(); cachedMaps.put("TEST", nnc); clientConfig.setNearCacheConfigMap(cachedMaps); HazelcastInstance client = HazelcastClient.newHazelcastClient(clientConfig); ``` I have a Hazelcast 3.5 cluster with 2 nodes. I create 2 clients using the above code. 1) First client populates ReplicatedMap "TEST" with some values 2) Then second client starts getting one of the values in a loop 3) First client changes the value in ReplicatedMap 4) Using jvisualvm i can see that values in the ReplicatedMap have changed 5) But second client only sees obsolete values
1.0
Near cache on client size for replicated map does not gets invalidated after replicated map changes - This issue was initially described in https://groups.google.com/forum/#!topic/hazelcast/QkSlH2UWhsI I'am trying to add NearCache to ReplicatedMap and configure it to invalidate values on change. ```java ClientConfig clientConfig = new ClientConfig(); NearCacheConfig nnc = new NearCacheConfig(); nnc.setInvalidateOnChange(true); nnc.setInMemoryFormat(InMemoryFormat.OBJECT); Map<String, NearCacheConfig> cachedMaps = new HashMap(); cachedMaps.put("TEST", nnc); clientConfig.setNearCacheConfigMap(cachedMaps); HazelcastInstance client = HazelcastClient.newHazelcastClient(clientConfig); ``` I have a Hazelcast 3.5 cluster with 2 nodes. I create 2 clients using the above code. 1) First client populates ReplicatedMap "TEST" with some values 2) Then second client starts getting one of the values in a loop 3) First client changes the value in ReplicatedMap 4) Using jvisualvm i can see that values in the ReplicatedMap have changed 5) But second client only sees obsolete values
defect
near cache on client size for replicated map does not gets invalidated after replicated map changes this issue was initially described in i am trying to add nearcache to replicatedmap and configure it to invalidate values on change java clientconfig clientconfig new clientconfig nearcacheconfig nnc new nearcacheconfig nnc setinvalidateonchange true nnc setinmemoryformat inmemoryformat object map cachedmaps new hashmap cachedmaps put test nnc clientconfig setnearcacheconfigmap cachedmaps hazelcastinstance client hazelcastclient newhazelcastclient clientconfig i have a hazelcast cluster with nodes i create clients using the above code first client populates replicatedmap test with some values then second client starts getting one of the values in a loop first client changes the value in replicatedmap using jvisualvm i can see that values in the replicatedmap have changed but second client only sees obsolete values
1
160,628
6,101,217,754
IssuesEvent
2017-06-20 14:13:28
kuzzleio/kuzzle
https://api.github.com/repos/kuzzleio/kuzzle
closed
Error on isActionAllowed when user is deleted and token still available
bug priority-blocking
When you delete an user, but the token is still available (so there is no more userId linked to this token) and you try to do actions with this token, the funnel controller triggers an error : ``` error: 2017-04-04T09:32:58+02:00 [LOG:ERROR] Cannot read property 'isActionAllowed' of null 1|kuzzle | TypeError: Cannot read property 'isActionAllowed' of null 1|kuzzle | at kuzzle.repositories.token.verifyToken.then.then.user (/home/elastic/kuzzle/kuzzle/lib/api/controllers/funnelController.js:237:20) ``` Since : ```return kuzzle.repositories.user.load(request.context.token.userId);``` can't load the user
1.0
Error on isActionAllowed when user is deleted and token still available - When you delete an user, but the token is still available (so there is no more userId linked to this token) and you try to do actions with this token, the funnel controller triggers an error : ``` error: 2017-04-04T09:32:58+02:00 [LOG:ERROR] Cannot read property 'isActionAllowed' of null 1|kuzzle | TypeError: Cannot read property 'isActionAllowed' of null 1|kuzzle | at kuzzle.repositories.token.verifyToken.then.then.user (/home/elastic/kuzzle/kuzzle/lib/api/controllers/funnelController.js:237:20) ``` Since : ```return kuzzle.repositories.user.load(request.context.token.userId);``` can't load the user
non_defect
error on isactionallowed when user is deleted and token still available when you delete an user but the token is still available so there is no more userid linked to this token and you try to do actions with this token the funnel controller triggers an error error cannot read property isactionallowed of null kuzzle typeerror cannot read property isactionallowed of null kuzzle at kuzzle repositories token verifytoken then then user home elastic kuzzle kuzzle lib api controllers funnelcontroller js since return kuzzle repositories user load request context token userid can t load the user
0
11,532
9,383,895,596
IssuesEvent
2019-04-05 06:01:40
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
[Service Bus] Compilation fails on addition of parameter to SessionMessageHandlerOptions
Client Service Bus
Steps to reproduce: - Add a new parameter to each MessageHandlerOptions and SessionMessageHandlerOptions Actual: ``` test/receiveAndDeleteMode.spec.ts:209:5 - error TS2349: Cannot invoke an expression whose type lacks a call signature. Type '((onMessage: OnMessage, onError: OnError, options?: MessageHandlerOptions | undefined) => void) | ((onMessage: OnMessage, onError: OnError, options?: SessionMessageHandlerOptions | undefined) => void)' has no compatible call signatures. 209 receiver.receive( ~~~~~~~~~~~~~~~~~ 210 (msg: ServiceBusMessage) => { ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ... 219 { autoComplete: autoCompleteFlag } ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 220 ); ~~~~~ ``` Expected: No test breaks/compilation failures.
1.0
[Service Bus] Compilation fails on addition of parameter to SessionMessageHandlerOptions - Steps to reproduce: - Add a new parameter to each MessageHandlerOptions and SessionMessageHandlerOptions Actual: ``` test/receiveAndDeleteMode.spec.ts:209:5 - error TS2349: Cannot invoke an expression whose type lacks a call signature. Type '((onMessage: OnMessage, onError: OnError, options?: MessageHandlerOptions | undefined) => void) | ((onMessage: OnMessage, onError: OnError, options?: SessionMessageHandlerOptions | undefined) => void)' has no compatible call signatures. 209 receiver.receive( ~~~~~~~~~~~~~~~~~ 210 (msg: ServiceBusMessage) => { ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ... 219 { autoComplete: autoCompleteFlag } ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 220 ); ~~~~~ ``` Expected: No test breaks/compilation failures.
non_defect
compilation fails on addition of parameter to sessionmessagehandleroptions steps to reproduce add a new parameter to each messagehandleroptions and sessionmessagehandleroptions actual test receiveanddeletemode spec ts error cannot invoke an expression whose type lacks a call signature type onmessage onmessage onerror onerror options messagehandleroptions undefined void onmessage onmessage onerror onerror options sessionmessagehandleroptions undefined void has no compatible call signatures receiver receive msg servicebusmessage autocomplete autocompleteflag expected no test breaks compilation failures
0
215,110
16,591,684,994
IssuesEvent
2021-06-01 08:29:53
trilobit-gmbh/contao-databaseformfieldoptions-bundle
https://api.github.com/repos/trilobit-gmbh/contao-databaseformfieldoptions-bundle
closed
Ausgabe eines bestimmten Zeitraums in "WHERE" ?
documentation
Hallo! erstmal "TOP EXTENSION" .. leider schaffe ich es nicht mir die AUsgabe für einen bestimmten Zeitraum nur auszugeben! Ich möchte z.B. nur zukünftige EInträge der nächsten 3 Monate anzeigen lassen, mit `DATE_SUB(CURDATE(), INTERVAL +3 MONTH) AND CURDATE()` klappt es nicht oder auch `startDate BETWEEN DATE_ADD(CURDATE(), INTERVAL +3 MONTH) AND CURDATE()` auch leider nicht! Ich hoffe Ihr könnt helfen! Vielen Dank - bleibt gesund!
1.0
Ausgabe eines bestimmten Zeitraums in "WHERE" ? - Hallo! erstmal "TOP EXTENSION" .. leider schaffe ich es nicht mir die AUsgabe für einen bestimmten Zeitraum nur auszugeben! Ich möchte z.B. nur zukünftige EInträge der nächsten 3 Monate anzeigen lassen, mit `DATE_SUB(CURDATE(), INTERVAL +3 MONTH) AND CURDATE()` klappt es nicht oder auch `startDate BETWEEN DATE_ADD(CURDATE(), INTERVAL +3 MONTH) AND CURDATE()` auch leider nicht! Ich hoffe Ihr könnt helfen! Vielen Dank - bleibt gesund!
non_defect
ausgabe eines bestimmten zeitraums in where hallo erstmal top extension leider schaffe ich es nicht mir die ausgabe für einen bestimmten zeitraum nur auszugeben ich möchte z b nur zukünftige einträge der nächsten monate anzeigen lassen mit date sub curdate interval month and curdate klappt es nicht oder auch startdate between date add curdate interval month and curdate auch leider nicht ich hoffe ihr könnt helfen vielen dank bleibt gesund
0
123,393
12,197,964,555
IssuesEvent
2020-04-29 21:48:41
kwk/test-llvm-bz-import-5
https://api.github.com/repos/kwk/test-llvm-bz-import-5
closed
Incorrect mention of Module method in programmer's manual
BZ-BUG-STATUS: RESOLVED BZ-RESOLUTION: FIXED Documentation/General docs dummy import from bugzilla
This issue was imported from Bugzilla https://bugs.llvm.org/show_bug.cgi?id=13268.
1.0
Incorrect mention of Module method in programmer's manual - This issue was imported from Bugzilla https://bugs.llvm.org/show_bug.cgi?id=13268.
non_defect
incorrect mention of module method in programmer s manual this issue was imported from bugzilla
0
8,512
2,611,515,664
IssuesEvent
2015-02-27 05:51:02
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Connecting to server with alternate nick when your username is already in use hangs frontend.
auto-migrated Priority-Medium Type-Defect
``` Connected to server in 0.9.19-dev when I was already connected as nemo in 0.9.18 Get prompted to enter another nick. I enter nemo_admin which is an alt I use for just such circumstances. Prompt goes away, client stays on "connecting" screen. Hit back button. Client quits. Server: ("CONNECTED", "Hedgewars server http://www.hedgewars.org/", "1") Client: ("NICK", "nemo") Client: ("PROTO", "44") Server: ("NICK", "nemo") Server: ("PROTO", "44") Server: ("NOTICE", "0") Server: ("PING") Client: ("PONG") Server: ("PING") Client: ("PONG") Client: ("NICK", "nemo_admin") Server: ("ERROR", "Nickname already chosen") I suppose it doesn't properly handle the 2nd error. ``` Original issue reported on code.google.com by `kyberneticist@gmail.com` on 1 Mar 2013 at 12:56 * Blocking: #580
1.0
Connecting to server with alternate nick when your username is already in use hangs frontend. - ``` Connected to server in 0.9.19-dev when I was already connected as nemo in 0.9.18 Get prompted to enter another nick. I enter nemo_admin which is an alt I use for just such circumstances. Prompt goes away, client stays on "connecting" screen. Hit back button. Client quits. Server: ("CONNECTED", "Hedgewars server http://www.hedgewars.org/", "1") Client: ("NICK", "nemo") Client: ("PROTO", "44") Server: ("NICK", "nemo") Server: ("PROTO", "44") Server: ("NOTICE", "0") Server: ("PING") Client: ("PONG") Server: ("PING") Client: ("PONG") Client: ("NICK", "nemo_admin") Server: ("ERROR", "Nickname already chosen") I suppose it doesn't properly handle the 2nd error. ``` Original issue reported on code.google.com by `kyberneticist@gmail.com` on 1 Mar 2013 at 12:56 * Blocking: #580
defect
connecting to server with alternate nick when your username is already in use hangs frontend connected to server in dev when i was already connected as nemo in get prompted to enter another nick i enter nemo admin which is an alt i use for just such circumstances prompt goes away client stays on connecting screen hit back button client quits server connected hedgewars server client nick nemo client proto server nick nemo server proto server notice server ping client pong server ping client pong client nick nemo admin server error nickname already chosen i suppose it doesn t properly handle the error original issue reported on code google com by kyberneticist gmail com on mar at blocking
1
58,450
16,541,541,936
IssuesEvent
2021-05-27 17:27:23
networkx/networkx
https://api.github.com/repos/networkx/networkx
closed
Communicability Betweenness Centrality Error For normalized=False
Defect
If you run the method communicability_betweenness_centrality(G, normalized) where normalized is false, you will get the following error: ``` File "../../test/driver_auto_communicability_betweenness_centrality_v01.py", line 186, in main libraryMethodReturnTypeVariable = nx.communicability_betweenness_centrality(Graph,normalized) local variable 'scale' referenced before assignment ``` --------------------------------------------- ``` def _rescale(cbc, normalized): # helper to rescale betweenness centrality if normalized is True: order = len(cbc) if order <= 2: scale = None else: scale = 1.0 / ((order - 1.0) ** 2 - (order - 1.0)) if scale is not None: for v in cbc: cbc[v] *= scale return cbc ``` This is because communicability_betweenness_centrality calls the above method, and as you can see, scale is not defined on the line `if scale is not None`. I'm not sure what scale should be if normalization is not True.
1.0
Communicability Betweenness Centrality Error For normalized=False - If you run the method communicability_betweenness_centrality(G, normalized) where normalized is false, you will get the following error: ``` File "../../test/driver_auto_communicability_betweenness_centrality_v01.py", line 186, in main libraryMethodReturnTypeVariable = nx.communicability_betweenness_centrality(Graph,normalized) local variable 'scale' referenced before assignment ``` --------------------------------------------- ``` def _rescale(cbc, normalized): # helper to rescale betweenness centrality if normalized is True: order = len(cbc) if order <= 2: scale = None else: scale = 1.0 / ((order - 1.0) ** 2 - (order - 1.0)) if scale is not None: for v in cbc: cbc[v] *= scale return cbc ``` This is because communicability_betweenness_centrality calls the above method, and as you can see, scale is not defined on the line `if scale is not None`. I'm not sure what scale should be if normalization is not True.
defect
communicability betweenness centrality error for normalized false if you run the method communicability betweenness centrality g normalized where normalized is false you will get the following error file test driver auto communicability betweenness centrality py line in main librarymethodreturntypevariable nx communicability betweenness centrality graph normalized local variable scale referenced before assignment def rescale cbc normalized helper to rescale betweenness centrality if normalized is true order len cbc if order scale none else scale order order if scale is not none for v in cbc cbc scale return cbc this is because communicability betweenness centrality calls the above method and as you can see scale is not defined on the line if scale is not none i m not sure what scale should be if normalization is not true
1
713,119
24,516,912,150
IssuesEvent
2022-10-11 06:16:42
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
[NSFW] www.xnxx.com - design is broken
nsfw priority-important browser-fenix engine-gecko
<!-- @browser: Firefox Mobile 107.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/112147 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.xnxx.com/video-16tvti84/my_pregnant_neighbor_invites_me_to_her_apartment_but_her_husband_was_about_to_arrive.#show-related **Browser / Version**: Firefox Mobile 107.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Edge **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: /// Is | Mmmmmmmmmmmmmmmmmmmmmm <details> <summary>View the screenshot</summary> Screenshot removed - possible explicit content. </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20221010033207</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/e66b5b35-74fc-4a87-9190-28748874b7aa) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
[NSFW] www.xnxx.com - design is broken - <!-- @browser: Firefox Mobile 107.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/112147 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.xnxx.com/video-16tvti84/my_pregnant_neighbor_invites_me_to_her_apartment_but_her_husband_was_about_to_arrive.#show-related **Browser / Version**: Firefox Mobile 107.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Edge **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: /// Is | Mmmmmmmmmmmmmmmmmmmmmm <details> <summary>View the screenshot</summary> Screenshot removed - possible explicit content. </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20221010033207</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/e66b5b35-74fc-4a87-9190-28748874b7aa) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
design is broken url browser version firefox mobile operating system android tested another browser yes edge problem type design is broken description items are overlapped steps to reproduce is mmmmmmmmmmmmmmmmmmmmmm view the screenshot screenshot removed possible explicit content browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
26,613
4,773,055,350
IssuesEvent
2016-10-26 22:51:44
wheeler-microfluidics/microdrop
https://api.github.com/repos/wheeler-microfluidics/microdrop
opened
No software limits on voltage, current or frequency (Trac #3)
defect Incomplete Migration microdrop Migrated from Trac
Migrated from http://microfluidics.utoronto.ca/ticket/3 ```json { "status": "closed", "changetime": "2014-11-06T21:45:19", "description": "The software should enforce limits on voltage and current (and maybe frequency) to prevent damage to the electronics. In the future, it may make sense warn users about potential damage to their device, though this would require more information about the modes of failure.", "reporter": "ryan", "cc": "", "resolution": "duplicate", "_ts": "1415310319042497", "component": "microdrop", "summary": "No software limits on voltage, current or frequency", "priority": "major", "keywords": "", "version": "0.1", "time": "2011-09-18T05:12:44", "milestone": "Microdrop 1.1", "owner": "ryan", "type": "defect" } ```
1.0
No software limits on voltage, current or frequency (Trac #3) - Migrated from http://microfluidics.utoronto.ca/ticket/3 ```json { "status": "closed", "changetime": "2014-11-06T21:45:19", "description": "The software should enforce limits on voltage and current (and maybe frequency) to prevent damage to the electronics. In the future, it may make sense warn users about potential damage to their device, though this would require more information about the modes of failure.", "reporter": "ryan", "cc": "", "resolution": "duplicate", "_ts": "1415310319042497", "component": "microdrop", "summary": "No software limits on voltage, current or frequency", "priority": "major", "keywords": "", "version": "0.1", "time": "2011-09-18T05:12:44", "milestone": "Microdrop 1.1", "owner": "ryan", "type": "defect" } ```
defect
no software limits on voltage current or frequency trac migrated from json status closed changetime description the software should enforce limits on voltage and current and maybe frequency to prevent damage to the electronics in the future it may make sense warn users about potential damage to their device though this would require more information about the modes of failure reporter ryan cc resolution duplicate ts component microdrop summary no software limits on voltage current or frequency priority major keywords version time milestone microdrop owner ryan type defect
1
69,822
22,686,807,253
IssuesEvent
2022-07-04 14:49:58
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Element Desktop incorrectly thinks I'm not joined in a room
T-Defect S-Major Z-Cache-Confusion O-Occasional
### Steps to reproduce Unclear how to reproduce this. I opened Element Desktop, and the Forge Federation chatroom was in the "Historical" section, was missing its room name, and couldn't be viewed. ### Outcome #### What did you expect? I'm a member of the Forge Federation room, and I can see it and read it in Element Web, so I expected it to be viewable and in the list of current rooms I'm in. #### What happened instead? The room was in the Historical section, and attempting to open it just showed me a pre-join screen. Clicking "Join" just produces a loading spinner that stays there until I switch to a different room. ![20220630_23h08m33s_grim](https://user-images.githubusercontent.com/6477010/176816386-e43186b9-62b1-4008-91fe-ff2b43e173eb.png) ![20220630_23h14m38s_grim](https://user-images.githubusercontent.com/6477010/176816486-65600d14-d4a9-4744-a6b6-c1643ac08ecf.png) ### Operating system NixOS ### Application version Element version: 1.10.15, Olm version: 3.2.8 ### How did you install the app? From nixpkgs. ### Homeserver Synapse 1.61.1 ### Will you send logs? Yes
1.0
Element Desktop incorrectly thinks I'm not joined in a room - ### Steps to reproduce Unclear how to reproduce this. I opened Element Desktop, and the Forge Federation chatroom was in the "Historical" section, was missing its room name, and couldn't be viewed. ### Outcome #### What did you expect? I'm a member of the Forge Federation room, and I can see it and read it in Element Web, so I expected it to be viewable and in the list of current rooms I'm in. #### What happened instead? The room was in the Historical section, and attempting to open it just showed me a pre-join screen. Clicking "Join" just produces a loading spinner that stays there until I switch to a different room. ![20220630_23h08m33s_grim](https://user-images.githubusercontent.com/6477010/176816386-e43186b9-62b1-4008-91fe-ff2b43e173eb.png) ![20220630_23h14m38s_grim](https://user-images.githubusercontent.com/6477010/176816486-65600d14-d4a9-4744-a6b6-c1643ac08ecf.png) ### Operating system NixOS ### Application version Element version: 1.10.15, Olm version: 3.2.8 ### How did you install the app? From nixpkgs. ### Homeserver Synapse 1.61.1 ### Will you send logs? Yes
defect
element desktop incorrectly thinks i m not joined in a room steps to reproduce unclear how to reproduce this i opened element desktop and the forge federation chatroom was in the historical section was missing its room name and couldn t be viewed outcome what did you expect i m a member of the forge federation room and i can see it and read it in element web so i expected it to be viewable and in the list of current rooms i m in what happened instead the room was in the historical section and attempting to open it just showed me a pre join screen clicking join just produces a loading spinner that stays there until i switch to a different room operating system nixos application version element version olm version how did you install the app from nixpkgs homeserver synapse will you send logs yes
1
43,045
9,368,048,017
IssuesEvent
2019-04-03 07:43:09
atomist/sdm-core
https://api.github.com/repos/atomist/sdm-core
closed
Code Inspection: npm audit on artifact-caching
code-inspection
### js-yaml:<3.13.0 - _(warn)_ [Denial of Service](https://npmjs.com/advisories/788) _Upgrade to version 3.13.0._ - `js-yaml:3.12.1`: - `@atomist/automation-client>graphql-code-generator>js-yaml` - `js-yaml:3.12.0`: - `mocha>js-yaml` [atomist:code-inspection:artifact-caching=@atomist/atomist-sdm]
1.0
Code Inspection: npm audit on artifact-caching - ### js-yaml:<3.13.0 - _(warn)_ [Denial of Service](https://npmjs.com/advisories/788) _Upgrade to version 3.13.0._ - `js-yaml:3.12.1`: - `@atomist/automation-client>graphql-code-generator>js-yaml` - `js-yaml:3.12.0`: - `mocha>js-yaml` [atomist:code-inspection:artifact-caching=@atomist/atomist-sdm]
non_defect
code inspection npm audit on artifact caching js yaml warn upgrade to version js yaml atomist automation client graphql code generator js yaml js yaml mocha js yaml
0
127,137
5,019,344,456
IssuesEvent
2016-12-14 11:24:41
hpi-swt2/workshop-portal
https://api.github.com/repos/hpi-swt2/workshop-portal
opened
Profilinfo
High Priority team-hendrik
- [ ] Name in Registration in Vornamem, Nachname aufteilen - [ ] aus Profilinfo entfernen und über Profil ändern - [ ] umbennenn im Dropdown in Einstellungen - [ ] Mail nicht änderbar - [ ] alles Deutsch
1.0
Profilinfo - - [ ] Name in Registration in Vornamem, Nachname aufteilen - [ ] aus Profilinfo entfernen und über Profil ändern - [ ] umbennenn im Dropdown in Einstellungen - [ ] Mail nicht änderbar - [ ] alles Deutsch
non_defect
profilinfo name in registration in vornamem nachname aufteilen aus profilinfo entfernen und über profil ändern umbennenn im dropdown in einstellungen mail nicht änderbar alles deutsch
0
53,541
13,261,852,001
IssuesEvent
2020-08-20 20:38:59
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[filterscripts] coordinate service -> astro (Trac #1641)
Migrated from Trac combo reconstruction defect
This import needs to change: ```text File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter from icecube import dataclasses, coordinate_service ImportError: cannot import name coordinate_service ``` And this line in the file: ```text add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1641">https://code.icecube.wisc.edu/projects/icecube/ticket/1641</a>, reported by david.schultzand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "_ts": "1550067117911749", "description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2016-04-17T17:21:55", "component": "combo reconstruction", "summary": "[filterscripts] coordinate service -> astro", "priority": "blocker", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
[filterscripts] coordinate service -> astro (Trac #1641) - This import needs to change: ```text File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter from icecube import dataclasses, coordinate_service ImportError: cannot import name coordinate_service ``` And this line in the file: ```text add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1641">https://code.icecube.wisc.edu/projects/icecube/ticket/1641</a>, reported by david.schultzand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "_ts": "1550067117911749", "description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2016-04-17T17:21:55", "component": "combo reconstruction", "summary": "[filterscripts] coordinate service -> astro", "priority": "blocker", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
coordinate service astro trac this import needs to change text file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter from icecube import dataclasses coordinate service importerror cannot import name coordinate service and this line in the file text add gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd migrated from json status closed changetime ts description this import needs to change n n file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter n from icecube import dataclasses coordinate service nimporterror cannot import name coordinate service n n nand this line in the file n nadd gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd n reporter david schultz cc resolution fixed time component combo reconstruction summary coordinate service astro priority blocker keywords milestone owner kjmeagher type defect
1
53,894
13,262,457,375
IssuesEvent
2020-08-20 21:50:08
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256)
Migrated from Trac combo core defect
It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None` ```text $ python CoincSuite/resources/test/AfterpulseTest.py NOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int)) NOTICE (TesterModule): Recombination Decisions by AfterpulseTester: YES : 1 NO : 0 UNDECIDED : 0 (TesterModule.cxx:45 in virtual void TesterModule::Finish()) CoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead. self.assert_(self.frame.Has(SplitName+"RecombAttempts")) CoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead. self.assert_(ra[0] == "AfterpulseTester") . ---------------------------------------------------------------------- Ran 1 test in 0.001s OK CoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead. self.assert_(eh.sub_event_stream=="split" and eh.sub_event_id==0) . ---------------------------------------------------------------------- Ran 1 test in 0.000s OK ERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)())) Traceback (most recent call last): File "CoincSuite/resources/test/AfterpulseTest.py", line 194, in <module> tray.Execute() File "/Users/kmeagher/icecube/combo/release/lib/I3Tray.py", line 256, in Execute super(I3Tray, self).Execute() File "/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py", line 83, in RunTests test_result = self.test_runner.run(self.test_suites) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py", line 176, in run test(result) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py", line 84, in __call__ return self.run(*args, **kwds) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py", line 122, in run test(result) TypeError: 'NoneType' object is not callable ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2256">https://code.icecube.wisc.edu/projects/icecube/ticket/2256</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "_ts": "1593001902142004", "description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2019-03-18T19:19:29", "component": "combo core", "summary": "[icetray] I3TestModuleFactory doesn't work in python3", "priority": "blocker", "keywords": "", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256) - It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None` ```text $ python CoincSuite/resources/test/AfterpulseTest.py NOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int)) NOTICE (TesterModule): Recombination Decisions by AfterpulseTester: YES : 1 NO : 0 UNDECIDED : 0 (TesterModule.cxx:45 in virtual void TesterModule::Finish()) CoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead. self.assert_(self.frame.Has(SplitName+"RecombAttempts")) CoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead. self.assert_(ra[0] == "AfterpulseTester") . ---------------------------------------------------------------------- Ran 1 test in 0.001s OK CoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead. self.assert_(eh.sub_event_stream=="split" and eh.sub_event_id==0) . ---------------------------------------------------------------------- Ran 1 test in 0.000s OK ERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)())) Traceback (most recent call last): File "CoincSuite/resources/test/AfterpulseTest.py", line 194, in <module> tray.Execute() File "/Users/kmeagher/icecube/combo/release/lib/I3Tray.py", line 256, in Execute super(I3Tray, self).Execute() File "/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py", line 83, in RunTests test_result = self.test_runner.run(self.test_suites) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py", line 176, in run test(result) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py", line 84, in __call__ return self.run(*args, **kwds) File "/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py", line 122, in run test(result) TypeError: 'NoneType' object is not callable ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2256">https://code.icecube.wisc.edu/projects/icecube/ticket/2256</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "_ts": "1593001902142004", "description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2019-03-18T19:19:29", "component": "combo core", "summary": "[icetray] I3TestModuleFactory doesn't work in python3", "priority": "blocker", "keywords": "", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
doesn t work in trac it looks like the unittest works differently in python and in you can keep running the same unit test over and over again but in after you run it once it gets deleted and becomes a none text python coincsuite resources test afterpulsetest py notice finishing cxx in void execute unsigned int notice testermodule recombination decisions by afterpulsetester yes no undecided testermodule cxx in virtual void testermodule finish coincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead self assert self frame has splitname recombattempts coincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead self assert ra afterpulsetester ran test in ok coincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead self assert eh sub event stream split and eh sub event id ran test in ok error testphysicssequence exception thrown cxx in void do void traceback most recent call last file coincsuite resources test afterpulsetest py line in tray execute file users kmeagher icecube combo release lib py line in execute super self execute file users kmeagher icecube combo release lib icecube icetray test module py line in runtests test result self test runner run self test suites file usr local cellar python frameworks python framework versions lib unittest runner py line in run test result file usr local cellar python frameworks python framework versions lib unittest suite py line in call return self run args kwds file usr local cellar python frameworks python framework versions lib unittest suite py line in run test result typeerror nonetype object is not callable migrated from json status closed changetime ts description it looks like the unittest works differently in python and in you can keep running the same unit test over and over again but in after you run it once it gets deleted and becomes a none n n python coincsuite resources test afterpulsetest py nnotice finishing cxx in void execute unsigned int nnotice testermodule nrecombination decisions by afterpulsetester n yes n no n undecided n testermodule cxx in virtual void testermodule finish ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert self frame has splitname recombattempts ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert ra afterpulsetester n n nran test in n nok ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert eh sub event stream split and eh sub event id n n nran test in n nok nerror testphysicssequence exception thrown cxx in void do void ntraceback most recent call last n file coincsuite resources test afterpulsetest py line in n tray execute n file users kmeagher icecube combo release lib py line in execute n super self execute n file users kmeagher icecube combo release lib icecube icetray test module py line in runtests n test result self test runner run self test suites n file usr local cellar python frameworks python framework versions lib unittest runner py line in run n test result n file usr local cellar python frameworks python framework versions lib unittest suite py line in call n return self run args kwds n file usr local cellar python frameworks python framework versions lib unittest suite py line in run n test result ntypeerror nonetype object is not callable n n n reporter kjmeagher cc resolution fixed time component combo core summary doesn t work in priority blocker keywords milestone autumnal equinox owner olivas type defect
1
81,061
30,690,870,367
IssuesEvent
2023-07-26 15:05:58
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Read marker updates cause notifications
T-Defect X-Needs-Info
### Steps to reproduce 1. Send a message in a room 2. Switch to another room 3. Wait for the other person to read your message from 1. 4. Watch the room from 1. be marked as unread 5. If you have all notifications to noisy etc, receive notifications for it ### Outcome #### What did you expect? No notifications for read markers #### What happened instead? Unread markers everywhere ### Operating system arch ### Browser information Firefox 115.0.2 ### URL for webapp develop.element.io ### Application version Element version: cc8afed1968f-react-c57a4cb09090-js-de7959de6c5c Olm version: 3.2.14 ### Homeserver matrix.org ### Will you send logs? Yes
1.0
Read marker updates cause notifications - ### Steps to reproduce 1. Send a message in a room 2. Switch to another room 3. Wait for the other person to read your message from 1. 4. Watch the room from 1. be marked as unread 5. If you have all notifications to noisy etc, receive notifications for it ### Outcome #### What did you expect? No notifications for read markers #### What happened instead? Unread markers everywhere ### Operating system arch ### Browser information Firefox 115.0.2 ### URL for webapp develop.element.io ### Application version Element version: cc8afed1968f-react-c57a4cb09090-js-de7959de6c5c Olm version: 3.2.14 ### Homeserver matrix.org ### Will you send logs? Yes
defect
read marker updates cause notifications steps to reproduce send a message in a room switch to another room wait for the other person to read your message from watch the room from be marked as unread if you have all notifications to noisy etc receive notifications for it outcome what did you expect no notifications for read markers what happened instead unread markers everywhere operating system arch browser information firefox url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs yes
1
7,395
9,647,271,837
IssuesEvent
2019-05-17 13:34:12
Yoast/wordpress-seo
https://api.github.com/repos/Yoast/wordpress-seo
closed
Storefront Pricing Tables conflict
WooCommerce backlog bug compatibility
<!-- Please use this template when creating an issue. - Please check the boxes after you've created your issue. - Please use the latest version of Yoast SEO.--> * [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md). * [x] I've searched for any related issues and avoided creating a duplicate issue. ### Please give us a description of what happened. Published post with Woocommerce Storefront Pricing Table shortcode in it. Error message appeared on an otherwise blank page: Fatal error: Uncaught Error: Call to undefined function wc_print_notices() in /home/alexg392/staging/1/wp-content/plugins/woocommerce/includes/class-wc-shortcodes.php:578 Stack trace: #0 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(319): WC_Shortcodes::shop_messages('', '', 'woocommerce_mes...') #1 [internal function]: do_shortcode_tag(Array) #2 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(197): preg_replace_callback('/\\[(\\[?)(woocom...', 'do_shortcode_ta...', '[woocommerce_me...') #3 /home/xxxxxx/staging/1/wp-content/plugins/storefront-pricing-tables/storefront-pricing-tables.php(451): do_shortcode('[woocommerce_me...') #4 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(319): Storefront_Pricing_Tables->spt_pricing_table(Array, '</p>\n<p>[pricin...', 'pricing_table') #5 [internal function]: do_shortcode_tag(Array) #6 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(197): preg_replace_callback('/\\[(\\[?)(pricin...', 'do_shortcode_ta...', '<p>See at a gla...') #7 /home/xxxxxx/staging/1/wp-include in /home/xxxxxx/staging/1/wp-content/plugins/woocommerce/includes/class-wc-shortcodes.php on line 578 However, the post is published correctly, and the Pricing Table does display. So I tested for plugin conflict, Yoast SEO was in conflict with this plugin. Woocommerce support say: 'Well it looks like something is triggering the shortcode to be parsed, which is generally not recommended to do when saving the post. Which causes a function to be called which is normally only available on the frontend. So it looks like Yoast is initially triggering this, and then the Pricing Tables plugin is the one that is calling the function, because it expects to only render the shortcode on the frontend.' ### Please describe what you expected to happen and why. The post to be published without an error message. ### How can we reproduce this behavior? 1. Install Woocommerce + Storefront theme + Storefront Pricing Tables plugin + Yoast SEO 2. Add some Pricing Tables to a post. 3. Publish post. ### Technical info * WordPress version: 4.9.1 * Yoast SEO version: 6.0 * Relevant plugins in case of a bug: Woocommerce 3.2.6 / Storefront Pricing Tables 1.0.3 / Theme: Storefront (+ Child theme 'Bookshop')
True
Storefront Pricing Tables conflict - <!-- Please use this template when creating an issue. - Please check the boxes after you've created your issue. - Please use the latest version of Yoast SEO.--> * [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md). * [x] I've searched for any related issues and avoided creating a duplicate issue. ### Please give us a description of what happened. Published post with Woocommerce Storefront Pricing Table shortcode in it. Error message appeared on an otherwise blank page: Fatal error: Uncaught Error: Call to undefined function wc_print_notices() in /home/alexg392/staging/1/wp-content/plugins/woocommerce/includes/class-wc-shortcodes.php:578 Stack trace: #0 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(319): WC_Shortcodes::shop_messages('', '', 'woocommerce_mes...') #1 [internal function]: do_shortcode_tag(Array) #2 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(197): preg_replace_callback('/\\[(\\[?)(woocom...', 'do_shortcode_ta...', '[woocommerce_me...') #3 /home/xxxxxx/staging/1/wp-content/plugins/storefront-pricing-tables/storefront-pricing-tables.php(451): do_shortcode('[woocommerce_me...') #4 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(319): Storefront_Pricing_Tables->spt_pricing_table(Array, '</p>\n<p>[pricin...', 'pricing_table') #5 [internal function]: do_shortcode_tag(Array) #6 /home/xxxxxx/staging/1/wp-includes/shortcodes.php(197): preg_replace_callback('/\\[(\\[?)(pricin...', 'do_shortcode_ta...', '<p>See at a gla...') #7 /home/xxxxxx/staging/1/wp-include in /home/xxxxxx/staging/1/wp-content/plugins/woocommerce/includes/class-wc-shortcodes.php on line 578 However, the post is published correctly, and the Pricing Table does display. So I tested for plugin conflict, Yoast SEO was in conflict with this plugin. Woocommerce support say: 'Well it looks like something is triggering the shortcode to be parsed, which is generally not recommended to do when saving the post. Which causes a function to be called which is normally only available on the frontend. So it looks like Yoast is initially triggering this, and then the Pricing Tables plugin is the one that is calling the function, because it expects to only render the shortcode on the frontend.' ### Please describe what you expected to happen and why. The post to be published without an error message. ### How can we reproduce this behavior? 1. Install Woocommerce + Storefront theme + Storefront Pricing Tables plugin + Yoast SEO 2. Add some Pricing Tables to a post. 3. Publish post. ### Technical info * WordPress version: 4.9.1 * Yoast SEO version: 6.0 * Relevant plugins in case of a bug: Woocommerce 3.2.6 / Storefront Pricing Tables 1.0.3 / Theme: Storefront (+ Child theme 'Bookshop')
non_defect
storefront pricing tables conflict please use this template when creating an issue please check the boxes after you ve created your issue please use the latest version of yoast seo i ve read and understood the i ve searched for any related issues and avoided creating a duplicate issue please give us a description of what happened published post with woocommerce storefront pricing table shortcode in it error message appeared on an otherwise blank page fatal error uncaught error call to undefined function wc print notices in home staging wp content plugins woocommerce includes class wc shortcodes php stack trace home xxxxxx staging wp includes shortcodes php wc shortcodes shop messages woocommerce mes do shortcode tag array home xxxxxx staging wp includes shortcodes php preg replace callback do shortcode tag array home xxxxxx staging wp includes shortcodes php preg replace callback pricin do shortcode ta see at a gla home xxxxxx staging wp include in home xxxxxx staging wp content plugins woocommerce includes class wc shortcodes php on line however the post is published correctly and the pricing table does display so i tested for plugin conflict yoast seo was in conflict with this plugin woocommerce support say well it looks like something is triggering the shortcode to be parsed which is generally not recommended to do when saving the post which causes a function to be called which is normally only available on the frontend so it looks like yoast is initially triggering this and then the pricing tables plugin is the one that is calling the function because it expects to only render the shortcode on the frontend please describe what you expected to happen and why the post to be published without an error message how can we reproduce this behavior install woocommerce storefront theme storefront pricing tables plugin yoast seo add some pricing tables to a post publish post technical info wordpress version yoast seo version relevant plugins in case of a bug woocommerce storefront pricing tables theme storefront child theme bookshop
0
263,434
8,289,507,699
IssuesEvent
2018-09-19 14:51:18
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
opened
Loading of checkpoints through yaml can fail for cyclic dependencies
priority/important topic/Workflows type/bug
For the creation of checkpoints of `Process` instances in the engine, `yaml` is used to dump the object. The object can then be recreated in memory, in principle, through `yaml.load` of the dump. However, for certain instances of `WorkChain` this fails. This can be reproduced with this MWE: ``` import yaml from aiida.common.extendeddicts import AttributeDict from aiida.orm import construct_backend backend = construct_backend() d = AttributeDict() d['computer'] = computer bundle = yaml.dump(d) loaded = yaml.load(bundle) ``` This will trigger an exception at the last line: ``` ConstructorError: found unconstructable recursive node in "<string>", line 4, column 15: _backend: &id001 !!python/object:aiida.orm ... ``` The bundle that was generated by `yaml.dump` looks as follows: ``` !!python/object/new:aiida.common.extendeddicts.AttributeDict dictitems: computer: !!python/object:aiida.orm.implementation.django.computer.DjangoComputer _backend: &id001 !!python/object:aiida.orm.implementation.django.backend.DjangoBackend _authinfos: !!python/object:aiida.orm.implementation.django.authinfo.DjangoAuthInfoCollection _backend: *id001 _computers: !!python/object:aiida.orm.implementation.django.computer.DjangoComputerCollection _backend: *id001 _logs: !!python/object:aiida.orm.implementation.django.log.DjangoLogCollection _backend: *id001 _query_manager: !!python/object:aiida.backends.djsite.queries.DjangoQueryManager _backend: *id001 _users: !!python/object:aiida.orm.implementation.django.user.DjangoUserCollection _backend: *id001 _dbcomputer: !!python/object:aiida.orm.implementation.django.utils.ModelWrapper _model: !!python/object/apply:django.db.models.base.model_unpickle args: - !!python/tuple [db, DbComputer] - [] - !!python/name:django.db.models.base.simple_class_factory '' state: _django_version: 1.8.19 _state: !!python/object:django.db.models.base.ModelState {adding: false, db: default} description: !!python/unicode 'localhost' enabled: true hostname: !!python/unicode 'localhost' id: 1 metadata: !!python/unicode '{"workdir": "/scratch/", "prepend_text": "", "mpirun_command": [], "shebang": "#!/bin/bash", "append_text": "", "default_mpiprocs_per_machine": 1, "test": 1}' name: !!python/unicode 'localhost' scheduler_type: !!python/unicode 'direct' transport_params: !!python/unicode '{}' transport_type: !!python/unicode 'local' uuid: !!python/unicode '555f3245-4f75-4afb-a7a0-009aee5e8dc6' ``` The problem has to do with the cyclic dependency of the `Computer` instance on the `Backend` instance, which in turn has various collections, that each also have that very backend instance as a member. In principle this recursiveness is supported by `yaml`, partly made possible by the use of these `&id001` and `*id001` references. Interestingly, even though the problem seems to originate in the reconstruction of the `Computer` object, the problem disappears if the wrapping `AttributeDict` is removed. The following, only slightly differing MWE, works just fine: ``` import yaml from aiida.orm import construct_backend backend = construct_backend() computer = backend.computers.get(name='localhost') bundle = yaml.dump(computer) loaded = yaml.load(bundle) ``` After further investigation, it seems that it actually has to do with the `AttributeDict`, because replacing it with a normal dictionary in the first MWE, makes the problem disappear as well.
1.0
Loading of checkpoints through yaml can fail for cyclic dependencies - For the creation of checkpoints of `Process` instances in the engine, `yaml` is used to dump the object. The object can then be recreated in memory, in principle, through `yaml.load` of the dump. However, for certain instances of `WorkChain` this fails. This can be reproduced with this MWE: ``` import yaml from aiida.common.extendeddicts import AttributeDict from aiida.orm import construct_backend backend = construct_backend() d = AttributeDict() d['computer'] = computer bundle = yaml.dump(d) loaded = yaml.load(bundle) ``` This will trigger an exception at the last line: ``` ConstructorError: found unconstructable recursive node in "<string>", line 4, column 15: _backend: &id001 !!python/object:aiida.orm ... ``` The bundle that was generated by `yaml.dump` looks as follows: ``` !!python/object/new:aiida.common.extendeddicts.AttributeDict dictitems: computer: !!python/object:aiida.orm.implementation.django.computer.DjangoComputer _backend: &id001 !!python/object:aiida.orm.implementation.django.backend.DjangoBackend _authinfos: !!python/object:aiida.orm.implementation.django.authinfo.DjangoAuthInfoCollection _backend: *id001 _computers: !!python/object:aiida.orm.implementation.django.computer.DjangoComputerCollection _backend: *id001 _logs: !!python/object:aiida.orm.implementation.django.log.DjangoLogCollection _backend: *id001 _query_manager: !!python/object:aiida.backends.djsite.queries.DjangoQueryManager _backend: *id001 _users: !!python/object:aiida.orm.implementation.django.user.DjangoUserCollection _backend: *id001 _dbcomputer: !!python/object:aiida.orm.implementation.django.utils.ModelWrapper _model: !!python/object/apply:django.db.models.base.model_unpickle args: - !!python/tuple [db, DbComputer] - [] - !!python/name:django.db.models.base.simple_class_factory '' state: _django_version: 1.8.19 _state: !!python/object:django.db.models.base.ModelState {adding: false, db: default} description: !!python/unicode 'localhost' enabled: true hostname: !!python/unicode 'localhost' id: 1 metadata: !!python/unicode '{"workdir": "/scratch/", "prepend_text": "", "mpirun_command": [], "shebang": "#!/bin/bash", "append_text": "", "default_mpiprocs_per_machine": 1, "test": 1}' name: !!python/unicode 'localhost' scheduler_type: !!python/unicode 'direct' transport_params: !!python/unicode '{}' transport_type: !!python/unicode 'local' uuid: !!python/unicode '555f3245-4f75-4afb-a7a0-009aee5e8dc6' ``` The problem has to do with the cyclic dependency of the `Computer` instance on the `Backend` instance, which in turn has various collections, that each also have that very backend instance as a member. In principle this recursiveness is supported by `yaml`, partly made possible by the use of these `&id001` and `*id001` references. Interestingly, even though the problem seems to originate in the reconstruction of the `Computer` object, the problem disappears if the wrapping `AttributeDict` is removed. The following, only slightly differing MWE, works just fine: ``` import yaml from aiida.orm import construct_backend backend = construct_backend() computer = backend.computers.get(name='localhost') bundle = yaml.dump(computer) loaded = yaml.load(bundle) ``` After further investigation, it seems that it actually has to do with the `AttributeDict`, because replacing it with a normal dictionary in the first MWE, makes the problem disappear as well.
non_defect
loading of checkpoints through yaml can fail for cyclic dependencies for the creation of checkpoints of process instances in the engine yaml is used to dump the object the object can then be recreated in memory in principle through yaml load of the dump however for certain instances of workchain this fails this can be reproduced with this mwe import yaml from aiida common extendeddicts import attributedict from aiida orm import construct backend backend construct backend d attributedict d computer bundle yaml dump d loaded yaml load bundle this will trigger an exception at the last line constructorerror found unconstructable recursive node in line column backend python object aiida orm the bundle that was generated by yaml dump looks as follows python object new aiida common extendeddicts attributedict dictitems computer python object aiida orm implementation django computer djangocomputer backend python object aiida orm implementation django backend djangobackend authinfos python object aiida orm implementation django authinfo djangoauthinfocollection backend computers python object aiida orm implementation django computer djangocomputercollection backend logs python object aiida orm implementation django log djangologcollection backend query manager python object aiida backends djsite queries djangoquerymanager backend users python object aiida orm implementation django user djangousercollection backend dbcomputer python object aiida orm implementation django utils modelwrapper model python object apply django db models base model unpickle args python tuple python name django db models base simple class factory state django version state python object django db models base modelstate adding false db default description python unicode localhost enabled true hostname python unicode localhost id metadata python unicode workdir scratch prepend text mpirun command shebang bin bash append text default mpiprocs per machine test name python unicode localhost scheduler type python unicode direct transport params python unicode transport type python unicode local uuid python unicode the problem has to do with the cyclic dependency of the computer instance on the backend instance which in turn has various collections that each also have that very backend instance as a member in principle this recursiveness is supported by yaml partly made possible by the use of these and references interestingly even though the problem seems to originate in the reconstruction of the computer object the problem disappears if the wrapping attributedict is removed the following only slightly differing mwe works just fine import yaml from aiida orm import construct backend backend construct backend computer backend computers get name localhost bundle yaml dump computer loaded yaml load bundle after further investigation it seems that it actually has to do with the attributedict because replacing it with a normal dictionary in the first mwe makes the problem disappear as well
0
334,504
29,871,036,801
IssuesEvent
2023-06-20 08:30:50
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix nn.test_tensorflow_embedding_lookup
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |jax|<a href="null"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix nn.test_tensorflow_embedding_lookup - | | | |---|---| |jax|<a href="null"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5281339622/jobs/9554772858"><img src=https://img.shields.io/badge/-success-success></a>
non_defect
fix nn test tensorflow embedding lookup jax img src numpy img src tensorflow a href src torch a href src paddle a href src
0
73,956
24,886,793,173
IssuesEvent
2022-10-28 08:29:08
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]:Selenium grid 4 console is in loading state intermittently
C-grid R-awaiting answer C-java I-defect
### What happened? Have migrated to selenium grid 4 (4.1.3) and working fine. But sometimes when open the console, cannot see nodes, rather seeing " loading " message on the screen. Attached screenshot. Whenever we see this issue, have to go to hub and restart the task for this environment and it works. But would like to get this issue fixed. <img width="960" alt="selenium4 console error" src="https://user-images.githubusercontent.com/16626518/187320123-b9442441-4294-473b-9c8a-daaf5aa39d33.PNG"> ### How can we reproduce the issue? ```shell set up selenium nodes for selenium grid 4 . Intermittently get the loading page. port used to set up is 5955. ``` ### Relevant log output ```shell Getting "loading " message in console ``` ### Operating System windows10 ### Selenium version java 8 ### What are the browser(s) and version(s) where you see this issue? chrome 104, edge 104 ### What are the browser driver(s) and version(s) where you see this issue? chrome driver 104. ### Are you using Selenium Grid? 4.1.3
1.0
[🐛 Bug]:Selenium grid 4 console is in loading state intermittently - ### What happened? Have migrated to selenium grid 4 (4.1.3) and working fine. But sometimes when open the console, cannot see nodes, rather seeing " loading " message on the screen. Attached screenshot. Whenever we see this issue, have to go to hub and restart the task for this environment and it works. But would like to get this issue fixed. <img width="960" alt="selenium4 console error" src="https://user-images.githubusercontent.com/16626518/187320123-b9442441-4294-473b-9c8a-daaf5aa39d33.PNG"> ### How can we reproduce the issue? ```shell set up selenium nodes for selenium grid 4 . Intermittently get the loading page. port used to set up is 5955. ``` ### Relevant log output ```shell Getting "loading " message in console ``` ### Operating System windows10 ### Selenium version java 8 ### What are the browser(s) and version(s) where you see this issue? chrome 104, edge 104 ### What are the browser driver(s) and version(s) where you see this issue? chrome driver 104. ### Are you using Selenium Grid? 4.1.3
defect
selenium grid console is in loading state intermittently what happened have migrated to selenium grid and working fine but sometimes when open the console cannot see nodes rather seeing loading message on the screen attached screenshot whenever we see this issue have to go to hub and restart the task for this environment and it works but would like to get this issue fixed img width alt console error src how can we reproduce the issue shell set up selenium nodes for selenium grid intermittently get the loading page port used to set up is relevant log output shell getting loading message in console operating system selenium version java what are the browser s and version s where you see this issue chrome edge what are the browser driver s and version s where you see this issue chrome driver are you using selenium grid
1
582,583
17,365,126,106
IssuesEvent
2021-07-30 05:53:36
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Update GenerateMavenDependenciesList.kt to call out first co-ordinate name that should be updated to update all occurences of the license
Priority: Essential Type: Task Where: Settings
Currently, the `GenerateMavenDependenciesList.kt` only prints the licenses whose links have not been verified manually. It should also point to a particular coordinate that should be updated in order to update all occurrences of those licenses. Reference Comment - https://github.com/oppia/oppia-android/pull/3434#discussion_r670979579
1.0
Update GenerateMavenDependenciesList.kt to call out first co-ordinate name that should be updated to update all occurences of the license - Currently, the `GenerateMavenDependenciesList.kt` only prints the licenses whose links have not been verified manually. It should also point to a particular coordinate that should be updated in order to update all occurrences of those licenses. Reference Comment - https://github.com/oppia/oppia-android/pull/3434#discussion_r670979579
non_defect
update generatemavendependencieslist kt to call out first co ordinate name that should be updated to update all occurences of the license currently the generatemavendependencieslist kt only prints the licenses whose links have not been verified manually it should also point to a particular coordinate that should be updated in order to update all occurrences of those licenses reference comment
0
25,818
4,466,747,008
IssuesEvent
2016-08-25 00:16:30
deriamis/autokey
https://api.github.com/repos/deriamis/autokey
opened
[CLOSED] Cannot "stop" autokey daemon on non-English locales
auto-migrated Priority-Medium Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Saturday Mar 28, 2015 at 10:54 GMT_ _Originally opened as https://github.com/timoc/autokey/issues/17_ ---- ``` What steps will reproduce the problem? 1. Install another language, such as fr_FR 2. Execute the following in a terminal: :: TEST CASE :: root@karmic-vm:~# echo $LANG en_US.UTF-8 root@karmic-vm:~# invoke-rc.d autokey start root@karmic-vm:~# pkill -9 -f autokey root@karmic-vm:~# LC_ALL=fr_FR.UTF-8 invoke-rc.d autokey stop [Errno 3] Aucun processus de ce type invoke-rc.d: initscript autokey, action "stop" failed. root@karmic-vm:~# :: END TEST CASE :: The error lies on http://code.google.com/p/autokey/source/browse/trunk/src/lib/daemon.py#111 , where the daemon checks the error text rather than the error code. This error seems to primarily be presenting itself when removing the autokey package, but can occur under the situation described above. See https://launchpad.net/bugs/479131 for more information. ``` Original issue reported on code.google.com by `Luke.Faraone` on 1 Dec 2009 at 1:35
1.0
[CLOSED] Cannot "stop" autokey daemon on non-English locales - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Saturday Mar 28, 2015 at 10:54 GMT_ _Originally opened as https://github.com/timoc/autokey/issues/17_ ---- ``` What steps will reproduce the problem? 1. Install another language, such as fr_FR 2. Execute the following in a terminal: :: TEST CASE :: root@karmic-vm:~# echo $LANG en_US.UTF-8 root@karmic-vm:~# invoke-rc.d autokey start root@karmic-vm:~# pkill -9 -f autokey root@karmic-vm:~# LC_ALL=fr_FR.UTF-8 invoke-rc.d autokey stop [Errno 3] Aucun processus de ce type invoke-rc.d: initscript autokey, action "stop" failed. root@karmic-vm:~# :: END TEST CASE :: The error lies on http://code.google.com/p/autokey/source/browse/trunk/src/lib/daemon.py#111 , where the daemon checks the error text rather than the error code. This error seems to primarily be presenting itself when removing the autokey package, but can occur under the situation described above. See https://launchpad.net/bugs/479131 for more information. ``` Original issue reported on code.google.com by `Luke.Faraone` on 1 Dec 2009 at 1:35
defect
cannot stop autokey daemon on non english locales issue by saturday mar at gmt originally opened as what steps will reproduce the problem install another language such as fr fr execute the following in a terminal test case root karmic vm echo lang en us utf root karmic vm invoke rc d autokey start root karmic vm pkill f autokey root karmic vm lc all fr fr utf invoke rc d autokey stop aucun processus de ce type invoke rc d initscript autokey action stop failed root karmic vm end test case the error lies on where the daemon checks the error text rather than the error code this error seems to primarily be presenting itself when removing the autokey package but can occur under the situation described above see for more information original issue reported on code google com by luke faraone on dec at
1
40,145
9,855,246,073
IssuesEvent
2019-06-19 18:53:54
zfsonlinux/zfs
https://api.github.com/repos/zfsonlinux/zfs
closed
memory leak in zpool_vdev.c check_disk function
Type: Defect good first issue
<!-- Please fill out the following template, which will help other contributors address your issue. --> After open the path and call blkid_get_tag_value function, should call free(value) before return <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Distribution Version | Linux Kernel | Architecture | ZFS Version | SPL Version | <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing ### Describe how to reproduce the problem ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
memory leak in zpool_vdev.c check_disk function - <!-- Please fill out the following template, which will help other contributors address your issue. --> After open the path and call blkid_get_tag_value function, should call free(value) before return <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Distribution Version | Linux Kernel | Architecture | ZFS Version | SPL Version | <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing ### Describe how to reproduce the problem ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
memory leak in zpool vdev c check disk function after open the path and call blkid get tag value function should call free value before return thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information type version name distribution name distribution version linux kernel architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing describe how to reproduce the problem include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
386,821
11,451,149,589
IssuesEvent
2020-02-06 10:59:16
Son-Guhun/Titan-Land-Lands-of-Plenty
https://api.github.com/repos/Son-Guhun/Titan-Land-Lands-of-Plenty
closed
Create a program that makes adding new units easier
priority
## Model - [x] Create Race Selector - [x] Name - [x] Parent - [x] Create Worker - [x] Race (Human/Orc/Undead/Nelf/Creep) - [x] Name - [x] Spawner - [x] Is Reforged? (campaign classification) - [x] Create Structure - [x] Worker - [x] Inherits Race and IsReforged from Worker - [x] Create Units - [x] Base Unit - [x] Structure - [x] Model - [x] Inherits Race and IsReforged from Structure ## View https://pysimplegui.readthedocs.io/en/latest/cookbook/ https://kivy.org/doc/stable/gettingstarted/first_app.html
1.0
Create a program that makes adding new units easier - ## Model - [x] Create Race Selector - [x] Name - [x] Parent - [x] Create Worker - [x] Race (Human/Orc/Undead/Nelf/Creep) - [x] Name - [x] Spawner - [x] Is Reforged? (campaign classification) - [x] Create Structure - [x] Worker - [x] Inherits Race and IsReforged from Worker - [x] Create Units - [x] Base Unit - [x] Structure - [x] Model - [x] Inherits Race and IsReforged from Structure ## View https://pysimplegui.readthedocs.io/en/latest/cookbook/ https://kivy.org/doc/stable/gettingstarted/first_app.html
non_defect
create a program that makes adding new units easier model create race selector name parent create worker race human orc undead nelf creep name spawner is reforged campaign classification create structure worker inherits race and isreforged from worker create units base unit structure model inherits race and isreforged from structure view
0
174,645
27,702,281,871
IssuesEvent
2023-03-14 08:55:14
elementor/elementor
https://api.github.com/repos/elementor/elementor
closed
100 VH is not equal to "fit to screen". / elementor dev 3.6
request/enhancement type/ui/ux component/editor type/design
hello @shilo-ey ,good time ! As you know, in version 3.6, the fit to screen height has been removed. And we can only set the height of the container to 100vh, so that the height of the container is equal to the screen size. But 100VH has a drawback that would be great if fixed in version 3.6. **This problem only exists on mobile phones and works properly on the desktop.** This problem in the mobile phone causes the content to not be in the center after loading the page. We have to scroll a bit to hide the address of the bar and in this case the content will be in the center. This is not really good. The problem is that if we set the container height to 100VH, the container height will be equal to the screen, but due to the browser's load address, the screen can be scrolled ! This is a big weakness. But this problem does not exist in the "fit to screen" height. This means that if we set the section height to fit to screen, there is a browser's bar address, but it is not possible to navigate the page, and that's great. I hope I have explained the problem well. **But if I want to summarize this problem, it is that if we set the container height to 100VH, on the tablet and mobile phone, the screen is able to scroll (vertically). This is because the browser's address bar is not fixed. While it should not be scrollable** I hope this is fixed in 3.6. Thanks
1.0
100 VH is not equal to "fit to screen". / elementor dev 3.6 - hello @shilo-ey ,good time ! As you know, in version 3.6, the fit to screen height has been removed. And we can only set the height of the container to 100vh, so that the height of the container is equal to the screen size. But 100VH has a drawback that would be great if fixed in version 3.6. **This problem only exists on mobile phones and works properly on the desktop.** This problem in the mobile phone causes the content to not be in the center after loading the page. We have to scroll a bit to hide the address of the bar and in this case the content will be in the center. This is not really good. The problem is that if we set the container height to 100VH, the container height will be equal to the screen, but due to the browser's load address, the screen can be scrolled ! This is a big weakness. But this problem does not exist in the "fit to screen" height. This means that if we set the section height to fit to screen, there is a browser's bar address, but it is not possible to navigate the page, and that's great. I hope I have explained the problem well. **But if I want to summarize this problem, it is that if we set the container height to 100VH, on the tablet and mobile phone, the screen is able to scroll (vertically). This is because the browser's address bar is not fixed. While it should not be scrollable** I hope this is fixed in 3.6. Thanks
non_defect
vh is not equal to fit to screen elementor dev hello shilo ey good time as you know in version the fit to screen height has been removed and we can only set the height of the container to so that the height of the container is equal to the screen size but has a drawback that would be great if fixed in version this problem only exists on mobile phones and works properly on the desktop this problem in the mobile phone causes the content to not be in the center after loading the page we have to scroll a bit to hide the address of the bar and in this case the content will be in the center this is not really good the problem is that if we set the container height to the container height will be equal to the screen but due to the browser s load address the screen can be scrolled this is a big weakness but this problem does not exist in the fit to screen height this means that if we set the section height to fit to screen there is a browser s bar address but it is not possible to navigate the page and that s great i hope i have explained the problem well but if i want to summarize this problem it is that if we set the container height to on the tablet and mobile phone the screen is able to scroll vertically this is because the browser s address bar is not fixed while it should not be scrollable i hope this is fixed in thanks
0
10,501
2,622,168,552
IssuesEvent
2015-03-04 00:13:28
byzhang/rapidjson
https://api.github.com/repos/byzhang/rapidjson
closed
Incorrect parsing of unsigned int number types
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Decode number field < int32_max {"bar" : 1596811002 } 2. Decode number field > int32_max with bit 0x80000000 set {"bar" : 2740789000 } 3. Attempt to extract both fields from document as Int64() What is the expected output? What do you see instead? Expected output is correctly parsing to int64_t in both cases. Actual behavior is an Assert that the second case does not have the kInt64Flag. Set flags in this case include kNumber, kInt, and kUint. What version of the product are you using? On what operating system? 0.1 Please provide any additional information below. The error is at line 80 of rapidjson/document.h: if (!(u & 0x80000000)) flags_ |= kIntFlag | kInt64Flag I've fixed it locally by changing the code to this: GenericValue(unsigned u) : flags_(kNumberUintFlag) { data_.n.u64 = u; if (!(u & 0x80000000)) flags_ |= kIntFlag; flags_ |= kInt64Flag; } I didn't attach a patch due to the small scope of the change, but would be happy to if you'd like. ``` Original issue reported on code.google.com by `eric....@gmail.com` on 15 Oct 2012 at 8:34
1.0
Incorrect parsing of unsigned int number types - ``` What steps will reproduce the problem? 1. Decode number field < int32_max {"bar" : 1596811002 } 2. Decode number field > int32_max with bit 0x80000000 set {"bar" : 2740789000 } 3. Attempt to extract both fields from document as Int64() What is the expected output? What do you see instead? Expected output is correctly parsing to int64_t in both cases. Actual behavior is an Assert that the second case does not have the kInt64Flag. Set flags in this case include kNumber, kInt, and kUint. What version of the product are you using? On what operating system? 0.1 Please provide any additional information below. The error is at line 80 of rapidjson/document.h: if (!(u & 0x80000000)) flags_ |= kIntFlag | kInt64Flag I've fixed it locally by changing the code to this: GenericValue(unsigned u) : flags_(kNumberUintFlag) { data_.n.u64 = u; if (!(u & 0x80000000)) flags_ |= kIntFlag; flags_ |= kInt64Flag; } I didn't attach a patch due to the small scope of the change, but would be happy to if you'd like. ``` Original issue reported on code.google.com by `eric....@gmail.com` on 15 Oct 2012 at 8:34
defect
incorrect parsing of unsigned int number types what steps will reproduce the problem decode number field max bar decode number field max with bit set bar attempt to extract both fields from document as what is the expected output what do you see instead expected output is correctly parsing to t in both cases actual behavior is an assert that the second case does not have the set flags in this case include knumber kint and kuint what version of the product are you using on what operating system please provide any additional information below the error is at line of rapidjson document h if u flags kintflag i ve fixed it locally by changing the code to this genericvalue unsigned u flags knumberuintflag data n u if u flags kintflag flags i didn t attach a patch due to the small scope of the change but would be happy to if you d like original issue reported on code google com by eric gmail com on oct at
1
72,351
24,067,780,300
IssuesEvent
2022-09-17 18:47:10
krischik/vim-scripts
https://api.github.com/repos/krischik/vim-scripts
closed
Mispelling
Type-Defect Priority-Medium auto-migrated
``` The autoload script is named "rainbow_parenthsis" and the functions defined inside use the same spelling. However, the variable "rainbow_parenthesis#active" uses the wrong spelling, causing an error. ``` Original issue reported on code.google.com by `gra...@gmail.com` on 21 Sep 2010 at 6:12
1.0
Mispelling - ``` The autoload script is named "rainbow_parenthsis" and the functions defined inside use the same spelling. However, the variable "rainbow_parenthesis#active" uses the wrong spelling, causing an error. ``` Original issue reported on code.google.com by `gra...@gmail.com` on 21 Sep 2010 at 6:12
defect
mispelling the autoload script is named rainbow parenthsis and the functions defined inside use the same spelling however the variable rainbow parenthesis active uses the wrong spelling causing an error original issue reported on code google com by gra gmail com on sep at
1
175,100
6,547,012,748
IssuesEvent
2017-09-04 13:00:02
status-im/status-go
https://api.github.com/repos/status-im/status-go
closed
Wait for CompleteTransaction or DiscardTransaction for requests to the upstream server
advanced bug high-priority
Currently, the wait for `CompleteTransaction` or `DiscardTransaction` is done deeply in `go-ethereum` and it only works for transactions signed and sent locally. However, if an upstream json-rpc server is specified, this won't happen as it should. Internally, I suppose, we should move that waiting logic from `go-ethereum` into Status thus reverting this commit: https://github.com/status-im/go-ethereum/commit/b939a561e547c6815d302ade11812c069f3e3cbb This is not the final decision but rather just a suggestion. This task must be finished fast and under this milestone. You should also be prepared for a large portion of potential refactoring. Please, also add validation rules for `NodeConfig.UpstreamConfig`. TODO: Those URLs https://github.com/status-im/status-go/pull/193/files#diff-41b02904df72dc739704e64f91cdc759R92 seems a bit private to Status. Is it so or we can expose them?
1.0
Wait for CompleteTransaction or DiscardTransaction for requests to the upstream server - Currently, the wait for `CompleteTransaction` or `DiscardTransaction` is done deeply in `go-ethereum` and it only works for transactions signed and sent locally. However, if an upstream json-rpc server is specified, this won't happen as it should. Internally, I suppose, we should move that waiting logic from `go-ethereum` into Status thus reverting this commit: https://github.com/status-im/go-ethereum/commit/b939a561e547c6815d302ade11812c069f3e3cbb This is not the final decision but rather just a suggestion. This task must be finished fast and under this milestone. You should also be prepared for a large portion of potential refactoring. Please, also add validation rules for `NodeConfig.UpstreamConfig`. TODO: Those URLs https://github.com/status-im/status-go/pull/193/files#diff-41b02904df72dc739704e64f91cdc759R92 seems a bit private to Status. Is it so or we can expose them?
non_defect
wait for completetransaction or discardtransaction for requests to the upstream server currently the wait for completetransaction or discardtransaction is done deeply in go ethereum and it only works for transactions signed and sent locally however if an upstream json rpc server is specified this won t happen as it should internally i suppose we should move that waiting logic from go ethereum into status thus reverting this commit this is not the final decision but rather just a suggestion this task must be finished fast and under this milestone you should also be prepared for a large portion of potential refactoring please also add validation rules for nodeconfig upstreamconfig todo those urls seems a bit private to status is it so or we can expose them
0
116,308
24,896,339,163
IssuesEvent
2022-10-28 16:09:10
eclipse-theia/theia
https://api.github.com/repos/eclipse-theia/theia
opened
vscode: add proper `extensions` support in `theia.d.ts`
vscode
<!-- Please fill out the following content for a feature request. --> <!-- Please provide a clear description of the feature and any relevant information. --> ### Feature Description: The `extensions` namespace is currently marked as unsupported since we refer to this api as `plugins`: <details> <summary>Comparator Screenshot</summary> <br /> ![image](https://user-images.githubusercontent.com/40359487/198682866-34fdf4e6-2b64-41e6-945e-dc4a469eeba2.png) </details> The implementation seems backwards, `theia.d.ts` should follow `vscode.d.ts` (use `extensions` as a namespace and so on), and customizations for plugins should exist somewhere else: https://github.com/eclipse-theia/theia/blob/0a67da25bfd6b45f3e9c605e3a358f3a1a1f14be/packages/plugin-ext/src/plugin/plugin-context.ts#L1108-L1196 cc @tsmaeder
1.0
vscode: add proper `extensions` support in `theia.d.ts` - <!-- Please fill out the following content for a feature request. --> <!-- Please provide a clear description of the feature and any relevant information. --> ### Feature Description: The `extensions` namespace is currently marked as unsupported since we refer to this api as `plugins`: <details> <summary>Comparator Screenshot</summary> <br /> ![image](https://user-images.githubusercontent.com/40359487/198682866-34fdf4e6-2b64-41e6-945e-dc4a469eeba2.png) </details> The implementation seems backwards, `theia.d.ts` should follow `vscode.d.ts` (use `extensions` as a namespace and so on), and customizations for plugins should exist somewhere else: https://github.com/eclipse-theia/theia/blob/0a67da25bfd6b45f3e9c605e3a358f3a1a1f14be/packages/plugin-ext/src/plugin/plugin-context.ts#L1108-L1196 cc @tsmaeder
non_defect
vscode add proper extensions support in theia d ts feature description the extensions namespace is currently marked as unsupported since we refer to this api as plugins comparator screenshot the implementation seems backwards theia d ts should follow vscode d ts use extensions as a namespace and so on and customizations for plugins should exist somewhere else cc tsmaeder
0
918
3,377,770,813
IssuesEvent
2015-11-25 06:43:09
onyx-platform/onyx
https://api.github.com/repos/onyx-platform/onyx
closed
Fix the plugin release process
release-process
Currently the plugin releases do not update the READMEs which breaks the X.x.x.x releases. The version should be set back to a snapshot release. I'd prefer we always release 4 digit releases of plugins. For example, if we release 0.8.0, then onyx-metrics 0.8.0.0 should be released, then master should be set to 0.8.0.1-SNAPSHOT with the README and any other files reflecting this.
1.0
Fix the plugin release process - Currently the plugin releases do not update the READMEs which breaks the X.x.x.x releases. The version should be set back to a snapshot release. I'd prefer we always release 4 digit releases of plugins. For example, if we release 0.8.0, then onyx-metrics 0.8.0.0 should be released, then master should be set to 0.8.0.1-SNAPSHOT with the README and any other files reflecting this.
non_defect
fix the plugin release process currently the plugin releases do not update the readmes which breaks the x x x x releases the version should be set back to a snapshot release i d prefer we always release digit releases of plugins for example if we release then onyx metrics should be released then master should be set to snapshot with the readme and any other files reflecting this
0
336,518
30,199,944,734
IssuesEvent
2023-07-05 04:06:03
marcpage/libernet
https://api.github.com/repos/marcpage/libernet
closed
Add code in tests/Path_test.cpp to test const Path &Path::copyContentsTo(const Path&) const
good first issue test
Call and verify that copyContentsTo works correctly. ``` const Path &Path::copyContentsTo(const Path&) const ```
1.0
Add code in tests/Path_test.cpp to test const Path &Path::copyContentsTo(const Path&) const - Call and verify that copyContentsTo works correctly. ``` const Path &Path::copyContentsTo(const Path&) const ```
non_defect
add code in tests path test cpp to test const path path copycontentsto const path const call and verify that copycontentsto works correctly const path path copycontentsto const path const
0
34,847
7,460,910,863
IssuesEvent
2018-03-30 22:02:48
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
pisiviga: KÜ detailvaates on "Näita kõik" / "Peida kõik", aga kadunud on gruppide ükshaaval sisse-välja lülitamine
C: AIS P: low R: fixed T: defect
**Reported by aadikaljuvee on 4 Sep 2017 10:41 UTC** Lehel http://www.dev-ais-web.arhiiv.ee/et/description_unit/view/?id=110000010411 on KÜ andmete grupid: Põhiandmed Kirjeldus Sisu ja struktuur Juurdepääs Publikatsioonid ja koostamise andmed Neid gruppe peaks saama igaüht ükshaaval avada/sulgeda kui vastaval grupipealkirjal klõpsata (ja see avatus/suletus tuleks meelde jätta ka teiste KÜde jaoks), aga saab ainult kõiki korraga avada/sulgeda
1.0
pisiviga: KÜ detailvaates on "Näita kõik" / "Peida kõik", aga kadunud on gruppide ükshaaval sisse-välja lülitamine - **Reported by aadikaljuvee on 4 Sep 2017 10:41 UTC** Lehel http://www.dev-ais-web.arhiiv.ee/et/description_unit/view/?id=110000010411 on KÜ andmete grupid: Põhiandmed Kirjeldus Sisu ja struktuur Juurdepääs Publikatsioonid ja koostamise andmed Neid gruppe peaks saama igaüht ükshaaval avada/sulgeda kui vastaval grupipealkirjal klõpsata (ja see avatus/suletus tuleks meelde jätta ka teiste KÜde jaoks), aga saab ainult kõiki korraga avada/sulgeda
defect
pisiviga kü detailvaates on näita kõik peida kõik aga kadunud on gruppide ükshaaval sisse välja lülitamine reported by aadikaljuvee on sep utc lehel on kü andmete grupid põhiandmed kirjeldus sisu ja struktuur juurdepääs publikatsioonid ja koostamise andmed neid gruppe peaks saama igaüht ükshaaval avada sulgeda kui vastaval grupipealkirjal klõpsata ja see avatus suletus tuleks meelde jätta ka teiste küde jaoks aga saab ainult kõiki korraga avada sulgeda
1
333,262
29,519,507,467
IssuesEvent
2023-06-04 22:17:49
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Examine `test_jax_numpy_creation` tests
JAX Frontend Testing Test Sweep ToDo_internal
Ensure a comprehensive examination of all tests to confirm their alignment with the current test writing policy. Ensure the tests follow the following guidelines, among others: 1. Avoid restricting the use of `get_dtypes` to specific kinds (`float`, `int`) unless absolutely necessary; the default value should be used when possible. 2. Avoid the use of `assume` to bypass specific data types. 3. Unless necessary, avoid restricting the generation of numbers with `min_value` or `max_value`. For example, it is acceptable to set a maximum value for the `pad` function if the `pad_value` value can halt the test. 4. Ensure proper use of strategies. 5. Verify that all tests are executed rather than skipped by `test_values=False`, unless the test logic is implemented within the test body itself. 6. Confirm that testing is comprehensive and complete, meaning that all possible combinations of inputs are tested. 7. Avoid manually setting test parameters, including native array flags, container flags, the number of positional arguments, etc. 8. Ensure that no parameters are skipped during testing. > Always refer to the documentation for **Ivy Tests** and **Ivy Frontend Tests** for complete test writing policy. If a test function does not require update, simply mark the function in the list as completed, otherwise create an issue for that test function. - [ ] test_jax_numpy_array - [ ] test_jax_numpy_zeros_like - [ ] test_jax_numpy_arange - [ ] test_jax_numpy_zeros - [ ] test_jax_numpy_ones - [ ] test_jax_numpy_ones_like - [ ] test_jax_numpy_asarray - [ ] test_jax_numpy_hstack - [ ] test_jax_numpy_eye - [ ] test_numpy_triu - [ ] test_jax_numpy_empty - [ ] test_jax_numpy_vander - [ ] test_jax_numpy_full_like - [ ] test_jax_numpy_ndim - [ ] test_jax_numpy_empty_like - [ ] test_jax_numpy_full
2.0
Examine `test_jax_numpy_creation` tests - Ensure a comprehensive examination of all tests to confirm their alignment with the current test writing policy. Ensure the tests follow the following guidelines, among others: 1. Avoid restricting the use of `get_dtypes` to specific kinds (`float`, `int`) unless absolutely necessary; the default value should be used when possible. 2. Avoid the use of `assume` to bypass specific data types. 3. Unless necessary, avoid restricting the generation of numbers with `min_value` or `max_value`. For example, it is acceptable to set a maximum value for the `pad` function if the `pad_value` value can halt the test. 4. Ensure proper use of strategies. 5. Verify that all tests are executed rather than skipped by `test_values=False`, unless the test logic is implemented within the test body itself. 6. Confirm that testing is comprehensive and complete, meaning that all possible combinations of inputs are tested. 7. Avoid manually setting test parameters, including native array flags, container flags, the number of positional arguments, etc. 8. Ensure that no parameters are skipped during testing. > Always refer to the documentation for **Ivy Tests** and **Ivy Frontend Tests** for complete test writing policy. If a test function does not require update, simply mark the function in the list as completed, otherwise create an issue for that test function. - [ ] test_jax_numpy_array - [ ] test_jax_numpy_zeros_like - [ ] test_jax_numpy_arange - [ ] test_jax_numpy_zeros - [ ] test_jax_numpy_ones - [ ] test_jax_numpy_ones_like - [ ] test_jax_numpy_asarray - [ ] test_jax_numpy_hstack - [ ] test_jax_numpy_eye - [ ] test_numpy_triu - [ ] test_jax_numpy_empty - [ ] test_jax_numpy_vander - [ ] test_jax_numpy_full_like - [ ] test_jax_numpy_ndim - [ ] test_jax_numpy_empty_like - [ ] test_jax_numpy_full
non_defect
examine test jax numpy creation tests ensure a comprehensive examination of all tests to confirm their alignment with the current test writing policy ensure the tests follow the following guidelines among others avoid restricting the use of get dtypes to specific kinds float int unless absolutely necessary the default value should be used when possible avoid the use of assume to bypass specific data types unless necessary avoid restricting the generation of numbers with min value or max value for example it is acceptable to set a maximum value for the pad function if the pad value value can halt the test ensure proper use of strategies verify that all tests are executed rather than skipped by test values false unless the test logic is implemented within the test body itself confirm that testing is comprehensive and complete meaning that all possible combinations of inputs are tested avoid manually setting test parameters including native array flags container flags the number of positional arguments etc ensure that no parameters are skipped during testing always refer to the documentation for ivy tests and ivy frontend tests for complete test writing policy if a test function does not require update simply mark the function in the list as completed otherwise create an issue for that test function test jax numpy array test jax numpy zeros like test jax numpy arange test jax numpy zeros test jax numpy ones test jax numpy ones like test jax numpy asarray test jax numpy hstack test jax numpy eye test numpy triu test jax numpy empty test jax numpy vander test jax numpy full like test jax numpy ndim test jax numpy empty like test jax numpy full
0
32,476
8,868,673,854
IssuesEvent
2019-01-11 01:06:57
CodaProtocol/coda
https://api.github.com/repos/CodaProtocol/coda
closed
Make helper function in Ktree non-recursive
area-ledger-builder-controller category-bug category-mentored category-quick-fix effort-easy impact-crash priority-high
## Current Behavior A [helper function](https://github.com/CodaProtocol/coda/blob/68e3556abe88ca73a66ad26705646cfadb866b06/src/lib/ledger_builder_controller/ktree.ml#L104) within Ledger-builder-controller's `add` function will recurse, at worst case, through every entry in the `Ktree`, adding 3 call stack entries per resursive call due to the nesting in `List.map`. For larger `Security.max_depth` values, this will cause a stack overflow. ## Desired Behavior This helper function should be rewritten to either be tail recursive (preferably) or an imperative loop (if you have a good reason) so that we can scale `Ktree` to large maximum depths. A better name for the helper function would be nice as well. ## How to fix Go to the [helper function](https://github.com/CodaProtocol/coda/blob/68e3556abe88ca73a66ad26705646cfadb866b06/src/lib/ledger_builder_controller/ktree.ml#L104) inside `add`. This task is very localized to just this function.
1.0
Make helper function in Ktree non-recursive - ## Current Behavior A [helper function](https://github.com/CodaProtocol/coda/blob/68e3556abe88ca73a66ad26705646cfadb866b06/src/lib/ledger_builder_controller/ktree.ml#L104) within Ledger-builder-controller's `add` function will recurse, at worst case, through every entry in the `Ktree`, adding 3 call stack entries per resursive call due to the nesting in `List.map`. For larger `Security.max_depth` values, this will cause a stack overflow. ## Desired Behavior This helper function should be rewritten to either be tail recursive (preferably) or an imperative loop (if you have a good reason) so that we can scale `Ktree` to large maximum depths. A better name for the helper function would be nice as well. ## How to fix Go to the [helper function](https://github.com/CodaProtocol/coda/blob/68e3556abe88ca73a66ad26705646cfadb866b06/src/lib/ledger_builder_controller/ktree.ml#L104) inside `add`. This task is very localized to just this function.
non_defect
make helper function in ktree non recursive current behavior a within ledger builder controller s add function will recurse at worst case through every entry in the ktree adding call stack entries per resursive call due to the nesting in list map for larger security max depth values this will cause a stack overflow desired behavior this helper function should be rewritten to either be tail recursive preferably or an imperative loop if you have a good reason so that we can scale ktree to large maximum depths a better name for the helper function would be nice as well how to fix go to the inside add this task is very localized to just this function
0
40,522
12,799,555,670
IssuesEvent
2020-07-02 15:34:38
TreyM-WSS/WhiteSource-Demo
https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo
opened
CVE-2017-5650 (High) detected in tomcat-embed-core-8.5.4.jar
security vulnerability
## CVE-2017-5650 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p> <p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.4.0.RELEASE.jar - :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tomcat 9.0.0.M1 to 9.0.0.M18 and 8.5.0 to 8.5.12, the handling of an HTTP/2 GOAWAY frame for a connection did not close streams associated with that connection that were currently waiting for a WINDOW_UPDATE before allowing the application to write more data. These waiting streams each consumed a thread. A malicious client could therefore construct a series of HTTP/2 requests that would consume all available processing threads. <p>Publish Date: 2017-04-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5650>CVE-2017-5650</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5650">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5650</a></p> <p>Release Date: 2017-04-17</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M19,8.5.13,org.apache.tomcat:tomcat-coyote:9.0.0.M19,8.5.13</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.4","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.4.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M19,8.5.13,org.apache.tomcat:tomcat-coyote:9.0.0.M19,8.5.13"}],"vulnerabilityIdentifier":"CVE-2017-5650","vulnerabilityDetails":"In Apache Tomcat 9.0.0.M1 to 9.0.0.M18 and 8.5.0 to 8.5.12, the handling of an HTTP/2 GOAWAY frame for a connection did not close streams associated with that connection that were currently waiting for a WINDOW_UPDATE before allowing the application to write more data. These waiting streams each consumed a thread. A malicious client could therefore construct a series of HTTP/2 requests that would consume all available processing threads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5650","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-5650 (High) detected in tomcat-embed-core-8.5.4.jar - ## CVE-2017-5650 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p> <p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.4.0.RELEASE.jar - :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tomcat 9.0.0.M1 to 9.0.0.M18 and 8.5.0 to 8.5.12, the handling of an HTTP/2 GOAWAY frame for a connection did not close streams associated with that connection that were currently waiting for a WINDOW_UPDATE before allowing the application to write more data. These waiting streams each consumed a thread. A malicious client could therefore construct a series of HTTP/2 requests that would consume all available processing threads. <p>Publish Date: 2017-04-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5650>CVE-2017-5650</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5650">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5650</a></p> <p>Release Date: 2017-04-17</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M19,8.5.13,org.apache.tomcat:tomcat-coyote:9.0.0.M19,8.5.13</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.4","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.4.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M19,8.5.13,org.apache.tomcat:tomcat-coyote:9.0.0.M19,8.5.13"}],"vulnerabilityIdentifier":"CVE-2017-5650","vulnerabilityDetails":"In Apache Tomcat 9.0.0.M1 to 9.0.0.M18 and 8.5.0 to 8.5.12, the handling of an HTTP/2 GOAWAY frame for a connection did not close streams associated with that connection that were currently waiting for a WINDOW_UPDATE before allowing the application to write more data. These waiting streams each consumed a thread. A malicious client could therefore construct a series of HTTP/2 requests that would consume all available processing threads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5650","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file tmp ws scm whitesource demo pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in head commit a href vulnerability details in apache tomcat to and to the handling of an http goaway frame for a connection did not close streams associated with that connection that were currently waiting for a window update before allowing the application to write more data these waiting streams each consumed a thread a malicious client could therefore construct a series of http requests that would consume all available processing threads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in apache tomcat to and to the handling of an http goaway frame for a connection did not close streams associated with that connection that were currently waiting for a window update before allowing the application to write more data these waiting streams each consumed a thread a malicious client could therefore construct a series of http requests that would consume all available processing threads vulnerabilityurl
0