Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,488
| 2,610,094,959
|
IssuesEvent
|
2015-02-26 18:28:39
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳红蓝光怎样祛除青春痘
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳红蓝光怎样祛除青春痘【深圳韩方科颜全国热线400-869-181
8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构��
�韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳�
��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不
反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创��
�内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客�
��上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:20
|
1.0
|
深圳红蓝光怎样祛除青春痘 - ```
深圳红蓝光怎样祛除青春痘【深圳韩方科颜全国热线400-869-181
8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构��
�韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳�
��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不
反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创��
�内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客�
��上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:20
|
defect
|
深圳红蓝光怎样祛除青春痘 深圳红蓝光怎样祛除青春痘【 , 】深圳韩方科颜专业祛痘连锁机构,机构�� �韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳� ��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不 反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创�� �内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客� ��上的痘痘。 original issue reported on code google com by szft com on may at
| 1
|
37,833
| 8,530,200,388
|
IssuesEvent
|
2018-11-03 19:58:17
|
jadrake75/stamp-imageparsing
|
https://api.github.com/repos/jadrake75/stamp-imageparsing
|
closed
|
Setting a file extension does not change the mimeType
|
Defect
|
changing the extension should set the mimeType but as it is the mimeType is set to JPEG
see image-handler.saveRegions()
|
1.0
|
Setting a file extension does not change the mimeType - changing the extension should set the mimeType but as it is the mimeType is set to JPEG
see image-handler.saveRegions()
|
defect
|
setting a file extension does not change the mimetype changing the extension should set the mimetype but as it is the mimetype is set to jpeg see image handler saveregions
| 1
|
336,159
| 10,172,463,940
|
IssuesEvent
|
2019-08-08 10:45:06
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
colonialforestapts.securecafe.com - site is not usable
|
browser-fenix engine-gecko priority-normal
|
<!-- @browser: Firefox Mobile 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://colonialforestapts.securecafe.com/residentservices/ResidentCafeHandler.ashx
**Browser / Version**: Firefox Mobile 69.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Page won't load
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
colonialforestapts.securecafe.com - site is not usable - <!-- @browser: Firefox Mobile 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://colonialforestapts.securecafe.com/residentservices/ResidentCafeHandler.ashx
**Browser / Version**: Firefox Mobile 69.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Page won't load
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
colonialforestapts securecafe com site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description page won t load steps to reproduce browser configuration none from with ❤️
| 0
|
51,121
| 13,188,136,177
|
IssuesEvent
|
2020-08-13 05:39:08
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
[docs] switch to python-based documentation builder (Trac #2025)
|
Migrated from Trac analysis defect
|
speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo
http://code.icecube.wisc.edu/icetray/projects/docs/trunk
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2025">https://code.icecube.wisc.edu/ticket/2025</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:57",
"description": "speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo \nhttp://code.icecube.wisc.edu/icetray/projects/docs/trunk",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067237750774",
"component": "analysis",
"summary": "[docs] switch to python-based documentation builder",
"priority": "normal",
"keywords": "",
"time": "2017-05-18T07:11:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[docs] switch to python-based documentation builder (Trac #2025) - speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo
http://code.icecube.wisc.edu/icetray/projects/docs/trunk
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2025">https://code.icecube.wisc.edu/ticket/2025</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:57",
"description": "speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo \nhttp://code.icecube.wisc.edu/icetray/projects/docs/trunk",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067237750774",
"component": "analysis",
"summary": "[docs] switch to python-based documentation builder",
"priority": "normal",
"keywords": "",
"time": "2017-05-18T07:11:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
switch to python based documentation builder trac speed up documentation build by switching to python based doc builder docs build use the jn option on multicore systems this is in its own project which was added to combo migrated from json status closed changetime description speed up documentation build by switching to python based doc builder docs build use the jn option on multicore systems this is in its own project which was added to combo n reporter kjmeagher cc resolution fixed ts component analysis summary switch to python based documentation builder priority normal keywords time milestone owner nega type defect
| 1
|
134,316
| 5,224,164,280
|
IssuesEvent
|
2017-01-27 14:44:15
|
changer/sc-messages
|
https://api.github.com/repos/changer/sc-messages
|
closed
|
[IncidentChanger] Spellingissue in email (very-high)
|
02 - Status - Assigned 13 - Priority - Low
|
_From @changertesting on July 14, 2015 11:33_
Windows IE11
In a receiving email there is writte 'very-high' (with stripe). In the application there is everywhere 'very high' written (without stripe).

Please write this in all emails also as 'very high' (without stripe).
_Copied from original issue: changer/sc-client#459_
|
1.0
|
[IncidentChanger] Spellingissue in email (very-high) - _From @changertesting on July 14, 2015 11:33_
Windows IE11
In a receiving email there is writte 'very-high' (with stripe). In the application there is everywhere 'very high' written (without stripe).

Please write this in all emails also as 'very high' (without stripe).
_Copied from original issue: changer/sc-client#459_
|
non_defect
|
spellingissue in email very high from changertesting on july windows in a receiving email there is writte very high with stripe in the application there is everywhere very high written without stripe please write this in all emails also as very high without stripe copied from original issue changer sc client
| 0
|
6,452
| 8,744,950,807
|
IssuesEvent
|
2018-12-13 00:19:34
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
--incompatible_use_aapt2_by_default: Enable Bazel to use aapt2 by default
|
incompatible-change
|
Tracking issue for the incompatible change to switch Bazel to use aapt2 by default.
FR issue: https://github.com/bazelbuild/bazel/issues/4103
To switch between aapt versions, Bazel uses the following in order of precedence:
1) Value of `--android_aapt` flag.
2) Value of `aapt_version` attribute in the `android_binary` and `android_local_test` target.
3) `--incompatible_use_aapt2_by_default`: false -> aapt, true -> aapt2.
To resolve issues when migrating your app to build with aapt2, see https://developer.android.com/studio/command-line/aapt2#aapt2_changes
|
True
|
--incompatible_use_aapt2_by_default: Enable Bazel to use aapt2 by default - Tracking issue for the incompatible change to switch Bazel to use aapt2 by default.
FR issue: https://github.com/bazelbuild/bazel/issues/4103
To switch between aapt versions, Bazel uses the following in order of precedence:
1) Value of `--android_aapt` flag.
2) Value of `aapt_version` attribute in the `android_binary` and `android_local_test` target.
3) `--incompatible_use_aapt2_by_default`: false -> aapt, true -> aapt2.
To resolve issues when migrating your app to build with aapt2, see https://developer.android.com/studio/command-line/aapt2#aapt2_changes
|
non_defect
|
incompatible use by default enable bazel to use by default tracking issue for the incompatible change to switch bazel to use by default fr issue to switch between aapt versions bazel uses the following in order of precedence value of android aapt flag value of aapt version attribute in the android binary and android local test target incompatible use by default false aapt true to resolve issues when migrating your app to build with see
| 0
|
83,589
| 24,094,932,592
|
IssuesEvent
|
2022-09-19 17:51:34
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Kitty fails to build on Hydra for `x86_64-darwin`
|
6.topic: darwin 0.kind: build failure
|
Kitty (v0.25.1) seem to be having issues building on Hydra for `x86_64-darwin` ([log](https://hydra.nixos.org/build/179644390/nixlog/1)).
This is strange since,
* Hydra has no issues building the same package on `aarch64-darwin` ([log](https://hydra.nixos.org/build/179646101/nixlog/2));
* @ofborg didn't have any issues building the package on `x86_64-darwin` ([log](https://logs.nix.ci/?key=nixos/nixpkgs.175243&attempt_id=527ccead-4551-4cb3-a56c-a4ccc1a535f7)); and
* I don't have any issues building it locally on `x86_64-darwin`.
I don't understand enough about Hydra, and the configuration of the different build machines to really understand what's going on here, but it seems like the issue is likely due to some configuration difference between the build machines for `aarch64-darwin` and those for `x86_64-darwin`.
|
1.0
|
Kitty fails to build on Hydra for `x86_64-darwin` - Kitty (v0.25.1) seem to be having issues building on Hydra for `x86_64-darwin` ([log](https://hydra.nixos.org/build/179644390/nixlog/1)).
This is strange since,
* Hydra has no issues building the same package on `aarch64-darwin` ([log](https://hydra.nixos.org/build/179646101/nixlog/2));
* @ofborg didn't have any issues building the package on `x86_64-darwin` ([log](https://logs.nix.ci/?key=nixos/nixpkgs.175243&attempt_id=527ccead-4551-4cb3-a56c-a4ccc1a535f7)); and
* I don't have any issues building it locally on `x86_64-darwin`.
I don't understand enough about Hydra, and the configuration of the different build machines to really understand what's going on here, but it seems like the issue is likely due to some configuration difference between the build machines for `aarch64-darwin` and those for `x86_64-darwin`.
|
non_defect
|
kitty fails to build on hydra for darwin kitty seem to be having issues building on hydra for darwin this is strange since hydra has no issues building the same package on darwin ofborg didn t have any issues building the package on darwin and i don t have any issues building it locally on darwin i don t understand enough about hydra and the configuration of the different build machines to really understand what s going on here but it seems like the issue is likely due to some configuration difference between the build machines for darwin and those for darwin
| 0
|
21,516
| 3,516,551,640
|
IssuesEvent
|
2016-01-12 00:22:40
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
opened
|
generic method comments don't work for list/map literals
|
analyzer-strong-mode area-analyzer priority-low Type-Defect
|
`/*<T>*/[]` does not work, but `new List/*<T>*/()` does.
The workaround is easy: either the `new List` form or use downwards inference.
|
1.0
|
generic method comments don't work for list/map literals - `/*<T>*/[]` does not work, but `new List/*<T>*/()` does.
The workaround is easy: either the `new List` form or use downwards inference.
|
defect
|
generic method comments don t work for list map literals does not work but new list does the workaround is easy either the new list form or use downwards inference
| 1
|
67,449
| 20,961,613,433
|
IssuesEvent
|
2022-03-27 21:49:40
|
abedmaatalla/sipdroid
|
https://api.github.com/repos/abedmaatalla/sipdroid
|
closed
|
Sipdroid doesn't work with miui dialer
|
Priority-Medium Type-Defect auto-migrated
|
```
With MIUI installed, sipdroid loses control when you click into contacts. I.e.
if you press the contacts list button in the sipdroid dialer, it takes you into
the MIUI contacts pane, from which point the MIUI dialer takes over.
```
Original issue reported on code.google.com by `r...@flavor8.com` on 16 Oct 2011 at 2:37
|
1.0
|
Sipdroid doesn't work with miui dialer - ```
With MIUI installed, sipdroid loses control when you click into contacts. I.e.
if you press the contacts list button in the sipdroid dialer, it takes you into
the MIUI contacts pane, from which point the MIUI dialer takes over.
```
Original issue reported on code.google.com by `r...@flavor8.com` on 16 Oct 2011 at 2:37
|
defect
|
sipdroid doesn t work with miui dialer with miui installed sipdroid loses control when you click into contacts i e if you press the contacts list button in the sipdroid dialer it takes you into the miui contacts pane from which point the miui dialer takes over original issue reported on code google com by r com on oct at
| 1
|
23,929
| 3,873,906,380
|
IssuesEvent
|
2016-04-11 18:37:10
|
jfabry/LiveRobotProgramming
|
https://api.github.com/repos/jfabry/LiveRobotProgramming
|
closed
|
Var state is not copied on machine change
|
Component-Logic Priority-High Type-Defect
|
Code:
(machine m
(var x := [0])
(state foo (onentry [x:=42]))
(event fake [1+2])
)
(spawn m foo)
The value of x is 42, you can see it in the UI. But when you edit the transition the value is reset to 0.
Don't forget the test cases please :-)
|
1.0
|
Var state is not copied on machine change - Code:
(machine m
(var x := [0])
(state foo (onentry [x:=42]))
(event fake [1+2])
)
(spawn m foo)
The value of x is 42, you can see it in the UI. But when you edit the transition the value is reset to 0.
Don't forget the test cases please :-)
|
defect
|
var state is not copied on machine change code machine m var x state foo onentry event fake spawn m foo the value of x is you can see it in the ui but when you edit the transition the value is reset to don t forget the test cases please
| 1
|
665,353
| 22,310,140,187
|
IssuesEvent
|
2022-06-13 16:13:08
|
GoogleCloudPlatform/emblem
|
https://api.github.com/repos/GoogleCloudPlatform/emblem
|
closed
|
API Service 500 errors when following setup-quickstart.md & new terraform configuration
|
type: bug priority: p1
|
## Problem
When following [setup-quickstart](https://github.com/GoogleCloudPlatform/emblem/docs/tutorials/setup-quickstart.md) and using the new terraform configuration, the resulting API service throw 500 errors.
The error logs & responses are unclear, and some of the logs produced are parsed as multiple log entries which makes scanning difficult.
## Analysis
It looks like the Content API as deployed lacks IAM read access to Firestore, and this is handled with convoluted log output that makes verifying the problem difficult. If so, a change to fix this would be to adjust the IAM privileges on the Content API service identity.
Follow-up tasks to improve log-based troubleshooting would help. Because this is a demo, it would be worth evaluating if exposing the error more clearly in the HTTP response would be helpful and not lead folks to a production anti-pattern.
## Steps to Reproduce
Log entries from my installation:
```
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 2077, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1525, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1523, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1509, in dispatch_request
return self.ensure_sync(self.view_functions[rule.endpoint])(**req.view_args)
File "/app/main.py", line 85, in handle_list
return methods.list(resource_name)
File "/app/resources/methods.py", line 50, in list
if not auth.allowed("GET", resource_kind):
File "/app/resources/auth.py", line 55, in allowed
is_approver = user_is_approver(email)
File "/app/resources/auth.py", line 26, in user_is_approver
matching_approvers = db.list_matching(
File "/app/data/cloud_firestore.py", line 47, in list_matching
for resource_ref in query.stream():
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/query.py", line 273, in stream
response_iterator, expected_prefix = self._get_stream_iterator(
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/query.py", line 218, in _get_stream_iterator
response_iterator = self._client._firestore_api.run_query(
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/services/firestore/client.py", line 1236, in run_query
response = rpc(
File "/usr/local/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__
return wrapped_func(*args, **kwargs)
File "/usr/local/lib/python3.10/site-packages/google/api_core/retry.py", line 283, in retry_wrapped_func
return retry_target(
File "/usr/local/lib/python3.10/site-packages/google/api_core/retry.py", line 190, in retry_target
return target()
File "/usr/local/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 151, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.PermissionDenied: 403 Missing or insufficient permissions.
```
and
```
Reason: Internal Server Error
HTTP response headers: HTTPHeaderDict({'content-type': 'text/html; charset=utf-8', 'X-Cloud-Trace-Context': 'b43e6df168c33c04bdb1eee2b06bb27b;o=1', 'Date': 'Mon, 06 Jun 2022 15:58:21 GMT', 'Server': 'Google Frontend', 'Content-Length': '265', 'Alt-Svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000,h3-Q050=":443"; ma=2592000,h3-Q046=":443"; ma=2592000,h3-Q043=":443"; ma=2592000,quic=":443"; ma=2592000; v="46,43"'})
HTTP response body: <!doctype html>
<html lang=en>
<title>500 Internal Server Error</title>
<h1>Internal Server Error</h1>
<p>The server encountered an internal error and was unable to complete your request. Either the server is overloaded or there is an error in the application.</p>
```
|
1.0
|
API Service 500 errors when following setup-quickstart.md & new terraform configuration - ## Problem
When following [setup-quickstart](https://github.com/GoogleCloudPlatform/emblem/docs/tutorials/setup-quickstart.md) and using the new terraform configuration, the resulting API service throw 500 errors.
The error logs & responses are unclear, and some of the logs produced are parsed as multiple log entries which makes scanning difficult.
## Analysis
It looks like the Content API as deployed lacks IAM read access to Firestore, and this is handled with convoluted log output that makes verifying the problem difficult. If so, a change to fix this would be to adjust the IAM privileges on the Content API service identity.
Follow-up tasks to improve log-based troubleshooting would help. Because this is a demo, it would be worth evaluating if exposing the error more clearly in the HTTP response would be helpful and not lead folks to a production anti-pattern.
## Steps to Reproduce
Log entries from my installation:
```
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 2077, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1525, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1523, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1509, in dispatch_request
return self.ensure_sync(self.view_functions[rule.endpoint])(**req.view_args)
File "/app/main.py", line 85, in handle_list
return methods.list(resource_name)
File "/app/resources/methods.py", line 50, in list
if not auth.allowed("GET", resource_kind):
File "/app/resources/auth.py", line 55, in allowed
is_approver = user_is_approver(email)
File "/app/resources/auth.py", line 26, in user_is_approver
matching_approvers = db.list_matching(
File "/app/data/cloud_firestore.py", line 47, in list_matching
for resource_ref in query.stream():
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/query.py", line 273, in stream
response_iterator, expected_prefix = self._get_stream_iterator(
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/query.py", line 218, in _get_stream_iterator
response_iterator = self._client._firestore_api.run_query(
File "/usr/local/lib/python3.10/site-packages/google/cloud/firestore_v1/services/firestore/client.py", line 1236, in run_query
response = rpc(
File "/usr/local/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__
return wrapped_func(*args, **kwargs)
File "/usr/local/lib/python3.10/site-packages/google/api_core/retry.py", line 283, in retry_wrapped_func
return retry_target(
File "/usr/local/lib/python3.10/site-packages/google/api_core/retry.py", line 190, in retry_target
return target()
File "/usr/local/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 151, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.PermissionDenied: 403 Missing or insufficient permissions.
```
and
```
Reason: Internal Server Error
HTTP response headers: HTTPHeaderDict({'content-type': 'text/html; charset=utf-8', 'X-Cloud-Trace-Context': 'b43e6df168c33c04bdb1eee2b06bb27b;o=1', 'Date': 'Mon, 06 Jun 2022 15:58:21 GMT', 'Server': 'Google Frontend', 'Content-Length': '265', 'Alt-Svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000,h3-Q050=":443"; ma=2592000,h3-Q046=":443"; ma=2592000,h3-Q043=":443"; ma=2592000,quic=":443"; ma=2592000; v="46,43"'})
HTTP response body: <!doctype html>
<html lang=en>
<title>500 Internal Server Error</title>
<h1>Internal Server Error</h1>
<p>The server encountered an internal error and was unable to complete your request. Either the server is overloaded or there is an error in the application.</p>
```
|
non_defect
|
api service errors when following setup quickstart md new terraform configuration problem when following and using the new terraform configuration the resulting api service throw errors the error logs responses are unclear and some of the logs produced are parsed as multiple log entries which makes scanning difficult analysis it looks like the content api as deployed lacks iam read access to firestore and this is handled with convoluted log output that makes verifying the problem difficult if so a change to fix this would be to adjust the iam privileges on the content api service identity follow up tasks to improve log based troubleshooting would help because this is a demo it would be worth evaluating if exposing the error more clearly in the http response would be helpful and not lead folks to a production anti pattern steps to reproduce log entries from my installation traceback most recent call last file usr local lib site packages flask app py line in wsgi app response self full dispatch request file usr local lib site packages flask app py line in full dispatch request rv self handle user exception e file usr local lib site packages flask app py line in full dispatch request rv self dispatch request file usr local lib site packages flask app py line in dispatch request return self ensure sync self view functions req view args file app main py line in handle list return methods list resource name file app resources methods py line in list if not auth allowed get resource kind file app resources auth py line in allowed is approver user is approver email file app resources auth py line in user is approver matching approvers db list matching file app data cloud firestore py line in list matching for resource ref in query stream file usr local lib site packages google cloud firestore query py line in stream response iterator expected prefix self get stream iterator file usr local lib site packages google cloud firestore query py line in get stream iterator response iterator self client firestore api run query file usr local lib site packages google cloud firestore services firestore client py line in run query response rpc file usr local lib site packages google api core gapic method py line in call return wrapped func args kwargs file usr local lib site packages google api core retry py line in retry wrapped func return retry target file usr local lib site packages google api core retry py line in retry target return target file usr local lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions permissiondenied missing or insufficient permissions and reason internal server error http response headers httpheaderdict content type text html charset utf x cloud trace context o date mon jun gmt server google frontend content length alt svc ma ma ma ma ma quic ma v http response body internal server error internal server error the server encountered an internal error and was unable to complete your request either the server is overloaded or there is an error in the application
| 0
|
9,883
| 2,616,006,101
|
IssuesEvent
|
2015-03-02 00:50:22
|
jasonhall/bwapi
|
https://api.github.com/repos/jasonhall/bwapi
|
closed
|
Terrain Analyzer does not work without CompleteMapInformation
|
auto-migrated Priority-Medium Type-Defect Usability
|
```
It appears that the Terrain Analyzer will not work without the
CompleteMapInformation flag on, which makes sense. Unfortunately, this also
gives my AI complete information about every enemy unit outside of its fog of
war, which is unfortunate. I was wondering if there is any way to give the AI
knowledge about the terrain for the sake of BWTA, but deny it information
about the units-- as if it had "studied the map" before playing.
```
Original issue reported on code.google.com by `Clck...@gmail.com` on 15 Nov 2009 at 5:13
|
1.0
|
Terrain Analyzer does not work without CompleteMapInformation - ```
It appears that the Terrain Analyzer will not work without the
CompleteMapInformation flag on, which makes sense. Unfortunately, this also
gives my AI complete information about every enemy unit outside of its fog of
war, which is unfortunate. I was wondering if there is any way to give the AI
knowledge about the terrain for the sake of BWTA, but deny it information
about the units-- as if it had "studied the map" before playing.
```
Original issue reported on code.google.com by `Clck...@gmail.com` on 15 Nov 2009 at 5:13
|
defect
|
terrain analyzer does not work without completemapinformation it appears that the terrain analyzer will not work without the completemapinformation flag on which makes sense unfortunately this also gives my ai complete information about every enemy unit outside of its fog of war which is unfortunate i was wondering if there is any way to give the ai knowledge about the terrain for the sake of bwta but deny it information about the units as if it had studied the map before playing original issue reported on code google com by clck gmail com on nov at
| 1
|
39,919
| 9,744,989,956
|
IssuesEvent
|
2019-06-03 08:32:32
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
keyFilter on Chrome Android
|
defect
|
Hello;
The KeyFilter feature does not work on chrome for Android Mobile Devices.
Version 6.0.0 RC1
Thanks for the great effort
|
1.0
|
keyFilter on Chrome Android - Hello;
The KeyFilter feature does not work on chrome for Android Mobile Devices.
Version 6.0.0 RC1
Thanks for the great effort
|
defect
|
keyfilter on chrome android hello the keyfilter feature does not work on chrome for android mobile devices version thanks for the great effort
| 1
|
36,987
| 4,771,026,045
|
IssuesEvent
|
2016-10-26 16:47:35
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
opened
|
API: Expand read_csv dtype for categoricals
|
API Design Categorical Difficulty Intermediate Effort Low IO CSV
|
In https://github.com/pandas-dev/pandas/pull/13406 Chris added support for `dtype={'col': 'category'}` (thanks!). This issue is for expanding that syntax to allow a more complete specification of the resulting categorical.
```python
# Your code here
df = pd.read_csv(path, dtype={'col': pd.Categorical(['a', 'b', 'c'], ordered=True})
df = pd.read_csv(path, dtype={'col': ['a', 'b', 'c']}) # shorthand, but unordered only
# we would still accept `dtype={'col': 'category'}` as well, to infer categories
```
Implementation-wise, I think we can keep all the parsing logic as is, and simply loop over `dtype` and call `set_categories` (and maybe `as_ordered`) on all the categoricals just before returning to the user.
This would help a bit in dask, where their category type inference can fail if the first partition doesn't contain all the categories (see https://github.com/dask/dask/issues/1705). This is why it'd be preferable to do it as an option to `read_csv`, rather than putting in on the user to followup with a `set_categories`.
|
1.0
|
API: Expand read_csv dtype for categoricals - In https://github.com/pandas-dev/pandas/pull/13406 Chris added support for `dtype={'col': 'category'}` (thanks!). This issue is for expanding that syntax to allow a more complete specification of the resulting categorical.
```python
# Your code here
df = pd.read_csv(path, dtype={'col': pd.Categorical(['a', 'b', 'c'], ordered=True})
df = pd.read_csv(path, dtype={'col': ['a', 'b', 'c']}) # shorthand, but unordered only
# we would still accept `dtype={'col': 'category'}` as well, to infer categories
```
Implementation-wise, I think we can keep all the parsing logic as is, and simply loop over `dtype` and call `set_categories` (and maybe `as_ordered`) on all the categoricals just before returning to the user.
This would help a bit in dask, where their category type inference can fail if the first partition doesn't contain all the categories (see https://github.com/dask/dask/issues/1705). This is why it'd be preferable to do it as an option to `read_csv`, rather than putting in on the user to followup with a `set_categories`.
|
non_defect
|
api expand read csv dtype for categoricals in chris added support for dtype col category thanks this issue is for expanding that syntax to allow a more complete specification of the resulting categorical python your code here df pd read csv path dtype col pd categorical ordered true df pd read csv path dtype col shorthand but unordered only we would still accept dtype col category as well to infer categories implementation wise i think we can keep all the parsing logic as is and simply loop over dtype and call set categories and maybe as ordered on all the categoricals just before returning to the user this would help a bit in dask where their category type inference can fail if the first partition doesn t contain all the categories see this is why it d be preferable to do it as an option to read csv rather than putting in on the user to followup with a set categories
| 0
|
19,853
| 10,428,319,387
|
IssuesEvent
|
2019-09-16 22:11:53
|
gate5/test2
|
https://api.github.com/repos/gate5/test2
|
opened
|
CVE-2018-10054 (High) detected in h2-1.4.187.jar
|
security vulnerability
|
## CVE-2018-10054 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.187.jar</b></p></summary>
<p>null</p>
<p>Path to dependency file: /tmp/ws-scm/test2/pom.xml</p>
<p>Path to vulnerable library: epository/com/h2database/h2/1.4.187/h2-1.4.187.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.187.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/9caf3d214cc15f500423a2e431ea111cf9526739">9caf3d214cc15f500423a2e431ea111cf9526739</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
H2 1.4.197, as used in Datomic before 0.9.5697 and other products, allows remote code execution because CREATE ALIAS can execute arbitrary Java code.
<p>Publish Date: 2018-04-11
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054>CVE-2018-10054</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054</a></p>
<p>Release Date: 2018-04-11</p>
<p>Fix Resolution: 1.4.198</p>
</p>
</details>
<p></p>
|
True
|
CVE-2018-10054 (High) detected in h2-1.4.187.jar - ## CVE-2018-10054 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.187.jar</b></p></summary>
<p>null</p>
<p>Path to dependency file: /tmp/ws-scm/test2/pom.xml</p>
<p>Path to vulnerable library: epository/com/h2database/h2/1.4.187/h2-1.4.187.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.187.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/9caf3d214cc15f500423a2e431ea111cf9526739">9caf3d214cc15f500423a2e431ea111cf9526739</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
H2 1.4.197, as used in Datomic before 0.9.5697 and other products, allows remote code execution because CREATE ALIAS can execute arbitrary Java code.
<p>Publish Date: 2018-04-11
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054>CVE-2018-10054</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054</a></p>
<p>Release Date: 2018-04-11</p>
<p>Fix Resolution: 1.4.198</p>
</p>
</details>
<p></p>
|
non_defect
|
cve high detected in jar cve high severity vulnerability vulnerable library jar null path to dependency file tmp ws scm pom xml path to vulnerable library epository com jar dependency hierarchy x jar vulnerable library found in head commit a href vulnerability details as used in datomic before and other products allows remote code execution because create alias can execute arbitrary java code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
| 0
|
250,336
| 27,074,299,011
|
IssuesEvent
|
2023-02-14 09:30:06
|
AnswerConsulting/AnswerKing-Infrastructure
|
https://api.github.com/repos/AnswerConsulting/AnswerKing-Infrastructure
|
closed
|
Instance does not require IMDS access to require a token
|
security high module:ec2_instance
|
https://aquasecurity.github.io/tfsec/v0.61.3/checks/aws/ec2/enforce-http-token-imds/#possible-impact
Do not think there's any cost behind this one
|
True
|
Instance does not require IMDS access to require a token - https://aquasecurity.github.io/tfsec/v0.61.3/checks/aws/ec2/enforce-http-token-imds/#possible-impact
Do not think there's any cost behind this one
|
non_defect
|
instance does not require imds access to require a token do not think there s any cost behind this one
| 0
|
2,813
| 2,533,548,826
|
IssuesEvent
|
2015-01-24 00:41:21
|
pybox2d/pybox2d
|
https://api.github.com/repos/pybox2d/pybox2d
|
closed
|
Examples using pgu will crash with pygame 1.8.1
|
bug imported Priority-Medium
|
_From [ccanepacc](https://code.google.com/u/ccanepacc@gmail.com/) on September 04, 2008 21:07:03_
What steps will reproduce the problem? 1. install pygame 1.8.1
2. run demos.py , it will crash
3. Its a known compatibility problem with pgu - pygame 1.8.1 What is the expected output? What do you see instead? crash What version of the product are you using? On what operating system? Box2D-2.0.1b4.win32-py2.4.exe - win XP + SP3 Please provide any additional information below. Its a pgu incompatibility with pygame 1.8.1 . There is a pgu patch for
this, plus aditional compatibility notes in a thread of pygame users: http://thread.gmane.org/gmane.comp.python.pygame/15580
_Original issue: http://code.google.com/p/pybox2d/issues/detail?id=8_
|
1.0
|
Examples using pgu will crash with pygame 1.8.1 - _From [ccanepacc](https://code.google.com/u/ccanepacc@gmail.com/) on September 04, 2008 21:07:03_
What steps will reproduce the problem? 1. install pygame 1.8.1
2. run demos.py , it will crash
3. Its a known compatibility problem with pgu - pygame 1.8.1 What is the expected output? What do you see instead? crash What version of the product are you using? On what operating system? Box2D-2.0.1b4.win32-py2.4.exe - win XP + SP3 Please provide any additional information below. Its a pgu incompatibility with pygame 1.8.1 . There is a pgu patch for
this, plus aditional compatibility notes in a thread of pygame users: http://thread.gmane.org/gmane.comp.python.pygame/15580
_Original issue: http://code.google.com/p/pybox2d/issues/detail?id=8_
|
non_defect
|
examples using pgu will crash with pygame from on september what steps will reproduce the problem install pygame run demos py it will crash its a known compatibility problem with pgu pygame what is the expected output what do you see instead crash what version of the product are you using on what operating system exe win xp please provide any additional information below its a pgu incompatibility with pygame there is a pgu patch for this plus aditional compatibility notes in a thread of pygame users original issue
| 0
|
12,382
| 9,761,731,942
|
IssuesEvent
|
2019-06-05 09:29:11
|
rust-lang/wg-governance
|
https://api.github.com/repos/rust-lang/wg-governance
|
opened
|
Customize loomio theme
|
Infrastructure
|
We can customize the [loomio instance](https://github.com/loomio/loomio-deploy/blob/master/scripts/default_env#L73-L93) - we probably should and want to, to give it more feeling of rust-lang. However, what are our design resources? Like guidelines, colors etc...?
|
1.0
|
Customize loomio theme - We can customize the [loomio instance](https://github.com/loomio/loomio-deploy/blob/master/scripts/default_env#L73-L93) - we probably should and want to, to give it more feeling of rust-lang. However, what are our design resources? Like guidelines, colors etc...?
|
non_defect
|
customize loomio theme we can customize the we probably should and want to to give it more feeling of rust lang however what are our design resources like guidelines colors etc
| 0
|
223,185
| 24,711,670,906
|
IssuesEvent
|
2022-10-20 01:38:15
|
raindigi/site-preview
|
https://api.github.com/repos/raindigi/site-preview
|
closed
|
WS-2019-0424 (Medium) detected in elliptic-6.4.0.tgz - autoclosed
|
security vulnerability
|
## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-1.9.277.tgz (Root Library)
- node-libs-browser-2.1.0.tgz
- crypto-browserify-3.12.0.tgz
- create-ecdh-4.0.3.tgz
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/raindigi/site-preview/git/commits/5b58d9941528c1a41f80dabfe33e36195928235b">5b58d9941528c1a41f80dabfe33e36195928235b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0424">https://nvd.nist.gov/vuln/detail/WS-2019-0424</a></p>
<p>Release Date: 2019-11-13</p>
<p>Fix Resolution: GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105;Romano.Vue - 1.0.1;org.webjars.npm:elliptic - 6.5.4,6.3.3;VueJS.NetCore - 1.1.1;elliptic - 6.5.3;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;Fable.Template.Elmish.React - 0.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0424 (Medium) detected in elliptic-6.4.0.tgz - autoclosed - ## WS-2019-0424 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-1.9.277.tgz (Root Library)
- node-libs-browser-2.1.0.tgz
- crypto-browserify-3.12.0.tgz
- create-ecdh-4.0.3.tgz
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/raindigi/site-preview/git/commits/5b58d9941528c1a41f80dabfe33e36195928235b">5b58d9941528c1a41f80dabfe33e36195928235b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
all versions of elliptic are vulnerable to Timing Attack through side-channels.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0424">https://nvd.nist.gov/vuln/detail/WS-2019-0424</a></p>
<p>Release Date: 2019-11-13</p>
<p>Fix Resolution: GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105;Romano.Vue - 1.0.1;org.webjars.npm:elliptic - 6.5.4,6.3.3;VueJS.NetCore - 1.1.1;elliptic - 6.5.3;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;Fable.Template.Elmish.React - 0.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in elliptic tgz autoclosed ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file package json path to vulnerable library node modules elliptic package json dependency hierarchy gatsby tgz root library node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href vulnerability details all versions of elliptic are vulnerable to timing attack through side channels publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution gr pagerender razor midiator webclient romano vue org webjars npm elliptic vuejs netcore elliptic indianadavy vuejswebapitemplate csharp nordron angulartemplate corevuewebtest dotnetng template fable template elmish react step up your open source security game with whitesource
| 0
|
132,117
| 18,266,126,676
|
IssuesEvent
|
2021-10-04 08:40:02
|
artsking/linux-3.0.35_CVE-2020-15436_withPatch
|
https://api.github.com/repos/artsking/linux-3.0.35_CVE-2020-15436_withPatch
|
closed
|
CVE-2014-8159 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed
|
security vulnerability
|
## CVE-2014-8159 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35_CVE-2020-15436_withPatch/commit/87eecd735a2e4c02ba0c4dc61594d4311e35d5d9">87eecd735a2e4c02ba0c4dc61594d4311e35d5d9</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The InfiniBand (IB) implementation in the Linux kernel package before 2.6.32-504.12.2 on Red Hat Enterprise Linux (RHEL) 6 does not properly restrict use of User Verbs for registration of memory regions, which allows local users to access arbitrary physical memory locations, and consequently cause a denial of service (system crash) or gain privileges, by leveraging permissions on a uverbs device under /dev/infiniband/.
<p>Publish Date: 2015-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-8159>CVE-2014-8159</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2014-8159">https://www.linuxkernelcves.com/cves/CVE-2014-8159</a></p>
<p>Release Date: 2015-03-16</p>
<p>Fix Resolution: v3.2.69,v3.12.41,v3.14.39,v3.16.35,v3.18.13,v4.0-rc7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2014-8159 (Medium) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2014-8159 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-3.0.35_CVE-2020-15436_withPatch/commit/87eecd735a2e4c02ba0c4dc61594d4311e35d5d9">87eecd735a2e4c02ba0c4dc61594d4311e35d5d9</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The InfiniBand (IB) implementation in the Linux kernel package before 2.6.32-504.12.2 on Red Hat Enterprise Linux (RHEL) 6 does not properly restrict use of User Verbs for registration of memory regions, which allows local users to access arbitrary physical memory locations, and consequently cause a denial of service (system crash) or gain privileges, by leveraging permissions on a uverbs device under /dev/infiniband/.
<p>Publish Date: 2015-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-8159>CVE-2014-8159</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2014-8159">https://www.linuxkernelcves.com/cves/CVE-2014-8159</a></p>
<p>Release Date: 2015-03-16</p>
<p>Fix Resolution: v3.2.69,v3.12.41,v3.14.39,v3.16.35,v3.18.13,v4.0-rc7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the infiniband ib implementation in the linux kernel package before on red hat enterprise linux rhel does not properly restrict use of user verbs for registration of memory regions which allows local users to access arbitrary physical memory locations and consequently cause a denial of service system crash or gain privileges by leveraging permissions on a uverbs device under dev infiniband publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
53,856
| 13,262,394,566
|
IssuesEvent
|
2020-08-20 21:42:12
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
Missing symbolic link to data run (Trac #2201)
|
Migrated from Trac defect infrastructure
|
Hi,
there is a symbolic link missing for run 00131184.
/data/exp/IceCube/2018/filtered/level2/0619/Run00131184_73
does exist but
/data/exp/IceCube/2018/filtered/level2/0619/Run00131184
does not.
That would not be too annoying but the goodRun file references the /data/exp/IceCube/2018/filtered/level2/0619/Run00131184 folder.
As the other folders for 2018 seem to be mostly symlinked, I would like this folder to be symlinked as well.
Cheers,
Frederik
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2201">https://code.icecube.wisc.edu/projects/icecube/ticket/2201</a>, reported by flauberand owned by rsnihur</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:18:07",
"_ts": "1550067487593768",
"description": "Hi,\n\nthere is a symbolic link missing for run 00131184.\n/data/exp/IceCube/2018/filtered/level2/0619/Run00131184_73\ndoes exist but\n/data/exp/IceCube/2018/filtered/level2/0619/Run00131184\ndoes not.\n\nThat would not be too annoying but the goodRun file references the /data/exp/IceCube/2018/filtered/level2/0619/Run00131184 folder. \nAs the other folders for 2018 seem to be mostly symlinked, I would like this folder to be symlinked as well.\n\nCheers,\nFrederik\n\n",
"reporter": "flauber",
"cc": "",
"resolution": "wontfix",
"time": "2018-10-29T17:24:29",
"component": "infrastructure",
"summary": "Missing symbolic link to data run",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "rsnihur",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Missing symbolic link to data run (Trac #2201) - Hi,
there is a symbolic link missing for run 00131184.
/data/exp/IceCube/2018/filtered/level2/0619/Run00131184_73
does exist but
/data/exp/IceCube/2018/filtered/level2/0619/Run00131184
does not.
That would not be too annoying but the goodRun file references the /data/exp/IceCube/2018/filtered/level2/0619/Run00131184 folder.
As the other folders for 2018 seem to be mostly symlinked, I would like this folder to be symlinked as well.
Cheers,
Frederik
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2201">https://code.icecube.wisc.edu/projects/icecube/ticket/2201</a>, reported by flauberand owned by rsnihur</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:18:07",
"_ts": "1550067487593768",
"description": "Hi,\n\nthere is a symbolic link missing for run 00131184.\n/data/exp/IceCube/2018/filtered/level2/0619/Run00131184_73\ndoes exist but\n/data/exp/IceCube/2018/filtered/level2/0619/Run00131184\ndoes not.\n\nThat would not be too annoying but the goodRun file references the /data/exp/IceCube/2018/filtered/level2/0619/Run00131184 folder. \nAs the other folders for 2018 seem to be mostly symlinked, I would like this folder to be symlinked as well.\n\nCheers,\nFrederik\n\n",
"reporter": "flauber",
"cc": "",
"resolution": "wontfix",
"time": "2018-10-29T17:24:29",
"component": "infrastructure",
"summary": "Missing symbolic link to data run",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "rsnihur",
"type": "defect"
}
```
</p>
</details>
|
defect
|
missing symbolic link to data run trac hi there is a symbolic link missing for run data exp icecube filtered does exist but data exp icecube filtered does not that would not be too annoying but the goodrun file references the data exp icecube filtered folder as the other folders for seem to be mostly symlinked i would like this folder to be symlinked as well cheers frederik migrated from json status closed changetime ts description hi n nthere is a symbolic link missing for run n data exp icecube filtered ndoes exist but n data exp icecube filtered ndoes not n nthat would not be too annoying but the goodrun file references the data exp icecube filtered folder nas the other folders for seem to be mostly symlinked i would like this folder to be symlinked as well n ncheers nfrederik n n reporter flauber cc resolution wontfix time component infrastructure summary missing symbolic link to data run priority normal keywords milestone owner rsnihur type defect
| 1
|
32,410
| 12,126,069,238
|
IssuesEvent
|
2020-04-22 16:26:39
|
mixcore/mix.core
|
https://api.github.com/repos/mixcore/mix.core
|
closed
|
CVE-2018-20822 (Medium) detected in node-sass-4.13.0.tgz, node-sass-0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70
|
security vulnerability
|
## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/mix.core/src/WebApplication1/ClientApp/package.json</p>
<p>Path to vulnerable library: /mix.core/src/WebApplication1/ClientApp/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mixcore/mix.core/commit/bd312c87103111cf005a36a25823c65e1c23669f">bd312c87103111cf005a36a25823c65e1c23669f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20822 (Medium) detected in node-sass-4.13.0.tgz, node-sass-0bd48bbad6fccb0da16d3bdf76ad541f5f45ec70 - ## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/mix.core/src/WebApplication1/ClientApp/package.json</p>
<p>Path to vulnerable library: /mix.core/src/WebApplication1/ClientApp/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mixcore/mix.core/commit/bd312c87103111cf005a36a25823c65e1c23669f">bd312c87103111cf005a36a25823c65e1c23669f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in node sass tgz node sass cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm mix core src clientapp package json path to vulnerable library mix core src clientapp node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details libsass allows attackers to cause a denial of service uncontrolled recursion in sass complex selector perform in ast hpp and sass inspect operator in inspect cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
| 0
|
35,999
| 14,916,040,725
|
IssuesEvent
|
2021-01-22 17:35:39
|
hashicorp/terraform-provider-aws
|
https://api.github.com/repos/hashicorp/terraform-provider-aws
|
closed
|
No Certificate found for this domain issue
|
service/acm waiting-response
|
<!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Terraform Version
0.12.29
Terraform Provider version
2.70.0
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
aws_acm_certificate_validation
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
resource "aws_acm_certificate" "cert" {
domain_name = "example123.com"
validation_method = "DNS"
tags = {
Name = "certificate_01"
}
}
data "aws_acm_certificate" "domain_cert" {
domain = "example123.com"
types = ["AMAZON_ISSUED"]
statuses = ["PENDING_VALIDATION"]
}
resource "aws_acm_certificate_validation" "cert_validation" {
certificate_arn = data.aws_acm_certificate.domain_cert.arn
}
### Debug Output
### Expected Behavior
<!--- What should have happened? --->
DataSource should return arn value for certificate_arn parameter
### Actual Behavior
<!--- What actually happened? --->
Error: No certificate for domain "example123.com" found in this region
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Create the certificate using aws_acm_certificate resource.
2. Once the certificate is created, it will be in **PENDING_VALIDATION** state, now perform terraform plan/apply using
aws_acm_certificate_validation resource.
|
1.0
|
No Certificate found for this domain issue - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Terraform Version
0.12.29
Terraform Provider version
2.70.0
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
aws_acm_certificate_validation
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
resource "aws_acm_certificate" "cert" {
domain_name = "example123.com"
validation_method = "DNS"
tags = {
Name = "certificate_01"
}
}
data "aws_acm_certificate" "domain_cert" {
domain = "example123.com"
types = ["AMAZON_ISSUED"]
statuses = ["PENDING_VALIDATION"]
}
resource "aws_acm_certificate_validation" "cert_validation" {
certificate_arn = data.aws_acm_certificate.domain_cert.arn
}
### Debug Output
### Expected Behavior
<!--- What should have happened? --->
DataSource should return arn value for certificate_arn parameter
### Actual Behavior
<!--- What actually happened? --->
Error: No certificate for domain "example123.com" found in this region
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Create the certificate using aws_acm_certificate resource.
2. Once the certificate is created, it will be in **PENDING_VALIDATION** state, now perform terraform plan/apply using
aws_acm_certificate_validation resource.
|
non_defect
|
no certificate found for this domain issue please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version terraform version terraform provider version affected resource s aws acm certificate validation terraform configuration files resource aws acm certificate cert domain name com validation method dns tags name certificate data aws acm certificate domain cert domain com types statuses resource aws acm certificate validation cert validation certificate arn data aws acm certificate domain cert arn debug output expected behavior datasource should return arn value for certificate arn parameter actual behavior error no certificate for domain com found in this region steps to reproduce create the certificate using aws acm certificate resource once the certificate is created it will be in pending validation state now perform terraform plan apply using aws acm certificate validation resource
| 0
|
9,051
| 2,615,125,497
|
IssuesEvent
|
2015-03-01 05:53:37
|
chrsmith/google-api-java-client
|
https://api.github.com/repos/chrsmith/google-api-java-client
|
closed
|
picasa-android-sample does not work with command-line mvn
|
auto-migrated Milestone-Version1.9.0 Priority-Medium Type-Defect
|
```
Version of google-api-java-client (e.g. 1.8.0-beta)?
1.8.0-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
Describe the problem.
I follow the instructions to compile picasa-android-sample using mvn ( 3.0.4)
and it gives the following error message about the syntax of pom.xml after 'mvn
compile' (Note: I follow the instruction and uncommented the pom.xml
accordingly. I had tasks-android-sample working by following the instructions,
though failed with the picasa one. The repository version is : f1001d3e79f0 for
https://code.google.com/p/google-api-java-client.samples/
[INFO] Scanning for projects...
[ERROR] The build could not read 1 project -> [Help 1]
[ERROR]
[ERROR] The project com.google.apis-samples:picasa-android-sample:1.0.0
(/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml)
has 1 error
[ERROR] Malformed POM
/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml:
Unrecognised tag: 'exclusion' (position: START_TAG seen ...</version>\n
<exclusion>... @328:18) @
/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml,
line 328, column 18 -> [Help 2]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/ProjectBuildingException
[ERROR] [Help 2]
http://cwiki.apache.org/confluence/display/MAVEN/ModelParseException
How would you expect it to be fixed?
Have the sample work to get started with picasa access api.
```
Original issue reported on code.google.com by `jiang.long@gmail.com` on 5 Apr 2012 at 5:14
|
1.0
|
picasa-android-sample does not work with command-line mvn - ```
Version of google-api-java-client (e.g. 1.8.0-beta)?
1.8.0-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
Describe the problem.
I follow the instructions to compile picasa-android-sample using mvn ( 3.0.4)
and it gives the following error message about the syntax of pom.xml after 'mvn
compile' (Note: I follow the instruction and uncommented the pom.xml
accordingly. I had tasks-android-sample working by following the instructions,
though failed with the picasa one. The repository version is : f1001d3e79f0 for
https://code.google.com/p/google-api-java-client.samples/
[INFO] Scanning for projects...
[ERROR] The build could not read 1 project -> [Help 1]
[ERROR]
[ERROR] The project com.google.apis-samples:picasa-android-sample:1.0.0
(/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml)
has 1 error
[ERROR] Malformed POM
/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml:
Unrecognised tag: 'exclusion' (position: START_TAG seen ...</version>\n
<exclusion>... @328:18) @
/home/fll/work/google-api-java-client-samples/picasa-android-sample/pom.xml,
line 328, column 18 -> [Help 2]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/ProjectBuildingException
[ERROR] [Help 2]
http://cwiki.apache.org/confluence/display/MAVEN/ModelParseException
How would you expect it to be fixed?
Have the sample work to get started with picasa access api.
```
Original issue reported on code.google.com by `jiang.long@gmail.com` on 5 Apr 2012 at 5:14
|
defect
|
picasa android sample does not work with command line mvn version of google api java client e g beta beta java environment e g java android app engine java describe the problem i follow the instructions to compile picasa android sample using mvn and it gives the following error message about the syntax of pom xml after mvn compile note i follow the instruction and uncommented the pom xml accordingly i had tasks android sample working by following the instructions though failed with the picasa one the repository version is for scanning for projects the build could not read project the project com google apis samples picasa android sample home fll work google api java client samples picasa android sample pom xml has error malformed pom home fll work google api java client samples picasa android sample pom xml unrecognised tag exclusion position start tag seen n home fll work google api java client samples picasa android sample pom xml line column to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles how would you expect it to be fixed have the sample work to get started with picasa access api original issue reported on code google com by jiang long gmail com on apr at
| 1
|
120,442
| 17,644,197,157
|
IssuesEvent
|
2021-08-20 01:55:53
|
logbie/HyperGAN
|
https://api.github.com/repos/logbie/HyperGAN
|
opened
|
CVE-2021-37692 (Medium) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl
|
security vulnerability
|
## CVE-2021-37692 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: HyperGAN/requirements.txt</p>
<p>Path to vulnerable library: HyperGAN/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions under certain conditions, Go code can trigger a segfault in string deallocation. For string tensors, `C.TF_TString_Dealloc` is called during garbage collection within a finalizer function. However, tensor structure isn't checked until encoding to avoid a performance penalty. The current method for dealloc assumes that encoding succeeded, but segfaults when a string tensor is garbage collected whose encoding failed (e.g., due to mismatched dimensions). To fix this, the call to set the finalizer function is deferred until `NewTensor` returns and, if encoding failed for a string tensor, deallocs are determined based on bytes written. We have patched the issue in GitHub commit 8721ba96e5760c229217b594f6d2ba332beedf22. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, which is the other affected version.
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37692>CVE-2021-37692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cmgw-8vpc-rc59">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cmgw-8vpc-rc59</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: tensorflow - 2.5.1, 2.6.0, tensorflow-cpu - 2.5.1, 2.6.0, tensorflow-gpu - 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-37692 (Medium) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl - ## CVE-2021-37692 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: HyperGAN/requirements.txt</p>
<p>Path to vulnerable library: HyperGAN/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions under certain conditions, Go code can trigger a segfault in string deallocation. For string tensors, `C.TF_TString_Dealloc` is called during garbage collection within a finalizer function. However, tensor structure isn't checked until encoding to avoid a performance penalty. The current method for dealloc assumes that encoding succeeded, but segfaults when a string tensor is garbage collected whose encoding failed (e.g., due to mismatched dimensions). To fix this, the call to set the finalizer function is deferred until `NewTensor` returns and, if encoding failed for a string tensor, deallocs are determined based on bytes written. We have patched the issue in GitHub commit 8721ba96e5760c229217b594f6d2ba332beedf22. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, which is the other affected version.
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37692>CVE-2021-37692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cmgw-8vpc-rc59">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cmgw-8vpc-rc59</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: tensorflow - 2.5.1, 2.6.0, tensorflow-cpu - 2.5.1, 2.6.0, tensorflow-gpu - 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in tensorflow gpu whl cve medium severity vulnerability vulnerable library tensorflow gpu whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file hypergan requirements txt path to vulnerable library hypergan requirements txt dependency hierarchy x tensorflow gpu whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning in affected versions under certain conditions go code can trigger a segfault in string deallocation for string tensors c tf tstring dealloc is called during garbage collection within a finalizer function however tensor structure isn t checked until encoding to avoid a performance penalty the current method for dealloc assumes that encoding succeeded but segfaults when a string tensor is garbage collected whose encoding failed e g due to mismatched dimensions to fix this the call to set the finalizer function is deferred until newtensor returns and if encoding failed for a string tensor deallocs are determined based on bytes written we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow which is the other affected version publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
5,522
| 2,610,189,236
|
IssuesEvent
|
2015-02-26 18:59:55
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
介绍遗传性色斑怎样治
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
回首,恍然间似梦如幻。轮回的四季,美丽依旧,不知,心��
�否依旧。轻轻坲过逝去的故事,转身,任思念的伤在九月微�
��的风里,寂寞盛开。夜半无语,思忆飞扬,挑起半点忧伤,
装饰整个夜晚。为谁无眠,心中自愁,不曾哭泣,唯觉千行��
�,虽然,很累很累,却没有心累,总是努力让心静静休息,�
��知它太任性,就想在深夜徘徊,寻找着眼睛里遗留的那滴泪
珠……聆听着同一首音乐,一段朦胧的故事紧紧抱着孤单的��
�,一个童话靠在我冰冷的肩旁上,那么熟悉,我深深感觉到�
��一次呼吸里都充满了思念的韵味。它们抬头凝望着我,眼神
是那么的无力,很怕会在脆弱的瞬间逝去了那一束残弱的余��
�,相对无言,却能读懂彼此的心语。其实好想让你莫要流泪�
��只怕藏在心底会跟难受,你的眼角终于有了一滴泪珠,湿润
了整个岑寂的深夜。
一簟清秋,追不上消逝的年轮,风的季节里,感觉不到那股��
�去的飒爽,回忆里装满了两个月的悲欢离合,从未想过短暂�
��会有无言的告别,意念与远走,从此杳无音讯,黑夜中的你
是否还在哭泣,记得那首为你写的诗,而今验证了诗中的预��
�:“暮色吻风男儿心若云,晓岚逐烟伊人愁似水。去年舞榭�
��台烟花起,今夕窀穸石前白草依。朱红依旧红颜只是谁,谢
意如歌浪子为谁怜。艳阳思伤挥洒雨中情,悠悠之水逐尽心��
�愁。身随寒江罡风几棵松,心归何处今朝已枉然。”,其实�
��初应该在乎这短短的几句话,那是属于彼此不为人知的秘密
诗词。从别后,未相逢,也许是永远,最后一眼的对视太短��
�,彼此的泪水挡住了彼此的视线,假如爱有天意,让时间停�
��在一束流光之外,我愿意去承载岁月忧伤。我不知道是否会
有人在下一站等这我,但是现在长满色斑的我,是否还有爱��
�权利!遗传性色斑怎样治,
《客户案例》
我是我们公司的销售经理,这个工作就是天天要在外面��
�的,经常要见客户什么的,其实工作压力也挺大的,再加上�
��龄也越来越大了,皮肤就大不如从前了。一脸的黄褐斑,其
实女人么都是爱美的,没办法就只能靠美容啊,用高档化妆��
�来弥补。有一段时间听朋友说有一种化妆品特别好,说是可�
��让皮肤变的白而且还可以去掉我脸上的黄褐斑,我看人家用
的都挺好的,我就用了,哎!没想到这东西对我不管用,斑是�
��长越多啊,真是上火死我了。真是想美一下还没美成,还成
了大花脸了,以后真的是不敢再瞎用了。但是心里也急啊,��
�底如何去黄褐斑啊?当时用了一些淡斑的祛斑霜,也没见有什
么作用,后来也不用了。</br>
有天回家,妈妈给我买回了黛芙薇尔,说是她朋友用过��
�都说祛斑美白效果特别好,开始我还不相信,就去官网问了�
��下他们的专家,专家说是法国、美国、台湾三地的专家通力
协作,通过生物高科技手段,全面修复老化、病变细胞,彻��
�逆转受损基因,从而达到了激活肌肤细胞分裂再生、增强细�
��活性、强效美白的功效,特别是暗沉、干燥、暗黄、粗糙、
肤色不匀的肌肤。于是我就放下心来使用。</br>
使用完第一个周期的时候感觉皮肤润润的,感觉黛芙薇��
�祛斑还是很好的。第二个周期斑就开始淡了,看着我脸上的�
��化,老公对对我的太多也慢慢变化了。我又重新体会到了家
的温馨。后面2个周期以老客户的身份定的。脸上大面积的斑�
��有大变小,由深变浅,直至消失,由衷的感谢黛芙薇尔。
阅读了遗传性色斑怎样治,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
遗传性色斑怎样治,同时为您分享祛斑小方法
1、米醋浸白术,每日搽面。
2、生姜50克(干姜减半)入500毫升50%酒精中密闭浸泡15天,�
��晚擦于洗净的患处。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:41
|
1.0
|
介绍遗传性色斑怎样治 - ```
《摘要》
回首,恍然间似梦如幻。轮回的四季,美丽依旧,不知,心��
�否依旧。轻轻坲过逝去的故事,转身,任思念的伤在九月微�
��的风里,寂寞盛开。夜半无语,思忆飞扬,挑起半点忧伤,
装饰整个夜晚。为谁无眠,心中自愁,不曾哭泣,唯觉千行��
�,虽然,很累很累,却没有心累,总是努力让心静静休息,�
��知它太任性,就想在深夜徘徊,寻找着眼睛里遗留的那滴泪
珠……聆听着同一首音乐,一段朦胧的故事紧紧抱着孤单的��
�,一个童话靠在我冰冷的肩旁上,那么熟悉,我深深感觉到�
��一次呼吸里都充满了思念的韵味。它们抬头凝望着我,眼神
是那么的无力,很怕会在脆弱的瞬间逝去了那一束残弱的余��
�,相对无言,却能读懂彼此的心语。其实好想让你莫要流泪�
��只怕藏在心底会跟难受,你的眼角终于有了一滴泪珠,湿润
了整个岑寂的深夜。
一簟清秋,追不上消逝的年轮,风的季节里,感觉不到那股��
�去的飒爽,回忆里装满了两个月的悲欢离合,从未想过短暂�
��会有无言的告别,意念与远走,从此杳无音讯,黑夜中的你
是否还在哭泣,记得那首为你写的诗,而今验证了诗中的预��
�:“暮色吻风男儿心若云,晓岚逐烟伊人愁似水。去年舞榭�
��台烟花起,今夕窀穸石前白草依。朱红依旧红颜只是谁,谢
意如歌浪子为谁怜。艳阳思伤挥洒雨中情,悠悠之水逐尽心��
�愁。身随寒江罡风几棵松,心归何处今朝已枉然。”,其实�
��初应该在乎这短短的几句话,那是属于彼此不为人知的秘密
诗词。从别后,未相逢,也许是永远,最后一眼的对视太短��
�,彼此的泪水挡住了彼此的视线,假如爱有天意,让时间停�
��在一束流光之外,我愿意去承载岁月忧伤。我不知道是否会
有人在下一站等这我,但是现在长满色斑的我,是否还有爱��
�权利!遗传性色斑怎样治,
《客户案例》
我是我们公司的销售经理,这个工作就是天天要在外面��
�的,经常要见客户什么的,其实工作压力也挺大的,再加上�
��龄也越来越大了,皮肤就大不如从前了。一脸的黄褐斑,其
实女人么都是爱美的,没办法就只能靠美容啊,用高档化妆��
�来弥补。有一段时间听朋友说有一种化妆品特别好,说是可�
��让皮肤变的白而且还可以去掉我脸上的黄褐斑,我看人家用
的都挺好的,我就用了,哎!没想到这东西对我不管用,斑是�
��长越多啊,真是上火死我了。真是想美一下还没美成,还成
了大花脸了,以后真的是不敢再瞎用了。但是心里也急啊,��
�底如何去黄褐斑啊?当时用了一些淡斑的祛斑霜,也没见有什
么作用,后来也不用了。</br>
有天回家,妈妈给我买回了黛芙薇尔,说是她朋友用过��
�都说祛斑美白效果特别好,开始我还不相信,就去官网问了�
��下他们的专家,专家说是法国、美国、台湾三地的专家通力
协作,通过生物高科技手段,全面修复老化、病变细胞,彻��
�逆转受损基因,从而达到了激活肌肤细胞分裂再生、增强细�
��活性、强效美白的功效,特别是暗沉、干燥、暗黄、粗糙、
肤色不匀的肌肤。于是我就放下心来使用。</br>
使用完第一个周期的时候感觉皮肤润润的,感觉黛芙薇��
�祛斑还是很好的。第二个周期斑就开始淡了,看着我脸上的�
��化,老公对对我的太多也慢慢变化了。我又重新体会到了家
的温馨。后面2个周期以老客户的身份定的。脸上大面积的斑�
��有大变小,由深变浅,直至消失,由衷的感谢黛芙薇尔。
阅读了遗传性色斑怎样治,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
遗传性色斑怎样治,同时为您分享祛斑小方法
1、米醋浸白术,每日搽面。
2、生姜50克(干姜减半)入500毫升50%酒精中密闭浸泡15天,�
��晚擦于洗净的患处。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:41
|
defect
|
介绍遗传性色斑怎样治 《摘要》 回首,恍然间似梦如幻。轮回的四季,美丽依旧,不知,心�� �否依旧。轻轻坲过逝去的故事,转身,任思念的伤在九月微� ��的风里,寂寞盛开。夜半无语,思忆飞扬,挑起半点忧伤, 装饰整个夜晚。为谁无眠,心中自愁,不曾哭泣,唯觉千行�� �,虽然,很累很累,却没有心累,总是努力让心静静休息,� ��知它太任性,就想在深夜徘徊,寻找着眼睛里遗留的那滴泪 珠……聆听着同一首音乐,一段朦胧的故事紧紧抱着孤单的�� �,一个童话靠在我冰冷的肩旁上,那么熟悉,我深深感觉到� ��一次呼吸里都充满了思念的韵味。它们抬头凝望着我,眼神 是那么的无力,很怕会在脆弱的瞬间逝去了那一束残弱的余�� �,相对无言,却能读懂彼此的心语。其实好想让你莫要流泪� ��只怕藏在心底会跟难受,你的眼角终于有了一滴泪珠,湿润 了整个岑寂的深夜。 一簟清秋,追不上消逝的年轮,风的季节里,感觉不到那股�� �去的飒爽,回忆里装满了两个月的悲欢离合,从未想过短暂� ��会有无言的告别,意念与远走,从此杳无音讯,黑夜中的你 是否还在哭泣,记得那首为你写的诗,而今验证了诗中的预�� �:“暮色吻风男儿心若云,晓岚逐烟伊人愁似水。去年舞榭� ��台烟花起,今夕窀穸石前白草依。朱红依旧红颜只是谁,谢 意如歌浪子为谁怜。艳阳思伤挥洒雨中情,悠悠之水逐尽心�� �愁。身随寒江罡风几棵松,心归何处今朝已枉然。”,其实� ��初应该在乎这短短的几句话,那是属于彼此不为人知的秘密 诗词。从别后,未相逢,也许是永远,最后一眼的对视太短�� �,彼此的泪水挡住了彼此的视线,假如爱有天意,让时间停� ��在一束流光之外,我愿意去承载岁月忧伤。我不知道是否会 有人在下一站等这我,但是现在长满色斑的我,是否还有爱�� �权利!遗传性色斑怎样治, 《客户案例》 我是我们公司的销售经理,这个工作就是天天要在外面�� �的,经常要见客户什么的,其实工作压力也挺大的,再加上� ��龄也越来越大了,皮肤就大不如从前了。一脸的黄褐斑,其 实女人么都是爱美的,没办法就只能靠美容啊,用高档化妆�� �来弥补。有一段时间听朋友说有一种化妆品特别好,说是可� ��让皮肤变的白而且还可以去掉我脸上的黄褐斑,我看人家用 的都挺好的,我就用了,哎 没想到这东西对我不管用,斑是� ��长越多啊,真是上火死我了。真是想美一下还没美成,还成 了大花脸了,以后真的是不敢再瞎用了。但是心里也急啊,�� �底如何去黄褐斑啊 当时用了一些淡斑的祛斑霜,也没见有什 么作用,后来也不用了。 有天回家,妈妈给我买回了黛芙薇尔,说是她朋友用过�� �都说祛斑美白效果特别好,开始我还不相信,就去官网问了� ��下他们的专家,专家说是法国、美国、台湾三地的专家通力 协作,通过生物高科技手段,全面修复老化、病变细胞,彻�� �逆转受损基因,从而达到了激活肌肤细胞分裂再生、增强细� ��活性、强效美白的功效,特别是暗沉、干燥、暗黄、粗糙、 肤色不匀的肌肤。于是我就放下心来使用。 使用完第一个周期的时候感觉皮肤润润的,感觉黛芙薇�� �祛斑还是很好的。第二个周期斑就开始淡了,看着我脸上的� ��化,老公对对我的太多也慢慢变化了。我又重新体会到了家 的温馨。 。脸上大面积的斑� ��有大变小,由深变浅,直至消失,由衷的感谢黛芙薇尔。 阅读了遗传性色斑怎样治,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 遗传性色斑怎样治,同时为您分享祛斑小方法 、米醋浸白术,每日搽面。 、 (干姜减半) % ,� ��晚擦于洗净的患处。 original issue reported on code google com by additive gmail com on jul at
| 1
|
17,741
| 3,012,950,227
|
IssuesEvent
|
2015-07-29 04:33:38
|
yawlfoundation/yawl
|
https://api.github.com/repos/yawlfoundation/yawl
|
closed
|
minOccurs and maxOccurs don't work correct
|
auto-migrated Component-Engine OpSys-All Priority-High Type-Defect
|
```
What steps will reproduce the problem?
1. start the enclosed spec
2. By ticking on 'further' the workitem gets back to 'Enter notice' task
3. The boolean 'further' disappears after the workitem ended more than once
at task 'Enter notice'
What is the expected output? What do you see instead?
'further' should always be displayed.
```
Original issue reported on code.google.com by `stephan....@googlemail.com` on 18 Feb 2009 at 9:49
Attachments:
* [test2.yawl](https://storage.googleapis.com/google-code-attachments/yawl/issue-262/comment-0/test2.yawl)
* [YAWLOrgDataExport.ybkp](https://storage.googleapis.com/google-code-attachments/yawl/issue-262/comment-0/YAWLOrgDataExport.ybkp)
|
1.0
|
minOccurs and maxOccurs don't work correct - ```
What steps will reproduce the problem?
1. start the enclosed spec
2. By ticking on 'further' the workitem gets back to 'Enter notice' task
3. The boolean 'further' disappears after the workitem ended more than once
at task 'Enter notice'
What is the expected output? What do you see instead?
'further' should always be displayed.
```
Original issue reported on code.google.com by `stephan....@googlemail.com` on 18 Feb 2009 at 9:49
Attachments:
* [test2.yawl](https://storage.googleapis.com/google-code-attachments/yawl/issue-262/comment-0/test2.yawl)
* [YAWLOrgDataExport.ybkp](https://storage.googleapis.com/google-code-attachments/yawl/issue-262/comment-0/YAWLOrgDataExport.ybkp)
|
defect
|
minoccurs and maxoccurs don t work correct what steps will reproduce the problem start the enclosed spec by ticking on further the workitem gets back to enter notice task the boolean further disappears after the workitem ended more than once at task enter notice what is the expected output what do you see instead further should always be displayed original issue reported on code google com by stephan googlemail com on feb at attachments
| 1
|
206,203
| 23,367,611,078
|
IssuesEvent
|
2022-08-10 16:42:48
|
turkdevops/codecov-action
|
https://api.github.com/repos/turkdevops/codecov-action
|
closed
|
CVE-2021-44906 (Medium) detected in minimist-1.2.5.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- ts-jest-26.4.4.tgz (Root Library)
- json5-2.1.3.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/2484da9b2f1f466b4e954c828ba2397a23e54652">2484da9b2f1f466b4e954c828ba2397a23e54652</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution (minimist): 1.2.6</p>
<p>Direct dependency fix Resolution (ts-jest): 26.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-44906 (Medium) detected in minimist-1.2.5.tgz - autoclosed - ## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- ts-jest-26.4.4.tgz (Root Library)
- json5-2.1.3.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/2484da9b2f1f466b4e954c828ba2397a23e54652">2484da9b2f1f466b4e954c828ba2397a23e54652</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution (minimist): 1.2.6</p>
<p>Direct dependency fix Resolution (ts-jest): 26.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in minimist tgz autoclosed cve medium severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file package json path to vulnerable library node modules minimist package json dependency hierarchy ts jest tgz root library tgz x minimist tgz vulnerable library found in head commit a href found in base branch master vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist direct dependency fix resolution ts jest step up your open source security game with mend
| 0
|
21,090
| 3,460,525,139
|
IssuesEvent
|
2015-12-19 07:24:51
|
ariya/phantomjs
|
https://api.github.com/repos/ariya/phantomjs
|
closed
|
Crashes on specific site.
|
old.Priority-Medium old.Status-New old.Type-Defect
|
_**[m...@easel.io](http://code.google.com/u/103546247535701456586/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
> <b>What steps will reproduce the problem?</b>
1. Visit a specific site.
>
> <b>What is the expected output? What do you see instead?</b>
Crashes.
>
> <b>Which operating system are you using?</b>
OSX 10.7 and Linux
>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
Compiled from source (1.6.1)
>
> <b>Please provide any additional information below.</b>
Stacktrace attached, unfortunately I can't provide the site itself.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #698](http://code.google.com/p/phantomjs/issues/detail?id=698).
:star2: **2** people had starred this issue at the time of migration.
|
1.0
|
Crashes on specific site. - _**[m...@easel.io](http://code.google.com/u/103546247535701456586/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
> <b>What steps will reproduce the problem?</b>
1. Visit a specific site.
>
> <b>What is the expected output? What do you see instead?</b>
Crashes.
>
> <b>Which operating system are you using?</b>
OSX 10.7 and Linux
>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
Compiled from source (1.6.1)
>
> <b>Please provide any additional information below.</b>
Stacktrace attached, unfortunately I can't provide the site itself.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #698](http://code.google.com/p/phantomjs/issues/detail?id=698).
:star2: **2** people had starred this issue at the time of migration.
|
defect
|
crashes on specific site commented which version of phantomjs are you using tip run phantomjs version what steps will reproduce the problem visit a specific site what is the expected output what do you see instead crashes which operating system are you using osx and linux did you use binary phantomjs or did you compile it from source compiled from source please provide any additional information below stacktrace attached unfortunately i can t provide the site itself disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration
| 1
|
54,400
| 13,645,909,047
|
IssuesEvent
|
2020-09-25 21:49:29
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
ZTS: l2arc/persist_l2arc_007_pos failure
|
Component: Test Suite Type: Defect
|
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Any
Distribution Version | Any
Linux Kernel | Any
Architecture | x86_64
ZFS Version | zfs-2.0.0-rc1-93-gcf26677
### Describe the problem you're observing
```
Tests with results other than PASS that are unexpected:
FAIL l2arc/persist_l2arc_007_pos (expected PASS)
```
### Describe how to reproduce the problem
Occasionally reproduced by the CI during testing.
### Include any warning/errors/backtraces from the system logs
http://build.zfsonlinux.org/builders/CentOS%207%20x86_64%20%28TEST%29/builds/13298/steps/shell_9/logs/summary
|
1.0
|
ZTS: l2arc/persist_l2arc_007_pos failure - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Any
Distribution Version | Any
Linux Kernel | Any
Architecture | x86_64
ZFS Version | zfs-2.0.0-rc1-93-gcf26677
### Describe the problem you're observing
```
Tests with results other than PASS that are unexpected:
FAIL l2arc/persist_l2arc_007_pos (expected PASS)
```
### Describe how to reproduce the problem
Occasionally reproduced by the CI during testing.
### Include any warning/errors/backtraces from the system logs
http://build.zfsonlinux.org/builders/CentOS%207%20x86_64%20%28TEST%29/builds/13298/steps/shell_9/logs/summary
|
defect
|
zts persist pos failure system information type version name distribution name any distribution version any linux kernel any architecture zfs version zfs describe the problem you re observing tests with results other than pass that are unexpected fail persist pos expected pass describe how to reproduce the problem occasionally reproduced by the ci during testing include any warning errors backtraces from the system logs
| 1
|
71,014
| 23,411,112,662
|
IssuesEvent
|
2022-08-12 17:38:13
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
508-defect-1: Benefit service accordion has critical heading, labelling, and responsive defects
|
508/Accessibility 508-defect-1 508-issue-headings Letters benefits-crew benefits-team-1 squad-1
|
### Point of contact
Josh Kim
### Severity level
1, Launchblocker. Critical. Must be fixed before launch.
### Details
### Headings aren't logically ordered
"Choose the information..." is a `h3` despite its parent being a `h3`; depending on intent, this may not be a heading at all.
<img width="1537" alt="headings" src="https://user-images.githubusercontent.com/14154792/184410904-effee184-6438-4033-ba4a-ba6a62bc5e82.png">
### No group labels
Several labelling issues exist with the way this accordion was designed:
- "Choose the information" isn't a `legend` within a `fieldset` nor a group label for the checkboxes below; screen reader users may critically miss important context on the purpose of those checkboxes as the key instruction "Choose the information" is not programmatically linked to the checkboxes.
- Similar to the rest of this page, the button is generically labelled
### Table isn't responsive
The first table is not responsive and doesn't follow VA design system guidance
<img width="740" alt="responsivetables" src="https://user-images.githubusercontent.com/14154792/184412722-c1ae3049-6bb7-41fc-8bcb-8e289d2cca41.png">
### Reproduction steps
1. [Log in as user 299 and get to step 2](https://staging.va.gov/records/download-va-letters/letters/letter-list)
2. Open the benefit summary accordion
### Proposed solution or next steps
TBD, this requires further exploration with design, engineering, and content.
### References, articles, or WCAG support
_No response_
### Type of issue
- [ ] Screenreader
- [ ] Keyboard
- [ ] Focus
- [ ] Headings
- [ ] Color
- [ ] Zoom
- [ ] Semantics
- [ ] Axe-core
- [ ] Something else
|
1.0
|
508-defect-1: Benefit service accordion has critical heading, labelling, and responsive defects - ### Point of contact
Josh Kim
### Severity level
1, Launchblocker. Critical. Must be fixed before launch.
### Details
### Headings aren't logically ordered
"Choose the information..." is a `h3` despite its parent being a `h3`; depending on intent, this may not be a heading at all.
<img width="1537" alt="headings" src="https://user-images.githubusercontent.com/14154792/184410904-effee184-6438-4033-ba4a-ba6a62bc5e82.png">
### No group labels
Several labelling issues exist with the way this accordion was designed:
- "Choose the information" isn't a `legend` within a `fieldset` nor a group label for the checkboxes below; screen reader users may critically miss important context on the purpose of those checkboxes as the key instruction "Choose the information" is not programmatically linked to the checkboxes.
- Similar to the rest of this page, the button is generically labelled
### Table isn't responsive
The first table is not responsive and doesn't follow VA design system guidance
<img width="740" alt="responsivetables" src="https://user-images.githubusercontent.com/14154792/184412722-c1ae3049-6bb7-41fc-8bcb-8e289d2cca41.png">
### Reproduction steps
1. [Log in as user 299 and get to step 2](https://staging.va.gov/records/download-va-letters/letters/letter-list)
2. Open the benefit summary accordion
### Proposed solution or next steps
TBD, this requires further exploration with design, engineering, and content.
### References, articles, or WCAG support
_No response_
### Type of issue
- [ ] Screenreader
- [ ] Keyboard
- [ ] Focus
- [ ] Headings
- [ ] Color
- [ ] Zoom
- [ ] Semantics
- [ ] Axe-core
- [ ] Something else
|
defect
|
defect benefit service accordion has critical heading labelling and responsive defects point of contact josh kim severity level launchblocker critical must be fixed before launch details headings aren t logically ordered choose the information is a despite its parent being a depending on intent this may not be a heading at all img width alt headings src no group labels several labelling issues exist with the way this accordion was designed choose the information isn t a legend within a fieldset nor a group label for the checkboxes below screen reader users may critically miss important context on the purpose of those checkboxes as the key instruction choose the information is not programmatically linked to the checkboxes similar to the rest of this page the button is generically labelled table isn t responsive the first table is not responsive and doesn t follow va design system guidance img width alt responsivetables src reproduction steps open the benefit summary accordion proposed solution or next steps tbd this requires further exploration with design engineering and content references articles or wcag support no response type of issue screenreader keyboard focus headings color zoom semantics axe core something else
| 1
|
366,354
| 10,819,676,898
|
IssuesEvent
|
2019-11-08 14:52:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
za.yahoo.com - see bug description
|
browser-firefox-mobile engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.1.2; Mobile; rv:65.0) Gecko/65.0 Firefox/65.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://za.yahoo.com/?p=us
**Browser / Version**: Firefox Mobile 65.0
**Operating System**: Android 4.1.2
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Ghostery asks to remember my login password and when I say yes it just carries on repeating the question.
**Steps to Reproduce**:
As above.
[](https://webcompat.com/uploads/2019/10/f82d4451-ba46-4a56-ac57-459a4adfc809.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190411133757</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
['[console.log(DARLA notice: 450) https://s.yimg.com/rq/darla/3-20-0/js/g-r-min.js:3:22566]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
za.yahoo.com - see bug description - <!-- @browser: Firefox Mobile 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 4.1.2; Mobile; rv:65.0) Gecko/65.0 Firefox/65.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://za.yahoo.com/?p=us
**Browser / Version**: Firefox Mobile 65.0
**Operating System**: Android 4.1.2
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Ghostery asks to remember my login password and when I say yes it just carries on repeating the question.
**Steps to Reproduce**:
As above.
[](https://webcompat.com/uploads/2019/10/f82d4451-ba46-4a56-ac57-459a4adfc809.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190411133757</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
['[console.log(DARLA notice: 450) https://s.yimg.com/rq/darla/3-20-0/js/g-r-min.js:3:22566]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
za yahoo com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description ghostery asks to remember my login password and when i say yes it just carries on repeating the question steps to reproduce as above browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false console messages from with ❤️
| 0
|
3,456
| 2,610,063,094
|
IssuesEvent
|
2015-02-26 18:18:32
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
黄岩治不育去哪里比较好
|
auto-migrated Priority-Medium Type-Defect
|
```
黄岩治不育去哪里比较好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:47
|
1.0
|
黄岩治不育去哪里比较好 - ```
黄岩治不育去哪里比较好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:47
|
defect
|
黄岩治不育去哪里比较好 黄岩治不育去哪里比较好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
385,590
| 26,644,852,124
|
IssuesEvent
|
2023-01-25 09:10:45
|
albert-marrero/Yu-Gi-Oh-Scraper
|
https://api.github.com/repos/albert-marrero/Yu-Gi-Oh-Scraper
|
closed
|
Fix broken contributors link
|
documentation
|
**Describe the bug**
When I click contributors in the READ.MD, I am directed to a 404: Page Not Found.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'https://github.com/albert-marrero/Yu-Gi-Oh-Scraper'
2. Click on 'contributors' under Authors
4. See the error
**Expected behavior**
When I click on the link, I want to direct to the correct page.
|
1.0
|
Fix broken contributors link - **Describe the bug**
When I click contributors in the READ.MD, I am directed to a 404: Page Not Found.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'https://github.com/albert-marrero/Yu-Gi-Oh-Scraper'
2. Click on 'contributors' under Authors
4. See the error
**Expected behavior**
When I click on the link, I want to direct to the correct page.
|
non_defect
|
fix broken contributors link describe the bug when i click contributors in the read md i am directed to a page not found to reproduce steps to reproduce the behavior go to click on contributors under authors see the error expected behavior when i click on the link i want to direct to the correct page
| 0
|
20,960
| 31,798,915,149
|
IssuesEvent
|
2023-09-13 09:47:57
|
storybookjs/storybook
|
https://api.github.com/repos/storybookjs/storybook
|
closed
|
[Bug]: Hot reload broken in Storybook 7 with Stencil 3.2
|
bug compatibility with other tools html windows
|
### Describe the bug
Hello,
I'm on windows 10, node 16.3, npm 9.1.2
I use storybook + stencil for two years now and it always worked great.
I updated storybook from 6.5 to 7.0.5 and now the hot reload is broken for no storybook files.
That means I can edit a xxx.stories.js file but not a xxx.scss or xxx.tsx file (provided by stencil)
The terminal has no error so it's difficult to detect if something doesn't work correctly.
`<i> [webpack-dev-middleware] wait until bundle finished: /runtime~main.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /vendors-node_modules_storybook_addon-essentials_dist_actions_preview_mjs-node_modules_storybo-21683d.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /main.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /runtime_main.b0bb9eaf7bd2c874c260.hot-update.json`
I followed the classic migration guide when I updated storybook version : npx storybook@latest upgrade
Anyone could help me please?
### To Reproduce
I can provide a repo if necessary.
### System
```shell
System:
OS: Windows 10 10.0.19045
CPU: (8) x64 Intel(R) Core(TM) i7-7700HQ CPU @ 2.80GHz
Binaries:
Node: 16.13.0 - C:\Program Files\nodejs\node.EXE
Yarn: 1.22.19 - ~\AppData\Roaming\npm\yarn.CMD
npm: 9.1.2 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: Spartan (44.19041.1266.0), Chromium (112.0.1722.48)
npmPackages:
@storybook/addon-actions: ^7.0.5 => 7.0.5
@storybook/addon-essentials: ^7.0.5 => 7.0.5
@storybook/addon-interactions: ^7.0.5 => 7.0.5
@storybook/addon-links: ^7.0.5 => 7.0.5
@storybook/html: ^7.0.5 => 7.0.5
@storybook/react: ^7.0.5 => 7.0.5
@storybook/react-webpack5: ^7.0.5 => 7.0.5
@storybook/testing-library: ^0.1.0 => 0.1.0
```
### Additional context
_No response_
|
True
|
[Bug]: Hot reload broken in Storybook 7 with Stencil 3.2 - ### Describe the bug
Hello,
I'm on windows 10, node 16.3, npm 9.1.2
I use storybook + stencil for two years now and it always worked great.
I updated storybook from 6.5 to 7.0.5 and now the hot reload is broken for no storybook files.
That means I can edit a xxx.stories.js file but not a xxx.scss or xxx.tsx file (provided by stencil)
The terminal has no error so it's difficult to detect if something doesn't work correctly.
`<i> [webpack-dev-middleware] wait until bundle finished: /runtime~main.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /vendors-node_modules_storybook_addon-essentials_dist_actions_preview_mjs-node_modules_storybo-21683d.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /main.iframe.bundle.js
<i> [webpack-dev-middleware] wait until bundle finished: /runtime_main.b0bb9eaf7bd2c874c260.hot-update.json`
I followed the classic migration guide when I updated storybook version : npx storybook@latest upgrade
Anyone could help me please?
### To Reproduce
I can provide a repo if necessary.
### System
```shell
System:
OS: Windows 10 10.0.19045
CPU: (8) x64 Intel(R) Core(TM) i7-7700HQ CPU @ 2.80GHz
Binaries:
Node: 16.13.0 - C:\Program Files\nodejs\node.EXE
Yarn: 1.22.19 - ~\AppData\Roaming\npm\yarn.CMD
npm: 9.1.2 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: Spartan (44.19041.1266.0), Chromium (112.0.1722.48)
npmPackages:
@storybook/addon-actions: ^7.0.5 => 7.0.5
@storybook/addon-essentials: ^7.0.5 => 7.0.5
@storybook/addon-interactions: ^7.0.5 => 7.0.5
@storybook/addon-links: ^7.0.5 => 7.0.5
@storybook/html: ^7.0.5 => 7.0.5
@storybook/react: ^7.0.5 => 7.0.5
@storybook/react-webpack5: ^7.0.5 => 7.0.5
@storybook/testing-library: ^0.1.0 => 0.1.0
```
### Additional context
_No response_
|
non_defect
|
hot reload broken in storybook with stencil describe the bug hello i m on windows node npm i use storybook stencil for two years now and it always worked great i updated storybook from to and now the hot reload is broken for no storybook files that means i can edit a xxx stories js file but not a xxx scss or xxx tsx file provided by stencil the terminal has no error so it s difficult to detect if something doesn t work correctly wait until bundle finished runtime main iframe bundle js wait until bundle finished vendors node modules storybook addon essentials dist actions preview mjs node modules storybo iframe bundle js wait until bundle finished main iframe bundle js wait until bundle finished runtime main hot update json i followed the classic migration guide when i updated storybook version npx storybook latest upgrade anyone could help me please to reproduce i can provide a repo if necessary system shell system os windows cpu intel r core tm cpu binaries node c program files nodejs node exe yarn appdata roaming npm yarn cmd npm c program files nodejs npm cmd browsers edge spartan chromium npmpackages storybook addon actions storybook addon essentials storybook addon interactions storybook addon links storybook html storybook react storybook react storybook testing library additional context no response
| 0
|
72,896
| 31,774,220,741
|
IssuesEvent
|
2023-09-12 13:33:56
|
vre-hub/status
|
https://api.github.com/repos/vre-hub/status
|
closed
|
🛑 VRE Notebook Service is down
|
status vre-notebook-service
|
In [`31cf265`](https://github.com/vre-hub/status/commit/31cf265cfda5a2a9a09e09002d133e9410c72b40
), VRE Notebook Service (https://jhub-vre.cern.ch/) was **down**:
- HTTP code: 503
- Response time: 921 ms
|
1.0
|
🛑 VRE Notebook Service is down - In [`31cf265`](https://github.com/vre-hub/status/commit/31cf265cfda5a2a9a09e09002d133e9410c72b40
), VRE Notebook Service (https://jhub-vre.cern.ch/) was **down**:
- HTTP code: 503
- Response time: 921 ms
|
non_defect
|
🛑 vre notebook service is down in vre notebook service was down http code response time ms
| 0
|
70,387
| 8,530,728,508
|
IssuesEvent
|
2018-11-04 02:29:00
|
quicwg/base-drafts
|
https://api.github.com/repos/quicwg/base-drafts
|
closed
|
What is the default root of the priority tree
|
-http design
|
In http/2 the default root is stream 0x00, however this is now a request stream so we clearly can't depend on that any more.
|
1.0
|
What is the default root of the priority tree - In http/2 the default root is stream 0x00, however this is now a request stream so we clearly can't depend on that any more.
|
non_defect
|
what is the default root of the priority tree in http the default root is stream however this is now a request stream so we clearly can t depend on that any more
| 0
|
14,875
| 2,831,390,048
|
IssuesEvent
|
2015-05-24 15:54:55
|
nobodyguy/dslrdashboard
|
https://api.github.com/repos/nobodyguy/dslrdashboard
|
closed
|
Press LRTimelapse button makes the app crash
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Connect Canon EOS 5D Mark II to the application.
2. Press LRTimelapse button, whether LiveView is active or not.
What is the expected output? What do you see instead?
Expected: Display of LRTimelapse settings view.
Instead: Instant crash of the application before any display of this view.
What version of the product are you using? On what operating system?
V0.30.28
Android 4.1.2 (smartphone) and 4.0.4 (tablet)
Please provide any additional information below.
Problem detected with Canon EOS 5D Mark II and:
- Samsung Galaxy S3 smartphone,
- Archos G9 101 G9 with internal 250Gb disk tablet.
```
Original issue reported on code.google.com by `alain.in...@free.fr` on 3 Sep 2013 at 12:46
|
1.0
|
Press LRTimelapse button makes the app crash - ```
What steps will reproduce the problem?
1. Connect Canon EOS 5D Mark II to the application.
2. Press LRTimelapse button, whether LiveView is active or not.
What is the expected output? What do you see instead?
Expected: Display of LRTimelapse settings view.
Instead: Instant crash of the application before any display of this view.
What version of the product are you using? On what operating system?
V0.30.28
Android 4.1.2 (smartphone) and 4.0.4 (tablet)
Please provide any additional information below.
Problem detected with Canon EOS 5D Mark II and:
- Samsung Galaxy S3 smartphone,
- Archos G9 101 G9 with internal 250Gb disk tablet.
```
Original issue reported on code.google.com by `alain.in...@free.fr` on 3 Sep 2013 at 12:46
|
defect
|
press lrtimelapse button makes the app crash what steps will reproduce the problem connect canon eos mark ii to the application press lrtimelapse button whether liveview is active or not what is the expected output what do you see instead expected display of lrtimelapse settings view instead instant crash of the application before any display of this view what version of the product are you using on what operating system android smartphone and tablet please provide any additional information below problem detected with canon eos mark ii and samsung galaxy smartphone archos with internal disk tablet original issue reported on code google com by alain in free fr on sep at
| 1
|
211,761
| 7,204,425,054
|
IssuesEvent
|
2018-02-06 12:40:15
|
Icinga/icinga2
|
https://api.github.com/repos/Icinga/icinga2
|
closed
|
[dev.icinga.com #10987] Nested hostgroups do not work
|
API bug low-priority
|
This issue has been migrated from Redmine: https://dev.icinga.com/issues/10987
**Created by cniemann on 2016-01-19 14:44:33 +00:00**
Assignee: _(none)_
Status: _New_
Target Version: _Backlog_
Last Update: _2016-01-22 15:14:38 +00:00 (in Redmine)_
Icinga Version: 2.4.1
Backport?: Not yet backported
Include in Changelog: 1
---
It's not possible to add nested hostgroups in icinga2.4.1.
The API knows about the hostGroup object and it's nested groups but the groups attribute of the host object will not be changed.
I added the objects.cache output to clearify this problem.
My testing environment is the current vagrant box with icinga2x single machine.
I added to "/etc/icinga2/conf.d/groups.conf"
object HostGroup "lin-win" {
display\_name = "Linux and Windows Servers"
groups = \[ "linux-servers", "windows-servers" \]
}
API Output:
[root@icinga2 icinga2]# ICINGA2_API_PASSWORD=icinga icinga2 console --connect 'https://root@localhost:5665/' --eval 'get_host("c1-mysql-1").groups' | python -m json.tool
[
"bp-hosts-mysql",
"linux-servers"
]
[root@icinga2 icinga2]# ICINGA2_API_PASSWORD=icinga icinga2 console --connect 'https://root@localhost:5665/' --eval 'get_host_group("lin-win")' | python -m json.tool
{
"__name": "lin-win",
"action_url": "",
"active": true,
"display_name": "Linux and Windows Servers",
"extensions": {
"DbObject": {
"type": "Object"
}
},
"groups": [
"linux-servers",
"windows-servers"
],
"ha_mode": 0.0,
"name": "lin-win",
"notes": "",
"notes_url": "",
"original_attributes": null,
"package": "_etc",
"pause_called": false,
"paused": false,
"resume_called": true,
"start_called": true,
"state_loaded": true,
"stop_called": false,
"templates": [
"lin-win"
],
"type": "HostGroup",
"vars": null,
"version": 0.0,
"zone": ""
}
\[root@icinga2 icinga2\]\# icinga2 -V
icinga2 - The Icinga 2 network monitoring daemon (version: v2.4.1-110-gda85c7a)
Copyright (c) 2012-2016 Icinga Development Team (https://www.icinga.org/)
License GPLv2+: GNU GPL version 2 or later
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.
Application information:
Installation root: /usr
Sysconf directory: /etc
Run directory: /run
Local state directory: /var
Package data directory: /usr/share/icinga2
State path: /var/lib/icinga2/icinga2.state
Modified attributes path: /var/lib/icinga2/modified-attributes.conf
Objects path: /var/cache/icinga2/icinga2.debug
Vars path: /var/cache/icinga2/icinga2.vars
PID path: /run/icinga2/icinga2.pid
System information:
Platform: ME="cpe:/o:centos:centos:7
Platform version: UPPORT_PRODUCT_VERSION="7
Kernel: Linux
Kernel version: 3.10.0-327.4.4.el7.x86_64
Architecture: x86_64
- from /var/cache/icinga2/objects.cache
define hostgroup {
hostgroup_name lin-win
alias Linux and Windows Servers
members
}
define hostgroup {
hostgroup_name linux-servers
alias Linux Servers
members c1-mysql-1,c1-web-1,c2-mysql-2,icinga2
}
define hostgroup {
hostgroup_name windows-servers
alias Windows Servers
members c2-web-1
}
|
1.0
|
[dev.icinga.com #10987] Nested hostgroups do not work - This issue has been migrated from Redmine: https://dev.icinga.com/issues/10987
**Created by cniemann on 2016-01-19 14:44:33 +00:00**
Assignee: _(none)_
Status: _New_
Target Version: _Backlog_
Last Update: _2016-01-22 15:14:38 +00:00 (in Redmine)_
Icinga Version: 2.4.1
Backport?: Not yet backported
Include in Changelog: 1
---
It's not possible to add nested hostgroups in icinga2.4.1.
The API knows about the hostGroup object and it's nested groups but the groups attribute of the host object will not be changed.
I added the objects.cache output to clearify this problem.
My testing environment is the current vagrant box with icinga2x single machine.
I added to "/etc/icinga2/conf.d/groups.conf"
object HostGroup "lin-win" {
display\_name = "Linux and Windows Servers"
groups = \[ "linux-servers", "windows-servers" \]
}
API Output:
[root@icinga2 icinga2]# ICINGA2_API_PASSWORD=icinga icinga2 console --connect 'https://root@localhost:5665/' --eval 'get_host("c1-mysql-1").groups' | python -m json.tool
[
"bp-hosts-mysql",
"linux-servers"
]
[root@icinga2 icinga2]# ICINGA2_API_PASSWORD=icinga icinga2 console --connect 'https://root@localhost:5665/' --eval 'get_host_group("lin-win")' | python -m json.tool
{
"__name": "lin-win",
"action_url": "",
"active": true,
"display_name": "Linux and Windows Servers",
"extensions": {
"DbObject": {
"type": "Object"
}
},
"groups": [
"linux-servers",
"windows-servers"
],
"ha_mode": 0.0,
"name": "lin-win",
"notes": "",
"notes_url": "",
"original_attributes": null,
"package": "_etc",
"pause_called": false,
"paused": false,
"resume_called": true,
"start_called": true,
"state_loaded": true,
"stop_called": false,
"templates": [
"lin-win"
],
"type": "HostGroup",
"vars": null,
"version": 0.0,
"zone": ""
}
\[root@icinga2 icinga2\]\# icinga2 -V
icinga2 - The Icinga 2 network monitoring daemon (version: v2.4.1-110-gda85c7a)
Copyright (c) 2012-2016 Icinga Development Team (https://www.icinga.org/)
License GPLv2+: GNU GPL version 2 or later
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.
Application information:
Installation root: /usr
Sysconf directory: /etc
Run directory: /run
Local state directory: /var
Package data directory: /usr/share/icinga2
State path: /var/lib/icinga2/icinga2.state
Modified attributes path: /var/lib/icinga2/modified-attributes.conf
Objects path: /var/cache/icinga2/icinga2.debug
Vars path: /var/cache/icinga2/icinga2.vars
PID path: /run/icinga2/icinga2.pid
System information:
Platform: ME="cpe:/o:centos:centos:7
Platform version: UPPORT_PRODUCT_VERSION="7
Kernel: Linux
Kernel version: 3.10.0-327.4.4.el7.x86_64
Architecture: x86_64
- from /var/cache/icinga2/objects.cache
define hostgroup {
hostgroup_name lin-win
alias Linux and Windows Servers
members
}
define hostgroup {
hostgroup_name linux-servers
alias Linux Servers
members c1-mysql-1,c1-web-1,c2-mysql-2,icinga2
}
define hostgroup {
hostgroup_name windows-servers
alias Windows Servers
members c2-web-1
}
|
non_defect
|
nested hostgroups do not work this issue has been migrated from redmine created by cniemann on assignee none status new target version backlog last update in redmine icinga version backport not yet backported include in changelog it s not possible to add nested hostgroups in the api knows about the hostgroup object and it s nested groups but the groups attribute of the host object will not be changed i added the objects cache output to clearify this problem my testing environment is the current vagrant box with single machine i added to etc conf d groups conf object hostgroup lin win display name linux and windows servers groups api output api password icinga console connect eval get host mysql groups python m json tool bp hosts mysql linux servers api password icinga console connect eval get host group lin win python m json tool name lin win action url active true display name linux and windows servers extensions dbobject type object groups linux servers windows servers ha mode name lin win notes notes url original attributes null package etc pause called false paused false resume called true start called true state loaded true stop called false templates lin win type hostgroup vars null version zone v the icinga network monitoring daemon version copyright c icinga development team license gnu gpl version or later this is free software you are free to change and redistribute it there is no warranty to the extent permitted by law application information installation root usr sysconf directory etc run directory run local state directory var package data directory usr share state path var lib state modified attributes path var lib modified attributes conf objects path var cache debug vars path var cache vars pid path run pid system information platform me cpe o centos centos platform version upport product version kernel linux kernel version architecture from var cache objects cache define hostgroup hostgroup name lin win alias linux and windows servers members define hostgroup hostgroup name linux servers alias linux servers members mysql web mysql define hostgroup hostgroup name windows servers alias windows servers members web
| 0
|
31,503
| 11,944,336,796
|
IssuesEvent
|
2020-04-03 02:08:07
|
drakeg/udemy_django_vue
|
https://api.github.com/repos/drakeg/udemy_django_vue
|
closed
|
WS-2019-0381 (Medium) detected in kind-of-6.0.2.tgz
|
security vulnerability
|
## WS-2019-0381 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/udemy_django_vue/final_project/frontend/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/udemy_django_vue/final_project/frontend/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-babel-3.12.1.tgz (Root Library)
- webpack-4.41.2.tgz
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/drakeg/udemy_django_vue/commit/f99ecfe0710e84e6fe73a96c26c4108ed903d0fe">f99ecfe0710e84e6fe73a96c26c4108ed903d0fe</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of kind-of 6.x prior to 6.0.3 are vulnerable to a Validation Bypass. A maliciously crafted object can alter the result of the type check, allowing attackers to bypass the type checking validation.
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8>WS-2019-0381</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8">https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: kind-of - 6.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0381 (Medium) detected in kind-of-6.0.2.tgz - ## WS-2019-0381 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/udemy_django_vue/final_project/frontend/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/udemy_django_vue/final_project/frontend/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-babel-3.12.1.tgz (Root Library)
- webpack-4.41.2.tgz
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/drakeg/udemy_django_vue/commit/f99ecfe0710e84e6fe73a96c26c4108ed903d0fe">f99ecfe0710e84e6fe73a96c26c4108ed903d0fe</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of kind-of 6.x prior to 6.0.3 are vulnerable to a Validation Bypass. A maliciously crafted object can alter the result of the type check, allowing attackers to bypass the type checking validation.
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8>WS-2019-0381</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8">https://github.com/jonschlinkert/kind-of/commit/975c13a7cfaf25d811475823824af3a9c04b0ba8</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: kind-of - 6.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in kind of tgz ws medium severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file tmp ws scm udemy django vue final project frontend package json path to vulnerable library tmp ws scm udemy django vue final project frontend node modules kind of package json dependency hierarchy cli plugin babel tgz root library webpack tgz micromatch tgz x kind of tgz vulnerable library found in head commit a href vulnerability details versions of kind of x prior to are vulnerable to a validation bypass a maliciously crafted object can alter the result of the type check allowing attackers to bypass the type checking validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution kind of step up your open source security game with whitesource
| 0
|
46,299
| 13,055,886,944
|
IssuesEvent
|
2020-07-30 03:01:38
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[steamshovel] - I3Particle from MC doesn't show up (Trac #939)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/939
```json
{
"status": "closed",
"changetime": "2015-05-06T15:59:33",
"description": "Clicking on the I3Particle from the new MC frame object won't create a line for tracks.\nThis is for Gen2 Simulation files:\nhttps://wiki.icecube.wisc.edu/index.php/IceCube_Extensions_SimPage.NuGen_Simulation",
"reporter": "naoko",
"cc": "",
"resolution": "fixed",
"_ts": "1430927973153413",
"component": "combo reconstruction",
"summary": "[steamshovel] - I3Particle from MC doesn't show up",
"priority": "normal",
"keywords": "",
"time": "2015-04-16T14:11:47",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
|
1.0
|
[steamshovel] - I3Particle from MC doesn't show up (Trac #939) - Migrated from https://code.icecube.wisc.edu/ticket/939
```json
{
"status": "closed",
"changetime": "2015-05-06T15:59:33",
"description": "Clicking on the I3Particle from the new MC frame object won't create a line for tracks.\nThis is for Gen2 Simulation files:\nhttps://wiki.icecube.wisc.edu/index.php/IceCube_Extensions_SimPage.NuGen_Simulation",
"reporter": "naoko",
"cc": "",
"resolution": "fixed",
"_ts": "1430927973153413",
"component": "combo reconstruction",
"summary": "[steamshovel] - I3Particle from MC doesn't show up",
"priority": "normal",
"keywords": "",
"time": "2015-04-16T14:11:47",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
|
defect
|
from mc doesn t show up trac migrated from json status closed changetime description clicking on the from the new mc frame object won t create a line for tracks nthis is for simulation files n reporter naoko cc resolution fixed ts component combo reconstruction summary from mc doesn t show up priority normal keywords time milestone owner hdembinski type defect
| 1
|
17,602
| 3,012,747,745
|
IssuesEvent
|
2015-07-29 02:09:38
|
yawlfoundation/yawl
|
https://api.github.com/repos/yawlfoundation/yawl
|
closed
|
[CLOSED] Cyclic role/position/org group relations
|
auto-migrated Milestone-Release2.0 Priority-Critical Type-Defect
|
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:20 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/3_
----
```
At the moment it is possible to create cyclic dependencies (except
one-cycles) in the Org Data Management for roles, positions and Org
Groupings. This should not be allowed.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 18 Jul 2008 at 4:04
|
1.0
|
[CLOSED] Cyclic role/position/org group relations - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:20 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/3_
----
```
At the moment it is possible to create cyclic dependencies (except
one-cycles) in the Org Data Management for roles, positions and Org
Groupings. This should not be allowed.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 18 Jul 2008 at 4:04
|
defect
|
cyclic role position org group relations issue by monday jul at gmt originally opened as at the moment it is possible to create cyclic dependencies except one cycles in the org data management for roles positions and org groupings this should not be allowed original issue reported on code google com by arthurte gmail com on jul at
| 1
|
68,839
| 21,920,530,773
|
IssuesEvent
|
2022-05-22 13:56:40
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
closed
|
Problems with HPX serialization as a standalone feature. Testcase provided.
|
type: defect category: serialization category: modules
|
## Expected Behavior
Not really expected, but rather hoped for: Compile correctly.
## Actual Behavior
Inclusion of hpx/serialization.hpp causes compile errors
## Steps to Reproduce the Problem
Use the provided testcase and read its readme.md.
## Specifications
Windows, MSVC 17.2.1
HPX Version: 1.7.1
[vcpkg.zip](https://github.com/STEllAR-GROUP/hpx/files/8738998/vcpkg.zip)
|
1.0
|
Problems with HPX serialization as a standalone feature. Testcase provided. - ## Expected Behavior
Not really expected, but rather hoped for: Compile correctly.
## Actual Behavior
Inclusion of hpx/serialization.hpp causes compile errors
## Steps to Reproduce the Problem
Use the provided testcase and read its readme.md.
## Specifications
Windows, MSVC 17.2.1
HPX Version: 1.7.1
[vcpkg.zip](https://github.com/STEllAR-GROUP/hpx/files/8738998/vcpkg.zip)
|
defect
|
problems with hpx serialization as a standalone feature testcase provided expected behavior not really expected but rather hoped for compile correctly actual behavior inclusion of hpx serialization hpp causes compile errors steps to reproduce the problem use the provided testcase and read its readme md specifications windows msvc hpx version
| 1
|
331,091
| 10,060,097,971
|
IssuesEvent
|
2019-07-22 17:59:59
|
SETI/pds-opus
|
https://api.github.com/repos/SETI/pds-opus
|
closed
|
User confusion on Cart tab
|
A-Enhancement B-OPUS Django B-OPUS JS Effort 2 Medium Priority 4 Useful
|
From Michael Aye:
Hi Rob,
I think at the stage in the screenshot, the GUI guidance is confusing.
The links to “Select Metadata” and “View Table” are flow controllers, while the links to “MetaData CSV”, “Data Archive” and “URL Archive” are downlink buttons, and would be better named “Get Metadata”, Get Data” and “Get URLs” and get their own color and their own row, b/c they are conceptually very different from the other buttons.
If not another row for saving space, at least a different color, b/c they are really for different actions than the other two links.
Like now, these actions are all blurred together, I was staring at this page for about a minute until I figured out what I was supposed to do (and remember I’m fast ;) ).
Maybe the mental misguiding comes from the fact that the left side has a very big white “Download options” section but then does NOT have a download button, I think that’s what surprised or puzzled me the most.
[...]
I experienced some added confusion that pressing the “Data Archive” button only created a zip file on the left and it wasn’t very clear that i still need to click on that zip link.
|
1.0
|
User confusion on Cart tab - From Michael Aye:
Hi Rob,
I think at the stage in the screenshot, the GUI guidance is confusing.
The links to “Select Metadata” and “View Table” are flow controllers, while the links to “MetaData CSV”, “Data Archive” and “URL Archive” are downlink buttons, and would be better named “Get Metadata”, Get Data” and “Get URLs” and get their own color and their own row, b/c they are conceptually very different from the other buttons.
If not another row for saving space, at least a different color, b/c they are really for different actions than the other two links.
Like now, these actions are all blurred together, I was staring at this page for about a minute until I figured out what I was supposed to do (and remember I’m fast ;) ).
Maybe the mental misguiding comes from the fact that the left side has a very big white “Download options” section but then does NOT have a download button, I think that’s what surprised or puzzled me the most.
[...]
I experienced some added confusion that pressing the “Data Archive” button only created a zip file on the left and it wasn’t very clear that i still need to click on that zip link.
|
non_defect
|
user confusion on cart tab from michael aye hi rob i think at the stage in the screenshot the gui guidance is confusing the links to “select metadata” and “view table” are flow controllers while the links to “metadata csv” “data archive” and “url archive” are downlink buttons and would be better named “get metadata” get data” and “get urls” and get their own color and their own row b c they are conceptually very different from the other buttons if not another row for saving space at least a different color b c they are really for different actions than the other two links like now these actions are all blurred together i was staring at this page for about a minute until i figured out what i was supposed to do and remember i’m fast maybe the mental misguiding comes from the fact that the left side has a very big white “download options” section but then does not have a download button i think that’s what surprised or puzzled me the most i experienced some added confusion that pressing the “data archive” button only created a zip file on the left and it wasn’t very clear that i still need to click on that zip link
| 0
|
17,728
| 9,884,474,121
|
IssuesEvent
|
2019-06-24 22:17:06
|
keras-team/keras
|
https://api.github.com/repos/keras-team/keras
|
closed
|
tf.keras.Model.evaluate skews score in custom tf.keras.callbacks.Callback
|
backend:tensorflow type:bug/performance
|
**System information**
- Have I written custom code (as opposed to using example directory): YES
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04.2 LTS
- TensorFlow backend (yes / no): YES
- TensorFlow version: 2.0.0-dev20190601
- Keras version: 2.2.4-tf
- Python version: 3.6
- CUDA/cuDNN version: 7.3.1
- GPU model and memory: 4 GeForce GTX 1080 w/8119MiB
**Describe the current behavior**
I am using the TensorFlow Dataset api with the Keras Model.fit function. When `x` is a Dataset (and therefor no `y` is provided), an error is thrown if we provide a `batch_size` argument. However, a `batch_size` is required is we want to use `validation_data` (I would already classify this as an error).
So, I wrote my own custom callback to perform a call to Model.evaluate with the validation dataset I had already prepared after every `n`-th batch and after every single epoch.
During training, at, say, step 1000 (and before the first epoch), the accuracy reported by Tensorboard (I use the Tensorboard Keras callback for train loss and accuracy), as well as the accuracy reported by the metrics display next to the progress bar, might be at ~19%.
However, after the CustomCallback runs it's function on_batch_end(self, batch, logs=None) and a call to self.model.evaluate is made, the training accuracy displayed by the progress bar and by Tensorboard picks up where the validation accuracy left off. That is, if val accuracy happened to be ~28% (higher due to no Dropout), then the training metric would also read 28% from then on.
**Describe the expected behavior**
Calling the model's evaluate function from within a custom callback should not change what is displayed for train accuracy. Also, built-in Dataset validation would be nice.
**Code to reproduce the issue**
This is the custom callback. Instantiating this callback with a subdirectory for valdiation events and a Dataset instance, and then including this callback in a call to model.fit should reproduce the error. I invite you to provide the data. I am abandoning the Keras Model.fit pipeline for now due to time constraints but I thought I would still log the error.
For this bug to be noticeable, the training accuracy must be noticeably lower due to dropout than the val acc. (the color is blue but this is train accuracy, I promise. The jump is at step 1000, at which point Model.evaluate was called).

`
class Validation(tf.keras.callbacks.Callback):
def __init__(self, log_dir, dataset, *args, **kwargs):
self.dataset = dataset
self.writer = tf.summary.create_file_writer(f"{log_dir}/validation")
def on_train_begin(self, logs=None):
self.history = {}
self.step = 0
def on_batch_end(self, batch, logs=None):
self.step += 1
if self.step % Config.val_log_freq == 0:
metrics = self.model.evaluate(self.dataset, verbose=0)
for k, v in zip(self.model.metrics_names, metrics):
with self.writer.as_default():
tf.summary.scalar(f"batch_{k}", v, step=self.step)
self.writer.flush()
def on_epoch_end(self, epoch, logs=None):
metrics = self.model.evaluate(self.dataset, verbose=0)
for k, v in zip(self.model.metrics_names, metrics):
self.history.setdefault(k, []).append(v)
with self.writer.as_default():
tf.summary.scalar(f"epoch_{k}", v, step=self.step)
self.writer.flush()
validation_callback = Validation(log_dir, valid_dataset)`
**Other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
True
|
tf.keras.Model.evaluate skews score in custom tf.keras.callbacks.Callback - **System information**
- Have I written custom code (as opposed to using example directory): YES
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04.2 LTS
- TensorFlow backend (yes / no): YES
- TensorFlow version: 2.0.0-dev20190601
- Keras version: 2.2.4-tf
- Python version: 3.6
- CUDA/cuDNN version: 7.3.1
- GPU model and memory: 4 GeForce GTX 1080 w/8119MiB
**Describe the current behavior**
I am using the TensorFlow Dataset api with the Keras Model.fit function. When `x` is a Dataset (and therefor no `y` is provided), an error is thrown if we provide a `batch_size` argument. However, a `batch_size` is required is we want to use `validation_data` (I would already classify this as an error).
So, I wrote my own custom callback to perform a call to Model.evaluate with the validation dataset I had already prepared after every `n`-th batch and after every single epoch.
During training, at, say, step 1000 (and before the first epoch), the accuracy reported by Tensorboard (I use the Tensorboard Keras callback for train loss and accuracy), as well as the accuracy reported by the metrics display next to the progress bar, might be at ~19%.
However, after the CustomCallback runs it's function on_batch_end(self, batch, logs=None) and a call to self.model.evaluate is made, the training accuracy displayed by the progress bar and by Tensorboard picks up where the validation accuracy left off. That is, if val accuracy happened to be ~28% (higher due to no Dropout), then the training metric would also read 28% from then on.
**Describe the expected behavior**
Calling the model's evaluate function from within a custom callback should not change what is displayed for train accuracy. Also, built-in Dataset validation would be nice.
**Code to reproduce the issue**
This is the custom callback. Instantiating this callback with a subdirectory for valdiation events and a Dataset instance, and then including this callback in a call to model.fit should reproduce the error. I invite you to provide the data. I am abandoning the Keras Model.fit pipeline for now due to time constraints but I thought I would still log the error.
For this bug to be noticeable, the training accuracy must be noticeably lower due to dropout than the val acc. (the color is blue but this is train accuracy, I promise. The jump is at step 1000, at which point Model.evaluate was called).

`
class Validation(tf.keras.callbacks.Callback):
def __init__(self, log_dir, dataset, *args, **kwargs):
self.dataset = dataset
self.writer = tf.summary.create_file_writer(f"{log_dir}/validation")
def on_train_begin(self, logs=None):
self.history = {}
self.step = 0
def on_batch_end(self, batch, logs=None):
self.step += 1
if self.step % Config.val_log_freq == 0:
metrics = self.model.evaluate(self.dataset, verbose=0)
for k, v in zip(self.model.metrics_names, metrics):
with self.writer.as_default():
tf.summary.scalar(f"batch_{k}", v, step=self.step)
self.writer.flush()
def on_epoch_end(self, epoch, logs=None):
metrics = self.model.evaluate(self.dataset, verbose=0)
for k, v in zip(self.model.metrics_names, metrics):
self.history.setdefault(k, []).append(v)
with self.writer.as_default():
tf.summary.scalar(f"epoch_{k}", v, step=self.step)
self.writer.flush()
validation_callback = Validation(log_dir, valid_dataset)`
**Other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
non_defect
|
tf keras model evaluate skews score in custom tf keras callbacks callback system information have i written custom code as opposed to using example directory yes os platform and distribution e g linux ubuntu ubuntu lts tensorflow backend yes no yes tensorflow version keras version tf python version cuda cudnn version gpu model and memory geforce gtx w describe the current behavior i am using the tensorflow dataset api with the keras model fit function when x is a dataset and therefor no y is provided an error is thrown if we provide a batch size argument however a batch size is required is we want to use validation data i would already classify this as an error so i wrote my own custom callback to perform a call to model evaluate with the validation dataset i had already prepared after every n th batch and after every single epoch during training at say step and before the first epoch the accuracy reported by tensorboard i use the tensorboard keras callback for train loss and accuracy as well as the accuracy reported by the metrics display next to the progress bar might be at however after the customcallback runs it s function on batch end self batch logs none and a call to self model evaluate is made the training accuracy displayed by the progress bar and by tensorboard picks up where the validation accuracy left off that is if val accuracy happened to be higher due to no dropout then the training metric would also read from then on describe the expected behavior calling the model s evaluate function from within a custom callback should not change what is displayed for train accuracy also built in dataset validation would be nice code to reproduce the issue this is the custom callback instantiating this callback with a subdirectory for valdiation events and a dataset instance and then including this callback in a call to model fit should reproduce the error i invite you to provide the data i am abandoning the keras model fit pipeline for now due to time constraints but i thought i would still log the error for this bug to be noticeable the training accuracy must be noticeably lower due to dropout than the val acc the color is blue but this is train accuracy i promise the jump is at step at which point model evaluate was called class validation tf keras callbacks callback def init self log dir dataset args kwargs self dataset dataset self writer tf summary create file writer f log dir validation def on train begin self logs none self history self step def on batch end self batch logs none self step if self step config val log freq metrics self model evaluate self dataset verbose for k v in zip self model metrics names metrics with self writer as default tf summary scalar f batch k v step self step self writer flush def on epoch end self epoch logs none metrics self model evaluate self dataset verbose for k v in zip self model metrics names metrics self history setdefault k append v with self writer as default tf summary scalar f epoch k v step self step self writer flush validation callback validation log dir valid dataset other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 0
|
15,029
| 2,838,956,286
|
IssuesEvent
|
2015-05-27 10:53:46
|
ibus/ibus
|
https://api.github.com/repos/ibus/ibus
|
closed
|
ibus-hangul: setup doesn't work with ibus 1.5.x
|
Component-ibus-hangul Priority-Medium Type-Defect
|
```
What version of the product are you using? On what operating system?
OS (Linux distributions, UNIX or ...):
Architecture (i386, x86_64):
IBus version: 1.5.1
Input method name and version: ibus-hangul 1.4.2
Python version: 2.7
dbus version: 1.5.8
dbus-python version: 1.1.1
$ dpkg -s ibus|grep Version
Version: 1.5.1-1
$ dpkg -s ibus-hangul|grep Version
Version: 1.4.2-1
$ ibus-setup-hangul
Traceback (most recent call last):
File "/usr/share/ibus-hangul/setup/main.py", line 208, in <module>
Setup(bus).run()
File "/usr/share/ibus-hangul/setup/main.py", line 38, in __init__
self.__config.connect("value-changed", self.on_value_changed, None)
AttributeError: 'NoneType' object has no attribute 'connect'
$
```
Original issue reported on code.google.com by `ryu.changwoo` on 2013-02-10 02:19:54
|
1.0
|
ibus-hangul: setup doesn't work with ibus 1.5.x - ```
What version of the product are you using? On what operating system?
OS (Linux distributions, UNIX or ...):
Architecture (i386, x86_64):
IBus version: 1.5.1
Input method name and version: ibus-hangul 1.4.2
Python version: 2.7
dbus version: 1.5.8
dbus-python version: 1.1.1
$ dpkg -s ibus|grep Version
Version: 1.5.1-1
$ dpkg -s ibus-hangul|grep Version
Version: 1.4.2-1
$ ibus-setup-hangul
Traceback (most recent call last):
File "/usr/share/ibus-hangul/setup/main.py", line 208, in <module>
Setup(bus).run()
File "/usr/share/ibus-hangul/setup/main.py", line 38, in __init__
self.__config.connect("value-changed", self.on_value_changed, None)
AttributeError: 'NoneType' object has no attribute 'connect'
$
```
Original issue reported on code.google.com by `ryu.changwoo` on 2013-02-10 02:19:54
|
defect
|
ibus hangul setup doesn t work with ibus x what version of the product are you using on what operating system os linux distributions unix or architecture ibus version input method name and version ibus hangul python version dbus version dbus python version dpkg s ibus grep version version dpkg s ibus hangul grep version version ibus setup hangul traceback most recent call last file usr share ibus hangul setup main py line in setup bus run file usr share ibus hangul setup main py line in init self config connect value changed self on value changed none attributeerror nonetype object has no attribute connect original issue reported on code google com by ryu changwoo on
| 1
|
6,541
| 2,610,256,509
|
IssuesEvent
|
2015-02-26 19:21:55
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳激光祛痘费用
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳激光祛痘费用【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:34
|
1.0
|
深圳激光祛痘费用 - ```
深圳激光祛痘费用【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:34
|
defect
|
深圳激光祛痘费用 深圳激光祛痘费用【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 original issue reported on code google com by szft com on may at
| 1
|
256,181
| 8,127,025,788
|
IssuesEvent
|
2018-08-17 06:13:37
|
aowen87/BAR
|
https://api.github.com/repos/aowen87/BAR
|
closed
|
Fix python exception raise in EvalCubicSpline
|
Bug Likelihood: 3 - Occasional Priority: Normal Severity: 2 - Minor Irritation
|
Current code:
if((allX[0] > t) or (allX[n-1] < t)):
raise 't must be in the range between the first and last X'\n"
If the raise occurs, the message isn't shown b/c python's exception mechanism wants a class or a non-string instance.
All you see is:
"TypeError: exceptions must be classes or instances, not str" (@ line 4 in EvalCubicSpline)
There may be other instances of this in the python module as well.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 500
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Fix python exception raise in EvalCubicSpline
Assigned to: Cyrus Harrison
Category:
Target version: 2.1.2
Author: Cyrus Harrison
Start: 12/03/2010
Due date:
% Done: 0
Estimated time:
Created: 12/03/2010 02:22 pm
Updated: 12/08/2010 02:53 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.1.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Current code:
if((allX[0] > t) or (allX[n-1] < t)):
raise 't must be in the range between the first and last X'\n"
If the raise occurs, the message isn't shown b/c python's exception mechanism wants a class or a non-string instance.
All you see is:
"TypeError: exceptions must be classes or instances, not str" (@ line 4 in EvalCubicSpline)
There may be other instances of this in the python module as well.
Comments:
Update from LLNL Meeting.
Hi Everyone,I improved the exception raised in EvalCubicSpline when the user passes a bad interpolation value.It now raises a ValueError exception & ends up printing detailed info.RC:Sending visitpy/common/visitmodule.CTransmitting file data .Committed revision r13218.Trunk:Sending visitpy/common/visitmodule.CTransmitting file data .Committed revision r13220.
|
1.0
|
Fix python exception raise in EvalCubicSpline - Current code:
if((allX[0] > t) or (allX[n-1] < t)):
raise 't must be in the range between the first and last X'\n"
If the raise occurs, the message isn't shown b/c python's exception mechanism wants a class or a non-string instance.
All you see is:
"TypeError: exceptions must be classes or instances, not str" (@ line 4 in EvalCubicSpline)
There may be other instances of this in the python module as well.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 500
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Fix python exception raise in EvalCubicSpline
Assigned to: Cyrus Harrison
Category:
Target version: 2.1.2
Author: Cyrus Harrison
Start: 12/03/2010
Due date:
% Done: 0
Estimated time:
Created: 12/03/2010 02:22 pm
Updated: 12/08/2010 02:53 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.1.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Current code:
if((allX[0] > t) or (allX[n-1] < t)):
raise 't must be in the range between the first and last X'\n"
If the raise occurs, the message isn't shown b/c python's exception mechanism wants a class or a non-string instance.
All you see is:
"TypeError: exceptions must be classes or instances, not str" (@ line 4 in EvalCubicSpline)
There may be other instances of this in the python module as well.
Comments:
Update from LLNL Meeting.
Hi Everyone,I improved the exception raised in EvalCubicSpline when the user passes a bad interpolation value.It now raises a ValueError exception & ends up printing detailed info.RC:Sending visitpy/common/visitmodule.CTransmitting file data .Committed revision r13218.Trunk:Sending visitpy/common/visitmodule.CTransmitting file data .Committed revision r13220.
|
non_defect
|
fix python exception raise in evalcubicspline current code if allx t or allx t raise t must be in the range between the first and last x n if the raise occurs the message isn t shown b c python s exception mechanism wants a class or a non string instance all you see is typeerror exceptions must be classes or instances not str line in evalcubicspline there may be other instances of this in the python module as well redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject fix python exception raise in evalcubicspline assigned to cyrus harrison category target version author cyrus harrison start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description current code if allx t or allx t raise t must be in the range between the first and last x n if the raise occurs the message isn t shown b c python s exception mechanism wants a class or a non string instance all you see is typeerror exceptions must be classes or instances not str line in evalcubicspline there may be other instances of this in the python module as well comments update from llnl meeting hi everyone i improved the exception raised in evalcubicspline when the user passes a bad interpolation value it now raises a valueerror exception ends up printing detailed info rc sending visitpy common visitmodule ctransmitting file data committed revision trunk sending visitpy common visitmodule ctransmitting file data committed revision
| 0
|
75,670
| 25,989,652,818
|
IssuesEvent
|
2022-12-20 05:57:41
|
DependencyTrack/dependency-track
|
https://api.github.com/repos/DependencyTrack/dependency-track
|
closed
|
UTF8 encoded with ByteOrderMark BOM uploaded using PUT /v1/bom/ fails
|
defect p2 in triage
|
### Current Behavior
Uploading [this BOM](https://github.com/DependencyTrack/dependency-track/files/10263188/bom-with-utf8-byte-order-mark.zip) using the PUT /v1/bom endpoint fails with the following warning on the server, but returns a successful response to the client.
> The BOM uploaded is not in a supported format. Supported formats include CycloneDX XML and JSON
This is the same issue as https://github.com/DependencyTrack/dependency-track/issues/1214, although in that issue, only the POST endpoint was corrected.
### Steps to Reproduce
This python snip-it with [this BOM](https://github.com/DependencyTrack/dependency-track/files/10263188/bom-with-utf8-byte-order-mark.zip) will reproduce the issue:
```python
import requests
import base64
f = open("with-utf8-byte-order-mark.xml", "rb")
d = {
'project': '<projectId>',
'bom': base64.b64encode(f.read()).decode('ascii')
}
h = {
'Content-type':'application/json',
'Accept':'application/json',
'X-API-Key': "<APIKey>"
}
r = requests.put('<DependencyTrackURL>/api/v1/bom', json=d, headers=h)
print(r)
print(r.content)
```
### Expected Behavior
The PUT /v1/bom endpoint should strip the Byte-Order-Mark to have the same behaviour as the POST /v1/bom endpoint.
### Dependency-Track Version
4.7.0
### Dependency-Track Distribution
Container Image
### Database Server
N/A
### Database Server Version
_No response_
### Browser
N/A
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
|
1.0
|
UTF8 encoded with ByteOrderMark BOM uploaded using PUT /v1/bom/ fails - ### Current Behavior
Uploading [this BOM](https://github.com/DependencyTrack/dependency-track/files/10263188/bom-with-utf8-byte-order-mark.zip) using the PUT /v1/bom endpoint fails with the following warning on the server, but returns a successful response to the client.
> The BOM uploaded is not in a supported format. Supported formats include CycloneDX XML and JSON
This is the same issue as https://github.com/DependencyTrack/dependency-track/issues/1214, although in that issue, only the POST endpoint was corrected.
### Steps to Reproduce
This python snip-it with [this BOM](https://github.com/DependencyTrack/dependency-track/files/10263188/bom-with-utf8-byte-order-mark.zip) will reproduce the issue:
```python
import requests
import base64
f = open("with-utf8-byte-order-mark.xml", "rb")
d = {
'project': '<projectId>',
'bom': base64.b64encode(f.read()).decode('ascii')
}
h = {
'Content-type':'application/json',
'Accept':'application/json',
'X-API-Key': "<APIKey>"
}
r = requests.put('<DependencyTrackURL>/api/v1/bom', json=d, headers=h)
print(r)
print(r.content)
```
### Expected Behavior
The PUT /v1/bom endpoint should strip the Byte-Order-Mark to have the same behaviour as the POST /v1/bom endpoint.
### Dependency-Track Version
4.7.0
### Dependency-Track Distribution
Container Image
### Database Server
N/A
### Database Server Version
_No response_
### Browser
N/A
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
|
defect
|
encoded with byteordermark bom uploaded using put bom fails current behavior uploading using the put bom endpoint fails with the following warning on the server but returns a successful response to the client the bom uploaded is not in a supported format supported formats include cyclonedx xml and json this is the same issue as although in that issue only the post endpoint was corrected steps to reproduce this python snip it with will reproduce the issue python import requests import f open with byte order mark xml rb d project bom f read decode ascii h content type application json accept application json x api key r requests put api bom json d headers h print r print r content expected behavior the put bom endpoint should strip the byte order mark to have the same behaviour as the post bom endpoint dependency track version dependency track distribution container image database server n a database server version no response browser n a checklist i have read and understand the i have checked the for whether this defect was already reported
| 1
|
74,775
| 25,316,486,658
|
IssuesEvent
|
2022-11-17 22:09:05
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
zpool not importable and ASSERT at zdb.c:3908:load_concrete_ms_allocatable_trees()
|
Type: Defect
|
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04
Kernel Version | 5.4.0-132-generic #148-Ubuntu SMP Mon Oct 17 16:02:06 UTC 2022
Architecture | x86_64
OpenZFS Version | zfs-0.8.3-1ubuntu12.14, zfs-kmod-0.8.3-1ubuntu12.14
### Describe the problem you're observing
I have a zpool with raidz on three 5T Seagate disks, running on Ubuntu 20.04 happily for ~2years
which can't be imported anymore.
SMART looks unsuspicious for all three devices, e.g.
```
SMART Self-test log structure revision number 1
Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
# 1 Short offline Completed without error 00% 58 -
```
The troubling pool shows up in
```
zpool import -d /dev/disk/by-id
pool: three
id: 7585041214268964297
state: ONLINE
action: The pool can be imported using its name or numeric identifier.
config:
three ONLINE
raidz1-0 ONLINE
wwn-0x5000c500d3e2552b ONLINE
wwn-0x5000c500d3f965c2 ONLINE
wwn-0x5000c500d3f92a29 ONLINE
```
but it can not be imported:
```
zpool import three
cannot import 'three': I/O error
Destroy and re-create the pool from
a backup source.
```
The ZDB looks ok (no expert here)
```
zdb -e three
Configuration for import:
vdev_children: 1
version: 5000
pool_guid: 7585041214268964297
name: 'three'
state: 0
hostid: 2677716145
hostname: 'kaestchen'
vdev_tree:
type: 'root'
id: 0
guid: 7585041214268964297
children[0]:
type: 'raidz'
id: 0
guid: 5612826130591374815
nparity: 1
metaslab_array: 69
metaslab_shift: 34
ashift: 9
asize: 15002900103168
is_log: 0
create_txg: 4
children[0]:
type: 'disk'
id: 0
guid: 4245064270772150798
whole_disk: 1
DTL: 373
create_txg: 4
path: '/dev/sda1'
devid: 'ata-ST5000LM000-2AN170_WCJ3V9AZ-part1'
phys_path: 'pci-0000:00:1f.2-ata-2'
children[1]:
type: 'disk'
id: 1
guid: 3847351160452222273
whole_disk: 1
DTL: 391
create_txg: 4
path: '/dev/sdb1'
devid: 'ata-ST5000LM000-2AN170_WCJ3VQCA-part1'
phys_path: 'pci-0000:00:1f.2-ata-3'
children[2]:
type: 'disk'
id: 2
guid: 8390181302815008501
whole_disk: 1
DTL: 390
create_txg: 4
path: '/dev/sdc1'
devid: 'ata-ST5000LM000-2AN170_WCJ3VV39-part1'
phys_path: 'pci-0000:00:1f.2-ata-4'
load-policy:
load-request-txg: 18446744073709551615
load-rewind-policy: 2
zdb: can't open 'three': File exists
ZFS_DBGMSG(zdb) START:
ZFS_DBGMSG(zdb) END
```
but checking the datasets shows something if off:
```
zdb -d three
Dataset mos [META], ID 0, cr_txg 4, 174M, 965 objects
failed to own dataset 'three/video': Input/output error <<<============
Dataset three/www [ZPL], ID 146, cr_txg 505, 967M, 15841 objects
Dataset three/EMUmirror [ZPL], ID 626, cr_txg 52188, 8.71G, 480 objects
Dataset three/home [ZPL], ID 140, cr_txg 179, 231G, 116891 objects
Dataset three/urbackup [ZPL], ID 134, cr_txg 99, 211G, 30484298 objects
Dataset three/mayan [ZPL], ID 962, cr_txg 765918, 345M, 2466 objects
Dataset three/bionic [ZPL], ID 956, cr_txg 124432, 15.6G, 3146 objects
Dataset three/photos [ZPL], ID 14, cr_txg 87, 71.9G, 17318 objects
Dataset three [ZPL], ID 54, cr_txg 1, 106M, 47 objects
MOS object 18 (DSL directory) leaked <<<=========
MOS object 19 (DSL props) leaked <<<=========
MOS object 20 (DSL directory child map) leaked <<<=========
MOS object 21 (zap) leaked <<<=========
MOS object 22 (DSL dataset snap map) leaked <<<=========
MOS object 23 (DSL deadlist map) leaked <<<=========
Verified large_blocks feature refcount of 0 is correct
Verified large_dnode feature refcount of 0 is correct
Verified sha512 feature refcount of 0 is correct
Verified skein feature refcount of 0 is correct
Verified edonr feature refcount of 0 is correct
userobj_accounting feature refcount mismatch: 8 datasets != 9 refcount <<=====
Verified encryption feature refcount of 0 is correct
project_quota feature refcount mismatch: 8 datasets != 9 refcount <<<=========
```
I did a
```
zdb -mc three
Metaslabs:
vdev 0
metaslabs 873 offset spacemap free
--------------- ------------------- --------------- ------------
metaslab 0 offset 0 spacemap 78 free 9.25G
space map object 78:
smp_length = 0x49a68
smp_alloc = 0x1afc81800
metaslab 1 offset 400000000 spacemap 77 free 14.5G
space map object 77:
smp_length = 0x1082f8
smp_alloc = 0x5e023c00
metaslab 2 offset 800000000 spacemap 76 free 10.6G
[...]
space map object 952:
smp_length = 0x53b0
smp_alloc = 0xe000
metaslab 872 offset da000000000 spacemap 951 free 16.0G
space map object 951:
smp_length = 0x530
smp_alloc = 0x20800
Traversing all blocks to verify metadata checksums and verify nothing leaked ...
loading concrete vdev 0, metaslab 172 of 873 ...space_map_load(msp->ms_sm, msp->ms_allocatable, maptype) == 0 (0x5 == 0)
ASSERT at zdb.c:3908:load_concrete_ms_allocatable_trees()
```
I have also tried unsuccessfully to import the pool inside an Openmediavault VM that has all three disks passed in, with `zfs-0.8.5-pve1, zfs-kmod-0.8.5-pve1`.
|
1.0
|
zpool not importable and ASSERT at zdb.c:3908:load_concrete_ms_allocatable_trees() - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04
Kernel Version | 5.4.0-132-generic #148-Ubuntu SMP Mon Oct 17 16:02:06 UTC 2022
Architecture | x86_64
OpenZFS Version | zfs-0.8.3-1ubuntu12.14, zfs-kmod-0.8.3-1ubuntu12.14
### Describe the problem you're observing
I have a zpool with raidz on three 5T Seagate disks, running on Ubuntu 20.04 happily for ~2years
which can't be imported anymore.
SMART looks unsuspicious for all three devices, e.g.
```
SMART Self-test log structure revision number 1
Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
# 1 Short offline Completed without error 00% 58 -
```
The troubling pool shows up in
```
zpool import -d /dev/disk/by-id
pool: three
id: 7585041214268964297
state: ONLINE
action: The pool can be imported using its name or numeric identifier.
config:
three ONLINE
raidz1-0 ONLINE
wwn-0x5000c500d3e2552b ONLINE
wwn-0x5000c500d3f965c2 ONLINE
wwn-0x5000c500d3f92a29 ONLINE
```
but it can not be imported:
```
zpool import three
cannot import 'three': I/O error
Destroy and re-create the pool from
a backup source.
```
The ZDB looks ok (no expert here)
```
zdb -e three
Configuration for import:
vdev_children: 1
version: 5000
pool_guid: 7585041214268964297
name: 'three'
state: 0
hostid: 2677716145
hostname: 'kaestchen'
vdev_tree:
type: 'root'
id: 0
guid: 7585041214268964297
children[0]:
type: 'raidz'
id: 0
guid: 5612826130591374815
nparity: 1
metaslab_array: 69
metaslab_shift: 34
ashift: 9
asize: 15002900103168
is_log: 0
create_txg: 4
children[0]:
type: 'disk'
id: 0
guid: 4245064270772150798
whole_disk: 1
DTL: 373
create_txg: 4
path: '/dev/sda1'
devid: 'ata-ST5000LM000-2AN170_WCJ3V9AZ-part1'
phys_path: 'pci-0000:00:1f.2-ata-2'
children[1]:
type: 'disk'
id: 1
guid: 3847351160452222273
whole_disk: 1
DTL: 391
create_txg: 4
path: '/dev/sdb1'
devid: 'ata-ST5000LM000-2AN170_WCJ3VQCA-part1'
phys_path: 'pci-0000:00:1f.2-ata-3'
children[2]:
type: 'disk'
id: 2
guid: 8390181302815008501
whole_disk: 1
DTL: 390
create_txg: 4
path: '/dev/sdc1'
devid: 'ata-ST5000LM000-2AN170_WCJ3VV39-part1'
phys_path: 'pci-0000:00:1f.2-ata-4'
load-policy:
load-request-txg: 18446744073709551615
load-rewind-policy: 2
zdb: can't open 'three': File exists
ZFS_DBGMSG(zdb) START:
ZFS_DBGMSG(zdb) END
```
but checking the datasets shows something if off:
```
zdb -d three
Dataset mos [META], ID 0, cr_txg 4, 174M, 965 objects
failed to own dataset 'three/video': Input/output error <<<============
Dataset three/www [ZPL], ID 146, cr_txg 505, 967M, 15841 objects
Dataset three/EMUmirror [ZPL], ID 626, cr_txg 52188, 8.71G, 480 objects
Dataset three/home [ZPL], ID 140, cr_txg 179, 231G, 116891 objects
Dataset three/urbackup [ZPL], ID 134, cr_txg 99, 211G, 30484298 objects
Dataset three/mayan [ZPL], ID 962, cr_txg 765918, 345M, 2466 objects
Dataset three/bionic [ZPL], ID 956, cr_txg 124432, 15.6G, 3146 objects
Dataset three/photos [ZPL], ID 14, cr_txg 87, 71.9G, 17318 objects
Dataset three [ZPL], ID 54, cr_txg 1, 106M, 47 objects
MOS object 18 (DSL directory) leaked <<<=========
MOS object 19 (DSL props) leaked <<<=========
MOS object 20 (DSL directory child map) leaked <<<=========
MOS object 21 (zap) leaked <<<=========
MOS object 22 (DSL dataset snap map) leaked <<<=========
MOS object 23 (DSL deadlist map) leaked <<<=========
Verified large_blocks feature refcount of 0 is correct
Verified large_dnode feature refcount of 0 is correct
Verified sha512 feature refcount of 0 is correct
Verified skein feature refcount of 0 is correct
Verified edonr feature refcount of 0 is correct
userobj_accounting feature refcount mismatch: 8 datasets != 9 refcount <<=====
Verified encryption feature refcount of 0 is correct
project_quota feature refcount mismatch: 8 datasets != 9 refcount <<<=========
```
I did a
```
zdb -mc three
Metaslabs:
vdev 0
metaslabs 873 offset spacemap free
--------------- ------------------- --------------- ------------
metaslab 0 offset 0 spacemap 78 free 9.25G
space map object 78:
smp_length = 0x49a68
smp_alloc = 0x1afc81800
metaslab 1 offset 400000000 spacemap 77 free 14.5G
space map object 77:
smp_length = 0x1082f8
smp_alloc = 0x5e023c00
metaslab 2 offset 800000000 spacemap 76 free 10.6G
[...]
space map object 952:
smp_length = 0x53b0
smp_alloc = 0xe000
metaslab 872 offset da000000000 spacemap 951 free 16.0G
space map object 951:
smp_length = 0x530
smp_alloc = 0x20800
Traversing all blocks to verify metadata checksums and verify nothing leaked ...
loading concrete vdev 0, metaslab 172 of 873 ...space_map_load(msp->ms_sm, msp->ms_allocatable, maptype) == 0 (0x5 == 0)
ASSERT at zdb.c:3908:load_concrete_ms_allocatable_trees()
```
I have also tried unsuccessfully to import the pool inside an Openmediavault VM that has all three disks passed in, with `zfs-0.8.5-pve1, zfs-kmod-0.8.5-pve1`.
|
defect
|
zpool not importable and assert at zdb c load concrete ms allocatable trees system information type version name distribution name ubuntu distribution version kernel version generic ubuntu smp mon oct utc architecture openzfs version zfs zfs kmod describe the problem you re observing i have a zpool with raidz on three seagate disks running on ubuntu happily for which can t be imported anymore smart looks unsuspicious for all three devices e g smart self test log structure revision number num test description status remaining lifetime hours lba of first error short offline completed without error the troubling pool shows up in zpool import d dev disk by id pool three id state online action the pool can be imported using its name or numeric identifier config three online online wwn online wwn online wwn online but it can not be imported zpool import three cannot import three i o error destroy and re create the pool from a backup source the zdb looks ok no expert here zdb e three configuration for import vdev children version pool guid name three state hostid hostname kaestchen vdev tree type root id guid children type raidz id guid nparity metaslab array metaslab shift ashift asize is log create txg children type disk id guid whole disk dtl create txg path dev devid ata phys path pci ata children type disk id guid whole disk dtl create txg path dev devid ata phys path pci ata children type disk id guid whole disk dtl create txg path dev devid ata phys path pci ata load policy load request txg load rewind policy zdb can t open three file exists zfs dbgmsg zdb start zfs dbgmsg zdb end but checking the datasets shows something if off zdb d three dataset mos id cr txg objects failed to own dataset three video input output error dataset three www id cr txg objects dataset three emumirror id cr txg objects dataset three home id cr txg objects dataset three urbackup id cr txg objects dataset three mayan id cr txg objects dataset three bionic id cr txg objects dataset three photos id cr txg objects dataset three id cr txg objects mos object dsl directory leaked mos object dsl props leaked mos object dsl directory child map leaked mos object zap leaked mos object dsl dataset snap map leaked mos object dsl deadlist map leaked verified large blocks feature refcount of is correct verified large dnode feature refcount of is correct verified feature refcount of is correct verified skein feature refcount of is correct verified edonr feature refcount of is correct userobj accounting feature refcount mismatch datasets refcount verified encryption feature refcount of is correct project quota feature refcount mismatch datasets refcount i did a zdb mc three metaslabs vdev metaslabs offset spacemap free metaslab offset spacemap free space map object smp length smp alloc metaslab offset spacemap free space map object smp length smp alloc metaslab offset spacemap free space map object smp length smp alloc metaslab offset spacemap free space map object smp length smp alloc traversing all blocks to verify metadata checksums and verify nothing leaked loading concrete vdev metaslab of space map load msp ms sm msp ms allocatable maptype assert at zdb c load concrete ms allocatable trees i have also tried unsuccessfully to import the pool inside an openmediavault vm that has all three disks passed in with zfs zfs kmod
| 1
|
829,933
| 31,930,685,505
|
IssuesEvent
|
2023-09-19 07:11:59
|
GhiocelAndrei/IssueTrackerAPI
|
https://api.github.com/repos/GhiocelAndrei/IssueTrackerAPI
|
closed
|
Performance testing
|
high priority
|
The database will be populated with a high volume of data to assess the performance of the application.
Projects: 100K rows
Issues: 1M rows, randomly distributed across projects
Sprints: 1K rows, randomly distributed across projects
Users: 10K rows
A SQL script will be created for this purpose.
We'll use a third-party tool for load-testing: https://learn.microsoft.com/en-us/aspnet/core/test/load-tests?view=aspnetcore-7.0#third-party-tools then we'll test the most used endpoints:
- GET search
- GET issues/:id
- GET issues/Sprint/:id
- PATCH issues/:id
- PUT issues/AssignSprint/:id
- GET projects/all
- GET projects/:id
- GET users/all
- POST sprints/CreateWithIssues
- POST sprints/:id/Close
|
1.0
|
Performance testing - The database will be populated with a high volume of data to assess the performance of the application.
Projects: 100K rows
Issues: 1M rows, randomly distributed across projects
Sprints: 1K rows, randomly distributed across projects
Users: 10K rows
A SQL script will be created for this purpose.
We'll use a third-party tool for load-testing: https://learn.microsoft.com/en-us/aspnet/core/test/load-tests?view=aspnetcore-7.0#third-party-tools then we'll test the most used endpoints:
- GET search
- GET issues/:id
- GET issues/Sprint/:id
- PATCH issues/:id
- PUT issues/AssignSprint/:id
- GET projects/all
- GET projects/:id
- GET users/all
- POST sprints/CreateWithIssues
- POST sprints/:id/Close
|
non_defect
|
performance testing the database will be populated with a high volume of data to assess the performance of the application projects rows issues rows randomly distributed across projects sprints rows randomly distributed across projects users rows a sql script will be created for this purpose we ll use a third party tool for load testing then we ll test the most used endpoints get search get issues id get issues sprint id patch issues id put issues assignsprint id get projects all get projects id get users all post sprints createwithissues post sprints id close
| 0
|
46,771
| 13,055,973,775
|
IssuesEvent
|
2020-07-30 03:16:42
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
PolyplopiaSegment issue at Detector Level (Trac #1851)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1851
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:54",
"description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1550067174476394",
"component": "combo simulation",
"summary": "PolyplopiaSegment issue at Detector Level",
"priority": "normal",
"keywords": "",
"time": "2016-09-06T14:31:56",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
|
1.0
|
PolyplopiaSegment issue at Detector Level (Trac #1851) - Migrated from https://code.icecube.wisc.edu/ticket/1851
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:54",
"description": "Hi,\n\nI'm trying to run PolyplopiaSegment on the simulation release V04-01-13 at detector level. I end up with this error:\n\n`/data/user/saxani/environments/buildDet//env-shell.sh: line 155: 26646 Bus error PATH=$_PATH LD_LIBRARY_PATH=$_LD_LIBRARY_PATH DYLD_LIBRARY_PATH=$_DYLD_LIBRARY_PATH PYTHONPATH=$_PYTHONPATH ICETRAY_CLASSPATH=$_ICETRAY_CLASSPATH I3_PLATFORM=\"\" I3_SRC=$_I3_SRC I3_BUILD=$_I3_BUILD I3_PORTS=$_I3_PORTS I3_TESTDATA=$_I3_TESTDATA ROOTSYS=/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6.0_amd64/i3ports/root-v5.30.06 I3_SHELL=$_I3_SHELL $NEW_SHELL $ARGV`\n\nCan we reimplement the older version of PolyplopiaSegment into this release so that it can run after photon propagation?\n\nATTN: Juan Carlos\n\nThanks,\n\nSpencer",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1550067174476394",
"component": "combo simulation",
"summary": "PolyplopiaSegment issue at Detector Level",
"priority": "normal",
"keywords": "",
"time": "2016-09-06T14:31:56",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
|
defect
|
polyplopiasegment issue at detector level trac migrated from json status closed changetime description hi n ni m trying to run polyplopiasegment on the simulation release at detector level i end up with this error n n data user saxani environments builddet env shell sh line bus error path path ld library path ld library path dyld library path dyld library path pythonpath pythonpath icetray classpath icetray classpath platform src src build build ports ports testdata testdata rootsys cvmfs icecube opensciencegrid org rhel root shell shell new shell argv n ncan we reimplement the older version of polyplopiasegment into this release so that it can run after photon propagation n nattn juan carlos n nthanks n nspencer reporter saxani cc resolution fixed ts component combo simulation summary polyplopiasegment issue at detector level priority normal keywords time milestone owner juancarlos type defect
| 1
|
422,161
| 28,375,479,292
|
IssuesEvent
|
2023-04-12 20:29:03
|
RedHat-UX/red-hat-design-system
|
https://api.github.com/repos/RedHat-UX/red-hat-design-system
|
closed
|
[docs] Add icon to best practices example container
|
documentation
|
Need to add the exclamation mark in a red circle to the CSS for best practices.
<img width="774" alt="Screenshot 2023-04-12 at 10 54 57 AM" src="https://user-images.githubusercontent.com/95588923/231497481-2ab69800-f988-46cf-9873-9a1e7f132a26.png">
- Icon size: 24px
- Spacing from top left corner: 8px
The CSS class that's pulling in the border and transparent bg is `.example--palette-wrong2`.
|
1.0
|
[docs] Add icon to best practices example container - Need to add the exclamation mark in a red circle to the CSS for best practices.
<img width="774" alt="Screenshot 2023-04-12 at 10 54 57 AM" src="https://user-images.githubusercontent.com/95588923/231497481-2ab69800-f988-46cf-9873-9a1e7f132a26.png">
- Icon size: 24px
- Spacing from top left corner: 8px
The CSS class that's pulling in the border and transparent bg is `.example--palette-wrong2`.
|
non_defect
|
add icon to best practices example container need to add the exclamation mark in a red circle to the css for best practices img width alt screenshot at am src icon size spacing from top left corner the css class that s pulling in the border and transparent bg is example palette
| 0
|
340,912
| 24,677,001,112
|
IssuesEvent
|
2022-10-18 17:51:19
|
VDVde/OJP
|
https://api.github.com/repos/VDVde/OJP
|
closed
|
complexType `ProductCategoryRefStructure` is defined but never used
|
bug documentation
|
The commit https://github.com/VDVde/OJP/commit/3708ad4fbc36fd71ee5b0443426aec0990a82cb9 introduced the complexType `ProductCategoryRefStructure`, but it is not used anywhere, as the `ProductCategoryRef` uses `siri:ProductCategoryRefStructure` instead of `ProductCategoryRefStructure`.
```<xs:element name="ProductCategoryRef" type="siri:ProductCategoryRefStructure">```
As the definition of `ProductCategoryRefStructure` and `ProductCategoryCodeType` are identical in OJP and SIRI I don't see a reason to keep those unused definitions.
OJP_Common.xsd:
```
<xs:complexType name="ProductCategoryRefStructure">
<xs:annotation>
<xs:documentation>Reference to a product category. As defined in NeTEx and SIRI, a product category is a classification for VEHICLE JOURNEYs to express some common properties of journeys for marketing and fare products, e.g. to identify boats that run with steam in Switzerland ("BAV")</xs:documentation>
</xs:annotation>
<xs:simpleContent>
<xs:extension base="ProductCategoryCodeType"/>
</xs:simpleContent>
</xs:complexType>
<xs:simpleType name="ProductCategoryCodeType">
<xs:annotation>
<xs:documentation>Type for identifier of an product category code.</xs:documentation>
</xs:annotation>
<xs:restriction base="xs:NMTOKEN"/>
</xs:simpleType>
<xs:element name="ProductCategoryRef" type="siri:ProductCategoryRefStructure">
<xs:annotation>
<xs:documentation>Reference to a product category. Product categories should be defined once and used uniformly in all channels (e.g. NeTEx, SIRI, OJP)</xs:documentation>
</xs:annotation>
</xs:element>
```
siri_feature_support-v0.2.0.xsd:
```
<xsd:simpleType name="ProductCategoryCodeType">
<xsd:annotation>
<xsd:documentation>Type for identifier of a TYPE OF PRODUCT CATEGORY.</xsd:documentation>
</xsd:annotation>
<xsd:restriction base="xsd:NMTOKEN"/>
</xsd:simpleType>
<xsd:complexType name="ProductCategoryRefStructure">
<xsd:annotation>
<xsd:documentation>Type for reference to a TYPE OF PRODUCT CATEGORY.</xsd:documentation>
</xsd:annotation>
<xsd:simpleContent>
<xsd:extension base="ProductCategoryCodeType"/>
</xsd:simpleContent>
</xsd:complexType>
```
|
1.0
|
complexType `ProductCategoryRefStructure` is defined but never used - The commit https://github.com/VDVde/OJP/commit/3708ad4fbc36fd71ee5b0443426aec0990a82cb9 introduced the complexType `ProductCategoryRefStructure`, but it is not used anywhere, as the `ProductCategoryRef` uses `siri:ProductCategoryRefStructure` instead of `ProductCategoryRefStructure`.
```<xs:element name="ProductCategoryRef" type="siri:ProductCategoryRefStructure">```
As the definition of `ProductCategoryRefStructure` and `ProductCategoryCodeType` are identical in OJP and SIRI I don't see a reason to keep those unused definitions.
OJP_Common.xsd:
```
<xs:complexType name="ProductCategoryRefStructure">
<xs:annotation>
<xs:documentation>Reference to a product category. As defined in NeTEx and SIRI, a product category is a classification for VEHICLE JOURNEYs to express some common properties of journeys for marketing and fare products, e.g. to identify boats that run with steam in Switzerland ("BAV")</xs:documentation>
</xs:annotation>
<xs:simpleContent>
<xs:extension base="ProductCategoryCodeType"/>
</xs:simpleContent>
</xs:complexType>
<xs:simpleType name="ProductCategoryCodeType">
<xs:annotation>
<xs:documentation>Type for identifier of an product category code.</xs:documentation>
</xs:annotation>
<xs:restriction base="xs:NMTOKEN"/>
</xs:simpleType>
<xs:element name="ProductCategoryRef" type="siri:ProductCategoryRefStructure">
<xs:annotation>
<xs:documentation>Reference to a product category. Product categories should be defined once and used uniformly in all channels (e.g. NeTEx, SIRI, OJP)</xs:documentation>
</xs:annotation>
</xs:element>
```
siri_feature_support-v0.2.0.xsd:
```
<xsd:simpleType name="ProductCategoryCodeType">
<xsd:annotation>
<xsd:documentation>Type for identifier of a TYPE OF PRODUCT CATEGORY.</xsd:documentation>
</xsd:annotation>
<xsd:restriction base="xsd:NMTOKEN"/>
</xsd:simpleType>
<xsd:complexType name="ProductCategoryRefStructure">
<xsd:annotation>
<xsd:documentation>Type for reference to a TYPE OF PRODUCT CATEGORY.</xsd:documentation>
</xsd:annotation>
<xsd:simpleContent>
<xsd:extension base="ProductCategoryCodeType"/>
</xsd:simpleContent>
</xsd:complexType>
```
|
non_defect
|
complextype productcategoryrefstructure is defined but never used the commit introduced the complextype productcategoryrefstructure but it is not used anywhere as the productcategoryref uses siri productcategoryrefstructure instead of productcategoryrefstructure as the definition of productcategoryrefstructure and productcategorycodetype are identical in ojp and siri i don t see a reason to keep those unused definitions ojp common xsd reference to a product category as defined in netex and siri a product category is a classification for vehicle journeys to express some common properties of journeys for marketing and fare products e g to identify boats that run with steam in switzerland bav type for identifier of an product category code reference to a product category product categories should be defined once and used uniformly in all channels e g netex siri ojp siri feature support xsd type for identifier of a type of product category type for reference to a type of product category
| 0
|
114,201
| 4,621,420,494
|
IssuesEvent
|
2016-09-27 01:11:29
|
CascadesCarnivoreProject/Carnassial
|
https://api.github.com/repos/CascadesCarnivoreProject/Carnassial
|
closed
|
Timelapse UX: filter is not applied when file properties are changed
|
Low Priority fix
|
When a filter other than all is selected editing a file's properties can remove the file from the set the filter would select. For example,
- changing any property when a custom filter considers that property
- changing image quality with light, dark, ok, missing, or corrupted filters
This behaviour seems both a feature and a bug. It allows a user to trivially undo accidental or erroneous edits. But it's also confusing in that the status bar continues to indicate the image matches the filter and when the user might be thinking enough about the filter behaviour to be expecting Timelapse to figure out the edit and move on to another image in response.
As an initial response I'm tempted to say the core behaviour here is by design and that if the user wants the currently viewed file to drop out of the selected set they need to reapply the filter. In that context some type of indicator might be appropriate, perhaps along the lines of "Image Number: foo of bar files matching filter when it was applied, file no longer matches filter" in the status bar.
|
1.0
|
Timelapse UX: filter is not applied when file properties are changed - When a filter other than all is selected editing a file's properties can remove the file from the set the filter would select. For example,
- changing any property when a custom filter considers that property
- changing image quality with light, dark, ok, missing, or corrupted filters
This behaviour seems both a feature and a bug. It allows a user to trivially undo accidental or erroneous edits. But it's also confusing in that the status bar continues to indicate the image matches the filter and when the user might be thinking enough about the filter behaviour to be expecting Timelapse to figure out the edit and move on to another image in response.
As an initial response I'm tempted to say the core behaviour here is by design and that if the user wants the currently viewed file to drop out of the selected set they need to reapply the filter. In that context some type of indicator might be appropriate, perhaps along the lines of "Image Number: foo of bar files matching filter when it was applied, file no longer matches filter" in the status bar.
|
non_defect
|
timelapse ux filter is not applied when file properties are changed when a filter other than all is selected editing a file s properties can remove the file from the set the filter would select for example changing any property when a custom filter considers that property changing image quality with light dark ok missing or corrupted filters this behaviour seems both a feature and a bug it allows a user to trivially undo accidental or erroneous edits but it s also confusing in that the status bar continues to indicate the image matches the filter and when the user might be thinking enough about the filter behaviour to be expecting timelapse to figure out the edit and move on to another image in response as an initial response i m tempted to say the core behaviour here is by design and that if the user wants the currently viewed file to drop out of the selected set they need to reapply the filter in that context some type of indicator might be appropriate perhaps along the lines of image number foo of bar files matching filter when it was applied file no longer matches filter in the status bar
| 0
|
25,763
| 4,441,174,780
|
IssuesEvent
|
2016-08-19 08:13:20
|
PSICQUIC/psicquic-registry
|
https://api.github.com/repos/PSICQUIC/psicquic-registry
|
opened
|
baderlab REST/SOAP services refusing connections despite being marked as active in registry
|
auto-migrated Priority-Medium Type-Defect
|
_From @GoogleCodeExporter on March 18, 2015 11:1_
```
http://webservice.baderlab.org:8180/psicquic-ws/webservices/current/search/query/brca2?format=tab25
http://webservice.baderlab.org:8180/psi-gm/webservices/current/search/query/brca2?format=tab25
http://www.baderlab.org:8180/psicquic-interoporc-ws/webservices/current/search/query/brca2?format=tab25
```
Original issue reported on code.google.com by `kenny.br...@gmail.com` on 1 Aug 2012 at 1:07
_Copied from original issue: MICommunity/psicquic#31_
|
1.0
|
baderlab REST/SOAP services refusing connections despite being marked as active in registry - _From @GoogleCodeExporter on March 18, 2015 11:1_
```
http://webservice.baderlab.org:8180/psicquic-ws/webservices/current/search/query/brca2?format=tab25
http://webservice.baderlab.org:8180/psi-gm/webservices/current/search/query/brca2?format=tab25
http://www.baderlab.org:8180/psicquic-interoporc-ws/webservices/current/search/query/brca2?format=tab25
```
Original issue reported on code.google.com by `kenny.br...@gmail.com` on 1 Aug 2012 at 1:07
_Copied from original issue: MICommunity/psicquic#31_
|
defect
|
baderlab rest soap services refusing connections despite being marked as active in registry from googlecodeexporter on march original issue reported on code google com by kenny br gmail com on aug at copied from original issue micommunity psicquic
| 1
|
82,384
| 23,760,423,671
|
IssuesEvent
|
2022-09-01 08:24:35
|
Crocoblock/suggestions
|
https://api.github.com/repos/Crocoblock/suggestions
|
closed
|
Hover effect on JetWooBuilder product thumbnails causes a bug when enabling webp images with <picture>
|
JetWooBuilder
|
Hello, the hover effect on JetWooBuilder product thumbnails causes a bug when enabling webp images with <picture>
|
1.0
|
Hover effect on JetWooBuilder product thumbnails causes a bug when enabling webp images with <picture> - Hello, the hover effect on JetWooBuilder product thumbnails causes a bug when enabling webp images with <picture>
|
non_defect
|
hover effect on jetwoobuilder product thumbnails causes a bug when enabling webp images with hello the hover effect on jetwoobuilder product thumbnails causes a bug when enabling webp images with
| 0
|
46,832
| 13,055,984,757
|
IssuesEvent
|
2020-07-30 03:18:28
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[MuonGun] Surfaces refactor broke deserialization of pre-IceSim5 S frames (Trac #1956)
|
Incomplete Migration Migrated from Trac combo core defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1956
```json
{
"status": "closed",
"changetime": "2017-03-14T20:23:10",
"description": "Trying to deserialize an S frame written with IceSim 4 with current software fails with \n{{{\nFATAL (phys-services): Version 117 is from the future (SamplingSurface.cxx:50 in void I3Surfaces::SamplingSurface::serialize(Archive&, unsigned int) [with Archive = icecube::archive::portable_binary_iarchive])\n}}}\n\nThis is probably because the refactor added a new layer in the inheritance tree, the current code tries to read a class ID and version from the stream that are not there. While empty base classes do not take up space in memory, they turn out to matter quite a bit for serialization.",
"reporter": "jvansanten",
"cc": "",
"resolution": "invalid",
"_ts": "1489522990898099",
"component": "combo core",
"summary": "[MuonGun] Surfaces refactor broke deserialization of pre-IceSim5 S frames",
"priority": "critical",
"keywords": "",
"time": "2017-03-14T15:16:30",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
1.0
|
[MuonGun] Surfaces refactor broke deserialization of pre-IceSim5 S frames (Trac #1956) - Migrated from https://code.icecube.wisc.edu/ticket/1956
```json
{
"status": "closed",
"changetime": "2017-03-14T20:23:10",
"description": "Trying to deserialize an S frame written with IceSim 4 with current software fails with \n{{{\nFATAL (phys-services): Version 117 is from the future (SamplingSurface.cxx:50 in void I3Surfaces::SamplingSurface::serialize(Archive&, unsigned int) [with Archive = icecube::archive::portable_binary_iarchive])\n}}}\n\nThis is probably because the refactor added a new layer in the inheritance tree, the current code tries to read a class ID and version from the stream that are not there. While empty base classes do not take up space in memory, they turn out to matter quite a bit for serialization.",
"reporter": "jvansanten",
"cc": "",
"resolution": "invalid",
"_ts": "1489522990898099",
"component": "combo core",
"summary": "[MuonGun] Surfaces refactor broke deserialization of pre-IceSim5 S frames",
"priority": "critical",
"keywords": "",
"time": "2017-03-14T15:16:30",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
defect
|
surfaces refactor broke deserialization of pre s frames trac migrated from json status closed changetime description trying to deserialize an s frame written with icesim with current software fails with n nfatal phys services version is from the future samplingsurface cxx in void samplingsurface serialize archive unsigned int n n nthis is probably because the refactor added a new layer in the inheritance tree the current code tries to read a class id and version from the stream that are not there while empty base classes do not take up space in memory they turn out to matter quite a bit for serialization reporter jvansanten cc resolution invalid ts component combo core summary surfaces refactor broke deserialization of pre s frames priority critical keywords time milestone owner type defect
| 1
|
217,684
| 16,726,000,351
|
IssuesEvent
|
2021-06-10 13:04:59
|
catboost/catboost
|
https://api.github.com/repos/catboost/catboost
|
closed
|
Catboost ranking objective
|
documentation
|
Problem: I was comparing the models using different objectives for ranking. It seems that catboost implements several novel objectives that other algorithms haven't used before. I would love to learn more about them but the documentation on how those objective are computed are not detailed enough (https://tech.yandex.com/catboost/doc/dg/concepts/loss-functions-docpage/#loss-functions__ranking) . Could anyone point me to other resources regarding the objectives?
More specifically, I want to know the difference between PairLogit/yetiRank and PairLogitPairwise/YetiRankPairwise. Also some more documentations on QuerySoftMax, QueryRMSE and QueryCrossEntropy would be great.
catboost version: {version}
Operating System: {OS}
CPU: {CPU}
# GPU: {GPU}
|
1.0
|
Catboost ranking objective - Problem: I was comparing the models using different objectives for ranking. It seems that catboost implements several novel objectives that other algorithms haven't used before. I would love to learn more about them but the documentation on how those objective are computed are not detailed enough (https://tech.yandex.com/catboost/doc/dg/concepts/loss-functions-docpage/#loss-functions__ranking) . Could anyone point me to other resources regarding the objectives?
More specifically, I want to know the difference between PairLogit/yetiRank and PairLogitPairwise/YetiRankPairwise. Also some more documentations on QuerySoftMax, QueryRMSE and QueryCrossEntropy would be great.
catboost version: {version}
Operating System: {OS}
CPU: {CPU}
# GPU: {GPU}
|
non_defect
|
catboost ranking objective problem i was comparing the models using different objectives for ranking it seems that catboost implements several novel objectives that other algorithms haven t used before i would love to learn more about them but the documentation on how those objective are computed are not detailed enough could anyone point me to other resources regarding the objectives more specifically i want to know the difference between pairlogit yetirank and pairlogitpairwise yetirankpairwise also some more documentations on querysoftmax queryrmse and querycrossentropy would be great catboost version version operating system os cpu cpu gpu gpu
| 0
|
252,805
| 21,632,106,647
|
IssuesEvent
|
2022-05-05 10:48:43
|
Tencent/bk-ci
|
https://api.github.com/repos/Tencent/bk-ci
|
closed
|
feat:编译加速支持gcc参数 -fprofile-dir
|
for gray for test done streams/for gray
|
-fprofile-dir=path 把程序运行后产生的gcda文件放在path目录下,为避免同名文件冲突,编译产生的gcno文件名是由#组织的绝对路径
|
1.0
|
feat:编译加速支持gcc参数 -fprofile-dir - -fprofile-dir=path 把程序运行后产生的gcda文件放在path目录下,为避免同名文件冲突,编译产生的gcno文件名是由#组织的绝对路径
|
non_defect
|
feat 编译加速支持gcc参数 fprofile dir fprofile dir path 把程序运行后产生的gcda文件放在path目录下,为避免同名文件冲突,编译产生的gcno文件名是由 组织的绝对路径
| 0
|
33,642
| 9,199,270,349
|
IssuesEvent
|
2019-03-07 14:36:08
|
quicklisp/quicklisp-projects
|
https://api.github.com/repos/quicklisp/quicklisp-projects
|
closed
|
Please add SPELL
|
canbuild
|
https://github.com/robert-strandh/Spell
This is a spellchecking library in standard Common Lisp without any external dependencies.
|
1.0
|
Please add SPELL - https://github.com/robert-strandh/Spell
This is a spellchecking library in standard Common Lisp without any external dependencies.
|
non_defect
|
please add spell this is a spellchecking library in standard common lisp without any external dependencies
| 0
|
400,596
| 11,777,543,099
|
IssuesEvent
|
2020-03-16 14:58:29
|
prometheus/prometheus
|
https://api.github.com/repos/prometheus/prometheus
|
closed
|
Allow exporting graphed data (as CSV?)
|
component/ui kind/enhancement priority/Pmaybe
|
I've talked to people introducing Prometheus in a more traditional big company - when trying to convince deciders there that Prometheus is the right tool for them, the more Excel-type of people have been asking for simple ways to export graph data as CSV. Grafana supports some kind of graph export, but it's not ideal for their needs.
We could think about adding a query/graph data export button to the Prometheus web UI. Questions are:
- do we want that?
- what formats should it support? (I can imagine JSON and CSV being the obvious ones)
- if we do support CSV, do we want to drag in some JS dependency to be sure to get all the quoting right?
|
1.0
|
Allow exporting graphed data (as CSV?) - I've talked to people introducing Prometheus in a more traditional big company - when trying to convince deciders there that Prometheus is the right tool for them, the more Excel-type of people have been asking for simple ways to export graph data as CSV. Grafana supports some kind of graph export, but it's not ideal for their needs.
We could think about adding a query/graph data export button to the Prometheus web UI. Questions are:
- do we want that?
- what formats should it support? (I can imagine JSON and CSV being the obvious ones)
- if we do support CSV, do we want to drag in some JS dependency to be sure to get all the quoting right?
|
non_defect
|
allow exporting graphed data as csv i ve talked to people introducing prometheus in a more traditional big company when trying to convince deciders there that prometheus is the right tool for them the more excel type of people have been asking for simple ways to export graph data as csv grafana supports some kind of graph export but it s not ideal for their needs we could think about adding a query graph data export button to the prometheus web ui questions are do we want that what formats should it support i can imagine json and csv being the obvious ones if we do support csv do we want to drag in some js dependency to be sure to get all the quoting right
| 0
|
13,409
| 2,755,621,308
|
IssuesEvent
|
2015-04-26 20:19:01
|
zerfgog/zinger
|
https://api.github.com/repos/zerfgog/zinger
|
closed
|
Lua errors :)
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. finish all holes
2. enter "zd forcenext" in console
3. ???
4. Profit.
It'll start reporting:
[gamemodes\zinger\gamemode\sv_gameplay.lua:405] attempt to index field '?' (a
nil value)
[gamemodes\zinger\gamemode\sv_gameplay.lua:435] attempt to index field '?' (a
nil value)(Hook: Think)
```
Original issue reported on code.google.com by `Borstym...@gmail.com` on 10 Sep 2010 at 6:56
|
1.0
|
Lua errors :) - ```
What steps will reproduce the problem?
1. finish all holes
2. enter "zd forcenext" in console
3. ???
4. Profit.
It'll start reporting:
[gamemodes\zinger\gamemode\sv_gameplay.lua:405] attempt to index field '?' (a
nil value)
[gamemodes\zinger\gamemode\sv_gameplay.lua:435] attempt to index field '?' (a
nil value)(Hook: Think)
```
Original issue reported on code.google.com by `Borstym...@gmail.com` on 10 Sep 2010 at 6:56
|
defect
|
lua errors what steps will reproduce the problem finish all holes enter zd forcenext in console profit it ll start reporting attempt to index field a nil value attempt to index field a nil value hook think original issue reported on code google com by borstym gmail com on sep at
| 1
|
326,398
| 24,082,450,967
|
IssuesEvent
|
2022-09-19 07:57:43
|
mbucchia/OpenXR-Toolkit
|
https://api.github.com/repos/mbucchia/OpenXR-Toolkit
|
closed
|
[Doc] OpenXR Toolkit does not detect Eye tracking of the Reverb G2 Omnicept
|
documentation
|
**Describe the bug**
The OpenXR Toolkit Version 1.1.0 does not detect the eye tracking feature of the connected HP Reverb G2 Omnicept. The last 12 month I used the normal HP Reverb G2. During the upgrade to the Omnicept I just disconneted the G2 and connected the Omnicept and downloaded all available updates provided by Microsoft. At the beginning everythings looks normal. As I tried to setup the Omnicept with the HP Setup tool, I got the message "connect your VR device using a USB 3.0 port" even the VR device (the Omnicept) was connected. So, I realised there is something wrong. After starting the MSFS2020 to test if the eye tracking is possible/shown in the OpenXR Setup, I got the validation, it isn´t working.
As conclusion I deinstalled all app´s like WMR, OpenXR, OpenXR Toolkit and all HP Apps/Software. Than restart the PC and run the Registry Cleaner of aShampoo, followed by another restart of the PC.
Than new installation of the WMR and the HP Setuptool for the Omnicept, so far everything worked fine. Finally I installed the OpenXR and the OpenXR Toolkit, checked if new Windows updates are available (Windows System and Microsoft Store). Once again in the MSFS2020, opening the OpenXR Toolkit setup in VR, no eye tracking.
**Environment**
- Application/Game and its version: Microsoft Flight Simulator 2020, all available updates installed
- Make and model of VR device: HP Reverb G2 Omnicept
- Version of the OpenXR runtime: 110.2204.7007
- Make and model of GPU: Nvidia GeForce 3090 RTX TUF OC
**To Reproduce**
Steps to reproduce the behavior: see above discrip
1. Run '....' game
2. Enable option '....'
3. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem. Use the Companion app to enable screenshots, then use Ctrl+F12 to capture. The file is placed in `%LocalAppData%\OpenXR-Toolkit\screenshots`.
[XR_APILAYER_NOVENDOR_toolkit.log](https://github.com/mbucchia/OpenXR-Toolkit/files/8560864/XR_APILAYER_NOVENDOR_toolkit.log)
|
1.0
|
[Doc] OpenXR Toolkit does not detect Eye tracking of the Reverb G2 Omnicept - **Describe the bug**
The OpenXR Toolkit Version 1.1.0 does not detect the eye tracking feature of the connected HP Reverb G2 Omnicept. The last 12 month I used the normal HP Reverb G2. During the upgrade to the Omnicept I just disconneted the G2 and connected the Omnicept and downloaded all available updates provided by Microsoft. At the beginning everythings looks normal. As I tried to setup the Omnicept with the HP Setup tool, I got the message "connect your VR device using a USB 3.0 port" even the VR device (the Omnicept) was connected. So, I realised there is something wrong. After starting the MSFS2020 to test if the eye tracking is possible/shown in the OpenXR Setup, I got the validation, it isn´t working.
As conclusion I deinstalled all app´s like WMR, OpenXR, OpenXR Toolkit and all HP Apps/Software. Than restart the PC and run the Registry Cleaner of aShampoo, followed by another restart of the PC.
Than new installation of the WMR and the HP Setuptool for the Omnicept, so far everything worked fine. Finally I installed the OpenXR and the OpenXR Toolkit, checked if new Windows updates are available (Windows System and Microsoft Store). Once again in the MSFS2020, opening the OpenXR Toolkit setup in VR, no eye tracking.
**Environment**
- Application/Game and its version: Microsoft Flight Simulator 2020, all available updates installed
- Make and model of VR device: HP Reverb G2 Omnicept
- Version of the OpenXR runtime: 110.2204.7007
- Make and model of GPU: Nvidia GeForce 3090 RTX TUF OC
**To Reproduce**
Steps to reproduce the behavior: see above discrip
1. Run '....' game
2. Enable option '....'
3. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem. Use the Companion app to enable screenshots, then use Ctrl+F12 to capture. The file is placed in `%LocalAppData%\OpenXR-Toolkit\screenshots`.
[XR_APILAYER_NOVENDOR_toolkit.log](https://github.com/mbucchia/OpenXR-Toolkit/files/8560864/XR_APILAYER_NOVENDOR_toolkit.log)
|
non_defect
|
openxr toolkit does not detect eye tracking of the reverb omnicept describe the bug the openxr toolkit version does not detect the eye tracking feature of the connected hp reverb omnicept the last month i used the normal hp reverb during the upgrade to the omnicept i just disconneted the and connected the omnicept and downloaded all available updates provided by microsoft at the beginning everythings looks normal as i tried to setup the omnicept with the hp setup tool i got the message connect your vr device using a usb port even the vr device the omnicept was connected so i realised there is something wrong after starting the to test if the eye tracking is possible shown in the openxr setup i got the validation it isn´t working as conclusion i deinstalled all app´s like wmr openxr openxr toolkit and all hp apps software than restart the pc and run the registry cleaner of ashampoo followed by another restart of the pc than new installation of the wmr and the hp setuptool for the omnicept so far everything worked fine finally i installed the openxr and the openxr toolkit checked if new windows updates are available windows system and microsoft store once again in the opening the openxr toolkit setup in vr no eye tracking environment application game and its version microsoft flight simulator all available updates installed make and model of vr device hp reverb omnicept version of the openxr runtime make and model of gpu nvidia geforce rtx tuf oc to reproduce steps to reproduce the behavior see above discrip run game enable option see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem use the companion app to enable screenshots then use ctrl to capture the file is placed in localappdata openxr toolkit screenshots
| 0
|
379,447
| 26,371,577,779
|
IssuesEvent
|
2023-01-11 21:11:22
|
p2-inc/phasetwo-docs
|
https://api.github.com/repos/p2-inc/phasetwo-docs
|
closed
|
[Docs] Clarify licensing in intro->open source section
|
documentation priority
|
We use AGPL for our Keycloak extensions:
- To allow anyone to use and host them in an unmodified fashion.
- To prevent a competitor from using them without releasing modifications and their own Keycloak improvements.
Our position is that the use of AGPL *does not* require users to use the same license for their whole stack, but *only* for extension and Keycloak modifications.
Perhaps the most important point is that "dual" licensing is available to paying customers, where they can use and host our extensions without the AGPL license requirements.
|
1.0
|
[Docs] Clarify licensing in intro->open source section - We use AGPL for our Keycloak extensions:
- To allow anyone to use and host them in an unmodified fashion.
- To prevent a competitor from using them without releasing modifications and their own Keycloak improvements.
Our position is that the use of AGPL *does not* require users to use the same license for their whole stack, but *only* for extension and Keycloak modifications.
Perhaps the most important point is that "dual" licensing is available to paying customers, where they can use and host our extensions without the AGPL license requirements.
|
non_defect
|
clarify licensing in intro open source section we use agpl for our keycloak extensions to allow anyone to use and host them in an unmodified fashion to prevent a competitor from using them without releasing modifications and their own keycloak improvements our position is that the use of agpl does not require users to use the same license for their whole stack but only for extension and keycloak modifications perhaps the most important point is that dual licensing is available to paying customers where they can use and host our extensions without the agpl license requirements
| 0
|
260,498
| 22,626,155,410
|
IssuesEvent
|
2022-06-30 10:50:20
|
Djazaanti/portfolio
|
https://api.github.com/repos/Djazaanti/portfolio
|
closed
|
Homepage : Rename 'posts' in HomeController as 'parcours'
|
Latest changes
|
That's not a posts but my career.
rename in : HomeController,
rename the function in UserModel as : getCareer
and, see if can do qomething for DB : may be just write the informations in portfolioModal ...
|
1.0
|
Homepage : Rename 'posts' in HomeController as 'parcours' - That's not a posts but my career.
rename in : HomeController,
rename the function in UserModel as : getCareer
and, see if can do qomething for DB : may be just write the informations in portfolioModal ...
|
non_defect
|
homepage rename posts in homecontroller as parcours that s not a posts but my career rename in homecontroller rename the function in usermodel as getcareer and see if can do qomething for db may be just write the informations in portfoliomodal
| 0
|
63,063
| 17,361,246,383
|
IssuesEvent
|
2021-07-29 20:59:44
|
Questie/Questie
|
https://api.github.com/repos/Questie/Questie
|
opened
|
Journey appends a quest abandonment only when the quest is not completed before
|
Type - Defect
|
This might be low prioriy unintentional behavior.
"Quest Abandoned" are appended to journey only if the player has not yet completed the quest.
If a quest is repeatable, then a player can complete it once and take it again and abandon this time.
https://github.com/Questie/Questie/blob/60ef402a359f54970862e3f80d82024d0d7191ab/Modules/Journey/QuestieJourney.lua#L158-L186
v6.5.1
|
1.0
|
Journey appends a quest abandonment only when the quest is not completed before - This might be low prioriy unintentional behavior.
"Quest Abandoned" are appended to journey only if the player has not yet completed the quest.
If a quest is repeatable, then a player can complete it once and take it again and abandon this time.
https://github.com/Questie/Questie/blob/60ef402a359f54970862e3f80d82024d0d7191ab/Modules/Journey/QuestieJourney.lua#L158-L186
v6.5.1
|
defect
|
journey appends a quest abandonment only when the quest is not completed before this might be low prioriy unintentional behavior quest abandoned are appended to journey only if the player has not yet completed the quest if a quest is repeatable then a player can complete it once and take it again and abandon this time
| 1
|
747,350
| 26,082,492,334
|
IssuesEvent
|
2022-12-25 15:36:21
|
cov-lineages/pango-designation
|
https://api.github.com/repos/cov-lineages/pango-designation
|
closed
|
BA.2.75.6 with S:F486V [16 seqs as of 2022-09-29]
|
BA.2.75 low priority
|
Noticed that BA.2.75.6 (S:346T) picked up S:F486V in 4 sequences in England and US
```txt
hCoV-19/England/PHEC-YYG4WPZ/2022|EPI_ISL_15021194|2022-09-06
hCoV-19/England/QEUH-3F89F82/2022|EPI_ISL_14948777|2022-09-05
hCoV-19/England/LSPA-3F7C438/2022|EPI_ISL_14810559|2022-08-30
hCoV-19/Australia/NSW-ICPMR-33405/2022|EPI_ISL_14993456|2022-09-05
hCoV-19/USA/NJ-GBW-LWPOAAAB19540/2022|EPI_ISL_15017960|2022-09-10
```
The AUS sequence has S:357K as well. Not sure whether these are monophyletic or not. Needs to be seen.
Two of them were uploaded today.
GISAID query: `NSP3_S403L,Spike_R346T, Spike_N460K, Spike_F486V`
The RBD is very similar to BQ.1.1 - more than any other BA.2.75 so far.
|
1.0
|
BA.2.75.6 with S:F486V [16 seqs as of 2022-09-29] - Noticed that BA.2.75.6 (S:346T) picked up S:F486V in 4 sequences in England and US
```txt
hCoV-19/England/PHEC-YYG4WPZ/2022|EPI_ISL_15021194|2022-09-06
hCoV-19/England/QEUH-3F89F82/2022|EPI_ISL_14948777|2022-09-05
hCoV-19/England/LSPA-3F7C438/2022|EPI_ISL_14810559|2022-08-30
hCoV-19/Australia/NSW-ICPMR-33405/2022|EPI_ISL_14993456|2022-09-05
hCoV-19/USA/NJ-GBW-LWPOAAAB19540/2022|EPI_ISL_15017960|2022-09-10
```
The AUS sequence has S:357K as well. Not sure whether these are monophyletic or not. Needs to be seen.
Two of them were uploaded today.
GISAID query: `NSP3_S403L,Spike_R346T, Spike_N460K, Spike_F486V`
The RBD is very similar to BQ.1.1 - more than any other BA.2.75 so far.
|
non_defect
|
ba with s noticed that ba s picked up s in sequences in england and us txt hcov england phec epi isl hcov england qeuh epi isl hcov england lspa epi isl hcov australia nsw icpmr epi isl hcov usa nj gbw epi isl the aus sequence has s as well not sure whether these are monophyletic or not needs to be seen two of them were uploaded today gisaid query spike spike spike the rbd is very similar to bq more than any other ba so far
| 0
|
54,892
| 14,007,481,934
|
IssuesEvent
|
2020-10-28 21:39:25
|
AeroScripts/QuestieDev
|
https://api.github.com/repos/AeroScripts/QuestieDev
|
closed
|
Errors showing due to potentially bad data for the quest 'The Challenge' - part of the Tier 0.5 dungeon set questline.
|
Questie - Quest Issue Type - Defect
|
## Bug description
The bug/issue is that upon login and/or switching to different zones while having the quest 'The Challange' active the chat is filled with error messages.
## Screenshots


## Questie version
Running version 6.1.0
|
1.0
|
Errors showing due to potentially bad data for the quest 'The Challenge' - part of the Tier 0.5 dungeon set questline. - ## Bug description
The bug/issue is that upon login and/or switching to different zones while having the quest 'The Challange' active the chat is filled with error messages.
## Screenshots


## Questie version
Running version 6.1.0
|
defect
|
errors showing due to potentially bad data for the quest the challenge part of the tier dungeon set questline bug description the bug issue is that upon login and or switching to different zones while having the quest the challange active the chat is filled with error messages screenshots questie version running version
| 1
|
41,020
| 12,812,506,847
|
IssuesEvent
|
2020-07-04 06:53:47
|
shrivastava-prateek/angularjs-es6-webpack
|
https://api.github.com/repos/shrivastava-prateek/angularjs-es6-webpack
|
opened
|
CVE-2017-1000048 (High) detected in qs-6.2.0.tgz
|
security vulnerability
|
## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.2.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.0.tgz">https://registry.npmjs.org/qs/-/qs-6.2.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/angularjs-es6-webpack/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/angularjs-es6-webpack/node_modules/browser-sync/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.12.12.tgz (Root Library)
- :x: **qs-6.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shrivastava-prateek/angularjs-es6-webpack/commit/5a7519c9340d9d27cd18c80cc9093d3b1193db9d">5a7519c9340d9d27cd18c80cc9093d3b1193db9d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/ljharb/qs/commit/c709f6e3ef2ed324f17c43369e1d45ad351e86e6">https://github.com/ljharb/qs/commit/c709f6e3ef2ed324f17c43369e1d45ad351e86e6</a></p>
<p>Release Date: 2017-03-06</p>
<p>Fix Resolution: Replace or update the following file: v6.0.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-1000048 (High) detected in qs-6.2.0.tgz - ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.2.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.0.tgz">https://registry.npmjs.org/qs/-/qs-6.2.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/angularjs-es6-webpack/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/angularjs-es6-webpack/node_modules/browser-sync/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.12.12.tgz (Root Library)
- :x: **qs-6.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shrivastava-prateek/angularjs-es6-webpack/commit/5a7519c9340d9d27cd18c80cc9093d3b1193db9d">5a7519c9340d9d27cd18c80cc9093d3b1193db9d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/ljharb/qs/commit/c709f6e3ef2ed324f17c43369e1d45ad351e86e6">https://github.com/ljharb/qs/commit/c709f6e3ef2ed324f17c43369e1d45ad351e86e6</a></p>
<p>Release Date: 2017-03-06</p>
<p>Fix Resolution: Replace or update the following file: v6.0.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file tmp ws scm angularjs webpack package json path to vulnerable library tmp ws scm angularjs webpack node modules browser sync node modules qs package json dependency hierarchy browser sync tgz root library x qs tgz vulnerable library found in head commit a href vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file step up your open source security game with whitesource
| 0
|
244,312
| 7,873,416,248
|
IssuesEvent
|
2018-06-25 14:13:59
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[studio] catch exception if dependency on ContentServiceImpl.contentExists in dependency resolver
|
CI bug priority: highest!
|
### Expected behavior
Catch and report exceptions in this method:
`org.craftercms.studio.impl.v1.service.dependency.RegexDependencyResolver.resolve(RegexDependencyResolver.java:79)
`
That are a result of this method call:
`org.craftercms.studio.impl.v1.service.content.ContentServiceImpl.contentExists(ContentServiceImpl.java:159)
`
An ERROR log message should say that:
Dependencies in File {current file being scanned} could not be resolved for path {path that failed exist} due to the following error: {error}
### Actual behavior
Exceptions with contentExists exceptions are not caught during dependency resolution operation and the operation (site import, write operation) fails as a result.
### Steps to reproduce the problem
in progress: https://github.com/craftercms/craftercms/issues/2228
### Log/stack trace (use https://gist.github.com)
https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da
### Specs
#### Version
Studio Version Number: 3.0.14-SNAPSHOT-4838ba
Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e
Build Date/Time: 06-14-2018 14:17:47 -0400
#### OS
Any
#### Browser
Any
|
1.0
|
[studio] catch exception if dependency on ContentServiceImpl.contentExists in dependency resolver - ### Expected behavior
Catch and report exceptions in this method:
`org.craftercms.studio.impl.v1.service.dependency.RegexDependencyResolver.resolve(RegexDependencyResolver.java:79)
`
That are a result of this method call:
`org.craftercms.studio.impl.v1.service.content.ContentServiceImpl.contentExists(ContentServiceImpl.java:159)
`
An ERROR log message should say that:
Dependencies in File {current file being scanned} could not be resolved for path {path that failed exist} due to the following error: {error}
### Actual behavior
Exceptions with contentExists exceptions are not caught during dependency resolution operation and the operation (site import, write operation) fails as a result.
### Steps to reproduce the problem
in progress: https://github.com/craftercms/craftercms/issues/2228
### Log/stack trace (use https://gist.github.com)
https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da
### Specs
#### Version
Studio Version Number: 3.0.14-SNAPSHOT-4838ba
Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e
Build Date/Time: 06-14-2018 14:17:47 -0400
#### OS
Any
#### Browser
Any
|
non_defect
|
catch exception if dependency on contentserviceimpl contentexists in dependency resolver expected behavior catch and report exceptions in this method org craftercms studio impl service dependency regexdependencyresolver resolve regexdependencyresolver java that are a result of this method call org craftercms studio impl service content contentserviceimpl contentexists contentserviceimpl java an error log message should say that dependencies in file current file being scanned could not be resolved for path path that failed exist due to the following error error actual behavior exceptions with contentexists exceptions are not caught during dependency resolution operation and the operation site import write operation fails as a result steps to reproduce the problem in progress log stack trace use specs version studio version number snapshot build number build date time os any browser any
| 0
|
10,202
| 26,503,468,252
|
IssuesEvent
|
2023-01-18 12:07:42
|
OasisLMF/OasisPlatform
|
https://api.github.com/repos/OasisLMF/OasisPlatform
|
closed
|
Enable parquet storage
|
feature scalable architecture
|
<!--- IMPORTANT: Please apply the relevant labels, for example if this issue is needed as a backported fix add the label `LTS fix` (Long term support fix) -->
## Issue Description
Update `PORTFOLIO_PARQUET_STORAGE=True`
- [ ] With Parquet exposure storage on, `PORTFOLIO_PARQUET_STORAGE=True`, converted csv files are stored twice in the inputs directory
```
oasis-worker(2.1.0) 9df6678e3e29 /tmp/run/analysis-65_files-2c287d7ac2de420cab322ef92630ed25 # ls -la
-rw-r--r-- 1 root root 9902 Jun 9 09:16 account.csv.parquet
-rw-r--r-- 1 root root 9902 Jun 9 09:16 account.parquet
-rw-r--r-- 1 root root 862 Jun 9 09:16 analysis_settings.json
-rw-r--r-- 1 root root 44 Jun 9 09:41 keys-errors.csv
-rw-r--r-- 1 root root 534706 Jun 9 09:41 keys.csv
-rw-r--r-- 1 root root 364056 Jun 9 09:16 location.csv.parquet
-rw-r--r-- 1 root root 364056 Jun 9 09:16 location.parquet
-rw-r--r-- 1 root root 1488 May 27 08:58 lookup_config.json
-rw-r--r-- 1 root root 13925 Jun 9 09:16 reinsinfo.parquet
-rw-r--r-- 1 root root 9920 Jun 9 09:16 reinsscope.parquet
-rw-r--r-- 1 root root 13925 Jun 9 09:16 ri_info.csv.parquet
-rw-r--r-- 1 root root 9920 Jun 9 09:16 ri_scope.csv.parquet
```
- [ ] Needs OasisLMF fix https://github.com/OasisLMF/OasisLMF/issues/1054
|
1.0
|
Enable parquet storage - <!--- IMPORTANT: Please apply the relevant labels, for example if this issue is needed as a backported fix add the label `LTS fix` (Long term support fix) -->
## Issue Description
Update `PORTFOLIO_PARQUET_STORAGE=True`
- [ ] With Parquet exposure storage on, `PORTFOLIO_PARQUET_STORAGE=True`, converted csv files are stored twice in the inputs directory
```
oasis-worker(2.1.0) 9df6678e3e29 /tmp/run/analysis-65_files-2c287d7ac2de420cab322ef92630ed25 # ls -la
-rw-r--r-- 1 root root 9902 Jun 9 09:16 account.csv.parquet
-rw-r--r-- 1 root root 9902 Jun 9 09:16 account.parquet
-rw-r--r-- 1 root root 862 Jun 9 09:16 analysis_settings.json
-rw-r--r-- 1 root root 44 Jun 9 09:41 keys-errors.csv
-rw-r--r-- 1 root root 534706 Jun 9 09:41 keys.csv
-rw-r--r-- 1 root root 364056 Jun 9 09:16 location.csv.parquet
-rw-r--r-- 1 root root 364056 Jun 9 09:16 location.parquet
-rw-r--r-- 1 root root 1488 May 27 08:58 lookup_config.json
-rw-r--r-- 1 root root 13925 Jun 9 09:16 reinsinfo.parquet
-rw-r--r-- 1 root root 9920 Jun 9 09:16 reinsscope.parquet
-rw-r--r-- 1 root root 13925 Jun 9 09:16 ri_info.csv.parquet
-rw-r--r-- 1 root root 9920 Jun 9 09:16 ri_scope.csv.parquet
```
- [ ] Needs OasisLMF fix https://github.com/OasisLMF/OasisLMF/issues/1054
|
non_defect
|
enable parquet storage issue description update portfolio parquet storage true with parquet exposure storage on portfolio parquet storage true converted csv files are stored twice in the inputs directory oasis worker tmp run analysis files ls la rw r r root root jun account csv parquet rw r r root root jun account parquet rw r r root root jun analysis settings json rw r r root root jun keys errors csv rw r r root root jun keys csv rw r r root root jun location csv parquet rw r r root root jun location parquet rw r r root root may lookup config json rw r r root root jun reinsinfo parquet rw r r root root jun reinsscope parquet rw r r root root jun ri info csv parquet rw r r root root jun ri scope csv parquet needs oasislmf fix
| 0
|
61,947
| 15,109,694,367
|
IssuesEvent
|
2021-02-08 18:11:30
|
project-chip/connectedhomeip
|
https://api.github.com/repos/project-chip/connectedhomeip
|
closed
|
Add building pigweed-app to the CI
|
build issue testing tools
|
#### Problem
There is no build validation in CI for pigweed-app.
#### Proposed Solution
Add building pigweed-app to the github workflows
|
1.0
|
Add building pigweed-app to the CI - #### Problem
There is no build validation in CI for pigweed-app.
#### Proposed Solution
Add building pigweed-app to the github workflows
|
non_defect
|
add building pigweed app to the ci problem there is no build validation in ci for pigweed app proposed solution add building pigweed app to the github workflows
| 0
|
771,484
| 27,086,629,104
|
IssuesEvent
|
2023-02-14 17:27:22
|
rzshrote/pybrops
|
https://api.github.com/repos/rzshrote/pybrops
|
opened
|
Consider better ways to create variance matrices
|
enhancement medium priority
|
The current method is to create them via a classmethod, but perhaps a factory object would make more sense, especially for selection routines that use them. The classmethod strategy to create objects is also not mutually exclusive with the factory strategy. Both can be used.
|
1.0
|
Consider better ways to create variance matrices - The current method is to create them via a classmethod, but perhaps a factory object would make more sense, especially for selection routines that use them. The classmethod strategy to create objects is also not mutually exclusive with the factory strategy. Both can be used.
|
non_defect
|
consider better ways to create variance matrices the current method is to create them via a classmethod but perhaps a factory object would make more sense especially for selection routines that use them the classmethod strategy to create objects is also not mutually exclusive with the factory strategy both can be used
| 0
|
10,027
| 4,700,073,713
|
IssuesEvent
|
2016-10-12 17:28:26
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
[0] NuGet Package Task & Targets
|
Area:PJ2MsBuild CLI 1.1 Type:Feature
|
As part of the move from Project.Json/xproj to csproj, we need to enable msbuild to be able to generate .nupkgs as part of its build output, in an integrated way.
|
1.0
|
[0] NuGet Package Task & Targets - As part of the move from Project.Json/xproj to csproj, we need to enable msbuild to be able to generate .nupkgs as part of its build output, in an integrated way.
|
non_defect
|
nuget package task targets as part of the move from project json xproj to csproj we need to enable msbuild to be able to generate nupkgs as part of its build output in an integrated way
| 0
|
1,867
| 2,603,972,679
|
IssuesEvent
|
2015-02-24 19:00:39
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳疣病医院
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳疣病医院〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:01
|
1.0
|
沈阳疣病医院 - ```
沈阳疣病医院〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:01
|
defect
|
沈阳疣病医院 沈阳疣病医院〓沈陽軍區政治部醫院性病〓tel: 〓�� � , 。位于沈陽� �� 。是一所與新中國同建立共輝煌的歷史悠� ��、設備精良、技術權威、專家云集,是預防、保健、醫療、 科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫�� �、全國首批醫療規范定點單位,是第四軍醫大學、東南大學� ��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部 衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。 original issue reported on code google com by gmail com on jun at
| 1
|
66,489
| 7,001,813,500
|
IssuesEvent
|
2017-12-18 11:41:16
|
Kademi/kademi-dev
|
https://api.github.com/repos/Kademi/kademi-dev
|
closed
|
Proposal pages for admin
|
enhancement Ready to Test - Dev
|
@bradmac We have a link from Lead page in admin console which is linked creating proposal page. So I think we also need page for proposal in admin too.
|
1.0
|
Proposal pages for admin - @bradmac We have a link from Lead page in admin console which is linked creating proposal page. So I think we also need page for proposal in admin too.
|
non_defect
|
proposal pages for admin bradmac we have a link from lead page in admin console which is linked creating proposal page so i think we also need page for proposal in admin too
| 0
|
41,433
| 10,458,319,581
|
IssuesEvent
|
2019-09-20 08:17:33
|
scemino/xnamugen
|
https://api.github.com/repos/scemino/xnamugen
|
closed
|
Fork for Mono?
|
Priority-Medium Type-Defect auto-migrated
|
```
Perhaps since Mono has its own variant of XNA, MonoXNA, perhaps we can move
to fork it to run with OpenGL and OpenAL?
I have a Macbook, and have compiled MonoXNA to run on my Machine, Imagine
the concept of an Open-Source, Cross-Platform Mugen.
Cheers,
ChaosCharza.
```
Original issue reported on code.google.com by `ChaosCha...@gmail.com` on 25 Mar 2010 at 11:25
|
1.0
|
Fork for Mono? - ```
Perhaps since Mono has its own variant of XNA, MonoXNA, perhaps we can move
to fork it to run with OpenGL and OpenAL?
I have a Macbook, and have compiled MonoXNA to run on my Machine, Imagine
the concept of an Open-Source, Cross-Platform Mugen.
Cheers,
ChaosCharza.
```
Original issue reported on code.google.com by `ChaosCha...@gmail.com` on 25 Mar 2010 at 11:25
|
defect
|
fork for mono perhaps since mono has its own variant of xna monoxna perhaps we can move to fork it to run with opengl and openal i have a macbook and have compiled monoxna to run on my machine imagine the concept of an open source cross platform mugen cheers chaoscharza original issue reported on code google com by chaoscha gmail com on mar at
| 1
|
23,619
| 3,851,864,605
|
IssuesEvent
|
2016-04-06 05:27:26
|
GPF/imame4all
|
https://api.github.com/repos/GPF/imame4all
|
closed
|
netplay not sync
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. i add some code of netplay on android, make it work.
2. test two player in wifi lan, connected,but the display not sync exactly
What version of the product are you using? On what operating system?
mod of MAME4droid 1.5 , android 4.1
Please provide any additional information below.
key event synced, but something else not.
please help me, What factors will affect the synchronization, thx!
```
Original issue reported on code.google.com by `qiaochen...@gmail.com` on 16 Nov 2013 at 8:17
|
1.0
|
netplay not sync - ```
What steps will reproduce the problem?
1. i add some code of netplay on android, make it work.
2. test two player in wifi lan, connected,but the display not sync exactly
What version of the product are you using? On what operating system?
mod of MAME4droid 1.5 , android 4.1
Please provide any additional information below.
key event synced, but something else not.
please help me, What factors will affect the synchronization, thx!
```
Original issue reported on code.google.com by `qiaochen...@gmail.com` on 16 Nov 2013 at 8:17
|
defect
|
netplay not sync what steps will reproduce the problem i add some code of netplay on android make it work test two player in wifi lan connected but the display not sync exactly what version of the product are you using on what operating system mod of android please provide any additional information below key event synced but something else not please help me what factors will affect the synchronization thx original issue reported on code google com by qiaochen gmail com on nov at
| 1
|
725,010
| 24,948,488,567
|
IssuesEvent
|
2022-11-01 03:49:26
|
AY2223S1-CS2103T-W08-1/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-W08-1/tp
|
closed
|
[PE-D][Tester C] Does not regenerate sample data
|
high priority
|
1) run command clear
2) close the application
3) no sample data generated
User might have accidentally cleared the application and now doesnt have an easy way to play around with the application

<!--session: 1666944149785-97cbf685-57d9-45ad-a313-8435d03f2880-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: jsincorporated/ped#3
|
1.0
|
[PE-D][Tester C] Does not regenerate sample data - 1) run command clear
2) close the application
3) no sample data generated
User might have accidentally cleared the application and now doesnt have an easy way to play around with the application

<!--session: 1666944149785-97cbf685-57d9-45ad-a313-8435d03f2880-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: jsincorporated/ped#3
|
non_defect
|
does not regenerate sample data run command clear close the application no sample data generated user might have accidentally cleared the application and now doesnt have an easy way to play around with the application labels severity medium type featureflaw original jsincorporated ped
| 0
|
57,296
| 6,542,477,378
|
IssuesEvent
|
2017-09-02 07:24:27
|
tcpd/surf
|
https://api.github.com/repos/tcpd/surf
|
closed
|
surf standalone app
|
Tested. Waiting for Release
|
package surf as a standalone app.
Integrate an embeddable tomcat into it.
|
1.0
|
surf standalone app - package surf as a standalone app.
Integrate an embeddable tomcat into it.
|
non_defect
|
surf standalone app package surf as a standalone app integrate an embeddable tomcat into it
| 0
|
60,916
| 7,428,737,390
|
IssuesEvent
|
2018-03-24 05:31:43
|
PalouseRobosub/robosub
|
https://api.github.com/repos/PalouseRobosub/robosub
|
closed
|
Enum bugs in localization.cpp
|
Localization (Senior Design)
|
Run tests on the localization.cpp code (and it's related subsystem code) and create bug issues for any crashes / failed initializations / unexpected outputs
|
1.0
|
Enum bugs in localization.cpp - Run tests on the localization.cpp code (and it's related subsystem code) and create bug issues for any crashes / failed initializations / unexpected outputs
|
non_defect
|
enum bugs in localization cpp run tests on the localization cpp code and it s related subsystem code and create bug issues for any crashes failed initializations unexpected outputs
| 0
|
7,578
| 2,610,406,171
|
IssuesEvent
|
2015-02-26 20:12:00
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
opened
|
Geonosis Build Pads
|
auto-migrated Priority-Medium Type-Defect
|
```
In the Geonosis land skirmish map, Team 2 have two tactical build pads.
```
-----
Original issue reported on code.google.com by `KillerHurdz@netscape.net` on 22 Aug 2011 at 4:51
|
1.0
|
Geonosis Build Pads - ```
In the Geonosis land skirmish map, Team 2 have two tactical build pads.
```
-----
Original issue reported on code.google.com by `KillerHurdz@netscape.net` on 22 Aug 2011 at 4:51
|
defect
|
geonosis build pads in the geonosis land skirmish map team have two tactical build pads original issue reported on code google com by killerhurdz netscape net on aug at
| 1
|
21,262
| 3,478,557,792
|
IssuesEvent
|
2015-12-28 13:36:40
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
reopened
|
InvocationFuture async calls don't detect lost operations
|
Team: Core Type: Defect
|
A sync call (future.get) will call is still running and if returns false, then timeout.
But future using an andThen won't trigger this behavior. So if a future is only used this way and the operation/response got lost, it won't be detected.
|
1.0
|
InvocationFuture async calls don't detect lost operations - A sync call (future.get) will call is still running and if returns false, then timeout.
But future using an andThen won't trigger this behavior. So if a future is only used this way and the operation/response got lost, it won't be detected.
|
defect
|
invocationfuture async calls don t detect lost operations a sync call future get will call is still running and if returns false then timeout but future using an andthen won t trigger this behavior so if a future is only used this way and the operation response got lost it won t be detected
| 1
|
179,911
| 6,635,686,495
|
IssuesEvent
|
2017-09-26 00:03:09
|
coreos/bugs
|
https://api.github.com/repos/coreos/bugs
|
closed
|
CT should use new passwd.users semantics
|
area/usability component/ct kind/cleanup priority/P1 team/tools
|
# Issue Report #
## Feature Request ##
### Desired Feature ###
Ignition recently deprecated the "create" object from `passwd.users`. CT should use these semantics as well.
|
1.0
|
CT should use new passwd.users semantics - # Issue Report #
## Feature Request ##
### Desired Feature ###
Ignition recently deprecated the "create" object from `passwd.users`. CT should use these semantics as well.
|
non_defect
|
ct should use new passwd users semantics issue report feature request desired feature ignition recently deprecated the create object from passwd users ct should use these semantics as well
| 0
|
311,953
| 9,540,891,083
|
IssuesEvent
|
2019-04-30 20:45:09
|
heptiolabs/wardroom
|
https://api.github.com/repos/heptiolabs/wardroom
|
closed
|
kubeadm token creation fails with permission error
|
bug p0 - Higher Priority
|
/kind bug
**What steps did you take and what happened:**
Ran the following command:
```
python provision.py -a install -o xenial examples/calico.yml
```
The kubeadm token creation play fails with the following error:
```
TASK [generate a kubeadm token] ************************************************************************************************************************************************
fatal: [master1 -> master1]: FAILED! => {"changed": true, "cmd": ["/usr/bin/kubeadm", "token", "create", "--config", "/etc/kubernetes/kubeadm.conf", "--kubeconfig", "/etc/kubernetes/admin.conf"], "delta": "0:00:00.680351", "end": "2019-04-30 12:28:29.203961", "msg": "non-zero return code", "rc": 1, "start": "2019-04-30 12:28:28.523610",
"stderr": "failed to load admin kubeconfig: open /etc/kubernetes/admin.conf: permission denied", "stderr_lines": ["failed to load admin kubeconfig: open /etc/kubernetes/admin.conf: permission denied"], "stdout": "", "stdout_lines": []}
```
We need to `become: yes` in the play.
|
1.0
|
kubeadm token creation fails with permission error - /kind bug
**What steps did you take and what happened:**
Ran the following command:
```
python provision.py -a install -o xenial examples/calico.yml
```
The kubeadm token creation play fails with the following error:
```
TASK [generate a kubeadm token] ************************************************************************************************************************************************
fatal: [master1 -> master1]: FAILED! => {"changed": true, "cmd": ["/usr/bin/kubeadm", "token", "create", "--config", "/etc/kubernetes/kubeadm.conf", "--kubeconfig", "/etc/kubernetes/admin.conf"], "delta": "0:00:00.680351", "end": "2019-04-30 12:28:29.203961", "msg": "non-zero return code", "rc": 1, "start": "2019-04-30 12:28:28.523610",
"stderr": "failed to load admin kubeconfig: open /etc/kubernetes/admin.conf: permission denied", "stderr_lines": ["failed to load admin kubeconfig: open /etc/kubernetes/admin.conf: permission denied"], "stdout": "", "stdout_lines": []}
```
We need to `become: yes` in the play.
|
non_defect
|
kubeadm token creation fails with permission error kind bug what steps did you take and what happened ran the following command python provision py a install o xenial examples calico yml the kubeadm token creation play fails with the following error task fatal failed changed true cmd delta end msg non zero return code rc start stderr failed to load admin kubeconfig open etc kubernetes admin conf permission denied stderr lines stdout stdout lines we need to become yes in the play
| 0
|
321,673
| 23,866,694,630
|
IssuesEvent
|
2022-09-07 11:40:34
|
dipzza/ultrastar-song2txt
|
https://api.github.com/repos/dipzza/ultrastar-song2txt
|
closed
|
Redactar el resumen de la memoria.
|
documentation
|
Falta el resumen inicial del trabajo realizado, relacionado con #9.
Debería ser corto, en inglés y español, y presentar las ideas principales que se tratan en la memoria.
|
1.0
|
Redactar el resumen de la memoria. - Falta el resumen inicial del trabajo realizado, relacionado con #9.
Debería ser corto, en inglés y español, y presentar las ideas principales que se tratan en la memoria.
|
non_defect
|
redactar el resumen de la memoria falta el resumen inicial del trabajo realizado relacionado con debería ser corto en inglés y español y presentar las ideas principales que se tratan en la memoria
| 0
|
209,487
| 23,726,982,734
|
IssuesEvent
|
2022-08-30 20:37:17
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Failures in SSL Stream tests: There are no more endpoints available in the endpoint mapper.
|
area-System.Net.Security
|
This happened in the `release/7.0` branch. Can you please confirm if this will require a backported fix?
PR: https://github.com/dotnet/runtime/pull/74808/
Queue: Libraries Test Run release coreclr windows x86 Release
Job: https://dev.azure.com/dnceng/public/_build/results?buildId=1976097&view=logs&j=457f7e88-dfa2-5bd9-f871-fdf124c2477d&t=bfe52dfb-2099-5c7f-ee52-70a1d81c544e
Log: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74808-merge-0a29e7160d114b13be/System.Net.Security.Tests/3/console.38a99609.log?helixlogtype=result
Failure message:
> CryptographicException : No hay más extremos disponibles desde el asignador de extremos.
Which translates to:
> CryptographicException : There are no more endpoints available in the endpoint mapper.
Tests with a similar failure:
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_SecondNegotiateClientCertificateAsync_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_UntrustedCaWithCustomTrust_OK](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_IncompleteIncomingTlsFrame_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_ClientWriteData](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncNoRenego_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NestedAuth_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncConcurrentIO_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_PendingDecryptedData_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncTls13_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_TargetHostName_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_RandomSizeWrites_OK](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.CertificateValidationRemoteServer.ConnectWithRevocation_WithCallback](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamMutualAuthenticationTest.SslStream_RequireClientCert_IsMutuallyAuthenticated_ReturnsTrue](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamCredentialCacheTest.SslStream_SameCertUsedForClientAndServer_Ok](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
Callstack example:
```
System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_SecondNegotiateClientCertificateAsync_Throws(sendClientCertificate: True) [FAIL]
System.AggregateException : One or more errors occurred. (No hay m�s extremos disponibles desde el asignador de extremos.) (The following constructor parameters did not have matching fixture data: CertificateSetup setup)
---- System.Security.Cryptography.CryptographicException : No hay m�s extremos disponibles desde el asignador de extremos.
---- The following constructor parameters did not have matching fixture data: CertificateSetup setup
Stack Trace:
----- Inner Stack Trace #1 (System.Security.Cryptography.CryptographicException) -----
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngHelpers.cs(33,0): at System.Security.Cryptography.CngHelpers.OpenStorageProvider(CngProvider provider)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngKey.Create.cs(41,0): at System.Security.Cryptography.CngKey.Create(CngAlgorithm algorithm, String keyName, CngKeyCreationParameters creationParameters)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngAlgorithmCore.cs(73,0): at System.Security.Cryptography.CngAlgorithmCore.GetOrGenerateKey(Int32 keySize, CngAlgorithm algorithm)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSACng.Key.cs(24,0): at System.Security.Cryptography.RSACng.get_Key()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSACng.ImportExport.cs(50,0): at System.Security.Cryptography.RSACng.ExportKeyBlob(Boolean includePrivateParameters)
/_/src/libraries/Common/src/System/Security/Cryptography/RSACng.ImportExport.cs(282,0): at System.Security.Cryptography.RSACng.ExportParameters(Boolean includePrivateParameters)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSA.cs(811,0): at System.Security.Cryptography.RSA.WritePkcs1PublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSA.cs(675,0): at System.Security.Cryptography.RSA.ExportRSAPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSAWrapper.cs(141,0): at System.Security.Cryptography.RSAWrapper.ExportRSAPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/RSAPkcs1X509SignatureGenerator.cs(37,0): at System.Security.Cryptography.X509Certificates.RSAPkcs1X509SignatureGenerator.BuildPublicKey(RSA rsa)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/RSAPkcs1X509SignatureGenerator.cs(27,0): at System.Security.Cryptography.X509Certificates.RSAPkcs1X509SignatureGenerator.BuildPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/X509SignatureGenerator.cs(10,0): at System.Security.Cryptography.X509Certificates.X509SignatureGenerator.get_PublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/CertificateRequest.cs(141,0): at System.Security.Cryptography.X509Certificates.CertificateRequest..ctor(String subjectName, RSA key, HashAlgorithmName hashAlgorithm, RSASignaturePadding padding)
/_/src/libraries/Common/tests/System/Security/Cryptography/X509Certificates/CertificateAuthority.cs(820,0): at System.Security.Cryptography.X509Certificates.Tests.Common.CertificateAuthority.BuildPrivatePki(PkiOptions pkiOptions, RevocationResponder& responder, CertificateAuthority& rootAuthority, CertificateAuthority[]& intermediateAuthorities, X509Certificate2& endEntityCert, Int32 intermediateAuthorityCount, String testName, Boolean registerAuthorities, Boolean pkiOptionsInSubject, String subjectName, Int32 keySize, X509ExtensionCollection extensions)
/_/src/libraries/System.Net.Security/tests/FunctionalTests/TestHelper.cs(168,0): at System.Net.Security.Tests.TestHelper.GenerateCertificates(String targetName, String testName, Boolean longChain, Boolean serverCertificate)
/_/src/libraries/System.Net.Security/tests/FunctionalTests/SslStreamNetworkStreamTest.cs(30,0): at System.Net.Security.Tests.CertificateSetup..ctor()
at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
/_/src/libraries/System.Private.CoreLib/src/System/Reflection/ConstructorInvoker.cs(86,0): at System.Reflection.ConstructorInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr)
```
|
True
|
Failures in SSL Stream tests: There are no more endpoints available in the endpoint mapper. - This happened in the `release/7.0` branch. Can you please confirm if this will require a backported fix?
PR: https://github.com/dotnet/runtime/pull/74808/
Queue: Libraries Test Run release coreclr windows x86 Release
Job: https://dev.azure.com/dnceng/public/_build/results?buildId=1976097&view=logs&j=457f7e88-dfa2-5bd9-f871-fdf124c2477d&t=bfe52dfb-2099-5c7f-ee52-70a1d81c544e
Log: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74808-merge-0a29e7160d114b13be/System.Net.Security.Tests/3/console.38a99609.log?helixlogtype=result
Failure message:
> CryptographicException : No hay más extremos disponibles desde el asignador de extremos.
Which translates to:
> CryptographicException : There are no more endpoints available in the endpoint mapper.
Tests with a similar failure:
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_SecondNegotiateClientCertificateAsync_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_UntrustedCaWithCustomTrust_OK](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_IncompleteIncomingTlsFrame_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_ClientWriteData](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncNoRenego_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NestedAuth_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncConcurrentIO_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_PendingDecryptedData_Throws](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsyncTls13_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_NegotiateClientCertificateAsync_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_TargetHostName_Succeeds](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_RandomSizeWrites_OK](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.CertificateValidationRemoteServer.ConnectWithRevocation_WithCallback](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamMutualAuthenticationTest.SslStream_RequireClientCert_IsMutuallyAuthenticated_ReturnsTrue](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
[System.Net.Security.Tests.SslStreamCredentialCacheTest.SslStream_SameCertUsedForClientAndServer_Ok](https://helix.dot.net/api/2019-06-17/jobs/0a29e716-0d11-4b13-be0e-1d71d340c394/workitems/System.Net.Security.Tests/console)
Callstack example:
```
System.Net.Security.Tests.SslStreamNetworkStreamTest.SslStream_SecondNegotiateClientCertificateAsync_Throws(sendClientCertificate: True) [FAIL]
System.AggregateException : One or more errors occurred. (No hay m�s extremos disponibles desde el asignador de extremos.) (The following constructor parameters did not have matching fixture data: CertificateSetup setup)
---- System.Security.Cryptography.CryptographicException : No hay m�s extremos disponibles desde el asignador de extremos.
---- The following constructor parameters did not have matching fixture data: CertificateSetup setup
Stack Trace:
----- Inner Stack Trace #1 (System.Security.Cryptography.CryptographicException) -----
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngHelpers.cs(33,0): at System.Security.Cryptography.CngHelpers.OpenStorageProvider(CngProvider provider)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngKey.Create.cs(41,0): at System.Security.Cryptography.CngKey.Create(CngAlgorithm algorithm, String keyName, CngKeyCreationParameters creationParameters)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/CngAlgorithmCore.cs(73,0): at System.Security.Cryptography.CngAlgorithmCore.GetOrGenerateKey(Int32 keySize, CngAlgorithm algorithm)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSACng.Key.cs(24,0): at System.Security.Cryptography.RSACng.get_Key()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSACng.ImportExport.cs(50,0): at System.Security.Cryptography.RSACng.ExportKeyBlob(Boolean includePrivateParameters)
/_/src/libraries/Common/src/System/Security/Cryptography/RSACng.ImportExport.cs(282,0): at System.Security.Cryptography.RSACng.ExportParameters(Boolean includePrivateParameters)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSA.cs(811,0): at System.Security.Cryptography.RSA.WritePkcs1PublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSA.cs(675,0): at System.Security.Cryptography.RSA.ExportRSAPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/RSAWrapper.cs(141,0): at System.Security.Cryptography.RSAWrapper.ExportRSAPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/RSAPkcs1X509SignatureGenerator.cs(37,0): at System.Security.Cryptography.X509Certificates.RSAPkcs1X509SignatureGenerator.BuildPublicKey(RSA rsa)
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/RSAPkcs1X509SignatureGenerator.cs(27,0): at System.Security.Cryptography.X509Certificates.RSAPkcs1X509SignatureGenerator.BuildPublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/X509SignatureGenerator.cs(10,0): at System.Security.Cryptography.X509Certificates.X509SignatureGenerator.get_PublicKey()
/_/src/libraries/System.Security.Cryptography/src/System/Security/Cryptography/X509Certificates/CertificateRequest.cs(141,0): at System.Security.Cryptography.X509Certificates.CertificateRequest..ctor(String subjectName, RSA key, HashAlgorithmName hashAlgorithm, RSASignaturePadding padding)
/_/src/libraries/Common/tests/System/Security/Cryptography/X509Certificates/CertificateAuthority.cs(820,0): at System.Security.Cryptography.X509Certificates.Tests.Common.CertificateAuthority.BuildPrivatePki(PkiOptions pkiOptions, RevocationResponder& responder, CertificateAuthority& rootAuthority, CertificateAuthority[]& intermediateAuthorities, X509Certificate2& endEntityCert, Int32 intermediateAuthorityCount, String testName, Boolean registerAuthorities, Boolean pkiOptionsInSubject, String subjectName, Int32 keySize, X509ExtensionCollection extensions)
/_/src/libraries/System.Net.Security/tests/FunctionalTests/TestHelper.cs(168,0): at System.Net.Security.Tests.TestHelper.GenerateCertificates(String targetName, String testName, Boolean longChain, Boolean serverCertificate)
/_/src/libraries/System.Net.Security/tests/FunctionalTests/SslStreamNetworkStreamTest.cs(30,0): at System.Net.Security.Tests.CertificateSetup..ctor()
at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
/_/src/libraries/System.Private.CoreLib/src/System/Reflection/ConstructorInvoker.cs(86,0): at System.Reflection.ConstructorInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr)
```
|
non_defect
|
failures in ssl stream tests there are no more endpoints available in the endpoint mapper this happened in the release branch can you please confirm if this will require a backported fix pr queue libraries test run release coreclr windows release job log failure message cryptographicexception no hay más extremos disponibles desde el asignador de extremos which translates to cryptographicexception there are no more endpoints available in the endpoint mapper tests with a similar failure callstack example system net security tests sslstreamnetworkstreamtest sslstream secondnegotiateclientcertificateasync throws sendclientcertificate true system aggregateexception one or more errors occurred no hay m�s extremos disponibles desde el asignador de extremos the following constructor parameters did not have matching fixture data certificatesetup setup system security cryptography cryptographicexception no hay m�s extremos disponibles desde el asignador de extremos the following constructor parameters did not have matching fixture data certificatesetup setup stack trace inner stack trace system security cryptography cryptographicexception src libraries system security cryptography src system security cryptography cnghelpers cs at system security cryptography cnghelpers openstorageprovider cngprovider provider src libraries system security cryptography src system security cryptography cngkey create cs at system security cryptography cngkey create cngalgorithm algorithm string keyname cngkeycreationparameters creationparameters src libraries system security cryptography src system security cryptography cngalgorithmcore cs at system security cryptography cngalgorithmcore getorgeneratekey keysize cngalgorithm algorithm src libraries system security cryptography src system security cryptography rsacng key cs at system security cryptography rsacng get key src libraries system security cryptography src system security cryptography rsacng importexport cs at system security cryptography rsacng exportkeyblob boolean includeprivateparameters src libraries common src system security cryptography rsacng importexport cs at system security cryptography rsacng exportparameters boolean includeprivateparameters src libraries system security cryptography src system security cryptography rsa cs at system security cryptography rsa src libraries system security cryptography src system security cryptography rsa cs at system security cryptography rsa exportrsapublickey src libraries system security cryptography src system security cryptography rsawrapper cs at system security cryptography rsawrapper exportrsapublickey src libraries system security cryptography src system security cryptography cs at system security cryptography buildpublickey rsa rsa src libraries system security cryptography src system security cryptography cs at system security cryptography buildpublickey src libraries system security cryptography src system security cryptography cs at system security cryptography get publickey src libraries system security cryptography src system security cryptography certificaterequest cs at system security cryptography certificaterequest ctor string subjectname rsa key hashalgorithmname hashalgorithm rsasignaturepadding padding src libraries common tests system security cryptography certificateauthority cs at system security cryptography tests common certificateauthority buildprivatepki pkioptions pkioptions revocationresponder responder certificateauthority rootauthority certificateauthority intermediateauthorities endentitycert intermediateauthoritycount string testname boolean registerauthorities boolean pkioptionsinsubject string subjectname keysize extensions src libraries system net security tests functionaltests testhelper cs at system net security tests testhelper generatecertificates string targetname string testname boolean longchain boolean servercertificate src libraries system net security tests functionaltests sslstreamnetworkstreamtest cs at system net security tests certificatesetup ctor at system runtimemethodhandle invokemethod object target void arguments signature sig boolean isconstructor src libraries system private corelib src system reflection constructorinvoker cs at system reflection constructorinvoker invoke object obj intptr args bindingflags invokeattr
| 0
|
56,216
| 14,982,574,673
|
IssuesEvent
|
2021-01-28 16:09:02
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
vdev_id not properly detecting multipath disk configuration
|
Status: Triage Needed Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.1
Linux Kernel | 5.4.0-51-generic
Architecture | x86_64
ZFS Version | 0.8.5-1
SPL Version | 0.8.5-1
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I have a system with 2x LSI SAS3008 PCI-Express Fusion-MPT SAS-3 connecting to two IO controllers on a Supermicro 90-disk JBOD: https://www.supermicro.com/en/products/chassis/4U/946/SC946ED-R2KJBOD in a crossover dual-channel config
```
LSI0 --> IO Controller 1 channel A
|
--> IO Controller 2 channel A
LSI1 --> IO Controller 1 channel B
|
--> IO controller 2 channel B
```
This is intended to provide cross-failover protection in case of card or IO controller failure. I currently have this configured with "multibus" in multipathd, and it is properly detecting and setting up 90 aggregated paths:
Snip of multipath -ll, some addresses greeked
```
35000cca26f633fd8 dm-31 HGST,HUH721212AL4200
size=11T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
|- 0:0:46:0 sdht 134:48 active ready running
|- 15:0:46:0 sdck 69:128 active ready running
|- 0:0:139:0 sdlf 67:464 active ready running
`- 15:0:139:0 sdfw 131:32 active ready running
35000cca27a3c9d58 dm-30 HGST,HUH721212AL4200
size=11T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
|- 0:0:45:0 sdhs 134:32 active ready running
|- 15:0:45:0 sdcj 69:112 active ready running
|- 0:0:138:0 sdle 67:448 active ready running
`- 15:0:138:0 sdfv 131:16 active ready running
```
When attempting to use ``vdev_id.conf`` using a standard multipath config:
```
multipath yes
# PCI_ID HBA PORT CHANNEL NAME
channel 5e:00.0 1 A
channel 5e:00.0 0 B
channel af:00.0 1 A
channel af:00.0 0 B
```
I get the following ``/dev/disk/by-vdev`` entries:
```
# ls -l
total 0
lrwxrwxrwx 1 root root 10 Oct 21 19:35 A7 -> ../../sdek
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B0 -> ../../dm-43
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B1 -> ../../dm-1
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B10 -> ../../dm-2
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B11 -> ../../dm-3
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B12 -> ../../dm-8
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B13 -> ../../dm-9
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B14 -> ../../dm-49
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B15 -> ../../dm-11
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B16 -> ../../dm-12
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B17 -> ../../dm-52
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B18 -> ../../dm-53
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B19 -> ../../dm-15
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B2 -> ../../dm-82
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B20 -> ../../dm-56
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B21 -> ../../dm-18
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B22 -> ../../dm-58
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B23 -> ../../dm-20
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B24 -> ../../dm-60
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B25 -> ../../dm-22
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B26 -> ../../dm-62
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B27 -> ../../dm-63
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B28 -> ../../dm-64
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B29 -> ../../dm-26
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B3 -> ../../dm-66
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B30 -> ../../dm-28
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B31 -> ../../dm-29
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B32 -> ../../dm-30
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B33 -> ../../dm-70
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B34 -> ../../dm-71
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B35 -> ../../dm-72
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B36 -> ../../dm-73
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B37 -> ../../dm-74
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B38 -> ../../dm-36
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B4 -> ../../dm-37
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B5 -> ../../dm-85
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B6 -> ../../dm-86
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B7 -> ../../dm-40
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B8 -> ../../dm-80
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B9 -> ../../dm-81
```
Note that this is missing many entries, currently totaling only 40.
When testing vdev_id for bugs, it does appear to throw an error:
```
# sh -x /lib/udev/vdev_id -d /dev/dm-1
+ PATH=/bin:/sbin:/usr/bin:/usr/sbin
+ CONFIG=/etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ DEV=
+ MULTIPATH=
+ TOPOLOGY=
+ BAY=
+ getopts c:d:eg:mp:h OPTION
+ DEV=/dev/dm-1
+ getopts c:d:eg:mp:h OPTION
+ [ ! -r /etc/zfs/vdev_id.conf ]
+ [ -z /dev/dm-1 -a -z ]
+ [ -z ]
+ awk $1 == "topology" {print $2; exit} /etc/zfs/vdev_id.conf
+ TOPOLOGY=
+ [ -z ]
+ awk $1 == "slot" {print $2; exit} /etc/zfs/vdev_id.conf
+ BAY=
+ TOPOLOGY=sas_direct
+ [ = yes -a sas_direct = sas_direct ]
+ alias_handler
+ local DM_PART=
+ echo
+ grep -q -E p[0-9][0-9]*$
+ ID_VDEV=
+ [ -z ]
+ BAY=bay
+ sas_handler
+ [ -z ]
+ awk $1 == "phys_per_port" {print $2; exit} /etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ PHYS_PER_PORT=4
+ echo 4
+ grep -q -E ^[0-9]+$
+ [ -z ]
+ awk $1 == "multipath" {print $2; exit} /etc/zfs/vdev_id.conf
+ MULTIPATH_MODE=yes
+ [ yes = yes ]
+ [ -z ]
+ ls -l --full-time /dev/mapper
+ awk /\//dev/dm-1$/{print $9}
awk: cmd. line:1: /\//dev/dm-1$/{print $9}
awk: cmd. line:1: ^ unterminated regexp
+ DM_NAME=
+ [ != partition ]
+ echo
+ awk -Fp /p/{print "-part"$2}
+ PART=
+ echo
+ sed s/p[0-9][0-9]*$//
+ DM_NAME=
+ [ -z ]
+ return
+ ID_VDEV=
+ [ -n ]
```
It returns the same error for mapper entries that do not get vdevs generated:
```
# sh -x /lib/udev/vdev_id -d /dev/dm-4
+ PATH=/bin:/sbin:/usr/bin:/usr/sbin
+ CONFIG=/etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ DEV=
+ MULTIPATH=
+ TOPOLOGY=
+ BAY=
+ getopts c:d:eg:mp:h OPTION
+ DEV=/dev/dm-4
+ getopts c:d:eg:mp:h OPTION
+ [ ! -r /etc/zfs/vdev_id.conf ]
+ [ -z /dev/dm-4 -a -z ]
+ [ -z ]
+ awk $1 == "topology" {print $2; exit} /etc/zfs/vdev_id.conf
+ TOPOLOGY=
+ [ -z ]
+ awk $1 == "slot" {print $2; exit} /etc/zfs/vdev_id.conf
+ BAY=
+ TOPOLOGY=sas_direct
+ [ = yes -a sas_direct = sas_direct ]
+ alias_handler
+ local DM_PART=
+ echo
+ grep -q -E p[0-9][0-9]*$
+ ID_VDEV=
+ [ -z ]
+ BAY=bay
+ sas_handler
+ [ -z ]
+ awk $1 == "phys_per_port" {print $2; exit} /etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ PHYS_PER_PORT=4
+ echo 4
+ grep -q -E ^[0-9]+$
+ [ -z ]
+ awk $1 == "multipath" {print $2; exit} /etc/zfs/vdev_id.conf
+ MULTIPATH_MODE=yes
+ [ yes = yes ]
+ [ -z ]
+ ls -l --full-time /dev/mapper
+ awk /\//dev/dm-4$/{print $9}
awk: cmd. line:1: /\//dev/dm-4$/{print $9}
awk: cmd. line:1: ^ unterminated regexp
+ DM_NAME=
+ [ != partition ]
+ echo
+ awk -Fp /p/{print "-part"$2}
+ PART=
+ echo
+ sed s/p[0-9][0-9]*$//
+ DM_NAME=
+ [ -z ]
+ return
+ ID_VDEV=
+ [ -n ]
```
I'm unclear from the documentation what a number of the terms are for the vdev_id.conf file in the context of real-world hardware configurations (how do I determine number of phys, how do I know if switch or direct, is a HBA port the actual physical wire, what does "CHANNEL NAME" refer to?), so I'm having much difficulty guessing as to what settings might need to be changed.
Finally, I offer the output of sas_discover from https://github.com/stanford-rc/sasutils which appears to have mapped the entire SAS topology, it does look a bit odd compared to what I expect to see, this may be be due to the internal wiring of the JBOD
```
hostname
|--host15
| |--4x--expander-15:0
| | |--1x--end_device-15:0:12
| | | `--enclosure
| | |--1x--end_device-15:0:0
| | | `-- 1 x disk
| | |--1x--end_device-15:0:1
| | | `-- 1 x disk
| | |--1x--end_device-15:0:2
| | | `-- 1 x disk
| | |--1x--end_device-15:0:3
| | | `-- 1 x disk
| | |--1x--end_device-15:0:4
| | | `-- 1 x disk
| | |--1x--end_device-15:0:5
| | | `-- 1 x disk
| | |--1x--end_device-15:0:6
| | | `-- 1 x disk
| | |--1x--end_device-15:0:7
| | | `-- 1 x disk
| | |--1x--end_device-15:0:8
| | | `-- 1 x disk
| | |--1x--end_device-15:0:9
| | | `-- 1 x disk
| | |--1x--end_device-15:0:10
| | | `-- 1 x disk
| | |--1x--end_device-15:0:11
| | | `-- 1 x disk
| | |--8x--expander-15:2
| | | |-- 39 x end_device -- disk
| | | `-- 1 x end_device -- enclosure
| | `--8x--expander-15:1
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--4x--expander-15:3
| |--1x--end_device-15:3:12
| | `--enclosure
| |--1x--end_device-15:3:0
| | `-- 1 x disk
| |--1x--end_device-15:3:1
| | `-- 1 x disk
| |--1x--end_device-15:3:2
| | `-- 1 x disk
| |--1x--end_device-15:3:3
| | `-- 1 x disk
| |--1x--end_device-15:3:4
| | `-- 1 x disk
| |--1x--end_device-15:3:5
| | `-- 1 x disk
| |--1x--end_device-15:3:6
| | `-- 1 x disk
| |--1x--end_device-15:3:7
| | `-- 1 x disk
| |--1x--end_device-15:3:8
| | `-- 1 x disk
| |--1x--end_device-15:3:9
| | `-- 1 x disk
| |--1x--end_device-15:3:10
| | `-- 1 x disk
| |--1x--end_device-15:3:11
| | `-- 1 x disk
| |--8x--expander-15:5
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--8x--expander-15:4
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--host0
|--4x--expander-0:3
| |--1x--end_device-0:3:12
| | `--enclosure
| |--1x--end_device-0:3:0
| | `-- 1 x disk
| |--1x--end_device-0:3:1
| | `-- 1 x disk
| |--1x--end_device-0:3:2
| | `-- 1 x disk
| |--1x--end_device-0:3:3
| | `-- 1 x disk
| |--1x--end_device-0:3:4
| | `-- 1 x disk
| |--1x--end_device-0:3:5
| | `-- 1 x disk
| |--1x--end_device-0:3:6
| | `-- 1 x disk
| |--1x--end_device-0:3:7
| | `-- 1 x disk
| |--1x--end_device-0:3:8
| | `-- 1 x disk
| |--1x--end_device-0:3:9
| | `-- 1 x disk
| |--1x--end_device-0:3:10
| | `-- 1 x disk
| |--1x--end_device-0:3:11
| | `-- 1 x disk
| |--8x--expander-0:5
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--8x--expander-0:4
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--4x--expander-0:0
|--1x--end_device-0:0:12
| `--enclosure
|--1x--end_device-0:0:0
| `-- 1 x disk
|--1x--end_device-0:0:1
| `-- 1 x disk
|--1x--end_device-0:0:2
| `-- 1 x disk
|--1x--end_device-0:0:3
| `-- 1 x disk
|--1x--end_device-0:0:4
| `-- 1 x disk
|--1x--end_device-0:0:5
| `-- 1 x disk
|--1x--end_device-0:0:6
| `-- 1 x disk
|--1x--end_device-0:0:7
| `-- 1 x disk
|--1x--end_device-0:0:8
| `-- 1 x disk
|--1x--end_device-0:0:9
| `-- 1 x disk
|--1x--end_device-0:0:10
| `-- 1 x disk
|--1x--end_device-0:0:11
| `-- 1 x disk
|--8x--expander-0:2
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--8x--expander-0:1
|-- 39 x end_device -- disk
`-- 1 x end_device -- enclosure
```
It looks like internally the IO controllers each have 1 expander going to 12 disks, and cascading 2 more expanders with 39 disks each, a very odd config, but each of the 90 disks is exposed properly on each physical wiring path, as expected.
|
1.0
|
vdev_id not properly detecting multipath disk configuration - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.1
Linux Kernel | 5.4.0-51-generic
Architecture | x86_64
ZFS Version | 0.8.5-1
SPL Version | 0.8.5-1
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I have a system with 2x LSI SAS3008 PCI-Express Fusion-MPT SAS-3 connecting to two IO controllers on a Supermicro 90-disk JBOD: https://www.supermicro.com/en/products/chassis/4U/946/SC946ED-R2KJBOD in a crossover dual-channel config
```
LSI0 --> IO Controller 1 channel A
|
--> IO Controller 2 channel A
LSI1 --> IO Controller 1 channel B
|
--> IO controller 2 channel B
```
This is intended to provide cross-failover protection in case of card or IO controller failure. I currently have this configured with "multibus" in multipathd, and it is properly detecting and setting up 90 aggregated paths:
Snip of multipath -ll, some addresses greeked
```
35000cca26f633fd8 dm-31 HGST,HUH721212AL4200
size=11T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
|- 0:0:46:0 sdht 134:48 active ready running
|- 15:0:46:0 sdck 69:128 active ready running
|- 0:0:139:0 sdlf 67:464 active ready running
`- 15:0:139:0 sdfw 131:32 active ready running
35000cca27a3c9d58 dm-30 HGST,HUH721212AL4200
size=11T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
|- 0:0:45:0 sdhs 134:32 active ready running
|- 15:0:45:0 sdcj 69:112 active ready running
|- 0:0:138:0 sdle 67:448 active ready running
`- 15:0:138:0 sdfv 131:16 active ready running
```
When attempting to use ``vdev_id.conf`` using a standard multipath config:
```
multipath yes
# PCI_ID HBA PORT CHANNEL NAME
channel 5e:00.0 1 A
channel 5e:00.0 0 B
channel af:00.0 1 A
channel af:00.0 0 B
```
I get the following ``/dev/disk/by-vdev`` entries:
```
# ls -l
total 0
lrwxrwxrwx 1 root root 10 Oct 21 19:35 A7 -> ../../sdek
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B0 -> ../../dm-43
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B1 -> ../../dm-1
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B10 -> ../../dm-2
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B11 -> ../../dm-3
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B12 -> ../../dm-8
lrwxrwxrwx 1 root root 10 Oct 21 19:38 B13 -> ../../dm-9
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B14 -> ../../dm-49
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B15 -> ../../dm-11
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B16 -> ../../dm-12
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B17 -> ../../dm-52
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B18 -> ../../dm-53
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B19 -> ../../dm-15
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B2 -> ../../dm-82
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B20 -> ../../dm-56
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B21 -> ../../dm-18
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B22 -> ../../dm-58
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B23 -> ../../dm-20
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B24 -> ../../dm-60
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B25 -> ../../dm-22
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B26 -> ../../dm-62
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B27 -> ../../dm-63
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B28 -> ../../dm-64
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B29 -> ../../dm-26
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B3 -> ../../dm-66
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B30 -> ../../dm-28
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B31 -> ../../dm-29
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B32 -> ../../dm-30
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B33 -> ../../dm-70
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B34 -> ../../dm-71
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B35 -> ../../dm-72
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B36 -> ../../dm-73
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B37 -> ../../dm-74
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B38 -> ../../dm-36
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B4 -> ../../dm-37
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B5 -> ../../dm-85
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B6 -> ../../dm-86
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B7 -> ../../dm-40
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B8 -> ../../dm-80
lrwxrwxrwx 1 root root 11 Oct 21 19:38 B9 -> ../../dm-81
```
Note that this is missing many entries, currently totaling only 40.
When testing vdev_id for bugs, it does appear to throw an error:
```
# sh -x /lib/udev/vdev_id -d /dev/dm-1
+ PATH=/bin:/sbin:/usr/bin:/usr/sbin
+ CONFIG=/etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ DEV=
+ MULTIPATH=
+ TOPOLOGY=
+ BAY=
+ getopts c:d:eg:mp:h OPTION
+ DEV=/dev/dm-1
+ getopts c:d:eg:mp:h OPTION
+ [ ! -r /etc/zfs/vdev_id.conf ]
+ [ -z /dev/dm-1 -a -z ]
+ [ -z ]
+ awk $1 == "topology" {print $2; exit} /etc/zfs/vdev_id.conf
+ TOPOLOGY=
+ [ -z ]
+ awk $1 == "slot" {print $2; exit} /etc/zfs/vdev_id.conf
+ BAY=
+ TOPOLOGY=sas_direct
+ [ = yes -a sas_direct = sas_direct ]
+ alias_handler
+ local DM_PART=
+ echo
+ grep -q -E p[0-9][0-9]*$
+ ID_VDEV=
+ [ -z ]
+ BAY=bay
+ sas_handler
+ [ -z ]
+ awk $1 == "phys_per_port" {print $2; exit} /etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ PHYS_PER_PORT=4
+ echo 4
+ grep -q -E ^[0-9]+$
+ [ -z ]
+ awk $1 == "multipath" {print $2; exit} /etc/zfs/vdev_id.conf
+ MULTIPATH_MODE=yes
+ [ yes = yes ]
+ [ -z ]
+ ls -l --full-time /dev/mapper
+ awk /\//dev/dm-1$/{print $9}
awk: cmd. line:1: /\//dev/dm-1$/{print $9}
awk: cmd. line:1: ^ unterminated regexp
+ DM_NAME=
+ [ != partition ]
+ echo
+ awk -Fp /p/{print "-part"$2}
+ PART=
+ echo
+ sed s/p[0-9][0-9]*$//
+ DM_NAME=
+ [ -z ]
+ return
+ ID_VDEV=
+ [ -n ]
```
It returns the same error for mapper entries that do not get vdevs generated:
```
# sh -x /lib/udev/vdev_id -d /dev/dm-4
+ PATH=/bin:/sbin:/usr/bin:/usr/sbin
+ CONFIG=/etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ DEV=
+ MULTIPATH=
+ TOPOLOGY=
+ BAY=
+ getopts c:d:eg:mp:h OPTION
+ DEV=/dev/dm-4
+ getopts c:d:eg:mp:h OPTION
+ [ ! -r /etc/zfs/vdev_id.conf ]
+ [ -z /dev/dm-4 -a -z ]
+ [ -z ]
+ awk $1 == "topology" {print $2; exit} /etc/zfs/vdev_id.conf
+ TOPOLOGY=
+ [ -z ]
+ awk $1 == "slot" {print $2; exit} /etc/zfs/vdev_id.conf
+ BAY=
+ TOPOLOGY=sas_direct
+ [ = yes -a sas_direct = sas_direct ]
+ alias_handler
+ local DM_PART=
+ echo
+ grep -q -E p[0-9][0-9]*$
+ ID_VDEV=
+ [ -z ]
+ BAY=bay
+ sas_handler
+ [ -z ]
+ awk $1 == "phys_per_port" {print $2; exit} /etc/zfs/vdev_id.conf
+ PHYS_PER_PORT=
+ PHYS_PER_PORT=4
+ echo 4
+ grep -q -E ^[0-9]+$
+ [ -z ]
+ awk $1 == "multipath" {print $2; exit} /etc/zfs/vdev_id.conf
+ MULTIPATH_MODE=yes
+ [ yes = yes ]
+ [ -z ]
+ ls -l --full-time /dev/mapper
+ awk /\//dev/dm-4$/{print $9}
awk: cmd. line:1: /\//dev/dm-4$/{print $9}
awk: cmd. line:1: ^ unterminated regexp
+ DM_NAME=
+ [ != partition ]
+ echo
+ awk -Fp /p/{print "-part"$2}
+ PART=
+ echo
+ sed s/p[0-9][0-9]*$//
+ DM_NAME=
+ [ -z ]
+ return
+ ID_VDEV=
+ [ -n ]
```
I'm unclear from the documentation what a number of the terms are for the vdev_id.conf file in the context of real-world hardware configurations (how do I determine number of phys, how do I know if switch or direct, is a HBA port the actual physical wire, what does "CHANNEL NAME" refer to?), so I'm having much difficulty guessing as to what settings might need to be changed.
Finally, I offer the output of sas_discover from https://github.com/stanford-rc/sasutils which appears to have mapped the entire SAS topology, it does look a bit odd compared to what I expect to see, this may be be due to the internal wiring of the JBOD
```
hostname
|--host15
| |--4x--expander-15:0
| | |--1x--end_device-15:0:12
| | | `--enclosure
| | |--1x--end_device-15:0:0
| | | `-- 1 x disk
| | |--1x--end_device-15:0:1
| | | `-- 1 x disk
| | |--1x--end_device-15:0:2
| | | `-- 1 x disk
| | |--1x--end_device-15:0:3
| | | `-- 1 x disk
| | |--1x--end_device-15:0:4
| | | `-- 1 x disk
| | |--1x--end_device-15:0:5
| | | `-- 1 x disk
| | |--1x--end_device-15:0:6
| | | `-- 1 x disk
| | |--1x--end_device-15:0:7
| | | `-- 1 x disk
| | |--1x--end_device-15:0:8
| | | `-- 1 x disk
| | |--1x--end_device-15:0:9
| | | `-- 1 x disk
| | |--1x--end_device-15:0:10
| | | `-- 1 x disk
| | |--1x--end_device-15:0:11
| | | `-- 1 x disk
| | |--8x--expander-15:2
| | | |-- 39 x end_device -- disk
| | | `-- 1 x end_device -- enclosure
| | `--8x--expander-15:1
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--4x--expander-15:3
| |--1x--end_device-15:3:12
| | `--enclosure
| |--1x--end_device-15:3:0
| | `-- 1 x disk
| |--1x--end_device-15:3:1
| | `-- 1 x disk
| |--1x--end_device-15:3:2
| | `-- 1 x disk
| |--1x--end_device-15:3:3
| | `-- 1 x disk
| |--1x--end_device-15:3:4
| | `-- 1 x disk
| |--1x--end_device-15:3:5
| | `-- 1 x disk
| |--1x--end_device-15:3:6
| | `-- 1 x disk
| |--1x--end_device-15:3:7
| | `-- 1 x disk
| |--1x--end_device-15:3:8
| | `-- 1 x disk
| |--1x--end_device-15:3:9
| | `-- 1 x disk
| |--1x--end_device-15:3:10
| | `-- 1 x disk
| |--1x--end_device-15:3:11
| | `-- 1 x disk
| |--8x--expander-15:5
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--8x--expander-15:4
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--host0
|--4x--expander-0:3
| |--1x--end_device-0:3:12
| | `--enclosure
| |--1x--end_device-0:3:0
| | `-- 1 x disk
| |--1x--end_device-0:3:1
| | `-- 1 x disk
| |--1x--end_device-0:3:2
| | `-- 1 x disk
| |--1x--end_device-0:3:3
| | `-- 1 x disk
| |--1x--end_device-0:3:4
| | `-- 1 x disk
| |--1x--end_device-0:3:5
| | `-- 1 x disk
| |--1x--end_device-0:3:6
| | `-- 1 x disk
| |--1x--end_device-0:3:7
| | `-- 1 x disk
| |--1x--end_device-0:3:8
| | `-- 1 x disk
| |--1x--end_device-0:3:9
| | `-- 1 x disk
| |--1x--end_device-0:3:10
| | `-- 1 x disk
| |--1x--end_device-0:3:11
| | `-- 1 x disk
| |--8x--expander-0:5
| | |-- 39 x end_device -- disk
| | `-- 1 x end_device -- enclosure
| `--8x--expander-0:4
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--4x--expander-0:0
|--1x--end_device-0:0:12
| `--enclosure
|--1x--end_device-0:0:0
| `-- 1 x disk
|--1x--end_device-0:0:1
| `-- 1 x disk
|--1x--end_device-0:0:2
| `-- 1 x disk
|--1x--end_device-0:0:3
| `-- 1 x disk
|--1x--end_device-0:0:4
| `-- 1 x disk
|--1x--end_device-0:0:5
| `-- 1 x disk
|--1x--end_device-0:0:6
| `-- 1 x disk
|--1x--end_device-0:0:7
| `-- 1 x disk
|--1x--end_device-0:0:8
| `-- 1 x disk
|--1x--end_device-0:0:9
| `-- 1 x disk
|--1x--end_device-0:0:10
| `-- 1 x disk
|--1x--end_device-0:0:11
| `-- 1 x disk
|--8x--expander-0:2
| |-- 39 x end_device -- disk
| `-- 1 x end_device -- enclosure
`--8x--expander-0:1
|-- 39 x end_device -- disk
`-- 1 x end_device -- enclosure
```
It looks like internally the IO controllers each have 1 expander going to 12 disks, and cascading 2 more expanders with 39 disks each, a very odd config, but each of the 90 disks is exposed properly on each physical wiring path, as expected.
|
defect
|
vdev id not properly detecting multipath disk configuration thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu distribution version linux kernel generic architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing i have a system with lsi pci express fusion mpt sas connecting to two io controllers on a supermicro disk jbod in a crossover dual channel config io controller channel a io controller channel a io controller channel b io controller channel b this is intended to provide cross failover protection in case of card or io controller failure i currently have this configured with multibus in multipathd and it is properly detecting and setting up aggregated paths snip of multipath ll some addresses greeked dm hgst size features hwhandler wp rw policy round robin prio status active sdht active ready running sdck active ready running sdlf active ready running sdfw active ready running dm hgst size features hwhandler wp rw policy round robin prio status active sdhs active ready running sdcj active ready running sdle active ready running sdfv active ready running when attempting to use vdev id conf using a standard multipath config multipath yes pci id hba port channel name channel a channel b channel af a channel af b i get the following dev disk by vdev entries ls l total lrwxrwxrwx root root oct sdek lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm lrwxrwxrwx root root oct dm note that this is missing many entries currently totaling only when testing vdev id for bugs it does appear to throw an error sh x lib udev vdev id d dev dm path bin sbin usr bin usr sbin config etc zfs vdev id conf phys per port dev multipath topology bay getopts c d eg mp h option dev dev dm getopts c d eg mp h option awk topology print exit etc zfs vdev id conf topology awk slot print exit etc zfs vdev id conf bay topology sas direct alias handler local dm part echo grep q e p id vdev bay bay sas handler awk phys per port print exit etc zfs vdev id conf phys per port phys per port echo grep q e awk multipath print exit etc zfs vdev id conf multipath mode yes ls l full time dev mapper awk dev dm print awk cmd line dev dm print awk cmd line unterminated regexp dm name echo awk fp p print part part echo sed s p dm name return id vdev it returns the same error for mapper entries that do not get vdevs generated sh x lib udev vdev id d dev dm path bin sbin usr bin usr sbin config etc zfs vdev id conf phys per port dev multipath topology bay getopts c d eg mp h option dev dev dm getopts c d eg mp h option awk topology print exit etc zfs vdev id conf topology awk slot print exit etc zfs vdev id conf bay topology sas direct alias handler local dm part echo grep q e p id vdev bay bay sas handler awk phys per port print exit etc zfs vdev id conf phys per port phys per port echo grep q e awk multipath print exit etc zfs vdev id conf multipath mode yes ls l full time dev mapper awk dev dm print awk cmd line dev dm print awk cmd line unterminated regexp dm name echo awk fp p print part part echo sed s p dm name return id vdev i m unclear from the documentation what a number of the terms are for the vdev id conf file in the context of real world hardware configurations how do i determine number of phys how do i know if switch or direct is a hba port the actual physical wire what does channel name refer to so i m having much difficulty guessing as to what settings might need to be changed finally i offer the output of sas discover from which appears to have mapped the entire sas topology it does look a bit odd compared to what i expect to see this may be be due to the internal wiring of the jbod hostname expander end device enclosure end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk expander x end device disk x end device enclosure expander x end device disk x end device enclosure expander end device enclosure end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk expander x end device disk x end device enclosure expander x end device disk x end device enclosure expander end device enclosure end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk expander x end device disk x end device enclosure expander x end device disk x end device enclosure expander end device enclosure end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk end device x disk expander x end device disk x end device enclosure expander x end device disk x end device enclosure it looks like internally the io controllers each have expander going to disks and cascading more expanders with disks each a very odd config but each of the disks is exposed properly on each physical wiring path as expected
| 1
|
187,566
| 14,428,279,328
|
IssuesEvent
|
2020-12-06 09:00:40
|
kalexmills/github-vet-tests-dec2020
|
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
|
closed
|
godeep/lzma: lzma_encoder_test.go; 29 LoC
|
fresh small test
|
Found a possible issue in [godeep/lzma](https://www.github.com/godeep/lzma) at [lzma_encoder_test.go](https://github.com/godeep/lzma/blob/63d93b6f6cbfa0d0a25d882bac75b40d8f22de91/lzma_encoder_test.go#L95-L123)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable tt used in defer or goroutine at line 106
[Click here to see the code in its original context.](https://github.com/godeep/lzma/blob/63d93b6f6cbfa0d0a25d882bac75b40d8f22de91/lzma_encoder_test.go#L95-L123)
<details>
<summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary>
```go
for _, tt := range lzmaTests {
if tt.err == nil {
pr, pw := io.Pipe()
defer pr.Close()
in := bytes.NewBuffer([]byte(tt.raw))
size := int64(-1)
if tt.size == true {
size = int64(len([]byte(tt.raw)))
}
go func() {
defer pw.Close()
w := NewWriterSizeLevel(pw, size, tt.level)
defer w.Close()
_, err := io.Copy(w, in)
if err != nil {
t.Errorf("%v", err)
}
}()
b.Reset()
_, err := io.Copy(b, pr)
if err != nil {
t.Errorf("%v", err)
}
res := b.Bytes()
if bytes.Equal(res, tt.lzma) == false {
t.Errorf("%s: got %d-byte %q, want %d-byte %q", tt.descr, len(res), string(res), len(tt.lzma), string(tt.lzma))
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 63d93b6f6cbfa0d0a25d882bac75b40d8f22de91
|
1.0
|
godeep/lzma: lzma_encoder_test.go; 29 LoC -
Found a possible issue in [godeep/lzma](https://www.github.com/godeep/lzma) at [lzma_encoder_test.go](https://github.com/godeep/lzma/blob/63d93b6f6cbfa0d0a25d882bac75b40d8f22de91/lzma_encoder_test.go#L95-L123)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable tt used in defer or goroutine at line 106
[Click here to see the code in its original context.](https://github.com/godeep/lzma/blob/63d93b6f6cbfa0d0a25d882bac75b40d8f22de91/lzma_encoder_test.go#L95-L123)
<details>
<summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary>
```go
for _, tt := range lzmaTests {
if tt.err == nil {
pr, pw := io.Pipe()
defer pr.Close()
in := bytes.NewBuffer([]byte(tt.raw))
size := int64(-1)
if tt.size == true {
size = int64(len([]byte(tt.raw)))
}
go func() {
defer pw.Close()
w := NewWriterSizeLevel(pw, size, tt.level)
defer w.Close()
_, err := io.Copy(w, in)
if err != nil {
t.Errorf("%v", err)
}
}()
b.Reset()
_, err := io.Copy(b, pr)
if err != nil {
t.Errorf("%v", err)
}
res := b.Bytes()
if bytes.Equal(res, tt.lzma) == false {
t.Errorf("%s: got %d-byte %q, want %d-byte %q", tt.descr, len(res), string(res), len(tt.lzma), string(tt.lzma))
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 63d93b6f6cbfa0d0a25d882bac75b40d8f22de91
|
non_defect
|
godeep lzma lzma encoder test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable tt used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for tt range lzmatests if tt err nil pr pw io pipe defer pr close in bytes newbuffer byte tt raw size if tt size true size len byte tt raw go func defer pw close w newwritersizelevel pw size tt level defer w close err io copy w in if err nil t errorf v err b reset err io copy b pr if err nil t errorf v err res b bytes if bytes equal res tt lzma false t errorf s got d byte q want d byte q tt descr len res string res len tt lzma string tt lzma leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
51,233
| 13,207,398,880
|
IssuesEvent
|
2020-08-14 22:57:22
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
all tickets by milestone report broken (Trac #74)
|
Incomplete Migration Migrated from Trac defect infrastructure
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/74">https://code.icecube.wisc.edu/projects/icecube/ticket/74</a>, reported by troyand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-07-06T14:18:10",
"_ts": "1183731490000000",
"description": "all tickets by milestone report broken.\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-27T18:25:30",
"component": "infrastructure",
"summary": "all tickets by milestone report broken",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
all tickets by milestone report broken (Trac #74) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/74">https://code.icecube.wisc.edu/projects/icecube/ticket/74</a>, reported by troyand owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-07-06T14:18:10",
"_ts": "1183731490000000",
"description": "all tickets by milestone report broken.\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-27T18:25:30",
"component": "infrastructure",
"summary": "all tickets by milestone report broken",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
defect
|
all tickets by milestone report broken trac migrated from json status closed changetime ts description all tickets by milestone report broken n reporter troy cc resolution duplicate time component infrastructure summary all tickets by milestone report broken priority normal keywords milestone owner cgils type defect
| 1
|
34,253
| 7,431,748,803
|
IssuesEvent
|
2018-03-25 17:43:46
|
Yahkal/replicaisland
|
https://api.github.com/repos/Yahkal/replicaisland
|
closed
|
Message window system causes hang in corner case
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Load ReplicantIsland
2. Enter Memory #023
3. Encounter Wanda in the upper-right corner of the map
What is the expected output? What do you see instead?
Expected output: Wanda will chat briefly about her partner.
Actual output: system hangs inside ConversationDialogActivity::formatPages
What version of the product are you using? On what operating system?
Tested on Google Nexus 7 (2012 wifi) with CyanogenMod 10.1.0-RC5 (analogue for
Android 4.2.2)
Please provide any additional information below.
Seems to be caused by the way paint::BreakString method is handled by OS.
Corner case is when lastSpace < currentOffset; causes an infinite loop.
Attached patch fixes issue and technically may make the
ConversationDialogActivity class slightly more robust.
```
Original issue reported on code.google.com by `lord.sop...@gmail.com` on 10 Dec 2013 at 1:36
Attachments:
- [ConversationDialogActivity-InfiniteLoopFix.patch](https://storage.googleapis.com/google-code-attachments/replicaisland/issue-80/comment-0/ConversationDialogActivity-InfiniteLoopFix.patch)
|
1.0
|
Message window system causes hang in corner case - ```
What steps will reproduce the problem?
1. Load ReplicantIsland
2. Enter Memory #023
3. Encounter Wanda in the upper-right corner of the map
What is the expected output? What do you see instead?
Expected output: Wanda will chat briefly about her partner.
Actual output: system hangs inside ConversationDialogActivity::formatPages
What version of the product are you using? On what operating system?
Tested on Google Nexus 7 (2012 wifi) with CyanogenMod 10.1.0-RC5 (analogue for
Android 4.2.2)
Please provide any additional information below.
Seems to be caused by the way paint::BreakString method is handled by OS.
Corner case is when lastSpace < currentOffset; causes an infinite loop.
Attached patch fixes issue and technically may make the
ConversationDialogActivity class slightly more robust.
```
Original issue reported on code.google.com by `lord.sop...@gmail.com` on 10 Dec 2013 at 1:36
Attachments:
- [ConversationDialogActivity-InfiniteLoopFix.patch](https://storage.googleapis.com/google-code-attachments/replicaisland/issue-80/comment-0/ConversationDialogActivity-InfiniteLoopFix.patch)
|
defect
|
message window system causes hang in corner case what steps will reproduce the problem load replicantisland enter memory encounter wanda in the upper right corner of the map what is the expected output what do you see instead expected output wanda will chat briefly about her partner actual output system hangs inside conversationdialogactivity formatpages what version of the product are you using on what operating system tested on google nexus wifi with cyanogenmod analogue for android please provide any additional information below seems to be caused by the way paint breakstring method is handled by os corner case is when lastspace currentoffset causes an infinite loop attached patch fixes issue and technically may make the conversationdialogactivity class slightly more robust original issue reported on code google com by lord sop gmail com on dec at attachments
| 1
|
63,292
| 17,571,460,384
|
IssuesEvent
|
2021-08-14 19:41:59
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Upgrade from 0.8.1 to 0.8.2 caused "Mismatch between pool hostid and system"
|
Type: Defect Status: Stale
|
### System information
<!-- add version after "|" character -->
Distribution Name | Scientific Linux
Distribution Version | 7.7
Linux Kernel | 3.10.0-1062.1.2.el7.x86_64
Architecture | x86-64
ZFS Version | 0.8.2
### Describe the problem you're observing
I was using ZoL version 0.8.1 from zfs-testing repo. After upgrade to 0.8.2 from EL7.7 repo and reboot, I get this
```
# zpool status storage
pool: storage
state: ONLINE
status: Mismatch between pool hostid and system hostid on imported pool.
This pool was previously imported into a system with a different hostid,
and then was verbatim imported into this system.
action: Export this pool on all systems on which it is imported.
Then import it to correct the mismatch.
see: http://zfsonlinux.org/msg/ZFS-8000-EY
scan: scrub repaired 0B in 0 days 07:40:32 with 0 errors on Sat Jun 29 18:33:32 2019
config:
NAME STATE READ WRITE CKSUM
storage ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KE5M ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KEB0 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KEQJ ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KHX1 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KJNK ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KR6D ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KS8R ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KV6J ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840L4A9 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840L4N4 ONLINE 0 0 0
logs
ata-M4-CT064M4SSD2_00000000111903078C20-part3 ONLINE 0 0 0
errors: No known data errors
```
The pool has survived multiple reboots and ZoL version upgrades in the past without such effect. The problem happened to all 3 pools on this system.
I cannot say if the problem is reproducible. I know how to fix it with export/import, I only wanted to point out that it happened by itself after version upgrade.
Host ID was never defined on this system: `/etc/hostid` does not exist, `cat /sys/module/spl/parameters/spl_hostid` returns 0 (just as before).
|
1.0
|
Upgrade from 0.8.1 to 0.8.2 caused "Mismatch between pool hostid and system" - ### System information
<!-- add version after "|" character -->
Distribution Name | Scientific Linux
Distribution Version | 7.7
Linux Kernel | 3.10.0-1062.1.2.el7.x86_64
Architecture | x86-64
ZFS Version | 0.8.2
### Describe the problem you're observing
I was using ZoL version 0.8.1 from zfs-testing repo. After upgrade to 0.8.2 from EL7.7 repo and reboot, I get this
```
# zpool status storage
pool: storage
state: ONLINE
status: Mismatch between pool hostid and system hostid on imported pool.
This pool was previously imported into a system with a different hostid,
and then was verbatim imported into this system.
action: Export this pool on all systems on which it is imported.
Then import it to correct the mismatch.
see: http://zfsonlinux.org/msg/ZFS-8000-EY
scan: scrub repaired 0B in 0 days 07:40:32 with 0 errors on Sat Jun 29 18:33:32 2019
config:
NAME STATE READ WRITE CKSUM
storage ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KE5M ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KEB0 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KEQJ ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KHX1 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KJNK ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KR6D ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KS8R ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840KV6J ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840L4A9 ONLINE 0 0 0
ata-ST8000AS0002-1NA17Z_Z840L4N4 ONLINE 0 0 0
logs
ata-M4-CT064M4SSD2_00000000111903078C20-part3 ONLINE 0 0 0
errors: No known data errors
```
The pool has survived multiple reboots and ZoL version upgrades in the past without such effect. The problem happened to all 3 pools on this system.
I cannot say if the problem is reproducible. I know how to fix it with export/import, I only wanted to point out that it happened by itself after version upgrade.
Host ID was never defined on this system: `/etc/hostid` does not exist, `cat /sys/module/spl/parameters/spl_hostid` returns 0 (just as before).
|
defect
|
upgrade from to caused mismatch between pool hostid and system system information distribution name scientific linux distribution version linux kernel architecture zfs version describe the problem you re observing i was using zol version from zfs testing repo after upgrade to from repo and reboot i get this zpool status storage pool storage state online status mismatch between pool hostid and system hostid on imported pool this pool was previously imported into a system with a different hostid and then was verbatim imported into this system action export this pool on all systems on which it is imported then import it to correct the mismatch see scan scrub repaired in days with errors on sat jun config name state read write cksum storage online online ata online ata online ata online ata online ata online ata online ata online ata online ata online ata online logs ata online errors no known data errors the pool has survived multiple reboots and zol version upgrades in the past without such effect the problem happened to all pools on this system i cannot say if the problem is reproducible i know how to fix it with export import i only wanted to point out that it happened by itself after version upgrade host id was never defined on this system etc hostid does not exist cat sys module spl parameters spl hostid returns just as before
| 1
|
224
| 2,520,400,954
|
IssuesEvent
|
2015-01-19 01:59:39
|
larcenists/larceny
|
https://api.github.com/repos/larcenists/larceny
|
closed
|
syntax definitions don't propagate into fasl files
|
C: TWOBIT C: twobit P: minor T: defect wontfix
|
**Reported by: pnkfelix on Mon Jun 12 03:04:23 2006**
Consider the following interactions with Larceny, where ```foo.sch``` just contains syntax definition:
```(define-syntax bazzle (syntax-rules () ((bazzle) 'basil)))```
```
% larceny
> (compile-file "foo.sch")
> (exit)
% larceny
> (load "foo.sch")
> (bazzle)
'basil
> (exit)
% larceny
> (load "foo.fasl")
> (bazzle)
Error: Reference to undefined global variable "bazzle".
```
The fasl file and the sch file have different effects when you load them. This essentially implies that users have to be very careful about when they compile files that contain syntax definitions. (Perhaps this was a deliberate design choice, but PnkFelix sees it as a bug).
PnkFelix is working around this problem for his designs on ExtendingRequire, but it would be a good thing to think about fixing in general.
|
1.0
|
syntax definitions don't propagate into fasl files - **Reported by: pnkfelix on Mon Jun 12 03:04:23 2006**
Consider the following interactions with Larceny, where ```foo.sch``` just contains syntax definition:
```(define-syntax bazzle (syntax-rules () ((bazzle) 'basil)))```
```
% larceny
> (compile-file "foo.sch")
> (exit)
% larceny
> (load "foo.sch")
> (bazzle)
'basil
> (exit)
% larceny
> (load "foo.fasl")
> (bazzle)
Error: Reference to undefined global variable "bazzle".
```
The fasl file and the sch file have different effects when you load them. This essentially implies that users have to be very careful about when they compile files that contain syntax definitions. (Perhaps this was a deliberate design choice, but PnkFelix sees it as a bug).
PnkFelix is working around this problem for his designs on ExtendingRequire, but it would be a good thing to think about fixing in general.
|
defect
|
syntax definitions don t propagate into fasl files reported by pnkfelix on mon jun consider the following interactions with larceny where foo sch just contains syntax definition define syntax bazzle syntax rules bazzle basil larceny compile file foo sch exit larceny load foo sch bazzle basil exit larceny load foo fasl bazzle error reference to undefined global variable bazzle the fasl file and the sch file have different effects when you load them this essentially implies that users have to be very careful about when they compile files that contain syntax definitions perhaps this was a deliberate design choice but pnkfelix sees it as a bug pnkfelix is working around this problem for his designs on extendingrequire but it would be a good thing to think about fixing in general
| 1
|
21,254
| 3,477,596,474
|
IssuesEvent
|
2015-12-28 02:17:40
|
hackdefendr/gps-gimp-paint-studio
|
https://api.github.com/repos/hackdefendr/gps-gimp-paint-studio
|
closed
|
Wrong file format in patterns directory
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Start Gimp with gps-gimp-paint-studio installed
2.
3.
What is the expected output? What do you see instead?
Couldn't recognize the image file format for file
'/usr/share/gimp/2.0/patterns/GPS-Pat/Canvas-Covered.xcf'
What version of the product are you using? On what operating system?
It looks like Canvas-Covered.xcf should not be in pattern directory
Please provide any additional information below.
```
Original issue reported on code.google.com by `tshim...@gmail.com` on 9 Nov 2014 at 7:31
|
1.0
|
Wrong file format in patterns directory - ```
What steps will reproduce the problem?
1. Start Gimp with gps-gimp-paint-studio installed
2.
3.
What is the expected output? What do you see instead?
Couldn't recognize the image file format for file
'/usr/share/gimp/2.0/patterns/GPS-Pat/Canvas-Covered.xcf'
What version of the product are you using? On what operating system?
It looks like Canvas-Covered.xcf should not be in pattern directory
Please provide any additional information below.
```
Original issue reported on code.google.com by `tshim...@gmail.com` on 9 Nov 2014 at 7:31
|
defect
|
wrong file format in patterns directory what steps will reproduce the problem start gimp with gps gimp paint studio installed what is the expected output what do you see instead couldn t recognize the image file format for file usr share gimp patterns gps pat canvas covered xcf what version of the product are you using on what operating system it looks like canvas covered xcf should not be in pattern directory please provide any additional information below original issue reported on code google com by tshim gmail com on nov at
| 1
|
120,776
| 12,097,140,317
|
IssuesEvent
|
2020-04-20 08:06:01
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
closed
|
недостаточная документация по clickhouse-local
|
comp-documentation
|
добрый день!
к примеру, есть такой файл
```
ilia@localhost:~$ head example.csv
"column1","column2","column3"
1,2,3
1,1,1
2,1,1
```
команда фейлится по непонятной причине. подозреваю, что не описан явно формат. но ведь он, согласно спецификации, описан в первой строке ?
```
ilia@localhost:~$ clickhouse-local -N example --file example.csv --input-format CSV --query "select column1 from example"
Code: 60. DB::Exception: Table _local.example doesn't exist.
ilia@localhost:~$
```
действительно, добавляем явно формат, запрос проходит. но что это ? зачем оно включило первую служебную строку ?
```
ilia@localhost:~$ clickhouse-local -N example --structure "column1 String, column2 String, column3 String" --file example.csv --input-format CSV --query "select column1 from example"
Read 4 rows, 138.00 B in 0.001 sec., 5748 rows/sec., 193.68 KiB/sec.
column1
1
1
2
ilia@localhost:~$
```
кажется, тут есть серьезные недоработки с поддержкой CSV (и с документацией)
|
1.0
|
недостаточная документация по clickhouse-local - добрый день!
к примеру, есть такой файл
```
ilia@localhost:~$ head example.csv
"column1","column2","column3"
1,2,3
1,1,1
2,1,1
```
команда фейлится по непонятной причине. подозреваю, что не описан явно формат. но ведь он, согласно спецификации, описан в первой строке ?
```
ilia@localhost:~$ clickhouse-local -N example --file example.csv --input-format CSV --query "select column1 from example"
Code: 60. DB::Exception: Table _local.example doesn't exist.
ilia@localhost:~$
```
действительно, добавляем явно формат, запрос проходит. но что это ? зачем оно включило первую служебную строку ?
```
ilia@localhost:~$ clickhouse-local -N example --structure "column1 String, column2 String, column3 String" --file example.csv --input-format CSV --query "select column1 from example"
Read 4 rows, 138.00 B in 0.001 sec., 5748 rows/sec., 193.68 KiB/sec.
column1
1
1
2
ilia@localhost:~$
```
кажется, тут есть серьезные недоработки с поддержкой CSV (и с документацией)
|
non_defect
|
недостаточная документация по clickhouse local добрый день к примеру есть такой файл ilia localhost head example csv команда фейлится по непонятной причине подозреваю что не описан явно формат но ведь он согласно спецификации описан в первой строке ilia localhost clickhouse local n example file example csv input format csv query select from example code db exception table local example doesn t exist ilia localhost действительно добавляем явно формат запрос проходит но что это зачем оно включило первую служебную строку ilia localhost clickhouse local n example structure string string string file example csv input format csv query select from example read rows b in sec rows sec kib sec ilia localhost кажется тут есть серьезные недоработки с поддержкой csv и с документацией
| 0
|
398,986
| 11,742,577,996
|
IssuesEvent
|
2020-03-12 01:18:41
|
thaliawww/concrexit
|
https://api.github.com/repos/thaliawww/concrexit
|
closed
|
Textual change on society page
|
activemembers feature priority: medium
|
In GitLab by rwechel on Dec 8, 2018, 17:22
### One-sentence description
A link on the society page to the society policy document + application form.
### Motivation
It is a must for anyone that wants to form a society that they read the policy document and fill in an application form. Currently there is no official way to know how to form one.
### Desired functionality
Adding a simple text + link on the society page, written below.
### Suggested implementation
"Gezelschap aanvragen? Lees het beleidsdocument goed door:
https://thalia.nu/documents/document/381/
Het aanvraagformulier kan onderaan het beleidsdocument worden gevonden.
Om een gezelschap aan te vragen dien je het ingevulde formulier bij het
bestuur aan te leveren."
"Interested in forming a society? Read the policy document carefully:
https://thalia.nu/documents/document/381/
The application form can be found at the bottom of the policy document.
To request a society you must submit the filled in form to the board."
|
1.0
|
Textual change on society page - In GitLab by rwechel on Dec 8, 2018, 17:22
### One-sentence description
A link on the society page to the society policy document + application form.
### Motivation
It is a must for anyone that wants to form a society that they read the policy document and fill in an application form. Currently there is no official way to know how to form one.
### Desired functionality
Adding a simple text + link on the society page, written below.
### Suggested implementation
"Gezelschap aanvragen? Lees het beleidsdocument goed door:
https://thalia.nu/documents/document/381/
Het aanvraagformulier kan onderaan het beleidsdocument worden gevonden.
Om een gezelschap aan te vragen dien je het ingevulde formulier bij het
bestuur aan te leveren."
"Interested in forming a society? Read the policy document carefully:
https://thalia.nu/documents/document/381/
The application form can be found at the bottom of the policy document.
To request a society you must submit the filled in form to the board."
|
non_defect
|
textual change on society page in gitlab by rwechel on dec one sentence description a link on the society page to the society policy document application form motivation it is a must for anyone that wants to form a society that they read the policy document and fill in an application form currently there is no official way to know how to form one desired functionality adding a simple text link on the society page written below suggested implementation gezelschap aanvragen lees het beleidsdocument goed door het aanvraagformulier kan onderaan het beleidsdocument worden gevonden om een gezelschap aan te vragen dien je het ingevulde formulier bij het bestuur aan te leveren interested in forming a society read the policy document carefully the application form can be found at the bottom of the policy document to request a society you must submit the filled in form to the board
| 0
|
43,808
| 11,851,078,457
|
IssuesEvent
|
2020-03-24 17:34:08
|
mestrade/go-hello
|
https://api.github.com/repos/mestrade/go-hello
|
opened
|
CVE-2019-17531 - Jackson-Databind-2.6.0(java)
|
security/defectDojo
|
*CVE-2019-17531 - Jackson-Databind-2.6.0(java)*
*Severity:* Critical
*Cve:* CVE-2019-17531
*Product/Engagement:* test / AdHoc Import - Tue, 24 Mar 2020 15:57:40
*Systems*:
*Description*:
Image hash: sha256:8a3e381ece363cb5f0187e5f24988a8febd98e76cd5bc0562d443845066d6e58
Package: jackson-databind-2.6.0
Package path: /usr/share/jenkins/jenkins.war:WEB-INF/plugins/github.hpi:WEB-INF/lib/jackson-databind-2.6.0.jar
Package type: java
Feed: nvdv2/nvdv2:cves
CVE: CVE-2019-17531
CPE: cpe:/a:-:jackson-databind:2.6.0:-:-
*Mitigation*:
Upgrade to jackson-databind None
URL: https://nvd.nist.gov/vuln/detail/CVE-2019-17531
*Impact*:
*References*:https://nvd.nist.gov/vuln/detail/CVE-2019-17531
|
1.0
|
CVE-2019-17531 - Jackson-Databind-2.6.0(java) - *CVE-2019-17531 - Jackson-Databind-2.6.0(java)*
*Severity:* Critical
*Cve:* CVE-2019-17531
*Product/Engagement:* test / AdHoc Import - Tue, 24 Mar 2020 15:57:40
*Systems*:
*Description*:
Image hash: sha256:8a3e381ece363cb5f0187e5f24988a8febd98e76cd5bc0562d443845066d6e58
Package: jackson-databind-2.6.0
Package path: /usr/share/jenkins/jenkins.war:WEB-INF/plugins/github.hpi:WEB-INF/lib/jackson-databind-2.6.0.jar
Package type: java
Feed: nvdv2/nvdv2:cves
CVE: CVE-2019-17531
CPE: cpe:/a:-:jackson-databind:2.6.0:-:-
*Mitigation*:
Upgrade to jackson-databind None
URL: https://nvd.nist.gov/vuln/detail/CVE-2019-17531
*Impact*:
*References*:https://nvd.nist.gov/vuln/detail/CVE-2019-17531
|
defect
|
cve jackson databind java cve jackson databind java severity critical cve cve product engagement test adhoc import tue mar systems description image hash package jackson databind package path usr share jenkins jenkins war web inf plugins github hpi web inf lib jackson databind jar package type java feed cves cve cve cpe cpe a jackson databind mitigation upgrade to jackson databind none url impact references
| 1
|
44,049
| 11,938,394,667
|
IssuesEvent
|
2020-04-02 13:43:38
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Timeline: edit event not working when editableGroup is false
|
defect
|
## 1) Environment
- PrimeFaces version: **8.0**
- Does it work on the newest released PrimeFaces version? **No** Version? **9.0-SNAPSHOT**
- Does it work on the newest sources in GitHub? **No**
- Application server + version: **Payara 5.201**
- Affected browsers: **All**
## 2) Expected behavior
When i disable editableGroup global on the timeline events related to a group should fire edit event when i double click them.
## 3) Actual behavior
Whe i disable editableGroup global, events related to a group not firing edit event when i double click the event.
## 4) Steps to reproduce
In showcase project timeline grouping add following line to the timeline tag:
``` xhtml
<p:ajax event="edit" oncomplete="alert('test');" />
```
Then when you run showcase and doubleclick an event the alert is showing up.
Next set editableGroup="false" on the timeline definition tag. So the alert is no longer showing up.
Full timeline definition in showcase timeline grouping page:
``` xhtml
<p:timeline id="timeline" value="#{groupingTimelineView.model}" var="order" varGroup="truck"
editable="true" editableGroup="false" eventMargin="0" eventMarginAxis="0" stackEvents="false"
orientationAxis="top" widgetVar="timelineWdgt">
<p:ajax event="changed" update="@none" listener="#{groupingTimelineView.onChange}"/>
<p:ajax event="delete" update="@none" listener="#{groupingTimelineView.onDelete}"/>
<p:ajax event="add" update="@none" onstart="PF('timelineWdgt').cancelAdd()"/>
<p:ajax event="edit" oncomplete="alert('test');" />
<f:facet name="group">
<h:graphicImage library="demo" name="images/timeline/truck.png" style="vertical-align:middle;" alt="Truck"/>
<h:outputText value="Truck #{truck.code}" style="font-weight:bold;"/>
</f:facet>
<h:graphicImage library="demo" name="#{order.imagePath}" rendered="#{not empty order.imagePath}"
style="display:inline; vertical-align:middle;" alt="Order"/>
<h:outputText value="Order #{order.number}"/>
</p:timeline>
```
## 5) Sample XHTML
..
## 6) Sample bean
..
|
1.0
|
Timeline: edit event not working when editableGroup is false - ## 1) Environment
- PrimeFaces version: **8.0**
- Does it work on the newest released PrimeFaces version? **No** Version? **9.0-SNAPSHOT**
- Does it work on the newest sources in GitHub? **No**
- Application server + version: **Payara 5.201**
- Affected browsers: **All**
## 2) Expected behavior
When i disable editableGroup global on the timeline events related to a group should fire edit event when i double click them.
## 3) Actual behavior
Whe i disable editableGroup global, events related to a group not firing edit event when i double click the event.
## 4) Steps to reproduce
In showcase project timeline grouping add following line to the timeline tag:
``` xhtml
<p:ajax event="edit" oncomplete="alert('test');" />
```
Then when you run showcase and doubleclick an event the alert is showing up.
Next set editableGroup="false" on the timeline definition tag. So the alert is no longer showing up.
Full timeline definition in showcase timeline grouping page:
``` xhtml
<p:timeline id="timeline" value="#{groupingTimelineView.model}" var="order" varGroup="truck"
editable="true" editableGroup="false" eventMargin="0" eventMarginAxis="0" stackEvents="false"
orientationAxis="top" widgetVar="timelineWdgt">
<p:ajax event="changed" update="@none" listener="#{groupingTimelineView.onChange}"/>
<p:ajax event="delete" update="@none" listener="#{groupingTimelineView.onDelete}"/>
<p:ajax event="add" update="@none" onstart="PF('timelineWdgt').cancelAdd()"/>
<p:ajax event="edit" oncomplete="alert('test');" />
<f:facet name="group">
<h:graphicImage library="demo" name="images/timeline/truck.png" style="vertical-align:middle;" alt="Truck"/>
<h:outputText value="Truck #{truck.code}" style="font-weight:bold;"/>
</f:facet>
<h:graphicImage library="demo" name="#{order.imagePath}" rendered="#{not empty order.imagePath}"
style="display:inline; vertical-align:middle;" alt="Order"/>
<h:outputText value="Order #{order.number}"/>
</p:timeline>
```
## 5) Sample XHTML
..
## 6) Sample bean
..
|
defect
|
timeline edit event not working when editablegroup is false environment primefaces version does it work on the newest released primefaces version no version snapshot does it work on the newest sources in github no application server version payara affected browsers all expected behavior when i disable editablegroup global on the timeline events related to a group should fire edit event when i double click them actual behavior whe i disable editablegroup global events related to a group not firing edit event when i double click the event steps to reproduce in showcase project timeline grouping add following line to the timeline tag xhtml then when you run showcase and doubleclick an event the alert is showing up next set editablegroup false on the timeline definition tag so the alert is no longer showing up full timeline definition in showcase timeline grouping page xhtml p timeline id timeline value groupingtimelineview model var order vargroup truck editable true editablegroup false eventmargin eventmarginaxis stackevents false orientationaxis top widgetvar timelinewdgt h graphicimage library demo name order imagepath rendered not empty order imagepath style display inline vertical align middle alt order sample xhtml sample bean
| 1
|
191,969
| 15,307,985,544
|
IssuesEvent
|
2021-02-24 21:44:05
|
anitab-org/bridge-in-tech-backend
|
https://api.github.com/repos/anitab-org/bridge-in-tech-backend
|
closed
|
Feat: Migrating documentation from wiki to docusaurus
|
Category: Documentation/Training Type: Maintenance
|
**Is your feature request related to a problem? Please describe.**
Current documentation is fine but with `docusaurus` , the documentation becomes more readable and maintainable.
**Describe the solution you'd like**
Moving the project's documentation to [docusaurus](https://v2.docusaurus.io).
**Additional context**
Mentorship System and Open source programs projects are moving documentation to docusaurus.
|
1.0
|
Feat: Migrating documentation from wiki to docusaurus - **Is your feature request related to a problem? Please describe.**
Current documentation is fine but with `docusaurus` , the documentation becomes more readable and maintainable.
**Describe the solution you'd like**
Moving the project's documentation to [docusaurus](https://v2.docusaurus.io).
**Additional context**
Mentorship System and Open source programs projects are moving documentation to docusaurus.
|
non_defect
|
feat migrating documentation from wiki to docusaurus is your feature request related to a problem please describe current documentation is fine but with docusaurus the documentation becomes more readable and maintainable describe the solution you d like moving the project s documentation to additional context mentorship system and open source programs projects are moving documentation to docusaurus
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.