Unnamed: 0 int64 3 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 2 430 | labels stringlengths 4 347 | body stringlengths 5 237k | index stringclasses 7
values | text_combine stringlengths 96 237k | label stringclasses 2
values | text stringlengths 96 219k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
208,723 | 16,135,389,564 | IssuesEvent | 2021-04-29 11:08:04 | vmware-samples/vcenter-event-broker-appliance | https://api.github.com/repos/vmware-samples/vcenter-event-broker-appliance | opened | Update documentation to include multiple trigger example | bug documentation | **Describe the bug**
Add an example where multiple Knative Triggers are defined which is required if you wish to have a function map to multiple vCenter Server Events | 1.0 | Update documentation to include multiple trigger example - **Describe the bug**
Add an example where multiple Knative Triggers are defined which is required if you wish to have a function map to multiple vCenter Server Events | non_comp | update documentation to include multiple trigger example describe the bug add an example where multiple knative triggers are defined which is required if you wish to have a function map to multiple vcenter server events | 0 |
355,466 | 10,581,014,261 | IssuesEvent | 2019-10-08 08:17:14 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Panic during ACL per parth merging | Priority: Medium bug | **Branch/Environment/Version**
master/release-2.9
**Describe the bug**
There is bug in ACL merging mechanism, which can cause panic if the number of APIs in one policy does not match another policy, and both of them define ACL rules per path for the same API.
**Actual behavior**
```
2019/10/07 20:12:57 http: panic serving 172.30.1.226:45638: runtime error: index out of range
goroutine 111003154 [running]:
net/http.(*conn).serve.func1(0xc075cc8f00)
/usr/local/go/src/net/http/server.go:1769 +0x139
panic(0x1020c20, 0x1d8a6e0)
/usr/local/go/src/runtime/panic.go:522 +0x1b5
github.com/TykTechnologies/tyk/gateway.BaseMiddleware.ApplyPolicies(0xc117c6ca80, 0x145bc60, 0xc07c86d630, 0xc07e981030, 0xc049fb24e8, 0x4059000000000000, 0x4059000000000000)
/src/github.com/TykTechnologies/tyk/gateway/middleware.go:365 +0x2920
github.com/TykTechnologies/tyk/gateway.BaseMiddleware.CheckSessionAndIdentityForValidKey(0xc117c6ca80, 0x145bc60, 0xc07c86d630, 0xc07e981030, 0xc0feb16e80, 0x38, 0xc05211e700, 0x0, 0x0, 0x0, ...)
/src/github.com/TykTechnologies/tyk/gateway/middleware.go:571 +0x2b7
github.com/TykTechnologies/tyk/gateway.(*AuthKey).ProcessRequest(0xc163cde0c0, 0x145e2e0, 0xc0f4e33ce0, 0xc05211e700, 0x0, 0x0, 0xc00029c001, 0xc072460f60, 0xc049fb3170)
/src/github.com/TykTechnologies/tyk/gateway/mw_auth_key.go:93 +0x44e
github.com/TykTechnologies/tyk/gateway.TraceMiddleware
```
**Expected behavior**
Should not panic
| 1.0 | Panic during ACL per parth merging - **Branch/Environment/Version**
master/release-2.9
**Describe the bug**
There is bug in ACL merging mechanism, which can cause panic if the number of APIs in one policy does not match another policy, and both of them define ACL rules per path for the same API.
**Actual behavior**
```
2019/10/07 20:12:57 http: panic serving 172.30.1.226:45638: runtime error: index out of range
goroutine 111003154 [running]:
net/http.(*conn).serve.func1(0xc075cc8f00)
/usr/local/go/src/net/http/server.go:1769 +0x139
panic(0x1020c20, 0x1d8a6e0)
/usr/local/go/src/runtime/panic.go:522 +0x1b5
github.com/TykTechnologies/tyk/gateway.BaseMiddleware.ApplyPolicies(0xc117c6ca80, 0x145bc60, 0xc07c86d630, 0xc07e981030, 0xc049fb24e8, 0x4059000000000000, 0x4059000000000000)
/src/github.com/TykTechnologies/tyk/gateway/middleware.go:365 +0x2920
github.com/TykTechnologies/tyk/gateway.BaseMiddleware.CheckSessionAndIdentityForValidKey(0xc117c6ca80, 0x145bc60, 0xc07c86d630, 0xc07e981030, 0xc0feb16e80, 0x38, 0xc05211e700, 0x0, 0x0, 0x0, ...)
/src/github.com/TykTechnologies/tyk/gateway/middleware.go:571 +0x2b7
github.com/TykTechnologies/tyk/gateway.(*AuthKey).ProcessRequest(0xc163cde0c0, 0x145e2e0, 0xc0f4e33ce0, 0xc05211e700, 0x0, 0x0, 0xc00029c001, 0xc072460f60, 0xc049fb3170)
/src/github.com/TykTechnologies/tyk/gateway/mw_auth_key.go:93 +0x44e
github.com/TykTechnologies/tyk/gateway.TraceMiddleware
```
**Expected behavior**
Should not panic
| non_comp | panic during acl per parth merging branch environment version master release describe the bug there is bug in acl merging mechanism which can cause panic if the number of apis in one policy does not match another policy and both of them define acl rules per path for the same api actual behavior http panic serving runtime error index out of range goroutine net http conn serve usr local go src net http server go panic usr local go src runtime panic go github com tyktechnologies tyk gateway basemiddleware applypolicies src github com tyktechnologies tyk gateway middleware go github com tyktechnologies tyk gateway basemiddleware checksessionandidentityforvalidkey src github com tyktechnologies tyk gateway middleware go github com tyktechnologies tyk gateway authkey processrequest src github com tyktechnologies tyk gateway mw auth key go github com tyktechnologies tyk gateway tracemiddleware expected behavior should not panic | 0 |
10,166 | 12,179,577,006 | IssuesEvent | 2020-04-28 10:55:52 | pbrezina/sssd-test | https://api.github.com/repos/pbrezina/sssd-test | closed | Remove the pysss.local interface | Breaks compatibility Closed: Fixed | Cloned from Pagure issue: https://pagure.io/SSSD/sssd/issue/3493
- Created at 2017-08-30 15:31:27 by [jhrozek](https://pagure.io/user/jhrozek)
- Closed at 2018-08-13 18:34:58 as Fixed
- Assigned to [fidencio](https://pagure.io/user/fidencio)
---
We are de-emphasizing the local domain in favor of the files domain. There is no point in offering a python interface for the local domain. We should just remove it.
### Comments
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-08-31 21:38:32](https://pagure.io/SSSD/sssd/issue/3493#462676)**
**Metadata Update from @jhrozek**:
- Issue set to the milestone: SSSD 2.0
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-09-14 09:01:18](https://pagure.io/SSSD/sssd/issue/3493#465416)**
**Metadata Update from @jhrozek**:
- Issue tagged with: breaks compatibility
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-10-31 12:41:24](https://pagure.io/SSSD/sssd/issue/3493#476085)**
PR:https://github.com/SSSD/sssd/pull/430
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-12-07 20:33:31](https://pagure.io/SSSD/sssd/issue/3493#483532)**
**Metadata Update from @jhrozek**:
- Issue priority set to: minor
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-06 19:29:56](https://pagure.io/SSSD/sssd/issue/3493#525141)**
PR: part of https://github.com/SSSD/sssd/pull/611
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-06 19:30:00](https://pagure.io/SSSD/sssd/issue/3493#525142)**
**Metadata Update from @fidencio**:
- Issue assigned to fidencio
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-07 10:13:53](https://pagure.io/SSSD/sssd/issue/3493#525186)**
**Metadata Update from @fidencio**:
- Issue tagged with: PR
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-13 18:34:48](https://pagure.io/SSSD/sssd/issue/3493#525962)**
Fixed as part of:
0e211b8ba30c3adcdeef21ca1339b194cbfffb04
82d51b7feba8df49df3f7a023d2b16aba26a858e
b8946c46e04e8b5329e4d95f9a67b4a76f883bb2
c075e2865b2fbedb6a7eeb6a90780c6dbffeb5fd
15342ebe8c22738e947728f06c6833b5be09eca8
5a87af9128c5468ce1f2e55602c9d3ec05987b8d
a24f0c202ed7ea159ad04eb19b506b7f21ad41cc
2e8fe6a3dd022721e1ff476c8fd7af317a535d63
728e4be10fc7a7b3173b74b1f597e835914c9abf
99b5bb54447287da5031dcd2bfebeef8b47d52d7
a8a9e66a852ba8aa1371e2e1897d49f3feee76da
064ca0b46eed06977dbd32c2bd894f71dc43be2e
6ebcc59b9458c5a118cd182c07a840fa6756a43e
35a200d5bd84d6a6742e77ffda9945e37134d721
2243b348976705ca81da954cb1ead1f406d01918
7d483737f3763947a5b8a43efbfaf31f54f956a4
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-13 18:35:00](https://pagure.io/SSSD/sssd/issue/3493#525963)**
**Metadata Update from @jhrozek**:
- Issue close_status updated to: Fixed
- Issue status updated to: Closed (was: Open)
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 09:11:38](https://pagure.io/SSSD/sssd/issue/3493#537608)**
Unfortunately, FreeIPA *does* use this interface and now we are broken:
```
[root@cloud-qe-19 ~]# ipa hbactest --host=cloud-qe-19.octy120.test --service=sshd --user='IPAAD2016\aduser1'
ipa: ERROR: an internal error has occurred
[root@cloud-qe-19 ~]# tail -30 /var/log/httpd/error_log
[Wed Oct 24 03:03:47.477681 2018] [wsgi:error] [pid 8771:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Destroyed connection context.ldap2_140573265369968
[Wed Oct 24 03:03:47.493424 2018] [:warn] [pid 8775:tid 140573163890432] [client 10.19.34.59:34812] failed to set perms (3140) on file (/var/run/ipa/ccaches/admin@OCTY120.TEST)!, referer: https://cloud-qe-19.octy120.test/ipa/xml
[Wed Oct 24 03:03:47.493951 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI wsgi_dispatch.__call__:
[Wed Oct 24 03:03:47.494025 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI jsonserver_session.__call__:
[Wed Oct 24 03:03:47.526543 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Created connection context.ldap2_140573265365760
[Wed Oct 24 03:03:47.526616 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI jsonserver.__call__:
[Wed Oct 24 03:03:47.526666 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI WSGIExecutioner.__call__:
[Wed Oct 24 03:03:47.526915 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: raw: hbactest(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', version='2.230')
[Wed Oct 24 03:03:47.527093 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: hbactest(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', nodetail=False, enabled=False, disabled=False, version='2.230')
[Wed Oct 24 03:03:47.527386 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: raw: hbacrule_find(None, sizelimit=None, version='2.230', no_members=False)
[Wed Oct 24 03:03:47.527566 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: hbacrule_find(None, sizelimit=None, all=False, raw=False, version='2.230', no_members=False, pkey_only=False)
[Wed Oct 24 03:03:47.529372 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: retrieving schema for SchemaCache url=ldapi://%2fvar%2frun%2fslapd-OCTY120-TEST.socket conn=<ldap.ldapobject.SimpleLDAPObject object at 0x7fd9c2fb3e80>
[Wed Oct 24 03:03:47.717473 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: ERROR: non-public: AttributeError: module 'pysss' has no attribute 'getgrouplist'
[Wed Oct 24 03:03:47.717492 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] Traceback (most recent call last):
[Wed Oct 24 03:03:47.717495 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/rpcserver.py", line 370, in wsgi_execute
[Wed Oct 24 03:03:47.717499 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] result = command(*args, **options)
[Wed Oct 24 03:03:47.717501 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 450, in __call__
[Wed Oct 24 03:03:47.717504 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] return self.__do_call(*args, **options)
[Wed Oct 24 03:03:47.717510 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 478, in __do_call
[Wed Oct 24 03:03:47.717514 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ret = self.run(*args, **options)
[Wed Oct 24 03:03:47.717516 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 800, in run
[Wed Oct 24 03:03:47.717519 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] return self.execute(*args, **options)
[Wed Oct 24 03:03:47.717522 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/plugins/hbactest.py", line 410, in execute
[Wed Oct 24 03:03:47.717525 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] user_sid, group_sids = domain_validator.get_trusted_domain_user_and_groups(options['user'])
[Wed Oct 24 03:03:47.717527 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/dcerpc.py", line 614, in get_trusted_domain_user_and_groups
[Wed Oct 24 03:03:47.717530 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] group_list = pysss.getgrouplist(object_name)
[Wed Oct 24 03:03:47.717534 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] AttributeError: module 'pysss' has no attribute 'getgrouplist'
[Wed Oct 24 03:03:47.717541 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812]
[Wed Oct 24 03:03:47.717668 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: INFO: [jsonserver_session] admin@OCTY120.TEST: hbactest/1(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', version='2.230'): InternalError
[Wed Oct 24 03:03:47.718159 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Destroyed connection context.ldap2_140573265365760
```
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:00:25](https://pagure.io/SSSD/sssd/issue/3493#537622)**
**Metadata Update from @jhrozek**:
- Custom field rhbz adjusted to https://bugzilla.redhat.com/show_bug.cgi?id=1642372
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:00:25](https://pagure.io/SSSD/sssd/issue/3493#537623)**
Issue linked to Bugzilla: [Bug 1642372](https://bugzilla.redhat.com/show_bug.cgi?id=1642372)
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:02:20](https://pagure.io/SSSD/sssd/issue/3493#537626)**
* master: 0603645f5ea5f707875807b4f815400f4b79e41
This re-adds the getgroupslist() api
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:02:21](https://pagure.io/SSSD/sssd/issue/3493#537627)**
**Metadata Update from @jhrozek**:
- Issue set to the milestone: None (was: SSSD 2.0)
| True | Remove the pysss.local interface - Cloned from Pagure issue: https://pagure.io/SSSD/sssd/issue/3493
- Created at 2017-08-30 15:31:27 by [jhrozek](https://pagure.io/user/jhrozek)
- Closed at 2018-08-13 18:34:58 as Fixed
- Assigned to [fidencio](https://pagure.io/user/fidencio)
---
We are de-emphasizing the local domain in favor of the files domain. There is no point in offering a python interface for the local domain. We should just remove it.
### Comments
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-08-31 21:38:32](https://pagure.io/SSSD/sssd/issue/3493#462676)**
**Metadata Update from @jhrozek**:
- Issue set to the milestone: SSSD 2.0
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-09-14 09:01:18](https://pagure.io/SSSD/sssd/issue/3493#465416)**
**Metadata Update from @jhrozek**:
- Issue tagged with: breaks compatibility
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-10-31 12:41:24](https://pagure.io/SSSD/sssd/issue/3493#476085)**
PR:https://github.com/SSSD/sssd/pull/430
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2017-12-07 20:33:31](https://pagure.io/SSSD/sssd/issue/3493#483532)**
**Metadata Update from @jhrozek**:
- Issue priority set to: minor
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-06 19:29:56](https://pagure.io/SSSD/sssd/issue/3493#525141)**
PR: part of https://github.com/SSSD/sssd/pull/611
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-06 19:30:00](https://pagure.io/SSSD/sssd/issue/3493#525142)**
**Metadata Update from @fidencio**:
- Issue assigned to fidencio
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-07 10:13:53](https://pagure.io/SSSD/sssd/issue/3493#525186)**
**Metadata Update from @fidencio**:
- Issue tagged with: PR
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-13 18:34:48](https://pagure.io/SSSD/sssd/issue/3493#525962)**
Fixed as part of:
0e211b8ba30c3adcdeef21ca1339b194cbfffb04
82d51b7feba8df49df3f7a023d2b16aba26a858e
b8946c46e04e8b5329e4d95f9a67b4a76f883bb2
c075e2865b2fbedb6a7eeb6a90780c6dbffeb5fd
15342ebe8c22738e947728f06c6833b5be09eca8
5a87af9128c5468ce1f2e55602c9d3ec05987b8d
a24f0c202ed7ea159ad04eb19b506b7f21ad41cc
2e8fe6a3dd022721e1ff476c8fd7af317a535d63
728e4be10fc7a7b3173b74b1f597e835914c9abf
99b5bb54447287da5031dcd2bfebeef8b47d52d7
a8a9e66a852ba8aa1371e2e1897d49f3feee76da
064ca0b46eed06977dbd32c2bd894f71dc43be2e
6ebcc59b9458c5a118cd182c07a840fa6756a43e
35a200d5bd84d6a6742e77ffda9945e37134d721
2243b348976705ca81da954cb1ead1f406d01918
7d483737f3763947a5b8a43efbfaf31f54f956a4
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-08-13 18:35:00](https://pagure.io/SSSD/sssd/issue/3493#525963)**
**Metadata Update from @jhrozek**:
- Issue close_status updated to: Fixed
- Issue status updated to: Closed (was: Open)
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 09:11:38](https://pagure.io/SSSD/sssd/issue/3493#537608)**
Unfortunately, FreeIPA *does* use this interface and now we are broken:
```
[root@cloud-qe-19 ~]# ipa hbactest --host=cloud-qe-19.octy120.test --service=sshd --user='IPAAD2016\aduser1'
ipa: ERROR: an internal error has occurred
[root@cloud-qe-19 ~]# tail -30 /var/log/httpd/error_log
[Wed Oct 24 03:03:47.477681 2018] [wsgi:error] [pid 8771:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Destroyed connection context.ldap2_140573265369968
[Wed Oct 24 03:03:47.493424 2018] [:warn] [pid 8775:tid 140573163890432] [client 10.19.34.59:34812] failed to set perms (3140) on file (/var/run/ipa/ccaches/admin@OCTY120.TEST)!, referer: https://cloud-qe-19.octy120.test/ipa/xml
[Wed Oct 24 03:03:47.493951 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI wsgi_dispatch.__call__:
[Wed Oct 24 03:03:47.494025 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI jsonserver_session.__call__:
[Wed Oct 24 03:03:47.526543 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Created connection context.ldap2_140573265365760
[Wed Oct 24 03:03:47.526616 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI jsonserver.__call__:
[Wed Oct 24 03:03:47.526666 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: WSGI WSGIExecutioner.__call__:
[Wed Oct 24 03:03:47.526915 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: raw: hbactest(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', version='2.230')
[Wed Oct 24 03:03:47.527093 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: hbactest(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', nodetail=False, enabled=False, disabled=False, version='2.230')
[Wed Oct 24 03:03:47.527386 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: raw: hbacrule_find(None, sizelimit=None, version='2.230', no_members=False)
[Wed Oct 24 03:03:47.527566 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: hbacrule_find(None, sizelimit=None, all=False, raw=False, version='2.230', no_members=False, pkey_only=False)
[Wed Oct 24 03:03:47.529372 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: retrieving schema for SchemaCache url=ldapi://%2fvar%2frun%2fslapd-OCTY120-TEST.socket conn=<ldap.ldapobject.SimpleLDAPObject object at 0x7fd9c2fb3e80>
[Wed Oct 24 03:03:47.717473 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: ERROR: non-public: AttributeError: module 'pysss' has no attribute 'getgrouplist'
[Wed Oct 24 03:03:47.717492 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] Traceback (most recent call last):
[Wed Oct 24 03:03:47.717495 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/rpcserver.py", line 370, in wsgi_execute
[Wed Oct 24 03:03:47.717499 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] result = command(*args, **options)
[Wed Oct 24 03:03:47.717501 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 450, in __call__
[Wed Oct 24 03:03:47.717504 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] return self.__do_call(*args, **options)
[Wed Oct 24 03:03:47.717510 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 478, in __do_call
[Wed Oct 24 03:03:47.717514 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ret = self.run(*args, **options)
[Wed Oct 24 03:03:47.717516 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipalib/frontend.py", line 800, in run
[Wed Oct 24 03:03:47.717519 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] return self.execute(*args, **options)
[Wed Oct 24 03:03:47.717522 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/plugins/hbactest.py", line 410, in execute
[Wed Oct 24 03:03:47.717525 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] user_sid, group_sids = domain_validator.get_trusted_domain_user_and_groups(options['user'])
[Wed Oct 24 03:03:47.717527 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] File "/usr/lib/python3.6/site-packages/ipaserver/dcerpc.py", line 614, in get_trusted_domain_user_and_groups
[Wed Oct 24 03:03:47.717530 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] group_list = pysss.getgrouplist(object_name)
[Wed Oct 24 03:03:47.717534 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] AttributeError: module 'pysss' has no attribute 'getgrouplist'
[Wed Oct 24 03:03:47.717541 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812]
[Wed Oct 24 03:03:47.717668 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: INFO: [jsonserver_session] admin@OCTY120.TEST: hbactest/1(user='IPAAD2016\\\\aduser1', targethost='cloud-qe-19.octy120.test', service='sshd', version='2.230'): InternalError
[Wed Oct 24 03:03:47.718159 2018] [wsgi:error] [pid 8774:tid 140573238957824] [remote 10.19.34.59:34812] ipa: DEBUG: Destroyed connection context.ldap2_140573265365760
```
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:00:25](https://pagure.io/SSSD/sssd/issue/3493#537622)**
**Metadata Update from @jhrozek**:
- Custom field rhbz adjusted to https://bugzilla.redhat.com/show_bug.cgi?id=1642372
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:00:25](https://pagure.io/SSSD/sssd/issue/3493#537623)**
Issue linked to Bugzilla: [Bug 1642372](https://bugzilla.redhat.com/show_bug.cgi?id=1642372)
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:02:20](https://pagure.io/SSSD/sssd/issue/3493#537626)**
* master: 0603645f5ea5f707875807b4f815400f4b79e41
This re-adds the getgroupslist() api
---
**Comment from [jhrozek](https://pagure.io/user/jhrozek) at [2018-10-24 12:02:21](https://pagure.io/SSSD/sssd/issue/3493#537627)**
**Metadata Update from @jhrozek**:
- Issue set to the milestone: None (was: SSSD 2.0)
| comp | remove the pysss local interface cloned from pagure issue created at by closed at as fixed assigned to we are de emphasizing the local domain in favor of the files domain there is no point in offering a python interface for the local domain we should just remove it comments comment from at metadata update from jhrozek issue set to the milestone sssd comment from at metadata update from jhrozek issue tagged with breaks compatibility comment from at pr comment from at metadata update from jhrozek issue priority set to minor comment from at pr part of comment from at metadata update from fidencio issue assigned to fidencio comment from at metadata update from fidencio issue tagged with pr comment from at fixed as part of comment from at metadata update from jhrozek issue close status updated to fixed issue status updated to closed was open comment from at unfortunately freeipa does use this interface and now we are broken ipa hbactest host cloud qe test service sshd user ipa error an internal error has occurred tail var log httpd error log ipa debug destroyed connection context failed to set perms on file var run ipa ccaches admin test referer ipa debug wsgi wsgi dispatch call ipa debug wsgi jsonserver session call ipa debug created connection context ipa debug wsgi jsonserver call ipa debug wsgi wsgiexecutioner call ipa debug raw hbactest user targethost cloud qe test service sshd version ipa debug hbactest user targethost cloud qe test service sshd nodetail false enabled false disabled false version ipa debug raw hbacrule find none sizelimit none version no members false ipa debug hbacrule find none sizelimit none all false raw false version no members false pkey only false ipa debug retrieving schema for schemacache url ldapi test socket conn ipa error non public attributeerror module pysss has no attribute getgrouplist traceback most recent call last file usr lib site packages ipaserver rpcserver py line in wsgi execute result command args options file usr lib site packages ipalib frontend py line in call return self do call args options file usr lib site packages ipalib frontend py line in do call ret self run args options file usr lib site packages ipalib frontend py line in run return self execute args options file usr lib site packages ipaserver plugins hbactest py line in execute user sid group sids domain validator get trusted domain user and groups options file usr lib site packages ipaserver dcerpc py line in get trusted domain user and groups group list pysss getgrouplist object name attributeerror module pysss has no attribute getgrouplist ipa info admin test hbactest user targethost cloud qe test service sshd version internalerror ipa debug destroyed connection context comment from at metadata update from jhrozek custom field rhbz adjusted to comment from at issue linked to bugzilla comment from at master this re adds the getgroupslist api comment from at metadata update from jhrozek issue set to the milestone none was sssd | 1 |
6,652 | 8,948,161,267 | IssuesEvent | 2019-01-25 01:05:25 | SpongePowered/SpongeForge | https://api.github.com/repos/SpongePowered/SpongeForge | closed | Mowzies Mobs Structures not spawning + Mobs not dropping items | system: world gen type: mod incompatibility version: 1.12 | **I am currently running**
<!-- If you don't use the latest version, please tell us why. -->
- SpongeForge version: 1.12.2-2768-7.1.5-RC3511
- Forge version: 1.12.2 - 14.23.5.2796
- Java version: Version 8 Update 191 [64 bit]
- Operating System: Windows 10 10.0.17134 Build 17134
<!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of
your mods and plugins by running "/sponge plugins" and/or "/sponge mods" -->
- Plugins/Mods: Minecraft, Minecraft Coder Pack, SpongeAPI, Sponge, Buycraft, CatClearLag, GriefPrevention, LuckPerms, Nucleus, MMCPrefix.
(Please note; even disabling all plugins- it still affects Mowzies Mobs.)
<!-- Please include as much information as possible. For the description, assume we have no idea how
mods work, be as detailed as possible and include a step by step reproduction. It is recommended
you try to reproduce the issue you are having yourself with as few mods as possible. -->
**Issue Description**
In a normal singleplayer/server without spongeforge, structures mowzies-mobs related like where the Sun Chief spawns. Are not generated. (this excludes Ferrous Wroughtnaught structures as they are underground)
As well as this; Sun Chiefs don't actually spawn, however, Frostmaws/Foliaaths do.
SpongeForge also affects drops on these mobs; eg:
Killing a frostmaw is meant to drop a Ice Crystal, this never happens in spongeforge servers. But occurs in servers without Sponge.
Killing a Ferrous Wroughtnaut is meant to drop a Axe of a Thousand Metals or a Wrought Helm, this never occurs once again, in sponge forge servers, but occurs in normal servers without sponge.
This also happens with the tribesmen, they never drop the masks of misery, etc.
This issue overall is related to ALL Mowzies-Mobs drops; such as the Helm and Masks, Ice Crystals, etc.
and the generation of *above-ground* structures within Mowzies-Mobs
<!-- Please provide a *full* server log (and crash-report if applicable).
Go to https://gist.github.com/ and upload them there, then paste the resulting link here!
Don't use hastebin/pastebin or other similar sites, as they have a history of quickly
deleting files before we can look at them. -->
Our server log does not provide any information about not spawning.
However; if a server log is absolutely needed. I can provide one.
| True | Mowzies Mobs Structures not spawning + Mobs not dropping items - **I am currently running**
<!-- If you don't use the latest version, please tell us why. -->
- SpongeForge version: 1.12.2-2768-7.1.5-RC3511
- Forge version: 1.12.2 - 14.23.5.2796
- Java version: Version 8 Update 191 [64 bit]
- Operating System: Windows 10 10.0.17134 Build 17134
<!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of
your mods and plugins by running "/sponge plugins" and/or "/sponge mods" -->
- Plugins/Mods: Minecraft, Minecraft Coder Pack, SpongeAPI, Sponge, Buycraft, CatClearLag, GriefPrevention, LuckPerms, Nucleus, MMCPrefix.
(Please note; even disabling all plugins- it still affects Mowzies Mobs.)
<!-- Please include as much information as possible. For the description, assume we have no idea how
mods work, be as detailed as possible and include a step by step reproduction. It is recommended
you try to reproduce the issue you are having yourself with as few mods as possible. -->
**Issue Description**
In a normal singleplayer/server without spongeforge, structures mowzies-mobs related like where the Sun Chief spawns. Are not generated. (this excludes Ferrous Wroughtnaught structures as they are underground)
As well as this; Sun Chiefs don't actually spawn, however, Frostmaws/Foliaaths do.
SpongeForge also affects drops on these mobs; eg:
Killing a frostmaw is meant to drop a Ice Crystal, this never happens in spongeforge servers. But occurs in servers without Sponge.
Killing a Ferrous Wroughtnaut is meant to drop a Axe of a Thousand Metals or a Wrought Helm, this never occurs once again, in sponge forge servers, but occurs in normal servers without sponge.
This also happens with the tribesmen, they never drop the masks of misery, etc.
This issue overall is related to ALL Mowzies-Mobs drops; such as the Helm and Masks, Ice Crystals, etc.
and the generation of *above-ground* structures within Mowzies-Mobs
<!-- Please provide a *full* server log (and crash-report if applicable).
Go to https://gist.github.com/ and upload them there, then paste the resulting link here!
Don't use hastebin/pastebin or other similar sites, as they have a history of quickly
deleting files before we can look at them. -->
Our server log does not provide any information about not spawning.
However; if a server log is absolutely needed. I can provide one.
| comp | mowzies mobs structures not spawning mobs not dropping items i am currently running spongeforge version forge version java version version update operating system windows build please include all mods plugins you had installed when your issue happened you can get a list of your mods and plugins by running sponge plugins and or sponge mods plugins mods minecraft minecraft coder pack spongeapi sponge buycraft catclearlag griefprevention luckperms nucleus mmcprefix please note even disabling all plugins it still affects mowzies mobs please include as much information as possible for the description assume we have no idea how mods work be as detailed as possible and include a step by step reproduction it is recommended you try to reproduce the issue you are having yourself with as few mods as possible issue description in a normal singleplayer server without spongeforge structures mowzies mobs related like where the sun chief spawns are not generated this excludes ferrous wroughtnaught structures as they are underground as well as this sun chiefs don t actually spawn however frostmaws foliaaths do spongeforge also affects drops on these mobs eg killing a frostmaw is meant to drop a ice crystal this never happens in spongeforge servers but occurs in servers without sponge killing a ferrous wroughtnaut is meant to drop a axe of a thousand metals or a wrought helm this never occurs once again in sponge forge servers but occurs in normal servers without sponge this also happens with the tribesmen they never drop the masks of misery etc this issue overall is related to all mowzies mobs drops such as the helm and masks ice crystals etc and the generation of above ground structures within mowzies mobs please provide a full server log and crash report if applicable go to and upload them there then paste the resulting link here don t use hastebin pastebin or other similar sites as they have a history of quickly deleting files before we can look at them our server log does not provide any information about not spawning however if a server log is absolutely needed i can provide one | 1 |
141,138 | 21,417,081,755 | IssuesEvent | 2022-04-22 12:01:25 | dpgraham4401/haztrak | https://api.github.com/repos/dpgraham4401/haztrak | opened | Migrate to bootstrap 5 | help wanted design decision | Migrate from the Bootstrap Framework version 4.6.1 to 5.x.x as the title says. This would be nice to get done before the front end becomes more than a placeholder. Right now (4/15/2022), I don't think migrating would be that difficult because we're not heavily using any bootstrap features or javascript.
There is a little JQuery being used but it's minimal, located in the [static content directory](https://github.com/dpgraham4401/haztrak/tree/main/static).
haztrak's frontend starting point was [SBadmin2](https://startbootstrap.com/theme/sb-admin-2)
[Bootstrap's documentation on migration to version 5](https://getbootstrap.com/docs/5.0/migration/) | 1.0 | Migrate to bootstrap 5 - Migrate from the Bootstrap Framework version 4.6.1 to 5.x.x as the title says. This would be nice to get done before the front end becomes more than a placeholder. Right now (4/15/2022), I don't think migrating would be that difficult because we're not heavily using any bootstrap features or javascript.
There is a little JQuery being used but it's minimal, located in the [static content directory](https://github.com/dpgraham4401/haztrak/tree/main/static).
haztrak's frontend starting point was [SBadmin2](https://startbootstrap.com/theme/sb-admin-2)
[Bootstrap's documentation on migration to version 5](https://getbootstrap.com/docs/5.0/migration/) | non_comp | migrate to bootstrap migrate from the bootstrap framework version to x x as the title says this would be nice to get done before the front end becomes more than a placeholder right now i don t think migrating would be that difficult because we re not heavily using any bootstrap features or javascript there is a little jquery being used but it s minimal located in the haztrak s frontend starting point was | 0 |
5,657 | 5,894,742,840 | IssuesEvent | 2017-05-18 03:32:20 | webcompat/webcompat.com | https://api.github.com/repos/webcompat/webcompat.com | closed | Add CSP policy | lang: Python scope: security status: ready | Probably something like:
```
Content-Security-Policy: default-src 'self'; img-src *; script-src 'self' www.google-analytics.com;
```
This allows images from anywhere, Google Analytics scripts to run, and everything else must come from webcompat.com.
We need to probably fix a few things so we don't have to enable `'unsafe-inline'` (there are a few inline scripts in templates, but we could just as easily put them in external scripts).
We might just deploy report-only mode: http://www.w3.org/TR/CSP/#content-security-policy-report-only-header-field and let that run for a few weeks to see if we would break anything with that policy. We'd just need to make a POST endpoint to log errors.
| True | Add CSP policy - Probably something like:
```
Content-Security-Policy: default-src 'self'; img-src *; script-src 'self' www.google-analytics.com;
```
This allows images from anywhere, Google Analytics scripts to run, and everything else must come from webcompat.com.
We need to probably fix a few things so we don't have to enable `'unsafe-inline'` (there are a few inline scripts in templates, but we could just as easily put them in external scripts).
We might just deploy report-only mode: http://www.w3.org/TR/CSP/#content-security-policy-report-only-header-field and let that run for a few weeks to see if we would break anything with that policy. We'd just need to make a POST endpoint to log errors.
| non_comp | add csp policy probably something like content security policy default src self img src script src self this allows images from anywhere google analytics scripts to run and everything else must come from webcompat com we need to probably fix a few things so we don t have to enable unsafe inline there are a few inline scripts in templates but we could just as easily put them in external scripts we might just deploy report only mode and let that run for a few weeks to see if we would break anything with that policy we d just need to make a post endpoint to log errors | 0 |
552,410 | 16,240,316,076 | IssuesEvent | 2021-05-07 08:44:31 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Small vertical aligment issue in detail pages with different badges | component: ux contrib: assigned priority: p3 | ### Describe the problem and steps to reproduce it:
1. Go to an add-on detail page with multiple "badges" like https://addons-dev.allizom.org/en-US/firefox/addon/webmail-ad-regression2/
### What happened?
There is more vertical space between the two last badges than between the two first ones, creating an unbalanced look:
<img width="390" alt="Screen Shot 2021-05-06 at 22 10 00" src="https://user-images.githubusercontent.com/217628/117359633-f524a600-aeb7-11eb-9de4-8d2098132fd5.png">
### What did you expect to happen?
I think the badges should be vertically aligned with the same space between each other. | 1.0 | Small vertical aligment issue in detail pages with different badges - ### Describe the problem and steps to reproduce it:
1. Go to an add-on detail page with multiple "badges" like https://addons-dev.allizom.org/en-US/firefox/addon/webmail-ad-regression2/
### What happened?
There is more vertical space between the two last badges than between the two first ones, creating an unbalanced look:
<img width="390" alt="Screen Shot 2021-05-06 at 22 10 00" src="https://user-images.githubusercontent.com/217628/117359633-f524a600-aeb7-11eb-9de4-8d2098132fd5.png">
### What did you expect to happen?
I think the badges should be vertically aligned with the same space between each other. | non_comp | small vertical aligment issue in detail pages with different badges describe the problem and steps to reproduce it go to an add on detail page with multiple badges like what happened there is more vertical space between the two last badges than between the two first ones creating an unbalanced look img width alt screen shot at src what did you expect to happen i think the badges should be vertically aligned with the same space between each other | 0 |
21,796 | 10,679,710,722 | IssuesEvent | 2019-10-21 19:50:40 | codice/ddf | https://api.github.com/repos/codice/ddf | closed | Add a second layer of protection for XSS reflection attack in GeoCoderEndpoint | :lock: Security | <!--
Have you read DDF's Code of Conduct? By filing an Issue, you are
expected to comply with it, including treating everyone with respect:
https://github.com/codice/ddf/blob/master/.github/CODE_OF_CONDUCT.md
Do you want to ask a question? Are you looking for support? The DDF
Developers group - https://groups.google.com/forum/#!forum/ddf-developers
is the best place for getting support.
-->
### Description
`GeoCoderEndpoint` is currently protected from XSS Reflection attacks by an underlying regex check for valid input for the name attached to a geo query. While the regex should be able to catch any issues, adding html sanitization on the provided string before it is returned is an extra layer of protection.
#### Expected behavior (if applicable):
`GeoCoderEndpoint` will ensure a sanitized copy of the input `jsonp` parameter should one somehow elude underlying regex checks.
### Version
2.16x, 2.17x, 2.18x, 2.19x
### Additional Information
<!-- Any additional information, or data that might be useful to know about -->
| True | Add a second layer of protection for XSS reflection attack in GeoCoderEndpoint - <!--
Have you read DDF's Code of Conduct? By filing an Issue, you are
expected to comply with it, including treating everyone with respect:
https://github.com/codice/ddf/blob/master/.github/CODE_OF_CONDUCT.md
Do you want to ask a question? Are you looking for support? The DDF
Developers group - https://groups.google.com/forum/#!forum/ddf-developers
is the best place for getting support.
-->
### Description
`GeoCoderEndpoint` is currently protected from XSS Reflection attacks by an underlying regex check for valid input for the name attached to a geo query. While the regex should be able to catch any issues, adding html sanitization on the provided string before it is returned is an extra layer of protection.
#### Expected behavior (if applicable):
`GeoCoderEndpoint` will ensure a sanitized copy of the input `jsonp` parameter should one somehow elude underlying regex checks.
### Version
2.16x, 2.17x, 2.18x, 2.19x
### Additional Information
<!-- Any additional information, or data that might be useful to know about -->
| non_comp | add a second layer of protection for xss reflection attack in geocoderendpoint have you read ddf s code of conduct by filing an issue you are expected to comply with it including treating everyone with respect do you want to ask a question are you looking for support the ddf developers group is the best place for getting support description geocoderendpoint is currently protected from xss reflection attacks by an underlying regex check for valid input for the name attached to a geo query while the regex should be able to catch any issues adding html sanitization on the provided string before it is returned is an extra layer of protection expected behavior if applicable geocoderendpoint will ensure a sanitized copy of the input jsonp parameter should one somehow elude underlying regex checks version additional information | 0 |
373,070 | 11,032,893,194 | IssuesEvent | 2019-12-06 21:22:49 | chef/automate | https://api.github.com/repos/chef/automate | closed | Add data to client run node detail pages + ui consistency | HIGH PRIORITY automate-ui client runs iamv2 ui | ## Overview
Add _Chef Organization_, _Chef Server_, _Chef Roles_, _Chef Tags_, and _Projects_ to client run node detail pages and some ui consistency work.
## Design Details
### General
- Remove _Node_ from the heading text
- Remove the subheading
### Visualization Changes
- Change the radial chart to be 175px instead of 200
- Change the label and percentage to use `chef-primary-dark`
### Data
#### Data Changes
- Add _Chef Organization_, _Chef Server_, _Chef Roles_, _Chef Tags_, and _Projects_
- Change _Deprecations_ to _Cookbook Deprecations_
- Change _Uptiime_ to _Node Up Time_
- If a node has zero _Cookbook Deprecations_ display _None_ instead of zero
- For _Chef Tags_, _Chef Roles_, and _Projects_ if a node has more than 1, then display the count instead of the names of those things (e.g., 2 Chef tags, or 3 projects, or 17 Chef roles)
- if a client run node does not have an FQDN, IP address, chef org, chef server, chef roles, chef tags, policy group, policy name, policy revision then those pieces of data along with their labels should be hidden
- _Projects_ should only be displayed if the iam version is 2.1
#### Data Structure
- sections, _Run Information_, _Node Information_, and _Metadata_
- use `text-transform` for the uppercase headings
##### Run Information
- _Run Duration_
- _Run ID_
- radial chart
##### Node Information
- _Chef Client Version_
- _FQDN_
- _IP Address_
- _Node ID_
- _Node Up Time_
- _Platform_
##### Metadata
- _Chef Organization_
- _Chef Roles_
- _Chef Server_
- _Chef Tags_
- _Cookbook Deprecations_
- _Enviornment_
- _Policy Group_
- _Policy Name_
- _Policy Revision_
- _Projects_
### Status Message
#### Status Content
- left align the content
- change content to read _Run failed <datetime>: <error>_ or _Run succeeded <datatime>. ... removing the rest of the filler words.
#### Accordion Behaviour
- status message should be transformed into an accordion, interacting anywhere on the status message should toggle the accordions state
- default when visiting a node detail page should be that the accordion is open
- when the accordion is open display _View less -_ on the right, when closed display _View more +_
- should be responsive, if there is room for 3 columns display then, otherwise display 2 columns, otherwise, display 1 column
- if the screen size dictates a single column, the accordion should be open and the closing behvaiour should be hidden
### Changes to error log button
- change label to _Error Log_
- base styles `color: #fff; border: #fff 1px solid; background-color: transparent;`
- hover styles `background-color: rgba(255, 255, 255, 0.25);` .. or similar

### Styles
- column headings `font-size: 15px; text-transform: uppercase;`
- data labels `font-weight: 600;`
- work with @susanev on space between data/labels and columns
- when columns wrap down, they should get at least 10px of top margin added
- add border around the entire accordion that uses the status color
### Screenshot
#### 3-col

#### 2-col

#### 1-col

| 1.0 | Add data to client run node detail pages + ui consistency - ## Overview
Add _Chef Organization_, _Chef Server_, _Chef Roles_, _Chef Tags_, and _Projects_ to client run node detail pages and some ui consistency work.
## Design Details
### General
- Remove _Node_ from the heading text
- Remove the subheading
### Visualization Changes
- Change the radial chart to be 175px instead of 200
- Change the label and percentage to use `chef-primary-dark`
### Data
#### Data Changes
- Add _Chef Organization_, _Chef Server_, _Chef Roles_, _Chef Tags_, and _Projects_
- Change _Deprecations_ to _Cookbook Deprecations_
- Change _Uptiime_ to _Node Up Time_
- If a node has zero _Cookbook Deprecations_ display _None_ instead of zero
- For _Chef Tags_, _Chef Roles_, and _Projects_ if a node has more than 1, then display the count instead of the names of those things (e.g., 2 Chef tags, or 3 projects, or 17 Chef roles)
- if a client run node does not have an FQDN, IP address, chef org, chef server, chef roles, chef tags, policy group, policy name, policy revision then those pieces of data along with their labels should be hidden
- _Projects_ should only be displayed if the iam version is 2.1
#### Data Structure
- sections, _Run Information_, _Node Information_, and _Metadata_
- use `text-transform` for the uppercase headings
##### Run Information
- _Run Duration_
- _Run ID_
- radial chart
##### Node Information
- _Chef Client Version_
- _FQDN_
- _IP Address_
- _Node ID_
- _Node Up Time_
- _Platform_
##### Metadata
- _Chef Organization_
- _Chef Roles_
- _Chef Server_
- _Chef Tags_
- _Cookbook Deprecations_
- _Enviornment_
- _Policy Group_
- _Policy Name_
- _Policy Revision_
- _Projects_
### Status Message
#### Status Content
- left align the content
- change content to read _Run failed <datetime>: <error>_ or _Run succeeded <datatime>. ... removing the rest of the filler words.
#### Accordion Behaviour
- status message should be transformed into an accordion, interacting anywhere on the status message should toggle the accordions state
- default when visiting a node detail page should be that the accordion is open
- when the accordion is open display _View less -_ on the right, when closed display _View more +_
- should be responsive, if there is room for 3 columns display then, otherwise display 2 columns, otherwise, display 1 column
- if the screen size dictates a single column, the accordion should be open and the closing behvaiour should be hidden
### Changes to error log button
- change label to _Error Log_
- base styles `color: #fff; border: #fff 1px solid; background-color: transparent;`
- hover styles `background-color: rgba(255, 255, 255, 0.25);` .. or similar

### Styles
- column headings `font-size: 15px; text-transform: uppercase;`
- data labels `font-weight: 600;`
- work with @susanev on space between data/labels and columns
- when columns wrap down, they should get at least 10px of top margin added
- add border around the entire accordion that uses the status color
### Screenshot
#### 3-col

#### 2-col

#### 1-col

| non_comp | add data to client run node detail pages ui consistency overview add chef organization chef server chef roles chef tags and projects to client run node detail pages and some ui consistency work design details general remove node from the heading text remove the subheading visualization changes change the radial chart to be instead of change the label and percentage to use chef primary dark data data changes add chef organization chef server chef roles chef tags and projects change deprecations to cookbook deprecations change uptiime to node up time if a node has zero cookbook deprecations display none instead of zero for chef tags chef roles and projects if a node has more than then display the count instead of the names of those things e g chef tags or projects or chef roles if a client run node does not have an fqdn ip address chef org chef server chef roles chef tags policy group policy name policy revision then those pieces of data along with their labels should be hidden projects should only be displayed if the iam version is data structure sections run information node information and metadata use text transform for the uppercase headings run information run duration run id radial chart node information chef client version fqdn ip address node id node up time platform metadata chef organization chef roles chef server chef tags cookbook deprecations enviornment policy group policy name policy revision projects status message status content left align the content change content to read run failed or run succeeded removing the rest of the filler words accordion behaviour status message should be transformed into an accordion interacting anywhere on the status message should toggle the accordions state default when visiting a node detail page should be that the accordion is open when the accordion is open display view less on the right when closed display view more should be responsive if there is room for columns display then otherwise display columns otherwise display column if the screen size dictates a single column the accordion should be open and the closing behvaiour should be hidden changes to error log button change label to error log base styles color fff border fff solid background color transparent hover styles background color rgba or similar styles column headings font size text transform uppercase data labels font weight work with susanev on space between data labels and columns when columns wrap down they should get at least of top margin added add border around the entire accordion that uses the status color screenshot col col col | 0 |
10,847 | 12,829,149,417 | IssuesEvent | 2020-07-06 22:07:46 | simpeg/simpeg | https://api.github.com/repos/simpeg/simpeg | closed | the problem with "Problem" | breaks backward compatibility community framework-bug | I had a chat with @dougoldenburg today and he brought to light some of the confusion introduced by the term `Problem` in the framework, in particular when first introducing the SimPEG framework to new community members, which is perhaps indication of a more systemic issue.

## Points of confusion
- often a first association with the word "Problem" is the thing you are trying to solve: the application, the reason you are considering inversion in the first place.
- "Problem" is over-used already. "Forward Problem", "Inverse Problem"
- even in the paper, we use "problem" to mean two different things \cite{SimPEGpaper}

## Potential Alternatives
- `Physics`
- `GoverningEquations`
- `PartialDifferentialEquations`
- `PDE`
- `Equations`
In looking at the current `Problem.py` and considering the construction of a forward simulation, I am (first pass!) partial to `Equations`.
- `BaseProblem` --> `BaseEquations`
- `BaseTimeProblem` --> `BaseTimeEquations`
- perhaps `Problem.py` --> `GoverningEquations.py`
Then in a module, if we consider EM
- `BaseEMequations`
- `BaseFDEMequations`
- `Maxwell3D_e(BaseFDEMequations)`
```
from SimPEG.EM import FDEM
equations = FDEM.Maxwell3D_e(mesh)
survey = FDEM.Survey(srcList)
equations.pair(survey) # or fwdSim = simulation(equations, survey) down the road
```
I look forward to hearing thoughts! | True | the problem with "Problem" - I had a chat with @dougoldenburg today and he brought to light some of the confusion introduced by the term `Problem` in the framework, in particular when first introducing the SimPEG framework to new community members, which is perhaps indication of a more systemic issue.

## Points of confusion
- often a first association with the word "Problem" is the thing you are trying to solve: the application, the reason you are considering inversion in the first place.
- "Problem" is over-used already. "Forward Problem", "Inverse Problem"
- even in the paper, we use "problem" to mean two different things \cite{SimPEGpaper}

## Potential Alternatives
- `Physics`
- `GoverningEquations`
- `PartialDifferentialEquations`
- `PDE`
- `Equations`
In looking at the current `Problem.py` and considering the construction of a forward simulation, I am (first pass!) partial to `Equations`.
- `BaseProblem` --> `BaseEquations`
- `BaseTimeProblem` --> `BaseTimeEquations`
- perhaps `Problem.py` --> `GoverningEquations.py`
Then in a module, if we consider EM
- `BaseEMequations`
- `BaseFDEMequations`
- `Maxwell3D_e(BaseFDEMequations)`
```
from SimPEG.EM import FDEM
equations = FDEM.Maxwell3D_e(mesh)
survey = FDEM.Survey(srcList)
equations.pair(survey) # or fwdSim = simulation(equations, survey) down the road
```
I look forward to hearing thoughts! | comp | the problem with problem i had a chat with dougoldenburg today and he brought to light some of the confusion introduced by the term problem in the framework in particular when first introducing the simpeg framework to new community members which is perhaps indication of a more systemic issue points of confusion often a first association with the word problem is the thing you are trying to solve the application the reason you are considering inversion in the first place problem is over used already forward problem inverse problem even in the paper we use problem to mean two different things cite simpegpaper potential alternatives physics governingequations partialdifferentialequations pde equations in looking at the current problem py and considering the construction of a forward simulation i am first pass partial to equations baseproblem baseequations basetimeproblem basetimeequations perhaps problem py governingequations py then in a module if we consider em baseemequations basefdemequations e basefdemequations from simpeg em import fdem equations fdem e mesh survey fdem survey srclist equations pair survey or fwdsim simulation equations survey down the road i look forward to hearing thoughts | 1 |
16,094 | 21,512,472,403 | IssuesEvent | 2022-04-28 06:41:37 | erosman/support | https://api.github.com/repos/erosman/support | closed | Google Search Sidebar UserCSS | 📜 userscript compatibility | UserCSS version of the [Google Search Sidebar](https://github.com/jmlntw/google-search-sidebar).
ref: [FireMonkey UserCSS version](https://github.com/jmlntw/google-search-sidebar/issues/32)
```css
/*
==UserCSS==
@name Google Search Sidebar
@namespace jmln.tw
@version 0.3.8
@description A user script and user style to move Google search tools to sidebar.
@author Jimmy Lin
@license MIT
@homepage https://github.com/jmlntw/google-search-sidebar
@supportURL https://github.com/jmlntw/google-search-sidebar/issues
@include https://www.google.*/search?*
@include https://www.google.*/webhp?*
@compatible FireMonkey on Firefox
==/UserCSS==
*/
/* CSS Variables
========================================================================== */
:root {
--user-sidebar-width: 200px;
--user-sidebar-spacer: 30px;
--user-sidebar-primary-color: #dd4b39;
--user-action-menu-spacer: 2px;
--user-action-menu-background: #eee;
--user-action-menu-font-size: 85%;
}
/* Search Tools Menu
========================================================================== */
/**
* Hide "Tools" toggle.
*/
#hdtb-tls {
display: none !important;
}
/**
* 1. Make menu container visible.
* 2. Set the sidebar width.
* 3. Remove menu container background.
*/
#hdtbMenus {
display: block !important; /* 1 */
position: absolute !important; /* 1 */
top: 68px !important; /* 1 */
width: var(--user-sidebar-width); /* 2 */
background: unset !important; /* 3 */
}
/**
* Remove original menu height.
*/
#hdtbMenus .hdtb-mn-cont {
height: 0 !important;
}
/**
* 1. Make dropdowns visible.
* 2. Add bottom space to each dropdown.
* 3. Remove dropdown background and shadow.
* 4. Wrap dropdowns text if too long.
*/
#hdtbMenus div.EwsJzb {
display: block !important; /* 1 */
position: static !important; /* 1 */
}
#hdtbMenus div.sAKBe {
box-shadow: none !important; /* 3 */
border-radius: 0 !important; /* 3 */
}
#hdtbMenus g-menu {
padding-top: 0 !important; /* 2 */
padding-bottom: var(--user-sidebar-spacer) !important; /* 2 */
background: unset !important; /* 3 */
}
#hdtbMenus div.znKVS {
white-space: normal !important; /* 4 */
}
/**
* Hide dropdown toggle by default.
*/
#hdtbMenus div.rIbAWc {
display: none !important;
}
/**
* 1. Remove checkmark on selected dropdown items.
* 2. Emphasize selected dropdown items.
*/
#hdtbMenus g-menu-item.nvELY {
background-image: none !important; /* 1 */
color: var(--user-sidebar-primary-color) !important; /* 2 */
font-weight: bolder !important; /* 2 */
}
/**
* Align the "Clear" button in the bottom of sidebar.
*/
#hdtbMenus #hdtb-rst {
display: block !important;
padding: 0 var(--user-sidebar-spacer) !important;
}
/**
* Make result status text ("About xxx,xxx results") visible.
*/
#extabar div.LHJvCe {
position: static !important;
opacity: 1 !important;
}
/* Main Content and Footer
========================================================================== */
/**
* Move main content and footer to the right.
*/
#rcnt,
#slim_appbar {
padding-inline-start: calc(
var(--user-sidebar-width) - 180px + var(--user-sidebar-spacer)
) !important;
}
/**
* Align the top carousel when searching movies.
*/
#kx {
margin-inline-start: calc(
var(--user-sidebar-width) + var(--user-sidebar-spacer)
) !important;
}
/**
* Align new material icons toolbar.
*/
#hdtb-msb-vis {
padding-inline-start: calc(
var(--user-sidebar-width) - 180px + var(--user-sidebar-spacer)
) !important;
}
/**
* Move "People also search for" carousel.
*/
#extabar div[role="navigation"] {
padding-inline-start: calc(
var(--user-sidebar-width) + var(--user-sidebar-spacer)
) !important;
}
#extabar div.LXqMce {
margin-left: 0 !important;
}
#extabar div.LXqMce:dir(rtl) {
margin-right: 0 !important;
}
#extabar div[role="list"] {
padding-left: 0 !important;
}
#extabar div[role="list"]:dir(rtl) {
padding-right: 0 !important;
}
/* Action Menu
========================================================================== */
/**
* Hide action menu toggle.
*/
.action-menu > a {
display: none !important;
}
/**
* 1. Make action menu visible.
* 2. Reset action menu position.
* 3. Remove decoration styles from action menu.
*/
.action-menu .action-menu-panel {
display: inline-block !important; /* 2 */
visibility: visible !important; /* 1 */
position: relative !important; /* 2 */
top: calc(var(--user-action-menu-spacer) * -1) !important; /* 2 */
border: 0 !important; /* 3 */
box-shadow: none !important; /* 3 */
background: transparent !important; /* 3 */
}
/**
* 1. Make action menu items display inline.
* 2. Reset menu items styles.
*/
.action-menu .action-menu-item {
display: inline-block !important; /* 1 */
margin: 0 var(--user-action-menu-spacer) !important; /* 2 */
background: var(--user-action-menu-background) !important; /* 2 */
}
/**
* Make action menu items smaller.
*/
.action-menu a.fl {
padding: var(--user-action-menu-spacer)
calc(var(--user-action-menu-spacer) * 2) !important;
font-size: var(--user-action-menu-font-size) !important;
}
/* Third-Party Compatibility
========================================================================== */
/**
* Align Evernote Similar Search block.
*/
#simSearchFrame {
margin-left: calc(-1 * (var(--user-sidebar-width) / 2)) !important;
}
/*
* Fix Google Search Region dropdown menu.
*/
.hdtb-mn-cont > .hide-focus-ring {
display: block !important;
width: var(--user-sidebar-width) !important;
min-width: var(--user-sidebar-width) !important;
max-width: var(--user-sidebar-width) !important;
}
ul.gm-region-menu-dropdown {
display: block;
position: static;
box-shadow: none;
}
``` | True | Google Search Sidebar UserCSS - UserCSS version of the [Google Search Sidebar](https://github.com/jmlntw/google-search-sidebar).
ref: [FireMonkey UserCSS version](https://github.com/jmlntw/google-search-sidebar/issues/32)
```css
/*
==UserCSS==
@name Google Search Sidebar
@namespace jmln.tw
@version 0.3.8
@description A user script and user style to move Google search tools to sidebar.
@author Jimmy Lin
@license MIT
@homepage https://github.com/jmlntw/google-search-sidebar
@supportURL https://github.com/jmlntw/google-search-sidebar/issues
@include https://www.google.*/search?*
@include https://www.google.*/webhp?*
@compatible FireMonkey on Firefox
==/UserCSS==
*/
/* CSS Variables
========================================================================== */
:root {
--user-sidebar-width: 200px;
--user-sidebar-spacer: 30px;
--user-sidebar-primary-color: #dd4b39;
--user-action-menu-spacer: 2px;
--user-action-menu-background: #eee;
--user-action-menu-font-size: 85%;
}
/* Search Tools Menu
========================================================================== */
/**
* Hide "Tools" toggle.
*/
#hdtb-tls {
display: none !important;
}
/**
* 1. Make menu container visible.
* 2. Set the sidebar width.
* 3. Remove menu container background.
*/
#hdtbMenus {
display: block !important; /* 1 */
position: absolute !important; /* 1 */
top: 68px !important; /* 1 */
width: var(--user-sidebar-width); /* 2 */
background: unset !important; /* 3 */
}
/**
* Remove original menu height.
*/
#hdtbMenus .hdtb-mn-cont {
height: 0 !important;
}
/**
* 1. Make dropdowns visible.
* 2. Add bottom space to each dropdown.
* 3. Remove dropdown background and shadow.
* 4. Wrap dropdowns text if too long.
*/
#hdtbMenus div.EwsJzb {
display: block !important; /* 1 */
position: static !important; /* 1 */
}
#hdtbMenus div.sAKBe {
box-shadow: none !important; /* 3 */
border-radius: 0 !important; /* 3 */
}
#hdtbMenus g-menu {
padding-top: 0 !important; /* 2 */
padding-bottom: var(--user-sidebar-spacer) !important; /* 2 */
background: unset !important; /* 3 */
}
#hdtbMenus div.znKVS {
white-space: normal !important; /* 4 */
}
/**
* Hide dropdown toggle by default.
*/
#hdtbMenus div.rIbAWc {
display: none !important;
}
/**
* 1. Remove checkmark on selected dropdown items.
* 2. Emphasize selected dropdown items.
*/
#hdtbMenus g-menu-item.nvELY {
background-image: none !important; /* 1 */
color: var(--user-sidebar-primary-color) !important; /* 2 */
font-weight: bolder !important; /* 2 */
}
/**
* Align the "Clear" button in the bottom of sidebar.
*/
#hdtbMenus #hdtb-rst {
display: block !important;
padding: 0 var(--user-sidebar-spacer) !important;
}
/**
* Make result status text ("About xxx,xxx results") visible.
*/
#extabar div.LHJvCe {
position: static !important;
opacity: 1 !important;
}
/* Main Content and Footer
========================================================================== */
/**
* Move main content and footer to the right.
*/
#rcnt,
#slim_appbar {
padding-inline-start: calc(
var(--user-sidebar-width) - 180px + var(--user-sidebar-spacer)
) !important;
}
/**
* Align the top carousel when searching movies.
*/
#kx {
margin-inline-start: calc(
var(--user-sidebar-width) + var(--user-sidebar-spacer)
) !important;
}
/**
* Align new material icons toolbar.
*/
#hdtb-msb-vis {
padding-inline-start: calc(
var(--user-sidebar-width) - 180px + var(--user-sidebar-spacer)
) !important;
}
/**
* Move "People also search for" carousel.
*/
#extabar div[role="navigation"] {
padding-inline-start: calc(
var(--user-sidebar-width) + var(--user-sidebar-spacer)
) !important;
}
#extabar div.LXqMce {
margin-left: 0 !important;
}
#extabar div.LXqMce:dir(rtl) {
margin-right: 0 !important;
}
#extabar div[role="list"] {
padding-left: 0 !important;
}
#extabar div[role="list"]:dir(rtl) {
padding-right: 0 !important;
}
/* Action Menu
========================================================================== */
/**
* Hide action menu toggle.
*/
.action-menu > a {
display: none !important;
}
/**
* 1. Make action menu visible.
* 2. Reset action menu position.
* 3. Remove decoration styles from action menu.
*/
.action-menu .action-menu-panel {
display: inline-block !important; /* 2 */
visibility: visible !important; /* 1 */
position: relative !important; /* 2 */
top: calc(var(--user-action-menu-spacer) * -1) !important; /* 2 */
border: 0 !important; /* 3 */
box-shadow: none !important; /* 3 */
background: transparent !important; /* 3 */
}
/**
* 1. Make action menu items display inline.
* 2. Reset menu items styles.
*/
.action-menu .action-menu-item {
display: inline-block !important; /* 1 */
margin: 0 var(--user-action-menu-spacer) !important; /* 2 */
background: var(--user-action-menu-background) !important; /* 2 */
}
/**
* Make action menu items smaller.
*/
.action-menu a.fl {
padding: var(--user-action-menu-spacer)
calc(var(--user-action-menu-spacer) * 2) !important;
font-size: var(--user-action-menu-font-size) !important;
}
/* Third-Party Compatibility
========================================================================== */
/**
* Align Evernote Similar Search block.
*/
#simSearchFrame {
margin-left: calc(-1 * (var(--user-sidebar-width) / 2)) !important;
}
/*
* Fix Google Search Region dropdown menu.
*/
.hdtb-mn-cont > .hide-focus-ring {
display: block !important;
width: var(--user-sidebar-width) !important;
min-width: var(--user-sidebar-width) !important;
max-width: var(--user-sidebar-width) !important;
}
ul.gm-region-menu-dropdown {
display: block;
position: static;
box-shadow: none;
}
``` | comp | google search sidebar usercss usercss version of the ref css usercss name google search sidebar namespace jmln tw version description a user script and user style to move google search tools to sidebar author jimmy lin license mit homepage supporturl include include compatible firemonkey on firefox usercss css variables root user sidebar width user sidebar spacer user sidebar primary color user action menu spacer user action menu background eee user action menu font size search tools menu hide tools toggle hdtb tls display none important make menu container visible set the sidebar width remove menu container background hdtbmenus display block important position absolute important top important width var user sidebar width background unset important remove original menu height hdtbmenus hdtb mn cont height important make dropdowns visible add bottom space to each dropdown remove dropdown background and shadow wrap dropdowns text if too long hdtbmenus div ewsjzb display block important position static important hdtbmenus div sakbe box shadow none important border radius important hdtbmenus g menu padding top important padding bottom var user sidebar spacer important background unset important hdtbmenus div znkvs white space normal important hide dropdown toggle by default hdtbmenus div ribawc display none important remove checkmark on selected dropdown items emphasize selected dropdown items hdtbmenus g menu item nvely background image none important color var user sidebar primary color important font weight bolder important align the clear button in the bottom of sidebar hdtbmenus hdtb rst display block important padding var user sidebar spacer important make result status text about xxx xxx results visible extabar div lhjvce position static important opacity important main content and footer move main content and footer to the right rcnt slim appbar padding inline start calc var user sidebar width var user sidebar spacer important align the top carousel when searching movies kx margin inline start calc var user sidebar width var user sidebar spacer important align new material icons toolbar hdtb msb vis padding inline start calc var user sidebar width var user sidebar spacer important move people also search for carousel extabar div padding inline start calc var user sidebar width var user sidebar spacer important extabar div lxqmce margin left important extabar div lxqmce dir rtl margin right important extabar div padding left important extabar div dir rtl padding right important action menu hide action menu toggle action menu a display none important make action menu visible reset action menu position remove decoration styles from action menu action menu action menu panel display inline block important visibility visible important position relative important top calc var user action menu spacer important border important box shadow none important background transparent important make action menu items display inline reset menu items styles action menu action menu item display inline block important margin var user action menu spacer important background var user action menu background important make action menu items smaller action menu a fl padding var user action menu spacer calc var user action menu spacer important font size var user action menu font size important third party compatibility align evernote similar search block simsearchframe margin left calc var user sidebar width important fix google search region dropdown menu hdtb mn cont hide focus ring display block important width var user sidebar width important min width var user sidebar width important max width var user sidebar width important ul gm region menu dropdown display block position static box shadow none | 1 |
643,466 | 20,957,723,137 | IssuesEvent | 2022-03-27 10:36:23 | AY2122S2-CS2103T-W12-2/tp | https://api.github.com/repos/AY2122S2-CS2103T-W12-2/tp | closed | Enchance the profiling of each person | priority.HIGH | Slightly modification to the current `Person `object needs to be done in order to cater to the university context.
New attributes to be added : Major, Telegram ID, Matriculation Card | 1.0 | Enchance the profiling of each person - Slightly modification to the current `Person `object needs to be done in order to cater to the university context.
New attributes to be added : Major, Telegram ID, Matriculation Card | non_comp | enchance the profiling of each person slightly modification to the current person object needs to be done in order to cater to the university context new attributes to be added major telegram id matriculation card | 0 |
18,760 | 26,124,637,747 | IssuesEvent | 2022-12-28 16:46:31 | davidpcaldwell/slime | https://api.github.com/repos/davidpcaldwell/slime | closed | Get Nashorn embedding working on JDK 11 | compatibility | Right now, Nashorn is not detected as compatible in jsh/launcher/main.js on line 139, where it checks for the existence of a `getContext` function on the `jdk.nashorn.internal.runtime.Context` class. The method is there, but can't be seen by Nashorn for some reason. (This was also the root cause for #203.)
It appears that some combination of the introduction of JPMS and changes to Nashorn make the class, but not its members, visible to Nashorn. We may need to add some --add-opens or something to our command line, or possibly to use Java reflection APIs; these do succeed in finding what we need, but in traversing the stack, we were not able to get all of the invocations working on a first attempt (could not figure out how to invoke Nashorn APIs with a `ScriptObject` given a Nashorn object).
This also may be related to the issue represented by https://github.com/davidpcaldwell/slime/issues/203, which was a test that appeared not to pass under JDK 11 at one point. Not sure the status of that test.
Pay attention to jsh/launcher/api.html when fixing this; it currently claims that JDK 11 downloads Rhino 1.7R3 at runtime if not installed; this may not be true currently (may be later version), and if it is true, might change if we get Nashorn working.
May want to consider standalone Nashorn for this; see https://github.com/szegedi/nashorn/wiki/Using-Nashorn-with-different-Java-versions
See https://blog.codefx.org/java/java-11-migration-guide/#Illegal-Access-To-Internal-APIs for one set of resources on reflection.
Test case:
```
env JSH_ENGINE=nashorn JSH_LAUNCHER_BASH_DEBUG=1 JSH_LAUNCHER_DEBUG=true ./jsh.bash jsh/test/jsh-data.jsh.js
```
Slideshow from Oracle explaining Nashorn / JPMS: https://www.oracle.com/technetwork/java/jvmls2016-haupt-nashorn-3125551.pdf | True | Get Nashorn embedding working on JDK 11 - Right now, Nashorn is not detected as compatible in jsh/launcher/main.js on line 139, where it checks for the existence of a `getContext` function on the `jdk.nashorn.internal.runtime.Context` class. The method is there, but can't be seen by Nashorn for some reason. (This was also the root cause for #203.)
It appears that some combination of the introduction of JPMS and changes to Nashorn make the class, but not its members, visible to Nashorn. We may need to add some --add-opens or something to our command line, or possibly to use Java reflection APIs; these do succeed in finding what we need, but in traversing the stack, we were not able to get all of the invocations working on a first attempt (could not figure out how to invoke Nashorn APIs with a `ScriptObject` given a Nashorn object).
This also may be related to the issue represented by https://github.com/davidpcaldwell/slime/issues/203, which was a test that appeared not to pass under JDK 11 at one point. Not sure the status of that test.
Pay attention to jsh/launcher/api.html when fixing this; it currently claims that JDK 11 downloads Rhino 1.7R3 at runtime if not installed; this may not be true currently (may be later version), and if it is true, might change if we get Nashorn working.
May want to consider standalone Nashorn for this; see https://github.com/szegedi/nashorn/wiki/Using-Nashorn-with-different-Java-versions
See https://blog.codefx.org/java/java-11-migration-guide/#Illegal-Access-To-Internal-APIs for one set of resources on reflection.
Test case:
```
env JSH_ENGINE=nashorn JSH_LAUNCHER_BASH_DEBUG=1 JSH_LAUNCHER_DEBUG=true ./jsh.bash jsh/test/jsh-data.jsh.js
```
Slideshow from Oracle explaining Nashorn / JPMS: https://www.oracle.com/technetwork/java/jvmls2016-haupt-nashorn-3125551.pdf | comp | get nashorn embedding working on jdk right now nashorn is not detected as compatible in jsh launcher main js on line where it checks for the existence of a getcontext function on the jdk nashorn internal runtime context class the method is there but can t be seen by nashorn for some reason this was also the root cause for it appears that some combination of the introduction of jpms and changes to nashorn make the class but not its members visible to nashorn we may need to add some add opens or something to our command line or possibly to use java reflection apis these do succeed in finding what we need but in traversing the stack we were not able to get all of the invocations working on a first attempt could not figure out how to invoke nashorn apis with a scriptobject given a nashorn object this also may be related to the issue represented by which was a test that appeared not to pass under jdk at one point not sure the status of that test pay attention to jsh launcher api html when fixing this it currently claims that jdk downloads rhino at runtime if not installed this may not be true currently may be later version and if it is true might change if we get nashorn working may want to consider standalone nashorn for this see see for one set of resources on reflection test case env jsh engine nashorn jsh launcher bash debug jsh launcher debug true jsh bash jsh test jsh data jsh js slideshow from oracle explaining nashorn jpms | 1 |
2,863 | 5,694,330,386 | IssuesEvent | 2017-04-15 12:10:20 | derhansen/sf_event_mgt | https://api.github.com/repos/derhansen/sf_event_mgt | closed | [TASK] Remove fluidStandaloneService | 8.7 compatibility | Remove fluidStandaloneService and use native fluid standalone rendering | True | [TASK] Remove fluidStandaloneService - Remove fluidStandaloneService and use native fluid standalone rendering | comp | remove fluidstandaloneservice remove fluidstandaloneservice and use native fluid standalone rendering | 1 |
358,471 | 25,193,184,021 | IssuesEvent | 2022-11-12 06:42:27 | crushten/github_workflow_repo | https://api.github.com/repos/crushten/github_workflow_repo | closed | Centralized the workflows | documentation | Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects.
https://docs.github.com/en/actions/using-workflows/reusing-workflows
https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/
https://github.blog/2022-02-10-using-reusable-workflows-github-actions/
https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/
https://docs.github.com/en/actions/creating-actions/creating-a-composite-action
https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258
https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8 | 1.0 | Centralized the workflows - Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects.
https://docs.github.com/en/actions/using-workflows/reusing-workflows
https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/
https://github.blog/2022-02-10-using-reusable-workflows-github-actions/
https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/
https://docs.github.com/en/actions/creating-actions/creating-a-composite-action
https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258
https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8 | non_comp | centralized the workflows would like to centralized all workflows here will need to make some sort of structure and then can use that in other projects | 0 |
3,691 | 6,555,362,542 | IssuesEvent | 2017-09-06 09:58:42 | scylladb/scylla | https://api.github.com/repos/scylladb/scylla | closed | Support termination of repair jobs. | cassandra 2.2 compatibility enhancement nodetool | ```
$ curl -X POST 'http://localhost:10000/storage_service/force_terminate'
{"message": "API call is not supported yet", "code": 500}
```
```
ss::force_terminate_all_repair_sessions.set(r, [](std::unique_ptr<request> req) {
//TBD
unimplemented();
return make_ready_future<json::json_return_type>(json_void());
});
```
Please support termination of running repair jobs.
| True | Support termination of repair jobs. - ```
$ curl -X POST 'http://localhost:10000/storage_service/force_terminate'
{"message": "API call is not supported yet", "code": 500}
```
```
ss::force_terminate_all_repair_sessions.set(r, [](std::unique_ptr<request> req) {
//TBD
unimplemented();
return make_ready_future<json::json_return_type>(json_void());
});
```
Please support termination of running repair jobs.
| comp | support termination of repair jobs curl x post message api call is not supported yet code ss force terminate all repair sessions set r std unique ptr req tbd unimplemented return make ready future json void please support termination of running repair jobs | 1 |
6,365 | 8,682,479,375 | IssuesEvent | 2018-12-02 08:50:09 | andstatus/andstatus | https://api.github.com/repos/andstatus/andstatus | opened | "less-than sign" and "greater-than sign" character entity references in content are unnecessarily unescaped | Compatibility | As a user noticed in GNU Social, HTML-formatted Note, which has "less-than sign" and "greater-than sign" character entity references in content are not shown properly in AndStatus.
Via web interface the note looks like this:

but in AndStatus some part of the content is not shown:

This is cause by unescaping of "less-than sign" and "greater-than sign" character entity references during note saving in AndStatus.
Root cause of the problem: mediaType ( https://www.w3.org/TR/activitystreams-vocabulary/#dfn-mediatype ) of content, received by AndStatus from a server (from a Social Network), is undefined (in AndStatus). It could be either "text/plain" or "text/html".
Moreover, when a User creates a new Note ("tweet", "notice", "toot", ...), mediaType is unknown also, so AndStatus cannot process typed text correctly without knowing if the text has HTML markup or it has simply a text, which may look like HTML markup (like in the above case, see screenshots).
We have "Allow HTML content" attribute for each "Social Network", but it is mostly about processing of the content, not about its input format. And changing this attribute doesn't change notes, already stored in a local database... | True | "less-than sign" and "greater-than sign" character entity references in content are unnecessarily unescaped - As a user noticed in GNU Social, HTML-formatted Note, which has "less-than sign" and "greater-than sign" character entity references in content are not shown properly in AndStatus.
Via web interface the note looks like this:

but in AndStatus some part of the content is not shown:

This is cause by unescaping of "less-than sign" and "greater-than sign" character entity references during note saving in AndStatus.
Root cause of the problem: mediaType ( https://www.w3.org/TR/activitystreams-vocabulary/#dfn-mediatype ) of content, received by AndStatus from a server (from a Social Network), is undefined (in AndStatus). It could be either "text/plain" or "text/html".
Moreover, when a User creates a new Note ("tweet", "notice", "toot", ...), mediaType is unknown also, so AndStatus cannot process typed text correctly without knowing if the text has HTML markup or it has simply a text, which may look like HTML markup (like in the above case, see screenshots).
We have "Allow HTML content" attribute for each "Social Network", but it is mostly about processing of the content, not about its input format. And changing this attribute doesn't change notes, already stored in a local database... | comp | less than sign and greater than sign character entity references in content are unnecessarily unescaped as a user noticed in gnu social html formatted note which has less than sign and greater than sign character entity references in content are not shown properly in andstatus via web interface the note looks like this but in andstatus some part of the content is not shown this is cause by unescaping of less than sign and greater than sign character entity references during note saving in andstatus root cause of the problem mediatype of content received by andstatus from a server from a social network is undefined in andstatus it could be either text plain or text html moreover when a user creates a new note tweet notice toot mediatype is unknown also so andstatus cannot process typed text correctly without knowing if the text has html markup or it has simply a text which may look like html markup like in the above case see screenshots we have allow html content attribute for each social network but it is mostly about processing of the content not about its input format and changing this attribute doesn t change notes already stored in a local database | 1 |
44,195 | 2,900,073,044 | IssuesEvent | 2015-06-17 14:47:49 | icatproject/topcat | https://api.github.com/repos/icatproject/topcat | closed | Use an ICAT authenticator to provide access to the "admin interface" | invalid Priority-Low Type-Enhancement | ```
This should simplify the topcat code and would be convenient for the installer, especially
if he has an authenticator installed already as will often be the case.
```
Original issue reported on code.google.com by `dr.s.m.fisher` on 2013-10-10 08:10:35 | 1.0 | Use an ICAT authenticator to provide access to the "admin interface" - ```
This should simplify the topcat code and would be convenient for the installer, especially
if he has an authenticator installed already as will often be the case.
```
Original issue reported on code.google.com by `dr.s.m.fisher` on 2013-10-10 08:10:35 | non_comp | use an icat authenticator to provide access to the admin interface this should simplify the topcat code and would be convenient for the installer especially if he has an authenticator installed already as will often be the case original issue reported on code google com by dr s m fisher on | 0 |
13,068 | 8,788,310,591 | IssuesEvent | 2018-12-20 21:43:01 | idno/known | https://api.github.com/repos/idno/known | closed | Reflected XSS when deleting a post | Security | ## While trying to do this:
create a post and use the following as the title:
`Tester <script>alert(123)</script>`
## I encountered this error:
The original scripting is kept, and allows for Reflected XSS when deleting a post - so when deleting the script is executed - this allows for a security breach
* It'd also be really handy if you could tell us the contents of your ```version.known``` file
version = "0.9.9-a"
build = 2018120201
* What database are you using? (e.g. mongo, mysql, postgres)
mysql
| True | Reflected XSS when deleting a post - ## While trying to do this:
create a post and use the following as the title:
`Tester <script>alert(123)</script>`
## I encountered this error:
The original scripting is kept, and allows for Reflected XSS when deleting a post - so when deleting the script is executed - this allows for a security breach
* It'd also be really handy if you could tell us the contents of your ```version.known``` file
version = "0.9.9-a"
build = 2018120201
* What database are you using? (e.g. mongo, mysql, postgres)
mysql
| non_comp | reflected xss when deleting a post while trying to do this create a post and use the following as the title tester alert i encountered this error the original scripting is kept and allows for reflected xss when deleting a post so when deleting the script is executed this allows for a security breach it d also be really handy if you could tell us the contents of your version known file version a build what database are you using e g mongo mysql postgres mysql | 0 |
57,132 | 15,704,318,395 | IssuesEvent | 2021-03-26 14:52:12 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | 508-defect-2 [SCREENREADER, COGNITION]: Bold text SHOULD be made headings, in this instance | 508-defect-2 508-issue-headings 508/Accessibility cms content frontend sitewide vsa vsa-public-websites | # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
```diff
! Team affected: VSA Public Websites; Project: AMA Phase 2
```
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
Within the subway steps content, the text that is currently `<p> <strong>Words</strong> … </p>`, these **should** be headings, since the screen reader does not indicate they are headings, and it isn't clear when it is read out — it sounds like a sentence isn't finished.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to understand the hierarchy and content of the page content using heading levels to save time and aid comprehension.
## Environment
- Operating System: all
- Browser: all
- Screenreading device: any
- Server destination: staging & production
## Steps to Recreate
1. Enter [https://www.va.gov/decision-reviews/supplemental-claim/](https://www.va.gov/decision-reviews/supplemental-claim/) in browser
2. Start screenreading device of your choice
3. Navigate to the subway steps
4. Listen to how the content is read by the screen reader, without looking at the screen
5. Verify that it is confusing, sounds like sentences end abruptly.
## Possible Fixes (optional)
Recommend changing these to headings, of the appropriate level.
| 1.0 | 508-defect-2 [SCREENREADER, COGNITION]: Bold text SHOULD be made headings, in this instance - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
```diff
! Team affected: VSA Public Websites; Project: AMA Phase 2
```
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
Within the subway steps content, the text that is currently `<p> <strong>Words</strong> … </p>`, these **should** be headings, since the screen reader does not indicate they are headings, and it isn't clear when it is read out — it sounds like a sentence isn't finished.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to understand the hierarchy and content of the page content using heading levels to save time and aid comprehension.
## Environment
- Operating System: all
- Browser: all
- Screenreading device: any
- Server destination: staging & production
## Steps to Recreate
1. Enter [https://www.va.gov/decision-reviews/supplemental-claim/](https://www.va.gov/decision-reviews/supplemental-claim/) in browser
2. Start screenreading device of your choice
3. Navigate to the subway steps
4. Listen to how the content is read by the screen reader, without looking at the screen
5. Verify that it is confusing, sounds like sentences end abruptly.
## Possible Fixes (optional)
Recommend changing these to headings, of the appropriate level.
| non_comp | defect bold text should be made headings in this instance diff team affected vsa public websites project ama phase feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description within the subway steps content the text that is currently words … these should be headings since the screen reader does not indicate they are headings and it isn t clear when it is read out — it sounds like a sentence isn t finished point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to understand the hierarchy and content of the page content using heading levels to save time and aid comprehension environment operating system all browser all screenreading device any server destination staging production steps to recreate enter in browser start screenreading device of your choice navigate to the subway steps listen to how the content is read by the screen reader without looking at the screen verify that it is confusing sounds like sentences end abruptly possible fixes optional recommend changing these to headings of the appropriate level | 0 |
11,068 | 13,095,684,610 | IssuesEvent | 2020-08-03 14:30:08 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | 'added cards' statistic does not reflect the number of added cards. | Anki Ecosystem Compatibility Bug Help Wanted Keep Open Reproduced Statistics | ###### Research
*Enter an [ x ] character to confirm the points below:*
[ x ] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
[ x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
[ x ] I have searched for similar existing issues here and on the user forum
###### Reproduction Steps
1. Open ankidroid
2. go to a stack, choose statistics and scroll down to see 'added cards'
3. note the number
4. Add some cards to the stack
5. look at statistics/ added cards again - no change
6 synchronize the stack
7 look at statistics/ added cards again - no change
sometimes the number fluctuates somewhat, but is is never correct. currently it reports 1112 cards as added in android in the last month, and my PC version gives 1846 cards added in the last month.
###### Expected Result
###### Actual Result
###### Debug info
Refer to the [support page](https://ankidroid.org/docs/help.html) if you are unsure where to get the "debug info".
| True | 'added cards' statistic does not reflect the number of added cards. - ###### Research
*Enter an [ x ] character to confirm the points below:*
[ x ] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
[ x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
[ x ] I have searched for similar existing issues here and on the user forum
###### Reproduction Steps
1. Open ankidroid
2. go to a stack, choose statistics and scroll down to see 'added cards'
3. note the number
4. Add some cards to the stack
5. look at statistics/ added cards again - no change
6 synchronize the stack
7 look at statistics/ added cards again - no change
sometimes the number fluctuates somewhat, but is is never correct. currently it reports 1112 cards as added in android in the last month, and my PC version gives 1846 cards added in the last month.
###### Expected Result
###### Actual Result
###### Debug info
Refer to the [support page](https://ankidroid.org/docs/help.html) if you are unsure where to get the "debug info".
| comp | added cards statistic does not reflect the number of added cards research enter an character to confirm the points below i have read the and am reporting a bug or enhancement request specific to ankidroid i have checked the and the and could not find a solution to my issue i have searched for similar existing issues here and on the user forum reproduction steps open ankidroid go to a stack choose statistics and scroll down to see added cards note the number add some cards to the stack look at statistics added cards again no change synchronize the stack look at statistics added cards again no change sometimes the number fluctuates somewhat but is is never correct currently it reports cards as added in android in the last month and my pc version gives cards added in the last month expected result actual result debug info refer to the if you are unsure where to get the debug info | 1 |
4,576 | 7,204,332,933 | IssuesEvent | 2018-02-06 12:19:27 | foxbunny/duckweed | https://api.github.com/repos/foxbunny/duckweed | opened | Remove custom events | backwards incompatible enhancement | Custom events are generally more specific to how we build applications, and should not be in the framework. They may be provided as separate modules but not as part of the Duckweed package. | True | Remove custom events - Custom events are generally more specific to how we build applications, and should not be in the framework. They may be provided as separate modules but not as part of the Duckweed package. | comp | remove custom events custom events are generally more specific to how we build applications and should not be in the framework they may be provided as separate modules but not as part of the duckweed package | 1 |
17,676 | 24,368,978,165 | IssuesEvent | 2022-10-03 17:31:12 | kellnerd/musicbrainz-scripts | https://api.github.com/repos/kellnerd/musicbrainz-scripts | closed | Regular expression look behind not handled in Firefox 68 | wontfix punctuation compatibility | It will most probably be a won't fix because I am in a niche use-case, but as it took me much trouble to pin point why the unicode punctuation user script does not work in Firefox for Android 68 (the last version to support user scripts):
Regex lookbehind `(?<=` and negative lookbehind `(?<!` is [not supported before Firefox 78](https://caniuse.com/?search=lookbehind).
Some tr[ai]cks:
- https://stackoverflow.com/questions/7376238/javascript-regex-look-behind-alternative
- https://cmsdk.com/javascript/javascript-negative-lookbehind-equivalent.html | True | Regular expression look behind not handled in Firefox 68 - It will most probably be a won't fix because I am in a niche use-case, but as it took me much trouble to pin point why the unicode punctuation user script does not work in Firefox for Android 68 (the last version to support user scripts):
Regex lookbehind `(?<=` and negative lookbehind `(?<!` is [not supported before Firefox 78](https://caniuse.com/?search=lookbehind).
Some tr[ai]cks:
- https://stackoverflow.com/questions/7376238/javascript-regex-look-behind-alternative
- https://cmsdk.com/javascript/javascript-negative-lookbehind-equivalent.html | comp | regular expression look behind not handled in firefox it will most probably be a won t fix because i am in a niche use case but as it took me much trouble to pin point why the unicode punctuation user script does not work in firefox for android the last version to support user scripts regex lookbehind and negative lookbehind is some tr cks | 1 |
16,968 | 23,377,188,194 | IssuesEvent | 2022-08-11 05:19:48 | WPChill/modula-lite | https://api.github.com/repos/WPChill/modula-lite | opened | test all modulas with new version of elementor and elementor pro (3.7.3) | compatibility |
Hi, Since I have updated Elementor 3.7 & Elementor Pro 3.7.3, there is a conflict with all of the Modula Pro plugins I have setup. It is causing the Elementor page builder to be constantly on a grey loading screen, but when I deactivate your plugins it works again.
| True | test all modulas with new version of elementor and elementor pro (3.7.3) -
Hi, Since I have updated Elementor 3.7 & Elementor Pro 3.7.3, there is a conflict with all of the Modula Pro plugins I have setup. It is causing the Elementor page builder to be constantly on a grey loading screen, but when I deactivate your plugins it works again.
| comp | test all modulas with new version of elementor and elementor pro hi since i have updated elementor elementor pro there is a conflict with all of the modula pro plugins i have setup it is causing the elementor page builder to be constantly on a grey loading screen but when i deactivate your plugins it works again | 1 |
343,809 | 30,691,022,165 | IssuesEvent | 2023-07-26 15:11:10 | goalapa/spring-cacamuca | https://api.github.com/repos/goalapa/spring-cacamuca | opened | [test] 4주차 - [신고] 테스트 코드 검증 | 🎯test | ✏️Description
-
신고, 블랙리스트의 테스트 코드 검증
✅TODO
-
- [ ] Repository Service Tests
- [ ] Report Check Null Service Tests
- [ ] BlackList Service Tests
- [ ] Report Controller Test
- [ ] BlackList Controller Test
🐾ETC
-
| 1.0 | [test] 4주차 - [신고] 테스트 코드 검증 - ✏️Description
-
신고, 블랙리스트의 테스트 코드 검증
✅TODO
-
- [ ] Repository Service Tests
- [ ] Report Check Null Service Tests
- [ ] BlackList Service Tests
- [ ] Report Controller Test
- [ ] BlackList Controller Test
🐾ETC
-
| non_comp | 테스트 코드 검증 ✏️description 신고 블랙리스트의 테스트 코드 검증 ✅todo repository service tests report check null service tests blacklist service tests report controller test blacklist controller test 🐾etc | 0 |
5,436 | 7,894,576,628 | IssuesEvent | 2018-06-28 22:03:13 | presscustomizr/hueman | https://api.github.com/repos/presscustomizr/hueman | opened | Possible customizer issue with a Google Analytics plugin | bug plugin compatibility question | > I just today updated Hueman and Add-Ons. It disabled my homepage sidebar layout. On trying to open the Customizer it would not open and the only way I was able to get it to function was by disabling Google Analytics plugin. It then forced me to utilise the Social Links in order to be able to save settings in the Customizer.
https://wordpress.org/support/topic/latest-update-42/#post-10446106 | True | Possible customizer issue with a Google Analytics plugin - > I just today updated Hueman and Add-Ons. It disabled my homepage sidebar layout. On trying to open the Customizer it would not open and the only way I was able to get it to function was by disabling Google Analytics plugin. It then forced me to utilise the Social Links in order to be able to save settings in the Customizer.
https://wordpress.org/support/topic/latest-update-42/#post-10446106 | comp | possible customizer issue with a google analytics plugin i just today updated hueman and add ons it disabled my homepage sidebar layout on trying to open the customizer it would not open and the only way i was able to get it to function was by disabling google analytics plugin it then forced me to utilise the social links in order to be able to save settings in the customizer | 1 |
238,179 | 18,235,352,743 | IssuesEvent | 2021-10-01 05:57:44 | podverse/podverse-rn | https://api.github.com/repos/podverse/podverse-rn | opened | Help update the podverse-rn onboarding documentation (Windows) | good first issue hacktoberfest documentation | **Windows**
We need to update our documentation to make it easier for people to start contributing. It'd be a huge help if we can work with someone who doesn't already have podverse-rn running on their computer, so we can go from a clean slate to up-and-running with podverse-rn, and document the steps we take along the way.
If you are interested in contributing to Podverse, and have at least some experience setting up a JavaScript dev environment, this could be a great way to get started with the Podverse code base. I'm happy to do a screen share with you over Discord if you would like, to walk you through the setup process and update the documentation as we go. | 1.0 | Help update the podverse-rn onboarding documentation (Windows) - **Windows**
We need to update our documentation to make it easier for people to start contributing. It'd be a huge help if we can work with someone who doesn't already have podverse-rn running on their computer, so we can go from a clean slate to up-and-running with podverse-rn, and document the steps we take along the way.
If you are interested in contributing to Podverse, and have at least some experience setting up a JavaScript dev environment, this could be a great way to get started with the Podverse code base. I'm happy to do a screen share with you over Discord if you would like, to walk you through the setup process and update the documentation as we go. | non_comp | help update the podverse rn onboarding documentation windows windows we need to update our documentation to make it easier for people to start contributing it d be a huge help if we can work with someone who doesn t already have podverse rn running on their computer so we can go from a clean slate to up and running with podverse rn and document the steps we take along the way if you are interested in contributing to podverse and have at least some experience setting up a javascript dev environment this could be a great way to get started with the podverse code base i m happy to do a screen share with you over discord if you would like to walk you through the setup process and update the documentation as we go | 0 |
17,651 | 10,741,054,603 | IssuesEvent | 2019-10-29 19:26:03 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Converting the json output to excel format | Pri3 cognitive-services/svc computer-vision/subsvc cxp product-question triaged | @JJ @shanselman @matsuu @dlapiduz @bigongiari
Hey, I just want to convert the output JSON to excel format I just want only the "Text" field in the JSON.
what changes should I do in the code? can you please help me with this? Is there any provision in Azure to do that?
I am performing the table extraction and I just want to extract the text into excel.
please checkout the input image:

---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ab15e2d6-7f06-3037-d47d-2d76fb8456ac
* Version Independent ID: 396d1582-f04b-ad66-88ee-ae614ecd0ece
* Content: [Quickstart: Extract printed text - REST, Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/Computer-vision/QuickStarts/python-print-text)
* Content Source: [articles/cognitive-services/Computer-vision/QuickStarts/python-print-text.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/QuickStarts/python-print-text.md)
* Service: **cognitive-services**
* Sub-service: **computer-vision**
* GitHub Login: @PatrickFarley
* Microsoft Alias: **pafarley** | 1.0 | Converting the json output to excel format - @JJ @shanselman @matsuu @dlapiduz @bigongiari
Hey, I just want to convert the output JSON to excel format I just want only the "Text" field in the JSON.
what changes should I do in the code? can you please help me with this? Is there any provision in Azure to do that?
I am performing the table extraction and I just want to extract the text into excel.
please checkout the input image:

---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ab15e2d6-7f06-3037-d47d-2d76fb8456ac
* Version Independent ID: 396d1582-f04b-ad66-88ee-ae614ecd0ece
* Content: [Quickstart: Extract printed text - REST, Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/Computer-vision/QuickStarts/python-print-text)
* Content Source: [articles/cognitive-services/Computer-vision/QuickStarts/python-print-text.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/QuickStarts/python-print-text.md)
* Service: **cognitive-services**
* Sub-service: **computer-vision**
* GitHub Login: @PatrickFarley
* Microsoft Alias: **pafarley** | non_comp | converting the json output to excel format jj shanselman matsuu dlapiduz bigongiari hey i just want to convert the output json to excel format i just want only the text field in the json what changes should i do in the code can you please help me with this is there any provision in azure to do that i am performing the table extraction and i just want to extract the text into excel please checkout the input image document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service computer vision github login patrickfarley microsoft alias pafarley | 0 |
317,149 | 27,216,836,524 | IssuesEvent | 2023-02-20 23:02:30 | arfc/openmcyclus | https://api.github.com/repos/arfc/openmcyclus | closed | CI Environment needs Cyclus and OpenMC | Difficulty:2-Challenging Priority:1-Critical Status:4-In Progress Type:Test | The CI testing environment needs to have Cyclus and OpenMC installed to ensure that all tests can be run.
Observations on why this doesn't exist right now:
- Cyclus (installed from conda or source) needs python 3.6
- OpenMC when installed from source requires at least python 3.7
- OpenMC when installed from Mamba can use python 3.6, but if this is the case then it doesn't have the `deplete` method
- When installing locally, these issues do not arise because when installing cyclus from source, it gets installed into `~/.local/` instead of in the conda environment. Therefore, cyclus can be installed in one conda environment (with python 3.6) then used in another one, independent of the python version.
Next steps:
- Install Cyclus from source, then change the python version in the environment (previously tried but got an error that has since been fixed) | 1.0 | CI Environment needs Cyclus and OpenMC - The CI testing environment needs to have Cyclus and OpenMC installed to ensure that all tests can be run.
Observations on why this doesn't exist right now:
- Cyclus (installed from conda or source) needs python 3.6
- OpenMC when installed from source requires at least python 3.7
- OpenMC when installed from Mamba can use python 3.6, but if this is the case then it doesn't have the `deplete` method
- When installing locally, these issues do not arise because when installing cyclus from source, it gets installed into `~/.local/` instead of in the conda environment. Therefore, cyclus can be installed in one conda environment (with python 3.6) then used in another one, independent of the python version.
Next steps:
- Install Cyclus from source, then change the python version in the environment (previously tried but got an error that has since been fixed) | non_comp | ci environment needs cyclus and openmc the ci testing environment needs to have cyclus and openmc installed to ensure that all tests can be run observations on why this doesn t exist right now cyclus installed from conda or source needs python openmc when installed from source requires at least python openmc when installed from mamba can use python but if this is the case then it doesn t have the deplete method when installing locally these issues do not arise because when installing cyclus from source it gets installed into local instead of in the conda environment therefore cyclus can be installed in one conda environment with python then used in another one independent of the python version next steps install cyclus from source then change the python version in the environment previously tried but got an error that has since been fixed | 0 |
15,115 | 18,984,083,706 | IssuesEvent | 2021-11-21 12:15:33 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | Pedantic readInt brakes use case behavior | backward compatibility st-accepted | **Describe the issue**
Moving to the last LTS version (21.8.10.19) I discover that [#24475](https://github.com/ClickHouse/ClickHouse/pull/24475) break the behavior I rely on in a fairly big pipeline at my company (~1MM inserts per sec).
The events are Nginx access logs, where some non-existent values can be logged as a hyphen (`-`). When the correct value is expected as integers this is indeed a consistency fault. But also it is the way Nginx works and will not change anytime soon.
¯\\_(ツ)_/¯
At the current version I'm using (21.4.4.30) my log-forwarders can do INSERT using `-` for `Int8|16|32|64` columns without any problem, the value is stored as `0` as expected.
But now at the LTS version, the exception below arises. I'm using Kafka engine tables in the environment but, if you look at the examples to reproduce below the exception you'll see that this is not a Kafka Engine-related problem.
````
DB::ParsingException: Cannot parse number with a sign character but without any numeric character: Cannot parse Int32 from String
[Complete stack at the end]
````
**How to reproduce**
This will work just fine in any version before [#24475](https://github.com/ClickHouse/ClickHouse/pull/24475), after that the exception mentioned arises.
```
CREATE TABLE numbers
(
`number` Int32
)
ENGINE = MergeTree
ORDER BY number
INSERT INTO numbers VALUES ('-');
```
**Error message and/or stacktrace**
```
2021.11.13 09:56:59.190791 [ 45086 ] {} <Error> void DB::StorageKafka::threadFunc(size_t): Code: 72, e.displayText() = DB::ParsingException: Cannot parse number with a sign character but without any numeric character: Cannot parse Int32 from String: while converting source column proxy_status to destination column proxy_status: while pushing to view eventstream.http_access_log_queue_v5_mv (5efff929-803b-4ab4-9eff-f929803b8ab4), Stack trace (when copying this message, always include the lines below):
0. DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool) @ 0x8fdbc9a in /usr/bin/clickhouse
1. void DB::readIntTextImpl<int, void, (DB::ReadIntTextCheckOverflow)0>(int&, DB::ReadBuffer&) @ 0x8fef791 in /usr/bin/clickhouse
2. COW<DB::IColumn>::immutable_ptr<DB::IColumn> DB::ConvertThroughParsing<DB::DataTypeString, DB::DataTypeNumber<int>, DB::NameToInt32, (DB::ConvertFromStringExceptionMode)0, (DB::ConvertFromStringParsingMode)0>::execute<void*>(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, void*) @ 0xb2d1738 in /usr/bin/clickhouse
3. bool DB::callOnIndexAndDataType<DB::DataTypeNumber<int>, DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeInternal(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const::'lambda'(auto const&, auto const&)&, DB::ConvertDefaultBehaviorTag>(DB::TypeIndex, auto&&, DB::ToNumberMonotonicity<int>&&...) @ 0xb2c5ac9 in /usr/bin/clickhouse
4. DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeInternal(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb2c4fec in /usr/bin/clickhouse
5. DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeImpl(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb2c463c in /usr/bin/clickhouse
6. DB::FunctionToExecutableFunctionAdaptor::executeImpl(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb16d12e in /usr/bin/clickhouse
7. DB::IExecutableFunction::executeWithoutLowCardinalityColumns(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabc4de in /usr/bin/clickhouse
8. DB::IExecutableFunction::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabcaf2 in /usr/bin/clickhouse
9. DB::IFunctionBase::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xb16cc27 in /usr/bin/clickhouse
10. COW<DB::IColumn>::immutable_ptr<DB::IColumn> std::__1::__function::__policy_invoker<COW<DB::IColumn>::immutable_ptr<DB::IColumn> (std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long)>::__call_impl<std::__1::__function::__default_alloc_func<DB::FunctionCast::createFunctionAdaptor(std::__1::shared_ptr<DB::IFunction>, std::__1::shared_ptr<DB::IDataType const> const&)::'lambda'(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long), COW<DB::IColumn>::immutable_ptr<DB::IColumn> (std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long)> >(std::__1::__function::__policy_storage const*, std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long) @ 0xb400b56 in /usr/bin/clickhouse
11. ? @ 0xb3f8120 in /usr/bin/clickhouse
12. DB::IExecutableFunction::executeWithoutLowCardinalityColumns(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabc4de in /usr/bin/clickhouse
13. DB::IExecutableFunction::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabcbfd in /usr/bin/clickhouse
14. DB::IFunctionBase::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xb16cc27 in /usr/bin/clickhouse
15. DB::castColumn(DB::ColumnWithTypeAndName const&, std::__1::shared_ptr<DB::IDataType const> const&) @ 0x10774bd1 in /usr/bin/clickhouse
16. ? @ 0xfdd587d in /usr/bin/clickhouse
17. DB::ConvertingBlockInputStream::readImpl() @ 0xfdd5ce5 in /usr/bin/clickhouse
18. DB::IBlockInputStream::read() @ 0xfdd6c67 in /usr/bin/clickhouse
19. DB::PushingToViewsBlockOutputStream::process(DB::Block const&, DB::PushingToViewsBlockOutputStream::ViewInfo&) @ 0x103da76c in /usr/bin/clickhouse
20. DB::PushingToViewsBlockOutputStream::write(DB::Block const&) @ 0x103d9cee in /usr/bin/clickhouse
21. DB::AddingDefaultBlockOutputStream::write(DB::Block const&) @ 0x103e34ab in /usr/bin/clickhouse
22. ? @ 0x100eaf3e in /usr/bin/clickhouse
23. DB::copyData(DB::IBlockInputStream&, DB::IBlockOutputStream&, std::__1::function<void (DB::Block const&)> const&, std::__1::atomic<bool>*) @ 0xfdfb14a in /usr/bin/clickhouse
24. DB::StorageKafka::streamToViews() @ 0x10bc7c59 in /usr/bin/clickhouse
25. DB::StorageKafka::threadFunc(unsigned long) @ 0x10bc6838 in /usr/bin/clickhouse
26. DB::BackgroundSchedulePoolTaskInfo::execute() @ 0x10042328 in /usr/bin/clickhouse
27. DB::BackgroundSchedulePool::threadFunction() @ 0x10044357 in /usr/bin/clickhouse
28. ? @ 0x100450d4 in /usr/bin/clickhouse
29. ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) @ 0x901c91f in /usr/bin/clickhouse
30. ? @ 0x9020203 in /usr/bin/clickhouse
31. start_thread @ 0x7ea5 in /usr/lib64/libpthread-2.17.so
(version 21.8.10.19 (official build))
```
**Additional context**
I know that I can solve this using some other function to get rid of the hyphen, but consider 1MM INSERT per second, this is a daily growing throughput, you know that any nanosecond on this will turn into seconds at some point.
My opinion here is that ReadInt should check the sign of course, but only when the value has a number. Any value missing a number should be treated as a string and, zero or null should be returned. Also, makes no sense for me use a not-unsigned integer conversion function that tries to validate the sign. | True | Pedantic readInt brakes use case behavior - **Describe the issue**
Moving to the last LTS version (21.8.10.19) I discover that [#24475](https://github.com/ClickHouse/ClickHouse/pull/24475) break the behavior I rely on in a fairly big pipeline at my company (~1MM inserts per sec).
The events are Nginx access logs, where some non-existent values can be logged as a hyphen (`-`). When the correct value is expected as integers this is indeed a consistency fault. But also it is the way Nginx works and will not change anytime soon.
¯\\_(ツ)_/¯
At the current version I'm using (21.4.4.30) my log-forwarders can do INSERT using `-` for `Int8|16|32|64` columns without any problem, the value is stored as `0` as expected.
But now at the LTS version, the exception below arises. I'm using Kafka engine tables in the environment but, if you look at the examples to reproduce below the exception you'll see that this is not a Kafka Engine-related problem.
````
DB::ParsingException: Cannot parse number with a sign character but without any numeric character: Cannot parse Int32 from String
[Complete stack at the end]
````
**How to reproduce**
This will work just fine in any version before [#24475](https://github.com/ClickHouse/ClickHouse/pull/24475), after that the exception mentioned arises.
```
CREATE TABLE numbers
(
`number` Int32
)
ENGINE = MergeTree
ORDER BY number
INSERT INTO numbers VALUES ('-');
```
**Error message and/or stacktrace**
```
2021.11.13 09:56:59.190791 [ 45086 ] {} <Error> void DB::StorageKafka::threadFunc(size_t): Code: 72, e.displayText() = DB::ParsingException: Cannot parse number with a sign character but without any numeric character: Cannot parse Int32 from String: while converting source column proxy_status to destination column proxy_status: while pushing to view eventstream.http_access_log_queue_v5_mv (5efff929-803b-4ab4-9eff-f929803b8ab4), Stack trace (when copying this message, always include the lines below):
0. DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool) @ 0x8fdbc9a in /usr/bin/clickhouse
1. void DB::readIntTextImpl<int, void, (DB::ReadIntTextCheckOverflow)0>(int&, DB::ReadBuffer&) @ 0x8fef791 in /usr/bin/clickhouse
2. COW<DB::IColumn>::immutable_ptr<DB::IColumn> DB::ConvertThroughParsing<DB::DataTypeString, DB::DataTypeNumber<int>, DB::NameToInt32, (DB::ConvertFromStringExceptionMode)0, (DB::ConvertFromStringParsingMode)0>::execute<void*>(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, void*) @ 0xb2d1738 in /usr/bin/clickhouse
3. bool DB::callOnIndexAndDataType<DB::DataTypeNumber<int>, DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeInternal(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const::'lambda'(auto const&, auto const&)&, DB::ConvertDefaultBehaviorTag>(DB::TypeIndex, auto&&, DB::ToNumberMonotonicity<int>&&...) @ 0xb2c5ac9 in /usr/bin/clickhouse
4. DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeInternal(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb2c4fec in /usr/bin/clickhouse
5. DB::FunctionConvert<DB::DataTypeNumber<int>, DB::NameToInt32, DB::ToNumberMonotonicity<int> >::executeImpl(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb2c463c in /usr/bin/clickhouse
6. DB::FunctionToExecutableFunctionAdaptor::executeImpl(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long) const @ 0xb16d12e in /usr/bin/clickhouse
7. DB::IExecutableFunction::executeWithoutLowCardinalityColumns(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabc4de in /usr/bin/clickhouse
8. DB::IExecutableFunction::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabcaf2 in /usr/bin/clickhouse
9. DB::IFunctionBase::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xb16cc27 in /usr/bin/clickhouse
10. COW<DB::IColumn>::immutable_ptr<DB::IColumn> std::__1::__function::__policy_invoker<COW<DB::IColumn>::immutable_ptr<DB::IColumn> (std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long)>::__call_impl<std::__1::__function::__default_alloc_func<DB::FunctionCast::createFunctionAdaptor(std::__1::shared_ptr<DB::IFunction>, std::__1::shared_ptr<DB::IDataType const> const&)::'lambda'(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long), COW<DB::IColumn>::immutable_ptr<DB::IColumn> (std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long)> >(std::__1::__function::__policy_storage const*, std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> >&, std::__1::shared_ptr<DB::IDataType const> const&, DB::ColumnNullable const*, unsigned long) @ 0xb400b56 in /usr/bin/clickhouse
11. ? @ 0xb3f8120 in /usr/bin/clickhouse
12. DB::IExecutableFunction::executeWithoutLowCardinalityColumns(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabc4de in /usr/bin/clickhouse
13. DB::IExecutableFunction::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xfabcbfd in /usr/bin/clickhouse
14. DB::IFunctionBase::execute(std::__1::vector<DB::ColumnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName> > const&, std::__1::shared_ptr<DB::IDataType const> const&, unsigned long, bool) const @ 0xb16cc27 in /usr/bin/clickhouse
15. DB::castColumn(DB::ColumnWithTypeAndName const&, std::__1::shared_ptr<DB::IDataType const> const&) @ 0x10774bd1 in /usr/bin/clickhouse
16. ? @ 0xfdd587d in /usr/bin/clickhouse
17. DB::ConvertingBlockInputStream::readImpl() @ 0xfdd5ce5 in /usr/bin/clickhouse
18. DB::IBlockInputStream::read() @ 0xfdd6c67 in /usr/bin/clickhouse
19. DB::PushingToViewsBlockOutputStream::process(DB::Block const&, DB::PushingToViewsBlockOutputStream::ViewInfo&) @ 0x103da76c in /usr/bin/clickhouse
20. DB::PushingToViewsBlockOutputStream::write(DB::Block const&) @ 0x103d9cee in /usr/bin/clickhouse
21. DB::AddingDefaultBlockOutputStream::write(DB::Block const&) @ 0x103e34ab in /usr/bin/clickhouse
22. ? @ 0x100eaf3e in /usr/bin/clickhouse
23. DB::copyData(DB::IBlockInputStream&, DB::IBlockOutputStream&, std::__1::function<void (DB::Block const&)> const&, std::__1::atomic<bool>*) @ 0xfdfb14a in /usr/bin/clickhouse
24. DB::StorageKafka::streamToViews() @ 0x10bc7c59 in /usr/bin/clickhouse
25. DB::StorageKafka::threadFunc(unsigned long) @ 0x10bc6838 in /usr/bin/clickhouse
26. DB::BackgroundSchedulePoolTaskInfo::execute() @ 0x10042328 in /usr/bin/clickhouse
27. DB::BackgroundSchedulePool::threadFunction() @ 0x10044357 in /usr/bin/clickhouse
28. ? @ 0x100450d4 in /usr/bin/clickhouse
29. ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) @ 0x901c91f in /usr/bin/clickhouse
30. ? @ 0x9020203 in /usr/bin/clickhouse
31. start_thread @ 0x7ea5 in /usr/lib64/libpthread-2.17.so
(version 21.8.10.19 (official build))
```
**Additional context**
I know that I can solve this using some other function to get rid of the hyphen, but consider 1MM INSERT per second, this is a daily growing throughput, you know that any nanosecond on this will turn into seconds at some point.
My opinion here is that ReadInt should check the sign of course, but only when the value has a number. Any value missing a number should be treated as a string and, zero or null should be returned. Also, makes no sense for me use a not-unsigned integer conversion function that tries to validate the sign. | comp | pedantic readint brakes use case behavior describe the issue moving to the last lts version i discover that break the behavior i rely on in a fairly big pipeline at my company inserts per sec the events are nginx access logs where some non existent values can be logged as a hyphen when the correct value is expected as integers this is indeed a consistency fault but also it is the way nginx works and will not change anytime soon ¯ ツ ¯ at the current version i m using my log forwarders can do insert using for columns without any problem the value is stored as as expected but now at the lts version the exception below arises i m using kafka engine tables in the environment but if you look at the examples to reproduce below the exception you ll see that this is not a kafka engine related problem db parsingexception cannot parse number with a sign character but without any numeric character cannot parse from string how to reproduce this will work just fine in any version before after that the exception mentioned arises create table numbers number engine mergetree order by number insert into numbers values error message and or stacktrace void db storagekafka threadfunc size t code e displaytext db parsingexception cannot parse number with a sign character but without any numeric character cannot parse from string while converting source column proxy status to destination column proxy status while pushing to view eventstream http access log queue mv stack trace when copying this message always include the lines below db exception exception std basic string std allocator const int bool in usr bin clickhouse void db readinttextimpl int db readbuffer in usr bin clickhouse cow immutable ptr db convertthroughparsing db db convertfromstringexceptionmode db convertfromstringparsingmode execute std vector const std shared ptr const unsigned long void in usr bin clickhouse bool db callonindexanddatatype db functionconvert db db tonumbermonotonicity executeinternal std vector const std shared ptr const unsigned long const lambda auto const auto const db convertdefaultbehaviortag db typeindex auto db tonumbermonotonicity in usr bin clickhouse db functionconvert db db tonumbermonotonicity executeinternal std vector const std shared ptr const unsigned long const in usr bin clickhouse db functionconvert db db tonumbermonotonicity executeimpl std vector const std shared ptr const unsigned long const in usr bin clickhouse db functiontoexecutablefunctionadaptor executeimpl std vector const std shared ptr const unsigned long const in usr bin clickhouse db iexecutablefunction executewithoutlowcardinalitycolumns std vector const std shared ptr const unsigned long bool const in usr bin clickhouse db iexecutablefunction execute std vector const std shared ptr const unsigned long bool const in usr bin clickhouse db ifunctionbase execute std vector const std shared ptr const unsigned long bool const in usr bin clickhouse cow immutable ptr std function policy invoker immutable ptr std vector std shared ptr const db columnnullable const unsigned long call impl std shared ptr const lambda std vector std shared ptr const db columnnullable const unsigned long cow immutable ptr std vector std shared ptr const db columnnullable const unsigned long std function policy storage const std vector std shared ptr const db columnnullable const unsigned long in usr bin clickhouse in usr bin clickhouse db iexecutablefunction executewithoutlowcardinalitycolumns std vector const std shared ptr const unsigned long bool const in usr bin clickhouse db iexecutablefunction execute std vector const std shared ptr const unsigned long bool const in usr bin clickhouse db ifunctionbase execute std vector const std shared ptr const unsigned long bool const in usr bin clickhouse db castcolumn db columnwithtypeandname const std shared ptr const in usr bin clickhouse in usr bin clickhouse db convertingblockinputstream readimpl in usr bin clickhouse db iblockinputstream read in usr bin clickhouse db pushingtoviewsblockoutputstream process db block const db pushingtoviewsblockoutputstream viewinfo in usr bin clickhouse db pushingtoviewsblockoutputstream write db block const in usr bin clickhouse db addingdefaultblockoutputstream write db block const in usr bin clickhouse in usr bin clickhouse db copydata db iblockinputstream db iblockoutputstream std function const std atomic in usr bin clickhouse db storagekafka streamtoviews in usr bin clickhouse db storagekafka threadfunc unsigned long in usr bin clickhouse db backgroundschedulepooltaskinfo execute in usr bin clickhouse db backgroundschedulepool threadfunction in usr bin clickhouse in usr bin clickhouse threadpoolimpl worker std list iterator in usr bin clickhouse in usr bin clickhouse start thread in usr libpthread so version official build additional context i know that i can solve this using some other function to get rid of the hyphen but consider insert per second this is a daily growing throughput you know that any nanosecond on this will turn into seconds at some point my opinion here is that readint should check the sign of course but only when the value has a number any value missing a number should be treated as a string and zero or null should be returned also makes no sense for me use a not unsigned integer conversion function that tries to validate the sign | 1 |
40,054 | 6,796,408,418 | IssuesEvent | 2017-11-01 18:54:41 | edgi-govdata-archiving/web-monitoring | https://api.github.com/repos/edgi-govdata-archiving/web-monitoring | closed | Identify component/area leads in readme | documentation in progress | It would be helpful to 'pin' the top to our repo:
1. Area leads (e.g. of the 2 parts, as user spokesperson, as pm)
1. Channels on Slack...? | 1.0 | Identify component/area leads in readme - It would be helpful to 'pin' the top to our repo:
1. Area leads (e.g. of the 2 parts, as user spokesperson, as pm)
1. Channels on Slack...? | non_comp | identify component area leads in readme it would be helpful to pin the top to our repo area leads e g of the parts as user spokesperson as pm channels on slack | 0 |
276,489 | 23,995,137,699 | IssuesEvent | 2022-09-14 06:51:56 | FreeRDP/FreeRDP | https://api.github.com/repos/FreeRDP/FreeRDP | closed | /dynamic-resolution does not working. | fixed-waiting-test client | **Describe the bug**
After running with /f, /dynamic-resolution does not work when using shortcut keys or float bar to cancel full screen.
**To Reproduce**
1. Run xfreerdp command.
$ xfreerdp /log-level:DEBUG /f /floatbar:sticky:off,default:hidden,show:always /dynamic-resolution /v:xxx.xxx.xx.x /u:xxxxx
2. Remove the full screen display by manipulating the floatbar (or press ctrl + alt + enter)
3. Resize the window.
4. Automatic resizing does not work.The window size can be adjusted, but the resolution is not changed.
**Expected behavior**
When the window size is adjusted, I would like the resolution to automatically change to match the window size.
**Screenshots**

**Additional context**
the problem did not occur in v2.2.0, but has been occurring since v2.3.0.
In addition to /f, the problem was also reproduced with /multimon.
I am connected to windows 10.
Specifically, the following commit was causing the problem:
d3ed42a7
| 1.0 | /dynamic-resolution does not working. - **Describe the bug**
After running with /f, /dynamic-resolution does not work when using shortcut keys or float bar to cancel full screen.
**To Reproduce**
1. Run xfreerdp command.
$ xfreerdp /log-level:DEBUG /f /floatbar:sticky:off,default:hidden,show:always /dynamic-resolution /v:xxx.xxx.xx.x /u:xxxxx
2. Remove the full screen display by manipulating the floatbar (or press ctrl + alt + enter)
3. Resize the window.
4. Automatic resizing does not work.The window size can be adjusted, but the resolution is not changed.
**Expected behavior**
When the window size is adjusted, I would like the resolution to automatically change to match the window size.
**Screenshots**

**Additional context**
the problem did not occur in v2.2.0, but has been occurring since v2.3.0.
In addition to /f, the problem was also reproduced with /multimon.
I am connected to windows 10.
Specifically, the following commit was causing the problem:
d3ed42a7
| non_comp | dynamic resolution does not working describe the bug after running with f dynamic resolution does not work when using shortcut keys or float bar to cancel full screen to reproduce run xfreerdp command xfreerdp log level debug f floatbar sticky off default hidden show always dynamic resolution v xxx xxx xx x u xxxxx remove the full screen display by manipulating the floatbar or press ctrl alt enter resize the window automatic resizing does not work the window size can be adjusted but the resolution is not changed expected behavior when the window size is adjusted i would like the resolution to automatically change to match the window size screenshots additional context the problem did not occur in but has been occurring since in addition to f the problem was also reproduced with multimon i am connected to windows specifically the following commit was causing the problem | 0 |
7,419 | 9,666,885,914 | IssuesEvent | 2019-05-21 11:59:38 | arcticicestudio/nord-docs | https://api.github.com/repos/arcticicestudio/nord-docs | closed | Fix MDX v1 remark/rehype plugin loading after migration | scope-compatibility scope-plugin-support type-bug | During the migration to [MDX 1.0.0][mdx-v1] in #137 the [now deprecated][mdx-depr] `mdPlugins` and `hastPlugins` options were not replaced with their (new named) respective
equivalents `remarkPlugins` and `rehypePlugins`. Even if the documentation states that the options will be removed in v2 and are still supported (only showing a warning in the console when still used), the defined plugins were not loaded anymore causing e.g. no more automatic generation of `id` attributes for headers in MDX content.
Therefore, to finish 100% of the migration, both options should be renamed.
[mdx-v1]: https://mdxjs.com/blog/v1
[mdx-depr]: https://mdxjs.com/blog/v1#deprecations | True | Fix MDX v1 remark/rehype plugin loading after migration - During the migration to [MDX 1.0.0][mdx-v1] in #137 the [now deprecated][mdx-depr] `mdPlugins` and `hastPlugins` options were not replaced with their (new named) respective
equivalents `remarkPlugins` and `rehypePlugins`. Even if the documentation states that the options will be removed in v2 and are still supported (only showing a warning in the console when still used), the defined plugins were not loaded anymore causing e.g. no more automatic generation of `id` attributes for headers in MDX content.
Therefore, to finish 100% of the migration, both options should be renamed.
[mdx-v1]: https://mdxjs.com/blog/v1
[mdx-depr]: https://mdxjs.com/blog/v1#deprecations | comp | fix mdx remark rehype plugin loading after migration during the migration to in the mdplugins and hastplugins options were not replaced with their new named respective equivalents remarkplugins and rehypeplugins even if the documentation states that the options will be removed in and are still supported only showing a warning in the console when still used the defined plugins were not loaded anymore causing e g no more automatic generation of id attributes for headers in mdx content therefore to finish of the migration both options should be renamed | 1 |
6,534 | 5,510,874,644 | IssuesEvent | 2017-03-17 01:51:04 | GoogleCloudPlatform/google-cloud-java | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-java | closed | Large memory usage by gcloud-logging API. | API: logging performance Status: Pending Release Status: Release Blocking | The logging API has a large memory foot print, with approximately 4KB of memory consumed for each log generated.
This was measured with the following program:
```java
package com.google.cloud.runtimes.jetty9;
import java.util.concurrent.TimeUnit;
import java.util.logging.Logger;
public class StackDriverLogging {
static void init() throws Exception {
Logger log = Logger.getLogger(StackDriverLogging.class.getName());
log.info("test info");
log.fine("test fine");
long last = Runtime.getRuntime().totalMemory() - Runtime.getRuntime().freeMemory();
// Run 50 times to ensure everything is hot and any queues are max allocated
for (int r=50;r-->0;) {
long total = Runtime.getRuntime().totalMemory();
long free = Runtime.getRuntime().freeMemory();
long used = total - free;
long delta = used - last;
last = used;
System.err.printf("total(%d)-free(%d)=used(%d) delta=%d%n", total, free, used, delta);
for (int i = 1000; i-- > 0;)
log.info("test info");
if (r==10) {
System.err.println("DRAINING...");
// Wait a minute so all queued logs can drain out
TimeUnit.MINUTES.sleep(1);
System.gc();
System.err.println("LAST 10...");
}
Thread.sleep(10);
}
System.err.println("DUMP!");
Thread.sleep(1000000);
}
public static void main(String... args) throws Exception {
init();
}
}
```
Eclipse MAT was used to analyse a head dump taken immediately after 10K log entries were generated, then again once the program had returned to idle and network monitor indicated that all data had been flushed.
After 10L log messages, the heap was 51MB, which reduced to 13.4MB once the logs had been flushed (but no forced GC, so the difference may be larger). Thus at least 38MB appears to be consumed by logging 10K messages, which is 3.9KB per message. The attached reports( [10K Top Consumers.pdf](https://github.com/GoogleCloudPlatform/google-cloud-java/files/632824/10K.Top.Consumers.pdf) & [Idle Top Consumers.pdf](https://github.com/GoogleCloudPlatform/google-cloud-java/files/632825/Idle.Top.Consumers.pdf) ) indicate that 22.9 MB is used by `io.grpc.netty.NettyClientTransport$2` and 12.1MB alone is consumed by `io.netty.handler.codec.http2.StreamBufferingEncoder`.
Ideally, the memory usage per log line could be greatly reduced. Eitherway, it would also be highly desirable to be able to put some limits on the memory consumed by the logging subsystem so that excessively verbose logs are either discarded, summarized, blocked or excepted.
| True | Large memory usage by gcloud-logging API. - The logging API has a large memory foot print, with approximately 4KB of memory consumed for each log generated.
This was measured with the following program:
```java
package com.google.cloud.runtimes.jetty9;
import java.util.concurrent.TimeUnit;
import java.util.logging.Logger;
public class StackDriverLogging {
static void init() throws Exception {
Logger log = Logger.getLogger(StackDriverLogging.class.getName());
log.info("test info");
log.fine("test fine");
long last = Runtime.getRuntime().totalMemory() - Runtime.getRuntime().freeMemory();
// Run 50 times to ensure everything is hot and any queues are max allocated
for (int r=50;r-->0;) {
long total = Runtime.getRuntime().totalMemory();
long free = Runtime.getRuntime().freeMemory();
long used = total - free;
long delta = used - last;
last = used;
System.err.printf("total(%d)-free(%d)=used(%d) delta=%d%n", total, free, used, delta);
for (int i = 1000; i-- > 0;)
log.info("test info");
if (r==10) {
System.err.println("DRAINING...");
// Wait a minute so all queued logs can drain out
TimeUnit.MINUTES.sleep(1);
System.gc();
System.err.println("LAST 10...");
}
Thread.sleep(10);
}
System.err.println("DUMP!");
Thread.sleep(1000000);
}
public static void main(String... args) throws Exception {
init();
}
}
```
Eclipse MAT was used to analyse a head dump taken immediately after 10K log entries were generated, then again once the program had returned to idle and network monitor indicated that all data had been flushed.
After 10L log messages, the heap was 51MB, which reduced to 13.4MB once the logs had been flushed (but no forced GC, so the difference may be larger). Thus at least 38MB appears to be consumed by logging 10K messages, which is 3.9KB per message. The attached reports( [10K Top Consumers.pdf](https://github.com/GoogleCloudPlatform/google-cloud-java/files/632824/10K.Top.Consumers.pdf) & [Idle Top Consumers.pdf](https://github.com/GoogleCloudPlatform/google-cloud-java/files/632825/Idle.Top.Consumers.pdf) ) indicate that 22.9 MB is used by `io.grpc.netty.NettyClientTransport$2` and 12.1MB alone is consumed by `io.netty.handler.codec.http2.StreamBufferingEncoder`.
Ideally, the memory usage per log line could be greatly reduced. Eitherway, it would also be highly desirable to be able to put some limits on the memory consumed by the logging subsystem so that excessively verbose logs are either discarded, summarized, blocked or excepted.
| non_comp | large memory usage by gcloud logging api the logging api has a large memory foot print with approximately of memory consumed for each log generated this was measured with the following program java package com google cloud runtimes import java util concurrent timeunit import java util logging logger public class stackdriverlogging static void init throws exception logger log logger getlogger stackdriverlogging class getname log info test info log fine test fine long last runtime getruntime totalmemory runtime getruntime freememory run times to ensure everything is hot and any queues are max allocated for int r r long total runtime getruntime totalmemory long free runtime getruntime freememory long used total free long delta used last last used system err printf total d free d used d delta d n total free used delta for int i i log info test info if r system err println draining wait a minute so all queued logs can drain out timeunit minutes sleep system gc system err println last thread sleep system err println dump thread sleep public static void main string args throws exception init eclipse mat was used to analyse a head dump taken immediately after log entries were generated then again once the program had returned to idle and network monitor indicated that all data had been flushed after log messages the heap was which reduced to once the logs had been flushed but no forced gc so the difference may be larger thus at least appears to be consumed by logging messages which is per message the attached reports indicate that mb is used by io grpc netty nettyclienttransport and alone is consumed by io netty handler codec streambufferingencoder ideally the memory usage per log line could be greatly reduced eitherway it would also be highly desirable to be able to put some limits on the memory consumed by the logging subsystem so that excessively verbose logs are either discarded summarized blocked or excepted | 0 |
104,144 | 16,613,455,706 | IssuesEvent | 2021-06-02 14:09:20 | Thanraj/linux-4.1.15 | https://api.github.com/repos/Thanraj/linux-4.1.15 | opened | CVE-2018-20856 (High) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2018-20856 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/block/blk-core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/block/blk-core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 4.18.7. In block/blk-core.c, there is an __blk_drain_queue() use-after-free because a certain error case is mishandled.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20856>CVE-2018-20856</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20856">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20856</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution: 4.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20856 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2018-20856 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/block/blk-core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/block/blk-core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 4.18.7. In block/blk-core.c, there is an __blk_drain_queue() use-after-free because a certain error case is mishandled.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20856>CVE-2018-20856</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20856">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20856</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution: 4.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_comp | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files linux block blk core c linux block blk core c vulnerability details an issue was discovered in the linux kernel before in block blk core c there is an blk drain queue use after free because a certain error case is mishandled publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
117,244 | 17,439,428,144 | IssuesEvent | 2021-08-05 01:15:32 | snowdensb/questdb | https://api.github.com/repos/snowdensb/questdb | opened | CVE-2021-32803 (High) detected in tar-6.0.5.tgz | security vulnerability | ## CVE-2021-32803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.0.5.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.0.5.tgz">https://registry.npmjs.org/tar/-/tar-6.0.5.tgz</a></p>
<p>Path to dependency file: questdb/ui/package.json</p>
<p>Path to vulnerable library: questdb/ui/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- copy-webpack-plugin-6.0.2.tgz (Root Library)
- cacache-15.0.5.tgz
- :x: **tar-6.0.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32803>CVE-2021-32803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-r628-mhmh-qjhw">https://github.com/npm/node-tar/security/advisories/GHSA-r628-mhmh-qjhw</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.3, 4.4.15, 5.0.7, 6.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"6.0.5","packageFilePaths":["/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"copy-webpack-plugin:6.0.2;cacache:15.0.5;tar:6.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.3, 4.4.15, 5.0.7, 6.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32803","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32803","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-32803 (High) detected in tar-6.0.5.tgz - ## CVE-2021-32803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.0.5.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.0.5.tgz">https://registry.npmjs.org/tar/-/tar-6.0.5.tgz</a></p>
<p>Path to dependency file: questdb/ui/package.json</p>
<p>Path to vulnerable library: questdb/ui/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- copy-webpack-plugin-6.0.2.tgz (Root Library)
- cacache-15.0.5.tgz
- :x: **tar-6.0.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32803>CVE-2021-32803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-r628-mhmh-qjhw">https://github.com/npm/node-tar/security/advisories/GHSA-r628-mhmh-qjhw</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.3, 4.4.15, 5.0.7, 6.1.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"6.0.5","packageFilePaths":["/ui/package.json"],"isTransitiveDependency":true,"dependencyTree":"copy-webpack-plugin:6.0.2;cacache:15.0.5;tar:6.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.3, 4.4.15, 5.0.7, 6.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32803","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.2, 5.0.7, 4.4.15, and 3.2.3 has an arbitrary File Creation/Overwrite vulnerability via insufficient symlink protection. `node-tar` aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary `stat` calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory. This order of operations resulted in the directory being created and added to the `node-tar` directory cache. When a directory is present in the directory cache, subsequent calls to mkdir for that directory are skipped. However, this is also where `node-tar` checks for symlinks occur. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass `node-tar` symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.3, 4.4.15, 5.0.7 and 6.1.2.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32803","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | non_comp | cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file questdb ui package json path to vulnerable library questdb ui node modules tar package json dependency hierarchy copy webpack plugin tgz root library cacache tgz x tar tgz vulnerable library found in base branch master vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite vulnerability via insufficient symlink protection node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory this order of operations resulted in the directory being created and added to the node tar directory cache when a directory is present in the directory cache subsequent calls to mkdir for that directory are skipped however this is also where node tar checks for symlinks occur by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite this issue was addressed in releases and publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree copy webpack plugin cacache tar isminimumfixversionavailable true minimumfixversion tar basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm package tar aka node tar before versions and has an arbitrary file creation overwrite vulnerability via insufficient symlink protection node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory this order of operations resulted in the directory being created and added to the node tar directory cache when a directory is present in the directory cache subsequent calls to mkdir for that directory are skipped however this is also where node tar checks for symlinks occur by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite this issue was addressed in releases and vulnerabilityurl | 0 |
11,345 | 13,281,675,479 | IssuesEvent | 2020-08-23 18:32:34 | grondag/canvas | https://api.github.com/repos/grondag/canvas | closed | Unnecessary fallback codepath for CBT | compatibility | Hello! I noticed the fallback code for CBT is still in there, I've since fixed the issue with it not liking Fabric models. Hopefully. Anyway, this codepath seems to segfault with Now Playing (see: https://github.com/Nuclearfarts/connected-block-textures/issues/12), so it would be nice if this could be fixed. | True | Unnecessary fallback codepath for CBT - Hello! I noticed the fallback code for CBT is still in there, I've since fixed the issue with it not liking Fabric models. Hopefully. Anyway, this codepath seems to segfault with Now Playing (see: https://github.com/Nuclearfarts/connected-block-textures/issues/12), so it would be nice if this could be fixed. | comp | unnecessary fallback codepath for cbt hello i noticed the fallback code for cbt is still in there i ve since fixed the issue with it not liking fabric models hopefully anyway this codepath seems to segfault with now playing see so it would be nice if this could be fixed | 1 |
14,252 | 17,135,923,004 | IssuesEvent | 2021-07-13 02:06:59 | PhenX/php-svg-lib | https://api.github.com/repos/PhenX/php-svg-lib | closed | PHP 7.4 compatibility | php compatibility | When used with DomPDF Cpdf.php file, the scripts sets arrays ($currentLineTransparency and $currentFillTransparency) to null:
https://github.com/PhenX/php-svg-lib/blob/master/src/Svg/Surface/SurfaceCpdf.php#L394
In Cpdf it checks first if attempted mode and opacity is same as set before, this produces "Trying to access array offset on value of type null" notice in Php 7.4, because SurfadeCpdf sets those arrays to null after setting mode | True | PHP 7.4 compatibility - When used with DomPDF Cpdf.php file, the scripts sets arrays ($currentLineTransparency and $currentFillTransparency) to null:
https://github.com/PhenX/php-svg-lib/blob/master/src/Svg/Surface/SurfaceCpdf.php#L394
In Cpdf it checks first if attempted mode and opacity is same as set before, this produces "Trying to access array offset on value of type null" notice in Php 7.4, because SurfadeCpdf sets those arrays to null after setting mode | comp | php compatibility when used with dompdf cpdf php file the scripts sets arrays currentlinetransparency and currentfilltransparency to null in cpdf it checks first if attempted mode and opacity is same as set before this produces trying to access array offset on value of type null notice in php because surfadecpdf sets those arrays to null after setting mode | 1 |
8,958 | 10,975,491,964 | IssuesEvent | 2019-11-29 11:37:10 | WOTCStrategyOverhaul/CovertInfiltration | https://api.github.com/repos/WOTCStrategyOverhaul/CovertInfiltration | closed | Error Message about missing rewards at the end of the mission | bug gameplay mod compatibility | Various missions like rescue a vip will result in errors about CI having troubles with the mission reward string.
https://pastebin.com/PiBRRyLp | True | Error Message about missing rewards at the end of the mission - Various missions like rescue a vip will result in errors about CI having troubles with the mission reward string.
https://pastebin.com/PiBRRyLp | comp | error message about missing rewards at the end of the mission various missions like rescue a vip will result in errors about ci having troubles with the mission reward string | 1 |
315,537 | 9,621,958,265 | IssuesEvent | 2019-05-14 11:58:13 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | opened | v9-gui: display of access level for users | Priority: Medium Type: Bug | When access level is set to NONE for a user, **new** GUI doesn't display this information in `Actions` tab of the user. Not the case with `ALL`.
As I mentioned in #4395, after a new creation of a user, access level is displayed. If I clear my browser cache, I lost this information. | 1.0 | v9-gui: display of access level for users - When access level is set to NONE for a user, **new** GUI doesn't display this information in `Actions` tab of the user. Not the case with `ALL`.
As I mentioned in #4395, after a new creation of a user, access level is displayed. If I clear my browser cache, I lost this information. | non_comp | gui display of access level for users when access level is set to none for a user new gui doesn t display this information in actions tab of the user not the case with all as i mentioned in after a new creation of a user access level is displayed if i clear my browser cache i lost this information | 0 |
12,467 | 14,703,362,718 | IssuesEvent | 2021-01-04 14:56:40 | svenhjol/CharmForge | https://api.github.com/repos/svenhjol/CharmForge | closed | Crash on startup, "Critical injection failure" | compatibility | Versions
* Charm version - `2.2.1`
* Minecraft version - `1.16.4`
Description
Crash on startup, not experienced enough to understand why
Optional: Steps To Reproduce
Installed Charm-reforged & charmonium-reforged
Optional: Stacktrace/Screenshots
[latest.log](https://pastebin.com/AaX9LGBm)
| True | Crash on startup, "Critical injection failure" - Versions
* Charm version - `2.2.1`
* Minecraft version - `1.16.4`
Description
Crash on startup, not experienced enough to understand why
Optional: Steps To Reproduce
Installed Charm-reforged & charmonium-reforged
Optional: Stacktrace/Screenshots
[latest.log](https://pastebin.com/AaX9LGBm)
| comp | crash on startup critical injection failure versions charm version minecraft version description crash on startup not experienced enough to understand why optional steps to reproduce installed charm reforged charmonium reforged optional stacktrace screenshots | 1 |
3,434 | 6,401,916,256 | IssuesEvent | 2017-08-06 02:52:40 | checkstyle/checkstyle | https://api.github.com/repos/checkstyle/checkstyle | closed | move AbstractTypeAwareCheck and ClassResolver to javadoc package | approved breaking compatibility | AbstractTypeAwareCheck is deprecated and used only in javadoc Checks for now.
it is required to move all that deprecated functionality close to almost deprecated javadoc Checks that are not based on AST. | True | move AbstractTypeAwareCheck and ClassResolver to javadoc package - AbstractTypeAwareCheck is deprecated and used only in javadoc Checks for now.
it is required to move all that deprecated functionality close to almost deprecated javadoc Checks that are not based on AST. | comp | move abstracttypeawarecheck and classresolver to javadoc package abstracttypeawarecheck is deprecated and used only in javadoc checks for now it is required to move all that deprecated functionality close to almost deprecated javadoc checks that are not based on ast | 1 |
10,872 | 12,866,896,512 | IssuesEvent | 2020-07-10 05:21:20 | select2/select2 | https://api.github.com/repos/select2/select2 | closed | Deprecate and remove untested AttachContainer decorator | 4.x compatibility feature: dropdown status: stale | This decorator appears to be largely unused. A GitHub scan is being performed to catalog the common use cases for this decorator and to determine the feasibility of its removal in a future version of Select2. | True | Deprecate and remove untested AttachContainer decorator - This decorator appears to be largely unused. A GitHub scan is being performed to catalog the common use cases for this decorator and to determine the feasibility of its removal in a future version of Select2. | comp | deprecate and remove untested attachcontainer decorator this decorator appears to be largely unused a github scan is being performed to catalog the common use cases for this decorator and to determine the feasibility of its removal in a future version of | 1 |
329,948 | 28,315,756,297 | IssuesEvent | 2023-04-10 19:25:43 | SynBioHub/synbiohub3 | https://api.github.com/repos/SynBioHub/synbiohub3 | closed | /setup fails on SBH3 | backend test failure admin | [synbiohub test] test_setup_post starting
http://localhost:7777/setup
SBH1
Setup Successful
(SBH3)
http://localhost:6789/setup
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 143, in post_request
response.raise_for_status()
File "/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 415 Client Error: for url: http://localhost:6789/setup
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_suite.py", line 8, in <module>
test_root.test_root()
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_root.py", line 17, in test_root
firstTimeSetup.test_post()
File "/home/runner/work/synbiohub3/synbiohub3/tests/first_time_setup.py", line 39, in test_post
compare_post_request('setup', setup, headers = {"Accept": "text/plain"}, route_parameters = [], files = None)
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 404, in compare_post_request
compare_request(post_request(request, 1, data, headers, route_parameters, files = files), post_request(request, 3, data, headers, route_parameters, files = files), request, "post request", test_type)
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 146, in post_request
raise HTTPError("Internal server error. Content of response was \n" + response.text)
requests.exceptions.HTTPError: Internal server error. Content of response was | 1.0 | /setup fails on SBH3 - [synbiohub test] test_setup_post starting
http://localhost:7777/setup
SBH1
Setup Successful
(SBH3)
http://localhost:6789/setup
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 143, in post_request
response.raise_for_status()
File "/opt/hostedtoolcache/Python/3.9.16/x64/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 415 Client Error: for url: http://localhost:6789/setup
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_suite.py", line 8, in <module>
test_root.test_root()
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_root.py", line 17, in test_root
firstTimeSetup.test_post()
File "/home/runner/work/synbiohub3/synbiohub3/tests/first_time_setup.py", line 39, in test_post
compare_post_request('setup', setup, headers = {"Accept": "text/plain"}, route_parameters = [], files = None)
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 404, in compare_post_request
compare_request(post_request(request, 1, data, headers, route_parameters, files = files), post_request(request, 3, data, headers, route_parameters, files = files), request, "post request", test_type)
File "/home/runner/work/synbiohub3/synbiohub3/tests/test_functions.py", line 146, in post_request
raise HTTPError("Internal server error. Content of response was \n" + response.text)
requests.exceptions.HTTPError: Internal server error. Content of response was | non_comp | setup fails on test setup post starting setup successful file home runner work tests test functions py line in post request response raise for status file opt hostedtoolcache python lib site packages requests models py line in raise for status raise httperror http error msg response self requests exceptions httperror client error for url during handling of the above exception another exception occurred traceback most recent call last file home runner work tests test suite py line in test root test root file home runner work tests test root py line in test root firsttimesetup test post file home runner work tests first time setup py line in test post compare post request setup setup headers accept text plain route parameters files none file home runner work tests test functions py line in compare post request compare request post request request data headers route parameters files files post request request data headers route parameters files files request post request test type file home runner work tests test functions py line in post request raise httperror internal server error content of response was n response text requests exceptions httperror internal server error content of response was | 0 |
74,778 | 9,806,316,831 | IssuesEvent | 2019-06-12 11:05:07 | ross-rotordynamics/ross | https://api.github.com/repos/ross-rotordynamics/ross | closed | Improvements to ROSS installation guide | documentation | In the ROSS installation instructions, make it clear that the command "pip install"
should be used when opening the **Anaconda prompt**. | 1.0 | Improvements to ROSS installation guide - In the ROSS installation instructions, make it clear that the command "pip install"
should be used when opening the **Anaconda prompt**. | non_comp | improvements to ross installation guide in the ross installation instructions make it clear that the command pip install should be used when opening the anaconda prompt | 0 |
14,799 | 18,234,764,299 | IssuesEvent | 2021-10-01 04:46:34 | inspircd/inspircd | https://api.github.com/repos/inspircd/inspircd | closed | Migration of regex_pcre to PCRE2. | enhancement link compatibility | Currently the `regex_pcre` module uses the original PCRE library. This is no longer the most recent version of PCRE but PCRE2 regular expressions are not 100% compatible with PCRE regular expressions which will break any filters that use it.
The most sensible way to handle this migration seems to be to:
1. [x] Write a separate `regex_pcre2` module for master.
2. [x] Backport `regex_pcre2` to v3 contrib.
3. [x] Move `regex_pcre` to master contrib.
4. [x] Add some kind of compatibility option for `regex_pcre2` to pretend to be `regex_pcre`. | True | Migration of regex_pcre to PCRE2. - Currently the `regex_pcre` module uses the original PCRE library. This is no longer the most recent version of PCRE but PCRE2 regular expressions are not 100% compatible with PCRE regular expressions which will break any filters that use it.
The most sensible way to handle this migration seems to be to:
1. [x] Write a separate `regex_pcre2` module for master.
2. [x] Backport `regex_pcre2` to v3 contrib.
3. [x] Move `regex_pcre` to master contrib.
4. [x] Add some kind of compatibility option for `regex_pcre2` to pretend to be `regex_pcre`. | comp | migration of regex pcre to currently the regex pcre module uses the original pcre library this is no longer the most recent version of pcre but regular expressions are not compatible with pcre regular expressions which will break any filters that use it the most sensible way to handle this migration seems to be to write a separate regex module for master backport regex to contrib move regex pcre to master contrib add some kind of compatibility option for regex to pretend to be regex pcre | 1 |
4,551 | 7,195,060,336 | IssuesEvent | 2018-02-04 13:18:02 | shaarli/Shaarli | https://api.github.com/repos/shaarli/Shaarli | closed | Drop support for legacy PHP v5.5 | php compatibility | PHP 5.5 is [retired (EOL)](https://secure.php.net/eol.php).
Please don't [list it as supported](https://shaarli.readthedocs.io/en/master/Server-requirements/#supported-versions) anymore and allow plugin and Shaarli devs to use more modern stuff :wink: | True | Drop support for legacy PHP v5.5 - PHP 5.5 is [retired (EOL)](https://secure.php.net/eol.php).
Please don't [list it as supported](https://shaarli.readthedocs.io/en/master/Server-requirements/#supported-versions) anymore and allow plugin and Shaarli devs to use more modern stuff :wink: | comp | drop support for legacy php php is please don t anymore and allow plugin and shaarli devs to use more modern stuff wink | 1 |
16,145 | 21,627,708,558 | IssuesEvent | 2022-05-05 05:56:18 | magento/magento2 | https://api.github.com/repos/magento/magento2 | closed | Migration Tool conflict with 2.4.4 (Monolog) | Reproduced on 2.4.x Issue: needs update Area: Framework Area: Upgrades - Upgrade Compatibility Tool Reported on 2.4.4 | <!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
Unable to install the prerelease of Magento 2.4.4 while the Data Migration Tool is installed due to conflicting Monolog packages.
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.4.0) and any important information on the environment where bug is reproducible.
-->
1. Magento 2.4.3 installed
2. magento/data-migration-tool (latest, 2.4.3) installed
3. composer access to prerelease of 2.4.4
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. composer require-commerce magento/product-community-edition 2.4.4 --no-update --interactive-root-conflicts
2. composer update
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. Dependencies should resolve and Magento 2.4.4 and its dependencies should be installed.
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
```
Problem 1
- Root composer.json requires magento/data-migration-tool ^2.4 -> satisfiable by magento/data-migration-tool[2.4.0, 2.4.1, 2.4.2, 2.4.3].
- magento/product-community-edition 2.4.4 requires monolog/monolog ^2.3 -> satisfiable by monolog/monolog[2.3.0, ..., 2.4.0].
- You can only install one version of a package, so only one of these can be installed: monolog/monolog[1.0.0, ..., 1.27.0, 2.0.0, ..., 2.4.0].
- magento/product-enterprise-edition 2.4.4 requires magento/product-community-edition 2.4.4 -> satisfiable by magento/product-community-edition[2.4.4].
- magento/data-migration-tool[2.4.0, ..., 2.4.3] require monolog/monolog ^1.17 -> satisfiable by monolog/monolog[1.17.0, ..., 1.27.0].
- Root composer.json requires magento/product-enterprise-edition 2.4.4 -> satisfiable by magento/product-enterprise-edition[2.4.4].
```
### References
1. First, upgrade process for above was obtained from the Adobe doc [here](https://experienceleague.adobe.com/docs/commerce-operations/upgrade-guide/implementation/perform-upgrade.html).
2. This was caused in [this Pull Request](https://github.com/magento/magento2/pull/33497), while [this Issue](https://github.com/magento/magento2/issues/32868) was being addressed.
4. To resolve, data-migration-tool will need to update to Monolog ^2. A Monolog-written guide on that can be found [here](https://github.com/Seldaek/monolog/blob/main/UPGRADE.md).
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [X] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
| True | Migration Tool conflict with 2.4.4 (Monolog) - <!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
Unable to install the prerelease of Magento 2.4.4 while the Data Migration Tool is installed due to conflicting Monolog packages.
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.4.0) and any important information on the environment where bug is reproducible.
-->
1. Magento 2.4.3 installed
2. magento/data-migration-tool (latest, 2.4.3) installed
3. composer access to prerelease of 2.4.4
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. composer require-commerce magento/product-community-edition 2.4.4 --no-update --interactive-root-conflicts
2. composer update
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. Dependencies should resolve and Magento 2.4.4 and its dependencies should be installed.
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
```
Problem 1
- Root composer.json requires magento/data-migration-tool ^2.4 -> satisfiable by magento/data-migration-tool[2.4.0, 2.4.1, 2.4.2, 2.4.3].
- magento/product-community-edition 2.4.4 requires monolog/monolog ^2.3 -> satisfiable by monolog/monolog[2.3.0, ..., 2.4.0].
- You can only install one version of a package, so only one of these can be installed: monolog/monolog[1.0.0, ..., 1.27.0, 2.0.0, ..., 2.4.0].
- magento/product-enterprise-edition 2.4.4 requires magento/product-community-edition 2.4.4 -> satisfiable by magento/product-community-edition[2.4.4].
- magento/data-migration-tool[2.4.0, ..., 2.4.3] require monolog/monolog ^1.17 -> satisfiable by monolog/monolog[1.17.0, ..., 1.27.0].
- Root composer.json requires magento/product-enterprise-edition 2.4.4 -> satisfiable by magento/product-enterprise-edition[2.4.4].
```
### References
1. First, upgrade process for above was obtained from the Adobe doc [here](https://experienceleague.adobe.com/docs/commerce-operations/upgrade-guide/implementation/perform-upgrade.html).
2. This was caused in [this Pull Request](https://github.com/magento/magento2/pull/33497), while [this Issue](https://github.com/magento/magento2/issues/32868) was being addressed.
4. To resolve, data-migration-tool will need to update to Monolog ^2. A Monolog-written guide on that can be found [here](https://github.com/Seldaek/monolog/blob/main/UPGRADE.md).
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [X] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
| comp | migration tool conflict with monolog please review our guidelines before adding a new issue fields marked with are required please don t remove the template unable to install the prerelease of magento while the data migration tool is installed due to conflicting monolog packages preconditions provide the exact magento version example and any important information on the environment where bug is reproducible magento installed magento data migration tool latest installed composer access to prerelease of steps to reproduce important provide a set of clear steps to reproduce this bug we can not provide support without clear instructions on how to reproduce composer require commerce magento product community edition no update interactive root conflicts composer update expected result dependencies should resolve and magento and its dependencies should be installed actual result problem root composer json requires magento data migration tool satisfiable by magento data migration tool magento product community edition requires monolog monolog satisfiable by monolog monolog you can only install one version of a package so only one of these can be installed monolog monolog magento product enterprise edition requires magento product community edition satisfiable by magento product community edition magento data migration tool require monolog monolog satisfiable by monolog monolog root composer json requires magento product enterprise edition satisfiable by magento product enterprise edition references first upgrade process for above was obtained from the adobe doc this was caused in while was being addressed to resolve data migration tool will need to update to monolog a monolog written guide on that can be found please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability” | 1 |
10,540 | 12,476,377,142 | IssuesEvent | 2020-05-29 13:22:43 | oracle/truffleruby | https://api.github.com/repos/oracle/truffleruby | closed | SyntaxError: unexpected tIDENTIFIER with RHS or-equal assignment | compatibility | This method runs in MRI, but fails with a `SyntaxError` on TruffleRuby:
```
def update_object(obj)
parent = obj.parent ||= fetch_object_by_path obj.parent_path # <-- unexpected tIDENTIFIER
obj.resolve
parent.set_component obj
end
``` | True | SyntaxError: unexpected tIDENTIFIER with RHS or-equal assignment - This method runs in MRI, but fails with a `SyntaxError` on TruffleRuby:
```
def update_object(obj)
parent = obj.parent ||= fetch_object_by_path obj.parent_path # <-- unexpected tIDENTIFIER
obj.resolve
parent.set_component obj
end
``` | comp | syntaxerror unexpected tidentifier with rhs or equal assignment this method runs in mri but fails with a syntaxerror on truffleruby def update object obj parent obj parent fetch object by path obj parent path unexpected tidentifier obj resolve parent set component obj end | 1 |
4,372 | 7,063,984,190 | IssuesEvent | 2018-01-06 00:53:27 | danielbachhuber/gutenberg-plugin-compatibility | https://api.github.com/repos/danielbachhuber/gutenberg-plugin-compatibility | opened | really-simple-ssl | compatible:non-editor state:compatible | Really Simple SSL only adds an admin notice to the editor:

This is garbled in Gutenberg:

| True | really-simple-ssl - Really Simple SSL only adds an admin notice to the editor:

This is garbled in Gutenberg:

| comp | really simple ssl really simple ssl only adds an admin notice to the editor this is garbled in gutenberg | 1 |
18,748 | 26,116,212,848 | IssuesEvent | 2022-12-28 06:33:38 | Fabricators-of-Create/Create | https://api.github.com/repos/Fabricators-of-Create/Create | closed | KubeJS compat and better tags | enhancement compatibility | ### Describe the Suggestion
I was trying to get create to work with other mods' ores, ingots etc. Things like getting the crushing wheel to properly process other ores. All the trial and error to no avail. My concern is that since create (fabric) uses forge tags, it doesn't work at all with anything else besides vanilla. I already assume you guys are aware of this and I'm curious about the plans of mitigation.
### Screenshots and Videos
_No response_
### Additional Context
_No response_ | True | KubeJS compat and better tags - ### Describe the Suggestion
I was trying to get create to work with other mods' ores, ingots etc. Things like getting the crushing wheel to properly process other ores. All the trial and error to no avail. My concern is that since create (fabric) uses forge tags, it doesn't work at all with anything else besides vanilla. I already assume you guys are aware of this and I'm curious about the plans of mitigation.
### Screenshots and Videos
_No response_
### Additional Context
_No response_ | comp | kubejs compat and better tags describe the suggestion i was trying to get create to work with other mods ores ingots etc things like getting the crushing wheel to properly process other ores all the trial and error to no avail my concern is that since create fabric uses forge tags it doesn t work at all with anything else besides vanilla i already assume you guys are aware of this and i m curious about the plans of mitigation screenshots and videos no response additional context no response | 1 |
18,977 | 26,384,603,438 | IssuesEvent | 2023-01-12 11:15:09 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | New ipv4 / ipv6 partition names | backward compatibility | ```sql
create table test ( ipv4 IPv4, ipv6 IPv6 ) Engine MergeTree partition by ipv4 order by ipv4 as select '1.2.3.4', '::ffff:1.2.3.4';
select *, _part from test;
22.12
┌─ipv4────┬─ipv6───────────┬─_part──────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 16909060_1_1_0 │
└─────────┴────────────────┴────────────────┘
22.13
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ bb9df6e3b66e8909ecb142a41a3e6323_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
```
```sql
create table test ( ipv4 IPv4, ipv6 IPv6 ) Engine MergeTree partition by ipv6 order by ipv6 as select '1.2.3.4', '::ffff:1.2.3.4';
select *, _part from test;
22.12
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 1334d7cc23ffb5a5c0262304b3313426_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
22.13
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 45c632749fb22384025ef5d0c817e71e_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
``` | True | New ipv4 / ipv6 partition names - ```sql
create table test ( ipv4 IPv4, ipv6 IPv6 ) Engine MergeTree partition by ipv4 order by ipv4 as select '1.2.3.4', '::ffff:1.2.3.4';
select *, _part from test;
22.12
┌─ipv4────┬─ipv6───────────┬─_part──────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 16909060_1_1_0 │
└─────────┴────────────────┴────────────────┘
22.13
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ bb9df6e3b66e8909ecb142a41a3e6323_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
```
```sql
create table test ( ipv4 IPv4, ipv6 IPv6 ) Engine MergeTree partition by ipv6 order by ipv6 as select '1.2.3.4', '::ffff:1.2.3.4';
select *, _part from test;
22.12
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 1334d7cc23ffb5a5c0262304b3313426_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
22.13
┌─ipv4────┬─ipv6───────────┬─_part──────────────────────────────────┐
│ 1.2.3.4 │ ::ffff:1.2.3.4 │ 45c632749fb22384025ef5d0c817e71e_1_1_0 │
└─────────┴────────────────┴────────────────────────────────────────┘
``` | comp | new partition names sql create table test engine mergetree partition by order by as select ffff select part from test ┌─ ────┬─ ───────────┬─ part──────────┐ │ │ ffff │ │ └─────────┴────────────────┴────────────────┘ ┌─ ────┬─ ───────────┬─ part──────────────────────────────────┐ │ │ ffff │ │ └─────────┴────────────────┴────────────────────────────────────────┘ sql create table test engine mergetree partition by order by as select ffff select part from test ┌─ ────┬─ ───────────┬─ part──────────────────────────────────┐ │ │ ffff │ │ └─────────┴────────────────┴────────────────────────────────────────┘ ┌─ ────┬─ ───────────┬─ part──────────────────────────────────┐ │ │ ffff │ │ └─────────┴────────────────┴────────────────────────────────────────┘ | 1 |
15,118 | 18,984,769,970 | IssuesEvent | 2021-11-21 14:29:45 | WayofTime/BloodMagic | https://api.github.com/repos/WayofTime/BloodMagic | closed | [1.16] Mod Interactions -- Blood Magic Dagger Death X Draconic Evolution Undying Module | bug compatibility 1.16 | #### Issue Description:
User report:
"_When using the Sacrificial Knife with the Blood Altar from Blood Magic while the Undying Module is installed on an equipped Draconic Evolution Chest Piece, you die instead of being kept alive by the module, which activates, and your entire inventory is deleted except for your inventory. No gravestone is made._"
Our modpack dev talked to DE coders and they said this:
"_I suspect blood magic is trying to prevent mods from 'saving you' and it's backfiring somehow.
All i do is cancel the death event._"
I know BM is supposed to kill you as punishment for using the dagger too much, but at the worst you can allow DE's undying (DE always has been overpowered, and it's honestly fine, both your mod and theirs are in almost every modpack).
If this is on your side, I would suggest either that the LP gain is 0 once you don't have any HP or if you want to be evil, give the player your status effect Soul Fray. Just like if they died basically. From a technical point of view, you could consider that they died and got "revived" instantly by Undying Module. I hope that'll help in deciding how to make changes.
Thank you.
We have a report made on DE's side before coming on here:
brandon3055/Draconic-Evolution#1562
You can check that out if you wish, although the info is super similar to here.
Below are versions:
Draconic-Evolution-1.16.5-3.0.16.417-universal.jar
BloodMagic-1.16.4-3.1.5-25.jar (https://www.curseforge.com/minecraft/mc-mods/blood-magic)
gravestone-1.16.5-1.0.7.jar (https://www.curseforge.com/minecraft/mc-mods/gravestone-mod)
No report on Gravestone mod yet, as normally it works just fine.
#### What happens:
You die and you lose everything but Curios slots.
#### What you expected to happen:
Not die, and that a gravestone would generate if the player died.
#### Steps to reproduce:
1. Wear DE Draconic Armor with Undying Module
2. Prick your player with Sacrifial Dagger to generate blood
3. Completely overshoot pricking and attempt to die with it
4. You die even so DE's armor should prevent it
5. No Gravestone gets generated, lose stuff
---IMPORTANT---
See the original modpack issue at
FTBTeam/FTB-Modpack-Issues#156
IT CONTAINS SCREENSHOTS within the procedure being executed by our user, step by step.
____
#### Affected Versions (Do *not* use "latest"):
- BloodMagic: BloodMagic-1.16.4-3.1.5-25.jar
- Minecraft: 1.16.5
- Forge: 1.16.5-forge-36.2.8
- Mod Pack: FTB OceanBlock 1.4.0 | True | [1.16] Mod Interactions -- Blood Magic Dagger Death X Draconic Evolution Undying Module - #### Issue Description:
User report:
"_When using the Sacrificial Knife with the Blood Altar from Blood Magic while the Undying Module is installed on an equipped Draconic Evolution Chest Piece, you die instead of being kept alive by the module, which activates, and your entire inventory is deleted except for your inventory. No gravestone is made._"
Our modpack dev talked to DE coders and they said this:
"_I suspect blood magic is trying to prevent mods from 'saving you' and it's backfiring somehow.
All i do is cancel the death event._"
I know BM is supposed to kill you as punishment for using the dagger too much, but at the worst you can allow DE's undying (DE always has been overpowered, and it's honestly fine, both your mod and theirs are in almost every modpack).
If this is on your side, I would suggest either that the LP gain is 0 once you don't have any HP or if you want to be evil, give the player your status effect Soul Fray. Just like if they died basically. From a technical point of view, you could consider that they died and got "revived" instantly by Undying Module. I hope that'll help in deciding how to make changes.
Thank you.
We have a report made on DE's side before coming on here:
brandon3055/Draconic-Evolution#1562
You can check that out if you wish, although the info is super similar to here.
Below are versions:
Draconic-Evolution-1.16.5-3.0.16.417-universal.jar
BloodMagic-1.16.4-3.1.5-25.jar (https://www.curseforge.com/minecraft/mc-mods/blood-magic)
gravestone-1.16.5-1.0.7.jar (https://www.curseforge.com/minecraft/mc-mods/gravestone-mod)
No report on Gravestone mod yet, as normally it works just fine.
#### What happens:
You die and you lose everything but Curios slots.
#### What you expected to happen:
Not die, and that a gravestone would generate if the player died.
#### Steps to reproduce:
1. Wear DE Draconic Armor with Undying Module
2. Prick your player with Sacrifial Dagger to generate blood
3. Completely overshoot pricking and attempt to die with it
4. You die even so DE's armor should prevent it
5. No Gravestone gets generated, lose stuff
---IMPORTANT---
See the original modpack issue at
FTBTeam/FTB-Modpack-Issues#156
IT CONTAINS SCREENSHOTS within the procedure being executed by our user, step by step.
____
#### Affected Versions (Do *not* use "latest"):
- BloodMagic: BloodMagic-1.16.4-3.1.5-25.jar
- Minecraft: 1.16.5
- Forge: 1.16.5-forge-36.2.8
- Mod Pack: FTB OceanBlock 1.4.0 | comp | mod interactions blood magic dagger death x draconic evolution undying module issue description user report when using the sacrificial knife with the blood altar from blood magic while the undying module is installed on an equipped draconic evolution chest piece you die instead of being kept alive by the module which activates and your entire inventory is deleted except for your inventory no gravestone is made our modpack dev talked to de coders and they said this i suspect blood magic is trying to prevent mods from saving you and it s backfiring somehow all i do is cancel the death event i know bm is supposed to kill you as punishment for using the dagger too much but at the worst you can allow de s undying de always has been overpowered and it s honestly fine both your mod and theirs are in almost every modpack if this is on your side i would suggest either that the lp gain is once you don t have any hp or if you want to be evil give the player your status effect soul fray just like if they died basically from a technical point of view you could consider that they died and got revived instantly by undying module i hope that ll help in deciding how to make changes thank you we have a report made on de s side before coming on here draconic evolution you can check that out if you wish although the info is super similar to here below are versions draconic evolution universal jar bloodmagic jar gravestone jar no report on gravestone mod yet as normally it works just fine what happens you die and you lose everything but curios slots what you expected to happen not die and that a gravestone would generate if the player died steps to reproduce wear de draconic armor with undying module prick your player with sacrifial dagger to generate blood completely overshoot pricking and attempt to die with it you die even so de s armor should prevent it no gravestone gets generated lose stuff important see the original modpack issue at ftbteam ftb modpack issues it contains screenshots within the procedure being executed by our user step by step affected versions do not use latest bloodmagic bloodmagic jar minecraft forge forge mod pack ftb oceanblock | 1 |
6,739 | 9,037,843,701 | IssuesEvent | 2019-02-09 14:46:22 | MeGysssTaa/ReflexIssueTracker | https://api.github.com/repos/MeGysssTaa/ReflexIssueTracker | closed | Denizen support | Compatibility Feature Request | ## General
**ISSUE TYPE**
feature request
> Greeting, my server use the Denizen scripting langague and I would like Reflex to support this plugin :
eg: Everything that is detected/fired/modified by this plugin
Plugin page : https://www.spigotmc.org/resources/denizen.21039/
Regards | True | Denizen support - ## General
**ISSUE TYPE**
feature request
> Greeting, my server use the Denizen scripting langague and I would like Reflex to support this plugin :
eg: Everything that is detected/fired/modified by this plugin
Plugin page : https://www.spigotmc.org/resources/denizen.21039/
Regards | comp | denizen support general issue type feature request > greeting my server use the denizen scripting langague and i would like reflex to support this plugin eg everything that is detected fired modified by this plugin plugin page regards | 1 |
17,668 | 24,356,667,626 | IssuesEvent | 2022-10-03 08:06:42 | opensrp/web | https://api.github.com/repos/opensrp/web | closed | [FHIR Road Map]: User Management package update | Ready for Engineering User management FHIR compatibility | ### Issue Context?
- FHIR web current has a User Management package.
- We are introducing the changes to the UI of the User Management package.
### Issue Implemetation details?
#### User Management
##### User Management display page
- [x] Remove the email column from the table.
##### User Management Add/Edit Page
- [x] The **_Enable User_** field option **_Yes_** should always be selected for new user creation. Currently, **_No_** is always selected.

- [x] Rename the **_Group_** field label to **_KeyCloak User Group_**

#### User management details section
- [x] Update the Linked CareTeams to display the CareTeam names and not the IDs

- [x] Add the KeyCloak UUID display on the section. The label should read **_KeyCloak UUID_**.
- [x] Add the KeyCloak User Group to the display section. The label should read **_KeyCloak User Group_**.
- [x] Re-Order the section details. They should be ordered as follows **_KeyCloak UUID_**, **_KeyCloak User Group_**, **_Practitioner ID_**, **_Practitioner UUID_**, **_Practitioner Staus_**, **_Linked CareTeams_**
#### User Group Management
##### User Group display page
- [x] Align the table headers to the table body. i.e the **_Actions_** column title should be aligned to its body.
##### User Group Role Assingment Page.
- [x] #1096
[Screencast from 09-16-2022 01:33:27 AM.webm](https://user-images.githubusercontent.com/4817531/190537428-fa42f1d7-3051-49d6-b7b2-011d2f7fcf91.webm)
##### User group display page details section
- [x] Update the Members section to display the Firstname, Lastname & Username. The format should be Firstname Lastname (Username)
- [x] Re-Order the section details **Group UUID**, **Name**, **Roles** & **Members**
### Issue Acceptance criteria?
#### User Management
##### User Management display page
- [ ] The email columns should be removed from the table.
##### User Management Add/Edit Page
- [ ] The **_Enable User_** field option **_Yes_** should always be selected for new user creation.
- [ ] The **_Group_** field label should be renamed to **_KeyCloak User Group_**.
#### User management details section
- [ ] The Linked CareTeams should display the CareTeam names and not the IDs
- [ ] The KeyCloak UUID should be displayed in the section. The label should read **_KeyCloak UUID_**.
- [ ] The KeyCloak User Group should be displayed in the section. The label should read **_KeyCloak User Group_**.
- [ ] The section details should be re-ordered. They should be ordered as follows **_KeyCloak UUID_**, **_KeyCloak User Group_**, **_Practitioner ID_**, **_Practitioner UUID_**, **_Practitioner Staus_**, **_Linked CareTeams_**
#### User Group Management
##### User Group display page
- [ ] The table headers should be aligned to the table body. i.e the **_Actions_** column title should be aligned to its body.
##### User Group Role Assingment Page.
- [ ] The User Role selection checkbox should be clickable.
##### User group display page details section
- [ ] The Members section should display the Firstname, Lastname & Username. The format should be Firstname Lastname (Username)
- [ ] The section details should be re-ordered **Group UUID**, **Name**, **Roles** & **Members**
### FHIR resources to be used?
_No response_
### Relevant Information
_No response_ | True | [FHIR Road Map]: User Management package update - ### Issue Context?
- FHIR web current has a User Management package.
- We are introducing the changes to the UI of the User Management package.
### Issue Implemetation details?
#### User Management
##### User Management display page
- [x] Remove the email column from the table.
##### User Management Add/Edit Page
- [x] The **_Enable User_** field option **_Yes_** should always be selected for new user creation. Currently, **_No_** is always selected.

- [x] Rename the **_Group_** field label to **_KeyCloak User Group_**

#### User management details section
- [x] Update the Linked CareTeams to display the CareTeam names and not the IDs

- [x] Add the KeyCloak UUID display on the section. The label should read **_KeyCloak UUID_**.
- [x] Add the KeyCloak User Group to the display section. The label should read **_KeyCloak User Group_**.
- [x] Re-Order the section details. They should be ordered as follows **_KeyCloak UUID_**, **_KeyCloak User Group_**, **_Practitioner ID_**, **_Practitioner UUID_**, **_Practitioner Staus_**, **_Linked CareTeams_**
#### User Group Management
##### User Group display page
- [x] Align the table headers to the table body. i.e the **_Actions_** column title should be aligned to its body.
##### User Group Role Assingment Page.
- [x] #1096
[Screencast from 09-16-2022 01:33:27 AM.webm](https://user-images.githubusercontent.com/4817531/190537428-fa42f1d7-3051-49d6-b7b2-011d2f7fcf91.webm)
##### User group display page details section
- [x] Update the Members section to display the Firstname, Lastname & Username. The format should be Firstname Lastname (Username)
- [x] Re-Order the section details **Group UUID**, **Name**, **Roles** & **Members**
### Issue Acceptance criteria?
#### User Management
##### User Management display page
- [ ] The email columns should be removed from the table.
##### User Management Add/Edit Page
- [ ] The **_Enable User_** field option **_Yes_** should always be selected for new user creation.
- [ ] The **_Group_** field label should be renamed to **_KeyCloak User Group_**.
#### User management details section
- [ ] The Linked CareTeams should display the CareTeam names and not the IDs
- [ ] The KeyCloak UUID should be displayed in the section. The label should read **_KeyCloak UUID_**.
- [ ] The KeyCloak User Group should be displayed in the section. The label should read **_KeyCloak User Group_**.
- [ ] The section details should be re-ordered. They should be ordered as follows **_KeyCloak UUID_**, **_KeyCloak User Group_**, **_Practitioner ID_**, **_Practitioner UUID_**, **_Practitioner Staus_**, **_Linked CareTeams_**
#### User Group Management
##### User Group display page
- [ ] The table headers should be aligned to the table body. i.e the **_Actions_** column title should be aligned to its body.
##### User Group Role Assingment Page.
- [ ] The User Role selection checkbox should be clickable.
##### User group display page details section
- [ ] The Members section should display the Firstname, Lastname & Username. The format should be Firstname Lastname (Username)
- [ ] The section details should be re-ordered **Group UUID**, **Name**, **Roles** & **Members**
### FHIR resources to be used?
_No response_
### Relevant Information
_No response_ | comp | user management package update issue context fhir web current has a user management package we are introducing the changes to the ui of the user management package issue implemetation details user management user management display page remove the email column from the table user management add edit page the enable user field option yes should always be selected for new user creation currently no is always selected rename the group field label to keycloak user group user management details section update the linked careteams to display the careteam names and not the ids add the keycloak uuid display on the section the label should read keycloak uuid add the keycloak user group to the display section the label should read keycloak user group re order the section details they should be ordered as follows keycloak uuid keycloak user group practitioner id practitioner uuid practitioner staus linked careteams user group management user group display page align the table headers to the table body i e the actions column title should be aligned to its body user group role assingment page user group display page details section update the members section to display the firstname lastname username the format should be firstname lastname username re order the section details group uuid name roles members issue acceptance criteria user management user management display page the email columns should be removed from the table user management add edit page the enable user field option yes should always be selected for new user creation the group field label should be renamed to keycloak user group user management details section the linked careteams should display the careteam names and not the ids the keycloak uuid should be displayed in the section the label should read keycloak uuid the keycloak user group should be displayed in the section the label should read keycloak user group the section details should be re ordered they should be ordered as follows keycloak uuid keycloak user group practitioner id practitioner uuid practitioner staus linked careteams user group management user group display page the table headers should be aligned to the table body i e the actions column title should be aligned to its body user group role assingment page the user role selection checkbox should be clickable user group display page details section the members section should display the firstname lastname username the format should be firstname lastname username the section details should be re ordered group uuid name roles members fhir resources to be used no response relevant information no response | 1 |
12,052 | 14,224,536,777 | IssuesEvent | 2020-11-17 19:49:31 | juancarloscp52/BedrockIfy | https://api.github.com/repos/juancarloscp52/BedrockIfy | closed | Mod conflict with Haema | bug mod incompatibility | MC crashes when loading BedrockIfy and Haema mod.
Reported to williambl too - https://github.com/williambl/haema/issues/19
crash report - https://hatebin.com/tbgimqyrle
latest.log - https://hatebin.com/ithepnggws | True | Mod conflict with Haema - MC crashes when loading BedrockIfy and Haema mod.
Reported to williambl too - https://github.com/williambl/haema/issues/19
crash report - https://hatebin.com/tbgimqyrle
latest.log - https://hatebin.com/ithepnggws | comp | mod conflict with haema mc crashes when loading bedrockify and haema mod reported to williambl too crash report latest log | 1 |
3,381 | 6,346,720,339 | IssuesEvent | 2017-07-28 03:33:37 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | opened | Built-in function CHARSET not supported | compatibility rc3.1 | Please answer these questions before submitting your issue. Thanks!
1. What did you do?
```sql
CREATE TABLE t1 (user varchar(255) NOT NULL default '') ENGINE=MyISAM DEFAULT CHARSET=latin1;
INSERT INTO t1 VALUES ('one'),('two');
SELECT CHARSET('a');
```
2. What did you expect to see?
```
+--------------+
| CHARSET('a') |
+--------------+
| utf8 |
+--------------+
1 row in set (0.00 sec)
```
3. What did you see instead?
```
ERROR 1305 (42000): FUNCTION CHARSET does not exist
```
4. What version of TiDB are you using (`tidb-server -V`)?
| True | Built-in function CHARSET not supported - Please answer these questions before submitting your issue. Thanks!
1. What did you do?
```sql
CREATE TABLE t1 (user varchar(255) NOT NULL default '') ENGINE=MyISAM DEFAULT CHARSET=latin1;
INSERT INTO t1 VALUES ('one'),('two');
SELECT CHARSET('a');
```
2. What did you expect to see?
```
+--------------+
| CHARSET('a') |
+--------------+
| utf8 |
+--------------+
1 row in set (0.00 sec)
```
3. What did you see instead?
```
ERROR 1305 (42000): FUNCTION CHARSET does not exist
```
4. What version of TiDB are you using (`tidb-server -V`)?
| comp | built in function charset not supported please answer these questions before submitting your issue thanks what did you do sql create table user varchar not null default engine myisam default charset insert into values one two select charset a what did you expect to see charset a row in set sec what did you see instead error function charset does not exist what version of tidb are you using tidb server v | 1 |
5,766 | 8,214,957,702 | IssuesEvent | 2018-09-05 02:29:12 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | closed | parser failed to parse comment starts with two dash without whitespace but follows by other control character | component/parser type/compatibility | Please answer these questions before submitting your issue. Thanks!
1. What did you do?
If possible, provide a recipe for reproducing the error.
```
--
-- this is a comment.
--
CREATE TABLE `test`.`test_table_with_comment` (`id` int);
```
Note: The is no whitespace after the first `--` and third `--`.
If we use "%q" in Go to format, it will look like:
```
"--\n-- this is a comment.\n--\nCREATE TABLE `test`.`test_table_with_comment` (`id` int);\n"
```
Btw, I test with parser code, not execution in mysql client.
2. What did you expect to see?
```
Query OK, 0 rows affected (0.20 sec)
```
3. What did you see instead?
```
line 1 column 1 near "-
-- this is a comment.
--
CREATE TABLE `test`.`test_table_with_comment` (`id` int);
" (total length 86)
```
4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
418cdf2308c8ebedadbf356142f9afbab9aea111
According to mysql doc: https://dev.mysql.com/doc/refman/5.7/en/ansi-diff-comments.html
>Standard SQL uses “--” as a start-comment sequence. MySQL Server uses # as the start comment character. MySQL Server also supports a variant of the -- comment style. That is, the -- start-comment sequence must be followed by a space (or by a control character such as a newline).
I found that we only support whitespace in `lexer.go`
```
func startWithDash(s *Scanner) (tok int, pos Pos, lit string) {
pos = s.r.pos()
// should we support more control character(space character) ?
if strings.HasPrefix(s.r.s[pos.Offset:], "-- ") {
s.r.incN(3)
s.r.incAsLongAs(func(ch rune) bool {
return ch != '\n'
})
return s.scan()
}
if strings.HasPrefix(s.r.s[pos.Offset:], "->>") {
tok = juss
s.r.incN(3)
return
}
if strings.HasPrefix(s.r.s[pos.Offset:], "->") {
tok = jss
s.r.incN(2)
return
}
tok = int('-')
s.r.inc()
return
}
``` | True | parser failed to parse comment starts with two dash without whitespace but follows by other control character - Please answer these questions before submitting your issue. Thanks!
1. What did you do?
If possible, provide a recipe for reproducing the error.
```
--
-- this is a comment.
--
CREATE TABLE `test`.`test_table_with_comment` (`id` int);
```
Note: The is no whitespace after the first `--` and third `--`.
If we use "%q" in Go to format, it will look like:
```
"--\n-- this is a comment.\n--\nCREATE TABLE `test`.`test_table_with_comment` (`id` int);\n"
```
Btw, I test with parser code, not execution in mysql client.
2. What did you expect to see?
```
Query OK, 0 rows affected (0.20 sec)
```
3. What did you see instead?
```
line 1 column 1 near "-
-- this is a comment.
--
CREATE TABLE `test`.`test_table_with_comment` (`id` int);
" (total length 86)
```
4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
418cdf2308c8ebedadbf356142f9afbab9aea111
According to mysql doc: https://dev.mysql.com/doc/refman/5.7/en/ansi-diff-comments.html
>Standard SQL uses “--” as a start-comment sequence. MySQL Server uses # as the start comment character. MySQL Server also supports a variant of the -- comment style. That is, the -- start-comment sequence must be followed by a space (or by a control character such as a newline).
I found that we only support whitespace in `lexer.go`
```
func startWithDash(s *Scanner) (tok int, pos Pos, lit string) {
pos = s.r.pos()
// should we support more control character(space character) ?
if strings.HasPrefix(s.r.s[pos.Offset:], "-- ") {
s.r.incN(3)
s.r.incAsLongAs(func(ch rune) bool {
return ch != '\n'
})
return s.scan()
}
if strings.HasPrefix(s.r.s[pos.Offset:], "->>") {
tok = juss
s.r.incN(3)
return
}
if strings.HasPrefix(s.r.s[pos.Offset:], "->") {
tok = jss
s.r.incN(2)
return
}
tok = int('-')
s.r.inc()
return
}
``` | comp | parser failed to parse comment starts with two dash without whitespace but follows by other control character please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error this is a comment create table test test table with comment id int note the is no whitespace after the first and third if we use q in go to format it will look like n this is a comment n ncreate table test test table with comment id int n btw i test with parser code not execution in mysql client what did you expect to see query ok rows affected sec what did you see instead line column near this is a comment create table test test table with comment id int total length what version of tidb are you using tidb server v or run select tidb version on tidb according to mysql doc standard sql uses “ ” as a start comment sequence mysql server uses as the start comment character mysql server also supports a variant of the comment style that is the start comment sequence must be followed by a space or by a control character such as a newline i found that we only support whitespace in lexer go func startwithdash s scanner tok int pos pos lit string pos s r pos should we support more control character space character if strings hasprefix s r s s r incn s r incaslongas func ch rune bool return ch n return s scan if strings hasprefix s r s tok juss s r incn return if strings hasprefix s r s tok jss s r incn return tok int s r inc return | 1 |
7,424 | 2,601,762,392 | IssuesEvent | 2015-02-24 00:35:39 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | getRace() reveals "Random" selection | auto-migrated Milestone-Tournament NewFeature Priority-High Type-Enhancement | ```
getRace() will return a player's race even if they have chosen Random and have
not yet been scouted.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 14 Jul 2011 at 4:00 | 1.0 | getRace() reveals "Random" selection - ```
getRace() will return a player's race even if they have chosen Random and have
not yet been scouted.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 14 Jul 2011 at 4:00 | non_comp | getrace reveals random selection getrace will return a player s race even if they have chosen random and have not yet been scouted original issue reported on code google com by aheinerm on jul at | 0 |
179,921 | 21,606,571,012 | IssuesEvent | 2022-05-04 04:25:52 | testdemo1227/Demo-WebGoat1 | https://api.github.com/repos/testdemo1227/Demo-WebGoat1 | opened | jquery-1.10.2.min.js: 4 vulnerabilities (highest severity is: 6.1) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | ❌ |
| [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2015-9251](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jQuery - v3.0.0 | ❌ |
| [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0;jquery-rails - 4.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | True | jquery-1.10.2.min.js: 4 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | ❌ |
| [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2015-9251](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | jQuery - v3.0.0 | ❌ |
| [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.10.2.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary>
### Vulnerable Library - <b>jquery-1.10.2.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: /newDesign/index.html</p>
<p>Path to vulnerable library: /newDesign/assets/js/jquery-1.10.2.min.js,/src/main/webapp/js/jquery/jquery-1.10.2.min.js,/newDesign/assets/js/jquery-1.10.2.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0;jquery-rails - 4.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}},{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/newDesign/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | non_comp | jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file newdesign index html path to vulnerable library newdesign assets js jquery min js src main webapp js jquery jquery min js newdesign assets js jquery min js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium jquery min js direct jquery jquery rails medium jquery min js direct jquery medium jquery min js direct jquery medium jquery min js direct details cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file newdesign index html path to vulnerable library newdesign assets js jquery min js src main webapp js jquery jquery min js newdesign assets js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with whitesource cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file newdesign index html path to vulnerable library newdesign assets js jquery min js src main webapp js jquery jquery min js newdesign assets js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file newdesign index html path to vulnerable library newdesign assets js jquery min js src main webapp js jquery jquery min js newdesign assets js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file newdesign index html path to vulnerable library newdesign assets js jquery min js src main webapp js jquery jquery min js newdesign assets js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery jquery rails isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed vulnerabilityurl istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl | 0 |
202,874 | 15,305,191,154 | IssuesEvent | 2021-02-24 17:48:29 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: tpccbench/nodes=9/cpu=4/chaos/partition failed | C-test-failure O-roachtest O-robot branch-release-20.2 release-blocker | [(roachtest).tpccbench/nodes=9/cpu=4/chaos/partition failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=buildLog) on [release-20.2@8c79e2bc4b35d36c8527f4c40c974f03d9034f46](https://github.com/cockroachdb/cockroach/commits/8c79e2bc4b35d36c8527f4c40c974f03d9034f46):
```
The test failed on branch=release-20.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=9/cpu=4/chaos/partition/run_1
cluster.go:2654,tpcc.go:735,tpcc.go:576,test_runner.go:755: monitor failure: monitor task failed: failed with output "./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)\nError: COMMAND_PROBLEM: exit status 1\n(1) COMMAND_PROBLEM\nWraps: (2) Node 10. Command with error:\n | ```\n | ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}\n | ```\nWraps: (3) exit status 1\nError types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError\n": /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657161-1612856692-131-n10cpu4:10 -- ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}: exit status 20
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2642
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2650
| main.runTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:735
| main.registerTPCCBenchSpec.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:576
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:755
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2698
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.loadTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:663
| [...repeated from below...]
Wraps: (6) failed with output "./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)\nError: COMMAND_PROBLEM: exit status 1\n(1) COMMAND_PROBLEM\nWraps: (2) Node 10. Command with error:\n | ```\n | ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}\n | ```\nWraps: (3) exit status 1\nError types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError\n"
Wraps: (7) attached stack trace
-- stack trace:
| main.execCmdWithBuffer
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:564
| main.(*cluster).RunWithBuffer
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2311
| main.loadTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:662
| main.runTPCCBench.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:733
| main.(*monitor).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2632
| golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/pkg/mod/golang.org/x/sync@v0.0.0-20190911185100-cd5d95a43a6e/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (8) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657161-1612856692-131-n10cpu4:10 -- ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}
Wraps: (9) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.withPrefix (7) *withstack.withStack (8) *errutil.withPrefix (9) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/tpccbench/nodes=9/cpu=4/chaos/partition](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D9%2Fcpu%3D4%2Fchaos%2Fpartition.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: tpccbench/nodes=9/cpu=4/chaos/partition failed - [(roachtest).tpccbench/nodes=9/cpu=4/chaos/partition failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=buildLog) on [release-20.2@8c79e2bc4b35d36c8527f4c40c974f03d9034f46](https://github.com/cockroachdb/cockroach/commits/8c79e2bc4b35d36c8527f4c40c974f03d9034f46):
```
The test failed on branch=release-20.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=9/cpu=4/chaos/partition/run_1
cluster.go:2654,tpcc.go:735,tpcc.go:576,test_runner.go:755: monitor failure: monitor task failed: failed with output "./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)\nError: COMMAND_PROBLEM: exit status 1\n(1) COMMAND_PROBLEM\nWraps: (2) Node 10. Command with error:\n | ```\n | ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}\n | ```\nWraps: (3) exit status 1\nError types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError\n": /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657161-1612856692-131-n10cpu4:10 -- ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}: exit status 20
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2642
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2650
| main.runTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:735
| main.registerTPCCBenchSpec.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:576
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:755
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2698
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.loadTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:663
| [...repeated from below...]
Wraps: (6) failed with output "./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)\nError: COMMAND_PROBLEM: exit status 1\n(1) COMMAND_PROBLEM\nWraps: (2) Node 10. Command with error:\n | ```\n | ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}\n | ```\nWraps: (3) exit status 1\nError types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError\n"
Wraps: (7) attached stack trace
-- stack trace:
| main.execCmdWithBuffer
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:564
| main.(*cluster).RunWithBuffer
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2311
| main.loadTPCCBench
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:662
| main.runTPCCBench.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:733
| main.(*monitor).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2632
| golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/pkg/mod/golang.org/x/sync@v0.0.0-20190911185100-cd5d95a43a6e/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (8) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657161-1612856692-131-n10cpu4:10 -- ./workload run tpcc --warehouses=2000 --workers=2000 --max-rate=1000 --wait=false --duration=16m0s --scatter --tolerate-errors {pgurl:1-9}
Wraps: (9) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.withPrefix (7) *withstack.withStack (8) *errutil.withPrefix (9) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/tpccbench/nodes=9/cpu=4/chaos/partition](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657161&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D9%2Fcpu%3D4%2Fchaos%2Fpartition.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| non_comp | roachtest tpccbench nodes cpu chaos partition failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpccbench nodes cpu chaos partition run cluster go tpcc go tpcc go test runner go monitor failure monitor task failed failed with output workload lib linux gnu libm so version glibc not found required by workload nerror command problem exit status n command problem nwraps node command with error n n workload run tpcc warehouses workers max rate wait false duration scatter tolerate errors pgurl n nwraps exit status nerror types errors cmd hintdetail withdetail exec exiterror n home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run tpcc warehouses workers max rate wait false duration scatter tolerate errors pgurl exit status attached stack trace stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runtpccbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main registertpccbenchspec home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace stack trace main loadtpccbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go wraps failed with output workload lib linux gnu libm so version glibc not found required by workload nerror command problem exit status n command problem nwraps node command with error n n workload run tpcc warehouses workers max rate wait false duration scatter tolerate errors pgurl n nwraps exit status nerror types errors cmd hintdetail withdetail exec exiterror n wraps attached stack trace stack trace main execcmdwithbuffer home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster runwithbuffer home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main loadtpccbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main runtpccbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main monitor go home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go golang org x sync errgroup group go home agent work go pkg mod golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run tpcc warehouses workers max rate wait false duration scatter tolerate errors pgurl wraps exit status error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil withprefix exec exiterror more artifacts powered by | 0 |
17,645 | 24,328,406,789 | IssuesEvent | 2022-09-30 16:53:30 | dotnet/docs | https://api.github.com/repos/dotnet/docs | closed | [Breaking change]: Dropping older framework versions | doc-idea breaking-change Pri1 binary incompatible source incompatible :checkered_flag: Release: .NET 7 in-pr | ### Description
Starting with .NET 7 Preview 1, installing the core libraries packages into projects whose target framework is older than the following, is no longer supported:
- .NET Framework 4.6.2
- .NET 6
- .NET Standard 2.0
### Version
Other (please put exact version in description textbox)
### Previous behavior
The latest non-prerelease core libraries packages which are part of the ".NET 6" wave are supported when used from a project targeting .NET Framework 4.6.1, .NET Core 3.1, or .NET Standard 2.0 or above.
### New behavior
.NET 7 core libraries packages are supported to be used by projects targeting .NET Framework 4.6.2, .NET 6, .NET Standard 2.0 or newer.
### Type of breaking change
- [X] **Binary incompatible**: Existing binaries may encounter a breaking change in behavior, such as failure to load/execute or different run-time behavior.
- [X] **Source incompatible**: Source code may encounter a breaking change in behavior when targeting the new runtime/component/SDK, such as compile errors or different run-time behavior.
### Reason for change
Dropping a framework from a package is a source breaking change. At the same time, continuing to build for all frameworks we ever shipped increases the complexity and size of a package.
Starting with .NET 7, we no longer support frameworks which we previously supported but are now out-of-support:
- .NET Framework 4.6.1
- .NET Core 3.1
- .NET 5
### Recommended action
If your project is no longer being evolved but only maintained, simply don't upgrade the impacted packages. If your project is actively evolved, upgrade it to a later framework version to one of the following:
- .NET Framework 4.6.2
- .NET Core 6
- .NET Standard 2.0
### Feature area
Core .NET libraries
### Affected APIs
These packages no longer ship old frameworks:
- Microsoft.Bcl.AsyncInterfaces
- Microsoft.Extensions.Caching.Abstractions
- Microsoft.Extensions.Caching.Memory
- Microsoft.Extensions.Configuration
- Microsoft.Extensions.Configuration.Abstractions
- Microsoft.Extensions.Configuration.Binder
- Microsoft.Extensions.Configuration.CommandLine
- Microsoft.Extensions.Configuration.EnvironmentVariables
- Microsoft.Extensions.Configuration.FileExtensions
- Microsoft.Extensions.Configuration.Ini
- Microsoft.Extensions.Configuration.Json
- Microsoft.Extensions.Configuration.UserSecrets
- Microsoft.Extensions.Configuration.Xml
- Microsoft.Extensions.DependencyInjection
- Microsoft.Extensions.DependencyInjection.Abstractions
- Microsoft.Extensions.DependencyInjection.Specification.Tests
- Microsoft.Extensions.DependencyModel
- Microsoft.Extensions.FileProviders.Abstractions
- Microsoft.Extensions.FileProviders.Composite
- Microsoft.Extensions.FileProviders.Physical
- Microsoft.Extensions.FileSystemGlobbing
- Microsoft.Extensions.Hosting
- Microsoft.Extensions.Hosting.Abstractions
- Microsoft.Extensions.Hosting.Systemd
- Microsoft.Extensions.Hosting.WindowsServices
- Microsoft.Extensions.Http
- Microsoft.Extensions.Logging
- Microsoft.Extensions.Logging.Abstractions
- Microsoft.Extensions.Logging.Configuration
- Microsoft.Extensions.Logging.Console
- Microsoft.Extensions.Logging.Debug
- Microsoft.Extensions.Logging.EventLog
- Microsoft.Extensions.Logging.EventSource
- Microsoft.Extensions.Logging.TraceSource
- Microsoft.Extensions.Options
- Microsoft.Extensions.Options.ConfigurationExtensions
- Microsoft.Extensions.Options.DataAnnotations
- Microsoft.Extensions.Primitives
- Microsoft.NET.WebAssembly.Threading
- Microsoft.NETCore.Platforms
- Microsoft.Win32.Registry.AccessControl
- Microsoft.Win32.SystemEvents
- Microsoft.Windows.Compatibility
- Microsoft.XmlSerializer.Generator
- System.CodeDom
- System.Collections.Immutable
- System.ComponentModel.Composition
- System.ComponentModel.Composition.Registration
- System.Composition
- System.Composition.AttributedModel
- System.Composition.Convention
- System.Composition.Hosting
- System.Composition.Runtime
- System.Composition.TypedParts
- System.Configuration.ConfigurationManager
- System.Data.Odbc
- System.Data.OleDb
- System.Diagnostics.DiagnosticSource
- System.Diagnostics.EventLog
- System.Diagnostics.PerformanceCounter
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.Drawing.Common
- System.Formats.Asn1
- System.Formats.Cbor
- System.IO.Hashing
- System.IO.Packaging
- System.IO.Pipelines
- System.IO.Ports
- System.Management
- System.Memory.Data
- System.Net.Http.Json
- System.Net.Http.WinHttpHandler
- System.Numerics.Tensors
- System.Reflection.Context
- System.Reflection.Metadata
- System.Reflection.MetadataLoadContext
- System.Resources.Extensions
- System.Runtime.Caching
- System.Runtime.Serialization.Schema
- System.Security.Cryptography.Cose
- System.Security.Cryptography.Pkcs
- System.Security.Cryptography.ProtectedData
- System.Security.Cryptography.Xml
- System.Security.Permissions
- System.ServiceModel.Syndication
- System.ServiceProcess.ServiceController
- System.Speech
- System.Text.Encoding.CodePages
- System.Text.Encodings.Web
- System.Text.Json
- System.Threading.AccessControl
- System.Threading.Channels
- System.Threading.RateLimiting
- System.Threading.Tasks.Dataflow
- System.Windows.Extensions | True | [Breaking change]: Dropping older framework versions - ### Description
Starting with .NET 7 Preview 1, installing the core libraries packages into projects whose target framework is older than the following, is no longer supported:
- .NET Framework 4.6.2
- .NET 6
- .NET Standard 2.0
### Version
Other (please put exact version in description textbox)
### Previous behavior
The latest non-prerelease core libraries packages which are part of the ".NET 6" wave are supported when used from a project targeting .NET Framework 4.6.1, .NET Core 3.1, or .NET Standard 2.0 or above.
### New behavior
.NET 7 core libraries packages are supported to be used by projects targeting .NET Framework 4.6.2, .NET 6, .NET Standard 2.0 or newer.
### Type of breaking change
- [X] **Binary incompatible**: Existing binaries may encounter a breaking change in behavior, such as failure to load/execute or different run-time behavior.
- [X] **Source incompatible**: Source code may encounter a breaking change in behavior when targeting the new runtime/component/SDK, such as compile errors or different run-time behavior.
### Reason for change
Dropping a framework from a package is a source breaking change. At the same time, continuing to build for all frameworks we ever shipped increases the complexity and size of a package.
Starting with .NET 7, we no longer support frameworks which we previously supported but are now out-of-support:
- .NET Framework 4.6.1
- .NET Core 3.1
- .NET 5
### Recommended action
If your project is no longer being evolved but only maintained, simply don't upgrade the impacted packages. If your project is actively evolved, upgrade it to a later framework version to one of the following:
- .NET Framework 4.6.2
- .NET Core 6
- .NET Standard 2.0
### Feature area
Core .NET libraries
### Affected APIs
These packages no longer ship old frameworks:
- Microsoft.Bcl.AsyncInterfaces
- Microsoft.Extensions.Caching.Abstractions
- Microsoft.Extensions.Caching.Memory
- Microsoft.Extensions.Configuration
- Microsoft.Extensions.Configuration.Abstractions
- Microsoft.Extensions.Configuration.Binder
- Microsoft.Extensions.Configuration.CommandLine
- Microsoft.Extensions.Configuration.EnvironmentVariables
- Microsoft.Extensions.Configuration.FileExtensions
- Microsoft.Extensions.Configuration.Ini
- Microsoft.Extensions.Configuration.Json
- Microsoft.Extensions.Configuration.UserSecrets
- Microsoft.Extensions.Configuration.Xml
- Microsoft.Extensions.DependencyInjection
- Microsoft.Extensions.DependencyInjection.Abstractions
- Microsoft.Extensions.DependencyInjection.Specification.Tests
- Microsoft.Extensions.DependencyModel
- Microsoft.Extensions.FileProviders.Abstractions
- Microsoft.Extensions.FileProviders.Composite
- Microsoft.Extensions.FileProviders.Physical
- Microsoft.Extensions.FileSystemGlobbing
- Microsoft.Extensions.Hosting
- Microsoft.Extensions.Hosting.Abstractions
- Microsoft.Extensions.Hosting.Systemd
- Microsoft.Extensions.Hosting.WindowsServices
- Microsoft.Extensions.Http
- Microsoft.Extensions.Logging
- Microsoft.Extensions.Logging.Abstractions
- Microsoft.Extensions.Logging.Configuration
- Microsoft.Extensions.Logging.Console
- Microsoft.Extensions.Logging.Debug
- Microsoft.Extensions.Logging.EventLog
- Microsoft.Extensions.Logging.EventSource
- Microsoft.Extensions.Logging.TraceSource
- Microsoft.Extensions.Options
- Microsoft.Extensions.Options.ConfigurationExtensions
- Microsoft.Extensions.Options.DataAnnotations
- Microsoft.Extensions.Primitives
- Microsoft.NET.WebAssembly.Threading
- Microsoft.NETCore.Platforms
- Microsoft.Win32.Registry.AccessControl
- Microsoft.Win32.SystemEvents
- Microsoft.Windows.Compatibility
- Microsoft.XmlSerializer.Generator
- System.CodeDom
- System.Collections.Immutable
- System.ComponentModel.Composition
- System.ComponentModel.Composition.Registration
- System.Composition
- System.Composition.AttributedModel
- System.Composition.Convention
- System.Composition.Hosting
- System.Composition.Runtime
- System.Composition.TypedParts
- System.Configuration.ConfigurationManager
- System.Data.Odbc
- System.Data.OleDb
- System.Diagnostics.DiagnosticSource
- System.Diagnostics.EventLog
- System.Diagnostics.PerformanceCounter
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.Drawing.Common
- System.Formats.Asn1
- System.Formats.Cbor
- System.IO.Hashing
- System.IO.Packaging
- System.IO.Pipelines
- System.IO.Ports
- System.Management
- System.Memory.Data
- System.Net.Http.Json
- System.Net.Http.WinHttpHandler
- System.Numerics.Tensors
- System.Reflection.Context
- System.Reflection.Metadata
- System.Reflection.MetadataLoadContext
- System.Resources.Extensions
- System.Runtime.Caching
- System.Runtime.Serialization.Schema
- System.Security.Cryptography.Cose
- System.Security.Cryptography.Pkcs
- System.Security.Cryptography.ProtectedData
- System.Security.Cryptography.Xml
- System.Security.Permissions
- System.ServiceModel.Syndication
- System.ServiceProcess.ServiceController
- System.Speech
- System.Text.Encoding.CodePages
- System.Text.Encodings.Web
- System.Text.Json
- System.Threading.AccessControl
- System.Threading.Channels
- System.Threading.RateLimiting
- System.Threading.Tasks.Dataflow
- System.Windows.Extensions | comp | dropping older framework versions description starting with net preview installing the core libraries packages into projects whose target framework is older than the following is no longer supported net framework net net standard version other please put exact version in description textbox previous behavior the latest non prerelease core libraries packages which are part of the net wave are supported when used from a project targeting net framework net core or net standard or above new behavior net core libraries packages are supported to be used by projects targeting net framework net net standard or newer type of breaking change binary incompatible existing binaries may encounter a breaking change in behavior such as failure to load execute or different run time behavior source incompatible source code may encounter a breaking change in behavior when targeting the new runtime component sdk such as compile errors or different run time behavior reason for change dropping a framework from a package is a source breaking change at the same time continuing to build for all frameworks we ever shipped increases the complexity and size of a package starting with net we no longer support frameworks which we previously supported but are now out of support net framework net core net recommended action if your project is no longer being evolved but only maintained simply don t upgrade the impacted packages if your project is actively evolved upgrade it to a later framework version to one of the following net framework net core net standard feature area core net libraries affected apis these packages no longer ship old frameworks microsoft bcl asyncinterfaces microsoft extensions caching abstractions microsoft extensions caching memory microsoft extensions configuration microsoft extensions configuration abstractions microsoft extensions configuration binder microsoft extensions configuration commandline microsoft extensions configuration environmentvariables microsoft extensions configuration fileextensions microsoft extensions configuration ini microsoft extensions configuration json microsoft extensions configuration usersecrets microsoft extensions configuration xml microsoft extensions dependencyinjection microsoft extensions dependencyinjection abstractions microsoft extensions dependencyinjection specification tests microsoft extensions dependencymodel microsoft extensions fileproviders abstractions microsoft extensions fileproviders composite microsoft extensions fileproviders physical microsoft extensions filesystemglobbing microsoft extensions hosting microsoft extensions hosting abstractions microsoft extensions hosting systemd microsoft extensions hosting windowsservices microsoft extensions http microsoft extensions logging microsoft extensions logging abstractions microsoft extensions logging configuration microsoft extensions logging console microsoft extensions logging debug microsoft extensions logging eventlog microsoft extensions logging eventsource microsoft extensions logging tracesource microsoft extensions options microsoft extensions options configurationextensions microsoft extensions options dataannotations microsoft extensions primitives microsoft net webassembly threading microsoft netcore platforms microsoft registry accesscontrol microsoft systemevents microsoft windows compatibility microsoft xmlserializer generator system codedom system collections immutable system componentmodel composition system componentmodel composition registration system composition system composition attributedmodel system composition convention system composition hosting system composition runtime system composition typedparts system configuration configurationmanager system data odbc system data oledb system diagnostics diagnosticsource system diagnostics eventlog system diagnostics performancecounter system directoryservices system directoryservices accountmanagement system directoryservices protocols system drawing common system formats system formats cbor system io hashing system io packaging system io pipelines system io ports system management system memory data system net http json system net http winhttphandler system numerics tensors system reflection context system reflection metadata system reflection metadataloadcontext system resources extensions system runtime caching system runtime serialization schema system security cryptography cose system security cryptography pkcs system security cryptography protecteddata system security cryptography xml system security permissions system servicemodel syndication system serviceprocess servicecontroller system speech system text encoding codepages system text encodings web system text json system threading accesscontrol system threading channels system threading ratelimiting system threading tasks dataflow system windows extensions | 1 |
19,547 | 27,140,802,374 | IssuesEvent | 2023-02-16 16:12:40 | SkyrimLL/Skyrim | https://api.github.com/repos/SkyrimLL/Skyrim | closed | Drinking Fountains of Skyrim for SSE fixes | compatibility SMALL PATCHES | Got tired with broken Drinking Fountain patches with DynDoLod.. making my own patch with direct edits to the main esp file.
| True | Drinking Fountains of Skyrim for SSE fixes - Got tired with broken Drinking Fountain patches with DynDoLod.. making my own patch with direct edits to the main esp file.
| comp | drinking fountains of skyrim for sse fixes got tired with broken drinking fountain patches with dyndolod making my own patch with direct edits to the main esp file | 1 |
505,051 | 14,626,377,750 | IssuesEvent | 2020-12-23 10:14:54 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | spidersweb.pl - site is not usable | browser-fenix engine-gecko priority-normal | <!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64206 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://spidersweb.pl/2020/12/epic-games-store-darmowe-gry-lista.html
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/bd54bb77-9a27-4548-837e-8ea9b8631790.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201220193140</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/3cee0675-b3cb-46f3-8b78-f236268d3227)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | spidersweb.pl - site is not usable - <!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64206 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://spidersweb.pl/2020/12/epic-games-store-darmowe-gry-lista.html
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/bd54bb77-9a27-4548-837e-8ea9b8631790.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201220193140</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/3cee0675-b3cb-46f3-8b78-f236268d3227)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_comp | spidersweb pl site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description page not loading correctly steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
81,742 | 31,478,298,990 | IssuesEvent | 2023-08-30 12:19:27 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Read receipt avatars look a tiny bit off | T-Defect X-Regression S-Tolerable A-Read-Receipts A-Avatar O-Frequent | ### Steps to reproduce
1. Zoom way in on a read receipt to really see it with clarity
### Outcome
#### What did you expect?
I expect read receipts to look the same as they did prior to https://github.com/matrix-org/matrix-react-sdk/pull/11448
#### What happened instead?
They have become smaller due to the above update, but the images displayed inside them remain the same size, which means that the images are now too large for their containers. This is best exemplified by looking at an avatar that has a centered icon:

Compare that to the avatar as seen next to a message:

### Operating system
NixOS unstable
### Browser information
Firefox 115.0.2
### URL for webapp
develop.element.io
### Application version
Element version: fcd805255c26-react-64f762d7b0f3-js-5788d9744b76 Olm version: 3.2.14
### Homeserver
Not relevant
### Will you send logs?
No | 1.0 | Read receipt avatars look a tiny bit off - ### Steps to reproduce
1. Zoom way in on a read receipt to really see it with clarity
### Outcome
#### What did you expect?
I expect read receipts to look the same as they did prior to https://github.com/matrix-org/matrix-react-sdk/pull/11448
#### What happened instead?
They have become smaller due to the above update, but the images displayed inside them remain the same size, which means that the images are now too large for their containers. This is best exemplified by looking at an avatar that has a centered icon:

Compare that to the avatar as seen next to a message:

### Operating system
NixOS unstable
### Browser information
Firefox 115.0.2
### URL for webapp
develop.element.io
### Application version
Element version: fcd805255c26-react-64f762d7b0f3-js-5788d9744b76 Olm version: 3.2.14
### Homeserver
Not relevant
### Will you send logs?
No | non_comp | read receipt avatars look a tiny bit off steps to reproduce zoom way in on a read receipt to really see it with clarity outcome what did you expect i expect read receipts to look the same as they did prior to what happened instead they have become smaller due to the above update but the images displayed inside them remain the same size which means that the images are now too large for their containers this is best exemplified by looking at an avatar that has a centered icon compare that to the avatar as seen next to a message operating system nixos unstable browser information firefox url for webapp develop element io application version element version react js olm version homeserver not relevant will you send logs no | 0 |
151,362 | 13,399,745,356 | IssuesEvent | 2020-09-03 14:53:23 | UnBArqDsw/2020.1_G7_TCM | https://api.github.com/repos/UnBArqDsw/2020.1_G7_TCM | opened | [DOC]Design Sprint | documentation | ### Descrição
Elaboração do documento de design sprint
### Coisas a se fazer
- [ ] Desenvolvimento do documento de design sprint contendo o planejamento e fases. | 1.0 | [DOC]Design Sprint - ### Descrição
Elaboração do documento de design sprint
### Coisas a se fazer
- [ ] Desenvolvimento do documento de design sprint contendo o planejamento e fases. | non_comp | design sprint descrição elaboração do documento de design sprint coisas a se fazer desenvolvimento do documento de design sprint contendo o planejamento e fases | 0 |
10,187 | 4,717,163,259 | IssuesEvent | 2016-10-16 13:31:19 | DMSC-Instrument-Data/plankton | https://api.github.com/repos/DMSC-Instrument-Data/plankton | opened | Introduce "system tests" | build tools enhancement unit tests | We should think of a good way to implement system tests, meaning tests that involve communication with a device in a certain setup using the device protocol. We might do this either by spinning up a docker container with the device running in the configuration that is to be tested.
We could use the python unit test framework and combine it with something like [docker-py](https://docker-py.readthedocs.io/en/latest/) to start containers etc. | 1.0 | Introduce "system tests" - We should think of a good way to implement system tests, meaning tests that involve communication with a device in a certain setup using the device protocol. We might do this either by spinning up a docker container with the device running in the configuration that is to be tested.
We could use the python unit test framework and combine it with something like [docker-py](https://docker-py.readthedocs.io/en/latest/) to start containers etc. | non_comp | introduce system tests we should think of a good way to implement system tests meaning tests that involve communication with a device in a certain setup using the device protocol we might do this either by spinning up a docker container with the device running in the configuration that is to be tested we could use the python unit test framework and combine it with something like to start containers etc | 0 |
18,830 | 26,178,087,933 | IssuesEvent | 2023-01-02 12:19:24 | spring-projects-experimental/spring-native | https://api.github.com/repos/spring-projects-experimental/spring-native | closed | Introduce initial caching support | type: compatibility for: external-project | We have basic configuration for caching in `CachingHints`, but EhCache is not supported, and enabling caching support in PetClinic JPA with the following configuration:
```java
@Configuration
@EnableCaching
class CacheConfiguration {
@Bean
public JCacheManagerCustomizer petclinicCacheConfigurationCustomizer() {
return cm -> {
cm.createCache("vets", cacheConfiguration());
};
}
private javax.cache.configuration.Configuration<Object, Object> cacheConfiguration() {
return new MutableConfiguration<>().setStatisticsEnabled(true);
}
}
```
Generates this error that seems to indicate the need for a processor:
```
org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'vetController' defined in class path resource [org/springframework/samples/petclinic/vet/VetController.class]: Unsatisfied dependency expressed through constructor parameter 0; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'vetRepository': Post-processing of FactoryBean's singleton object failed; nested exception is com.oracle.svm.core.jdk.UnsupportedFeatureError: Proxy class defined by interfaces [interface org.springframework.samples.petclinic.vet.VetRepository, interface org.springframework.data.repository.Repository, interface org.springframework.transaction.interceptor.TransactionalProxy, interface org.springframework.aop.framework.Advised, interface org.springframework.core.DecoratingProxy, interface java.io.Serializable] not found. Generating proxy classes at runtime is not supported. Proxy classes need to be defined at image build time by specifying the list of interfaces that they implement. To define proxy classes use -H:DynamicProxyConfigurationFiles=<comma-separated-config-files> and -H:DynamicProxyConfigurationResources=<comma-separated-config-resources> options.
```
We should also document what kind of caching is supported (see the list in `CacheType`) and bring back caching support in the `petclinic-jpa` sample as configured in https://github.com/spring-projects/spring-petclinic. | True | Introduce initial caching support - We have basic configuration for caching in `CachingHints`, but EhCache is not supported, and enabling caching support in PetClinic JPA with the following configuration:
```java
@Configuration
@EnableCaching
class CacheConfiguration {
@Bean
public JCacheManagerCustomizer petclinicCacheConfigurationCustomizer() {
return cm -> {
cm.createCache("vets", cacheConfiguration());
};
}
private javax.cache.configuration.Configuration<Object, Object> cacheConfiguration() {
return new MutableConfiguration<>().setStatisticsEnabled(true);
}
}
```
Generates this error that seems to indicate the need for a processor:
```
org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'vetController' defined in class path resource [org/springframework/samples/petclinic/vet/VetController.class]: Unsatisfied dependency expressed through constructor parameter 0; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'vetRepository': Post-processing of FactoryBean's singleton object failed; nested exception is com.oracle.svm.core.jdk.UnsupportedFeatureError: Proxy class defined by interfaces [interface org.springframework.samples.petclinic.vet.VetRepository, interface org.springframework.data.repository.Repository, interface org.springframework.transaction.interceptor.TransactionalProxy, interface org.springframework.aop.framework.Advised, interface org.springframework.core.DecoratingProxy, interface java.io.Serializable] not found. Generating proxy classes at runtime is not supported. Proxy classes need to be defined at image build time by specifying the list of interfaces that they implement. To define proxy classes use -H:DynamicProxyConfigurationFiles=<comma-separated-config-files> and -H:DynamicProxyConfigurationResources=<comma-separated-config-resources> options.
```
We should also document what kind of caching is supported (see the list in `CacheType`) and bring back caching support in the `petclinic-jpa` sample as configured in https://github.com/spring-projects/spring-petclinic. | comp | introduce initial caching support we have basic configuration for caching in cachinghints but ehcache is not supported and enabling caching support in petclinic jpa with the following configuration java configuration enablecaching class cacheconfiguration bean public jcachemanagercustomizer petcliniccacheconfigurationcustomizer return cm cm createcache vets cacheconfiguration private javax cache configuration configuration cacheconfiguration return new mutableconfiguration setstatisticsenabled true generates this error that seems to indicate the need for a processor org springframework beans factory unsatisfieddependencyexception error creating bean with name vetcontroller defined in class path resource unsatisfied dependency expressed through constructor parameter nested exception is org springframework beans factory beancreationexception error creating bean with name vetrepository post processing of factorybean s singleton object failed nested exception is com oracle svm core jdk unsupportedfeatureerror proxy class defined by interfaces not found generating proxy classes at runtime is not supported proxy classes need to be defined at image build time by specifying the list of interfaces that they implement to define proxy classes use h dynamicproxyconfigurationfiles and h dynamicproxyconfigurationresources options we should also document what kind of caching is supported see the list in cachetype and bring back caching support in the petclinic jpa sample as configured in | 1 |
763,513 | 26,760,699,303 | IssuesEvent | 2023-01-31 06:30:41 | dkschdcks/repo-setup-sample | https://api.github.com/repos/dkschdcks/repo-setup-sample | closed | Sample Backlog1 | For: CI/CD Priority: Medium Type: idea Status: Available | ##Description
프로젝트 작업 전 수행
## Tasks
- [ ] 저녁먹기
- [ ] 약먹기
- [ ] AWS 다시 반복
## References
- [link test](link address)
| 1.0 | Sample Backlog1 - ##Description
프로젝트 작업 전 수행
## Tasks
- [ ] 저녁먹기
- [ ] 약먹기
- [ ] AWS 다시 반복
## References
- [link test](link address)
| non_comp | sample description 프로젝트 작업 전 수행 tasks 저녁먹기 약먹기 aws 다시 반복 references link address | 0 |
172,333 | 21,043,367,872 | IssuesEvent | 2022-03-31 14:07:21 | jgeraigery/dynatrace-service-broker | https://api.github.com/repos/jgeraigery/dynatrace-service-broker | closed | CVE-2017-5664 (High) detected in tomcat-embed-core-8.0.15.jar - autoclosed | security vulnerability | ## CVE-2017-5664 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.0.15.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.0.15/tomcat-embed-core-8.0.15.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.1.9.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.1.9.RELEASE.jar
- :x: **tomcat-embed-core-8.0.15.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.
<p>Publish Date: 2017-06-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p>
<p>Release Date: 2017-06-06</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.0.15","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.1.9.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.1.9.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.0.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-5664","vulnerabilityDetails":"The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-5664 (High) detected in tomcat-embed-core-8.0.15.jar - autoclosed - ## CVE-2017-5664 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.0.15.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.0.15/tomcat-embed-core-8.0.15.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.1.9.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.1.9.RELEASE.jar
- :x: **tomcat-embed-core-8.0.15.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.
<p>Publish Date: 2017-06-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p>
<p>Release Date: 2017-06-06</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.0.15","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.1.9.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.1.9.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.0.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-5664","vulnerabilityDetails":"The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_comp | cve high detected in tomcat embed core jar autoclosed cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details the error page mechanism of the java servlet specification requires that when an error occurs and an error page is configured for the error that occurred the original request and response are forwarded to the error page this means that the request is presented to the error page with the original http method if the error page is a static file expected behaviour is to serve content of the file as if processing a get request regardless of the actual http method the default servlet in apache tomcat to to to and to did not do this depending on the original request this could lead to unexpected and undesirable results for static error pages including if the defaultservlet is configured to permit writes the replacement or removal of the custom error page notes for other user provided error pages unless explicitly coded otherwise jsps ignore the http method jsps used as error pages must must ensure that they handle any error dispatch as a get request regardless of the actual method by default the response generated by a servlet does depend on the http method custom servlets used as error pages must ensure that they handle any error dispatch as a get request regardless of the actual method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the error page mechanism of the java servlet specification requires that when an error occurs and an error page is configured for the error that occurred the original request and response are forwarded to the error page this means that the request is presented to the error page with the original http method if the error page is a static file expected behaviour is to serve content of the file as if processing a get request regardless of the actual http method the default servlet in apache tomcat to to to and to did not do this depending on the original request this could lead to unexpected and undesirable results for static error pages including if the defaultservlet is configured to permit writes the replacement or removal of the custom error page notes for other user provided error pages unless explicitly coded otherwise jsps ignore the http method jsps used as error pages must must ensure that they handle any error dispatch as a get request regardless of the actual method by default the response generated by a servlet does depend on the http method custom servlets used as error pages must ensure that they handle any error dispatch as a get request regardless of the actual method vulnerabilityurl | 0 |
5,723 | 8,182,303,128 | IssuesEvent | 2018-08-29 04:12:31 | summernote/summernote | https://api.github.com/repos/summernote/summernote | closed | v0.6.0 Cannot upload image on IE9 | - Category: Compatibility | When use summernote v0.6.0 on IE9 upload image , isn't supported .
I try to upload image from my local directory and not work "Insert Image" button. I use IE 9, get this error:
Error: unable to get the value of attribute 'length': object is null or undefined
| True | v0.6.0 Cannot upload image on IE9 - When use summernote v0.6.0 on IE9 upload image , isn't supported .
I try to upload image from my local directory and not work "Insert Image" button. I use IE 9, get this error:
Error: unable to get the value of attribute 'length': object is null or undefined
| comp | cannot upload image on when use summernote on upload image isn t supported i try to upload image from my local directory and not work insert image button i use ie get this error error unable to get the value of attribute length object is null or undefined | 1 |
19,777 | 27,425,535,884 | IssuesEvent | 2023-03-01 19:59:35 | openxla/stablehlo | https://api.github.com/repos/openxla/stablehlo | opened | Consider providing source compatibility | Compatibility | To quote the compatibility RFC: "(G3) Source compatibility for C, C++ and Python APIs within libStablehlo is an aspirational goal. At the moment, we don't offer source compatibility guarantees, but please let us know if this is an important use case for you, and we can have a discussion about supporting it". | True | Consider providing source compatibility - To quote the compatibility RFC: "(G3) Source compatibility for C, C++ and Python APIs within libStablehlo is an aspirational goal. At the moment, we don't offer source compatibility guarantees, but please let us know if this is an important use case for you, and we can have a discussion about supporting it". | comp | consider providing source compatibility to quote the compatibility rfc source compatibility for c c and python apis within libstablehlo is an aspirational goal at the moment we don t offer source compatibility guarantees but please let us know if this is an important use case for you and we can have a discussion about supporting it | 1 |
713,850 | 24,541,526,787 | IssuesEvent | 2022-10-12 04:30:35 | googleapis/google-cloud-java | https://api.github.com/repos/googleapis/google-cloud-java | closed | Every module duplicates `<reporting>` configuration | type: bug priority: p2 | @lqiu96 Can we move it up to the parent pom? Would javadoc generation break? | 1.0 | Every module duplicates `<reporting>` configuration - @lqiu96 Can we move it up to the parent pom? Would javadoc generation break? | non_comp | every module duplicates configuration can we move it up to the parent pom would javadoc generation break | 0 |
1,032 | 3,487,460,680 | IssuesEvent | 2016-01-01 22:55:53 | AdguardTeam/AdguardForWindows | https://api.github.com/repos/AdguardTeam/AdguardForWindows | opened | Check all AG files on virustotal and report false positives | Compatibility | There are a few false positives right now:
https://a.disquscdn.com/uploads/mediaembed/images/3006/1475/original.jpg
https://a.disquscdn.com/uploads/mediaembed/images/3005/772/original.jpg | True | Check all AG files on virustotal and report false positives - There are a few false positives right now:
https://a.disquscdn.com/uploads/mediaembed/images/3006/1475/original.jpg
https://a.disquscdn.com/uploads/mediaembed/images/3005/772/original.jpg | comp | check all ag files on virustotal and report false positives there are a few false positives right now | 1 |
610 | 3,025,519,228 | IssuesEvent | 2015-08-03 09:12:21 | jesus2099/konami-command | https://api.github.com/repos/jesus2099/konami-command | closed | MASS MERGE on Safari : Error on dynamic updater | browser compatibility | [Cheezmo](https://musicbrainz.org/user/Cheezmo) reported, on Safari 9 NinjaKit 0.9.1 (Mac OS) the locking of the dynamic updater on removal. It happens either on *OFF* or *Remove from…* clicks, when it comes to check an artist, after having successfully checked labels.
> Refreshing memory…
> Checking release-group [2b98e6d7-a521-332f-961e-d281ba33ba3d](https://musicbrainz.org/release-group/2b98e6d7-a521-332f-961e-d281ba33ba3d) against all its releases…
> still used: keeping.
> Checking label [A&M Records](https://musicbrainz.org/label/35515729-1f2c-4cc9-9390-9af2764bc56c) against all its releases…
> (several pages but this check will stop as soon as it finds something)
> Checking label A&M Records against all its releases (page 2)…
> Checking label A&M Records against all its releases (page 3)…
> Checking label A&M Records against all its releases (page 4)…
> Checking label A&M Records against all its releases (page 5)…
> Checking label A&M Records against all its releases (page 6)…
> Checking label A&M Records against all its releases (page 7)…
> Checking label A&M Records against all its releases (page 8)…
> Checking label A&M Records against all its releases (page 9)…
> Checking label A&M Records against all its releases (page 10)…
> Checking label A&M Records against all its releases (page 11)…
> Checking label A&M Records against all its releases (page 12)…
> Checking label A&M Records against all its releases (page 13)…
> Checking label A&M Records against all its releases (page 14)…
> Checking label A&M Records against all its releases (page 15)…
> Checking label A&M Records against all its releases (page 16)…
> still used: keeping.
> Checking artist [The Police](https://musicbrainz.org/artist/9e0e2b01-41db-4008-bd8b-988977d6019a) against all its recordings…
onreadystatechangeTypeError:
null is not an object (evaluating 'stuff[checkAgainst].rawids.indexOf')
| True | MASS MERGE on Safari : Error on dynamic updater - [Cheezmo](https://musicbrainz.org/user/Cheezmo) reported, on Safari 9 NinjaKit 0.9.1 (Mac OS) the locking of the dynamic updater on removal. It happens either on *OFF* or *Remove from…* clicks, when it comes to check an artist, after having successfully checked labels.
> Refreshing memory…
> Checking release-group [2b98e6d7-a521-332f-961e-d281ba33ba3d](https://musicbrainz.org/release-group/2b98e6d7-a521-332f-961e-d281ba33ba3d) against all its releases…
> still used: keeping.
> Checking label [A&M Records](https://musicbrainz.org/label/35515729-1f2c-4cc9-9390-9af2764bc56c) against all its releases…
> (several pages but this check will stop as soon as it finds something)
> Checking label A&M Records against all its releases (page 2)…
> Checking label A&M Records against all its releases (page 3)…
> Checking label A&M Records against all its releases (page 4)…
> Checking label A&M Records against all its releases (page 5)…
> Checking label A&M Records against all its releases (page 6)…
> Checking label A&M Records against all its releases (page 7)…
> Checking label A&M Records against all its releases (page 8)…
> Checking label A&M Records against all its releases (page 9)…
> Checking label A&M Records against all its releases (page 10)…
> Checking label A&M Records against all its releases (page 11)…
> Checking label A&M Records against all its releases (page 12)…
> Checking label A&M Records against all its releases (page 13)…
> Checking label A&M Records against all its releases (page 14)…
> Checking label A&M Records against all its releases (page 15)…
> Checking label A&M Records against all its releases (page 16)…
> still used: keeping.
> Checking artist [The Police](https://musicbrainz.org/artist/9e0e2b01-41db-4008-bd8b-988977d6019a) against all its recordings…
onreadystatechangeTypeError:
null is not an object (evaluating 'stuff[checkAgainst].rawids.indexOf')
| comp | mass merge on safari error on dynamic updater reported on safari ninjakit mac os the locking of the dynamic updater on removal it happens either on off or remove from… clicks when it comes to check an artist after having successfully checked labels refreshing memory… checking release group against all its releases… still used keeping checking label against all its releases… several pages but this check will stop as soon as it finds something checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … checking label a m records against all its releases page … still used keeping checking artist against all its recordings… onreadystatechangetypeerror null is not an object evaluating stuff rawids indexof | 1 |
10,138 | 12,148,596,745 | IssuesEvent | 2020-04-24 14:49:59 | bitkeks/python-netflow-v9-softflowd | https://api.github.com/repos/bitkeks/python-netflow-v9-softflowd | closed | Fails to run on python 3.7 due to lru_cache syntax | bug python/compatibility | Both pip install 0.10.2 and from git clone fails to run python3 -m netflow.collector on python 3.7 with exception
$ python -m netflow.collector -h
Traceback (most recent call last):
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\runpy.py", line 183, in _run_module_as_main
mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\runpy.py", line 109, in _get_module_details
__import__(pkg_name)
File "C:\techno\git\python-netflow-v9-softflowd\netflow\__init__.py", line 10, in <module>
from .utils import parse_packet
File "C:\techno\git\python-netflow-v9-softflowd\netflow\utils.py", line 16, in <module>
from .ipfix import IPFIXExportPacket
File "C:\techno\git\python-netflow-v9-softflowd\netflow\ipfix.py", line 526, in <module>
class IPFIXDataTypes:
File "C:\techno\git\python-netflow-v9-softflowd\netflow\ipfix.py", line 559, in IPFIXDataTypes
def by_name(cls, key: str) -> Optional[DataType]:
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\functools.py", line 477, in lru_cache
raise TypeError('Expected maxsize to be an integer or None')
TypeError: Expected maxsize to be an integer or None
The root cause is lru_cache before 3.8 has to have at least to be "lru_cache()".
Maybe should bump up the dependency to 3.8 from 3.5 in the setup.py | True | Fails to run on python 3.7 due to lru_cache syntax - Both pip install 0.10.2 and from git clone fails to run python3 -m netflow.collector on python 3.7 with exception
$ python -m netflow.collector -h
Traceback (most recent call last):
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\runpy.py", line 183, in _run_module_as_main
mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\runpy.py", line 109, in _get_module_details
__import__(pkg_name)
File "C:\techno\git\python-netflow-v9-softflowd\netflow\__init__.py", line 10, in <module>
from .utils import parse_packet
File "C:\techno\git\python-netflow-v9-softflowd\netflow\utils.py", line 16, in <module>
from .ipfix import IPFIXExportPacket
File "C:\techno\git\python-netflow-v9-softflowd\netflow\ipfix.py", line 526, in <module>
class IPFIXDataTypes:
File "C:\techno\git\python-netflow-v9-softflowd\netflow\ipfix.py", line 559, in IPFIXDataTypes
def by_name(cls, key: str) -> Optional[DataType]:
File "C:\Users\dhristov\AppData\Local\Continuum\anaconda3\envs\py37\lib\functools.py", line 477, in lru_cache
raise TypeError('Expected maxsize to be an integer or None')
TypeError: Expected maxsize to be an integer or None
The root cause is lru_cache before 3.8 has to have at least to be "lru_cache()".
Maybe should bump up the dependency to 3.8 from 3.5 in the setup.py | comp | fails to run on python due to lru cache syntax both pip install and from git clone fails to run m netflow collector on python with exception python m netflow collector h traceback most recent call last file c users dhristov appdata local continuum envs lib runpy py line in run module as main mod name mod spec code get module details mod name error file c users dhristov appdata local continuum envs lib runpy py line in get module details import pkg name file c techno git python netflow softflowd netflow init py line in from utils import parse packet file c techno git python netflow softflowd netflow utils py line in from ipfix import ipfixexportpacket file c techno git python netflow softflowd netflow ipfix py line in class ipfixdatatypes file c techno git python netflow softflowd netflow ipfix py line in ipfixdatatypes def by name cls key str optional file c users dhristov appdata local continuum envs lib functools py line in lru cache raise typeerror expected maxsize to be an integer or none typeerror expected maxsize to be an integer or none the root cause is lru cache before has to have at least to be lru cache maybe should bump up the dependency to from in the setup py | 1 |
17,590 | 24,259,836,251 | IssuesEvent | 2022-09-27 21:21:05 | MetaMask/metamask-extension | https://api.github.com/repos/MetaMask/metamask-extension | closed | Get working on Android Firefox | type-enhancement browser-firefox mobile-compatibility | Android Firefox has extension support, so this should be feasible, but apparently doesn't work yet. | True | Get working on Android Firefox - Android Firefox has extension support, so this should be feasible, but apparently doesn't work yet. | comp | get working on android firefox android firefox has extension support so this should be feasible but apparently doesn t work yet | 1 |
283,282 | 21,316,084,413 | IssuesEvent | 2022-04-16 09:49:25 | shunyao643/pe | https://api.github.com/repos/shunyao643/pe | opened | Extra comma in Display Monthly Schedule explanation (UG, page 15) | severity.VeryLow type.DocumentationBug | The comma does not add any value.

<!--session: 1650095847074-c9d211e4-3808-474e-9cbf-de43abb79c19-->
<!--Version: Web v3.4.2--> | 1.0 | Extra comma in Display Monthly Schedule explanation (UG, page 15) - The comma does not add any value.

<!--session: 1650095847074-c9d211e4-3808-474e-9cbf-de43abb79c19-->
<!--Version: Web v3.4.2--> | non_comp | extra comma in display monthly schedule explanation ug page the comma does not add any value | 0 |
409,097 | 11,956,669,389 | IssuesEvent | 2020-04-04 11:35:26 | Thorium-Sim/thorium | https://api.github.com/repos/Thorium-Sim/thorium | opened | WASD Move-able videos for Tactical Maps | priority/low type/feature | ### Requested By: Maeson Busk (Maesonbusk@gmail.com)
### Priority: Low
### Version: 2.8.0
Could we have the possibility to make videos controllable by the WASD keys on tactical maps? Ultimately, it would be cool to be able to link a video moving to thruster inputs, similar to objects.
Why?
We could create graphics (like a nebula path map) that are animated, that move. The crew's ship is static, in the center of the screen. Moving the video allows the crew to fly through the nebula, without seeing the whole thing. This is similar to the old "Canada-esque" wormhole videos, where they had to steer through the top to bottom scrolling video, but it would be legitimately controlled.
Also, with a large enough 3d asteroid field, we could have the crew's ship in the middle, with asteroids coming at the camera. The video could move, thus allowing the crew to dodge asteroids. I realize the video would have to be quite large, to prevent the crew from just continuously flying down, and getting "out of bounds"
I know this is crazy, but I woke up from a dream at 5:30 in the morning and this came to mind.
Also, I understand that we will get real 3D navigation at some point. This just seems like a cool way to revolutionize how we do graphics currently.
| 1.0 | WASD Move-able videos for Tactical Maps - ### Requested By: Maeson Busk (Maesonbusk@gmail.com)
### Priority: Low
### Version: 2.8.0
Could we have the possibility to make videos controllable by the WASD keys on tactical maps? Ultimately, it would be cool to be able to link a video moving to thruster inputs, similar to objects.
Why?
We could create graphics (like a nebula path map) that are animated, that move. The crew's ship is static, in the center of the screen. Moving the video allows the crew to fly through the nebula, without seeing the whole thing. This is similar to the old "Canada-esque" wormhole videos, where they had to steer through the top to bottom scrolling video, but it would be legitimately controlled.
Also, with a large enough 3d asteroid field, we could have the crew's ship in the middle, with asteroids coming at the camera. The video could move, thus allowing the crew to dodge asteroids. I realize the video would have to be quite large, to prevent the crew from just continuously flying down, and getting "out of bounds"
I know this is crazy, but I woke up from a dream at 5:30 in the morning and this came to mind.
Also, I understand that we will get real 3D navigation at some point. This just seems like a cool way to revolutionize how we do graphics currently.
| non_comp | wasd move able videos for tactical maps requested by maeson busk maesonbusk gmail com priority low version could we have the possibility to make videos controllable by the wasd keys on tactical maps ultimately it would be cool to be able to link a video moving to thruster inputs similar to objects why we could create graphics like a nebula path map that are animated that move the crew s ship is static in the center of the screen moving the video allows the crew to fly through the nebula without seeing the whole thing this is similar to the old canada esque wormhole videos where they had to steer through the top to bottom scrolling video but it would be legitimately controlled also with a large enough asteroid field we could have the crew s ship in the middle with asteroids coming at the camera the video could move thus allowing the crew to dodge asteroids i realize the video would have to be quite large to prevent the crew from just continuously flying down and getting out of bounds i know this is crazy but i woke up from a dream at in the morning and this came to mind also i understand that we will get real navigation at some point this just seems like a cool way to revolutionize how we do graphics currently | 0 |
14,267 | 17,144,768,949 | IssuesEvent | 2021-07-13 13:34:47 | wp-media/imagify-plugin | https://api.github.com/repos/wp-media/imagify-plugin | closed | Imagify replaces img tags inside existing picture tags resulting in nested picture tags | 3rd party compatibility effort [S] severity: major type: bug | **Describe the bug**
When webp images display option with picture tags is used from the settings and theme already has picture tags present, then Imagify replaces img tags inside existing picture tags resulting in nested <picture> tags.
**Expected behavior**
Imagify should ignore already existing picture tags with webp images.
**Screenshots**

**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| True | Imagify replaces img tags inside existing picture tags resulting in nested picture tags - **Describe the bug**
When webp images display option with picture tags is used from the settings and theme already has picture tags present, then Imagify replaces img tags inside existing picture tags resulting in nested <picture> tags.
**Expected behavior**
Imagify should ignore already existing picture tags with webp images.
**Screenshots**

**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| comp | imagify replaces img tags inside existing picture tags resulting in nested picture tags describe the bug when webp images display option with picture tags is used from the settings and theme already has picture tags present then imagify replaces img tags inside existing picture tags resulting in nested tags expected behavior imagify should ignore already existing picture tags with webp images screenshots backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
98,780 | 11,096,192,193 | IssuesEvent | 2019-12-16 10:38:04 | bob-cd/wendy | https://api.github.com/repos/bob-cd/wendy | closed | document .wendy.edn | documentation | when running bob on a different server or host one has to specify this as config (or as env-var). that this is already possible via `~/.wendy.edn` should be documented. | 1.0 | document .wendy.edn - when running bob on a different server or host one has to specify this as config (or as env-var). that this is already possible via `~/.wendy.edn` should be documented. | non_comp | document wendy edn when running bob on a different server or host one has to specify this as config or as env var that this is already possible via wendy edn should be documented | 0 |
287,702 | 31,856,302,103 | IssuesEvent | 2023-09-15 07:42:54 | Trinadh465/linux-4.1.15_CVE-2023-26607 | https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-26607 | opened | CVE-2016-6197 (Medium) detected in linuxlinux-4.6 | Mend: dependency security vulnerability | ## CVE-2016-6197 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
fs/overlayfs/dir.c in the OverlayFS filesystem implementation in the Linux kernel before 4.6 does not properly verify the upper dentry before proceeding with unlink and rename system-call processing, which allows local users to cause a denial of service (system crash) via a rename system call that specifies a self-hardlink.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6197>CVE-2016-6197</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6197">https://nvd.nist.gov/vuln/detail/CVE-2016-6197</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: 4.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-6197 (Medium) detected in linuxlinux-4.6 - ## CVE-2016-6197 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-26607/commit/6fca0e3f2f14e1e851258fd815766531370084b0">6fca0e3f2f14e1e851258fd815766531370084b0</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
fs/overlayfs/dir.c in the OverlayFS filesystem implementation in the Linux kernel before 4.6 does not properly verify the upper dentry before proceeding with unlink and rename system-call processing, which allows local users to cause a denial of service (system crash) via a rename system call that specifies a self-hardlink.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6197>CVE-2016-6197</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6197">https://nvd.nist.gov/vuln/detail/CVE-2016-6197</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: 4.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_comp | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details fs overlayfs dir c in the overlayfs filesystem implementation in the linux kernel before does not properly verify the upper dentry before proceeding with unlink and rename system call processing which allows local users to cause a denial of service system crash via a rename system call that specifies a self hardlink publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
14,592 | 17,773,980,768 | IssuesEvent | 2021-08-30 16:45:28 | SkinsRestorer/SkinsRestorerX | https://api.github.com/repos/SkinsRestorer/SkinsRestorerX | closed | [plugin incompatibility] fastlogin*bungee* (newer versions) | plugin incompatibility | It seems to be that newer versions of skinsrestorer stopped supporting fastlogin.
- [ ] check since which version skinsrestorer stopped co-working
- [ ] Check since which commit skinsrestorer stopped co-working
- [ ] what is the version of fastlogin, was it a older version? | True | [plugin incompatibility] fastlogin*bungee* (newer versions) - It seems to be that newer versions of skinsrestorer stopped supporting fastlogin.
- [ ] check since which version skinsrestorer stopped co-working
- [ ] Check since which commit skinsrestorer stopped co-working
- [ ] what is the version of fastlogin, was it a older version? | comp | fastlogin bungee newer versions it seems to be that newer versions of skinsrestorer stopped supporting fastlogin check since which version skinsrestorer stopped co working check since which commit skinsrestorer stopped co working what is the version of fastlogin was it a older version | 1 |
8,572 | 10,561,545,497 | IssuesEvent | 2019-10-04 16:06:32 | sebastianbergmann/phpunit | https://api.github.com/repos/sebastianbergmann/phpunit | closed | Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker() | backward-compatibility-break bug | <!--
- Before reporting a backward compatibility break, please consult the release announcement that is available at https://phpunit.de/announcements (as well as the ChangeLog) to make sure it is not an expected change.
- Please do not report an issue for a version of PHPUnit that is no longer supported. A list of currently supported versions of PHPUnit is available at https://phpunit.de/supported-versions.html.
- Please do not report an issue if you are using a version of PHP that is not supported by the version of PHPUnit you are using. A list that shows which version of PHP is supported by which version of PHPUnit is available at https://phpunit.de/supported-versions.html.
- Please fill in this template according to your issue.
- Please keep the table shown below at the top of your issue.
- Please include the output of "composer info | sort" if you installed PHPUnit using Composer.
- Please post code as text (using proper markup). Do not post screenshots of code.
- Visit https://phpunit.de/support.html if you are looking for support.
- Please remove this comment before submitting your issue.
-->
| Q | A
| --------------------| ---------------
| PHPUnit version | 8.4.0
| PHP version | 7.3.9
| Installation Method | Composer
#### Summary
```
1) Chubbyphp\Tests\Mock\MockByCallsTraitTest::testInterfaceWithToManyCalls
Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker()
/vagrant/chubbyphp-mock/tests/MockByCallsTraitTest.php:152
2) Chubbyphp\Tests\Mock\MockByCallsTraitTest::testInterfaceWithToLessCalls
Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker()
/vagrant/chubbyphp-mock/tests/MockByCallsTraitTest.php:187
```
https://github.com/chubbyphp/chubbyphp-mock/issues/4
Related to: https://github.com/sebastianbergmann/phpunit/pull/3849
Other vendors with issues:
https://github.com/php-mock/php-mock-phpunit/issues/39
#### Previous Behavior
tests are up and running
#### Current Behavior
tests are broken as mentioned
#### How to reproduce
git clone git@github.com:chubbyphp/chubbyphp-mock.git && cd chubbyphp-mock && composer install && vendor/bin/phpunit
| True | Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker() - <!--
- Before reporting a backward compatibility break, please consult the release announcement that is available at https://phpunit.de/announcements (as well as the ChangeLog) to make sure it is not an expected change.
- Please do not report an issue for a version of PHPUnit that is no longer supported. A list of currently supported versions of PHPUnit is available at https://phpunit.de/supported-versions.html.
- Please do not report an issue if you are using a version of PHP that is not supported by the version of PHPUnit you are using. A list that shows which version of PHP is supported by which version of PHPUnit is available at https://phpunit.de/supported-versions.html.
- Please fill in this template according to your issue.
- Please keep the table shown below at the top of your issue.
- Please include the output of "composer info | sort" if you installed PHPUnit using Composer.
- Please post code as text (using proper markup). Do not post screenshots of code.
- Visit https://phpunit.de/support.html if you are looking for support.
- Please remove this comment before submitting your issue.
-->
| Q | A
| --------------------| ---------------
| PHPUnit version | 8.4.0
| PHP version | 7.3.9
| Installation Method | Composer
#### Summary
```
1) Chubbyphp\Tests\Mock\MockByCallsTraitTest::testInterfaceWithToManyCalls
Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker()
/vagrant/chubbyphp-mock/tests/MockByCallsTraitTest.php:152
2) Chubbyphp\Tests\Mock\MockByCallsTraitTest::testInterfaceWithToLessCalls
Error: Call to undefined method Mock_SampleInterface_8dc42d9c::__phpunit_getInvocationMocker()
/vagrant/chubbyphp-mock/tests/MockByCallsTraitTest.php:187
```
https://github.com/chubbyphp/chubbyphp-mock/issues/4
Related to: https://github.com/sebastianbergmann/phpunit/pull/3849
Other vendors with issues:
https://github.com/php-mock/php-mock-phpunit/issues/39
#### Previous Behavior
tests are up and running
#### Current Behavior
tests are broken as mentioned
#### How to reproduce
git clone git@github.com:chubbyphp/chubbyphp-mock.git && cd chubbyphp-mock && composer install && vendor/bin/phpunit
| comp | error call to undefined method mock sampleinterface phpunit getinvocationmocker before reporting a backward compatibility break please consult the release announcement that is available at as well as the changelog to make sure it is not an expected change please do not report an issue for a version of phpunit that is no longer supported a list of currently supported versions of phpunit is available at please do not report an issue if you are using a version of php that is not supported by the version of phpunit you are using a list that shows which version of php is supported by which version of phpunit is available at please fill in this template according to your issue please keep the table shown below at the top of your issue please include the output of composer info sort if you installed phpunit using composer please post code as text using proper markup do not post screenshots of code visit if you are looking for support please remove this comment before submitting your issue q a phpunit version php version installation method composer summary chubbyphp tests mock mockbycallstraittest testinterfacewithtomanycalls error call to undefined method mock sampleinterface phpunit getinvocationmocker vagrant chubbyphp mock tests mockbycallstraittest php chubbyphp tests mock mockbycallstraittest testinterfacewithtolesscalls error call to undefined method mock sampleinterface phpunit getinvocationmocker vagrant chubbyphp mock tests mockbycallstraittest php related to other vendors with issues previous behavior tests are up and running current behavior tests are broken as mentioned how to reproduce git clone git github com chubbyphp chubbyphp mock git cd chubbyphp mock composer install vendor bin phpunit | 1 |
15,966 | 20,177,281,138 | IssuesEvent | 2022-02-10 15:27:54 | ossf/tac | https://api.github.com/repos/ossf/tac | closed | TAC Election Process: Member makeup | ElectionProcess | **Proposal:**
Keep TAC size at 7 members
- 4 members are elected by the governing board
- 3 members are elected by Technical Initiative contributors. Each TI determines who is a contributor and may vote. 1 vote per person across all Technical Initiatives
| 1.0 | TAC Election Process: Member makeup - **Proposal:**
Keep TAC size at 7 members
- 4 members are elected by the governing board
- 3 members are elected by Technical Initiative contributors. Each TI determines who is a contributor and may vote. 1 vote per person across all Technical Initiatives
| non_comp | tac election process member makeup proposal keep tac size at members members are elected by the governing board members are elected by technical initiative contributors each ti determines who is a contributor and may vote vote per person across all technical initiatives | 0 |
7,885 | 10,096,348,242 | IssuesEvent | 2019-07-27 17:15:01 | rejahtavi/rfp2 | https://api.github.com/repos/rejahtavi/rfp2 | closed | More Player Models 1.12.2 Compatibility | compatibility wontfix | **Versions**
Minecraft = 1.12.2
Forge =1.12.2-14.23.4.2836
RFP2 =1.0.2
Looking down while using "More player models" the head layer of the Skin renders during first person and gets in the way.
I think this was an issue that was decided to let the maker of "More player models" to fix before,
Though i'd feel much more comfortable if i asked and had a response from you saying if you could fix this issue or not, Thank you for your hard work!

| True | More Player Models 1.12.2 Compatibility - **Versions**
Minecraft = 1.12.2
Forge =1.12.2-14.23.4.2836
RFP2 =1.0.2
Looking down while using "More player models" the head layer of the Skin renders during first person and gets in the way.
I think this was an issue that was decided to let the maker of "More player models" to fix before,
Though i'd feel much more comfortable if i asked and had a response from you saying if you could fix this issue or not, Thank you for your hard work!

| comp | more player models compatibility versions minecraft forge looking down while using more player models the head layer of the skin renders during first person and gets in the way i think this was an issue that was decided to let the maker of more player models to fix before though i d feel much more comfortable if i asked and had a response from you saying if you could fix this issue or not thank you for your hard work | 1 |
19,890 | 27,637,095,895 | IssuesEvent | 2023-03-10 15:12:48 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | AUTO join algorithm is broken in 22.12 | backward compatibility comp-joins | Starting with ClickHouse 22.12 some queries fail when `join_algorithm` is set to `auto`.
```
Received exception from server (version 22.12.3):
Code: 48. DB::Exception: Received from localhost:9000. DB::Exception: Can't execute AUTO join algorithm for this strictness/kind and right storage type. (NOT_IMPLEMENTED)
```
Even the query loading suggestion data for clickhouse-client at startup fails:
```
$ clickhouse-client
ClickHouse client version 22.12.3.5 (official build).
Connecting to localhost:9000 as user default.
Connected to ClickHouse server version 22.12.3 revision 54461.
i102579200 :) Cannot load data for command line suggestions: Code: 48. DB::Exception: Received from localhost:9000. DB::Exception: Can't execute AUTO join algorithm for this strictness/kind and right storage type. (NOT_IMPLEMENTED) (version 22.12.3.5 (official build))
```
_clickhouse server logs_
```
2023.01.15 20:54:15.151966 [ 3529982 ] {5ddb45e8-7fa4-44fe-a108-28c0d625f7aa} <Error> executeQuery: Code: 48. DB::Exception: Can't execute AUTO join algorithm for this strictness
/kind and right storage type. (NOT_IMPLEMENTED) (version 22.12.3.5 (official build)) (from [::1]:60318) (in query: SELECT DISTINCT arrayJoin(extractAll(name, '[\\w_]{2,}')) AS re
s FROM (SELECT * FROM viewIfPermitted(SELECT name FROM system.functions ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.table_engines EL
SE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.formats ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM
system.table_functions ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.data_type_families ELSE null('name String')) UNION ALL SELECT *
FROM viewIfPermitted(SELECT name FROM system.merge_tree_settings ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.settings ELSE null('nam
e String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT cluster FROM system.clusters ELSE null('cluster String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT macro FROM sys
tem.macros ELSE null('macro String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT policy_name FROM system.storage_policies ELSE null('policy_name String')) UNION ALL SELECT *
FROM viewIfPermitted(SELECT concat(func.name, comb.name) AS x FROM system.functions AS func CROSS JOIN system.aggregate_function_combinators AS comb WHERE is_aggregate ELSE null(
'x String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.databases LIMIT 10000 ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTIN
CT name FROM system.tables LIMIT 10000 ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTINCT name FROM system.dictionaries LIMIT 10000 ELSE null('name
String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTINCT name FROM system.columns LIMIT 10000 ELSE null('name String'))) WHERE notEmpty(res)), Stack trace (when copying
this message, always include the lines below):
0. ./build_docker/../src/Common/Exception.cpp:77: DB::Exception::Exception(DB::Exception::MessageMasked const&, int, bool) @ 0xe750cda in /usr/lib/debug/usr/bin/clickhouse.debug
1. ./build_docker/../contrib/llvm-project/libcxx/include/string:1499: DB::Exception::Exception<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::cha
r_traits<char>, std::__1::allocator<char>>*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, char>>(int, fmt::v8::b
asic_format_string<char, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, s
td::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, char>>::type>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::bas
ic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<cha
r>>*>, char>&&) @ 0x13a46154 in /usr/lib/debug/usr/bin/clickhouse.debug
2. ./build_docker/../src/Interpreters/ExpressionAnalyzer.cpp:1085: DB::SelectQueryExpressionAnalyzer::makeJoin(DB::ASTTablesInSelectQueryElement const&, std::__1::vector<DB::Colu
mnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName>> const&, std::__1::shared_ptr<DB::ActionsDAG>&) @ 0x13a34310 in /usr/lib/debug/usr/bin/clickhouse.debug
3. ./build_docker/../src/Interpreters/ExpressionAnalyzer.cpp:0: DB::SelectQueryExpressionAnalyzer::appendJoin(DB::ExpressionActionsChain&, std::__1::shared_ptr<DB::ActionsDAG>&)
@ 0x13a3020d in /usr/lib/debug/usr/bin/clickhouse.debug
4. ./build_docker/../contrib/llvm-project/libcxx/include/__utility/swap.h:37: DB::ExpressionAnalysisResult::ExpressionAnalysisResult(DB::SelectQueryExpressionAnalyzer&, std::__1:
:shared_ptr<DB::StorageInMemoryMetadata const> const&, bool, bool, bool, std::__1::shared_ptr<DB::FilterDAGInfo> const&, std::__1::shared_ptr<DB::FilterDAGInfo> const&, DB::Block
const&) @ 0x13a3e986 in /usr/lib/debug/usr/bin/clickhouse.debug
5. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:812: DB::InterpreterSelectQuery::getSampleBlockImpl() @ 0x142f33e8 in /usr/lib/debug/usr/bin/clickhouse.debug
6. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:645: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_p
tr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std:
:__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1::
shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>)::$_2::operator()(bool) const @ 0x142ebcb5 in /usr/lib/debug/usr/bin/clickhouse.debug
7. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:815: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std:
:__1::shared_ptr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_str
ing<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> cons
t&, std::__1::shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e6a69 in /usr/lib/debug/usr/bin/clickhouse.debug
8. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, st
d::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>,
std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
9. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> c
onst&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, st
d::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
10. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:596: DB::InterpreterSelectWithUnionQuery::getSampleBlock(std::__1::shared_ptr<DB::IAST> const&, st
d::__1::shared_ptr<DB::Context const>, bool) @ 0x14388655 in /usr/lib/debug/usr/bin/clickhouse.debug
11. ./build_docker/../contrib/llvm-project/libcxx/include/vector:951: DB::TableFunctionViewIfPermitted::isPermitted(std::__1::shared_ptr<DB::Context const> const&, DB::ColumnsDes
cription const&) const @ 0x12e6e0bb in /usr/lib/debug/usr/bin/clickhouse.debug
12. ./build_docker/../src/TableFunctions/TableFunctionViewIfPermitted.cpp:63: DB::TableFunctionViewIfPermitted::executeImpl(std::__1::shared_ptr<DB::IAST> const&, std::__1::share
d_ptr<DB::Context const>, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, DB::ColumnsDescription) const @ 0x12e6dcc4 in /usr/lib/debu
g/usr/bin/clickhouse.debug
13. ./build_docker/../src/TableFunctions/ITableFunction.cpp:0: DB::ITableFunction::execute(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context const>, std::__
1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, DB::ColumnsDescription, bool) const @ 0x13156016 in /usr/lib/debug/usr/bin/clickhouse.debug
14. ./build_docker/../src/Interpreters/Context.cpp:0: DB::Context::executeTableFunction(std::__1::shared_ptr<DB::IAST> const&, DB::ASTSelectQuery const*) @ 0x1391adf7 in /usr/lib
/debug/usr/bin/clickhouse.debug
15. ./build_docker/../src/Interpreters/JoinedTables.cpp:0: DB::JoinedTables::getLeftTableStorage() @ 0x143c4417 in /usr/lib/debug/usr/bin/clickhouse.debug
16. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:0: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_pt
r<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::
__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1::s
hared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e4972 in /usr/lib/debug/usr/bin/clickhouse.debug
17. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, s
td::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>
, std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
18. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST>
const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, s
td::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
19. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:40: DB::InterpreterSelectWithUnionQuery::getSampleBlock(std::__1::shared_ptr<DB::IAST> const&, std::__1
::shared_ptr<DB::Context const>, bool) @ 0x1438821f in /usr/lib/debug/usr/bin/clickhouse.debug
20. ./build_docker/../contrib/llvm-project/libcxx/include/list:285: DB::getDatabaseAndTablesWithColumns(std::__1::vector<DB::ASTTableExpression const*, std::__1::allocator<DB::AS
TTableExpression const*>> const&, std::__1::shared_ptr<DB::Context const>, bool, bool) @ 0x146a98aa in /usr/lib/debug/usr/bin/clickhouse.debug
21. ./build_docker/../contrib/llvm-project/libcxx/include/vector:951: DB::JoinedTables::resolveTables() @ 0x143c4960 in /usr/lib/debug/usr/bin/clickhouse.debug
22. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:430: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_
ptr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std
::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1:
:shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e4e76 in /usr/lib/debug/usr/bin/clickhouse.debug
23. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, s
td::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>
, std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
24. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST>
const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, s
td::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
25. ./build_docker/../contrib/llvm-project/libcxx/include/vector:434: DB::InterpreterFactory::get(std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context>, DB::SelectQu
eryOptions const&) @ 0x142a46dd in /usr/lib/debug/usr/bin/clickhouse.debug
26. ./build_docker/../src/Interpreters/executeQuery.cpp:0: DB::executeQueryImpl(char const*, char const*, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum,
DB::ReadBuffer*) @ 0x146924e0 in /usr/lib/debug/usr/bin/clickhouse.debug
27. ./build_docker/../src/Interpreters/executeQuery.cpp:1083: DB::executeQuery(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, std::_
_1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum) @ 0x1468fdad in /usr/lib/debug/usr/bin/clickhouse.debug
28. ./build_docker/../src/Server/TCPHandler.cpp:0: DB::TCPHandler::runImpl() @ 0x1538858f in /usr/lib/debug/usr/bin/clickhouse.debug
29. ./build_docker/../src/Server/TCPHandler.cpp:1916: DB::TCPHandler::run() @ 0x1539be59 in /usr/lib/debug/usr/bin/clickhouse.debug
30. ./build_docker/../contrib/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x1821ac34 in /usr/lib/debug/usr/bin/clickhouse.debug
31. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x1821c5db in /usr/lib/debug/usr/bin/clickhouse.debug
```
----
In earlier versions the same query succeeds.
22.8: https://fiddle.clickhouse.com/11f3234f-7756-43a0-aa5e-54c793927bb3
22.11: https://fiddle.clickhouse.com/0259c39c-83a9-4cce-8ef7-7fc91fff6241
And here is minimized query:
22.8: https://fiddle.clickhouse.com/9af2a61c-40d2-42bf-8127-cc53f8f30289
22.12: https://fiddle.clickhouse.com/6ff9c0fb-4651-4556-87ed-d4071286ef6f
| True | AUTO join algorithm is broken in 22.12 - Starting with ClickHouse 22.12 some queries fail when `join_algorithm` is set to `auto`.
```
Received exception from server (version 22.12.3):
Code: 48. DB::Exception: Received from localhost:9000. DB::Exception: Can't execute AUTO join algorithm for this strictness/kind and right storage type. (NOT_IMPLEMENTED)
```
Even the query loading suggestion data for clickhouse-client at startup fails:
```
$ clickhouse-client
ClickHouse client version 22.12.3.5 (official build).
Connecting to localhost:9000 as user default.
Connected to ClickHouse server version 22.12.3 revision 54461.
i102579200 :) Cannot load data for command line suggestions: Code: 48. DB::Exception: Received from localhost:9000. DB::Exception: Can't execute AUTO join algorithm for this strictness/kind and right storage type. (NOT_IMPLEMENTED) (version 22.12.3.5 (official build))
```
_clickhouse server logs_
```
2023.01.15 20:54:15.151966 [ 3529982 ] {5ddb45e8-7fa4-44fe-a108-28c0d625f7aa} <Error> executeQuery: Code: 48. DB::Exception: Can't execute AUTO join algorithm for this strictness
/kind and right storage type. (NOT_IMPLEMENTED) (version 22.12.3.5 (official build)) (from [::1]:60318) (in query: SELECT DISTINCT arrayJoin(extractAll(name, '[\\w_]{2,}')) AS re
s FROM (SELECT * FROM viewIfPermitted(SELECT name FROM system.functions ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.table_engines EL
SE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.formats ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM
system.table_functions ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.data_type_families ELSE null('name String')) UNION ALL SELECT *
FROM viewIfPermitted(SELECT name FROM system.merge_tree_settings ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.settings ELSE null('nam
e String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT cluster FROM system.clusters ELSE null('cluster String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT macro FROM sys
tem.macros ELSE null('macro String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT policy_name FROM system.storage_policies ELSE null('policy_name String')) UNION ALL SELECT *
FROM viewIfPermitted(SELECT concat(func.name, comb.name) AS x FROM system.functions AS func CROSS JOIN system.aggregate_function_combinators AS comb WHERE is_aggregate ELSE null(
'x String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT name FROM system.databases LIMIT 10000 ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTIN
CT name FROM system.tables LIMIT 10000 ELSE null('name String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTINCT name FROM system.dictionaries LIMIT 10000 ELSE null('name
String')) UNION ALL SELECT * FROM viewIfPermitted(SELECT DISTINCT name FROM system.columns LIMIT 10000 ELSE null('name String'))) WHERE notEmpty(res)), Stack trace (when copying
this message, always include the lines below):
0. ./build_docker/../src/Common/Exception.cpp:77: DB::Exception::Exception(DB::Exception::MessageMasked const&, int, bool) @ 0xe750cda in /usr/lib/debug/usr/bin/clickhouse.debug
1. ./build_docker/../contrib/llvm-project/libcxx/include/string:1499: DB::Exception::Exception<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::cha
r_traits<char>, std::__1::allocator<char>>*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, char>>(int, fmt::v8::b
asic_format_string<char, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, s
td::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, char>>::type>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::bas
ic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<cha
r>>*>, char>&&) @ 0x13a46154 in /usr/lib/debug/usr/bin/clickhouse.debug
2. ./build_docker/../src/Interpreters/ExpressionAnalyzer.cpp:1085: DB::SelectQueryExpressionAnalyzer::makeJoin(DB::ASTTablesInSelectQueryElement const&, std::__1::vector<DB::Colu
mnWithTypeAndName, std::__1::allocator<DB::ColumnWithTypeAndName>> const&, std::__1::shared_ptr<DB::ActionsDAG>&) @ 0x13a34310 in /usr/lib/debug/usr/bin/clickhouse.debug
3. ./build_docker/../src/Interpreters/ExpressionAnalyzer.cpp:0: DB::SelectQueryExpressionAnalyzer::appendJoin(DB::ExpressionActionsChain&, std::__1::shared_ptr<DB::ActionsDAG>&)
@ 0x13a3020d in /usr/lib/debug/usr/bin/clickhouse.debug
4. ./build_docker/../contrib/llvm-project/libcxx/include/__utility/swap.h:37: DB::ExpressionAnalysisResult::ExpressionAnalysisResult(DB::SelectQueryExpressionAnalyzer&, std::__1:
:shared_ptr<DB::StorageInMemoryMetadata const> const&, bool, bool, bool, std::__1::shared_ptr<DB::FilterDAGInfo> const&, std::__1::shared_ptr<DB::FilterDAGInfo> const&, DB::Block
const&) @ 0x13a3e986 in /usr/lib/debug/usr/bin/clickhouse.debug
5. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:812: DB::InterpreterSelectQuery::getSampleBlockImpl() @ 0x142f33e8 in /usr/lib/debug/usr/bin/clickhouse.debug
6. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:645: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_p
tr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std:
:__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1::
shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>)::$_2::operator()(bool) const @ 0x142ebcb5 in /usr/lib/debug/usr/bin/clickhouse.debug
7. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:815: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std:
:__1::shared_ptr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_str
ing<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> cons
t&, std::__1::shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e6a69 in /usr/lib/debug/usr/bin/clickhouse.debug
8. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, st
d::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>,
std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
9. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> c
onst&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, st
d::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
10. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:596: DB::InterpreterSelectWithUnionQuery::getSampleBlock(std::__1::shared_ptr<DB::IAST> const&, st
d::__1::shared_ptr<DB::Context const>, bool) @ 0x14388655 in /usr/lib/debug/usr/bin/clickhouse.debug
11. ./build_docker/../contrib/llvm-project/libcxx/include/vector:951: DB::TableFunctionViewIfPermitted::isPermitted(std::__1::shared_ptr<DB::Context const> const&, DB::ColumnsDes
cription const&) const @ 0x12e6e0bb in /usr/lib/debug/usr/bin/clickhouse.debug
12. ./build_docker/../src/TableFunctions/TableFunctionViewIfPermitted.cpp:63: DB::TableFunctionViewIfPermitted::executeImpl(std::__1::shared_ptr<DB::IAST> const&, std::__1::share
d_ptr<DB::Context const>, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, DB::ColumnsDescription) const @ 0x12e6dcc4 in /usr/lib/debu
g/usr/bin/clickhouse.debug
13. ./build_docker/../src/TableFunctions/ITableFunction.cpp:0: DB::ITableFunction::execute(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context const>, std::__
1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, DB::ColumnsDescription, bool) const @ 0x13156016 in /usr/lib/debug/usr/bin/clickhouse.debug
14. ./build_docker/../src/Interpreters/Context.cpp:0: DB::Context::executeTableFunction(std::__1::shared_ptr<DB::IAST> const&, DB::ASTSelectQuery const*) @ 0x1391adf7 in /usr/lib
/debug/usr/bin/clickhouse.debug
15. ./build_docker/../src/Interpreters/JoinedTables.cpp:0: DB::JoinedTables::getLeftTableStorage() @ 0x143c4417 in /usr/lib/debug/usr/bin/clickhouse.debug
16. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:0: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_pt
r<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::
__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1::s
hared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e4972 in /usr/lib/debug/usr/bin/clickhouse.debug
17. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, s
td::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>
, std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
18. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST>
const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, s
td::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
19. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:40: DB::InterpreterSelectWithUnionQuery::getSampleBlock(std::__1::shared_ptr<DB::IAST> const&, std::__1
::shared_ptr<DB::Context const>, bool) @ 0x1438821f in /usr/lib/debug/usr/bin/clickhouse.debug
20. ./build_docker/../contrib/llvm-project/libcxx/include/list:285: DB::getDatabaseAndTablesWithColumns(std::__1::vector<DB::ASTTableExpression const*, std::__1::allocator<DB::AS
TTableExpression const*>> const&, std::__1::shared_ptr<DB::Context const>, bool, bool) @ 0x146a98aa in /usr/lib/debug/usr/bin/clickhouse.debug
21. ./build_docker/../contrib/llvm-project/libcxx/include/vector:951: DB::JoinedTables::resolveTables() @ 0x143c4960 in /usr/lib/debug/usr/bin/clickhouse.debug
22. ./build_docker/../src/Interpreters/InterpreterSelectQuery.cpp:430: DB::InterpreterSelectQuery::InterpreterSelectQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_
ptr<DB::Context> const&, std::__1::optional<DB::Pipe>, std::__1::shared_ptr<DB::IStorage> const&, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std
::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&, std::__1:
:shared_ptr<DB::StorageInMemoryMetadata const> const&, std::__1::shared_ptr<DB::PreparedSets>) @ 0x142e4e76 in /usr/lib/debug/usr/bin/clickhouse.debug
23. ./build_docker/../contrib/llvm-project/libcxx/include/optional:260: DB::InterpreterSelectWithUnionQuery::buildCurrentChildInterpreter(std::__1::shared_ptr<DB::IAST> const&, s
td::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>
, std::__1::allocator<char>>>> const&) @ 0x14386ce2 in /usr/lib/debug/usr/bin/clickhouse.debug
24. ./build_docker/../src/Interpreters/InterpreterSelectWithUnionQuery.cpp:0: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST>
const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>, s
td::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>>>> const&) @ 0x14384482 in /usr/lib/debug/usr/bin/clickhouse.debug
25. ./build_docker/../contrib/llvm-project/libcxx/include/vector:434: DB::InterpreterFactory::get(std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context>, DB::SelectQu
eryOptions const&) @ 0x142a46dd in /usr/lib/debug/usr/bin/clickhouse.debug
26. ./build_docker/../src/Interpreters/executeQuery.cpp:0: DB::executeQueryImpl(char const*, char const*, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum,
DB::ReadBuffer*) @ 0x146924e0 in /usr/lib/debug/usr/bin/clickhouse.debug
27. ./build_docker/../src/Interpreters/executeQuery.cpp:1083: DB::executeQuery(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char>> const&, std::_
_1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum) @ 0x1468fdad in /usr/lib/debug/usr/bin/clickhouse.debug
28. ./build_docker/../src/Server/TCPHandler.cpp:0: DB::TCPHandler::runImpl() @ 0x1538858f in /usr/lib/debug/usr/bin/clickhouse.debug
29. ./build_docker/../src/Server/TCPHandler.cpp:1916: DB::TCPHandler::run() @ 0x1539be59 in /usr/lib/debug/usr/bin/clickhouse.debug
30. ./build_docker/../contrib/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x1821ac34 in /usr/lib/debug/usr/bin/clickhouse.debug
31. ./build_docker/../contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x1821c5db in /usr/lib/debug/usr/bin/clickhouse.debug
```
----
In earlier versions the same query succeeds.
22.8: https://fiddle.clickhouse.com/11f3234f-7756-43a0-aa5e-54c793927bb3
22.11: https://fiddle.clickhouse.com/0259c39c-83a9-4cce-8ef7-7fc91fff6241
And here is minimized query:
22.8: https://fiddle.clickhouse.com/9af2a61c-40d2-42bf-8127-cc53f8f30289
22.12: https://fiddle.clickhouse.com/6ff9c0fb-4651-4556-87ed-d4071286ef6f
| comp | auto join algorithm is broken in starting with clickhouse some queries fail when join algorithm is set to auto received exception from server version code db exception received from localhost db exception can t execute auto join algorithm for this strictness kind and right storage type not implemented even the query loading suggestion data for clickhouse client at startup fails clickhouse client clickhouse client version official build connecting to localhost as user default connected to clickhouse server version revision cannot load data for command line suggestions code db exception received from localhost db exception can t execute auto join algorithm for this strictness kind and right storage type not implemented version official build clickhouse server logs executequery code db exception can t execute auto join algorithm for this strictness kind and right storage type not implemented version official build from in query select distinct arrayjoin extractall name as re s from select from viewifpermitted select name from system functions else null name string union all select from viewifpermitted select name from system table engines el se null name string union all select from viewifpermitted select name from system formats else null name string union all select from viewifpermitted select name from system table functions else null name string union all select from viewifpermitted select name from system data type families else null name string union all select from viewifpermitted select name from system merge tree settings else null name string union all select from viewifpermitted select name from system settings else null nam e string union all select from viewifpermitted select cluster from system clusters else null cluster string union all select from viewifpermitted select macro from sys tem macros else null macro string union all select from viewifpermitted select policy name from system storage policies else null policy name string union all select from viewifpermitted select concat func name comb name as x from system functions as func cross join system aggregate function combinators as comb where is aggregate else null x string union all select from viewifpermitted select name from system databases limit else null name string union all select from viewifpermitted select distin ct name from system tables limit else null name string union all select from viewifpermitted select distinct name from system dictionaries limit else null name string union all select from viewifpermitted select distinct name from system columns limit else null name string where notempty res stack trace when copying this message always include the lines below build docker src common exception cpp db exception exception db exception messagemasked const int bool in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include string db exception exception fmt join view std wrap iter std basic string char std cha r traits std allocator std wrap iter std allocator char int fmt b asic format string std allocator s td wrap iter std allocator char type fmt join view std wrap iter std bas ic string std allocator std wrap iter std allocator cha r char in usr lib debug usr bin clickhouse debug build docker src interpreters expressionanalyzer cpp db selectqueryexpressionanalyzer makejoin db asttablesinselectqueryelement const std vector db colu mnwithtypeandname std allocator const std shared ptr in usr lib debug usr bin clickhouse debug build docker src interpreters expressionanalyzer cpp db selectqueryexpressionanalyzer appendjoin db expressionactionschain std shared ptr in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include utility swap h db expressionanalysisresult expressionanalysisresult db selectqueryexpressionanalyzer std shared ptr const bool bool bool std shared ptr const std shared ptr const db block const in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectquery cpp db interpreterselectquery getsampleblockimpl in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectquery cpp db interpreterselectquery interpreterselectquery std shared ptr const std shared p tr const std optional std shared ptr const db selectqueryoptions const std vector std basic string char std char traits std allocator std allocator std allocator const std shared ptr const std shared ptr operator bool const in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include memory shared ptr h db interpreterselectquery interpreterselectquery std shared ptr const std shared ptr const std optional std shared ptr const db selectqueryoptions const std vector std basic str ing std allocator std allocator std allocator cons t std shared ptr const std shared ptr in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include optional db interpreterselectwithunionquery buildcurrentchildinterpreter std shared ptr const st d vector std allocator std allocator std allocator const in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery interpreterselectwithunionquery std shared ptr c onst std shared ptr db selectqueryoptions const std vector std allocator st d allocator std allocator const in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include memory shared ptr h db interpreterselectwithunionquery getsampleblock std shared ptr const st d shared ptr bool in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include vector db tablefunctionviewifpermitted ispermitted std shared ptr const db columnsdes cription const const in usr lib debug usr bin clickhouse debug build docker src tablefunctions tablefunctionviewifpermitted cpp db tablefunctionviewifpermitted executeimpl std shared ptr const std share d ptr std basic string std allocator const db columnsdescription const in usr lib debu g usr bin clickhouse debug build docker src tablefunctions itablefunction cpp db itablefunction execute std shared ptr const std shared ptr std basic string std allocator const db columnsdescription bool const in usr lib debug usr bin clickhouse debug build docker src interpreters context cpp db context executetablefunction std shared ptr const db astselectquery const in usr lib debug usr bin clickhouse debug build docker src interpreters joinedtables cpp db joinedtables getlefttablestorage in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectquery cpp db interpreterselectquery interpreterselectquery std shared ptr const std shared pt r const std optional std shared ptr const db selectqueryoptions const std vector std basic string char std char traits std allocator std allocator std allocator const std s hared ptr const std shared ptr in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include optional db interpreterselectwithunionquery buildcurrentchildinterpreter std shared ptr const s td vector std allocator std allocator std allocator const in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery interpreterselectwithunionquery std shared ptr const std shared ptr db selectqueryoptions const std vector std allocator s td allocator std allocator const in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery getsampleblock std shared ptr const std shared ptr bool in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include list db getdatabaseandtableswithcolumns std vector db asttableexpression const std allocator db as ttableexpression const const std shared ptr bool bool in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include vector db joinedtables resolvetables in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectquery cpp db interpreterselectquery interpreterselectquery std shared ptr const std shared ptr const std optional std shared ptr const db selectqueryoptions const std vector std basic string char std char traits std allocator std allocator std allocator const std shared ptr const std shared ptr in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include optional db interpreterselectwithunionquery buildcurrentchildinterpreter std shared ptr const s td vector std allocator std allocator std allocator const in usr lib debug usr bin clickhouse debug build docker src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery interpreterselectwithunionquery std shared ptr const std shared ptr db selectqueryoptions const std vector std allocator s td allocator std allocator const in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include vector db interpreterfactory get std shared ptr std shared ptr db selectqu eryoptions const in usr lib debug usr bin clickhouse debug build docker src interpreters executequery cpp db executequeryimpl char const char const std shared ptr bool db queryprocessingstage enum db readbuffer in usr lib debug usr bin clickhouse debug build docker src interpreters executequery cpp db executequery std basic string std allocator const std shared ptr bool db queryprocessingstage enum in usr lib debug usr bin clickhouse debug build docker src server tcphandler cpp db tcphandler runimpl in usr lib debug usr bin clickhouse debug build docker src server tcphandler cpp db tcphandler run in usr lib debug usr bin clickhouse debug build docker contrib poco net src tcpserverconnection cpp poco net tcpserverconnection start in usr lib debug usr bin clickhouse debug build docker contrib llvm project libcxx include memory unique ptr h poco net tcpserverdispatcher run in usr lib debug usr bin clickhouse debug in earlier versions the same query succeeds and here is minimized query | 1 |
751,865 | 26,261,744,665 | IssuesEvent | 2023-01-06 08:36:02 | arianne/stendhal | https://api.github.com/repos/arianne/stendhal | closed | Elias Thinks His Grandson is Still Missing | type: bug game: quest priority: low | ## 🐞 Describe the bug
If you cure the myling then ask Elias for a quest before talking to Niall, he will respond as if Niall is still missing.
```
[11:41] <seriph> hi
[11:41] <Elias Breland> You have returned my grandson to me. I cannot thank you enough.
I don't have much to offer for your kind service, but please speak to Niall. He
is in the basement.
[11:41] <seriph> quest
[11:41] <Elias Breland> Thank you for accepting my plea for help. Please tell me if you
hear any news about what has become of my grandson. He used to play with a little
girl named Marianne.
```
Not a major bug. It doesn't affect the player's ability to complete the quest & is an easy fix. | 1.0 | Elias Thinks His Grandson is Still Missing - ## 🐞 Describe the bug
If you cure the myling then ask Elias for a quest before talking to Niall, he will respond as if Niall is still missing.
```
[11:41] <seriph> hi
[11:41] <Elias Breland> You have returned my grandson to me. I cannot thank you enough.
I don't have much to offer for your kind service, but please speak to Niall. He
is in the basement.
[11:41] <seriph> quest
[11:41] <Elias Breland> Thank you for accepting my plea for help. Please tell me if you
hear any news about what has become of my grandson. He used to play with a little
girl named Marianne.
```
Not a major bug. It doesn't affect the player's ability to complete the quest & is an easy fix. | non_comp | elias thinks his grandson is still missing 🐞 describe the bug if you cure the myling then ask elias for a quest before talking to niall he will respond as if niall is still missing hi you have returned my grandson to me i cannot thank you enough i don t have much to offer for your kind service but please speak to niall he is in the basement quest thank you for accepting my plea for help please tell me if you hear any news about what has become of my grandson he used to play with a little girl named marianne not a major bug it doesn t affect the player s ability to complete the quest is an easy fix | 0 |
261,350 | 22,740,625,469 | IssuesEvent | 2022-07-07 03:09:13 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | Fix Vulnerability Detector IT: Tier 1 - test_scan_results | team/qa feature/vuln-detector type/fix subteam/qa-storm test/nightly type/nightly-test-failure status/not-tracked | |Related issue|
|---|
|https://github.com/wazuh/wazuh-qa/issues/3057
| Date| Commit | Commit title | Build | Version|
|--|--|--|--|--|
| 2022-07-06| [c62d985](https://github.com/wazuh/wazuh/commit/c62d985f729591fb2f7b0026ec08bb1756dd031b) | disable filebeat metrics (https://github.com/wazuh/wazuh/pull/14121) | [#28645](https://ci.wazuh.info/job/Test_integration/28645) | 4.3.6 |
## Case
| Tier| Wazuh Type | OS
|--|--|--|
| 1 | Manager | CentOS
### Description
```
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED11]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED12]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED15]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES11]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES12]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES15]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_provider_and_nvd_vulnerabilities.py::test_scan_provider_and_nvd_vulnerabilities[SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_provider_vulnerabilities.py::test_scan_provider_vulnerabilities[SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_update_package[Alert vulnerability removal - SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_delete_package[Alert vulnerability removal - SUSE]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_provider_and_nvd_vulnerabilities.py::test_scan_provider_and_nvd_vulnerabilities[ALAS]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_provider_vulnerabilities.py::test_scan_provider_vulnerabilities[ALAS]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_update_package[Alert vulnerability removal - ALAS 2022]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_delete_package[Alert vulnerability removal - ALAS 2022]
```
## Tasks
- [ ] Run tests in Jenkins
- [ ] Run tests in local enviroment
- [ ] Read test code & docs and run tests manually
- [ ] Fix failures | 2.0 | Fix Vulnerability Detector IT: Tier 1 - test_scan_results - |Related issue|
|---|
|https://github.com/wazuh/wazuh-qa/issues/3057
| Date| Commit | Commit title | Build | Version|
|--|--|--|--|--|
| 2022-07-06| [c62d985](https://github.com/wazuh/wazuh/commit/c62d985f729591fb2f7b0026ec08bb1756dd031b) | disable filebeat metrics (https://github.com/wazuh/wazuh/pull/14121) | [#28645](https://ci.wazuh.info/job/Test_integration/28645) | 4.3.6 |
## Case
| Tier| Wazuh Type | OS
|--|--|--|
| 1 | Manager | CentOS
### Description
```
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED11]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED12]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLED15]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES11]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES12]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_nvd_vulnerabilities.py::test_no_agent_data[SLES15]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_provider_and_nvd_vulnerabilities.py::test_scan_provider_and_nvd_vulnerabilities[SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_provider_vulnerabilities.py::test_scan_provider_vulnerabilities[SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_update_package[Alert vulnerability removal - SUSE]
09:47:51 FAILED test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_delete_package[Alert vulnerability removal - SUSE]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_provider_and_nvd_vulnerabilities.py::test_scan_provider_and_nvd_vulnerabilities[ALAS]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_provider_vulnerabilities.py::test_scan_provider_vulnerabilities[ALAS]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_update_package[Alert vulnerability removal - ALAS 2022]
09:47:51 ERROR test_vulnerability_detector/test_scan_results/test_scan_vulnerability_removal.py::test_vulnerability_removal_delete_package[Alert vulnerability removal - ALAS 2022]
```
## Tasks
- [ ] Run tests in Jenkins
- [ ] Run tests in local enviroment
- [ ] Read test code & docs and run tests manually
- [ ] Fix failures | non_comp | fix vulnerability detector it tier test scan results related issue date commit commit title build version disable filebeat metrics case tier wazuh type os manager centos description failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan nvd vulnerabilities py test no agent data failed test vulnerability detector test scan results test scan provider and nvd vulnerabilities py test scan provider and nvd vulnerabilities failed test vulnerability detector test scan results test scan provider vulnerabilities py test scan provider vulnerabilities failed test vulnerability detector test scan results test scan vulnerability removal py test vulnerability removal update package failed test vulnerability detector test scan results test scan vulnerability removal py test vulnerability removal delete package error test vulnerability detector test scan results test scan provider and nvd vulnerabilities py test scan provider and nvd vulnerabilities error test vulnerability detector test scan results test scan provider vulnerabilities py test scan provider vulnerabilities error test vulnerability detector test scan results test scan vulnerability removal py test vulnerability removal update package error test vulnerability detector test scan results test scan vulnerability removal py test vulnerability removal delete package tasks run tests in jenkins run tests in local enviroment read test code docs and run tests manually fix failures | 0 |
28,478 | 6,998,078,463 | IssuesEvent | 2017-12-16 22:46:46 | NetOfficeFw/NetOffice | https://api.github.com/repos/NetOfficeFw/NetOffice | closed | Excel Range Address Stops Working in NetOffice Release 1.5.1.2 | bug CodePlex ms-excel | <img src="https://avatars.githubusercontent.com/u/30236365?s=96" align="left" width="48" height="48" /> <strong>Issue by <a href="https://www.codeplex.com/site/users/view/Jon49" rel="author">Jon49</a></strong>
<em datetime="2012-11-07T21:05:56">Wed, 07 Nov 2012 21:05:56 GMT</em>
_Originally opened as https://netoffice.codeplex.com/workitem/18875_
----
Update: My computer was doing some funky things. I'll try the new release again tomorrow just to make sure it wasn't something I did.
Windows XP, Excel 2003, VB.NET 2010, .NET 4.0
I switched to the new dlls (NetOffice Release 1.5.1.2) and the address from the range shows only two options.
Address() As String
Chars(index As Integer) As Char
(I have no idea where the Chars comes from).
This is what I have written and it gives an error with the second parameter (rng is a NetOffice.ExcelAPI.Range):
Dim sAddress As String = rng.Address(True, True) | 1.0 | Excel Range Address Stops Working in NetOffice Release 1.5.1.2 - <img src="https://avatars.githubusercontent.com/u/30236365?s=96" align="left" width="48" height="48" /> <strong>Issue by <a href="https://www.codeplex.com/site/users/view/Jon49" rel="author">Jon49</a></strong>
<em datetime="2012-11-07T21:05:56">Wed, 07 Nov 2012 21:05:56 GMT</em>
_Originally opened as https://netoffice.codeplex.com/workitem/18875_
----
Update: My computer was doing some funky things. I'll try the new release again tomorrow just to make sure it wasn't something I did.
Windows XP, Excel 2003, VB.NET 2010, .NET 4.0
I switched to the new dlls (NetOffice Release 1.5.1.2) and the address from the range shows only two options.
Address() As String
Chars(index As Integer) As Char
(I have no idea where the Chars comes from).
This is what I have written and it gives an error with the second parameter (rng is a NetOffice.ExcelAPI.Range):
Dim sAddress As String = rng.Address(True, True) | non_comp | excel range address stops working in netoffice release issue by wed nov gmt originally opened as update my computer was doing some funky things i ll try the new release again tomorrow just to make sure it wasn t something i did windows xp excel vb net net i switched to the new dlls netoffice release and the address from the range shows only two options address as string chars index as integer as char i have no idea where the chars comes from this is what i have written and it gives an error with the second parameter rng is a netoffice excelapi range dim saddress as string rng address true true | 0 |
329,344 | 10,014,944,650 | IssuesEvent | 2019-07-15 18:48:44 | pravega/pravega | https://api.github.com/repos/pravega/pravega | closed | Cherry-pick merges from master to r0.5 | kind/release priority/P0 status/accepted version/0.5.0 | **Problem description**
We need to cherry-pick the following merges to `r0.5`:
```
Issue 4011: (SegmentStore) Making Commit Backlog Throttler more aggressive (#4016)
Issue 4011: Fixed ThrottlerCalculator.Delay.ToString() (#4023)
```
**Problem location**
Release 0.5
**Suggestions for an improvement**
Backport the changes. | 1.0 | Cherry-pick merges from master to r0.5 - **Problem description**
We need to cherry-pick the following merges to `r0.5`:
```
Issue 4011: (SegmentStore) Making Commit Backlog Throttler more aggressive (#4016)
Issue 4011: Fixed ThrottlerCalculator.Delay.ToString() (#4023)
```
**Problem location**
Release 0.5
**Suggestions for an improvement**
Backport the changes. | non_comp | cherry pick merges from master to problem description we need to cherry pick the following merges to issue segmentstore making commit backlog throttler more aggressive issue fixed throttlercalculator delay tostring problem location release suggestions for an improvement backport the changes | 0 |
8,727 | 10,679,884,661 | IssuesEvent | 2019-10-21 20:12:03 | blay09/CraftingTweaks | https://api.github.com/repos/blay09/CraftingTweaks | closed | Compatibility Request for 1.14 Mod. | compatibility fixed in next version | Awesome mod we love your work and use it in all our modpacks.
We are updating our modpacks to 1.14.4 and it seems you mod is not working with [Crafting Station](https://www.curseforge.com/minecraft/mc-mods/crafting-station) mod yet.
Could you code a compatibility w/this crafting table, please?
Using:
Forge-28.0.81.jar
craftingstation-2.0.2a.jar
CraftingTweaks_1.14.4-10.1.3.jar
Thank you so much. | True | Compatibility Request for 1.14 Mod. - Awesome mod we love your work and use it in all our modpacks.
We are updating our modpacks to 1.14.4 and it seems you mod is not working with [Crafting Station](https://www.curseforge.com/minecraft/mc-mods/crafting-station) mod yet.
Could you code a compatibility w/this crafting table, please?
Using:
Forge-28.0.81.jar
craftingstation-2.0.2a.jar
CraftingTweaks_1.14.4-10.1.3.jar
Thank you so much. | comp | compatibility request for mod awesome mod we love your work and use it in all our modpacks we are updating our modpacks to and it seems you mod is not working with mod yet could you code a compatibility w this crafting table please using forge jar craftingstation jar craftingtweaks jar thank you so much | 1 |
92,672 | 26,747,709,543 | IssuesEvent | 2023-01-30 17:07:08 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Test failure due to inconsistencies related to MKL | stat:awaiting response type:build/install stalled comp:mkl TF 2.9 | <details><summary>Click to expand!</summary>
### Issue Type
Build/Install
### Have you reproduced the bug with TF nightly?
No
### Source
source
### Tensorflow Version
2.9.1
### Custom Code
No
### OS Platform and Distribution
Linux RHEL 7
### Python version
3.10
### Bazel version
5.1.1
### GCC/Compiler version
11.3
### Current Behaviour?
```shell
I see failing tests while running `bazel test`.
The `testGetMemoryInfoCPU` is guarded by a skip `if test_util.IsMklEnabled():`
And the other tests has a similar condition.
So it looks like `IsMklEnabled` doesn't return true when it should. Further investigation leads to several macros in the build files:
- `IsMklEnabled` returns true when `defined(INTEL_MKL) && defined(ENABLE_MKL)`
- There is `if_mkl([":mkl_cpu_allocator"])` which is why `testGetMemoryInfoCPU` fails when true
- `if_mkl` is true on `linux_x86_64` unconditionally
- `if_mkl(["-DINTEL_MKL"])`
- `if_enable_mkl(["-DENABLE_MKL"])`
I'd suggest to simplify this so either there is MKL or there is not. Having multiple macros with different defaults doesn't make sense to me. Or maybe the default for "ENABLE_MKL" should be the same as for "INTEL_MKL", e.g. True on x86
```
### Standalone code to reproduce the issue
```shell
bazel test
```
### Relevant log output
```shell
FAIL: testGetMemoryInfoCPU (__main__.ContextTest)
ContextTest.testGetMemoryInfoCPU
----------------------------------------------------------------------
Traceback (most recent call last):
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/eager/context_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/test_util.py", line 2229, in decorated
return func(self, *args, **kwargs)
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/eager/context_test_cpu.runfiles/org_tensorflow/tensorflow/python/eager/context_test.py", line 141, in testGetMemoryInfoCPU
with self.assertRaisesRegex(ValueError, 'Allocator stats not available'):
AssertionError: ValueError not raised
FAIL: test_simple (__main__.NodeFileWriterTest)
NodeFileWriterTest.test_simple
----------------------------------------------------------------------
Traceback (most recent call last):
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/framework/node_file_writer_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/test_util.py", line 2157, in decorated
return func(self, *args, **kwargs)
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/framework/node_file_writer_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/node_file_writer_test.py", line 142, in test_simple
self.assertEqual(node_def1.op, 'MatMul')
AssertionError:
- _MklMatMul
? ----
+ MatMul
```
</details> | 1.0 | Test failure due to inconsistencies related to MKL - <details><summary>Click to expand!</summary>
### Issue Type
Build/Install
### Have you reproduced the bug with TF nightly?
No
### Source
source
### Tensorflow Version
2.9.1
### Custom Code
No
### OS Platform and Distribution
Linux RHEL 7
### Python version
3.10
### Bazel version
5.1.1
### GCC/Compiler version
11.3
### Current Behaviour?
```shell
I see failing tests while running `bazel test`.
The `testGetMemoryInfoCPU` is guarded by a skip `if test_util.IsMklEnabled():`
And the other tests has a similar condition.
So it looks like `IsMklEnabled` doesn't return true when it should. Further investigation leads to several macros in the build files:
- `IsMklEnabled` returns true when `defined(INTEL_MKL) && defined(ENABLE_MKL)`
- There is `if_mkl([":mkl_cpu_allocator"])` which is why `testGetMemoryInfoCPU` fails when true
- `if_mkl` is true on `linux_x86_64` unconditionally
- `if_mkl(["-DINTEL_MKL"])`
- `if_enable_mkl(["-DENABLE_MKL"])`
I'd suggest to simplify this so either there is MKL or there is not. Having multiple macros with different defaults doesn't make sense to me. Or maybe the default for "ENABLE_MKL" should be the same as for "INTEL_MKL", e.g. True on x86
```
### Standalone code to reproduce the issue
```shell
bazel test
```
### Relevant log output
```shell
FAIL: testGetMemoryInfoCPU (__main__.ContextTest)
ContextTest.testGetMemoryInfoCPU
----------------------------------------------------------------------
Traceback (most recent call last):
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/eager/context_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/test_util.py", line 2229, in decorated
return func(self, *args, **kwargs)
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/eager/context_test_cpu.runfiles/org_tensorflow/tensorflow/python/eager/context_test.py", line 141, in testGetMemoryInfoCPU
with self.assertRaisesRegex(ValueError, 'Allocator stats not available'):
AssertionError: ValueError not raised
FAIL: test_simple (__main__.NodeFileWriterTest)
NodeFileWriterTest.test_simple
----------------------------------------------------------------------
Traceback (most recent call last):
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/framework/node_file_writer_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/test_util.py", line 2157, in decorated
return func(self, *args, **kwargs)
File "/dev/shm/jfg508/TensorFlow/2.9.1/foss-2022a/TensorFlow/bazel-root/bac5198b45911f6921886c0013c301e8/execroot/org_tensorflow/bazel-out/k8-opt/bin/tensorflow/python/framework/node_file_writer_test_cpu.runfiles/org_tensorflow/tensorflow/python/framework/node_file_writer_test.py", line 142, in test_simple
self.assertEqual(node_def1.op, 'MatMul')
AssertionError:
- _MklMatMul
? ----
+ MatMul
```
</details> | non_comp | test failure due to inconsistencies related to mkl click to expand issue type build install have you reproduced the bug with tf nightly no source source tensorflow version custom code no os platform and distribution linux rhel python version bazel version gcc compiler version current behaviour shell i see failing tests while running bazel test the testgetmemoryinfocpu is guarded by a skip if test util ismklenabled and the other tests has a similar condition so it looks like ismklenabled doesn t return true when it should further investigation leads to several macros in the build files ismklenabled returns true when defined intel mkl defined enable mkl there is if mkl which is why testgetmemoryinfocpu fails when true if mkl is true on linux unconditionally if mkl if enable mkl i d suggest to simplify this so either there is mkl or there is not having multiple macros with different defaults doesn t make sense to me or maybe the default for enable mkl should be the same as for intel mkl e g true on standalone code to reproduce the issue shell bazel test relevant log output shell fail testgetmemoryinfocpu main contexttest contexttest testgetmemoryinfocpu traceback most recent call last file dev shm tensorflow foss tensorflow bazel root execroot org tensorflow bazel out opt bin tensorflow python eager context test cpu runfiles org tensorflow tensorflow python framework test util py line in decorated return func self args kwargs file dev shm tensorflow foss tensorflow bazel root execroot org tensorflow bazel out opt bin tensorflow python eager context test cpu runfiles org tensorflow tensorflow python eager context test py line in testgetmemoryinfocpu with self assertraisesregex valueerror allocator stats not available assertionerror valueerror not raised fail test simple main nodefilewritertest nodefilewritertest test simple traceback most recent call last file dev shm tensorflow foss tensorflow bazel root execroot org tensorflow bazel out opt bin tensorflow python framework node file writer test cpu runfiles org tensorflow tensorflow python framework test util py line in decorated return func self args kwargs file dev shm tensorflow foss tensorflow bazel root execroot org tensorflow bazel out opt bin tensorflow python framework node file writer test cpu runfiles org tensorflow tensorflow python framework node file writer test py line in test simple self assertequal node op matmul assertionerror mklmatmul matmul | 0 |
7,525 | 7,964,278,598 | IssuesEvent | 2018-07-13 20:47:26 | PATRIC3/patric3_website | https://api.github.com/repos/PATRIC3/patric3_website | closed | Genome Comparison: option to open image in a new tab | Service: Others | In the circular genome viewer, I can right click on the circular image and open it in a new tab, so that I get the full view and I can zoom in / out.
For the same circular view in the genome comparison viewer, when I right click, open image in new tab option is greyed out. Any idea how we can enable it?

| 1.0 | Genome Comparison: option to open image in a new tab - In the circular genome viewer, I can right click on the circular image and open it in a new tab, so that I get the full view and I can zoom in / out.
For the same circular view in the genome comparison viewer, when I right click, open image in new tab option is greyed out. Any idea how we can enable it?

| non_comp | genome comparison option to open image in a new tab in the circular genome viewer i can right click on the circular image and open it in a new tab so that i get the full view and i can zoom in out for the same circular view in the genome comparison viewer when i right click open image in new tab option is greyed out any idea how we can enable it | 0 |
10,796 | 12,799,445,408 | IssuesEvent | 2020-07-02 15:24:26 | Hacktix/ChromaBoy | https://api.github.com/repos/Hacktix/ChromaBoy | opened | [Wario Land II] - Game resets after intro sequence | bug game compatibility | After starting a new game and having the first intro sequence played the game resets to the opening sequence and loads back into the main menu. | True | [Wario Land II] - Game resets after intro sequence - After starting a new game and having the first intro sequence played the game resets to the opening sequence and loads back into the main menu. | comp | game resets after intro sequence after starting a new game and having the first intro sequence played the game resets to the opening sequence and loads back into the main menu | 1 |
831,615 | 32,056,757,768 | IssuesEvent | 2023-09-24 06:59:28 | AY2324S1-CS2103T-W08-1/tp | https://api.github.com/repos/AY2324S1-CS2103T-W08-1/tp | opened | Update AboutUs page | priority.high | <ul>
<li>Add own details and a suitable photo</li>
<li>Edit the filename of the profile photo to <code>docs/images/github_username_in_lower_case.png</code></li>
<li>Indicate the different roles played and responsibilities held by each team member.</li>
</ul> | 1.0 | Update AboutUs page - <ul>
<li>Add own details and a suitable photo</li>
<li>Edit the filename of the profile photo to <code>docs/images/github_username_in_lower_case.png</code></li>
<li>Indicate the different roles played and responsibilities held by each team member.</li>
</ul> | non_comp | update aboutus page add own details and a suitable photo edit the filename of the profile photo to docs images github username in lower case png indicate the different roles played and responsibilities held by each team member | 0 |
16,432 | 22,197,232,600 | IssuesEvent | 2022-06-07 08:05:10 | aesara-devs/aesara | https://api.github.com/repos/aesara-devs/aesara | opened | Update of random variables not working in JAX backend | JAX backend compatibility | ```python
import aesara
import aesara.tensor as at
import numpy as np
rng = aesara.shared(np.random.default_rng())
x = at.random.normal(rng=rng)
f = aesara.function([], x, updates={rng: x.owner.outputs[0]}, mode="JAX")
assert f() != f() # Fails
``` | True | Update of random variables not working in JAX backend - ```python
import aesara
import aesara.tensor as at
import numpy as np
rng = aesara.shared(np.random.default_rng())
x = at.random.normal(rng=rng)
f = aesara.function([], x, updates={rng: x.owner.outputs[0]}, mode="JAX")
assert f() != f() # Fails
``` | comp | update of random variables not working in jax backend python import aesara import aesara tensor as at import numpy as np rng aesara shared np random default rng x at random normal rng rng f aesara function x updates rng x owner outputs mode jax assert f f fails | 1 |
591 | 8,028,031,203 | IssuesEvent | 2018-07-27 11:17:34 | svaarala/duktape | https://api.github.com/repos/svaarala/duktape | closed | Check portability assumptions for realloc() with zero size | portability | Behavior for zero size in C99/C++11 is implemented defined: http://www.cplusplus.com/reference/cstdlib/realloc/:
> If size is zero, the return value depends on the particular library implementation: it may either be a null pointer or some other location that shall not be dereferenced.
For Duktape this means that:
- If the return value is NULL, free() can be called but is not required.
- If the return value is non-NULL, Duktape should call free() when the block is freed.
This should be the current behavior already, but go through the call sites to ensure that.
In particular, the library implementation can essentially ignore a realloc(0) and leave the memory allocated, returning a non-NULL value. Duktape is then required to call free(). | True | Check portability assumptions for realloc() with zero size - Behavior for zero size in C99/C++11 is implemented defined: http://www.cplusplus.com/reference/cstdlib/realloc/:
> If size is zero, the return value depends on the particular library implementation: it may either be a null pointer or some other location that shall not be dereferenced.
For Duktape this means that:
- If the return value is NULL, free() can be called but is not required.
- If the return value is non-NULL, Duktape should call free() when the block is freed.
This should be the current behavior already, but go through the call sites to ensure that.
In particular, the library implementation can essentially ignore a realloc(0) and leave the memory allocated, returning a non-NULL value. Duktape is then required to call free(). | non_comp | check portability assumptions for realloc with zero size behavior for zero size in c is implemented defined if size is zero the return value depends on the particular library implementation it may either be a null pointer or some other location that shall not be dereferenced for duktape this means that if the return value is null free can be called but is not required if the return value is non null duktape should call free when the block is freed this should be the current behavior already but go through the call sites to ensure that in particular the library implementation can essentially ignore a realloc and leave the memory allocated returning a non null value duktape is then required to call free | 0 |
84,171 | 7,893,298,090 | IssuesEvent | 2018-06-28 17:35:13 | mozilla/iris | https://api.github.com/repos/mozilla/iris | opened | Fix web_console_controls.py | bug test case | A general overhaul of this test case is required, This feature's UI has changed, the images for the following, no longer exist. The functionality for these buttons was moved into a menu.
dock_button = 'dock_to_side.png'
dock_button_activated = 'dock_to_side_activated.png'
separate_window_button = 'separate_window.png'
Please rework the test case to use the controls from inside the menu. Also, please discontinue use of hard coded regions. | 1.0 | Fix web_console_controls.py - A general overhaul of this test case is required, This feature's UI has changed, the images for the following, no longer exist. The functionality for these buttons was moved into a menu.
dock_button = 'dock_to_side.png'
dock_button_activated = 'dock_to_side_activated.png'
separate_window_button = 'separate_window.png'
Please rework the test case to use the controls from inside the menu. Also, please discontinue use of hard coded regions. | non_comp | fix web console controls py a general overhaul of this test case is required this feature s ui has changed the images for the following no longer exist the functionality for these buttons was moved into a menu dock button dock to side png dock button activated dock to side activated png separate window button separate window png please rework the test case to use the controls from inside the menu also please discontinue use of hard coded regions | 0 |
9,043 | 11,090,942,708 | IssuesEvent | 2019-12-15 08:28:28 | tenpy/tenpy | https://api.github.com/repos/tenpy/tenpy | closed | combine_legs qconj default argument | incompatible | It might be more consistent to make the default value of `qconj` the `qconj` value of the first leg included into the pipe, instead of just +1. However, this breaks backwards compatibility with existing code. So we should:
- [x] Change the default to `None` in TeNPy v. 0.5.0 and issue a Warning if not set explicitly that the behaviour will change.
- [x] Fix all the warnings inside TeNPy
- [ ] Change the behaviour in the next release. | True | combine_legs qconj default argument - It might be more consistent to make the default value of `qconj` the `qconj` value of the first leg included into the pipe, instead of just +1. However, this breaks backwards compatibility with existing code. So we should:
- [x] Change the default to `None` in TeNPy v. 0.5.0 and issue a Warning if not set explicitly that the behaviour will change.
- [x] Fix all the warnings inside TeNPy
- [ ] Change the behaviour in the next release. | comp | combine legs qconj default argument it might be more consistent to make the default value of qconj the qconj value of the first leg included into the pipe instead of just however this breaks backwards compatibility with existing code so we should change the default to none in tenpy v and issue a warning if not set explicitly that the behaviour will change fix all the warnings inside tenpy change the behaviour in the next release | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.