Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
72,620 | 24,202,637,875 | IssuesEvent | 2022-09-24 19:44:49 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | closed | Jitsi preferredDomain is not taken into account | T-Defect | ### Steps to reproduce
I'm encountering the exact same issue as https://github.com/vector-im/element-android/issues/4497 with Element Android 1.4.36.
I'm serving the following `.well-known` correctly under both
https://matrix.domain.tld/.well-known/matrix/client and https://domain.tld/.well-known/matrix/client:
```
{
"m.homeserver": {"base_url": "https://matrix.domain.tld:443"},
"im.vector.riot.jitsi": {"preferredDomain": "meet.domain.tld"}
}
```
When the call is initiated from Element Web, it uses the preferredDomain `meet.domain.tld` correctly. The following config is also defined in `element/config.json`:
```
"jitsi": {
"preferred_domain": "meet.domain.tld"
},
```
But when the call is initiated from Element Android, it uses `meet.element.io`. The event source when the call is initiated from Android is as follows:
```
{
"type": "im.vector.modular.widgets",
"sender": "@user:matrix.domain.tld",
"content": {
"url": "https://app.element.io/jitsi.html?confId=MmUsjWstqvCXVTQLp44b9f8e#conferenceDomain=$domain&conferenceId=$conferenceId&isAudioOnly=$isAudioOnly&displayName=$matrix_display_name&avatarUrl=$matrix_avatar_url&userId=$matrix_user_id&roomId=$matrix_room_id&theme=$theme",
"type": "jitsi",
"data": {
"domain": "meet.element.io",
"conferenceId": "MmUsjWstqvCXVTQLp44b9f8e",
"isAudioOnly": false
},
"creatorUserId": "@user:matrix.domain.tld",
"id": "m.jitsi_@user:matrix.domain.tld_1664044283605",
"name": "jitsi",
"roomId": "!MmUsjWstqvCXVTQLpD:matrix.domain.tld",
"eventId": "$c1hJiQNlpE6ATztH2KrIG3ydu9kmxOVBtd_p3MTTD-M"
},
"state_key": "m.jitsi_@user:matrix.domain.tld_1664044283605",
"origin_server_ts": 1664044282888,
"unsigned": {
"age": 64
},
"event_id": "$c1hJiQNlpE6ATztH2KrIG3ydu9kmxOVBtd_p3MTTD-M",
"room_id": "!MmUsjWstqvCXVTQLpD:matrix.domain.tld"
}
```
### Outcome
#### What did you expect?
According to [this](https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md#element-android), the call was supposed to use the custom Jitsi domain `meet.domain.tld`
#### What happened instead?
Call was setup using `meet.element.io`
### Your phone model
Pixel 5
### Operating system version
Android 12
### Application version and app store
Element 1.4.36
### Homeserver
Synapse 1.67.0+bullseye1
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | 1.0 | Jitsi preferredDomain is not taken into account - ### Steps to reproduce
I'm encountering the exact same issue as https://github.com/vector-im/element-android/issues/4497 with Element Android 1.4.36.
I'm serving the following `.well-known` correctly under both
https://matrix.domain.tld/.well-known/matrix/client and https://domain.tld/.well-known/matrix/client:
```
{
"m.homeserver": {"base_url": "https://matrix.domain.tld:443"},
"im.vector.riot.jitsi": {"preferredDomain": "meet.domain.tld"}
}
```
When the call is initiated from Element Web, it uses the preferredDomain `meet.domain.tld` correctly. The following config is also defined in `element/config.json`:
```
"jitsi": {
"preferred_domain": "meet.domain.tld"
},
```
But when the call is initiated from Element Android, it uses `meet.element.io`. The event source when the call is initiated from Android is as follows:
```
{
"type": "im.vector.modular.widgets",
"sender": "@user:matrix.domain.tld",
"content": {
"url": "https://app.element.io/jitsi.html?confId=MmUsjWstqvCXVTQLp44b9f8e#conferenceDomain=$domain&conferenceId=$conferenceId&isAudioOnly=$isAudioOnly&displayName=$matrix_display_name&avatarUrl=$matrix_avatar_url&userId=$matrix_user_id&roomId=$matrix_room_id&theme=$theme",
"type": "jitsi",
"data": {
"domain": "meet.element.io",
"conferenceId": "MmUsjWstqvCXVTQLp44b9f8e",
"isAudioOnly": false
},
"creatorUserId": "@user:matrix.domain.tld",
"id": "m.jitsi_@user:matrix.domain.tld_1664044283605",
"name": "jitsi",
"roomId": "!MmUsjWstqvCXVTQLpD:matrix.domain.tld",
"eventId": "$c1hJiQNlpE6ATztH2KrIG3ydu9kmxOVBtd_p3MTTD-M"
},
"state_key": "m.jitsi_@user:matrix.domain.tld_1664044283605",
"origin_server_ts": 1664044282888,
"unsigned": {
"age": 64
},
"event_id": "$c1hJiQNlpE6ATztH2KrIG3ydu9kmxOVBtd_p3MTTD-M",
"room_id": "!MmUsjWstqvCXVTQLpD:matrix.domain.tld"
}
```
### Outcome
#### What did you expect?
According to [this](https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md#element-android), the call was supposed to use the custom Jitsi domain `meet.domain.tld`
#### What happened instead?
Call was setup using `meet.element.io`
### Your phone model
Pixel 5
### Operating system version
Android 12
### Application version and app store
Element 1.4.36
### Homeserver
Synapse 1.67.0+bullseye1
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | defect | jitsi preferreddomain is not taken into account steps to reproduce i m encountering the exact same issue as with element android i m serving the following well known correctly under both and m homeserver base url im vector riot jitsi preferreddomain meet domain tld when the call is initiated from element web it uses the preferreddomain meet domain tld correctly the following config is also defined in element config json jitsi preferred domain meet domain tld but when the call is initiated from element android it uses meet element io the event source when the call is initiated from android is as follows type im vector modular widgets sender user matrix domain tld content url type jitsi data domain meet element io conferenceid isaudioonly false creatoruserid user matrix domain tld id m jitsi user matrix domain tld name jitsi roomid mmusjwstqvcxvtqlpd matrix domain tld eventid m state key m jitsi user matrix domain tld origin server ts unsigned age event id m room id mmusjwstqvcxvtqlpd matrix domain tld outcome what did you expect according to the call was supposed to use the custom jitsi domain meet domain tld what happened instead call was setup using meet element io your phone model pixel operating system version android application version and app store element homeserver synapse will you send logs no are you willing to provide a pr yes | 1 |
445,805 | 31,331,429,267 | IssuesEvent | 2023-08-24 00:10:38 | brett-buskirk/rc-journey | https://api.github.com/repos/brett-buskirk/rc-journey | opened | Requirements for new contributers | documentation | Since this project requires additional contributors to add their own pages, some requirements need to be set. Some of these will be more hardset, like those needed for the underlying codebase. Others will be more community and stylistically based, which will become more pertinent once the project gains some momemtum. | 1.0 | Requirements for new contributers - Since this project requires additional contributors to add their own pages, some requirements need to be set. Some of these will be more hardset, like those needed for the underlying codebase. Others will be more community and stylistically based, which will become more pertinent once the project gains some momemtum. | non_defect | requirements for new contributers since this project requires additional contributors to add their own pages some requirements need to be set some of these will be more hardset like those needed for the underlying codebase others will be more community and stylistically based which will become more pertinent once the project gains some momemtum | 0 |
34,450 | 2,781,001,478 | IssuesEvent | 2015-05-06 09:58:41 | kromkrom/wordcontrol | https://api.github.com/repos/kromkrom/wordcontrol | opened | Display matched spelling in lexeme list | issue:change_request priority:-1 | Lexeme in the list must be accompanied by a wordform, whose spelling it was filtered by. | 1.0 | Display matched spelling in lexeme list - Lexeme in the list must be accompanied by a wordform, whose spelling it was filtered by. | non_defect | display matched spelling in lexeme list lexeme in the list must be accompanied by a wordform whose spelling it was filtered by | 0 |
196,915 | 15,613,787,982 | IssuesEvent | 2021-03-19 16:54:18 | devfile/api | https://api.github.com/repos/devfile/api | closed | allow `.devfile.yaml` | area/documentation kind/enhancement severity/P1 | (not sure if this is part of kubernets-api for devfile but it should for sure be part of the devfile spec - let me know if any better place this issue)
originally opened at https://github.com/openshift/odo/issues/3126 and https://github.com/openshift/odo/issues/3126
Simply asking that `.devfile.yaml` should be allowed for the name of the devfile to allow it to be hidden by default and not take up "visual space" in projects that create these files. | 1.0 | allow `.devfile.yaml` - (not sure if this is part of kubernets-api for devfile but it should for sure be part of the devfile spec - let me know if any better place this issue)
originally opened at https://github.com/openshift/odo/issues/3126 and https://github.com/openshift/odo/issues/3126
Simply asking that `.devfile.yaml` should be allowed for the name of the devfile to allow it to be hidden by default and not take up "visual space" in projects that create these files. | non_defect | allow devfile yaml not sure if this is part of kubernets api for devfile but it should for sure be part of the devfile spec let me know if any better place this issue originally opened at and simply asking that devfile yaml should be allowed for the name of the devfile to allow it to be hidden by default and not take up visual space in projects that create these files | 0 |
294,712 | 22,161,162,402 | IssuesEvent | 2022-06-04 14:28:27 | LotusZadrotus/homepage | https://api.github.com/repos/LotusZadrotus/homepage | closed | Create a contact info block | documentation enhancement Git Homework | Add links to social networks (LinkedIn, Facebook, Instagram), phone number, email, Telegram username, etc.
| 1.0 | Create a contact info block - Add links to social networks (LinkedIn, Facebook, Instagram), phone number, email, Telegram username, etc.
| non_defect | create a contact info block add links to social networks linkedin facebook instagram phone number email telegram username etc | 0 |
32,189 | 6,733,662,147 | IssuesEvent | 2017-10-18 15:27:58 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | hit: linking fails on macs sometimes | C: MOOSE T: defect | ## Rationale
When building MOOSE on Macs under some circumstances linking fails with missing symbol errors for common things like ``std::string``.
## Description
Using the Mac's default system python on a machine with no 32 bit developer libraries installed (e.g. 32 bit compatible c++ stdlib), if you try to build MOOSE the step that builds/links the python hit bindings fails with linker errors. The ``python-config --cflags`` output contains a ``-arch i386`` entry that we don't want.
## Impact
Make MOOSE compile successfully under more diverse conditions so users don't have to do silly work-arounds.
| 1.0 | hit: linking fails on macs sometimes - ## Rationale
When building MOOSE on Macs under some circumstances linking fails with missing symbol errors for common things like ``std::string``.
## Description
Using the Mac's default system python on a machine with no 32 bit developer libraries installed (e.g. 32 bit compatible c++ stdlib), if you try to build MOOSE the step that builds/links the python hit bindings fails with linker errors. The ``python-config --cflags`` output contains a ``-arch i386`` entry that we don't want.
## Impact
Make MOOSE compile successfully under more diverse conditions so users don't have to do silly work-arounds.
| defect | hit linking fails on macs sometimes rationale when building moose on macs under some circumstances linking fails with missing symbol errors for common things like std string description using the mac s default system python on a machine with no bit developer libraries installed e g bit compatible c stdlib if you try to build moose the step that builds links the python hit bindings fails with linker errors the python config cflags output contains a arch entry that we don t want impact make moose compile successfully under more diverse conditions so users don t have to do silly work arounds | 1 |
3,834 | 2,610,069,861 | IssuesEvent | 2015-02-26 18:20:29 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 台州割包茎哪里效果最好 | auto-migrated Priority-Medium Type-Defect | ```
台州割包茎哪里效果最好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:50 | 1.0 | 台州割包茎哪里效果最好 - ```
台州割包茎哪里效果最好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:50 | defect | 台州割包茎哪里效果最好 台州割包茎哪里效果最好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
42,978 | 11,410,700,221 | IssuesEvent | 2020-02-01 00:30:09 | libarchive/libarchive | https://api.github.com/repos/libarchive/libarchive | closed | Use after free when creating ISO9660 or XAR with hardlinks | Type-Defect | When creating ISO9660 or XAR archives with hardlinked entries, the functions isofile_free_hardlinks() or file_free_hardlinks() end up in a use after free. | 1.0 | Use after free when creating ISO9660 or XAR with hardlinks - When creating ISO9660 or XAR archives with hardlinked entries, the functions isofile_free_hardlinks() or file_free_hardlinks() end up in a use after free. | defect | use after free when creating or xar with hardlinks when creating or xar archives with hardlinked entries the functions isofile free hardlinks or file free hardlinks end up in a use after free | 1 |
121,497 | 10,170,396,029 | IssuesEvent | 2019-08-08 05:07:59 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | teamcity: failed test: TestMergeQueue | C-test-failure O-robot | The following tests appear to have failed on master (testrace): TestMergeQueue/sanity, TestMergeQueue/combined-threshold, TestMergeQueue, TestMergeQueue/sticky-bit, TestMergeQueue/lhs-undersize, TestMergeQueue/both-empty, TestMergeQueue/non-collocated
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestMergeQueue).
[#1428523](https://teamcity.cockroachdb.com/viewLog.html?buildId=1428523):
```
TestMergeQueue/non-collocated
...orage.go:829 [s2,r10/?:{b-c}] applied PREEMPTIVE snapshot in 320ms [clear=0ms batch=229ms entries=0ms commit=90ms]
I190807 19:00:07.904950 157073 storage/replica_command.go:1188 [s1,r10/1:{b-c}] change replicas (ADD_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n1,s1):1, next=2, gen=15]
I190807 19:00:07.915726 157073 storage/replica_raft.go:290 [s1,r10/1:{b-c},txn=93480a64] proposing ADD_REPLICA((n2,s2):2): updated=(n1,s1):1,(n2,s2):2 next=3
I190807 19:00:08.332247 157073 storage/replica_command.go:1188 [s2,r10/2:{b-c}] change replicas (REMOVE_REPLICA (n1,s1):1): existing descriptor r10:{b-c} [(n1,s1):1, (n2,s2):2, next=3, gen=16]
I190807 19:00:08.360906 157073 storage/replica_raft.go:290 [s2,r10/2:{b-c},txn=f9bde25a] proposing REMOVE_REPLICA((n1,s1):1): updated=(n2,s2):2 next=3
I190807 19:00:08.414632 170862 storage/store.go:2530 [replicaGC,s1,r10/1:{b-c}] removing replica r10/1
I190807 19:00:08.416044 170862 storage/replica_destroy.go:146 [replicaGC,s1,r10/1:{b-c}] removed 6 (1+5) keys in 1ms [clear=0ms commit=0ms]
W190807 19:00:08.416810 170888 storage/replica_command.go:1725 [merge,s2,r10/2:{b-c}] while adding target n1,s1: snapshot failed: (n1,s1):?: remote couldn't accept PREEMPTIVE snapshot a0c5420f at applied index 21 with error: [n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})
E190807 19:00:08.425903 170713 storage/queue.go:1027 [replicate,s2,r10/2:{b-c}] 1 matching stores are currently throttled: [[n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})]
I190807 19:00:08.426036 170713 storage/queue.go:1127 [replicate] purgatory is now empty
I190807 19:00:08.426725 170888 storage/store_snapshot.go:775 [merge,s2,r10/2:{b-c}] sending PREEMPTIVE snapshot 5a6de673 at applied index 21
I190807 19:00:08.427752 170888 storage/store_snapshot.go:818 [merge,s2,r10/2:{b-c}] streamed snapshot to (n1,s1):?: kv pairs: 21, log entries: 0, rate-limit: 8.0 MiB/sec, 0.00s
I190807 19:00:08.430891 170882 storage/replica_raftstorage.go:823 [s1,r10/?:{-}] applying PREEMPTIVE snapshot at index 21 (id=5a6de673, encoded size=1102, 1 rocksdb batches, 0 log entries)
I190807 19:00:08.432871 170882 storage/replica_raftstorage.go:829 [s1,r10/?:{b-c}] applied PREEMPTIVE snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I190807 19:00:08.442618 170888 storage/replica_command.go:1188 [merge,s2,r10/2:{b-c}] change replicas (ADD_REPLICA (n1,s1):3): existing descriptor r10:{b-c} [(n2,s2):2, next=3, gen=17]
I190807 19:00:08.460010 170888 storage/replica_raft.go:290 [merge,s2,r10/2:{b-c},txn=4cb784e3] proposing ADD_REPLICA((n1,s1):3): updated=(n2,s2):2,(n1,s1):3 next=4
I190807 19:00:08.495173 170888 storage/replica_command.go:1188 [merge,s1,r10/3:{b-c}] change replicas (REMOVE_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n2,s2):2, (n1,s1):3, next=4, gen=18]
I190807 19:00:08.510995 170888 storage/replica_raft.go:290 [merge,s1,r10/3:{b-c},txn=647f6935] proposing REMOVE_REPLICA((n2,s2):2): updated=(n1,s1):3 next=4
I190807 19:00:08.526619 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.538611 171003 storage/store.go:2530 [replicaGC,s2,r10/2:{b-c}] removing replica r10/2
I190807 19:00:08.540067 171003 storage/replica_destroy.go:146 [replicaGC,s2,r10/2:{b-c}] removed 7 (0+7) keys in 1ms [clear=0ms commit=0ms]
I190807 19:00:08.689040 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.877603 85667 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=bac592ef] removing replica r10/3
TestMergeQueue
--- FAIL: testrace/TestMergeQueue (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:57:01.616659 85583 gossip/gossip.go:394 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39043" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
W190807 18:57:01.710337 85583 gossip/gossip.go:1498 [n2] no incoming or outgoing connections
I190807 18:57:01.711139 85583 gossip/gossip.go:394 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:33469" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
I190807 18:57:01.730521 85876 gossip/client.go:124 [n2] started gossip client to 127.0.0.1:39043
I190807 18:57:01.795336 85583 storage/client_test.go:495 gossip network initialized
I190807 18:57:01.803889 85583 storage/replica_command.go:283 [s1,r1/1:/M{in-ax}] initiating a split of this range at key "a" [r2] (manual)
I190807 18:57:01.850586 85583 storage/replica_command.go:283 [s1,r2/1:{a-/Max}] initiating a split of this range at key "b" [r3] (manual)
I190807 18:57:01.899827 85583 storage/replica_command.go:283 [s1,r3/1:{b-/Max}] initiating a split of this range at key "c" [r4] (manual)
I190807 18:57:03.850154 85892 gossip/gossip.go:1512 [n2] node has connected to cluster via gossip
TestMergeQueue/lhs-undersize
--- FAIL: testrace/TestMergeQueue/lhs-undersize (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:08.782829 85658 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:59:08.807560 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.808047 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:59:08.843036 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.847447 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:59:09.705064 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:11.884279 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:11.886797 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:59:11.887846 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:59:23.526013 85698 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.0s [applied=1, batches=1, state_assertions=0]
I190807 18:59:23.529053 134859 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r8] (manual)
I190807 18:59:25.341106 134859 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r8:{b-c} [(n1,s1):1, next=2, gen=11] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=4.62+0.00 --> 4.62qps) below threshold (size=32 MiB, qps=4.62))
I190807 18:59:25.537883 85653 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=f8f4916e] removing replica r8/1
TestMergeQueue/combined-threshold
--- FAIL: testrace/TestMergeQueue/combined-threshold (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:36.625896 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 9.2s [applied=1, batches=1, state_assertions=0]
I190807 18:59:40.539715 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:40.540850 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 32 MiB (reasons: size=false used=true avail=false)
I190807 18:59:40.541589 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 32 MiB in 0.0s
W190807 18:59:45.926124 85710 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 7.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:45.932898 147186 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r9] (manual)
I190807 18:59:46.913641 156969 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r9:{b-c} [(n1,s1):1, next=2, gen=13] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=0.00+0.00 --> 0.00qps) below threshold (size=32 MiB, qps=0.00))
I190807 18:59:47.128755 85651 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=3ac231da] removing replica r9/1
TestMergeQueue/sticky-bit
...heduler).Start.func2(0x5778e20, 0xc001f28660)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc000345e10, 0xc000dc55e0, 0xc000345e00)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 85697 [sync.Cond.Wait]:
runtime.goparkunlock(...)
/usr/local/go/src/runtime/proc.go:307
sync.runtime_notifyListWait(0xc0023d8290, 0xc000000a9c)
/usr/local/go/src/runtime/sema.go:510 +0xf9
sync.(*Cond).Wait(0xc0023d8280)
/usr/local/go/src/sync/cond.go:56 +0x8e
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc001140ab0, 0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:192 +0x9c
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0025a63a0, 0xc000dc55e0, 0xc0025a6390)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 171028 [select]:
github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher.(*RequestBatcher).Send(0xc0003acfc0, 0x5778da0, 0xc0000de010, 0xa, 0x57d9da0, 0xc000795200, 0x0, 0x1, 0x0, 0xc00038bcc0)
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher/batcher.go:240 +0x251
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).gcTxnRecord(0xc0003acd20, 0x5778da0, 0xc0000de010, 0xa, 0xc0016c7560, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:772 +0x571
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).cleanupFinishedTxnIntents.func2(0x5778da0, 0xc0000de010)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:812 +0x91
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1(0xc0011b6aa0, 0x5778da0, 0xc0000de010, 0xc000496880, 0x33, 0x0, 0x0, 0xc002cd4f60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:321 +0xf4
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:316 +0x14b
goroutine 85736 [select]:
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1(0x5778e20, 0xc001f29c80)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:748 +0x1ef
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0002faf00, 0xc000dc55e0, 0xc000416f40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
****************************************************************************
This node experienced a fatal error (printed above), and as a result the
process is terminating.
Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
problem in CockroachDB. With your help, the support team at Cockroach Labs
will try to determine the root cause, recommend next steps, and we can
improve CockroachDB based on your report.
Please submit a crash report by following the instructions here:
https://github.com/cockroachdb/cockroach/issues/new/choose
If you would rather not post publicly, please contact us directly at:
support@cockroachlabs.com
The Cockroach Labs team appreciates your feedback.
TestMergeQueue/sanity
...269] removing replica r5/1
I190807 18:57:56.238789 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:56.239373 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:58.884367 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:58.885544 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:59.419582 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:59.420066 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.237228 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.237968 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.416758 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.417527 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.061599 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.062161 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.064319 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.064742 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.540987 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1-2/2 ("a"-"c") for 32 MiB (reasons: size=false used=true avail=false)
W190807 18:58:11.103858 85661 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.5s [applied=1, batches=1, state_assertions=0]
W190807 18:58:11.560338 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 13.6s [applied=1, batches=1, state_assertions=0]
I190807 18:58:11.562378 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.562787 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.584071 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.584547 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.894549 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1-2/2 ("a"-"c") for 32 MiB in 2.4s
W190807 18:58:25.326862 85705 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.8s [applied=1, batches=1, state_assertions=0]
I190807 18:58:25.331543 96953 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r6] (manual)
TestMergeQueue/both-empty
--- FAIL: testrace/TestMergeQueue/both-empty (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:58:27.185018 122024 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r6:{b-c} [(n1,s1):1, next=2, gen=7] into this range (lhs+rhs has (size=0 B+0 B qps=4.82+0.00 --> 4.82qps) below threshold (size=0 B, qps=4.82))
I190807 18:58:27.790959 85705 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=7ac14de8] removing replica r6/1
W190807 18:58:38.204739 85668 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:38.213803 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.214397 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:38.266936 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.269692 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:39.805151 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:39.805623 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:58:39.807038 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:41.885436 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:58:41.888047 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:58:41.890478 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:58:54.874866 85670 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:54.877563 122008 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r7] (manual)
I190807 18:58:56.925560 134529 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r7:{b-c} [(n1,s1):1, next=2, gen=9] into this range (lhs+rhs has (size=0 B+0 B qps=3.95+0.00 --> 3.95qps) below threshold (size=0 B, qps=3.95))
I190807 18:58:57.181493 85682 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=42f7db6a] removing replica r7/1
```
Please assign, take a look and update the issue accordingly.
| 1.0 | teamcity: failed test: TestMergeQueue - The following tests appear to have failed on master (testrace): TestMergeQueue/sanity, TestMergeQueue/combined-threshold, TestMergeQueue, TestMergeQueue/sticky-bit, TestMergeQueue/lhs-undersize, TestMergeQueue/both-empty, TestMergeQueue/non-collocated
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestMergeQueue).
[#1428523](https://teamcity.cockroachdb.com/viewLog.html?buildId=1428523):
```
TestMergeQueue/non-collocated
...orage.go:829 [s2,r10/?:{b-c}] applied PREEMPTIVE snapshot in 320ms [clear=0ms batch=229ms entries=0ms commit=90ms]
I190807 19:00:07.904950 157073 storage/replica_command.go:1188 [s1,r10/1:{b-c}] change replicas (ADD_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n1,s1):1, next=2, gen=15]
I190807 19:00:07.915726 157073 storage/replica_raft.go:290 [s1,r10/1:{b-c},txn=93480a64] proposing ADD_REPLICA((n2,s2):2): updated=(n1,s1):1,(n2,s2):2 next=3
I190807 19:00:08.332247 157073 storage/replica_command.go:1188 [s2,r10/2:{b-c}] change replicas (REMOVE_REPLICA (n1,s1):1): existing descriptor r10:{b-c} [(n1,s1):1, (n2,s2):2, next=3, gen=16]
I190807 19:00:08.360906 157073 storage/replica_raft.go:290 [s2,r10/2:{b-c},txn=f9bde25a] proposing REMOVE_REPLICA((n1,s1):1): updated=(n2,s2):2 next=3
I190807 19:00:08.414632 170862 storage/store.go:2530 [replicaGC,s1,r10/1:{b-c}] removing replica r10/1
I190807 19:00:08.416044 170862 storage/replica_destroy.go:146 [replicaGC,s1,r10/1:{b-c}] removed 6 (1+5) keys in 1ms [clear=0ms commit=0ms]
W190807 19:00:08.416810 170888 storage/replica_command.go:1725 [merge,s2,r10/2:{b-c}] while adding target n1,s1: snapshot failed: (n1,s1):?: remote couldn't accept PREEMPTIVE snapshot a0c5420f at applied index 21 with error: [n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})
E190807 19:00:08.425903 170713 storage/queue.go:1027 [replicate,s2,r10/2:{b-c}] 1 matching stores are currently throttled: [[n1,s1],r10: cannot apply snapshot: snapshot intersects existing range; initiated GC: [n1,s1,r10/1:{b-c}] (incoming {b-c})]
I190807 19:00:08.426036 170713 storage/queue.go:1127 [replicate] purgatory is now empty
I190807 19:00:08.426725 170888 storage/store_snapshot.go:775 [merge,s2,r10/2:{b-c}] sending PREEMPTIVE snapshot 5a6de673 at applied index 21
I190807 19:00:08.427752 170888 storage/store_snapshot.go:818 [merge,s2,r10/2:{b-c}] streamed snapshot to (n1,s1):?: kv pairs: 21, log entries: 0, rate-limit: 8.0 MiB/sec, 0.00s
I190807 19:00:08.430891 170882 storage/replica_raftstorage.go:823 [s1,r10/?:{-}] applying PREEMPTIVE snapshot at index 21 (id=5a6de673, encoded size=1102, 1 rocksdb batches, 0 log entries)
I190807 19:00:08.432871 170882 storage/replica_raftstorage.go:829 [s1,r10/?:{b-c}] applied PREEMPTIVE snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I190807 19:00:08.442618 170888 storage/replica_command.go:1188 [merge,s2,r10/2:{b-c}] change replicas (ADD_REPLICA (n1,s1):3): existing descriptor r10:{b-c} [(n2,s2):2, next=3, gen=17]
I190807 19:00:08.460010 170888 storage/replica_raft.go:290 [merge,s2,r10/2:{b-c},txn=4cb784e3] proposing ADD_REPLICA((n1,s1):3): updated=(n2,s2):2,(n1,s1):3 next=4
I190807 19:00:08.495173 170888 storage/replica_command.go:1188 [merge,s1,r10/3:{b-c}] change replicas (REMOVE_REPLICA (n2,s2):2): existing descriptor r10:{b-c} [(n2,s2):2, (n1,s1):3, next=4, gen=18]
I190807 19:00:08.510995 170888 storage/replica_raft.go:290 [merge,s1,r10/3:{b-c},txn=647f6935] proposing REMOVE_REPLICA((n2,s2):2): updated=(n1,s1):3 next=4
I190807 19:00:08.526619 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.538611 171003 storage/store.go:2530 [replicaGC,s2,r10/2:{b-c}] removing replica r10/2
I190807 19:00:08.540067 171003 storage/replica_destroy.go:146 [replicaGC,s2,r10/2:{b-c}] removed 7 (0+7) keys in 1ms [clear=0ms commit=0ms]
I190807 19:00:08.689040 170888 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r10:{b-c} [(n1,s1):3, next=4, gen=19] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.00 --> 0.00qps) below threshold (size=0 B, qps=0.00))
I190807 19:00:08.877603 85667 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=bac592ef] removing replica r10/3
TestMergeQueue
--- FAIL: testrace/TestMergeQueue (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:57:01.616659 85583 gossip/gossip.go:394 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39043" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
W190807 18:57:01.710337 85583 gossip/gossip.go:1498 [n2] no incoming or outgoing connections
I190807 18:57:01.711139 85583 gossip/gossip.go:394 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:33469" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > build_tag:"" started_at:0 cluster_name:""
I190807 18:57:01.730521 85876 gossip/client.go:124 [n2] started gossip client to 127.0.0.1:39043
I190807 18:57:01.795336 85583 storage/client_test.go:495 gossip network initialized
I190807 18:57:01.803889 85583 storage/replica_command.go:283 [s1,r1/1:/M{in-ax}] initiating a split of this range at key "a" [r2] (manual)
I190807 18:57:01.850586 85583 storage/replica_command.go:283 [s1,r2/1:{a-/Max}] initiating a split of this range at key "b" [r3] (manual)
I190807 18:57:01.899827 85583 storage/replica_command.go:283 [s1,r3/1:{b-/Max}] initiating a split of this range at key "c" [r4] (manual)
I190807 18:57:03.850154 85892 gossip/gossip.go:1512 [n2] node has connected to cluster via gossip
TestMergeQueue/lhs-undersize
--- FAIL: testrace/TestMergeQueue/lhs-undersize (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:08.782829 85658 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:59:08.807560 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.808047 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:59:08.843036 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:59:08.847447 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:59:09.705064 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:11.884279 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:11.886797 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:59:11.887846 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:59:23.526013 85698 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.0s [applied=1, batches=1, state_assertions=0]
I190807 18:59:23.529053 134859 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r8] (manual)
I190807 18:59:25.341106 134859 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r8:{b-c} [(n1,s1):1, next=2, gen=11] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=4.62+0.00 --> 4.62qps) below threshold (size=32 MiB, qps=4.62))
I190807 18:59:25.537883 85653 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=f8f4916e] removing replica r8/1
TestMergeQueue/combined-threshold
--- FAIL: testrace/TestMergeQueue/combined-threshold (0.000s)
Test ended in panic.
------- Stdout: -------
W190807 18:59:36.625896 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 9.2s [applied=1, batches=1, state_assertions=0]
I190807 18:59:40.539715 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:59:40.540850 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 32 MiB (reasons: size=false used=true avail=false)
I190807 18:59:40.541589 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 32 MiB in 0.0s
W190807 18:59:45.926124 85710 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 7.1s [applied=1, batches=1, state_assertions=0]
I190807 18:59:45.932898 147186 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r9] (manual)
I190807 18:59:46.913641 156969 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r9:{b-c} [(n1,s1):1, next=2, gen=13] into this range (lhs+rhs has (size=16 MiB+16 MiB qps=0.00+0.00 --> 0.00qps) below threshold (size=32 MiB, qps=0.00))
I190807 18:59:47.128755 85651 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=3ac231da] removing replica r9/1
TestMergeQueue/sticky-bit
...heduler).Start.func2(0x5778e20, 0xc001f28660)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc000345e10, 0xc000dc55e0, 0xc000345e00)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 85697 [sync.Cond.Wait]:
runtime.goparkunlock(...)
/usr/local/go/src/runtime/proc.go:307
sync.runtime_notifyListWait(0xc0023d8290, 0xc000000a9c)
/usr/local/go/src/runtime/sema.go:510 +0xf9
sync.(*Cond).Wait(0xc0023d8280)
/usr/local/go/src/sync/cond.go:56 +0x8e
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc001140ab0, 0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:192 +0x9c
github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x5778e20, 0xc00193a1b0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:161 +0x56
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0025a63a0, 0xc000dc55e0, 0xc0025a6390)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
goroutine 171028 [select]:
github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher.(*RequestBatcher).Send(0xc0003acfc0, 0x5778da0, 0xc0000de010, 0xa, 0x57d9da0, 0xc000795200, 0x0, 0x1, 0x0, 0xc00038bcc0)
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/requestbatcher/batcher.go:240 +0x251
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).gcTxnRecord(0xc0003acd20, 0x5778da0, 0xc0000de010, 0xa, 0xc0016c7560, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:772 +0x571
github.com/cockroachdb/cockroach/pkg/storage/intentresolver.(*IntentResolver).cleanupFinishedTxnIntents.func2(0x5778da0, 0xc0000de010)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/intentresolver/intent_resolver.go:812 +0x91
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1(0xc0011b6aa0, 0x5778da0, 0xc0000de010, 0xc000496880, 0x33, 0x0, 0x0, 0xc002cd4f60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:321 +0xf4
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:316 +0x14b
goroutine 85736 [select]:
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1(0x5778e20, 0xc001f29c80)
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:748 +0x1ef
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc0002faf00, 0xc000dc55e0, 0xc000416f40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x160
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0xc4
****************************************************************************
This node experienced a fatal error (printed above), and as a result the
process is terminating.
Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
problem in CockroachDB. With your help, the support team at Cockroach Labs
will try to determine the root cause, recommend next steps, and we can
improve CockroachDB based on your report.
Please submit a crash report by following the instructions here:
https://github.com/cockroachdb/cockroach/issues/new/choose
If you would rather not post publicly, please contact us directly at:
support@cockroachlabs.com
The Cockroach Labs team appreciates your feedback.
TestMergeQueue/sanity
...269] removing replica r5/1
I190807 18:57:56.238789 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:56.239373 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:58.884367 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:58.885544 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:57:59.419582 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:57:59.420066 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.237228 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.237968 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:08.416758 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:08.417527 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.061599 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.062161 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.064319 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:09.064742 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:09.540987 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1-2/2 ("a"-"c") for 32 MiB (reasons: size=false used=true avail=false)
W190807 18:58:11.103858 85661 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.5s [applied=1, batches=1, state_assertions=0]
W190807 18:58:11.560338 85686 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 13.6s [applied=1, batches=1, state_assertions=0]
I190807 18:58:11.562378 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.562787 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.584071 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:11.584547 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:11.894549 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1-2/2 ("a"-"c") for 32 MiB in 2.4s
W190807 18:58:25.326862 85705 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.8s [applied=1, batches=1, state_assertions=0]
I190807 18:58:25.331543 96953 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r6] (manual)
TestMergeQueue/both-empty
--- FAIL: testrace/TestMergeQueue/both-empty (0.000s)
Test ended in panic.
------- Stdout: -------
I190807 18:58:27.185018 122024 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r6:{b-c} [(n1,s1):1, next=2, gen=7] into this range (lhs+rhs has (size=0 B+0 B qps=4.82+0.00 --> 4.82qps) below threshold (size=0 B, qps=4.82))
I190807 18:58:27.790959 85705 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=7ac14de8] removing replica r6/1
W190807 18:58:38.204739 85668 storage/store.go:3618 [s1,r1/1:{/Min-a}] handle raft ready: 0.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:38.213803 85893 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.214397 85893 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:38.266936 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:38.269692 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
I190807 18:58:39.805151 85596 storage/node_liveness.go:836 [liveness-hb] retrying liveness update after storage.errRetryLiveness: result is ambiguous (context deadline exceeded)
W190807 18:58:39.805623 85596 storage/node_liveness.go:484 [liveness-hb] failed node liveness heartbeat: operation "node liveness heartbeat" timed out after 450ms
W190807 18:58:39.807038 85676 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 10.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:41.885436 85724 storage/compactor/compactor.go:325 [s1,compactor] purging suggested compaction for range "a" - "b" that contains live data
I190807 18:58:41.888047 85724 storage/compactor/compactor.go:370 [s1,compactor] processing compaction #1/1 ("b"-"c") for 16 MiB (reasons: size=false used=true avail=false)
I190807 18:58:41.890478 85724 storage/compactor/compactor.go:386 [s1,compactor] processed compaction #1/1 ("b"-"c") for 16 MiB in 0.0s
W190807 18:58:54.874866 85670 storage/store.go:3618 [s1,r2/1:{a-c}] handle raft ready: 11.7s [applied=1, batches=1, state_assertions=0]
I190807 18:58:54.877563 122008 storage/replica_command.go:283 [s1,r2/1:{a-c}] initiating a split of this range at key "b" [r7] (manual)
I190807 18:58:56.925560 134529 storage/replica_command.go:597 [merge,s1,r2/1:{a-b}] initiating a merge of r7:{b-c} [(n1,s1):1, next=2, gen=9] into this range (lhs+rhs has (size=0 B+0 B qps=3.95+0.00 --> 3.95qps) below threshold (size=0 B, qps=3.95))
I190807 18:58:57.181493 85682 storage/store.go:2530 [merge,s1,r2/1:{a-b},txn=42f7db6a] removing replica r7/1
```
Please assign, take a look and update the issue accordingly.
| non_defect | teamcity failed test testmergequeue the following tests appear to have failed on master testrace testmergequeue sanity testmergequeue combined threshold testmergequeue testmergequeue sticky bit testmergequeue lhs undersize testmergequeue both empty testmergequeue non collocated you may want to check testmergequeue non collocated orage go applied preemptive snapshot in storage replica command go change replicas add replica existing descriptor b c storage replica raft go proposing add replica updated next storage replica command go change replicas remove replica existing descriptor b c storage replica raft go proposing remove replica updated next storage store go removing replica storage replica destroy go removed keys in storage replica command go while adding target snapshot failed remote couldn t accept preemptive snapshot at applied index with error cannot apply snapshot snapshot intersects existing range initiated gc incoming b c storage queue go matching stores are currently throttled cannot apply snapshot snapshot intersects existing range initiated gc incoming b c storage queue go purgatory is now empty storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica existing descriptor b c storage replica raft go proposing add replica updated next storage replica command go change replicas remove replica existing descriptor b c storage replica raft go proposing remove replica updated next storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica storage replica destroy go removed keys in storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica testmergequeue fail testrace testmergequeue test ended in panic stdout gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at cluster name gossip gossip go no incoming or outgoing connections gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at cluster name gossip client go started gossip client to storage client test go gossip network initialized storage replica command go initiating a split of this range at key a manual storage replica command go initiating a split of this range at key b manual storage replica command go initiating a split of this range at key c manual gossip gossip go node has connected to cluster via gossip testmergequeue lhs undersize fail testrace testmergequeue lhs undersize test ended in panic stdout storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size mib mib qps below threshold size mib qps storage store go removing replica testmergequeue combined threshold fail testrace testmergequeue combined threshold test ended in panic stdout storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size mib mib qps below threshold size mib qps storage store go removing replica testmergequeue sticky bit heduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine runtime goparkunlock usr local go src runtime proc go sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg internal client requestbatcher requestbatcher send go src github com cockroachdb cockroach pkg internal client requestbatcher batcher go github com cockroachdb cockroach pkg storage intentresolver intentresolver gctxnrecord go src github com cockroachdb cockroach pkg storage intentresolver intent resolver go github com cockroachdb cockroach pkg storage intentresolver intentresolver cleanupfinishedtxnintents go src github com cockroachdb cockroach pkg storage intentresolver intent resolver go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go this node experienced a fatal error printed above and as a result the process is terminating fatal errors can occur due to faulty hardware disks memory clocks or a problem in cockroachdb with your help the support team at cockroach labs will try to determine the root cause recommend next steps and we can improve cockroachdb based on your report please submit a crash report by following the instructions here if you would rather not post publicly please contact us directly at support cockroachlabs com the cockroach labs team appreciates your feedback testmergequeue sanity removing replica storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage compactor compactor go processing compaction a c for mib reasons size false used true avail false storage store go handle raft ready storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage compactor compactor go processed compaction a c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual testmergequeue both empty fail testrace testmergequeue both empty test ended in panic stdout storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica storage store go handle raft ready storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage node liveness go retrying liveness update after storage errretryliveness result is ambiguous context deadline exceeded storage node liveness go failed node liveness heartbeat operation node liveness heartbeat timed out after storage store go handle raft ready storage compactor compactor go purging suggested compaction for range a b that contains live data storage compactor compactor go processing compaction b c for mib reasons size false used true avail false storage compactor compactor go processed compaction b c for mib in storage store go handle raft ready storage replica command go initiating a split of this range at key b manual storage replica command go initiating a merge of b c into this range lhs rhs has size b b qps below threshold size b qps storage store go removing replica please assign take a look and update the issue accordingly | 0 |
4,181 | 2,610,088,806 | IssuesEvent | 2015-02-26 18:26:53 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳痤疮怎样去祛除 | auto-migrated Priority-Medium Type-Defect | ```
深圳痤疮怎样去祛除【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:30 | 1.0 | 深圳痤疮怎样去祛除 - ```
深圳痤疮怎样去祛除【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:30 | defect | 深圳痤疮怎样去祛除 深圳痤疮怎样去祛除【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 original issue reported on code google com by szft com on may at | 1 |
28,302 | 5,239,168,208 | IssuesEvent | 2017-01-31 08:53:24 | pexcn/tb-tun | https://api.github.com/repos/pexcn/tb-tun | closed | tb_userspace.c is not byte-order agnostic | auto-migrated Priority-Medium Type-Defect | ```
I was skimming through tb_userspace.c and noticed a lot of numeric constants
with a hard-coded little-endian byte order.
For example, 0x0220 is used to match "2002", which will certainly fail on any
big-endian architecture. Instead, you should match against htons(0x2002).
This bug exists all throughout the code.
For a reference on network programming, see:
http://beej.us/guide/bgnet/output/html/singlepage/bgnet.html
```
Original issue reported on code.google.com by `sparkm...@gmail.com` on 10 Dec 2010 at 3:30
| 1.0 | tb_userspace.c is not byte-order agnostic - ```
I was skimming through tb_userspace.c and noticed a lot of numeric constants
with a hard-coded little-endian byte order.
For example, 0x0220 is used to match "2002", which will certainly fail on any
big-endian architecture. Instead, you should match against htons(0x2002).
This bug exists all throughout the code.
For a reference on network programming, see:
http://beej.us/guide/bgnet/output/html/singlepage/bgnet.html
```
Original issue reported on code.google.com by `sparkm...@gmail.com` on 10 Dec 2010 at 3:30
| defect | tb userspace c is not byte order agnostic i was skimming through tb userspace c and noticed a lot of numeric constants with a hard coded little endian byte order for example is used to match which will certainly fail on any big endian architecture instead you should match against htons this bug exists all throughout the code for a reference on network programming see original issue reported on code google com by sparkm gmail com on dec at | 1 |
562,879 | 16,671,633,909 | IssuesEvent | 2021-06-07 11:39:25 | FirebirdSQL/firebird | https://api.github.com/repos/FirebirdSQL/firebird | closed | "Nested aggregate functions are not allowed" error on FB30 [CORE6256] | affect-version: 3.0.5 component: engine fix-version: 4.0.0 priority: major type: bug | Submitted by: Attila Molnár (e_pluribus_unum)
Hi\!
This works for FB25 \(Dialect1\), but not for FB30 \(Dialect3\)\.
SELECT
SUM\(
kf\.keszlet\_db \+
\(SELECT
SUM\(\(
CASE WHEN rt\.vdb IS NULL THEN rt\.db ELSE rt\.vdb END \- COALESCE\(rt\.ktdb, 0\)\) \*
\(SELECT kv\.szorzo
FROM ktfj\_valtoszam\(kf\.ktfj\_id, rt\.mert\_id, kf\.mert\_id\) kv\)\)
FROM ktrt rt\)\) szabad\_db
FROM ktkf kf
Our workaround : extract the subselect into CTE and then FB30 accepts it\.
WITH
sub AS
\(SELECT
\(CASE WHEN rt\.vdb IS NULL THEN rt\.db ELSE rt\.vdb END \- COALESCE\(rt\.ktdb, 0\)\) \*
\(SELECT kv\.szorzo
FROM ktfj\_valtoszam\(kf\.ktfj\_id, rt\.mert\_id, kf\.mert\_id\) kv\) db
FROM ktrt rt\)
SELECT
SUM\(
kf\.keszlet\_db \+
\(SELECT
SUM\(db\)
FROM sub\)\) szabad\_db
FROM ktkf kf
\(I can send the database on request\) | 1.0 | "Nested aggregate functions are not allowed" error on FB30 [CORE6256] - Submitted by: Attila Molnár (e_pluribus_unum)
Hi\!
This works for FB25 \(Dialect1\), but not for FB30 \(Dialect3\)\.
SELECT
SUM\(
kf\.keszlet\_db \+
\(SELECT
SUM\(\(
CASE WHEN rt\.vdb IS NULL THEN rt\.db ELSE rt\.vdb END \- COALESCE\(rt\.ktdb, 0\)\) \*
\(SELECT kv\.szorzo
FROM ktfj\_valtoszam\(kf\.ktfj\_id, rt\.mert\_id, kf\.mert\_id\) kv\)\)
FROM ktrt rt\)\) szabad\_db
FROM ktkf kf
Our workaround : extract the subselect into CTE and then FB30 accepts it\.
WITH
sub AS
\(SELECT
\(CASE WHEN rt\.vdb IS NULL THEN rt\.db ELSE rt\.vdb END \- COALESCE\(rt\.ktdb, 0\)\) \*
\(SELECT kv\.szorzo
FROM ktfj\_valtoszam\(kf\.ktfj\_id, rt\.mert\_id, kf\.mert\_id\) kv\) db
FROM ktrt rt\)
SELECT
SUM\(
kf\.keszlet\_db \+
\(SELECT
SUM\(db\)
FROM sub\)\) szabad\_db
FROM ktkf kf
\(I can send the database on request\) | non_defect | nested aggregate functions are not allowed error on submitted by attila molnár e pluribus unum hi this works for but not for select sum kf keszlet db select sum case when rt vdb is null then rt db else rt vdb end coalesce rt ktdb select kv szorzo from ktfj valtoszam kf ktfj id rt mert id kf mert id kv from ktrt rt szabad db from ktkf kf our workaround extract the subselect into cte and then accepts it with sub as select case when rt vdb is null then rt db else rt vdb end coalesce rt ktdb select kv szorzo from ktfj valtoszam kf ktfj id rt mert id kf mert id kv db from ktrt rt select sum kf keszlet db select sum db from sub szabad db from ktkf kf i can send the database on request | 0 |
21,254 | 2,637,226,561 | IssuesEvent | 2015-03-10 11:42:44 | bedita/bedita | https://api.github.com/repos/bedita/bedita | opened | General review of Newsletter module | Module - Newsletter Priority - Normal Topic - Modules Topic - UI Type - Enhancement | General improvements of newsletter module:
- [ ] create module specific CSS file
- [ ] remove inline CSS in invoice view
- [ ] various visual improvements
- [ ] javascript controls to avoid save and send without sender email
- [ ] send now feature: adds current date and time rounded by 5 minutes before saving
- [ ] sender name in mail template
- [ ] saving a template with empty body does not work | 1.0 | General review of Newsletter module - General improvements of newsletter module:
- [ ] create module specific CSS file
- [ ] remove inline CSS in invoice view
- [ ] various visual improvements
- [ ] javascript controls to avoid save and send without sender email
- [ ] send now feature: adds current date and time rounded by 5 minutes before saving
- [ ] sender name in mail template
- [ ] saving a template with empty body does not work | non_defect | general review of newsletter module general improvements of newsletter module create module specific css file remove inline css in invoice view various visual improvements javascript controls to avoid save and send without sender email send now feature adds current date and time rounded by minutes before saving sender name in mail template saving a template with empty body does not work | 0 |
198,134 | 6,970,349,846 | IssuesEvent | 2017-12-11 09:59:53 | PiKeyAr/sadx_dreamcast | https://api.github.com/repos/PiKeyAr/sadx_dreamcast | reopened | Windy Valley bridge model | low priority/wontfix SA1/SADX issue | Have a few more things, here.
1. There's a z-fighting problem with the top of the beams in Sky Deck 3, and also a texture looking quite warped. I believe this also happens in SADX.

2. Another handful of small z-fighting problems in Sky Deck Act 3.

There's a problem at each of the corners of these things.
3. Some more z-fighting in the corner here in Speed Highway Act 3.

On the opposite side of this area is yet another z-fighting issue.
4. The billboard objects in Speed Highway Act 3 also have layering issues.


Additionally, the entire object appears to be ignoring lighting. Only the ad itself is meant to ignore it. In the Dreamcast version, the border does not ignore lighting.

5. I noticed that the bridge model being used in Windy Valley isn't the same as the one in the Dreamcast version.
SADX:

Dreamcast:

It's possible to port the model over, right? Or is there something wrong with the model in SADX? Other levels use the same model as the Dreamcast version, with no ropes dangling below the bridge. Ex: Red Mountain uses the same model as the Dreamcast. | 1.0 | Windy Valley bridge model - Have a few more things, here.
1. There's a z-fighting problem with the top of the beams in Sky Deck 3, and also a texture looking quite warped. I believe this also happens in SADX.

2. Another handful of small z-fighting problems in Sky Deck Act 3.

There's a problem at each of the corners of these things.
3. Some more z-fighting in the corner here in Speed Highway Act 3.

On the opposite side of this area is yet another z-fighting issue.
4. The billboard objects in Speed Highway Act 3 also have layering issues.


Additionally, the entire object appears to be ignoring lighting. Only the ad itself is meant to ignore it. In the Dreamcast version, the border does not ignore lighting.

5. I noticed that the bridge model being used in Windy Valley isn't the same as the one in the Dreamcast version.
SADX:

Dreamcast:

It's possible to port the model over, right? Or is there something wrong with the model in SADX? Other levels use the same model as the Dreamcast version, with no ropes dangling below the bridge. Ex: Red Mountain uses the same model as the Dreamcast. | non_defect | windy valley bridge model have a few more things here there s a z fighting problem with the top of the beams in sky deck and also a texture looking quite warped i believe this also happens in sadx another handful of small z fighting problems in sky deck act there s a problem at each of the corners of these things some more z fighting in the corner here in speed highway act on the opposite side of this area is yet another z fighting issue the billboard objects in speed highway act also have layering issues additionally the entire object appears to be ignoring lighting only the ad itself is meant to ignore it in the dreamcast version the border does not ignore lighting i noticed that the bridge model being used in windy valley isn t the same as the one in the dreamcast version sadx dreamcast it s possible to port the model over right or is there something wrong with the model in sadx other levels use the same model as the dreamcast version with no ropes dangling below the bridge ex red mountain uses the same model as the dreamcast | 0 |
332,508 | 10,096,906,149 | IssuesEvent | 2019-07-27 22:55:23 | saesrpg/saesrpg | https://api.github.com/repos/saesrpg/saesrpg | closed | /sell bug after getting arrested while having respawning cooldown | Priority: Medium Status: Done Type: Bug | as the title say i pressed start new life and got arrested when i tried to sell after in the jb i got this

*scorpyo said there is another bug related to respawning he may tell after in a comment* | 1.0 | /sell bug after getting arrested while having respawning cooldown - as the title say i pressed start new life and got arrested when i tried to sell after in the jb i got this

*scorpyo said there is another bug related to respawning he may tell after in a comment* | non_defect | sell bug after getting arrested while having respawning cooldown as the title say i pressed start new life and got arrested when i tried to sell after in the jb i got this scorpyo said there is another bug related to respawning he may tell after in a comment | 0 |
827,758 | 31,794,026,256 | IssuesEvent | 2023-09-13 06:41:41 | S-ENDA/DMH | https://api.github.com/repos/S-ENDA/DMH | closed | Data Structure and documenting - Vocabulary | Priority | Include information about vocabulary for structuring and documenting data | 1.0 | Data Structure and documenting - Vocabulary - Include information about vocabulary for structuring and documenting data | non_defect | data structure and documenting vocabulary include information about vocabulary for structuring and documenting data | 0 |
97,216 | 8,651,576,785 | IssuesEvent | 2018-11-27 03:51:39 | humera987/FXLabs-Test-Automation | https://api.github.com/repos/humera987/FXLabs-Test-Automation | closed | projecttest16 : ApiV1ProjectsIdSearchAutoSuggestionsSearchStatusGetQueryParamPageInvalidDatatype | projecttest16 | Project : projecttest16
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ZmMxNDhkMzAtMjU2Mi00ZmM3LWE3N2MtODI5YTEyZjYxODZj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 27 Nov 2018 03:50:21 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/projects/HWBmVnnx/search-auto-suggestions/search/HWBmVnnx?page=oHfAPx
Request :
Response :
{
"timestamp" : "2018-11-27T03:50:22.255+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/HWBmVnnx/search-auto-suggestions/search/HWBmVnnx"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | 1.0 | projecttest16 : ApiV1ProjectsIdSearchAutoSuggestionsSearchStatusGetQueryParamPageInvalidDatatype - Project : projecttest16
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ZmMxNDhkMzAtMjU2Mi00ZmM3LWE3N2MtODI5YTEyZjYxODZj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 27 Nov 2018 03:50:21 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/projects/HWBmVnnx/search-auto-suggestions/search/HWBmVnnx?page=oHfAPx
Request :
Response :
{
"timestamp" : "2018-11-27T03:50:22.255+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/HWBmVnnx/search-auto-suggestions/search/HWBmVnnx"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | non_defect | project job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api projects hwbmvnnx search auto suggestions search hwbmvnnx logs assertion resolved to result assertion resolved to result fx bot | 0 |
281,108 | 21,315,374,583 | IssuesEvent | 2022-04-16 07:13:18 | tonishka/pe | https://api.github.com/repos/tonishka/pe | opened | Manual testing in DG was not updated | severity.VeryLow type.DocumentationBug | I believe for the PE both UG and DG (Instructions for Manual Testing) were to be used. However I found that the Manual testing section in DG had not been updated to match the current implementation so I tested wholly based on the UG.
<!--session: 1650089447412-597c822a-d7d2-4220-afc9-cd48c37af25f-->
<!--Version: Web v3.4.2--> | 1.0 | Manual testing in DG was not updated - I believe for the PE both UG and DG (Instructions for Manual Testing) were to be used. However I found that the Manual testing section in DG had not been updated to match the current implementation so I tested wholly based on the UG.
<!--session: 1650089447412-597c822a-d7d2-4220-afc9-cd48c37af25f-->
<!--Version: Web v3.4.2--> | non_defect | manual testing in dg was not updated i believe for the pe both ug and dg instructions for manual testing were to be used however i found that the manual testing section in dg had not been updated to match the current implementation so i tested wholly based on the ug | 0 |
18,045 | 6,542,585,446 | IssuesEvent | 2017-09-02 09:20:11 | moby/moby | https://api.github.com/repos/moby/moby | closed | COPY directory in multi-stage build corrupts binary files | area/builder version/17.06 | <!--
If you are reporting a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
If you suspect your issue is a bug, please edit your issue description to
include the BUG REPORT INFORMATION shown below. If you fail to provide this
information within 7 days, we cannot debug your issue and will close it. We
will, however, reopen it if you later provide the information.
For more information about reporting issues, see
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#reporting-other-issues
---------------------------------------------------
GENERAL SUPPORT INFORMATION
---------------------------------------------------
The GitHub issue tracker is for bug reports and feature requests.
General support can be found at the following locations:
- Docker Support Forums - https://forums.docker.com
- IRC - irc.freenode.net #docker channel
- Post a question on StackOverflow, using the Docker tag
---------------------------------------------------
BUG REPORT INFORMATION
---------------------------------------------------
Use the commands below to provide key information from your environment:
You do NOT have to include this information if this is a FEATURE REQUEST
-->
**Description**
<!--
Docker messes up binary files when when copying them from another build stage as part of a directory.
-->
**Steps to reproduce the issue:**
1. Create a multi-stage Dockerfile that builds a binary and copies it over to the next stage as part of a directory.
```
FROM haskell:8.0.2 AS builder
COPY migrations /var/migrations
WORKDIR /var/migrations
RUN ghc -O2 -threaded ./migrate.hs
FROM haskell:8.0.2
# Copying a single file works
COPY --from=builder /var/migrations/migrate /migrate
# The `migrate` file inside /var/migrations is messed up:
# bash: ./migrate: cannot execute binary file: Exec format error
COPY --from=builder /var/migrations /var/migrations
```
2. Run the binary file (it is corrupted).
**Describe the results you received:**
The `/var/migrations/migrate` executable is corrupted in the copied directory.
**Describe the results you expected:**
The executable in `/var/migrations/migrate` should be the same as the intact one (`/migrate`)
**Output of `docker version`:**
```
Client:
Version: 17.06.1-ce
API version: 1.30
Go version: go1.8.3
Git commit: 874a737
Built: Thu Aug 17 22:53:38 2017
OS/Arch: darwin/amd64
Server:
Version: 17.06.1-ce
API version: 1.30 (minimum version 1.12)
Go version: go1.8.3
Git commit: 874a737
Built: Thu Aug 17 22:54:55 2017
OS/Arch: linux/amd64
Experimental: true
```
**Output of `docker info`:**
```
Containers: 68
Running: 0
Paused: 0
Stopped: 68
Images: 217
Server Version: 17.06.1-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 6e23458c129b551d5c9871e5174f6b1b7f6d1170
runc version: 810190ceaa507aa2727d7ae6f4790c76ec150bd2
init version: 949e6fa
Security Options:
seccomp
Profile: default
Kernel Version: 4.9.41-moby
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 1.952GiB
Name: moby
ID: JEBM:ZKRD:5PUF:4GZ7:URXV:UYEI:SJ7Z:BAD5:CYZR:VES4:FNYI:SO6H
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): true
File Descriptors: 18
Goroutines: 31
System Time: 2017-09-02T09:04:56.9943268Z
EventsListeners: 1
Registry: https://index.docker.io/v1/
Experimental: true
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
Also fails on AWS (docker version and info output is from my local docker installation).
| 1.0 | COPY directory in multi-stage build corrupts binary files - <!--
If you are reporting a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
If you suspect your issue is a bug, please edit your issue description to
include the BUG REPORT INFORMATION shown below. If you fail to provide this
information within 7 days, we cannot debug your issue and will close it. We
will, however, reopen it if you later provide the information.
For more information about reporting issues, see
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#reporting-other-issues
---------------------------------------------------
GENERAL SUPPORT INFORMATION
---------------------------------------------------
The GitHub issue tracker is for bug reports and feature requests.
General support can be found at the following locations:
- Docker Support Forums - https://forums.docker.com
- IRC - irc.freenode.net #docker channel
- Post a question on StackOverflow, using the Docker tag
---------------------------------------------------
BUG REPORT INFORMATION
---------------------------------------------------
Use the commands below to provide key information from your environment:
You do NOT have to include this information if this is a FEATURE REQUEST
-->
**Description**
<!--
Docker messes up binary files when when copying them from another build stage as part of a directory.
-->
**Steps to reproduce the issue:**
1. Create a multi-stage Dockerfile that builds a binary and copies it over to the next stage as part of a directory.
```
FROM haskell:8.0.2 AS builder
COPY migrations /var/migrations
WORKDIR /var/migrations
RUN ghc -O2 -threaded ./migrate.hs
FROM haskell:8.0.2
# Copying a single file works
COPY --from=builder /var/migrations/migrate /migrate
# The `migrate` file inside /var/migrations is messed up:
# bash: ./migrate: cannot execute binary file: Exec format error
COPY --from=builder /var/migrations /var/migrations
```
2. Run the binary file (it is corrupted).
**Describe the results you received:**
The `/var/migrations/migrate` executable is corrupted in the copied directory.
**Describe the results you expected:**
The executable in `/var/migrations/migrate` should be the same as the intact one (`/migrate`)
**Output of `docker version`:**
```
Client:
Version: 17.06.1-ce
API version: 1.30
Go version: go1.8.3
Git commit: 874a737
Built: Thu Aug 17 22:53:38 2017
OS/Arch: darwin/amd64
Server:
Version: 17.06.1-ce
API version: 1.30 (minimum version 1.12)
Go version: go1.8.3
Git commit: 874a737
Built: Thu Aug 17 22:54:55 2017
OS/Arch: linux/amd64
Experimental: true
```
**Output of `docker info`:**
```
Containers: 68
Running: 0
Paused: 0
Stopped: 68
Images: 217
Server Version: 17.06.1-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 6e23458c129b551d5c9871e5174f6b1b7f6d1170
runc version: 810190ceaa507aa2727d7ae6f4790c76ec150bd2
init version: 949e6fa
Security Options:
seccomp
Profile: default
Kernel Version: 4.9.41-moby
Operating System: Alpine Linux v3.5
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 1.952GiB
Name: moby
ID: JEBM:ZKRD:5PUF:4GZ7:URXV:UYEI:SJ7Z:BAD5:CYZR:VES4:FNYI:SO6H
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): true
File Descriptors: 18
Goroutines: 31
System Time: 2017-09-02T09:04:56.9943268Z
EventsListeners: 1
Registry: https://index.docker.io/v1/
Experimental: true
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
Also fails on AWS (docker version and info output is from my local docker installation).
| non_defect | copy directory in multi stage build corrupts binary files if you are reporting a new issue make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead if you suspect your issue is a bug please edit your issue description to include the bug report information shown below if you fail to provide this information within days we cannot debug your issue and will close it we will however reopen it if you later provide the information for more information about reporting issues see general support information the github issue tracker is for bug reports and feature requests general support can be found at the following locations docker support forums irc irc freenode net docker channel post a question on stackoverflow using the docker tag bug report information use the commands below to provide key information from your environment you do not have to include this information if this is a feature request description docker messes up binary files when when copying them from another build stage as part of a directory steps to reproduce the issue create a multi stage dockerfile that builds a binary and copies it over to the next stage as part of a directory from haskell as builder copy migrations var migrations workdir var migrations run ghc threaded migrate hs from haskell copying a single file works copy from builder var migrations migrate migrate the migrate file inside var migrations is messed up bash migrate cannot execute binary file exec format error copy from builder var migrations var migrations run the binary file it is corrupted describe the results you received the var migrations migrate executable is corrupted in the copied directory describe the results you expected the executable in var migrations migrate should be the same as the intact one migrate output of docker version client version ce api version go version git commit built thu aug os arch darwin server version ce api version minimum version go version git commit built thu aug os arch linux experimental true output of docker info containers running paused stopped images server version ce storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host ipvlan macvlan null overlay log awslogs fluentd gcplogs gelf journald json file logentries splunk syslog swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options seccomp profile default kernel version moby operating system alpine linux ostype linux architecture cpus total memory name moby id jebm zkrd urxv uyei cyzr fnyi docker root dir var lib docker debug mode client false debug mode server true file descriptors goroutines system time eventslisteners registry experimental true insecure registries live restore enabled false additional environment details aws virtualbox physical etc also fails on aws docker version and info output is from my local docker installation | 0 |
134,351 | 30,004,553,910 | IssuesEvent | 2023-06-26 11:34:41 | cosmos/ibc-go | https://api.github.com/repos/cosmos/ibc-go | opened | Make unused block hashes in testing header creation explicit | code hygiene | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
A mocked tendermint header creation is a very important part of the testing package. It should be very readable and easy to understand. Create a `UnusedHash` value to indicate which fields in the header are explicitly unused by the testing pacakge.
## Problem Definition
static strings are set in the testing header creation [here](https://github.com/cosmos/ibc-go/blob/main/testing/chain.go#L478), these decrease readability as you might assume they do something
## Proposal
Replace static strings with a `UnusedHash` variable:
```go
tmHeader := tmtypes.Header{
Version: tmprotoversion.Consensus{Block: tmversion.BlockProtocol, App: 2},
ChainID: chainID,
Height: blockHeight,
Time: timestamp,
LastBlockID: MakeBlockID(make([]byte, tmhash.Size), 10_000, make([]byte, tmhash.Size)),
LastCommitHash: chain.App.LastCommitID().Hash,
DataHash: UnusedHash,
ValidatorsHash: vsetHash,
NextValidatorsHash: nextValHash,
ConsensusHash: UnusedHash,
AppHash: chain.CurrentHeader.AppHash,
LastResultsHash: UnusedHash,
EvidenceHash: UnusedHash,
ProposerAddress: tmValSet.Proposer.Address, //nolint:staticcheck
}
```
Open to other names for `UnusedHash`, maybe `EmptyHash`? The `UnusedHash` can be added to `values.go`
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| 1.0 | Make unused block hashes in testing header creation explicit - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
A mocked tendermint header creation is a very important part of the testing package. It should be very readable and easy to understand. Create a `UnusedHash` value to indicate which fields in the header are explicitly unused by the testing pacakge.
## Problem Definition
static strings are set in the testing header creation [here](https://github.com/cosmos/ibc-go/blob/main/testing/chain.go#L478), these decrease readability as you might assume they do something
## Proposal
Replace static strings with a `UnusedHash` variable:
```go
tmHeader := tmtypes.Header{
Version: tmprotoversion.Consensus{Block: tmversion.BlockProtocol, App: 2},
ChainID: chainID,
Height: blockHeight,
Time: timestamp,
LastBlockID: MakeBlockID(make([]byte, tmhash.Size), 10_000, make([]byte, tmhash.Size)),
LastCommitHash: chain.App.LastCommitID().Hash,
DataHash: UnusedHash,
ValidatorsHash: vsetHash,
NextValidatorsHash: nextValHash,
ConsensusHash: UnusedHash,
AppHash: chain.CurrentHeader.AppHash,
LastResultsHash: UnusedHash,
EvidenceHash: UnusedHash,
ProposerAddress: tmValSet.Proposer.Address, //nolint:staticcheck
}
```
Open to other names for `UnusedHash`, maybe `EmptyHash`? The `UnusedHash` can be added to `values.go`
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| non_defect | make unused block hashes in testing header creation explicit ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v word of caution poorly thought out proposals may be rejected v without deliberation ☺ summary a mocked tendermint header creation is a very important part of the testing package it should be very readable and easy to understand create a unusedhash value to indicate which fields in the header are explicitly unused by the testing pacakge problem definition static strings are set in the testing header creation these decrease readability as you might assume they do something proposal replace static strings with a unusedhash variable go tmheader tmtypes header version tmprotoversion consensus block tmversion blockprotocol app chainid chainid height blockheight time timestamp lastblockid makeblockid make byte tmhash size make byte tmhash size lastcommithash chain app lastcommitid hash datahash unusedhash validatorshash vsethash nextvalidatorshash nextvalhash consensushash unusedhash apphash chain currentheader apphash lastresultshash unusedhash evidencehash unusedhash proposeraddress tmvalset proposer address nolint staticcheck open to other names for unusedhash maybe emptyhash the unusedhash can be added to values go for admin use not duplicate issue appropriate labels applied appropriate contributors tagged assigned | 0 |
539,599 | 15,791,575,248 | IssuesEvent | 2021-04-02 04:57:23 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Introduce a flag to disable Script Based Adaptive Authentication in carbon console | Priority/Normal Type/Improvement WUM | ### Describe your problem(s)
This is to disable Script Based Adaptive Authentication in the carbon console as it is recommended to use IS on that.
<img width="1499" alt="Screen Shot 2021-04-02 at 10 20 42 AM" src="https://user-images.githubusercontent.com/6702569/113381657-267ff280-939d-11eb-8266-e1ddcf0670c2.png">
### Describe your solution
<!-- Describe the feature/improvement -->
### How will you implement it
<!-- If you like to suggest an approach or a design -->
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members--> | 1.0 | Introduce a flag to disable Script Based Adaptive Authentication in carbon console - ### Describe your problem(s)
This is to disable Script Based Adaptive Authentication in the carbon console as it is recommended to use IS on that.
<img width="1499" alt="Screen Shot 2021-04-02 at 10 20 42 AM" src="https://user-images.githubusercontent.com/6702569/113381657-267ff280-939d-11eb-8266-e1ddcf0670c2.png">
### Describe your solution
<!-- Describe the feature/improvement -->
### How will you implement it
<!-- If you like to suggest an approach or a design -->
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members--> | non_defect | introduce a flag to disable script based adaptive authentication in carbon console describe your problem s this is to disable script based adaptive authentication in the carbon console as it is recommended to use is on that img width alt screen shot at am src describe your solution how will you implement it optional fields related issues suggested labels suggested assignees | 0 |
490,793 | 14,140,232,216 | IssuesEvent | 2020-11-10 10:54:40 | gardener/docforge | https://api.github.com/repos/gardener/docforge | closed | Silence forge command usage on error except wrong flags/cmd/args | kind/enhancement priority/normal | **What would you like to be added**:
Silence forge command usage on error except wrong flags/cmd/args.
**Why is this needed / describe a real-world scenario**:
Displaying the usage on every error is useless and clutters the screen pushing up the stack the errors that a user is actually interested in.
**How to categorize this?**:
/kind enhancement
/priority normal | 1.0 | Silence forge command usage on error except wrong flags/cmd/args - **What would you like to be added**:
Silence forge command usage on error except wrong flags/cmd/args.
**Why is this needed / describe a real-world scenario**:
Displaying the usage on every error is useless and clutters the screen pushing up the stack the errors that a user is actually interested in.
**How to categorize this?**:
/kind enhancement
/priority normal | non_defect | silence forge command usage on error except wrong flags cmd args what would you like to be added silence forge command usage on error except wrong flags cmd args why is this needed describe a real world scenario displaying the usage on every error is useless and clutters the screen pushing up the stack the errors that a user is actually interested in how to categorize this kind enhancement priority normal | 0 |
1,082 | 2,594,525,926 | IssuesEvent | 2015-02-20 04:48:53 | paulirish/mutation-summary | https://api.github.com/repos/paulirish/mutation-summary | opened | Missing object presence check | auto-migrated Priority-Medium Type-Defect | ```
On line 490 of src/mutation-summary.js you have `if (!change.attributes)` where
you should have `if (!change || !change.attributes)`
```
-----
Original issue reported on code.google.com by `qstr...@gmail.com` on 17 Dec 2013 at 9:43
* Merged into: #25 | 1.0 | Missing object presence check - ```
On line 490 of src/mutation-summary.js you have `if (!change.attributes)` where
you should have `if (!change || !change.attributes)`
```
-----
Original issue reported on code.google.com by `qstr...@gmail.com` on 17 Dec 2013 at 9:43
* Merged into: #25 | defect | missing object presence check on line of src mutation summary js you have if change attributes where you should have if change change attributes original issue reported on code google com by qstr gmail com on dec at merged into | 1 |
311,067 | 9,527,840,272 | IssuesEvent | 2019-04-29 06:21:39 | orcidee/rpgconmanager | https://api.github.com/repos/orcidee/rpgconmanager | opened | Désinscription des joueurs après la fermeture du module | Priority : 1 | Pour le moment, il est possible de se désinscrire d'une partie après la fermeture du module pour le joueur. C'est un soucis, parce que je ne regarde pas mes mails pendant la conv'... Il faudrait donc s'arranger pour que ce ne soit plus possible. Et idéalement, rajouter dans le mail qu'à partir de 10h le vendredi avant la convention, il faut se désinscrire directement à l'accueil dès son arrivée, ou, si la personne ne peut pas venir, envoyer un mail directement au MJ ou à l'adresse webmaster@orcidee.ch. | 1.0 | Désinscription des joueurs après la fermeture du module - Pour le moment, il est possible de se désinscrire d'une partie après la fermeture du module pour le joueur. C'est un soucis, parce que je ne regarde pas mes mails pendant la conv'... Il faudrait donc s'arranger pour que ce ne soit plus possible. Et idéalement, rajouter dans le mail qu'à partir de 10h le vendredi avant la convention, il faut se désinscrire directement à l'accueil dès son arrivée, ou, si la personne ne peut pas venir, envoyer un mail directement au MJ ou à l'adresse webmaster@orcidee.ch. | non_defect | désinscription des joueurs après la fermeture du module pour le moment il est possible de se désinscrire d une partie après la fermeture du module pour le joueur c est un soucis parce que je ne regarde pas mes mails pendant la conv il faudrait donc s arranger pour que ce ne soit plus possible et idéalement rajouter dans le mail qu à partir de le vendredi avant la convention il faut se désinscrire directement à l accueil dès son arrivée ou si la personne ne peut pas venir envoyer un mail directement au mj ou à l adresse webmaster orcidee ch | 0 |
48,144 | 13,067,476,593 | IssuesEvent | 2020-07-31 00:34:51 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [fill-ratio] Test failure (Trac #1855) | Migrated from Trac combo reconstruction defect |
```text
Start 189: fill-ratio::FillRatioTest.py
189/424 Test #189: fill-ratio::FillRatioTest.py ...................................***Failed 5.17 sec
INFO (I3Module): Opened file /build/ports/test-data/event-viewer/Level3aGCD_IC79_EEData_Run00115990.i3 (I3Reader.cxx:180 in void I3Reader::OpenNextFile())
INFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): ResultName : FillRatioTest (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusRMS : 3.500000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): ResultName : FillRatioInfo (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusRMS : 3.000000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())
FATAL (I3FillRatio2StringDoubleMap): FillRatioInfo DOES NOT EXIST (I3FillRatio2StringDoubleMap.cxx:93 in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))
ERROR (I3Module): FR_to_Map_segment: Exception thrown (I3Module.cxx:118 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/source/fill-ratio/resources/tests/FillRatioTest.py", line 48, in <module>
tray.Execute()
File "/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/build/lib/I3Tray.py", line 240, in Execute
super(I3Tray, self).Execute()
RuntimeError: FillRatioInfo DOES NOT EXIST (in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))
```
Migrated from https://code.icecube.wisc.edu/ticket/1855
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:18",
"description": "{{{\n Start 189: fill-ratio::FillRatioTest.py\n189/424 Test #189: fill-ratio::FillRatioTest.py ...................................***Failed 5.17 sec\nINFO (I3Module): Opened file /build/ports/test-data/event-viewer/Level3aGCD_IC79_EEData_Run00115990.i3 (I3Reader.cxx:180 in void I3Reader::OpenNextFile())\nINFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): ResultName : FillRatioTest (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusRMS : 3.500000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): ResultName : FillRatioInfo (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusRMS : 3.000000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())\nFATAL (I3FillRatio2StringDoubleMap): FillRatioInfo DOES NOT EXIST (I3FillRatio2StringDoubleMap.cxx:93 in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))\nERROR (I3Module): FR_to_Map_segment: Exception thrown (I3Module.cxx:118 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/source/fill-ratio/resources/tests/FillRatioTest.py\", line 48, in <module>\n tray.Execute()\n File \"/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/build/lib/I3Tray.py\", line 240, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: FillRatioInfo DOES NOT EXIST (in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))\n}}}",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1550067198646291",
"component": "combo reconstruction",
"summary": "[fill-ratio] Test failure",
"priority": "critical",
"keywords": "",
"time": "2016-09-09T18:29:13",
"milestone": "",
"owner": "mjl5147",
"type": "defect"
}
```
| 1.0 | [fill-ratio] Test failure (Trac #1855) -
```text
Start 189: fill-ratio::FillRatioTest.py
189/424 Test #189: fill-ratio::FillRatioTest.py ...................................***Failed 5.17 sec
INFO (I3Module): Opened file /build/ports/test-data/event-viewer/Level3aGCD_IC79_EEData_Run00115990.i3 (I3Reader.cxx:180 in void I3Reader::OpenNextFile())
INFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): ResultName : FillRatioTest (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusRMS : 3.500000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): ResultName : FillRatioInfo (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusRMS : 3.000000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())
INFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())
FATAL (I3FillRatio2StringDoubleMap): FillRatioInfo DOES NOT EXIST (I3FillRatio2StringDoubleMap.cxx:93 in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))
ERROR (I3Module): FR_to_Map_segment: Exception thrown (I3Module.cxx:118 in void I3Module::Do(void (I3Module::*)()))
Traceback (most recent call last):
File "/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/source/fill-ratio/resources/tests/FillRatioTest.py", line 48, in <module>
tray.Execute()
File "/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/build/lib/I3Tray.py", line 240, in Execute
super(I3Tray, self).Execute()
RuntimeError: FillRatioInfo DOES NOT EXIST (in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))
```
Migrated from https://code.icecube.wisc.edu/ticket/1855
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:18",
"description": "{{{\n Start 189: fill-ratio::FillRatioTest.py\n189/424 Test #189: fill-ratio::FillRatioTest.py ...................................***Failed 5.17 sec\nINFO (I3Module): Opened file /build/ports/test-data/event-viewer/Level3aGCD_IC79_EEData_Run00115990.i3 (I3Reader.cxx:180 in void I3Reader::OpenNextFile())\nINFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): ResultName : FillRatioTest (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusRMS : 3.500000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): Configuring the I3FillRatio (I3FillRatioModule.cxx:97 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): VertexName : clast (I3FillRatioModule.cxx:112 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): ResultName : FillRatioInfo (I3FillRatioModule.cxx:113 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): RecoPulseName : MaskedOfflinePulses (I3FillRatioModule.cxx:114 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusRMS : 3.000000 (I3FillRatioModule.cxx:115 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): SphericalRadiusMean : 2.000000 (I3FillRatioModule.cxx:116 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): AmplitudeWeightingPower : 0.000000 => no weighting (I3FillRatioModule.cxx:120 in virtual void I3FillRatioModule::Configure())\nINFO (I3FillRatioModule): BadOms [size of list] : 0 (I3FillRatioModule.cxx:132 in virtual void I3FillRatioModule::Configure())\nFATAL (I3FillRatio2StringDoubleMap): FillRatioInfo DOES NOT EXIST (I3FillRatio2StringDoubleMap.cxx:93 in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))\nERROR (I3Module): FR_to_Map_segment: Exception thrown (I3Module.cxx:118 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/source/fill-ratio/resources/tests/FillRatioTest.py\", line 48, in <module>\n tray.Execute()\n File \"/build/buildslave/kipple_cvmfs/Scientific_Linux_6__new___cvmfs_/build/lib/I3Tray.py\", line 240, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: FillRatioInfo DOES NOT EXIST (in virtual void I3FillRatio2StringDoubleMap::Physics(I3FramePtr))\n}}}",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1550067198646291",
"component": "combo reconstruction",
"summary": "[fill-ratio] Test failure",
"priority": "critical",
"keywords": "",
"time": "2016-09-09T18:29:13",
"milestone": "",
"owner": "mjl5147",
"type": "defect"
}
```
| defect | test failure trac text start fill ratio fillratiotest py test fill ratio fillratiotest py failed sec info opened file build ports test data event viewer eedata cxx in void opennextfile info configuring the cxx in virtual void configure info vertexname clast cxx in virtual void configure info resultname fillratiotest cxx in virtual void configure info recopulsename maskedofflinepulses cxx in virtual void configure info sphericalradiusrms cxx in virtual void configure info sphericalradiusmean cxx in virtual void configure info amplitudeweightingpower no weighting cxx in virtual void configure info badoms cxx in virtual void configure info configuring the cxx in virtual void configure info vertexname clast cxx in virtual void configure info resultname fillratioinfo cxx in virtual void configure info recopulsename maskedofflinepulses cxx in virtual void configure info sphericalradiusrms cxx in virtual void configure info sphericalradiusmean cxx in virtual void configure info amplitudeweightingpower no weighting cxx in virtual void configure info badoms cxx in virtual void configure fatal fillratioinfo does not exist cxx in virtual void physics error fr to map segment exception thrown cxx in void do void traceback most recent call last file build buildslave kipple cvmfs scientific linux new cvmfs source fill ratio resources tests fillratiotest py line in tray execute file build buildslave kipple cvmfs scientific linux new cvmfs build lib py line in execute super self execute runtimeerror fillratioinfo does not exist in virtual void physics migrated from json status closed changetime description n start fill ratio fillratiotest py test fill ratio fillratiotest py failed sec ninfo opened file build ports test data event viewer eedata cxx in void opennextfile ninfo configuring the cxx in virtual void configure ninfo vertexname clast cxx in virtual void configure ninfo resultname fillratiotest cxx in virtual void configure ninfo recopulsename maskedofflinepulses cxx in virtual void configure ninfo sphericalradiusrms cxx in virtual void configure ninfo sphericalradiusmean cxx in virtual void configure ninfo amplitudeweightingpower no weighting cxx in virtual void configure ninfo badoms cxx in virtual void configure ninfo configuring the cxx in virtual void configure ninfo vertexname clast cxx in virtual void configure ninfo resultname fillratioinfo cxx in virtual void configure ninfo recopulsename maskedofflinepulses cxx in virtual void configure ninfo sphericalradiusrms cxx in virtual void configure ninfo sphericalradiusmean cxx in virtual void configure ninfo amplitudeweightingpower no weighting cxx in virtual void configure ninfo badoms cxx in virtual void configure nfatal fillratioinfo does not exist cxx in virtual void physics nerror fr to map segment exception thrown cxx in void do void ntraceback most recent call last n file build buildslave kipple cvmfs scientific linux new cvmfs source fill ratio resources tests fillratiotest py line in n tray execute n file build buildslave kipple cvmfs scientific linux new cvmfs build lib py line in execute n super self execute nruntimeerror fillratioinfo does not exist in virtual void physics n reporter olivas cc resolution fixed ts component combo reconstruction summary test failure priority critical keywords time milestone owner type defect | 1 |
77,645 | 15,569,821,971 | IssuesEvent | 2021-03-17 01:04:17 | tlkh/serverless-transformers | https://api.github.com/repos/tlkh/serverless-transformers | opened | CVE-2020-35653 (High) detected in Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2020-35653 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: serverless-transformers/requirements.txt</p>
<p>Path to vulnerable library: serverless-transformers/requirements.txt,serverless-transformers/backend/requirements.txt</p>
<p>
Dependency Hierarchy:
- streamlit-0.55.2-py2.py3-none-any.whl (Root Library)
- :x: **Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Pillow before 8.1.0, PcxDecode has a buffer over-read when decoding a crafted PCX file because the user-supplied stride value is trusted for buffer calculations.
<p>Publish Date: 2021-01-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35653>CVE-2020-35653</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35653">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35653</a></p>
<p>Release Date: 2021-01-12</p>
<p>Fix Resolution: 8.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-35653 (High) detected in Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2020-35653 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/12/ad/61f8dfba88c4e56196bf6d056cdbba64dc9c5dfdfbc97d02e6472feed913/Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: serverless-transformers/requirements.txt</p>
<p>Path to vulnerable library: serverless-transformers/requirements.txt,serverless-transformers/backend/requirements.txt</p>
<p>
Dependency Hierarchy:
- streamlit-0.55.2-py2.py3-none-any.whl (Root Library)
- :x: **Pillow-6.2.2-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Pillow before 8.1.0, PcxDecode has a buffer over-read when decoding a crafted PCX file because the user-supplied stride value is trusted for buffer calculations.
<p>Publish Date: 2021-01-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35653>CVE-2020-35653</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35653">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35653</a></p>
<p>Release Date: 2021-01-12</p>
<p>Fix Resolution: 8.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file serverless transformers requirements txt path to vulnerable library serverless transformers requirements txt serverless transformers backend requirements txt dependency hierarchy streamlit none any whl root library x pillow whl vulnerable library found in base branch main vulnerability details in pillow before pcxdecode has a buffer over read when decoding a crafted pcx file because the user supplied stride value is trusted for buffer calculations publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
7,267 | 2,610,361,039 | IssuesEvent | 2015-02-26 19:56:46 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | Invalid grant error, publishing post is impossible | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
Each time I try to publish a post, I get this error:
{ "error" : "invalid_grant" }
What browser are you using?
Firefox 11.0
What version of ScribeFire are you running?
Next 4.0
```
-----
Original issue reported on code.google.com by `g.cevol...@gmail.com` on 23 Mar 2012 at 9:52 | 1.0 | Invalid grant error, publishing post is impossible - ```
What's the problem?
Each time I try to publish a post, I get this error:
{ "error" : "invalid_grant" }
What browser are you using?
Firefox 11.0
What version of ScribeFire are you running?
Next 4.0
```
-----
Original issue reported on code.google.com by `g.cevol...@gmail.com` on 23 Mar 2012 at 9:52 | defect | invalid grant error publishing post is impossible what s the problem each time i try to publish a post i get this error error invalid grant what browser are you using firefox what version of scribefire are you running next original issue reported on code google com by g cevol gmail com on mar at | 1 |
121,797 | 16,024,269,695 | IssuesEvent | 2021-04-21 06:59:57 | cocoa-mhlw/cocoa | https://api.github.com/repos/cocoa-mhlw/cocoa | reopened | ヘッダロゴのスペルミス (Welfar) | bug design ready-for-release | **Describe the bug**
ヘッダロゴ中、厚生労働省さんの英語名表記にスペルミス (Welfar) があります。
**To Reproduce**
1. COCOA を起動する
2. メインページ上部のヘッダロゴを見る
**Expected behavior**
正しいスペル (Welfare) でロゴが表示される。
**Smartphone (please complete the following information):**
- Device: Essential Phone PH-1
- OS: Android 10
- Version: 1.2.2
| 1.0 | ヘッダロゴのスペルミス (Welfar) - **Describe the bug**
ヘッダロゴ中、厚生労働省さんの英語名表記にスペルミス (Welfar) があります。
**To Reproduce**
1. COCOA を起動する
2. メインページ上部のヘッダロゴを見る
**Expected behavior**
正しいスペル (Welfare) でロゴが表示される。
**Smartphone (please complete the following information):**
- Device: Essential Phone PH-1
- OS: Android 10
- Version: 1.2.2
| non_defect | ヘッダロゴのスペルミス welfar describe the bug ヘッダロゴ中、厚生労働省さんの英語名表記にスペルミス welfar があります。 to reproduce cocoa を起動する メインページ上部のヘッダロゴを見る expected behavior 正しいスペル welfare でロゴが表示される。 smartphone please complete the following information device essential phone ph os android version | 0 |
334,487 | 24,421,225,062 | IssuesEvent | 2022-10-05 20:30:21 | cloudflare/cloudflare-docs | https://api.github.com/repos/cloudflare/cloudflare-docs | closed | [workers] Include `FormData` example(s) | documentation product:workers | We don't currently have any in our [examples list](https://developers.cloudflare.com/workers/examples).
Potential Example: [Playground](https://cloudflareworkers.com/#91f10aa9e02b4e9a17134afe50b788be:https://tutorial.cloudflareworkers.com)
Should be added to the `/examples` list and maybe linked to from the [`Request`](https://developers.cloudflare.com/workers/runtime-apis/request) runtime API page?
---
- [x] Add example
- [ ] Link within Request API doc | 1.0 | [workers] Include `FormData` example(s) - We don't currently have any in our [examples list](https://developers.cloudflare.com/workers/examples).
Potential Example: [Playground](https://cloudflareworkers.com/#91f10aa9e02b4e9a17134afe50b788be:https://tutorial.cloudflareworkers.com)
Should be added to the `/examples` list and maybe linked to from the [`Request`](https://developers.cloudflare.com/workers/runtime-apis/request) runtime API page?
---
- [x] Add example
- [ ] Link within Request API doc | non_defect | include formdata example s we don t currently have any in our potential example should be added to the examples list and maybe linked to from the runtime api page add example link within request api doc | 0 |
201,202 | 7,025,501,861 | IssuesEvent | 2017-12-23 11:36:53 | HackerExperience/Helix | https://api.github.com/repos/HackerExperience/Helix | opened | Disconnect logged-in players when underlying server NIP changes | Kind:Enhancement Kind:Feature Priority:Intermediate | Implementation: broadcast DisconnectedEvent to all remotely connected players, and then stop the channel.
Related:
- Notify IP changed (HackedDatabase)
- ServerWebsocketChannelState may need to be updated (test it) | 1.0 | Disconnect logged-in players when underlying server NIP changes - Implementation: broadcast DisconnectedEvent to all remotely connected players, and then stop the channel.
Related:
- Notify IP changed (HackedDatabase)
- ServerWebsocketChannelState may need to be updated (test it) | non_defect | disconnect logged in players when underlying server nip changes implementation broadcast disconnectedevent to all remotely connected players and then stop the channel related notify ip changed hackeddatabase serverwebsocketchannelstate may need to be updated test it | 0 |
38,946 | 9,083,161,689 | IssuesEvent | 2019-02-17 18:17:02 | extnet/Ext.NET | https://api.github.com/repos/extnet/Ext.NET | closed | Ext.net.FieldNote.setNote() error if no initial note is set | 4.x defect | Found: 4.7.1
Ext.NET forum thread: [setNote JS function broken](https://forums.ext.net/showthread.php?62576)
The `Ext.net.FieldNote.setNote()` method references null/undefined noteEl.dom if called when a field is defined with no `Note="value"` (client-side `note: "value"`) at define time.
When the method is called, if there's no fieldNote element, it should be created the same way it is when defining the field with a note.
This is an Ext.NET-specific feature, so this issue should not be tagged after the _Sencha_ label. | 1.0 | Ext.net.FieldNote.setNote() error if no initial note is set - Found: 4.7.1
Ext.NET forum thread: [setNote JS function broken](https://forums.ext.net/showthread.php?62576)
The `Ext.net.FieldNote.setNote()` method references null/undefined noteEl.dom if called when a field is defined with no `Note="value"` (client-side `note: "value"`) at define time.
When the method is called, if there's no fieldNote element, it should be created the same way it is when defining the field with a note.
This is an Ext.NET-specific feature, so this issue should not be tagged after the _Sencha_ label. | defect | ext net fieldnote setnote error if no initial note is set found ext net forum thread the ext net fieldnote setnote method references null undefined noteel dom if called when a field is defined with no note value client side note value at define time when the method is called if there s no fieldnote element it should be created the same way it is when defining the field with a note this is an ext net specific feature so this issue should not be tagged after the sencha label | 1 |
267,067 | 8,378,857,206 | IssuesEvent | 2018-10-06 18:30:21 | civic-app/civic-app | https://api.github.com/repos/civic-app/civic-app | opened | Google account creation/sign-in not working | Priority: High Status: Available Type: Bug | Pressing "Continue with Google" on the sign in screen brings up Google authentication in webview, but pressing my account to continue just takes me back to the sign in screen.
Possibly related: I'm getting "The given sign-in provider is disabled for this Firebase project. Enable it in the Firebase console, under the sign-in method tab of the Auth section." as red error text underneath the email/password fields when I click Sign In or Register.

| 1.0 | Google account creation/sign-in not working - Pressing "Continue with Google" on the sign in screen brings up Google authentication in webview, but pressing my account to continue just takes me back to the sign in screen.
Possibly related: I'm getting "The given sign-in provider is disabled for this Firebase project. Enable it in the Firebase console, under the sign-in method tab of the Auth section." as red error text underneath the email/password fields when I click Sign In or Register.

| non_defect | google account creation sign in not working pressing continue with google on the sign in screen brings up google authentication in webview but pressing my account to continue just takes me back to the sign in screen possibly related i m getting the given sign in provider is disabled for this firebase project enable it in the firebase console under the sign in method tab of the auth section as red error text underneath the email password fields when i click sign in or register | 0 |
55,599 | 23,511,808,462 | IssuesEvent | 2022-08-18 17:18:02 | pingidentity/terraform-provider-pingone | https://api.github.com/repos/pingidentity/terraform-provider-pingone | opened | Support for FIDO policies | enhancement service mfa waiting on sdk | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please provide a helpful description of the feature request here. --->
Provide support for FIDO MFA policies
### New or Affected Resource(s)
<!--- Please provide a list of the new and/or affected resources/data sources, for example: -->
- pingone_mfa_fido_policy
<!--- Optionally include a brief description on the type of change required, but this isn't essential -->
### Potential Terraform Configuration
<!-- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code -->
```hcl
# Copy-paste your PingOne related Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
# Remember to replace any account/customer sensitive information in the configuration before submitting the issue
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://apidocs.pingidentity.com/pingone/platform/v1/api/#get-read-one-organization
- or -
* https://docs.pingidentity.com/bundle/pingone/page/cxs1575407884833.html
--->
* https://apidocs.pingidentity.com/pingone/platform/v1/api/#fido-policies
| 1.0 | Support for FIDO policies - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please provide a helpful description of the feature request here. --->
Provide support for FIDO MFA policies
### New or Affected Resource(s)
<!--- Please provide a list of the new and/or affected resources/data sources, for example: -->
- pingone_mfa_fido_policy
<!--- Optionally include a brief description on the type of change required, but this isn't essential -->
### Potential Terraform Configuration
<!-- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code -->
```hcl
# Copy-paste your PingOne related Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
# Remember to replace any account/customer sensitive information in the configuration before submitting the issue
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://apidocs.pingidentity.com/pingone/platform/v1/api/#get-read-one-organization
- or -
* https://docs.pingidentity.com/bundle/pingone/page/cxs1575407884833.html
--->
* https://apidocs.pingidentity.com/pingone/platform/v1/api/#fido-policies
| non_defect | support for fido policies community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description provide support for fido mfa policies new or affected resource s pingone mfa fido policy potential terraform configuration hcl copy paste your pingone related terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key remember to replace any account customer sensitive information in the configuration before submitting the issue references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example or | 0 |
6,775 | 6,599,996,875 | IssuesEvent | 2017-09-17 05:00:18 | Daniel-Mietchen/ideas | https://api.github.com/repos/Daniel-Mietchen/ideas | opened | Look into Humanitarian Data Exchange | emergency-response High-priority infrastructure | Came up multiple times at the HOT Summit (cf. https://github.com/Daniel-Mietchen/events/issues/71 ):
- https://data.humdata.org/search?fq=(((tags:cod%20OR%20title:HOTOSM)%20AND%20(groups:atg%20OR%20groups:aia%20OR%20groups:msr%20OR%20groups:kna%20OR%20groups:ant%20OR%20groups:maf%20OR%20groups:blm%20OR%20groups:glp%20OR%20groups:pri%20OR%20groups:vir%20OR%20groups:hti%20OR%20groups:aia%20OR%20groups:msr%20OR%20groups:glp))%20OR%20(tags:hurricaneirma%20OR%20tags:hurricane%20irma))&sort=metadata_modified+desc&ext_page_size=100
- https://data.humdata.org/dataset/damages-assessment-for-hurricane-irma
- https://data.humdata.org/organization/inform?sort=metadata_modified+desc
- ([from Twitter](https://twitter.com/lusdavo/status/908760372534022145)): https://data.humdata.org/organization/ocha-liberia
| 1.0 | Look into Humanitarian Data Exchange - Came up multiple times at the HOT Summit (cf. https://github.com/Daniel-Mietchen/events/issues/71 ):
- https://data.humdata.org/search?fq=(((tags:cod%20OR%20title:HOTOSM)%20AND%20(groups:atg%20OR%20groups:aia%20OR%20groups:msr%20OR%20groups:kna%20OR%20groups:ant%20OR%20groups:maf%20OR%20groups:blm%20OR%20groups:glp%20OR%20groups:pri%20OR%20groups:vir%20OR%20groups:hti%20OR%20groups:aia%20OR%20groups:msr%20OR%20groups:glp))%20OR%20(tags:hurricaneirma%20OR%20tags:hurricane%20irma))&sort=metadata_modified+desc&ext_page_size=100
- https://data.humdata.org/dataset/damages-assessment-for-hurricane-irma
- https://data.humdata.org/organization/inform?sort=metadata_modified+desc
- ([from Twitter](https://twitter.com/lusdavo/status/908760372534022145)): https://data.humdata.org/organization/ocha-liberia
| non_defect | look into humanitarian data exchange came up multiple times at the hot summit cf | 0 |
30,960 | 6,377,671,848 | IssuesEvent | 2017-08-02 10:37:51 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Type.GetGenericArguments() doesn't seem to work with v16-beta5 | defect | Unable to extract generic type arguments from a generic type with v16-beta5 while it worked correctly with v16-beta4. I am assuming reflection is enabled by default at deck.net since I am able to log the method name in the code sample below.
### Steps To Reproduce (v16-beta5)
[Deck Sample](https://deck.net/04a46660a63606202a908acc58146075)
```c#
public interface ISomeInterface
{
Task<string> GetString();
}
public class Program
{
public static void Main()
{
var type = typeof(ISomeInterface);
var methods = type.GetMethods();
var firstMethod = methods[0];
Console.WriteLine(firstMethod.Name);
var genericArgs = firstMethod.ReturnType.GetGenericArguments();
Console.WriteLine(genericArgs.Length);
}
}
```
### Expected Result
```js
GetString
1
```
### Actual Result
```js
GetString
0
```
| 1.0 | Type.GetGenericArguments() doesn't seem to work with v16-beta5 - Unable to extract generic type arguments from a generic type with v16-beta5 while it worked correctly with v16-beta4. I am assuming reflection is enabled by default at deck.net since I am able to log the method name in the code sample below.
### Steps To Reproduce (v16-beta5)
[Deck Sample](https://deck.net/04a46660a63606202a908acc58146075)
```c#
public interface ISomeInterface
{
Task<string> GetString();
}
public class Program
{
public static void Main()
{
var type = typeof(ISomeInterface);
var methods = type.GetMethods();
var firstMethod = methods[0];
Console.WriteLine(firstMethod.Name);
var genericArgs = firstMethod.ReturnType.GetGenericArguments();
Console.WriteLine(genericArgs.Length);
}
}
```
### Expected Result
```js
GetString
1
```
### Actual Result
```js
GetString
0
```
| defect | type getgenericarguments doesn t seem to work with unable to extract generic type arguments from a generic type with while it worked correctly with i am assuming reflection is enabled by default at deck net since i am able to log the method name in the code sample below steps to reproduce c public interface isomeinterface task getstring public class program public static void main var type typeof isomeinterface var methods type getmethods var firstmethod methods console writeline firstmethod name var genericargs firstmethod returntype getgenericarguments console writeline genericargs length expected result js getstring actual result js getstring | 1 |
69,495 | 22,390,686,258 | IssuesEvent | 2022-06-17 07:21:34 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | Inviting a user to a space they're already in shows "is already in the room" | T-Defect A-Invite S-Tolerable O-Uncommon Z-WTF Team: Delight | ### Steps to reproduce
Invite a user in a space they're already in
https://user-images.githubusercontent.com/20701752/174246092-b76b8d22-7780-4fa8-aa39-03d0c3d01f36.mp4
### Outcome
#### What did you expect?
Dialog says "[user] is already in the space"
#### What happened instead?
It says "[user] is already in the room"
### Your phone model
Pixel 6 Pro
### Operating system version
Android 12
### Application version and app store
1.4.24-dev
### Homeserver
matrix.org
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | 1.0 | Inviting a user to a space they're already in shows "is already in the room" - ### Steps to reproduce
Invite a user in a space they're already in
https://user-images.githubusercontent.com/20701752/174246092-b76b8d22-7780-4fa8-aa39-03d0c3d01f36.mp4
### Outcome
#### What did you expect?
Dialog says "[user] is already in the space"
#### What happened instead?
It says "[user] is already in the room"
### Your phone model
Pixel 6 Pro
### Operating system version
Android 12
### Application version and app store
1.4.24-dev
### Homeserver
matrix.org
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | defect | inviting a user to a space they re already in shows is already in the room steps to reproduce invite a user in a space they re already in outcome what did you expect dialog says is already in the space what happened instead it says is already in the room your phone model pixel pro operating system version android application version and app store dev homeserver matrix org will you send logs no are you willing to provide a pr yes | 1 |
235,010 | 7,733,846,743 | IssuesEvent | 2018-05-26 16:43:01 | Polymer/lit-html | https://api.github.com/repos/Polymer/lit-html | closed | Make the attribute/prop difference between lit-html and lit-extended clearer | Priority: Medium Status: Available Type: Maintenance | Docs say that lit-extended uses properties by default, but it doesn't point out that lit-html uses attributes, so maybe that could be made a bit more clear?
I at least had to look through the code to figure out why things wasn't working with lit-extended | 1.0 | Make the attribute/prop difference between lit-html and lit-extended clearer - Docs say that lit-extended uses properties by default, but it doesn't point out that lit-html uses attributes, so maybe that could be made a bit more clear?
I at least had to look through the code to figure out why things wasn't working with lit-extended | non_defect | make the attribute prop difference between lit html and lit extended clearer docs say that lit extended uses properties by default but it doesn t point out that lit html uses attributes so maybe that could be made a bit more clear i at least had to look through the code to figure out why things wasn t working with lit extended | 0 |
60,205 | 17,023,368,794 | IssuesEvent | 2021-07-03 01:39:52 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Merkaartor.pro has syntax error | Component: merkaartor Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.38pm, Thursday, 19th February 2009]**
There's a syntax error in the including of the translations in Merkaartor.pro where the \ to escape the newline are missing.
Either make these two backslashes manually or use the diff. | 1.0 | Merkaartor.pro has syntax error - **[Submitted to the original trac issue database at 9.38pm, Thursday, 19th February 2009]**
There's a syntax error in the including of the translations in Merkaartor.pro where the \ to escape the newline are missing.
Either make these two backslashes manually or use the diff. | defect | merkaartor pro has syntax error there s a syntax error in the including of the translations in merkaartor pro where the to escape the newline are missing either make these two backslashes manually or use the diff | 1 |
413,545 | 27,958,241,163 | IssuesEvent | 2023-03-24 13:55:55 | Heklaterriol/JEM-Project | https://api.github.com/repos/Heklaterriol/JEM-Project | opened | Add precise instructions for the users to update from JEM 2.3.5 to 4.0 | documentation update to Joomla 4 | Where?
- on the website
- in the package
- ... | 1.0 | Add precise instructions for the users to update from JEM 2.3.5 to 4.0 - Where?
- on the website
- in the package
- ... | non_defect | add precise instructions for the users to update from jem to where on the website in the package | 0 |
114,394 | 17,209,452,019 | IssuesEvent | 2021-07-19 00:11:24 | turkdevops/javascript-sdk | https://api.github.com/repos/turkdevops/javascript-sdk | opened | WS-2019-0331 (Medium) detected in handlebars-4.0.12.tgz | security vulnerability | ## WS-2019-0331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.12.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz</a></p>
<p>Path to dependency file: javascript-sdk/package.json</p>
<p>Path to vulnerable library: javascript-sdk/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jsdoc-to-markdown-4.0.1.tgz (Root Library)
- dmd-3.0.12.tgz
- :x: **handlebars-4.0.12.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability found in handlebars before 4.5.2. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: handlebars - 4.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0331 (Medium) detected in handlebars-4.0.12.tgz - ## WS-2019-0331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.12.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz</a></p>
<p>Path to dependency file: javascript-sdk/package.json</p>
<p>Path to vulnerable library: javascript-sdk/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jsdoc-to-markdown-4.0.1.tgz (Root Library)
- dmd-3.0.12.tgz
- :x: **handlebars-4.0.12.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability found in handlebars before 4.5.2. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.
<p>Publish Date: 2019-11-13
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: handlebars - 4.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file javascript sdk package json path to vulnerable library javascript sdk node modules handlebars package json dependency hierarchy jsdoc to markdown tgz root library dmd tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
46,115 | 13,055,854,201 | IssuesEvent | 2020-07-30 02:56:01 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | trunk of icerec vs. 2008 level2 data: data wins (Trac #601) | Incomplete Migration Migrated from Trac dataio defect | Migrated from https://code.icecube.wisc.edu/ticket/601
```json
{
"status": "closed",
"changetime": "2010-03-04T02:38:05",
"description": "jeisch@miskwa% dataio-shovel Level2_All_Run00111436_Part00000000.i3 \nterminate called after throwing an instance of 'boost::archive::archive_exception'\n what(): unsupported version\nAbort",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1267670285000000",
"component": "dataio",
"summary": "trunk of icerec vs. 2008 level2 data: data wins",
"priority": "normal",
"keywords": "",
"time": "2010-03-03T15:52:26",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | trunk of icerec vs. 2008 level2 data: data wins (Trac #601) - Migrated from https://code.icecube.wisc.edu/ticket/601
```json
{
"status": "closed",
"changetime": "2010-03-04T02:38:05",
"description": "jeisch@miskwa% dataio-shovel Level2_All_Run00111436_Part00000000.i3 \nterminate called after throwing an instance of 'boost::archive::archive_exception'\n what(): unsupported version\nAbort",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1267670285000000",
"component": "dataio",
"summary": "trunk of icerec vs. 2008 level2 data: data wins",
"priority": "normal",
"keywords": "",
"time": "2010-03-03T15:52:26",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| defect | trunk of icerec vs data data wins trac migrated from json status closed changetime description jeisch miskwa dataio shovel all nterminate called after throwing an instance of boost archive archive exception n what unsupported version nabort reporter troy cc resolution fixed ts component dataio summary trunk of icerec vs data data wins priority normal keywords time milestone owner troy type defect | 1 |
41,754 | 5,368,456,152 | IssuesEvent | 2017-02-22 08:51:01 | RestComm/Restcomm-Connect | https://api.github.com/repos/RestComm/Restcomm-Connect | opened | Provide onTimeout hook for RVD/ES | Visual App Designer | It makes sense to be able to handle external service timeouts in a customized manner. In UX terms, a new "OnTimeout" field with a module dropdown would work. In case the external service takes too long to respond the application flow would be diverted to the handlermodule. | 1.0 | Provide onTimeout hook for RVD/ES - It makes sense to be able to handle external service timeouts in a customized manner. In UX terms, a new "OnTimeout" field with a module dropdown would work. In case the external service takes too long to respond the application flow would be diverted to the handlermodule. | non_defect | provide ontimeout hook for rvd es it makes sense to be able to handle external service timeouts in a customized manner in ux terms a new ontimeout field with a module dropdown would work in case the external service takes too long to respond the application flow would be diverted to the handlermodule | 0 |
75,576 | 25,925,551,697 | IssuesEvent | 2022-12-16 03:54:51 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | new fs not mounted after "zfs clone" | Type: Defect Status: Stale Status: Triage Needed | ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Centos
Distribution Version | 8.3
Linux Kernel | 4.18.0-240.15.1.el8_3
Architecture | x86_64
ZFS Version | 2.0.3-1
SPL Version | 2.0.3-1
### Describe the problem you're observing
tank/myfs with >> files
OK zfs snapshot tank/myfs@current
NOK zfs clone tank/myfs@current tank/myfs_export // OK in a sense of no exception shown
OK zfs list // clone with mountpoint shown
NOK ls /tank/myfs_export or zfs mount -a ls /tank/myfs_export
0 files
### Describe how to reproduce the problem
zfs destroy tank/myfs_export
repeat
same result
### Include any warning/errors/backtraces from the system logs
no warnings/backtraces/kernel message buffer entries/syslog entries
| 1.0 | new fs not mounted after "zfs clone" - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Centos
Distribution Version | 8.3
Linux Kernel | 4.18.0-240.15.1.el8_3
Architecture | x86_64
ZFS Version | 2.0.3-1
SPL Version | 2.0.3-1
### Describe the problem you're observing
tank/myfs with >> files
OK zfs snapshot tank/myfs@current
NOK zfs clone tank/myfs@current tank/myfs_export // OK in a sense of no exception shown
OK zfs list // clone with mountpoint shown
NOK ls /tank/myfs_export or zfs mount -a ls /tank/myfs_export
0 files
### Describe how to reproduce the problem
zfs destroy tank/myfs_export
repeat
same result
### Include any warning/errors/backtraces from the system logs
no warnings/backtraces/kernel message buffer entries/syslog entries
| defect | new fs not mounted after zfs clone system information type version name distribution name centos distribution version linux kernel architecture zfs version spl version describe the problem you re observing tank myfs with files ok zfs snapshot tank myfs current nok zfs clone tank myfs current tank myfs export ok in a sense of no exception shown ok zfs list clone with mountpoint shown nok ls tank myfs export or zfs mount a ls tank myfs export files describe how to reproduce the problem zfs destroy tank myfs export repeat same result include any warning errors backtraces from the system logs no warnings backtraces kernel message buffer entries syslog entries | 1 |
41,580 | 10,524,806,526 | IssuesEvent | 2019-09-30 13:58:48 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Record-level ExpiryPolicy not available/wrong when using an EntryProcessor | Module: ICache Priority: Low Source: Internal Team: Core Type: Defect | If you insert a record with a custom `ExpiryPolicy` (using `put(key, value, expiryPolicy)`), then there are two issues if you attempt to access the record-level expiry policy using an `EntryProcessor`:
- `CacheEntryProcessorEntry` seems to always use the cache configuration level ExpiryPolicy and not the record-level one.
- `CacheRecord.getExpiryPolicy()` always returns `null`. | 1.0 | Record-level ExpiryPolicy not available/wrong when using an EntryProcessor - If you insert a record with a custom `ExpiryPolicy` (using `put(key, value, expiryPolicy)`), then there are two issues if you attempt to access the record-level expiry policy using an `EntryProcessor`:
- `CacheEntryProcessorEntry` seems to always use the cache configuration level ExpiryPolicy and not the record-level one.
- `CacheRecord.getExpiryPolicy()` always returns `null`. | defect | record level expirypolicy not available wrong when using an entryprocessor if you insert a record with a custom expirypolicy using put key value expirypolicy then there are two issues if you attempt to access the record level expiry policy using an entryprocessor cacheentryprocessorentry seems to always use the cache configuration level expirypolicy and not the record level one cacherecord getexpirypolicy always returns null | 1 |
418,528 | 28,118,754,763 | IssuesEvent | 2023-03-31 12:49:20 | OpenAssetIO/OpenAssetIO | https://api.github.com/repos/OpenAssetIO/OpenAssetIO | closed | Document entry point based plugin loading | documentation | ## What
Document how the new entry point based plugin loading can be used.
## Why
Or no-one will know!
## ACs
- Doxygen documentation updated with:
- An "end user" focused section that covers "how to install plugins", that explains both the search path and entry point based mechanism.
- Glossary entry for "entry point plugins"
- Any other references to plugin loading updated to include relevant links. | 1.0 | Document entry point based plugin loading - ## What
Document how the new entry point based plugin loading can be used.
## Why
Or no-one will know!
## ACs
- Doxygen documentation updated with:
- An "end user" focused section that covers "how to install plugins", that explains both the search path and entry point based mechanism.
- Glossary entry for "entry point plugins"
- Any other references to plugin loading updated to include relevant links. | non_defect | document entry point based plugin loading what document how the new entry point based plugin loading can be used why or no one will know acs doxygen documentation updated with an end user focused section that covers how to install plugins that explains both the search path and entry point based mechanism glossary entry for entry point plugins any other references to plugin loading updated to include relevant links | 0 |
470,100 | 13,530,913,402 | IssuesEvent | 2020-09-15 20:42:21 | cassproject/cass-editor | https://api.github.com/repos/cassproject/cass-editor | reopened | Set up a job to check O*NET frameworks for updates periodically | Credential Engine CaSS High Priority | So frameworks don't have to be manually checked and re-imported. | 1.0 | Set up a job to check O*NET frameworks for updates periodically - So frameworks don't have to be manually checked and re-imported. | non_defect | set up a job to check o net frameworks for updates periodically so frameworks don t have to be manually checked and re imported | 0 |
70,508 | 8,556,130,795 | IssuesEvent | 2018-11-08 12:11:31 | horizontalsystems/bank-wallet-ios | https://api.github.com/repos/horizontalsystems/bank-wallet-ios | closed | change the font color of "description texts" | minor design fixes | change the font color of "description texts" on following pages to #grey:
- [x] wallet intro page
- [x] backup wallet,
- [x] backup wallet -> paper keys page
- [x] backup wallet -> submit keys page
- [x] pin pages | 1.0 | change the font color of "description texts" - change the font color of "description texts" on following pages to #grey:
- [x] wallet intro page
- [x] backup wallet,
- [x] backup wallet -> paper keys page
- [x] backup wallet -> submit keys page
- [x] pin pages | non_defect | change the font color of description texts change the font color of description texts on following pages to grey wallet intro page backup wallet backup wallet paper keys page backup wallet submit keys page pin pages | 0 |
5,491 | 2,610,188,675 | IssuesEvent | 2015-02-26 18:59:46 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 揭秘脸上色斑如何消除 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
自信是高调做事的秘诀:信心对于做事成功者具有重要意义��
�成功的欲望是创造和拥有财富的源泉。人一旦有了这种欲望�
��并经由自我暗示和潜意识的激发后形成一种自信心,这种信
心就会转化成一种“积极的感情”,它能帮助人们释放出无��
�的热情、智慧和精力,进而帮助人们获得财富与事业上的巨�
��成就。脸上色斑如何消除,
《客户案例》
由于职业的需要,我经常在外奔波,跑来跑去难免要晒��
�阳光,去年开始脸上长了很多的色斑,本来好好的脸变得很�
��看,一块块的斑长在脸颊和下巴、眼角附近,对着镜子自己
看了都觉得碍眼。为了祛斑我使用了很多祛斑化妆品,有贵��
�有便宜的,也了不少祛斑保健品,但色斑纹丝不动。脸上长�
��么多的斑点,和客户谈业务难免受到影响,因此还差点失去
几笔生意,所以今年我下定决心一定要把脸上的色斑祛除干��
�。可是祛斑说起来容易行动起来难,试过了很多方法都没收�
��效果,心里很着急。现在想想幸亏知道了「黛芙薇尔精华液
」,不然现在我还处在祛斑无方的困扰中呢。<br>
今年一月份的时候我在网上查找到了「黛芙薇尔精华液��
�,看了产品说明知道它是外用的祛斑产品,在线专家对产品�
��介绍和说明我也比较满意,就直接在「黛芙薇尔精华液」商
城上定购了两个周期的祛斑产品。每天按照说明方法使用,��
�概使用了两个星期的时候,我就感觉自己脸上的皮肤好了一�
��,没以前那么干燥了,也不脱屑,色斑倒是还没怎么改变。
长斑不是一朝一夕的事情,也不指望一天两天就把脸上的斑��
�去掉,我继续坚持使用。
等到使用完一个周期的时候,脸上的斑很明显就得到淡化了��
�一块块虽然还长在那里,但是颜色没那么深了,边缘的颜色�
��肤色变得很接近,看起来没以前那么恐怖了,芝麻粒大小的
色斑慢慢融进皮肤里面,渐渐变得像颗小痣。到使用第二个��
�期的时候,祛斑效果更明显了一点,大块的色斑颜色变得浅�
��,而且斑块有缩小的趋势,小粒些的斑点基本上都消失了,
也没有留下色素印子和疤痕。皮肤状况也得到了一些改善,��
�肤有逐渐变白的效果,同事也说我脸上的色斑没那么明显了�
��肤质也好了不少。两个周期的套装使用完以后,我脸上的色
斑已经淡化了很多,皮肤也变得嫩白、光滑了,为了避免反��
�,在「黛芙薇尔精华液」祛斑专家的建议下我又继续使用了�
��个周期的产品,经过三个周期套装的作用,我脸上的斑点基
本上都祛除干净了,到现在半年都过去了,色斑也没有反弹��
�象。
阅读了脸上色斑如何消除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上色斑如何消除,同时为您分享祛斑小方法
去斑方法,柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁,30分钟��
�用水洗去,再涂抹护肤霜,
连续20天,有助于驱除斑点,增白皮肤。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:31 | 1.0 | 揭秘脸上色斑如何消除 - ```
《摘要》
自信是高调做事的秘诀:信心对于做事成功者具有重要意义��
�成功的欲望是创造和拥有财富的源泉。人一旦有了这种欲望�
��并经由自我暗示和潜意识的激发后形成一种自信心,这种信
心就会转化成一种“积极的感情”,它能帮助人们释放出无��
�的热情、智慧和精力,进而帮助人们获得财富与事业上的巨�
��成就。脸上色斑如何消除,
《客户案例》
由于职业的需要,我经常在外奔波,跑来跑去难免要晒��
�阳光,去年开始脸上长了很多的色斑,本来好好的脸变得很�
��看,一块块的斑长在脸颊和下巴、眼角附近,对着镜子自己
看了都觉得碍眼。为了祛斑我使用了很多祛斑化妆品,有贵��
�有便宜的,也了不少祛斑保健品,但色斑纹丝不动。脸上长�
��么多的斑点,和客户谈业务难免受到影响,因此还差点失去
几笔生意,所以今年我下定决心一定要把脸上的色斑祛除干��
�。可是祛斑说起来容易行动起来难,试过了很多方法都没收�
��效果,心里很着急。现在想想幸亏知道了「黛芙薇尔精华液
」,不然现在我还处在祛斑无方的困扰中呢。<br>
今年一月份的时候我在网上查找到了「黛芙薇尔精华液��
�,看了产品说明知道它是外用的祛斑产品,在线专家对产品�
��介绍和说明我也比较满意,就直接在「黛芙薇尔精华液」商
城上定购了两个周期的祛斑产品。每天按照说明方法使用,��
�概使用了两个星期的时候,我就感觉自己脸上的皮肤好了一�
��,没以前那么干燥了,也不脱屑,色斑倒是还没怎么改变。
长斑不是一朝一夕的事情,也不指望一天两天就把脸上的斑��
�去掉,我继续坚持使用。
等到使用完一个周期的时候,脸上的斑很明显就得到淡化了��
�一块块虽然还长在那里,但是颜色没那么深了,边缘的颜色�
��肤色变得很接近,看起来没以前那么恐怖了,芝麻粒大小的
色斑慢慢融进皮肤里面,渐渐变得像颗小痣。到使用第二个��
�期的时候,祛斑效果更明显了一点,大块的色斑颜色变得浅�
��,而且斑块有缩小的趋势,小粒些的斑点基本上都消失了,
也没有留下色素印子和疤痕。皮肤状况也得到了一些改善,��
�肤有逐渐变白的效果,同事也说我脸上的色斑没那么明显了�
��肤质也好了不少。两个周期的套装使用完以后,我脸上的色
斑已经淡化了很多,皮肤也变得嫩白、光滑了,为了避免反��
�,在「黛芙薇尔精华液」祛斑专家的建议下我又继续使用了�
��个周期的产品,经过三个周期套装的作用,我脸上的斑点基
本上都祛除干净了,到现在半年都过去了,色斑也没有反弹��
�象。
阅读了脸上色斑如何消除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上色斑如何消除,同时为您分享祛斑小方法
去斑方法,柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁,30分钟��
�用水洗去,再涂抹护肤霜,
连续20天,有助于驱除斑点,增白皮肤。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:31 | defect | 揭秘脸上色斑如何消除 《摘要》 自信是高调做事的秘诀:信心对于做事成功者具有重要意义�� �成功的欲望是创造和拥有财富的源泉。人一旦有了这种欲望� ��并经由自我暗示和潜意识的激发后形成一种自信心,这种信 心就会转化成一种“积极的感情”,它能帮助人们释放出无�� �的热情、智慧和精力,进而帮助人们获得财富与事业上的巨� ��成就。脸上色斑如何消除, 《客户案例》 由于职业的需要,我经常在外奔波,跑来跑去难免要晒�� �阳光,去年开始脸上长了很多的色斑,本来好好的脸变得很� ��看,一块块的斑长在脸颊和下巴、眼角附近,对着镜子自己 看了都觉得碍眼。为了祛斑我使用了很多祛斑化妆品,有贵�� �有便宜的,也了不少祛斑保健品,但色斑纹丝不动。脸上长� ��么多的斑点,和客户谈业务难免受到影响,因此还差点失去 几笔生意,所以今年我下定决心一定要把脸上的色斑祛除干�� �。可是祛斑说起来容易行动起来难,试过了很多方法都没收� ��效果,心里很着急。现在想想幸亏知道了「黛芙薇尔精华液 」,不然现在我还处在祛斑无方的困扰中呢。 今年一月份的时候我在网上查找到了「黛芙薇尔精华液�� �,看了产品说明知道它是外用的祛斑产品,在线专家对产品� ��介绍和说明我也比较满意,就直接在「黛芙薇尔精华液」商 城上定购了两个周期的祛斑产品。每天按照说明方法使用,�� �概使用了两个星期的时候,我就感觉自己脸上的皮肤好了一� ��,没以前那么干燥了,也不脱屑,色斑倒是还没怎么改变。 长斑不是一朝一夕的事情,也不指望一天两天就把脸上的斑�� �去掉,我继续坚持使用。 等到使用完一个周期的时候,脸上的斑很明显就得到淡化了�� �一块块虽然还长在那里,但是颜色没那么深了,边缘的颜色� ��肤色变得很接近,看起来没以前那么恐怖了,芝麻粒大小的 色斑慢慢融进皮肤里面,渐渐变得像颗小痣。到使用第二个�� �期的时候,祛斑效果更明显了一点,大块的色斑颜色变得浅� ��,而且斑块有缩小的趋势,小粒些的斑点基本上都消失了, 也没有留下色素印子和疤痕。皮肤状况也得到了一些改善,�� �肤有逐渐变白的效果,同事也说我脸上的色斑没那么明显了� ��肤质也好了不少。两个周期的套装使用完以后,我脸上的色 斑已经淡化了很多,皮肤也变得嫩白、光滑了,为了避免反�� �,在「黛芙薇尔精华液」祛斑专家的建议下我又继续使用了� ��个周期的产品,经过三个周期套装的作用,我脸上的斑点基 本上都祛除干净了,到现在半年都过去了,色斑也没有反弹�� �象。 阅读了脸上色斑如何消除,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上色斑如何消除,同时为您分享祛斑小方法 去斑方法 柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁, �� �用水洗去,再涂抹护肤霜, ,有助于驱除斑点,增白皮肤。 original issue reported on code google com by additive gmail com on jul at | 1 |
347,022 | 24,887,448,901 | IssuesEvent | 2022-10-28 09:00:17 | rexong/ped | https://api.github.com/repos/rexong/ped | opened | Showing task of exam: Index not specified in format | type.DocumentationBug severity.Medium | 
`INDEX` was not specified in the format.
<!--session: 1666944017294-04ed4a4d-d16b-4814-8dfd-0887c85486b8-->
<!--Version: Web v3.4.4--> | 1.0 | Showing task of exam: Index not specified in format - 
`INDEX` was not specified in the format.
<!--session: 1666944017294-04ed4a4d-d16b-4814-8dfd-0887c85486b8-->
<!--Version: Web v3.4.4--> | non_defect | showing task of exam index not specified in format index was not specified in the format | 0 |
5,671 | 2,610,193,029 | IssuesEvent | 2015-02-26 19:00:57 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 解密男人脸上色斑怎么去除 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
我不知道它从哪里来,要到哪里去,也不知道它因何而来,��
�何而去。时光的隧道里,它存在了千年,亦穿越了千年。期�
��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自
承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,��
�此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日�
��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划
痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的��
�寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗�
��脸上的雀斑真的无法去除吗!男人脸上色斑怎么去除,
《客户案例》
我的黄褐斑是前两年才长的,时间不是很长,斑也不是��
�严重,主要是前两年工作太忙,压力也比较大,那段时间还�
��烦躁,总是失眠,整个人几乎就是处于一种焦虑的状态,神
经每天都绷得紧紧的,就怕断了,这不,后来就长了斑。为��
�祛斑我是什么方法都用尽啦,食疗啊还有什么祛斑霜啊都用�
��,但是都没用,我是每天都为了这色斑发愁啊。</br>
后来经过朋友介绍,说黛芙薇尔不错,建议我使用一下��
�看看效果怎么样,直到后来使用了黛芙薇尔以后,我才开始�
��信原来坚持祛斑是完全可以成功的,关键在于是否选对方法
,选对方法祛斑如此简单。一旦误入歧途,祛斑确实挺不容��
�的。
后来上网看了看黛芙薇尔的官方网站以后,觉得还不错,而��
�成功案例里面就有个跟自己有着相同经历的长斑人士,既然�
��家使用产品有效,那相信我也可以,在跟黛芙薇尔网站的专
家联系以后,他们也问了我不少情况,非常关心我的病情,��
�也就此咨询了很多问题,在跟黛芙薇尔的专家沟通之后,感�
��专家的专业知识比较强,也没怎么考虑,抱着试一试的心态
。就购买了一个周期的黛芙薇尔。</br>
使用了一个周期之后,竟然发现脸上的晒斑就淡了?!真是
太让我兴奋了!用完两个周期的时候,我脸上的晒斑淡化的几�
��都看不出来了!呵呵,脸上终于又恢复到以前的白嫩肌肤啦��
�再照镜子的时候,我又可以自信的笑了!
阅读了男人脸上色斑怎么去除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
男人脸上色斑怎么去除,同时为您分享祛斑小方法
每天要保证充足的睡眠,劳累会导致皮肤紧张疲倦,血液偏��
�,新陈代谢减缓,那时皮肤将无法取得充足的养分;角质层因
缺乏水分而使皮肤黯然无光。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38 | 1.0 | 解密男人脸上色斑怎么去除 - ```
《摘要》
我不知道它从哪里来,要到哪里去,也不知道它因何而来,��
�何而去。时光的隧道里,它存在了千年,亦穿越了千年。期�
��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自
承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,��
�此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日�
��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划
痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的��
�寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗�
��脸上的雀斑真的无法去除吗!男人脸上色斑怎么去除,
《客户案例》
我的黄褐斑是前两年才长的,时间不是很长,斑也不是��
�严重,主要是前两年工作太忙,压力也比较大,那段时间还�
��烦躁,总是失眠,整个人几乎就是处于一种焦虑的状态,神
经每天都绷得紧紧的,就怕断了,这不,后来就长了斑。为��
�祛斑我是什么方法都用尽啦,食疗啊还有什么祛斑霜啊都用�
��,但是都没用,我是每天都为了这色斑发愁啊。</br>
后来经过朋友介绍,说黛芙薇尔不错,建议我使用一下��
�看看效果怎么样,直到后来使用了黛芙薇尔以后,我才开始�
��信原来坚持祛斑是完全可以成功的,关键在于是否选对方法
,选对方法祛斑如此简单。一旦误入歧途,祛斑确实挺不容��
�的。
后来上网看了看黛芙薇尔的官方网站以后,觉得还不错,而��
�成功案例里面就有个跟自己有着相同经历的长斑人士,既然�
��家使用产品有效,那相信我也可以,在跟黛芙薇尔网站的专
家联系以后,他们也问了我不少情况,非常关心我的病情,��
�也就此咨询了很多问题,在跟黛芙薇尔的专家沟通之后,感�
��专家的专业知识比较强,也没怎么考虑,抱着试一试的心态
。就购买了一个周期的黛芙薇尔。</br>
使用了一个周期之后,竟然发现脸上的晒斑就淡了?!真是
太让我兴奋了!用完两个周期的时候,我脸上的晒斑淡化的几�
��都看不出来了!呵呵,脸上终于又恢复到以前的白嫩肌肤啦��
�再照镜子的时候,我又可以自信的笑了!
阅读了男人脸上色斑怎么去除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
男人脸上色斑怎么去除,同时为您分享祛斑小方法
每天要保证充足的睡眠,劳累会导致皮肤紧张疲倦,血液偏��
�,新陈代谢减缓,那时皮肤将无法取得充足的养分;角质层因
缺乏水分而使皮肤黯然无光。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38 | defect | 解密男人脸上色斑怎么去除 《摘要》 我不知道它从哪里来,要到哪里去,也不知道它因何而来,�� �何而去。时光的隧道里,它存在了千年,亦穿越了千年。期� ��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自 承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,�� �此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日� ��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划 痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的�� �寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗� ��脸上的雀斑真的无法去除吗!男人脸上色斑怎么去除, 《客户案例》 我的黄褐斑是前两年才长的,时间不是很长,斑也不是�� �严重,主要是前两年工作太忙,压力也比较大,那段时间还� ��烦躁,总是失眠,整个人几乎就是处于一种焦虑的状态,神 经每天都绷得紧紧的,就怕断了,这不,后来就长了斑。为�� �祛斑我是什么方法都用尽啦,食疗啊还有什么祛斑霜啊都用� ��,但是都没用,我是每天都为了这色斑发愁啊。 后来经过朋友介绍,说黛芙薇尔不错,建议我使用一下�� �看看效果怎么样,直到后来使用了黛芙薇尔以后,我才开始� ��信原来坚持祛斑是完全可以成功的,关键在于是否选对方法 ,选对方法祛斑如此简单。一旦误入歧途,祛斑确实挺不容�� �的。 后来上网看了看黛芙薇尔的官方网站以后,觉得还不错,而�� �成功案例里面就有个跟自己有着相同经历的长斑人士,既然� ��家使用产品有效,那相信我也可以,在跟黛芙薇尔网站的专 家联系以后,他们也问了我不少情况,非常关心我的病情,�� �也就此咨询了很多问题,在跟黛芙薇尔的专家沟通之后,感� ��专家的专业知识比较强,也没怎么考虑,抱着试一试的心态 。就购买了一个周期的黛芙薇尔。 使用了一个周期之后,竟然发现脸上的晒斑就淡了 真是 太让我兴奋了 用完两个周期的时候,我脸上的晒斑淡化的几� ��都看不出来了 呵呵,脸上终于又恢复到以前的白嫩肌肤啦�� �再照镜子的时候,我又可以自信的笑了 阅读了男人脸上色斑怎么去除,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 男人脸上色斑怎么去除,同时为您分享祛斑小方法 每天要保证充足的睡眠,劳累会导致皮肤紧张疲倦,血液偏�� �,新陈代谢减缓,那时皮肤将无法取得充足的养分 角质层因 缺乏水分而使皮肤黯然无光。 original issue reported on code google com by additive gmail com on jul at | 1 |
82,003 | 7,809,356,377 | IssuesEvent | 2018-06-12 00:05:01 | equella/Equella | https://api.github.com/repos/equella/Equella | closed | New Permission editor: Modifying a privilege only effects bottom entry | Ready for Testing bug | Steps to reproduce:
* On the new course page, either add a new course or edit and existing one
* Click the Permissions tab
* Add some new privileges
* Revoke the top privilege (via the three dot menu)
* The override will actually happen on the bottom entry, not the one you wanted.
| 1.0 | New Permission editor: Modifying a privilege only effects bottom entry - Steps to reproduce:
* On the new course page, either add a new course or edit and existing one
* Click the Permissions tab
* Add some new privileges
* Revoke the top privilege (via the three dot menu)
* The override will actually happen on the bottom entry, not the one you wanted.
| non_defect | new permission editor modifying a privilege only effects bottom entry steps to reproduce on the new course page either add a new course or edit and existing one click the permissions tab add some new privileges revoke the top privilege via the three dot menu the override will actually happen on the bottom entry not the one you wanted | 0 |
72,227 | 3,374,491,813 | IssuesEvent | 2015-11-24 13:20:06 | FAC-GM/app | https://api.github.com/repos/FAC-GM/app | closed | Provide access to app for single new user | priority-1 technical | @SimonLab @heron2014 Please provide Fritjof access to the application as requested via email by Nick. | 1.0 | Provide access to app for single new user - @SimonLab @heron2014 Please provide Fritjof access to the application as requested via email by Nick. | non_defect | provide access to app for single new user simonlab please provide fritjof access to the application as requested via email by nick | 0 |
30,797 | 6,287,482,207 | IssuesEvent | 2017-07-19 15:02:52 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | OpenMPI error in FunctionDT | C: MOOSE Low Difficulty P: normal T: defect | ### Description of the enhancement or error report
Creating a FunctionDT TimeStepper with time_t and time_dt inputs of different sizes causes an unexpected MPI communication error such as
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
[unset]: write_line error; fd=-1 buf=:cmd=abort exitcode=1
:
system msg for write_line failure : Bad file descriptor
There should be a test in the constructor to check that time_t and time_dt have the same size (or where-ever that bugs occurs, I am not sure)
### Rationale for the enhancement or information for reproducing the error
This would provide the user a clearer (and less scary) explanation of why his/her input file causes an error, and what to modify in its input file.
### Identified impact
Changes to FunctionDT TimeStepper or one of its attributes (I am not sure where it happens). | 1.0 | OpenMPI error in FunctionDT - ### Description of the enhancement or error report
Creating a FunctionDT TimeStepper with time_t and time_dt inputs of different sizes causes an unexpected MPI communication error such as
application called MPI_Abort(MPI_COMM_WORLD, 1) - process 0
[unset]: write_line error; fd=-1 buf=:cmd=abort exitcode=1
:
system msg for write_line failure : Bad file descriptor
There should be a test in the constructor to check that time_t and time_dt have the same size (or where-ever that bugs occurs, I am not sure)
### Rationale for the enhancement or information for reproducing the error
This would provide the user a clearer (and less scary) explanation of why his/her input file causes an error, and what to modify in its input file.
### Identified impact
Changes to FunctionDT TimeStepper or one of its attributes (I am not sure where it happens). | defect | openmpi error in functiondt description of the enhancement or error report creating a functiondt timestepper with time t and time dt inputs of different sizes causes an unexpected mpi communication error such as application called mpi abort mpi comm world process write line error fd buf cmd abort exitcode system msg for write line failure bad file descriptor there should be a test in the constructor to check that time t and time dt have the same size or where ever that bugs occurs i am not sure rationale for the enhancement or information for reproducing the error this would provide the user a clearer and less scary explanation of why his her input file causes an error and what to modify in its input file identified impact changes to functiondt timestepper or one of its attributes i am not sure where it happens | 1 |
228,858 | 25,263,097,743 | IssuesEvent | 2022-11-16 01:05:56 | Satheesh575555/linux-3.0.35 | https://api.github.com/repos/Satheesh575555/linux-3.0.35 | opened | CVE-2014-4654 (Medium) detected in linuxlinux-3.0.40 | security vulnerability | ## CVE-2014-4654 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The snd_ctl_elem_add function in sound/core/control.c in the ALSA control implementation in the Linux kernel before 3.15.2 does not check authorization for SNDRV_CTL_IOCTL_ELEM_REPLACE commands, which allows local users to remove kernel controls and cause a denial of service (use-after-free and system crash) by leveraging /dev/snd/controlCX access for an ioctl call.
<p>Publish Date: 2014-07-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-4654>CVE-2014-4654</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-4654">https://nvd.nist.gov/vuln/detail/CVE-2014-4654</a></p>
<p>Release Date: 2014-07-03</p>
<p>Fix Resolution: 3.15.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2014-4654 (Medium) detected in linuxlinux-3.0.40 - ## CVE-2014-4654 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/sound/core/control.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The snd_ctl_elem_add function in sound/core/control.c in the ALSA control implementation in the Linux kernel before 3.15.2 does not check authorization for SNDRV_CTL_IOCTL_ELEM_REPLACE commands, which allows local users to remove kernel controls and cause a denial of service (use-after-free and system crash) by leveraging /dev/snd/controlCX access for an ioctl call.
<p>Publish Date: 2014-07-03
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-4654>CVE-2014-4654</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-4654">https://nvd.nist.gov/vuln/detail/CVE-2014-4654</a></p>
<p>Release Date: 2014-07-03</p>
<p>Fix Resolution: 3.15.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files sound core control c sound core control c sound core control c vulnerability details the snd ctl elem add function in sound core control c in the alsa control implementation in the linux kernel before does not check authorization for sndrv ctl ioctl elem replace commands which allows local users to remove kernel controls and cause a denial of service use after free and system crash by leveraging dev snd controlcx access for an ioctl call publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
48,987 | 13,185,183,980 | IssuesEvent | 2020-08-12 20:53:24 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | liblog4cplus.* not included in <tarball>/lib/tools as it should be. (Trac #567) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/567
, reported by blaufuss and owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "These libs are missing with a \"make tarball\"\n\nRepeatable on linux (El5) and Mac OS x",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876862109",
"component": "cmake",
"summary": "liblog4cplus.* not included in <tarball>/lib/tools as it should be.",
"priority": "normal",
"keywords": "",
"time": "2009-07-08T21:53:44",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | liblog4cplus.* not included in <tarball>/lib/tools as it should be. (Trac #567) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/567
, reported by blaufuss and owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "These libs are missing with a \"make tarball\"\n\nRepeatable on linux (El5) and Mac OS x",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876862109",
"component": "cmake",
"summary": "liblog4cplus.* not included in <tarball>/lib/tools as it should be.",
"priority": "normal",
"keywords": "",
"time": "2009-07-08T21:53:44",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | not included in lib tools as it should be trac migrated from reported by blaufuss and owned by troy json status closed changetime description these libs are missing with a make tarball n nrepeatable on linux and mac os x reporter blaufuss cc resolution fixed ts component cmake summary not included in lib tools as it should be priority normal keywords time milestone owner troy type defect | 1 |
69,295 | 22,319,744,233 | IssuesEvent | 2022-06-14 04:32:14 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Storage consistency checks fail | T-Defect Z-Platform-Specific Z-Rageshake | ```
2020-02-18T08:47:43.640Z I StorageManager: Checking storage consistency
2020-02-18T08:47:43.640Z I StorageManager: Local storage supported? true
2020-02-18T08:47:43.641Z I StorageManager: IndexedDB supported? true
2020-02-18T08:47:43.641Z I StorageManager: Local storage contains data? true
2020-02-18T08:47:43.642Z I StorageManager: Crypto initialised? true
2020-02-18T08:47:43.776Z I StorageManager: Sync store using IndexedDB contains data? false
2020-02-18T08:47:43.952Z I StorageManager: Crypto store using IndexedDB contains data? false
2020-02-18T08:47:43.953Z E StorageManager: Data exists in local storage and crypto is marked as initialised but no data found in crypto store. IndexedDB storage has likely been evicted by the browser!
2020-02-18T08:47:43.955Z E StorageManager: Storage consistency checks failed
``` | 1.0 | Storage consistency checks fail - ```
2020-02-18T08:47:43.640Z I StorageManager: Checking storage consistency
2020-02-18T08:47:43.640Z I StorageManager: Local storage supported? true
2020-02-18T08:47:43.641Z I StorageManager: IndexedDB supported? true
2020-02-18T08:47:43.641Z I StorageManager: Local storage contains data? true
2020-02-18T08:47:43.642Z I StorageManager: Crypto initialised? true
2020-02-18T08:47:43.776Z I StorageManager: Sync store using IndexedDB contains data? false
2020-02-18T08:47:43.952Z I StorageManager: Crypto store using IndexedDB contains data? false
2020-02-18T08:47:43.953Z E StorageManager: Data exists in local storage and crypto is marked as initialised but no data found in crypto store. IndexedDB storage has likely been evicted by the browser!
2020-02-18T08:47:43.955Z E StorageManager: Storage consistency checks failed
``` | defect | storage consistency checks fail i storagemanager checking storage consistency i storagemanager local storage supported true i storagemanager indexeddb supported true i storagemanager local storage contains data true i storagemanager crypto initialised true i storagemanager sync store using indexeddb contains data false i storagemanager crypto store using indexeddb contains data false e storagemanager data exists in local storage and crypto is marked as initialised but no data found in crypto store indexeddb storage has likely been evicted by the browser e storagemanager storage consistency checks failed | 1 |
56,177 | 6,965,588,035 | IssuesEvent | 2017-12-09 08:13:43 | netlify/netlify-cms | https://api.github.com/repos/netlify/netlify-cms | closed | Cleanup Icons | help wanted needs design | Right now we have multiple different icon sets that are used in different parts of the UI. We currently have [Material Icons](https://github.com/netlify/netlify-cms/blob/b2d4714c903b403eb4c5cbe0d0d218769408188f/src/material-icons.css), [Entypo (v2?), and Font Awesome editor icons](https://github.com/netlify/netlify-cms/tree/0521757b8c674d74974d5d68042bf0965a48f7f6/src/components/UI/icon). These should probably be combined and only the ones that are actually used be included. We could also set up a way to automate icon updating/adding if possible, maybe using an NPM script and a font/icon service like Fontello's API.
Related to #418.
Any comments or suggestions? | 1.0 | Cleanup Icons - Right now we have multiple different icon sets that are used in different parts of the UI. We currently have [Material Icons](https://github.com/netlify/netlify-cms/blob/b2d4714c903b403eb4c5cbe0d0d218769408188f/src/material-icons.css), [Entypo (v2?), and Font Awesome editor icons](https://github.com/netlify/netlify-cms/tree/0521757b8c674d74974d5d68042bf0965a48f7f6/src/components/UI/icon). These should probably be combined and only the ones that are actually used be included. We could also set up a way to automate icon updating/adding if possible, maybe using an NPM script and a font/icon service like Fontello's API.
Related to #418.
Any comments or suggestions? | non_defect | cleanup icons right now we have multiple different icon sets that are used in different parts of the ui we currently have these should probably be combined and only the ones that are actually used be included we could also set up a way to automate icon updating adding if possible maybe using an npm script and a font icon service like fontello s api related to any comments or suggestions | 0 |
111,446 | 24,129,778,016 | IssuesEvent | 2022-09-21 06:13:37 | unicode-org/icu4x | https://api.github.com/repos/unicode-org/icu4x | closed | Label experimental normalizer functions with the proper docs | T-docs-tests C-unicode S-tiny | Noticed by @markusicu:
> Add experimental flag in Rust docs to `ComposingNormalizer::try_new_uts46_…unstable()`. Since it’s unstable, we can remove the `_any_` and `_buffer_` versions of it. | 1.0 | Label experimental normalizer functions with the proper docs - Noticed by @markusicu:
> Add experimental flag in Rust docs to `ComposingNormalizer::try_new_uts46_…unstable()`. Since it’s unstable, we can remove the `_any_` and `_buffer_` versions of it. | non_defect | label experimental normalizer functions with the proper docs noticed by markusicu add experimental flag in rust docs to composingnormalizer try new …unstable since it’s unstable we can remove the any and buffer versions of it | 0 |
16,188 | 3,509,703,493 | IssuesEvent | 2016-01-09 00:30:29 | UCI-UAVForge/Avionics | https://api.github.com/repos/UCI-UAVForge/Avionics | closed | Integrate Radio Comms and Data Management Modules | scope: data management scope: radio comms type: testing | Once radio comms and minimal data management modules are written, test that they are able to communicate with each other as expected. | 1.0 | Integrate Radio Comms and Data Management Modules - Once radio comms and minimal data management modules are written, test that they are able to communicate with each other as expected. | non_defect | integrate radio comms and data management modules once radio comms and minimal data management modules are written test that they are able to communicate with each other as expected | 0 |
43,244 | 11,577,876,220 | IssuesEvent | 2020-02-21 14:56:04 | LiskHQ/lisk-desktop | https://api.github.com/repos/LiskHQ/lisk-desktop | closed | BTC transaction history is not working | severity: high type: bug type: defect type: unplanned | ### Expected behavior
In Bitcoin wallet & Dashboard page, user can select any of the BTC transaction items and view the complete details of the specific BTC transaction
### Actual behavior
As a user, i selected one of the BTC transaction item from BTC wallet and Dashboard page. Instead of transaction details page, i have been sent to the generic error page
### Which version(s) does this affect? (Environment, OS, etc...)
Lisk-Desktop | 1.0 | BTC transaction history is not working - ### Expected behavior
In Bitcoin wallet & Dashboard page, user can select any of the BTC transaction items and view the complete details of the specific BTC transaction
### Actual behavior
As a user, i selected one of the BTC transaction item from BTC wallet and Dashboard page. Instead of transaction details page, i have been sent to the generic error page
### Which version(s) does this affect? (Environment, OS, etc...)
Lisk-Desktop | defect | btc transaction history is not working expected behavior in bitcoin wallet dashboard page user can select any of the btc transaction items and view the complete details of the specific btc transaction actual behavior as a user i selected one of the btc transaction item from btc wallet and dashboard page instead of transaction details page i have been sent to the generic error page which version s does this affect environment os etc lisk desktop | 1 |
775,683 | 27,235,347,786 | IssuesEvent | 2023-02-21 15:57:01 | ascheid/itsg33-pbmm-issue-gen | https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen | opened | AC-2(-4): Account Management | Priority: P2 Class: Technical ITSG-33 Control: AC-2 Suggested Assignment: IT Projects | #Control Definition
ACCOUNT MANAGEMENT | AUTOMATED AUDIT ACTIONS
The information system automatically audits account creation, modification, enabling, disabling, and removal actions, and notifies [Assignment: organization-defined personnel or roles]. Related controls: AU-2, AU-12.
#Class
Technical
#Suggested Placeholder Values
X
#Suggested Assignment
IT Projects
#Support Teams
IT Security Function, IT Operations Group
| 1.0 | AC-2(-4): Account Management - #Control Definition
ACCOUNT MANAGEMENT | AUTOMATED AUDIT ACTIONS
The information system automatically audits account creation, modification, enabling, disabling, and removal actions, and notifies [Assignment: organization-defined personnel or roles]. Related controls: AU-2, AU-12.
#Class
Technical
#Suggested Placeholder Values
X
#Suggested Assignment
IT Projects
#Support Teams
IT Security Function, IT Operations Group
| non_defect | ac account management control definition account management automated audit actions the information system automatically audits account creation modification enabling disabling and removal actions and notifies related controls au au class technical suggested placeholder values x suggested assignment it projects support teams it security function it operations group | 0 |
49,094 | 13,185,227,701 | IssuesEvent | 2020-08-12 20:58:46 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | python 3.4 detection (Trac #732) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/732
, reported by david.schultz and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T23:10:39",
"description": "14:04:36 - JoaoPedr1: in cmake tools/python.cmake when looking for the python version the ERROR_VARIABLE is read. When I updated to python 3.4 with archlinux it stopped being able to get the version, changing ERROR_VARIABLE to OUTPUT_VARIABLE solved the problem\n\nSounds like a change from printing the version to stderr, to printing to stdout.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1547248239669835",
"component": "cmake",
"summary": "python 3.4 detection",
"priority": "normal",
"keywords": "",
"time": "2014-06-13T19:31:21",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | python 3.4 detection (Trac #732) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/732
, reported by david.schultz and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T23:10:39",
"description": "14:04:36 - JoaoPedr1: in cmake tools/python.cmake when looking for the python version the ERROR_VARIABLE is read. When I updated to python 3.4 with archlinux it stopped being able to get the version, changing ERROR_VARIABLE to OUTPUT_VARIABLE solved the problem\n\nSounds like a change from printing the version to stderr, to printing to stdout.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1547248239669835",
"component": "cmake",
"summary": "python 3.4 detection",
"priority": "normal",
"keywords": "",
"time": "2014-06-13T19:31:21",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | python detection trac migrated from reported by david schultz and owned by nega json status closed changetime description in cmake tools python cmake when looking for the python version the error variable is read when i updated to python with archlinux it stopped being able to get the version changing error variable to output variable solved the problem n nsounds like a change from printing the version to stderr to printing to stdout reporter david schultz cc resolution fixed ts component cmake summary python detection priority normal keywords time milestone owner nega type defect | 1 |
191,726 | 15,302,195,786 | IssuesEvent | 2021-02-24 14:29:07 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | Misleading explanation of input variable for numpy.random.poisson | 04 - Documentation component: numpy.random | The original documentation is: https://numpy.org/doc/stable/reference/random/generated/numpy.random.poisson.html
The references for correction are:
* Weisstein, Eric W. “Poisson Distribution.” From MathWorld–A Wolfram Web Resource. http://mathworld.wolfram.com/PoissonDistribution.html
* Wikipedia, “Poisson distribution”, https://en.wikipedia.org/wiki/Poisson_distribution
The original documentation for numpy.random.poisson explains parameter ***lam*** (short for lambda) as `Expectation of interval, must be >= 0. A sequence of expectation intervals must be broadcastable over the requested size.` However, ***lam*** is not the expected interval (or at least not the expected time interval of an event), but the expected number of event occurrences in a fixed time interval.
The explanations of ***lam*** throughout this API documentation is misleading.
***lam*** in numpy is identical to ***v*** in the MathWorld reference and ***lambda*** in the Wikipedia reference. | 1.0 | Misleading explanation of input variable for numpy.random.poisson - The original documentation is: https://numpy.org/doc/stable/reference/random/generated/numpy.random.poisson.html
The references for correction are:
* Weisstein, Eric W. “Poisson Distribution.” From MathWorld–A Wolfram Web Resource. http://mathworld.wolfram.com/PoissonDistribution.html
* Wikipedia, “Poisson distribution”, https://en.wikipedia.org/wiki/Poisson_distribution
The original documentation for numpy.random.poisson explains parameter ***lam*** (short for lambda) as `Expectation of interval, must be >= 0. A sequence of expectation intervals must be broadcastable over the requested size.` However, ***lam*** is not the expected interval (or at least not the expected time interval of an event), but the expected number of event occurrences in a fixed time interval.
The explanations of ***lam*** throughout this API documentation is misleading.
***lam*** in numpy is identical to ***v*** in the MathWorld reference and ***lambda*** in the Wikipedia reference. | non_defect | misleading explanation of input variable for numpy random poisson the original documentation is the references for correction are weisstein eric w “poisson distribution ” from mathworld–a wolfram web resource wikipedia “poisson distribution” the original documentation for numpy random poisson explains parameter lam short for lambda as expectation of interval must be a sequence of expectation intervals must be broadcastable over the requested size however lam is not the expected interval or at least not the expected time interval of an event but the expected number of event occurrences in a fixed time interval the explanations of lam throughout this api documentation is misleading lam in numpy is identical to v in the mathworld reference and lambda in the wikipedia reference | 0 |
19,883 | 3,273,281,781 | IssuesEvent | 2015-10-26 00:01:50 | jsr107/jsr107spec | https://api.github.com/repos/jsr107/jsr107spec | closed | javax.cache.configuration is not exported in manifest (OSGi issue) | Defect | Hi there.
It seems that cache-api can't be used in OSGi environment since javax.cache.event is not available in export-package. I checked bnd.bnd file and didn't find javax.cache.event package there.
Is it a bug or maybe I use cache-api in a wrong way ?
Thanks,
Vlad | 1.0 | javax.cache.configuration is not exported in manifest (OSGi issue) - Hi there.
It seems that cache-api can't be used in OSGi environment since javax.cache.event is not available in export-package. I checked bnd.bnd file and didn't find javax.cache.event package there.
Is it a bug or maybe I use cache-api in a wrong way ?
Thanks,
Vlad | defect | javax cache configuration is not exported in manifest osgi issue hi there it seems that cache api can t be used in osgi environment since javax cache event is not available in export package i checked bnd bnd file and didn t find javax cache event package there is it a bug or maybe i use cache api in a wrong way thanks vlad | 1 |
60,278 | 17,023,386,418 | IssuesEvent | 2021-07-03 01:45:39 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Redraw artefacts | Component: merkaartor Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 10.38am, Sunday, 19th April 2009]**
Merkaartor 0.12 and 0.13.1 are unusable for me as the redraw doesn't work properly. Merkaartor 0.10 still works.
Moving around or zooming causes a ghosting effect, and all or some of the map may have noise over it.
I'm using Linux 2.6.29 with Xorg 1.5.3 and fglrx 8.582. | 1.0 | Redraw artefacts - **[Submitted to the original trac issue database at 10.38am, Sunday, 19th April 2009]**
Merkaartor 0.12 and 0.13.1 are unusable for me as the redraw doesn't work properly. Merkaartor 0.10 still works.
Moving around or zooming causes a ghosting effect, and all or some of the map may have noise over it.
I'm using Linux 2.6.29 with Xorg 1.5.3 and fglrx 8.582. | defect | redraw artefacts merkaartor and are unusable for me as the redraw doesn t work properly merkaartor still works moving around or zooming causes a ghosting effect and all or some of the map may have noise over it i m using linux with xorg and fglrx | 1 |
701,456 | 24,098,786,398 | IssuesEvent | 2022-09-19 21:29:39 | CDCgov/prime-reportstream | https://api.github.com/repos/CDCgov/prime-reportstream | closed | CO - CDPHE ELR Onboarding automated error message [Error: Cannot Write File to DB] | onboarding-ops support Medium Priority Colorado | This is an automatic email from the CDPHE Electronic Lab Reporting (ELR) system. There was an issue with a file or message submitted by your organization.
Issue Report: Cannot insert the file into the database due to the following reason:
**************************************************
Violation of UNIQUE KEY constraint 'IX_ImportFile'. Cannot insert duplicate key in object 'dbo.ImportFile'. The duplicate key value is (CDC PRIME, `covid-19-dca4a58d-41ed-41c3-8eb9-6b8bc97a8a33-20220819013031.hl7`).
**************************************************
A common cause for the failure to write a file to the database is that the submitted file has the same name as a previously submitted file. Please note that each file submitted must have a unique name.
Message ID: null
Received by: CDPHE ELR Onboarding
Received on: 2022-08-19 01:30:27
File source: CDC PRIME
File name: `covid-19-dca4a58d-41ed-41c3-8eb9-6b8bc97a8a33-20220819013031.hl7`
Performing Org Name: null
Please review, correct, and resubmit the file.
Please contact cdphe_elr@state.co.us if you have any questions about this email. | 1.0 | CO - CDPHE ELR Onboarding automated error message [Error: Cannot Write File to DB] - This is an automatic email from the CDPHE Electronic Lab Reporting (ELR) system. There was an issue with a file or message submitted by your organization.
Issue Report: Cannot insert the file into the database due to the following reason:
**************************************************
Violation of UNIQUE KEY constraint 'IX_ImportFile'. Cannot insert duplicate key in object 'dbo.ImportFile'. The duplicate key value is (CDC PRIME, `covid-19-dca4a58d-41ed-41c3-8eb9-6b8bc97a8a33-20220819013031.hl7`).
**************************************************
A common cause for the failure to write a file to the database is that the submitted file has the same name as a previously submitted file. Please note that each file submitted must have a unique name.
Message ID: null
Received by: CDPHE ELR Onboarding
Received on: 2022-08-19 01:30:27
File source: CDC PRIME
File name: `covid-19-dca4a58d-41ed-41c3-8eb9-6b8bc97a8a33-20220819013031.hl7`
Performing Org Name: null
Please review, correct, and resubmit the file.
Please contact cdphe_elr@state.co.us if you have any questions about this email. | non_defect | co cdphe elr onboarding automated error message this is an automatic email from the cdphe electronic lab reporting elr system there was an issue with a file or message submitted by your organization issue report cannot insert the file into the database due to the following reason violation of unique key constraint ix importfile cannot insert duplicate key in object dbo importfile the duplicate key value is cdc prime covid a common cause for the failure to write a file to the database is that the submitted file has the same name as a previously submitted file please note that each file submitted must have a unique name message id null received by cdphe elr onboarding received on file source cdc prime file name covid performing org name null please review correct and resubmit the file please contact cdphe elr state co us if you have any questions about this email | 0 |
245,616 | 26,549,306,843 | IssuesEvent | 2023-01-20 05:30:27 | nidhi7598/linux-3.0.35_CVE-2022-45934 | https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2022-45934 | opened | CVE-2019-18198 (High) detected in linux-stable-rtv3.8.6 | security vulnerability | ## CVE-2019-18198 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.3.4, a reference count usage error in the fib6_rule_suppress() function in the fib6 suppression feature of net/ipv6/fib6_rules.c, when handling the FIB_LOOKUP_NOREF flag, can be exploited by a local attacker to corrupt memory, aka CID-ca7a03c41753.
<p>Publish Date: 2019-10-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18198>CVE-2019-18198</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198</a></p>
<p>Release Date: 2019-10-31</p>
<p>Fix Resolution: v5.4-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-18198 (High) detected in linux-stable-rtv3.8.6 - ## CVE-2019-18198 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/fib6_rules.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.3.4, a reference count usage error in the fib6_rule_suppress() function in the fib6 suppression feature of net/ipv6/fib6_rules.c, when handling the FIB_LOOKUP_NOREF flag, can be exploited by a local attacker to corrupt memory, aka CID-ca7a03c41753.
<p>Publish Date: 2019-10-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18198>CVE-2019-18198</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18198</a></p>
<p>Release Date: 2019-10-31</p>
<p>Fix Resolution: v5.4-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net rules c net rules c net rules c vulnerability details in the linux kernel before a reference count usage error in the rule suppress function in the suppression feature of net rules c when handling the fib lookup noref flag can be exploited by a local attacker to corrupt memory aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
7,850 | 2,611,052,683 | IssuesEvent | 2015-02-27 00:24:17 | alistairreilly/andors-trail | https://api.github.com/repos/alistairreilly/andors-trail | closed | Remove the "Pause" button? | auto-migrated Milestone-0.6.7 Priority-Medium Type-Defect | ```
There is currently a "Pause" button if you press menu.
Should we really keep this around? You can more easily pause by pressing back
instead.
```
Original issue reported on code.google.com by `oskar.wi...@gmail.com` on 8 Dec 2010 at 9:28 | 1.0 | Remove the "Pause" button? - ```
There is currently a "Pause" button if you press menu.
Should we really keep this around? You can more easily pause by pressing back
instead.
```
Original issue reported on code.google.com by `oskar.wi...@gmail.com` on 8 Dec 2010 at 9:28 | defect | remove the pause button there is currently a pause button if you press menu should we really keep this around you can more easily pause by pressing back instead original issue reported on code google com by oskar wi gmail com on dec at | 1 |
49 | 2,497,283,854 | IssuesEvent | 2015-01-07 03:56:46 | CompEvol/beast2 | https://api.github.com/repos/CompEvol/beast2 | closed | BEAUti TN93 problem | BEAUti defect | From the BEAST user list:
A note of caution for Beauti 2 users:
Choosing TrN as the substitution model for a partition with automatic clock rates unchecked:
If a partitioned analysis (but not *Beast) is being used and "automatic set clock rates" is unchecked, and then TrN is selected as a substitution model for a partition, the two kappas don't show up in the priors panel even though they appear in the model panel. Still, if the xml is generated, the kappas *are* included, and as estimated parameters as well (i.e., with their operators). The way to get the kappas to show in the priors panel is to choose GTR first, view the parameters (which appear in the priors window in that case), then switch the GTR partition to TrN. One confusing further problem, however, is that one of the kappas will be set as not estimated, which is easy to miss going forward.
IBeauti 2 is also inconsistent in how it handles the GTR parameters under the "automatic set clock rates unchecked" condition. If the GTR model is the first one selected for that partition, it leaves one of the instantaneous rates set to be not estimated, which seems correct. If TrN was first selected for that partition, however (and if both kappas were checked as estimated, which is the default), it will set all six GTR rates as estimated when GTR is chosen after TrN was initially set.
Obviously is good practice to check each element of the xml file generated by beauti to ensure that the analysis meets expectations.
This pertains to the version of Beauti that comes with Beast 2.1.3.
If there is a better place for bug reports please let me know. Thanks,
Dave Marshall
UConn | 1.0 | BEAUti TN93 problem - From the BEAST user list:
A note of caution for Beauti 2 users:
Choosing TrN as the substitution model for a partition with automatic clock rates unchecked:
If a partitioned analysis (but not *Beast) is being used and "automatic set clock rates" is unchecked, and then TrN is selected as a substitution model for a partition, the two kappas don't show up in the priors panel even though they appear in the model panel. Still, if the xml is generated, the kappas *are* included, and as estimated parameters as well (i.e., with their operators). The way to get the kappas to show in the priors panel is to choose GTR first, view the parameters (which appear in the priors window in that case), then switch the GTR partition to TrN. One confusing further problem, however, is that one of the kappas will be set as not estimated, which is easy to miss going forward.
IBeauti 2 is also inconsistent in how it handles the GTR parameters under the "automatic set clock rates unchecked" condition. If the GTR model is the first one selected for that partition, it leaves one of the instantaneous rates set to be not estimated, which seems correct. If TrN was first selected for that partition, however (and if both kappas were checked as estimated, which is the default), it will set all six GTR rates as estimated when GTR is chosen after TrN was initially set.
Obviously is good practice to check each element of the xml file generated by beauti to ensure that the analysis meets expectations.
This pertains to the version of Beauti that comes with Beast 2.1.3.
If there is a better place for bug reports please let me know. Thanks,
Dave Marshall
UConn | defect | beauti problem from the beast user list a note of caution for beauti users choosing trn as the substitution model for a partition with automatic clock rates unchecked if a partitioned analysis but not beast is being used and automatic set clock rates is unchecked and then trn is selected as a substitution model for a partition the two kappas don t show up in the priors panel even though they appear in the model panel still if the xml is generated the kappas are included and as estimated parameters as well i e with their operators the way to get the kappas to show in the priors panel is to choose gtr first view the parameters which appear in the priors window in that case then switch the gtr partition to trn one confusing further problem however is that one of the kappas will be set as not estimated which is easy to miss going forward ibeauti is also inconsistent in how it handles the gtr parameters under the automatic set clock rates unchecked condition if the gtr model is the first one selected for that partition it leaves one of the instantaneous rates set to be not estimated which seems correct if trn was first selected for that partition however and if both kappas were checked as estimated which is the default it will set all six gtr rates as estimated when gtr is chosen after trn was initially set obviously is good practice to check each element of the xml file generated by beauti to ensure that the analysis meets expectations this pertains to the version of beauti that comes with beast if there is a better place for bug reports please let me know thanks dave marshall uconn | 1 |
4,838 | 2,610,157,885 | IssuesEvent | 2015-02-26 18:50:09 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Map Issue | auto-migrated Priority-Medium Type-Defect | ```
Yellow Stripe on Buildable CIS turrets in GC
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:56 | 1.0 | Map Issue - ```
Yellow Stripe on Buildable CIS turrets in GC
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:56 | defect | map issue yellow stripe on buildable cis turrets in gc original issue reported on code google com by gmail com on jan at | 1 |
47,755 | 13,066,196,217 | IssuesEvent | 2020-07-30 21:11:27 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [payload-parsing] TODOs (Trac #1127) | Migrated from Trac combo core defect | Fix or eliminate TODOs:
```text
public/payload-parsing/utility.h: * @todo seems more natural to put 'start' at the end.
```
Migrated from https://code.icecube.wisc.edu/ticket/1127
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:10",
"description": "Fix or eliminate TODOs:\n\n{{{\npublic/payload-parsing/utility.h: * @todo seems more natural to put 'start' at the end.\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1458335650323600",
"component": "combo core",
"summary": "[payload-parsing] TODOs",
"priority": "critical",
"keywords": "",
"time": "2015-08-17T17:07:30",
"milestone": "",
"owner": "rmaunu",
"type": "defect"
}
```
| 1.0 | [payload-parsing] TODOs (Trac #1127) - Fix or eliminate TODOs:
```text
public/payload-parsing/utility.h: * @todo seems more natural to put 'start' at the end.
```
Migrated from https://code.icecube.wisc.edu/ticket/1127
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:10",
"description": "Fix or eliminate TODOs:\n\n{{{\npublic/payload-parsing/utility.h: * @todo seems more natural to put 'start' at the end.\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"_ts": "1458335650323600",
"component": "combo core",
"summary": "[payload-parsing] TODOs",
"priority": "critical",
"keywords": "",
"time": "2015-08-17T17:07:30",
"milestone": "",
"owner": "rmaunu",
"type": "defect"
}
```
| defect | todos trac fix or eliminate todos text public payload parsing utility h todo seems more natural to put start at the end migrated from json status closed changetime description fix or eliminate todos n n npublic payload parsing utility h todo seems more natural to put start at the end n reporter david schultz cc resolution wontfix ts component combo core summary todos priority critical keywords time milestone owner rmaunu type defect | 1 |
74,545 | 25,165,603,602 | IssuesEvent | 2022-11-10 20:32:14 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | opened | PHP 8.1 Testing: Errors found on Compare Revisions page | Defect Needs refining | ## Describe the defect
Errors were found on the Compare revisions page when testing PHP 8.1:
> Deprecated function: strlen(): Passing null to parameter #1 ($string) of type string is deprecated in Drupal\Component\Utility\Unicode::validateUtf8() (line 478 of core/lib/Drupal/Component/Utility/Unicode.php).
> Deprecated function: str_replace(): Passing null to parameter #3 ($subject) of type array|string is deprecated in Drupal\Component\Utility\Xss::filter() (line 69 of core/lib/Drupal/Component/Utility/Xss.php).
> Deprecated function: strlen(): Passing null to parameter #1 ($string) of type string is deprecated in Drupal\Component\Utility\Unicode::validateUtf8() (line 478 of core/lib/Drupal/Component/Utility/Unicode.php).
> Deprecated function: str_replace(): Passing null to parameter #3 ($subject) of type array|string is deprecated in Drupal\Component\Utility\Xss::filter() (line 69 of core/lib/Drupal/Component/Utility/Xss.php).
## To Reproduce
Steps to reproduce the behavior:
1. Go to https://vacms-10870-upgrade-to-php81-rh7pjindaxvyxvr35alvilr3mqpgaoko.ci.cms.va.gov/node/50765/revisions/view/721480/721481/visual_inline
2. See errors being reporting at the top of the page
## AC / Expected behavior
- [ ] Investigate/Resolve errors
## Screenshots

## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [x] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| 1.0 | PHP 8.1 Testing: Errors found on Compare Revisions page - ## Describe the defect
Errors were found on the Compare revisions page when testing PHP 8.1:
> Deprecated function: strlen(): Passing null to parameter #1 ($string) of type string is deprecated in Drupal\Component\Utility\Unicode::validateUtf8() (line 478 of core/lib/Drupal/Component/Utility/Unicode.php).
> Deprecated function: str_replace(): Passing null to parameter #3 ($subject) of type array|string is deprecated in Drupal\Component\Utility\Xss::filter() (line 69 of core/lib/Drupal/Component/Utility/Xss.php).
> Deprecated function: strlen(): Passing null to parameter #1 ($string) of type string is deprecated in Drupal\Component\Utility\Unicode::validateUtf8() (line 478 of core/lib/Drupal/Component/Utility/Unicode.php).
> Deprecated function: str_replace(): Passing null to parameter #3 ($subject) of type array|string is deprecated in Drupal\Component\Utility\Xss::filter() (line 69 of core/lib/Drupal/Component/Utility/Xss.php).
## To Reproduce
Steps to reproduce the behavior:
1. Go to https://vacms-10870-upgrade-to-php81-rh7pjindaxvyxvr35alvilr3mqpgaoko.ci.cms.va.gov/node/50765/revisions/view/721480/721481/visual_inline
2. See errors being reporting at the top of the page
## AC / Expected behavior
- [ ] Investigate/Resolve errors
## Screenshots

## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [x] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| defect | php testing errors found on compare revisions page describe the defect errors were found on the compare revisions page when testing php deprecated function strlen passing null to parameter string of type string is deprecated in drupal component utility unicode line of core lib drupal component utility unicode php deprecated function str replace passing null to parameter subject of type array string is deprecated in drupal component utility xss filter line of core lib drupal component utility xss php deprecated function strlen passing null to parameter string of type string is deprecated in drupal component utility unicode line of core lib drupal component utility unicode php deprecated function str replace passing null to parameter subject of type array string is deprecated in drupal component utility xss filter line of core lib drupal component utility xss php to reproduce steps to reproduce the behavior go to see errors being reporting at the top of the page ac expected behavior investigate resolve errors screenshots additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support | 1 |
48,730 | 13,184,725,594 | IssuesEvent | 2020-08-12 19:58:53 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | i3 files no longer readable on diablo. (Trac #109) | Incomplete Migration Migrated from Trac dataio defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/109
, reported by blaufuss and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:41:56",
"description": "PPC arch is complaining about reading all .i3 files:\n\nFATAL: File does not appear to be an .i3 file.\n\nSeems to be a big/little endian issue added with lazy frame.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1194648116000000",
"component": "dataio",
"summary": "i3 files no longer readable on diablo.",
"priority": "normal",
"keywords": "",
"time": "2007-09-06T14:36:02",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | i3 files no longer readable on diablo. (Trac #109) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/109
, reported by blaufuss and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:41:56",
"description": "PPC arch is complaining about reading all .i3 files:\n\nFATAL: File does not appear to be an .i3 file.\n\nSeems to be a big/little endian issue added with lazy frame.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1194648116000000",
"component": "dataio",
"summary": "i3 files no longer readable on diablo.",
"priority": "normal",
"keywords": "",
"time": "2007-09-06T14:36:02",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | files no longer readable on diablo trac migrated from reported by blaufuss and owned by troy json status closed changetime description ppc arch is complaining about reading all files n nfatal file does not appear to be an file n nseems to be a big little endian issue added with lazy frame n n reporter blaufuss cc resolution fixed ts component dataio summary files no longer readable on diablo priority normal keywords time milestone owner troy type defect | 1 |
446,269 | 31,466,169,097 | IssuesEvent | 2023-08-30 02:17:10 | DevOps-MBSE/AaC | https://api.github.com/repos/DevOps-MBSE/AaC | closed | I want user-specific documentation on what an AaC model is, represents, and can be used for | documentation | As an AaC user, I want user-specific documentation on what an AaC model is, represents, and can be used for, so that I can get a more intimate familiarity with the `model` root type.
## AC:
- [ ] There is user documentation for the AaC `model` concept
- [ ] The `model` documentation explains:
- [ ] What kind of data does the `model` definition represent?
- [ ] What kind of data is represented in `model` definition sub-structures such as `behaviors`
- [ ] How would a user incorporate `model` into a modeled system (e.g. using model to represent services) | 1.0 | I want user-specific documentation on what an AaC model is, represents, and can be used for - As an AaC user, I want user-specific documentation on what an AaC model is, represents, and can be used for, so that I can get a more intimate familiarity with the `model` root type.
## AC:
- [ ] There is user documentation for the AaC `model` concept
- [ ] The `model` documentation explains:
- [ ] What kind of data does the `model` definition represent?
- [ ] What kind of data is represented in `model` definition sub-structures such as `behaviors`
- [ ] How would a user incorporate `model` into a modeled system (e.g. using model to represent services) | non_defect | i want user specific documentation on what an aac model is represents and can be used for as an aac user i want user specific documentation on what an aac model is represents and can be used for so that i can get a more intimate familiarity with the model root type ac there is user documentation for the aac model concept the model documentation explains what kind of data does the model definition represent what kind of data is represented in model definition sub structures such as behaviors how would a user incorporate model into a modeled system e g using model to represent services | 0 |
54,035 | 13,338,131,856 | IssuesEvent | 2020-08-28 10:27:02 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | Calendar/DatePicker: Error with SimpleDatePattern with 1-letter year (eg. dd.MM.y) | defect | **Describe the defect**
So unless I understand the docs wrongly, the `pattern` option on <p:calendar> and <p:datePicker> allows for a [SimpleDateFormat](https://docs.oracle.com/javase/10/docs/api/java/text/SimpleDateFormat.html) pattern.
When the year portion of the pattern is `yyyy` (full year) or `yy` (2 digits), everything works. However, when it is only `y`, it won't work (see below).
This is because the [SimpleDateFormat](https://docs.oracle.com/javase/10/docs/api/java/text/SimpleDateFormat.html) interprets `y` as:
> For formatting, if the number of pattern letters is 2, the year is truncated to 2 digits; **otherwise it is interpreted as a number**.
While the [JQueryUI datepicker](https://api.jqueryui.com/datepicker/#utility-formatDate) (and thus p:calendar and p:datepicker which uses the same algorithm) interprets `y` as
> y - year (two digit)
So the Date gets formatted as `20 May 2020` by the server, but the client expects `20 May 20`.
Now I suppose most people would want the full year anyways, but how I even found out about this issue is because I had been using the default pattern from SimpleDataFormat.LONG. Observe how the following program gives different outputs for different Java versions:
```java
import java.text.SimpleDateFormat;
import java.util.Locale;
public class Test {
public static void main(String[] args) {
String s =((SimpleDateFormat)SimpleDateFormat.getDateTimeInstance(SimpleDateFormat.LONG, SimpleDateFormat.SHORT, Locale.GERMAN)).toPattern();
System.out.println(s);
}
}
```
With Java 14:
```bash
user@mypc:/tmp/javatest$ java -version
openjdk version "14.0.1" 2020-04-14
OpenJDK Runtime Environment (build 14.0.1+7-Ubuntu-1ubuntu1)
OpenJDK 64-Bit Server VM (build 14.0.1+7-Ubuntu-1ubuntu1, mixed mode, sharing)
user@mypc:/tmp/javatest$ javac Test.java && java Test
d. MMMM y, HH:mm
user@mypc:/tmp/javatest$ javac Test.java && java -Djava.locale.providers=COMPAT,CLDR Test
d. MMMM yyyy HH:mm
```
With Java 8:
```bash
user@mypc:/tmp/javatest$ java -version
openjdk version "1.8.0_265"
OpenJDK Runtime Environment (build 1.8.0_265-8u265-b01-0ubuntu2~20.04-b01)
OpenJDK 64-Bit Server VM (build 25.265-b01, mixed mode)
user@mypc:/tmp/javatest$ javac Test.java && java Test
d. MMMM yyyy HH:mm
```
So while I would not call this terribly important, would still be nice if this would work.
**Reproducer**
[reproducer.zip](https://github.com/primefaces/primefaces/files/5141453/reproducer.zip)
**Environment:**
* PF Version: _9.0-SNAPSHOT_
**To Reproduce**
Click in the date input field. A date picker should show up, but instead an error message is logged in the console.
```xhtml
<p:datePicker id="timeDe" value="#{testView.testDate}" showTime="false" pattern="dd.MM.y"/>
```

| 1.0 | Calendar/DatePicker: Error with SimpleDatePattern with 1-letter year (eg. dd.MM.y) - **Describe the defect**
So unless I understand the docs wrongly, the `pattern` option on <p:calendar> and <p:datePicker> allows for a [SimpleDateFormat](https://docs.oracle.com/javase/10/docs/api/java/text/SimpleDateFormat.html) pattern.
When the year portion of the pattern is `yyyy` (full year) or `yy` (2 digits), everything works. However, when it is only `y`, it won't work (see below).
This is because the [SimpleDateFormat](https://docs.oracle.com/javase/10/docs/api/java/text/SimpleDateFormat.html) interprets `y` as:
> For formatting, if the number of pattern letters is 2, the year is truncated to 2 digits; **otherwise it is interpreted as a number**.
While the [JQueryUI datepicker](https://api.jqueryui.com/datepicker/#utility-formatDate) (and thus p:calendar and p:datepicker which uses the same algorithm) interprets `y` as
> y - year (two digit)
So the Date gets formatted as `20 May 2020` by the server, but the client expects `20 May 20`.
Now I suppose most people would want the full year anyways, but how I even found out about this issue is because I had been using the default pattern from SimpleDataFormat.LONG. Observe how the following program gives different outputs for different Java versions:
```java
import java.text.SimpleDateFormat;
import java.util.Locale;
public class Test {
public static void main(String[] args) {
String s =((SimpleDateFormat)SimpleDateFormat.getDateTimeInstance(SimpleDateFormat.LONG, SimpleDateFormat.SHORT, Locale.GERMAN)).toPattern();
System.out.println(s);
}
}
```
With Java 14:
```bash
user@mypc:/tmp/javatest$ java -version
openjdk version "14.0.1" 2020-04-14
OpenJDK Runtime Environment (build 14.0.1+7-Ubuntu-1ubuntu1)
OpenJDK 64-Bit Server VM (build 14.0.1+7-Ubuntu-1ubuntu1, mixed mode, sharing)
user@mypc:/tmp/javatest$ javac Test.java && java Test
d. MMMM y, HH:mm
user@mypc:/tmp/javatest$ javac Test.java && java -Djava.locale.providers=COMPAT,CLDR Test
d. MMMM yyyy HH:mm
```
With Java 8:
```bash
user@mypc:/tmp/javatest$ java -version
openjdk version "1.8.0_265"
OpenJDK Runtime Environment (build 1.8.0_265-8u265-b01-0ubuntu2~20.04-b01)
OpenJDK 64-Bit Server VM (build 25.265-b01, mixed mode)
user@mypc:/tmp/javatest$ javac Test.java && java Test
d. MMMM yyyy HH:mm
```
So while I would not call this terribly important, would still be nice if this would work.
**Reproducer**
[reproducer.zip](https://github.com/primefaces/primefaces/files/5141453/reproducer.zip)
**Environment:**
* PF Version: _9.0-SNAPSHOT_
**To Reproduce**
Click in the date input field. A date picker should show up, but instead an error message is logged in the console.
```xhtml
<p:datePicker id="timeDe" value="#{testView.testDate}" showTime="false" pattern="dd.MM.y"/>
```

| defect | calendar datepicker error with simpledatepattern with letter year eg dd mm y describe the defect so unless i understand the docs wrongly the pattern option on and allows for a pattern when the year portion of the pattern is yyyy full year or yy digits everything works however when it is only y it won t work see below this is because the interprets y as for formatting if the number of pattern letters is the year is truncated to digits otherwise it is interpreted as a number while the and thus p calendar and p datepicker which uses the same algorithm interprets y as y year two digit so the date gets formatted as may by the server but the client expects may now i suppose most people would want the full year anyways but how i even found out about this issue is because i had been using the default pattern from simpledataformat long observe how the following program gives different outputs for different java versions java import java text simpledateformat import java util locale public class test public static void main string args string s simpledateformat simpledateformat getdatetimeinstance simpledateformat long simpledateformat short locale german topattern system out println s with java bash user mypc tmp javatest java version openjdk version openjdk runtime environment build ubuntu openjdk bit server vm build ubuntu mixed mode sharing user mypc tmp javatest javac test java java test d mmmm y hh mm user mypc tmp javatest javac test java java djava locale providers compat cldr test d mmmm yyyy hh mm with java bash user mypc tmp javatest java version openjdk version openjdk runtime environment build openjdk bit server vm build mixed mode user mypc tmp javatest javac test java java test d mmmm yyyy hh mm so while i would not call this terribly important would still be nice if this would work reproducer environment pf version snapshot to reproduce click in the date input field a date picker should show up but instead an error message is logged in the console xhtml | 1 |
47,079 | 13,056,028,368 | IssuesEvent | 2020-07-30 03:26:12 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Set default value of 'needs_wavedeform_spe_corr' in OnlineFilter to True (Trac #2400) | Incomplete Migration Migrated from Trac combo core defect | Migrated from https://code.icecube.wisc.edu/ticket/2400
```json
{
"status": "closed",
"changetime": "2020-01-13T21:09:13",
"description": "This should always be true as of PASS2 unless specifically indicated otherwise. \n\nApply to combo/trunk, combo/snobo.",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1578949753662650",
"component": "combo core",
"summary": "Set default value of 'needs_wavedeform_spe_corr' in OnlineFilter to True",
"priority": "major",
"keywords": "",
"time": "2020-01-13T20:59:56",
"milestone": "Winter Solstice 2020",
"owner": "juancarlos",
"type": "defect"
}
```
| 1.0 | Set default value of 'needs_wavedeform_spe_corr' in OnlineFilter to True (Trac #2400) - Migrated from https://code.icecube.wisc.edu/ticket/2400
```json
{
"status": "closed",
"changetime": "2020-01-13T21:09:13",
"description": "This should always be true as of PASS2 unless specifically indicated otherwise. \n\nApply to combo/trunk, combo/snobo.",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1578949753662650",
"component": "combo core",
"summary": "Set default value of 'needs_wavedeform_spe_corr' in OnlineFilter to True",
"priority": "major",
"keywords": "",
"time": "2020-01-13T20:59:56",
"milestone": "Winter Solstice 2020",
"owner": "juancarlos",
"type": "defect"
}
```
| defect | set default value of needs wavedeform spe corr in onlinefilter to true trac migrated from json status closed changetime description this should always be true as of unless specifically indicated otherwise n napply to combo trunk combo snobo reporter juancarlos cc resolution fixed ts component combo core summary set default value of needs wavedeform spe corr in onlinefilter to true priority major keywords time milestone winter solstice owner juancarlos type defect | 1 |
723,488 | 24,898,513,828 | IssuesEvent | 2022-10-28 18:14:32 | OffchainLabs/arb-token-bridge | https://api.github.com/repos/OffchainLabs/arb-token-bridge | closed | better handling of insufficient l1 or l2 gas | Type: UX confusion Priority: High | I.e., display total eth cost (estimate) somewhere in UI before depositing / withdrawing and disable deposit if insufficient; on failure, surface an error to user in the UI. | 1.0 | better handling of insufficient l1 or l2 gas - I.e., display total eth cost (estimate) somewhere in UI before depositing / withdrawing and disable deposit if insufficient; on failure, surface an error to user in the UI. | non_defect | better handling of insufficient or gas i e display total eth cost estimate somewhere in ui before depositing withdrawing and disable deposit if insufficient on failure surface an error to user in the ui | 0 |
45,783 | 13,055,748,717 | IssuesEvent | 2020-07-30 02:37:15 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | glshovel DOMLaunchSeriesMap crash on warehouse data (Trac #50) | Incomplete Migration Migrated from Trac defect glshovel | Migrated from https://code.icecube.wisc.edu/ticket/50
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"description": "\nHi,\n\nI'm trying to look at simulated runs with GLShovel.\nI tried with a file I downloaded from the Data Warehouse :\n\"nugen_nue_twr.000445.000485.i3\".\n\nAfter openning the file, I select an event, and then try to click on\n\"DOMLaunchSeriesMap\".\nThe program crash and I get the execution trace you can see below.\n\nDo you experience the same problem ?\n\n\nCheers,\nJB\n\n\n\n\nDate/Time: 2007-06-06 14:58:07.230 +0200\nOS Version: 10.4.9 (Build 8P2137)\nReport Version: 4\n\nCommand: glshovel-ui\nPath: /Applications/GLShovel.app/Contents/MacOS/glshovel-ui\nParent: WindowServer [60]\n\nVersion: 1.5.0.9 (0.0.1)\n\nPID: 2735\nThread: 0\n\nException: EXC_BAD_ACCESS (0x0001)\nCodes: KERN_PROTECTION_FAILURE (0x0002) at 0x0000000c\n\nThread 0 Crashed:\n0 libglshovel.dylib 0x0025bbfb std::_Rb_tree<OMKey,\nstd::pair<OMKey const, I3OMGeo>, std::_Select1st<std::pair<OMKey\nconst, I3OMGeo> >, std::less<OMKey>, std::allocator<std::pair<OMKey\nconst, I3OMGeo> > >::find(OMKey const&) const + 11\n1 libglshovel.dylib 0x00228619\nrender::DOMLaunchSeriesMap::render(unsigned long, double, double,\nboost::shared_ptr<I3Geometry const>, boost::shared_ptr<I3Calibration\nconst>, boost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3Map<OMKey, std::vector<I3DOMLaunch,\nstd::allocator<I3DOMLaunch> > > const>) + 429\n2 libglshovel.dylib 0x0025f5ee\nrender_dispatch<I3Map<OMKey, std::vector<I3DOMLaunch,\nstd::allocator<I3DOMLaunch> > >, render::DOMLaunchSeriesMap>::dispatch\n(unsigned long, double, double, boost::shared_ptr<I3Geometry const>,\nboost::shared_ptr<I3Calibration const>,\nboost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3FrameObject const>) + 258\n3 libglshovel.dylib 0x00220775 Renderer::render\n(unsigned long, double, double, boost::shared_ptr<I3Geometry const>,\nboost::shared_ptr<I3Calibration const>,\nboost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3FrameObject const>) + 295\n4 libglshovel.dylib 0x002122be GLDetectorView::redraw\n(std::basic_string<char, std::char_traits<char>, std::allocator<char>\n> const&, std::basic_string<char, std::char_traits<char>,\nstd::allocator<char> > const&) + 836\n5 libglshovel.dylib 0x0021ac58\nGLShovelMainWindowImpl::treeItemChanged(QTreeWidgetItem*, int) + 1122\n6 libglshovel.dylib 0x0021f663\nGLShovelMainWindowImpl::qt_metacall(QMetaObject::Call, int, void**) +\n373\n7 QtCore 0x005d3a6a QMetaObject::activate\n(QObject*, int, int, void**) + 1228\n8 QtGui 0x0130365c\nQTreeWidgetPrivate::_q_emitItemChanged(QModelIndex const&) + 96\n9 QtGui 0x01303bdc\nQTreeWidget::qt_metacall(QMetaObject::Call, int, void**) + 984\n10 QtCore 0x005d3a6a QMetaObject::activate\n(QObject*, int, int, void**) + 1228\n",
"reporter": "troy",
"cc": "julien.bolmont@ifh.de",
"resolution": "duplicate",
"_ts": "1194753078000000",
"component": "glshovel",
"summary": "glshovel DOMLaunchSeriesMap crash on warehouse data",
"priority": "normal",
"keywords": "",
"time": "2007-06-06T13:31:51",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | glshovel DOMLaunchSeriesMap crash on warehouse data (Trac #50) - Migrated from https://code.icecube.wisc.edu/ticket/50
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"description": "\nHi,\n\nI'm trying to look at simulated runs with GLShovel.\nI tried with a file I downloaded from the Data Warehouse :\n\"nugen_nue_twr.000445.000485.i3\".\n\nAfter openning the file, I select an event, and then try to click on\n\"DOMLaunchSeriesMap\".\nThe program crash and I get the execution trace you can see below.\n\nDo you experience the same problem ?\n\n\nCheers,\nJB\n\n\n\n\nDate/Time: 2007-06-06 14:58:07.230 +0200\nOS Version: 10.4.9 (Build 8P2137)\nReport Version: 4\n\nCommand: glshovel-ui\nPath: /Applications/GLShovel.app/Contents/MacOS/glshovel-ui\nParent: WindowServer [60]\n\nVersion: 1.5.0.9 (0.0.1)\n\nPID: 2735\nThread: 0\n\nException: EXC_BAD_ACCESS (0x0001)\nCodes: KERN_PROTECTION_FAILURE (0x0002) at 0x0000000c\n\nThread 0 Crashed:\n0 libglshovel.dylib 0x0025bbfb std::_Rb_tree<OMKey,\nstd::pair<OMKey const, I3OMGeo>, std::_Select1st<std::pair<OMKey\nconst, I3OMGeo> >, std::less<OMKey>, std::allocator<std::pair<OMKey\nconst, I3OMGeo> > >::find(OMKey const&) const + 11\n1 libglshovel.dylib 0x00228619\nrender::DOMLaunchSeriesMap::render(unsigned long, double, double,\nboost::shared_ptr<I3Geometry const>, boost::shared_ptr<I3Calibration\nconst>, boost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3Map<OMKey, std::vector<I3DOMLaunch,\nstd::allocator<I3DOMLaunch> > > const>) + 429\n2 libglshovel.dylib 0x0025f5ee\nrender_dispatch<I3Map<OMKey, std::vector<I3DOMLaunch,\nstd::allocator<I3DOMLaunch> > >, render::DOMLaunchSeriesMap>::dispatch\n(unsigned long, double, double, boost::shared_ptr<I3Geometry const>,\nboost::shared_ptr<I3Calibration const>,\nboost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3FrameObject const>) + 258\n3 libglshovel.dylib 0x00220775 Renderer::render\n(unsigned long, double, double, boost::shared_ptr<I3Geometry const>,\nboost::shared_ptr<I3Calibration const>,\nboost::shared_ptr<I3DetectorStatus const>,\nboost::shared_ptr<I3FrameObject const>) + 295\n4 libglshovel.dylib 0x002122be GLDetectorView::redraw\n(std::basic_string<char, std::char_traits<char>, std::allocator<char>\n> const&, std::basic_string<char, std::char_traits<char>,\nstd::allocator<char> > const&) + 836\n5 libglshovel.dylib 0x0021ac58\nGLShovelMainWindowImpl::treeItemChanged(QTreeWidgetItem*, int) + 1122\n6 libglshovel.dylib 0x0021f663\nGLShovelMainWindowImpl::qt_metacall(QMetaObject::Call, int, void**) +\n373\n7 QtCore 0x005d3a6a QMetaObject::activate\n(QObject*, int, int, void**) + 1228\n8 QtGui 0x0130365c\nQTreeWidgetPrivate::_q_emitItemChanged(QModelIndex const&) + 96\n9 QtGui 0x01303bdc\nQTreeWidget::qt_metacall(QMetaObject::Call, int, void**) + 984\n10 QtCore 0x005d3a6a QMetaObject::activate\n(QObject*, int, int, void**) + 1228\n",
"reporter": "troy",
"cc": "julien.bolmont@ifh.de",
"resolution": "duplicate",
"_ts": "1194753078000000",
"component": "glshovel",
"summary": "glshovel DOMLaunchSeriesMap crash on warehouse data",
"priority": "normal",
"keywords": "",
"time": "2007-06-06T13:31:51",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| defect | glshovel domlaunchseriesmap crash on warehouse data trac migrated from json status closed changetime description nhi n ni m trying to look at simulated runs with glshovel ni tried with a file i downloaded from the data warehouse n nugen nue twr n nafter openning the file i select an event and then try to click on n domlaunchseriesmap nthe program crash and i get the execution trace you can see below n ndo you experience the same problem n n ncheers njb n n n n ndate time nos version build nreport version n ncommand glshovel ui npath applications glshovel app contents macos glshovel ui nparent windowserver n nversion n npid nthread n nexception exc bad access ncodes kern protection failure at n nthread crashed libglshovel dylib std rb tree std std less std allocator find omkey const const libglshovel dylib nrender domlaunchseriesmap render unsigned long double double nboost shared ptr boost shared ptr boost shared ptr nboost shared ptr const libglshovel dylib nrender dispatch render domlaunchseriesmap dispatch n unsigned long double double boost shared ptr nboost shared ptr nboost shared ptr nboost shared ptr libglshovel dylib renderer render n unsigned long double double boost shared ptr nboost shared ptr nboost shared ptr nboost shared ptr libglshovel dylib gldetectorview redraw n std basic string std allocator n const std basic string nstd allocator const libglshovel dylib nglshovelmainwindowimpl treeitemchanged qtreewidgetitem int libglshovel dylib nglshovelmainwindowimpl qt metacall qmetaobject call int void qtcore qmetaobject activate n qobject int int void qtgui nqtreewidgetprivate q emititemchanged qmodelindex const qtgui nqtreewidget qt metacall qmetaobject call int void qtcore qmetaobject activate n qobject int int void n reporter troy cc julien bolmont ifh de resolution duplicate ts component glshovel summary glshovel domlaunchseriesmap crash on warehouse data priority normal keywords time milestone owner troy type defect | 1 |
3,737 | 2,610,067,945 | IssuesEvent | 2015-02-26 18:19:57 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 路桥看前列腺炎去哪里最好 | auto-migrated Priority-Medium Type-Defect | ```
路桥看前列腺炎去哪里最好【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:57 | 1.0 | 路桥看前列腺炎去哪里最好 - ```
路桥看前列腺炎去哪里最好【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:57 | defect | 路桥看前列腺炎去哪里最好 路桥看前列腺炎去哪里最好【台州五洲生殖医院】 咨询热线 微信号tzwzszyy 医院地址 台 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
58,615 | 24,507,178,637 | IssuesEvent | 2022-10-10 17:29:40 | pingidentity/terraform-provider-pingone | https://api.github.com/repos/pingidentity/terraform-provider-pingone | opened | `pingone_licenses` panic when using `data_filter` | bug service base | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
Thank you for opening an issue. Please note that we try to keep the Terraform issue tracker reserved for bug reports and feature requests. For general usage questions, please see: https://www.terraform.io/community.html.
### PingOne Terraform provider Version
<!--- Check the version you have configured in your .tf files. If you are not running the latest version of the provider, please upgrade because your issue may have already been fixed. -->
v0.6.0
### Terraform Version
<!--- Run `terraform -v` to show the version. If you are not running the latest version of Terraform, please upgrade because your issue may have already been fixed. -->
N/a
### Affected Resource(s)
<!--- Please list the resources as a list, for example: -->
- pingone_licenses
<!--- If this issue appears to affect multiple resources, it may be an issue with Terraform's core, so please mention this. -->
### Terraform Configuration Files
```hcl
# Copy-paste your PingOne related Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
# Remember to replace any account/customer sensitive information in the configuration before submitting the issue
data "pingone_licenses" "license" {
organization_id = data.pingone_environment.admins_org.organization_id
data_filter {
name = "package"
values = ["INTERNAL"]
}
data_filter {
name = "status"
values = ["ACTIVE"]
}
}
```
### Debug Output
<!--- Please provide your debug output with `TF_LOG=DEBUG` enabled on your `terraform plan` or `terraform apply` -->
### Panic Output
<!--- If Terraform produced a panic, please provide your debug output from the GO panic -->
```
Stack trace from the terraform-provider-pingone_v0.6.0 plugin:
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x1979609]
goroutine 22 [running]:
github.com/pingidentity/terraform-provider-pingone/internal/sdk.ParseResponseWithCustomTimeout({0x1de04c0, 0xc0000f54a0}, 0x1c1a880?, {0x1c682cc, 0xf}, 0x28?, 0x1bb5000?, 0xc00041a201?)
github.com/pingidentity/terraform-provider-pingone/internal/sdk/sdk.go:88 +0x389
github.com/pingidentity/terraform-provider-pingone/internal/sdk.ParseResponse({0x1de04c0?, 0xc0000f54a0?}, 0x1af6c20?, {0x1c682cc?, 0x1b1e0a0?}, 0xc0000f5470?, 0x1273638?)
github.com/pingidentity/terraform-provider-pingone/internal/sdk/sdk.go:56 +0x32
github.com/pingidentity/terraform-provider-pingone/internal/service/base.datasourcePingOneLicensesRead({0x1de0488, 0xc0000965a0}, 0xc0000e4380, {0x1a88a60?, 0xc000795c30})
github.com/pingidentity/terraform-provider-pingone/internal/service/base/data_source_licenses.go:99 +0x35b
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).read(0xc000255b20, {0x1de04c0, 0xc0002cb920}, 0xd?, {0x1a88a60, 0xc000795c30})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/resource.go:724 +0x12e
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).ReadDataApply(0xc000255b20, {0x1de04c0, 0xc0002cb920}, 0xc0000e4280, {0x1a88a60, 0xc000795c30})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/resource.go:943 +0x145
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*GRPCProviderServer).ReadDataSource(0xc0003fe4f8, {0x1de04c0?, 0xc0002cb7d0?}, 0xc000150260)
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/grpc_provider.go:1179 +0x38f
github.com/hashicorp/terraform-plugin-go/tfprotov5/tf5server.(*server).ReadDataSource(0xc0003e3540, {0x1de04c0?, 0xc0002ca540?}, 0xc0000d40a0)
github.com/hashicorp/terraform-plugin-go@v0.14.0/tfprotov5/tf5server/server.go:658 +0x3ef
github.com/hashicorp/terraform-plugin-go/tfprotov5/internal/tfplugin5._Provider_ReadDataSource_Handler({0x1c1a720?, 0xc0003e3540}, {0x1de04c0, 0xc0002ca540}, 0xc0000e0000, 0x0)
github.com/hashicorp/terraform-plugin-go@v0.14.0/tfprotov5/internal/tfplugin5/tfplugin5_grpc.pb.go:421 +0x170
google.golang.org/grpc.(*Server).processUnaryRPC(0xc0001e61e0, {0x1de3d00, 0xc00048a1a0}, 0xc0000da000, 0xc000431b30, 0x24870f0, 0x0)
google.golang.org/grpc@v1.49.0/server.go:1301 +0xb0b
google.golang.org/grpc.(*Server).handleStream(0xc0001e61e0, {0x1de3d00, 0xc00048a1a0}, 0xc0000da000, 0x0)
google.golang.org/grpc@v1.49.0/server.go:1642 +0xa1b
google.golang.org/grpc.(*Server).serveStreams.func1.2()
google.golang.org/grpc@v1.49.0/server.go:938 +0x98
created by google.golang.org/grpc.(*Server).serveStreams.func1
google.golang.org/grpc@v1.49.0/server.go:936 +0x28a
Error: The terraform-provider-pingone_v0.6.0 plugin crashed!
```
### Expected Behavior
<!--- What should have happened? -->
Friendly error returned
### Actual Behavior
<!--- What actually happened? -->
Panic crash
### Steps to Reproduce
<!---Please list the steps required to reproduce the issue, for example: -->
1. `terraform apply`
### Important Factoids
<!--- Are there anything you'd like to share about the general setup of your PingOne account? Please do not include sensitive information or account data -->
### References
<!--- Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here? For example: -->
| 1.0 | `pingone_licenses` panic when using `data_filter` - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
Thank you for opening an issue. Please note that we try to keep the Terraform issue tracker reserved for bug reports and feature requests. For general usage questions, please see: https://www.terraform.io/community.html.
### PingOne Terraform provider Version
<!--- Check the version you have configured in your .tf files. If you are not running the latest version of the provider, please upgrade because your issue may have already been fixed. -->
v0.6.0
### Terraform Version
<!--- Run `terraform -v` to show the version. If you are not running the latest version of Terraform, please upgrade because your issue may have already been fixed. -->
N/a
### Affected Resource(s)
<!--- Please list the resources as a list, for example: -->
- pingone_licenses
<!--- If this issue appears to affect multiple resources, it may be an issue with Terraform's core, so please mention this. -->
### Terraform Configuration Files
```hcl
# Copy-paste your PingOne related Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
# Remember to replace any account/customer sensitive information in the configuration before submitting the issue
data "pingone_licenses" "license" {
organization_id = data.pingone_environment.admins_org.organization_id
data_filter {
name = "package"
values = ["INTERNAL"]
}
data_filter {
name = "status"
values = ["ACTIVE"]
}
}
```
### Debug Output
<!--- Please provide your debug output with `TF_LOG=DEBUG` enabled on your `terraform plan` or `terraform apply` -->
### Panic Output
<!--- If Terraform produced a panic, please provide your debug output from the GO panic -->
```
Stack trace from the terraform-provider-pingone_v0.6.0 plugin:
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x1979609]
goroutine 22 [running]:
github.com/pingidentity/terraform-provider-pingone/internal/sdk.ParseResponseWithCustomTimeout({0x1de04c0, 0xc0000f54a0}, 0x1c1a880?, {0x1c682cc, 0xf}, 0x28?, 0x1bb5000?, 0xc00041a201?)
github.com/pingidentity/terraform-provider-pingone/internal/sdk/sdk.go:88 +0x389
github.com/pingidentity/terraform-provider-pingone/internal/sdk.ParseResponse({0x1de04c0?, 0xc0000f54a0?}, 0x1af6c20?, {0x1c682cc?, 0x1b1e0a0?}, 0xc0000f5470?, 0x1273638?)
github.com/pingidentity/terraform-provider-pingone/internal/sdk/sdk.go:56 +0x32
github.com/pingidentity/terraform-provider-pingone/internal/service/base.datasourcePingOneLicensesRead({0x1de0488, 0xc0000965a0}, 0xc0000e4380, {0x1a88a60?, 0xc000795c30})
github.com/pingidentity/terraform-provider-pingone/internal/service/base/data_source_licenses.go:99 +0x35b
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).read(0xc000255b20, {0x1de04c0, 0xc0002cb920}, 0xd?, {0x1a88a60, 0xc000795c30})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/resource.go:724 +0x12e
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).ReadDataApply(0xc000255b20, {0x1de04c0, 0xc0002cb920}, 0xc0000e4280, {0x1a88a60, 0xc000795c30})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/resource.go:943 +0x145
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*GRPCProviderServer).ReadDataSource(0xc0003fe4f8, {0x1de04c0?, 0xc0002cb7d0?}, 0xc000150260)
github.com/hashicorp/terraform-plugin-sdk/v2@v2.23.0/helper/schema/grpc_provider.go:1179 +0x38f
github.com/hashicorp/terraform-plugin-go/tfprotov5/tf5server.(*server).ReadDataSource(0xc0003e3540, {0x1de04c0?, 0xc0002ca540?}, 0xc0000d40a0)
github.com/hashicorp/terraform-plugin-go@v0.14.0/tfprotov5/tf5server/server.go:658 +0x3ef
github.com/hashicorp/terraform-plugin-go/tfprotov5/internal/tfplugin5._Provider_ReadDataSource_Handler({0x1c1a720?, 0xc0003e3540}, {0x1de04c0, 0xc0002ca540}, 0xc0000e0000, 0x0)
github.com/hashicorp/terraform-plugin-go@v0.14.0/tfprotov5/internal/tfplugin5/tfplugin5_grpc.pb.go:421 +0x170
google.golang.org/grpc.(*Server).processUnaryRPC(0xc0001e61e0, {0x1de3d00, 0xc00048a1a0}, 0xc0000da000, 0xc000431b30, 0x24870f0, 0x0)
google.golang.org/grpc@v1.49.0/server.go:1301 +0xb0b
google.golang.org/grpc.(*Server).handleStream(0xc0001e61e0, {0x1de3d00, 0xc00048a1a0}, 0xc0000da000, 0x0)
google.golang.org/grpc@v1.49.0/server.go:1642 +0xa1b
google.golang.org/grpc.(*Server).serveStreams.func1.2()
google.golang.org/grpc@v1.49.0/server.go:938 +0x98
created by google.golang.org/grpc.(*Server).serveStreams.func1
google.golang.org/grpc@v1.49.0/server.go:936 +0x28a
Error: The terraform-provider-pingone_v0.6.0 plugin crashed!
```
### Expected Behavior
<!--- What should have happened? -->
Friendly error returned
### Actual Behavior
<!--- What actually happened? -->
Panic crash
### Steps to Reproduce
<!---Please list the steps required to reproduce the issue, for example: -->
1. `terraform apply`
### Important Factoids
<!--- Are there anything you'd like to share about the general setup of your PingOne account? Please do not include sensitive information or account data -->
### References
<!--- Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here? For example: -->
| non_defect | pingone licenses panic when using data filter community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment thank you for opening an issue please note that we try to keep the terraform issue tracker reserved for bug reports and feature requests for general usage questions please see pingone terraform provider version terraform version n a affected resource s pingone licenses terraform configuration files hcl copy paste your pingone related terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key remember to replace any account customer sensitive information in the configuration before submitting the issue data pingone licenses license organization id data pingone environment admins org organization id data filter name package values data filter name status values debug output panic output stack trace from the terraform provider pingone plugin panic runtime error invalid memory address or nil pointer dereference goroutine github com pingidentity terraform provider pingone internal sdk parseresponsewithcustomtimeout github com pingidentity terraform provider pingone internal sdk sdk go github com pingidentity terraform provider pingone internal sdk parseresponse github com pingidentity terraform provider pingone internal sdk sdk go github com pingidentity terraform provider pingone internal service base datasourcepingonelicensesread github com pingidentity terraform provider pingone internal service base data source licenses go github com hashicorp terraform plugin sdk helper schema resource read github com hashicorp terraform plugin sdk helper schema resource go github com hashicorp terraform plugin sdk helper schema resource readdataapply github com hashicorp terraform plugin sdk helper schema resource go github com hashicorp terraform plugin sdk helper schema grpcproviderserver readdatasource github com hashicorp terraform plugin sdk helper schema grpc provider go github com hashicorp terraform plugin go server readdatasource github com hashicorp terraform plugin go server go github com hashicorp terraform plugin go internal provider readdatasource handler github com hashicorp terraform plugin go internal grpc pb go google golang org grpc server processunaryrpc google golang org grpc server go google golang org grpc server handlestream google golang org grpc server go google golang org grpc server servestreams google golang org grpc server go created by google golang org grpc server servestreams google golang org grpc server go error the terraform provider pingone plugin crashed expected behavior friendly error returned actual behavior panic crash steps to reproduce terraform apply important factoids references | 0 |
77,309 | 26,910,166,030 | IssuesEvent | 2023-02-06 22:46:14 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | [Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.3) | content design 508/Accessibility ia 508-defect-2 collab-cycle-feedback Staging CCIssue02.04 CC-Dashboard my-education-benefits education-letters | ### General Information
#### VFS team name
DGIB - My Education Benefits
#### VFS product name
VA Education Letters
#### VFS feature name
#### Point of Contact/Reviewers
Brian DeConinck - @briandeconinck - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
### Issue Details
In the "How do I download..." section, three of the four steps given are just a heading with no additional content following. Headings are typically used to indicate sections of content, and screen reader users often navigate page content by jumping to the heading of the section they want to read. Headings with no content could be confusing.
### Link, screenshot or steps to recreate
### VA.gov Experience Standard
[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
WCAG SC 2.4.6_AA
### Platform Recommendation
I think the process list component might not be a good fit for the instructions given here. The component works really well when you have high-level steps that can be described briefly with a heading, but require a paragraph or two of explanation to actually follow the steps.
Consider reworking the steps to be a little broader (eg. combine the last three steps into one step with a brief heading and details), or else just use a standard `ol` ordered list rather than the process list component.
If it's important to use a process list component with these steps specifically, you may be able to use `p` elements in place of the headings for each step. I'm not aware of any other places on VA.gov that do this, but it would avoid the heading navigation issue while preserving the visual style.
---
### VFS Guidance
- Close the ticket when the issue has been resolved or validated by your Product Owner
- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up. | 1.0 | [Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.3) - ### General Information
#### VFS team name
DGIB - My Education Benefits
#### VFS product name
VA Education Letters
#### VFS feature name
#### Point of Contact/Reviewers
Brian DeConinck - @briandeconinck - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
### Issue Details
In the "How do I download..." section, three of the four steps given are just a heading with no additional content following. Headings are typically used to indicate sections of content, and screen reader users often navigate page content by jumping to the heading of the section they want to read. Headings with no content could be confusing.
### Link, screenshot or steps to recreate
### VA.gov Experience Standard
[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
WCAG SC 2.4.6_AA
### Platform Recommendation
I think the process list component might not be a good fit for the instructions given here. The component works really well when you have high-level steps that can be described briefly with a heading, but require a paragraph or two of explanation to actually follow the steps.
Consider reworking the steps to be a little broader (eg. combine the last three steps into one step with a brief heading and details), or else just use a standard `ol` ordered list rather than the process list component.
If it's important to use a process list component with these steps specifically, you may be able to use `p` elements in place of the headings for each step. I'm not aware of any other places on VA.gov that do this, but it would avoid the heading navigation issue while preserving the visual style.
---
### VFS Guidance
- Close the ticket when the issue has been resolved or validated by your Product Owner
- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up. | defect | your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page general information vfs team name dgib my education benefits vfs product name va education letters vfs feature name point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page issue details in the how do i download section three of the four steps given are just a heading with no additional content following headings are typically used to indicate sections of content and screen reader users often navigate page content by jumping to the heading of the section they want to read headings with no content could be confusing link screenshot or steps to recreate va gov experience standard other references wcag sc a wcag sc aa platform recommendation i think the process list component might not be a good fit for the instructions given here the component works really well when you have high level steps that can be described briefly with a heading but require a paragraph or two of explanation to actually follow the steps consider reworking the steps to be a little broader eg combine the last three steps into one step with a brief heading and details or else just use a standard ol ordered list rather than the process list component if it s important to use a process list component with these steps specifically you may be able to use p elements in place of the headings for each step i m not aware of any other places on va gov that do this but it would avoid the heading navigation issue while preserving the visual style vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up | 1 |
345,857 | 24,877,355,980 | IssuesEvent | 2022-10-27 20:23:50 | solana-developers/developer-experience | https://api.github.com/repos/solana-developers/developer-experience | closed | Show adding additional fees to a transaction | documentation video byte-sized | Let's get a tutorial showcasing how to add additional fees to a transaction to get it prioritized.
Ideally a short video/tutorial | 1.0 | Show adding additional fees to a transaction - Let's get a tutorial showcasing how to add additional fees to a transaction to get it prioritized.
Ideally a short video/tutorial | non_defect | show adding additional fees to a transaction let s get a tutorial showcasing how to add additional fees to a transaction to get it prioritized ideally a short video tutorial | 0 |
245,395 | 26,549,190,438 | IssuesEvent | 2023-01-20 05:20:11 | nidhi7598/linux-3.0.35_CVE-2022-45934 | https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2022-45934 | opened | CVE-2014-0101 (High) detected in linux-stable-rtv3.8.6 | security vulnerability | ## CVE-2014-0101 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/sm_statefuns.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The sctp_sf_do_5_1D_ce function in net/sctp/sm_statefuns.c in the Linux kernel through 3.13.6 does not validate certain auth_enable and auth_capable fields before making an sctp_sf_authenticate call, which allows remote attackers to cause a denial of service (NULL pointer dereference and system crash) via an SCTP handshake with a modified INIT chunk and a crafted AUTH chunk before a COOKIE_ECHO chunk.
<p>Publish Date: 2014-03-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0101>CVE-2014-0101</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-0101">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-0101</a></p>
<p>Release Date: 2014-03-11</p>
<p>Fix Resolution: v3.14-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2014-0101 (High) detected in linux-stable-rtv3.8.6 - ## CVE-2014-0101 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2022-45934/commit/5e23b7f9d2dd0154edd54986754eecd5b5308571">5e23b7f9d2dd0154edd54986754eecd5b5308571</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/sm_statefuns.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The sctp_sf_do_5_1D_ce function in net/sctp/sm_statefuns.c in the Linux kernel through 3.13.6 does not validate certain auth_enable and auth_capable fields before making an sctp_sf_authenticate call, which allows remote attackers to cause a denial of service (NULL pointer dereference and system crash) via an SCTP handshake with a modified INIT chunk and a crafted AUTH chunk before a COOKIE_ECHO chunk.
<p>Publish Date: 2014-03-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0101>CVE-2014-0101</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-0101">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-0101</a></p>
<p>Release Date: 2014-03-11</p>
<p>Fix Resolution: v3.14-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net sctp sm statefuns c vulnerability details the sctp sf do ce function in net sctp sm statefuns c in the linux kernel through does not validate certain auth enable and auth capable fields before making an sctp sf authenticate call which allows remote attackers to cause a denial of service null pointer dereference and system crash via an sctp handshake with a modified init chunk and a crafted auth chunk before a cookie echo chunk publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
69,316 | 8,393,026,616 | IssuesEvent | 2018-10-09 19:19:02 | unfoldingWord-dev/translationCore | https://api.github.com/repos/unfoldingWord-dev/translationCore | closed | Display more languages in the Scripture Pane | Priority/Todd Status/NeedsDesign | ## User Story
As a multilingual translator, I want to compare more than three translations in the Scripture Pane so that I fully leverage my multilingual ability to gain the broadest possible understanding of the passage.
## Specs
- [ ] Show up to 10 translations
- [ ] Allow for horizontal scrolling to view the other translations
- [ ] If there are only 3 translations selected, use the full Scripture Pane real estate
- [ ] If there are more than 3 translations, narrow the scripture display boxes so that a small portion of the "off-screen" translation shows.


| 1.0 | Display more languages in the Scripture Pane - ## User Story
As a multilingual translator, I want to compare more than three translations in the Scripture Pane so that I fully leverage my multilingual ability to gain the broadest possible understanding of the passage.
## Specs
- [ ] Show up to 10 translations
- [ ] Allow for horizontal scrolling to view the other translations
- [ ] If there are only 3 translations selected, use the full Scripture Pane real estate
- [ ] If there are more than 3 translations, narrow the scripture display boxes so that a small portion of the "off-screen" translation shows.


| non_defect | display more languages in the scripture pane user story as a multilingual translator i want to compare more than three translations in the scripture pane so that i fully leverage my multilingual ability to gain the broadest possible understanding of the passage specs show up to translations allow for horizontal scrolling to view the other translations if there are only translations selected use the full scripture pane real estate if there are more than translations narrow the scripture display boxes so that a small portion of the off screen translation shows | 0 |
18,603 | 3,074,961,948 | IssuesEvent | 2015-08-20 10:41:18 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | Ошибка при сравнении ПИ с БД | NameSpace Priority-High Type-Defect | Сравниваем ПИ с базой, где этого ПИ нет. Например в ПИ есть запись TgdcAttrUserDefined, а таблицы такой нет. Тогда сравнение вылетит. | 1.0 | Ошибка при сравнении ПИ с БД - Сравниваем ПИ с базой, где этого ПИ нет. Например в ПИ есть запись TgdcAttrUserDefined, а таблицы такой нет. Тогда сравнение вылетит. | defect | ошибка при сравнении пи с бд сравниваем пи с базой где этого пи нет например в пи есть запись tgdcattruserdefined а таблицы такой нет тогда сравнение вылетит | 1 |
13,774 | 2,783,799,007 | IssuesEvent | 2015-05-07 03:42:26 | tlh24/myopen | https://api.github.com/repos/tlh24/myopen | closed | individual (hilighted) waveforms are not properly scaled | auto-migrated Priority-Medium Type-Defect | ```
Right-clicking on points in the PCA-space cluster highlights corresponding
waveform traces. However these hilighted traces are not scaled properly.
```
Original issue reported on code.google.com by `joseph.o...@gmail.com` on 9 May 2013 at 12:09 | 1.0 | individual (hilighted) waveforms are not properly scaled - ```
Right-clicking on points in the PCA-space cluster highlights corresponding
waveform traces. However these hilighted traces are not scaled properly.
```
Original issue reported on code.google.com by `joseph.o...@gmail.com` on 9 May 2013 at 12:09 | defect | individual hilighted waveforms are not properly scaled right clicking on points in the pca space cluster highlights corresponding waveform traces however these hilighted traces are not scaled properly original issue reported on code google com by joseph o gmail com on may at | 1 |
77,504 | 27,025,406,757 | IssuesEvent | 2023-02-11 14:39:43 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Joining a room in a space: "Something went wrong" | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see?
Log on as different user timblbot:matrix.org created a space (solidprojectspace:matrix.org)
2. Log on as timbl:matric.org
3. Go to the space solidprojecspace:matrix.org
4. See the list of rooms, some with "View" buttons, some with "Join" buttons
5. What do you click?
"Join" (the r"solid" room in the list)
7. More steps…
### Outcome
#### What did you expect?
To move to a a display of the room n a oned state
#### What happened instead?
<img width="983" alt="Screenshot 2023-02-11 at 14 36 40" src="https://user-images.githubusercontent.com/1254848/218263781-4642787a-b3be-45a0-895b-ac98e522f371.png">
### Operating system
Mac OS 13.1
### Application version
Element
### How did you install the app?
I forget
### Homeserver
matrix.org
### Will you send logs?
Yes | 1.0 | Joining a room in a space: "Something went wrong" - ### Steps to reproduce
1. Where are you starting? What can you see?
Log on as different user timblbot:matrix.org created a space (solidprojectspace:matrix.org)
2. Log on as timbl:matric.org
3. Go to the space solidprojecspace:matrix.org
4. See the list of rooms, some with "View" buttons, some with "Join" buttons
5. What do you click?
"Join" (the r"solid" room in the list)
7. More steps…
### Outcome
#### What did you expect?
To move to a a display of the room n a oned state
#### What happened instead?
<img width="983" alt="Screenshot 2023-02-11 at 14 36 40" src="https://user-images.githubusercontent.com/1254848/218263781-4642787a-b3be-45a0-895b-ac98e522f371.png">
### Operating system
Mac OS 13.1
### Application version
Element
### How did you install the app?
I forget
### Homeserver
matrix.org
### Will you send logs?
Yes | defect | joining a room in a space something went wrong steps to reproduce where are you starting what can you see log on as different user timblbot matrix org created a space solidprojectspace matrix org log on as timbl matric org go to the space solidprojecspace matrix org see the list of rooms some with view buttons some with join buttons what do you click join the r solid room in the list more steps… outcome what did you expect to move to a a display of the room n a oned state what happened instead img width alt screenshot at src operating system mac os application version element how did you install the app i forget homeserver matrix org will you send logs yes | 1 |
397,859 | 11,734,068,375 | IssuesEvent | 2020-03-11 08:37:57 | garden-io/garden | https://api.github.com/repos/garden-io/garden | closed | Cluster initialise failed with kaniko and external registry | bug priority:high | ## Bug
When initialising a fresh cluster with garden services and `kaniko` in cluster building and **an external docker registry (ECR),** we got an error messaging saying garden failed to communicate with the docker daemon.
We tried to get around this by changing our in cluster builder to `cluster-docker` to initialise the docker daemon. The docker daemon failed to be brought up because the proxy failed to resolve the in cluster registry.
### Current Behavior
Garden fails to initialise the cluster with this setup.
### Expected behavior
The proxy failing to talk to the registry should be ignored since we don't need the in cluster registry.
### Reproducible example
Our provider configuration looked like:
```
buildMode: kaniko
deploymentRegistry:
hostname: "${var.AWS_ACCOUNT_ID}.dkr.ecr.${var.AWS_REGION}.amazonaws.com"
```
### Workaround
We changed the buildMode to `cluster-docker` and removed the `deploymentRegistry` temporarily. Then we could add these both back in later.
### Suggested solution(s)
The failing proxy connection should be ignored/not checked, and when using kaniko, we shouldn't need a docker daemon running in the cluster. Unsure why garden failed with this.
### Your environment
0.11.5
| 1.0 | Cluster initialise failed with kaniko and external registry - ## Bug
When initialising a fresh cluster with garden services and `kaniko` in cluster building and **an external docker registry (ECR),** we got an error messaging saying garden failed to communicate with the docker daemon.
We tried to get around this by changing our in cluster builder to `cluster-docker` to initialise the docker daemon. The docker daemon failed to be brought up because the proxy failed to resolve the in cluster registry.
### Current Behavior
Garden fails to initialise the cluster with this setup.
### Expected behavior
The proxy failing to talk to the registry should be ignored since we don't need the in cluster registry.
### Reproducible example
Our provider configuration looked like:
```
buildMode: kaniko
deploymentRegistry:
hostname: "${var.AWS_ACCOUNT_ID}.dkr.ecr.${var.AWS_REGION}.amazonaws.com"
```
### Workaround
We changed the buildMode to `cluster-docker` and removed the `deploymentRegistry` temporarily. Then we could add these both back in later.
### Suggested solution(s)
The failing proxy connection should be ignored/not checked, and when using kaniko, we shouldn't need a docker daemon running in the cluster. Unsure why garden failed with this.
### Your environment
0.11.5
| non_defect | cluster initialise failed with kaniko and external registry bug when initialising a fresh cluster with garden services and kaniko in cluster building and an external docker registry ecr we got an error messaging saying garden failed to communicate with the docker daemon we tried to get around this by changing our in cluster builder to cluster docker to initialise the docker daemon the docker daemon failed to be brought up because the proxy failed to resolve the in cluster registry current behavior garden fails to initialise the cluster with this setup expected behavior the proxy failing to talk to the registry should be ignored since we don t need the in cluster registry reproducible example our provider configuration looked like buildmode kaniko deploymentregistry hostname var aws account id dkr ecr var aws region amazonaws com workaround we changed the buildmode to cluster docker and removed the deploymentregistry temporarily then we could add these both back in later suggested solution s the failing proxy connection should be ignored not checked and when using kaniko we shouldn t need a docker daemon running in the cluster unsure why garden failed with this your environment | 0 |
68,049 | 14,894,585,524 | IssuesEvent | 2021-01-21 07:48:20 | alexcloudstar/bored | https://api.github.com/repos/alexcloudstar/bored | closed | CVE-2020-11022 (Medium) detected in jquery-1.7.1.min.js | security vulnerability | ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: bored/node_modules/sockjs/examples/express-3.x/index.html</p>
<p>Path to vulnerable library: bored/node_modules/sockjs/examples/express-3.x/index.html,bored/node_modules/sockjs/examples/multiplex/index.html,bored/node_modules/sockjs/examples/echo/index.html,bored/node_modules/sockjs/examples/hapi/html/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alexcloudstar/bored/commit/8da32ae6539174960734b81ea6729d1d4e3e2bdf">8da32ae6539174960734b81ea6729d1d4e3e2bdf</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-11022 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: bored/node_modules/sockjs/examples/express-3.x/index.html</p>
<p>Path to vulnerable library: bored/node_modules/sockjs/examples/express-3.x/index.html,bored/node_modules/sockjs/examples/multiplex/index.html,bored/node_modules/sockjs/examples/echo/index.html,bored/node_modules/sockjs/examples/hapi/html/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alexcloudstar/bored/commit/8da32ae6539174960734b81ea6729d1d4e3e2bdf">8da32ae6539174960734b81ea6729d1d4e3e2bdf</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file bored node modules sockjs examples express x index html path to vulnerable library bored node modules sockjs examples express x index html bored node modules sockjs examples multiplex index html bored node modules sockjs examples echo index html bored node modules sockjs examples hapi html index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
11,993 | 9,546,251,915 | IssuesEvent | 2019-05-01 19:23:23 | terraform-providers/terraform-provider-aws | https://api.github.com/repos/terraform-providers/terraform-provider-aws | closed | Support Attaching and Detaching IAM Roles to RDS Database Instances (e.g. Oracle S3 Integration) | enhancement service/rds | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
Amazon recently announced an integration between Amazon RDS for Oracle instances and Amazon S3: https://aws.amazon.com/about-aws/whats-new/2019/02/Amazon-RDS-for-Oracle-Now-Supports-Amazon-S3-Integration/
> With Amazon S3 Integration, you can perform data ingress with Oracle Data Pump to migrate workloads into your RDS Oracle DB Instance. After exporting your data from your source instance, you upload your Data Pump files to your Amazon S3 bucket. You can then copy the files from your S3 bucket to the RDS Oracle host and finally use the DBMS_DATAPUMP package to import your data into the target RDS Oracle DB Instance.
> With Amazon S3 Integration, you can also perform data egress out of your RDS Oracle DB instance. First, back up the data locally on the RDS Oracle host. This local backup can either be an Oracle RMAN backup or Oracle Data Pump export of the database. Once these files are created locally, you can copy the files to the S3 bucket. Optionally, you can move these backups from Amazon S3 to Amazon S3 Glacier for long term storage that complements your automated backup strategy. By using this method, you can satisfy regulatory requirements for storing database backups in an off-site location and, at the same time, cost effectively support your long-term retention policies.
Details for configuring this integration can be found at: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/oracle-s3-integration.html
While we should already have the necessary support required for Option Groups, the DB Instance requires new management of IAM Role attachments. This is similar to DB Cluster attachments for IAM Roles, however DB Instance attachments also require a feature name.
### New or Affected Resource(s)
* `aws_db_instance` or potentially a new `aws_db_instance_role_attachment` resource
### Potential Terraform Configuration
```hcl
# If implemented within the resource
resource "aws_db_instance" "example" {
role {
feature_name = "..."
role_arn = "${aws_iam_role.example.arn}"
}
}
# If implemented as a separate resource
resource "aws_db_instance_role_attachment" "example" {
db_instance_identifier = "${aws_db_instance.example.id}"
feature_name = "..."
role_arn = "${aws_iam_role.example.arn}"
}
```
### References
* https://aws.amazon.com/about-aws/whats-new/2019/02/Amazon-RDS-for-Oracle-Now-Supports-Amazon-S3-Integration/
* https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/oracle-s3-integration.html
* Similar, but requires additional configuration as compared to: #382
* https://docs.aws.amazon.com/sdk-for-go/api/service/rds/#RDS.AddRoleToDBInstance
* https://docs.aws.amazon.com/sdk-for-go/api/service/rds/#RDS.RemoveRoleFromDBInstance | 1.0 | Support Attaching and Detaching IAM Roles to RDS Database Instances (e.g. Oracle S3 Integration) - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
Amazon recently announced an integration between Amazon RDS for Oracle instances and Amazon S3: https://aws.amazon.com/about-aws/whats-new/2019/02/Amazon-RDS-for-Oracle-Now-Supports-Amazon-S3-Integration/
> With Amazon S3 Integration, you can perform data ingress with Oracle Data Pump to migrate workloads into your RDS Oracle DB Instance. After exporting your data from your source instance, you upload your Data Pump files to your Amazon S3 bucket. You can then copy the files from your S3 bucket to the RDS Oracle host and finally use the DBMS_DATAPUMP package to import your data into the target RDS Oracle DB Instance.
> With Amazon S3 Integration, you can also perform data egress out of your RDS Oracle DB instance. First, back up the data locally on the RDS Oracle host. This local backup can either be an Oracle RMAN backup or Oracle Data Pump export of the database. Once these files are created locally, you can copy the files to the S3 bucket. Optionally, you can move these backups from Amazon S3 to Amazon S3 Glacier for long term storage that complements your automated backup strategy. By using this method, you can satisfy regulatory requirements for storing database backups in an off-site location and, at the same time, cost effectively support your long-term retention policies.
Details for configuring this integration can be found at: https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/oracle-s3-integration.html
While we should already have the necessary support required for Option Groups, the DB Instance requires new management of IAM Role attachments. This is similar to DB Cluster attachments for IAM Roles, however DB Instance attachments also require a feature name.
### New or Affected Resource(s)
* `aws_db_instance` or potentially a new `aws_db_instance_role_attachment` resource
### Potential Terraform Configuration
```hcl
# If implemented within the resource
resource "aws_db_instance" "example" {
role {
feature_name = "..."
role_arn = "${aws_iam_role.example.arn}"
}
}
# If implemented as a separate resource
resource "aws_db_instance_role_attachment" "example" {
db_instance_identifier = "${aws_db_instance.example.id}"
feature_name = "..."
role_arn = "${aws_iam_role.example.arn}"
}
```
### References
* https://aws.amazon.com/about-aws/whats-new/2019/02/Amazon-RDS-for-Oracle-Now-Supports-Amazon-S3-Integration/
* https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/oracle-s3-integration.html
* Similar, but requires additional configuration as compared to: #382
* https://docs.aws.amazon.com/sdk-for-go/api/service/rds/#RDS.AddRoleToDBInstance
* https://docs.aws.amazon.com/sdk-for-go/api/service/rds/#RDS.RemoveRoleFromDBInstance | non_defect | support attaching and detaching iam roles to rds database instances e g oracle integration community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description amazon recently announced an integration between amazon rds for oracle instances and amazon with amazon integration you can perform data ingress with oracle data pump to migrate workloads into your rds oracle db instance after exporting your data from your source instance you upload your data pump files to your amazon bucket you can then copy the files from your bucket to the rds oracle host and finally use the dbms datapump package to import your data into the target rds oracle db instance with amazon integration you can also perform data egress out of your rds oracle db instance first back up the data locally on the rds oracle host this local backup can either be an oracle rman backup or oracle data pump export of the database once these files are created locally you can copy the files to the bucket optionally you can move these backups from amazon to amazon glacier for long term storage that complements your automated backup strategy by using this method you can satisfy regulatory requirements for storing database backups in an off site location and at the same time cost effectively support your long term retention policies details for configuring this integration can be found at while we should already have the necessary support required for option groups the db instance requires new management of iam role attachments this is similar to db cluster attachments for iam roles however db instance attachments also require a feature name new or affected resource s aws db instance or potentially a new aws db instance role attachment resource potential terraform configuration hcl if implemented within the resource resource aws db instance example role feature name role arn aws iam role example arn if implemented as a separate resource resource aws db instance role attachment example db instance identifier aws db instance example id feature name role arn aws iam role example arn references similar but requires additional configuration as compared to | 0 |
48,075 | 19,915,516,836 | IssuesEvent | 2022-01-25 22:06:13 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Aqua - 6.5 Upgrade | aqua ops and shared services | ## Summary
Upgrade to the latest stable version of Aqua.
## Code
The code already exists as a PR in CCM
## Known Issues
Enforcers are failing and we are waiting on support from aqua.
| 1.0 | Aqua - 6.5 Upgrade - ## Summary
Upgrade to the latest stable version of Aqua.
## Code
The code already exists as a PR in CCM
## Known Issues
Enforcers are failing and we are waiting on support from aqua.
| non_defect | aqua upgrade summary upgrade to the latest stable version of aqua code the code already exists as a pr in ccm known issues enforcers are failing and we are waiting on support from aqua | 0 |
47,750 | 13,066,191,730 | IssuesEvent | 2020-07-30 21:10:52 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | DomTools examples don't run (Trac #1119) | Migrated from Trac combo reconstruction defect | DAQReTrigger.py and I3LCCleaning-DOMLaunches.py can'l load payloads-unpacker
/I3FirstPulsifierExample.py bad input to I3Reader
IC80toIC40RetriggerWithSLCFilter.py can't load libDOMcalibrator
IC80toIC60Retriggering.py can't load libtwr-decode
Migrated from https://code.icecube.wisc.edu/ticket/1119
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "DAQReTrigger.py and I3LCCleaning-DOMLaunches.py can'l load payloads-unpacker\n/I3FirstPulsifierExample.py bad input to I3Reader\nIC80toIC40RetriggerWithSLCFilter.py can't load libDOMcalibrator\nIC80toIC60Retriggering.py can't load libtwr-decode",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "DomTools examples don't run",
"priority": "normal",
"keywords": "",
"time": "2015-08-17T14:15:52",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | DomTools examples don't run (Trac #1119) - DAQReTrigger.py and I3LCCleaning-DOMLaunches.py can'l load payloads-unpacker
/I3FirstPulsifierExample.py bad input to I3Reader
IC80toIC40RetriggerWithSLCFilter.py can't load libDOMcalibrator
IC80toIC60Retriggering.py can't load libtwr-decode
Migrated from https://code.icecube.wisc.edu/ticket/1119
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "DAQReTrigger.py and I3LCCleaning-DOMLaunches.py can'l load payloads-unpacker\n/I3FirstPulsifierExample.py bad input to I3Reader\nIC80toIC40RetriggerWithSLCFilter.py can't load libDOMcalibrator\nIC80toIC60Retriggering.py can't load libtwr-decode",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "DomTools examples don't run",
"priority": "normal",
"keywords": "",
"time": "2015-08-17T14:15:52",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| defect | domtools examples don t run trac daqretrigger py and domlaunches py can l load payloads unpacker py bad input to py can t load libdomcalibrator py can t load libtwr decode migrated from json status closed changetime description daqretrigger py and domlaunches py can l load payloads unpacker n py bad input to py can t load libdomcalibrator py can t load libtwr decode reporter kjmeagher cc resolution fixed ts component combo reconstruction summary domtools examples don t run priority normal keywords time milestone owner olivas type defect | 1 |
262,852 | 27,989,448,770 | IssuesEvent | 2023-03-27 01:32:21 | Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting | https://api.github.com/repos/Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting | opened | CVE-2023-25660 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl | Mend: dependency security vulnerability | ## CVE-2023-25660 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, when the parameter `summarize` of `tf.raw_ops.Print` is zero, the new method `SummarizeArray<bool>` will reference to a nullptr, leading to a seg fault. A fix is included in TensorFlow version 2.12 and version 2.11.1.
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25660>CVE-2023-25660</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qjqc-vqcf-5qvj">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qjqc-vqcf-5qvj</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-25660 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2023-25660 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, when the parameter `summarize` of `tf.raw_ops.Print` is zero, the new method `SummarizeArray<bool>` will reference to a nullptr, leading to a seg fault. A fix is included in TensorFlow version 2.12 and version 2.11.1.
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25660>CVE-2023-25660</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qjqc-vqcf-5qvj">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qjqc-vqcf-5qvj</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x tensorflow whl vulnerable library found in base branch main vulnerability details tensorflow is an open source platform for machine learning prior to versions and when the parameter summarize of tf raw ops print is zero the new method summarizearray will reference to a nullptr leading to a seg fault a fix is included in tensorflow version and version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend | 0 |
80,855 | 30,562,370,818 | IssuesEvent | 2023-07-20 15:21:37 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: Process unexpectedly closed with status 1 | I-defect needs-triaging | ### What happened?
hi.bro:
from this code:
CAN YOU HELP ME WHAT IS ERROR AND HOW CAN I FIX IT ?
### How can we reproduce the issue?
```shell
# Import necessary modules
from selenium import webdriver
from selenium.webdriver.firefox.options import Options
from bs4 import BeautifulSoup
from datetime import datetime, timedelta
from decimal import Decimal
import csv
import sys
import time
import requests
import pandas as pd
import pytz
import os
# Set up the driver
options = webdriver.FirefoxOptions()
options.add_argument("--no-sandbox")
options.add_argument("--disable-dev-shm-usage")
options.add_argument("--disable-gpu")
options.add_argument("--disable-extensions")
options.add_argument('-headless') # run Firefox in headless mode
driver = webdriver.Firefox(options=options)
i have this ERROR:
---------------------------------------------------------------------------
WebDriverException Traceback (most recent call last)
[<ipython-input-40-04de7e1eabe5>](https://localhost:8080/#) in <cell line: 23>()
21 options.add_argument("--disable-extensions")
22 options.add_argument('-headless') # run Firefox in headless mode
---> 23 driver = webdriver.Firefox(options=options)
24 #########################################################################################
25 # URL
4 frames
[/usr/local/lib/python3.10/dist-packages/selenium/webdriver/remote/errorhandler.py](https://localhost:8080/#) in check_response(self, response)
243 alert_text = value["alert"].get("text")
244 raise exception_class(message, screen, stacktrace, alert_text) # type: ignore[call-arg] # mypy is not smart enough here
--> 245 raise exception_class(message, screen, stacktrace)
WebDriverException: Message: Process unexpectedly closed with status 1
```
### Relevant log output
```shell
# Import necessary modules
from selenium import webdriver
from selenium.webdriver.firefox.options import Options
from bs4 import BeautifulSoup
from datetime import datetime, timedelta
from decimal import Decimal
import csv
import sys
import time
import requests
import pandas as pd
import pytz
import os
# Set up the driver
options = webdriver.FirefoxOptions()
options.add_argument("--no-sandbox")
options.add_argument("--disable-dev-shm-usage")
options.add_argument("--disable-gpu")
options.add_argument("--disable-extensions")
options.add_argument('-headless') # run Firefox in headless mode
driver = webdriver.Firefox(options=options)
i have this ERROR:
---------------------------------------------------------------------------
WebDriverException Traceback (most recent call last)
[<ipython-input-40-04de7e1eabe5>](https://localhost:8080/#) in <cell line: 23>()
21 options.add_argument("--disable-extensions")
22 options.add_argument('-headless') # run Firefox in headless mode
---> 23 driver = webdriver.Firefox(options=options)
24 #########################################################################################
25 # URL
4 frames
[/usr/local/lib/python3.10/dist-packages/selenium/webdriver/remote/errorhandler.py](https://localhost:8080/#) in check_response(self, response)
243 alert_text = value["alert"].get("text")
244 raise exception_class(message, screen, stacktrace, alert_text) # type: ignore[call-arg] # mypy is not smart enough here
--> 245 raise exception_class(message, screen, stacktrace)
WebDriverException: Message: Process unexpectedly closed with status 1
```
### Operating System
colab
### Selenium version
python
### What are the browser(s) and version(s) where you see this issue?
firefox
### What are the browser driver(s) and version(s) where you see this issue?
firefox driver
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: Process unexpectedly closed with status 1 - ### What happened?
hi.bro:
from this code:
CAN YOU HELP ME WHAT IS ERROR AND HOW CAN I FIX IT ?
### How can we reproduce the issue?
```shell
# Import necessary modules
from selenium import webdriver
from selenium.webdriver.firefox.options import Options
from bs4 import BeautifulSoup
from datetime import datetime, timedelta
from decimal import Decimal
import csv
import sys
import time
import requests
import pandas as pd
import pytz
import os
# Set up the driver
options = webdriver.FirefoxOptions()
options.add_argument("--no-sandbox")
options.add_argument("--disable-dev-shm-usage")
options.add_argument("--disable-gpu")
options.add_argument("--disable-extensions")
options.add_argument('-headless') # run Firefox in headless mode
driver = webdriver.Firefox(options=options)
i have this ERROR:
---------------------------------------------------------------------------
WebDriverException Traceback (most recent call last)
[<ipython-input-40-04de7e1eabe5>](https://localhost:8080/#) in <cell line: 23>()
21 options.add_argument("--disable-extensions")
22 options.add_argument('-headless') # run Firefox in headless mode
---> 23 driver = webdriver.Firefox(options=options)
24 #########################################################################################
25 # URL
4 frames
[/usr/local/lib/python3.10/dist-packages/selenium/webdriver/remote/errorhandler.py](https://localhost:8080/#) in check_response(self, response)
243 alert_text = value["alert"].get("text")
244 raise exception_class(message, screen, stacktrace, alert_text) # type: ignore[call-arg] # mypy is not smart enough here
--> 245 raise exception_class(message, screen, stacktrace)
WebDriverException: Message: Process unexpectedly closed with status 1
```
### Relevant log output
```shell
# Import necessary modules
from selenium import webdriver
from selenium.webdriver.firefox.options import Options
from bs4 import BeautifulSoup
from datetime import datetime, timedelta
from decimal import Decimal
import csv
import sys
import time
import requests
import pandas as pd
import pytz
import os
# Set up the driver
options = webdriver.FirefoxOptions()
options.add_argument("--no-sandbox")
options.add_argument("--disable-dev-shm-usage")
options.add_argument("--disable-gpu")
options.add_argument("--disable-extensions")
options.add_argument('-headless') # run Firefox in headless mode
driver = webdriver.Firefox(options=options)
i have this ERROR:
---------------------------------------------------------------------------
WebDriverException Traceback (most recent call last)
[<ipython-input-40-04de7e1eabe5>](https://localhost:8080/#) in <cell line: 23>()
21 options.add_argument("--disable-extensions")
22 options.add_argument('-headless') # run Firefox in headless mode
---> 23 driver = webdriver.Firefox(options=options)
24 #########################################################################################
25 # URL
4 frames
[/usr/local/lib/python3.10/dist-packages/selenium/webdriver/remote/errorhandler.py](https://localhost:8080/#) in check_response(self, response)
243 alert_text = value["alert"].get("text")
244 raise exception_class(message, screen, stacktrace, alert_text) # type: ignore[call-arg] # mypy is not smart enough here
--> 245 raise exception_class(message, screen, stacktrace)
WebDriverException: Message: Process unexpectedly closed with status 1
```
### Operating System
colab
### Selenium version
python
### What are the browser(s) and version(s) where you see this issue?
firefox
### What are the browser driver(s) and version(s) where you see this issue?
firefox driver
### Are you using Selenium Grid?
_No response_ | defect | process unexpectedly closed with status what happened hi bro from this code can you help me what is error and how can i fix it how can we reproduce the issue shell import necessary modules from selenium import webdriver from selenium webdriver firefox options import options from import beautifulsoup from datetime import datetime timedelta from decimal import decimal import csv import sys import time import requests import pandas as pd import pytz import os set up the driver options webdriver firefoxoptions options add argument no sandbox options add argument disable dev shm usage options add argument disable gpu options add argument disable extensions options add argument headless run firefox in headless mode driver webdriver firefox options options i have this error webdriverexception traceback most recent call last in options add argument disable extensions options add argument headless run firefox in headless mode driver webdriver firefox options options url frames in check response self response alert text value get text raise exception class message screen stacktrace alert text type ignore mypy is not smart enough here raise exception class message screen stacktrace webdriverexception message process unexpectedly closed with status relevant log output shell import necessary modules from selenium import webdriver from selenium webdriver firefox options import options from import beautifulsoup from datetime import datetime timedelta from decimal import decimal import csv import sys import time import requests import pandas as pd import pytz import os set up the driver options webdriver firefoxoptions options add argument no sandbox options add argument disable dev shm usage options add argument disable gpu options add argument disable extensions options add argument headless run firefox in headless mode driver webdriver firefox options options i have this error webdriverexception traceback most recent call last in options add argument disable extensions options add argument headless run firefox in headless mode driver webdriver firefox options options url frames in check response self response alert text value get text raise exception class message screen stacktrace alert text type ignore mypy is not smart enough here raise exception class message screen stacktrace webdriverexception message process unexpectedly closed with status operating system colab selenium version python what are the browser s and version s where you see this issue firefox what are the browser driver s and version s where you see this issue firefox driver are you using selenium grid no response | 1 |
9,496 | 2,615,153,592 | IssuesEvent | 2015-03-01 06:31:22 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | very slow 50% in 20h or errors | auto-migrated Priority-Triage Type-Defect | ```
Reaver 1.4
backtrack 5 RC1
Alfa Awus036h monitor mode
signal -40 to -70
reaver runs on my pc and laptop
I tried it on my PC with i5 4x4ghz and my laptop Dell 1515 with c2d 2ghz
On both it's very slow.
First my laptop starts with 13sec/pin and now after 20h it's 22sec/pin.
My PC starts with 5sec/pin and then it's also 15sec/pin.
I stopped my PC but my laptop still runs with reaver turned on.
On my laptop I sometimes see that error 0x02 and 0x03 and retrying pin but it's
rare. The main issue is... it's super slow.
On the other hand my PC which as mentioned before is faster doesn't do anything
faster.
But most of the times reaver on my PC shows those errors which I found in other
posts like:
[!] WPS transaction failed (code: 0x02), re-trying last pin
...some retries
[!] WPS transaction failed (code: 0x03), re-trying last pin
..some retries
or
[!] WARNING: Receive timeout occurred
[+] Sending EAPOL START request
I tried :
reaver -i mon0 -b <mac> -vv
reaver -i mon0 -b <mac> -vv -a
reaver -i mon0 -b <mac> -vv -a --no-nacks
reaver -i mon0 -b <mac> -vv -a --no-nacks -d15
So I'm testing reaver just on my laptop.
```
Original issue reported on code.google.com by `piotrchm...@poczta.fm` on 11 Feb 2012 at 5:42 | 1.0 | very slow 50% in 20h or errors - ```
Reaver 1.4
backtrack 5 RC1
Alfa Awus036h monitor mode
signal -40 to -70
reaver runs on my pc and laptop
I tried it on my PC with i5 4x4ghz and my laptop Dell 1515 with c2d 2ghz
On both it's very slow.
First my laptop starts with 13sec/pin and now after 20h it's 22sec/pin.
My PC starts with 5sec/pin and then it's also 15sec/pin.
I stopped my PC but my laptop still runs with reaver turned on.
On my laptop I sometimes see that error 0x02 and 0x03 and retrying pin but it's
rare. The main issue is... it's super slow.
On the other hand my PC which as mentioned before is faster doesn't do anything
faster.
But most of the times reaver on my PC shows those errors which I found in other
posts like:
[!] WPS transaction failed (code: 0x02), re-trying last pin
...some retries
[!] WPS transaction failed (code: 0x03), re-trying last pin
..some retries
or
[!] WARNING: Receive timeout occurred
[+] Sending EAPOL START request
I tried :
reaver -i mon0 -b <mac> -vv
reaver -i mon0 -b <mac> -vv -a
reaver -i mon0 -b <mac> -vv -a --no-nacks
reaver -i mon0 -b <mac> -vv -a --no-nacks -d15
So I'm testing reaver just on my laptop.
```
Original issue reported on code.google.com by `piotrchm...@poczta.fm` on 11 Feb 2012 at 5:42 | defect | very slow in or errors reaver backtrack alfa monitor mode signal to reaver runs on my pc and laptop i tried it on my pc with and my laptop dell with on both it s very slow first my laptop starts with pin and now after it s pin my pc starts with pin and then it s also pin i stopped my pc but my laptop still runs with reaver turned on on my laptop i sometimes see that error and and retrying pin but it s rare the main issue is it s super slow on the other hand my pc which as mentioned before is faster doesn t do anything faster but most of the times reaver on my pc shows those errors which i found in other posts like wps transaction failed code re trying last pin some retries wps transaction failed code re trying last pin some retries or warning receive timeout occurred sending eapol start request i tried reaver i b vv reaver i b vv a reaver i b vv a no nacks reaver i b vv a no nacks so i m testing reaver just on my laptop original issue reported on code google com by piotrchm poczta fm on feb at | 1 |
359,934 | 10,682,605,886 | IssuesEvent | 2019-10-22 06:05:18 | data2health/website | https://api.github.com/repos/data2health/website | closed | add contact info | enhancement high priority (easy fix) | on the [google doc](https://docs.google.com/document/d/1rQ1_hzY4HilH2D9tPRFJvz8n--VZrxOU9gn2EjnsPDk/edit), there is a request for "contact", but there is a contact us button the bottom of the [webpage](https://ctsa.ncats.nih.gov/cd2h/about-us/)
do we want to add contact info for each individual person?
related to #24 | 1.0 | add contact info - on the [google doc](https://docs.google.com/document/d/1rQ1_hzY4HilH2D9tPRFJvz8n--VZrxOU9gn2EjnsPDk/edit), there is a request for "contact", but there is a contact us button the bottom of the [webpage](https://ctsa.ncats.nih.gov/cd2h/about-us/)
do we want to add contact info for each individual person?
related to #24 | non_defect | add contact info on the there is a request for contact but there is a contact us button the bottom of the do we want to add contact info for each individual person related to | 0 |
63,977 | 18,097,628,966 | IssuesEvent | 2021-09-22 10:50:06 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | opened | Media picker for attachments doesn't show all media | T-Defect A-Media S-Major | ### Steps to reproduce
1. Go to a room or a DM and click on +
2. Select picture
### What happened?
### What did you expect?
See all my media, starting with most recent
### What happened?
My first run experience was different from subsequent ones. I can see only one picture out of about 10 for attaching.
### Your phone model
_No response_
### Operating system version
14.6
### Application version
Gil's version for testing spaces, 22 Sept build
### Homeserver
_No response_
### Have you submitted a rageshake?
No | 1.0 | Media picker for attachments doesn't show all media - ### Steps to reproduce
1. Go to a room or a DM and click on +
2. Select picture
### What happened?
### What did you expect?
See all my media, starting with most recent
### What happened?
My first run experience was different from subsequent ones. I can see only one picture out of about 10 for attaching.
### Your phone model
_No response_
### Operating system version
14.6
### Application version
Gil's version for testing spaces, 22 Sept build
### Homeserver
_No response_
### Have you submitted a rageshake?
No | defect | media picker for attachments doesn t show all media steps to reproduce go to a room or a dm and click on select picture what happened what did you expect see all my media starting with most recent what happened my first run experience was different from subsequent ones i can see only one picture out of about for attaching your phone model no response operating system version application version gil s version for testing spaces sept build homeserver no response have you submitted a rageshake no | 1 |
24,904 | 4,128,497,339 | IssuesEvent | 2016-06-10 06:44:40 | jsr107/jsr107tck | https://api.github.com/repos/jsr107/jsr107tck | closed | TCK Exception Assertions wrong for CacheManager.getCache(String, Class<K>, Class<V>) | Defect TCK Challenge | ### TCK Challenge:
The TCK asserts a java.lang.ClassCastException in tests listed below but pg 23 of the spec and the javadoc for CacheManager.getCache(String, Class<K>, Class<V>) state that IllegalArgumentException is thrown if the specified key and/or value types are incompatible with the configured cache.
### From TCK report output:
Tests in error:
getIncorrectCacheValueType(org.jsr107.tck.CacheManagerTest): Unexpected exception, expected\<java.lang.ClassCastException\> but was\<java.lang.IllegalArgumentException\>
getIncorrectCacheType(org.jsr107.tck.CacheManagerTest): Unexpected exception, expected\<java.lang.ClassCastException\> but was\<java.lang.IllegalArgumentException\>
### Proposed Resolution:
Update the CacheManagerTest.java as follows:
> << @Test(expected = ClassCastException.class)
> \>\> @Test(expected = IllegalArgumentException.class)
> public void getIncorrectCacheType() {
>
> << @Test(expected = ClassCastException.class)
> \>\> @Test(expected = IllegalArgumentException.class)
> public void getIncorrectCacheValueType() { | 1.0 | TCK Exception Assertions wrong for CacheManager.getCache(String, Class<K>, Class<V>) - ### TCK Challenge:
The TCK asserts a java.lang.ClassCastException in tests listed below but pg 23 of the spec and the javadoc for CacheManager.getCache(String, Class<K>, Class<V>) state that IllegalArgumentException is thrown if the specified key and/or value types are incompatible with the configured cache.
### From TCK report output:
Tests in error:
getIncorrectCacheValueType(org.jsr107.tck.CacheManagerTest): Unexpected exception, expected\<java.lang.ClassCastException\> but was\<java.lang.IllegalArgumentException\>
getIncorrectCacheType(org.jsr107.tck.CacheManagerTest): Unexpected exception, expected\<java.lang.ClassCastException\> but was\<java.lang.IllegalArgumentException\>
### Proposed Resolution:
Update the CacheManagerTest.java as follows:
> << @Test(expected = ClassCastException.class)
> \>\> @Test(expected = IllegalArgumentException.class)
> public void getIncorrectCacheType() {
>
> << @Test(expected = ClassCastException.class)
> \>\> @Test(expected = IllegalArgumentException.class)
> public void getIncorrectCacheValueType() { | defect | tck exception assertions wrong for cachemanager getcache string class class tck challenge the tck asserts a java lang classcastexception in tests listed below but pg of the spec and the javadoc for cachemanager getcache string class class state that illegalargumentexception is thrown if the specified key and or value types are incompatible with the configured cache from tck report output tests in error getincorrectcachevaluetype org tck cachemanagertest unexpected exception expected but was getincorrectcachetype org tck cachemanagertest unexpected exception expected but was proposed resolution update the cachemanagertest java as follows test expected classcastexception class test expected illegalargumentexception class public void getincorrectcachetype test expected classcastexception class test expected illegalargumentexception class public void getincorrectcachevaluetype | 1 |
179,076 | 30,113,261,714 | IssuesEvent | 2023-06-30 09:25:51 | rust-lang/www.rust-lang.org | https://api.github.com/repos/rust-lang/www.rust-lang.org | closed | Not properly aligned text | C-Bug A-Design | ### Page(s) Affected
learn page
### What needs to be fixed?
Text position
### Suggested Improvement
Align to center

| 1.0 | Not properly aligned text - ### Page(s) Affected
learn page
### What needs to be fixed?
Text position
### Suggested Improvement
Align to center

| non_defect | not properly aligned text page s affected learn page what needs to be fixed text position suggested improvement align to center | 0 |
18,783 | 3,087,128,140 | IssuesEvent | 2015-08-25 09:34:51 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | closed | [Command Fault] - IF_CONTROLS | Priority-Medium Status-CantReproduce Type-Defect | Originally reported on Google Code with ID 426
```
What steps will reproduce the problem?
1. Start a level with these properties:
. An imprisoned enemy creature (i.e. Player 1 Spider)
. An IF_CONTROLS(PLAYER1,SPIDER == 1) command
. An observable trigger when the IF statement is true
. Prison surrounded by water/lava
2. Slap the Spider slightly outside the border of the Prison.
3. The Spider will now be out of Player control, and presumably in PLAYER1 control,
as the Spider fires at the Player's creatures.
What is the expected output? What do you see instead?
It is expected that the Spider be registered as under control of PLAYER1, and the command
triggered, but this does not happen. The command is broken for the rest of the game.
What version of the product are you using? On what operating system?
Windows 7 64-bit
KeeperFX 0.4.5 r1467
Please provide any additional information below.
In the campaign I'm creating, the player is to prevent the Lord of the Land from escaping
his Prison or dying. I have everything working, but the abuse of this command fault
can cause the player to avoid the objective and break the level entirely.
```
Reported by `KeeplarBooger` on 2014-11-22 02:54:10
| 1.0 | [Command Fault] - IF_CONTROLS - Originally reported on Google Code with ID 426
```
What steps will reproduce the problem?
1. Start a level with these properties:
. An imprisoned enemy creature (i.e. Player 1 Spider)
. An IF_CONTROLS(PLAYER1,SPIDER == 1) command
. An observable trigger when the IF statement is true
. Prison surrounded by water/lava
2. Slap the Spider slightly outside the border of the Prison.
3. The Spider will now be out of Player control, and presumably in PLAYER1 control,
as the Spider fires at the Player's creatures.
What is the expected output? What do you see instead?
It is expected that the Spider be registered as under control of PLAYER1, and the command
triggered, but this does not happen. The command is broken for the rest of the game.
What version of the product are you using? On what operating system?
Windows 7 64-bit
KeeperFX 0.4.5 r1467
Please provide any additional information below.
In the campaign I'm creating, the player is to prevent the Lord of the Land from escaping
his Prison or dying. I have everything working, but the abuse of this command fault
can cause the player to avoid the objective and break the level entirely.
```
Reported by `KeeplarBooger` on 2014-11-22 02:54:10
| defect | if controls originally reported on google code with id what steps will reproduce the problem start a level with these properties an imprisoned enemy creature i e player spider an if controls spider command an observable trigger when the if statement is true prison surrounded by water lava slap the spider slightly outside the border of the prison the spider will now be out of player control and presumably in control as the spider fires at the player s creatures what is the expected output what do you see instead it is expected that the spider be registered as under control of and the command triggered but this does not happen the command is broken for the rest of the game what version of the product are you using on what operating system windows bit keeperfx please provide any additional information below in the campaign i m creating the player is to prevent the lord of the land from escaping his prison or dying i have everything working but the abuse of this command fault can cause the player to avoid the objective and break the level entirely reported by keeplarbooger on | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.