Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
143,366 | 11,545,408,089 | IssuesEvent | 2020-02-18 13:23:40 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled | Module: IMap Source: Internal Team: Core Type: Test-Failure | http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/4021/testReport/junit/com.hazelcast.map.impl.mapstore.writebehind/WriteBehindEntryStoreQueueReplicationTest/queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up/
```
java.lang.AssertionError: Expected 'expirationTime' to be between 1572011647548 and 1572011651548, but was 1572011652000
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertBetween(HazelcastTestSupport.java:1461)
at com.hazelcast.map.impl.mapstore.TestEntryStore.assertRecordStored(TestEntryStore.java:171)
at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.lambda$queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up$2(WriteBehindEntryStoreQueueReplicationTest.java:116)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1318)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1420)
at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up(WriteBehindEntryStoreQueueReplicationTest.java:114)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | 1.0 | WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled - http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/4021/testReport/junit/com.hazelcast.map.impl.mapstore.writebehind/WriteBehindEntryStoreQueueReplicationTest/queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up/
```
java.lang.AssertionError: Expected 'expirationTime' to be between 1572011647548 and 1572011651548, but was 1572011652000
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertBetween(HazelcastTestSupport.java:1461)
at com.hazelcast.map.impl.mapstore.TestEntryStore.assertRecordStored(TestEntryStore.java:171)
at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.lambda$queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up$2(WriteBehindEntryStoreQueueReplicationTest.java:116)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1318)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1420)
at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up(WriteBehindEntryStoreQueueReplicationTest.java:114)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | non_priority | writebehindentrystorequeuereplicationtest queued entries with expirationtimes are replicated when cluster scaled java lang assertionerror expected expirationtime to be between and but was at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertbetween hazelcasttestsupport java at com hazelcast map impl mapstore testentrystore assertrecordstored testentrystore java at com hazelcast map impl mapstore writebehind writebehindentrystorequeuereplicationtest lambda queued entries with expirationtimes are replicated when cluster scaled up writebehindentrystorequeuereplicationtest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast map impl mapstore writebehind writebehindentrystorequeuereplicationtest queued entries with expirationtimes are replicated when cluster scaled up writebehindentrystorequeuereplicationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java | 0 |
206,961 | 16,062,439,546 | IssuesEvent | 2021-04-23 14:18:05 | poliastro/poliastro | https://api.github.com/repos/poliastro/poliastro | closed | Review documentation classification, take 2 | documentation | Now that I work at a documentation company 🤓 I have been thinking more and more about poliastro docs, and in particular within the Diátaxis framework https://diataxis.fr/ (formerly known as "The Documentation System").

Even though we did restructure the documentation as part of #975, I would like to revisit this with some ideas:
- Add a "Getting started" category that includes the document currently called "Getting started" renamed to "Installation", the "User guide" renamed to "Quickstart guide", and the contributor documentation
- Rename "Tutorials" to "Examples" (in the spirit of How-to Guides)
- Add a new "Background" category that includes the more theoretical examples, like https://docs.poliastro.space/en/latest/examples/Comparing%20Hohmann%20and%20bielliptic%20transfers.html, https://docs.poliastro.space/en/latest/examples/Propagation%20using%20Cowells%20formulation.html, https://docs.poliastro.space/en/latest/examples/Revisiting%20Lamberts%20problem%20in%20Python.html, https://docs.poliastro.space/en/latest/examples/Studying%20Hohmann%20transfers.html, and https://docs.poliastro.space/en/latest/examples/Multirevolutions%20solution%20in%20Lamberts%20problem.html
- Add two TOCs in https://docs.poliastro.space/en/latest/autoapi/index.html, which is something that we had in older versions and I think it's worth bringing back. Not sure how easy it is to do with autoapi. | 1.0 | Review documentation classification, take 2 - Now that I work at a documentation company 🤓 I have been thinking more and more about poliastro docs, and in particular within the Diátaxis framework https://diataxis.fr/ (formerly known as "The Documentation System").

Even though we did restructure the documentation as part of #975, I would like to revisit this with some ideas:
- Add a "Getting started" category that includes the document currently called "Getting started" renamed to "Installation", the "User guide" renamed to "Quickstart guide", and the contributor documentation
- Rename "Tutorials" to "Examples" (in the spirit of How-to Guides)
- Add a new "Background" category that includes the more theoretical examples, like https://docs.poliastro.space/en/latest/examples/Comparing%20Hohmann%20and%20bielliptic%20transfers.html, https://docs.poliastro.space/en/latest/examples/Propagation%20using%20Cowells%20formulation.html, https://docs.poliastro.space/en/latest/examples/Revisiting%20Lamberts%20problem%20in%20Python.html, https://docs.poliastro.space/en/latest/examples/Studying%20Hohmann%20transfers.html, and https://docs.poliastro.space/en/latest/examples/Multirevolutions%20solution%20in%20Lamberts%20problem.html
- Add two TOCs in https://docs.poliastro.space/en/latest/autoapi/index.html, which is something that we had in older versions and I think it's worth bringing back. Not sure how easy it is to do with autoapi. | non_priority | review documentation classification take now that i work at a documentation company 🤓 i have been thinking more and more about poliastro docs and in particular within the diátaxis framework formerly known as the documentation system even though we did restructure the documentation as part of i would like to revisit this with some ideas add a getting started category that includes the document currently called getting started renamed to installation the user guide renamed to quickstart guide and the contributor documentation rename tutorials to examples in the spirit of how to guides add a new background category that includes the more theoretical examples like and add two tocs in which is something that we had in older versions and i think it s worth bringing back not sure how easy it is to do with autoapi | 0 |
553,649 | 16,376,097,025 | IssuesEvent | 2021-05-16 05:36:16 | KShewengger/benefit-management | https://api.github.com/repos/KShewengger/benefit-management | opened | Setup Vouchers Module, Initial Functionality and Route | Priority: Medium Type: API Type: Functionality | Common:
- Entity
- Model
- Type
Providers:
- Resolver
- Service
Core:
- Controller
- Module
Routes:
`/vouchers` | 1.0 | Setup Vouchers Module, Initial Functionality and Route - Common:
- Entity
- Model
- Type
Providers:
- Resolver
- Service
Core:
- Controller
- Module
Routes:
`/vouchers` | priority | setup vouchers module initial functionality and route common entity model type providers resolver service core controller module routes vouchers | 1 |
21,938 | 7,098,106,501 | IssuesEvent | 2018-01-15 02:25:32 | hadithhouse/hadithhouse | https://api.github.com/repos/hadithhouse/hadithhouse | opened | Use tslint in build | build & deploy | To make sure that code not styled correctly doesn't get into the repository, I would like to add `tslint` to the build process so build fails if there are style warnings. | 1.0 | Use tslint in build - To make sure that code not styled correctly doesn't get into the repository, I would like to add `tslint` to the build process so build fails if there are style warnings. | non_priority | use tslint in build to make sure that code not styled correctly doesn t get into the repository i would like to add tslint to the build process so build fails if there are style warnings | 0 |
134,014 | 5,218,783,475 | IssuesEvent | 2017-01-26 17:17:47 | duckduckgo/zeroclickinfo-fathead | https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead | opened | Objective-C: Create new Objective-C Reference Fathead | Difficulty: High Mission: Programming New Instant Answer Priority: High Status: Needs a Developer Topic: Objective-C Topic: Reference | # Create a new C Fathead Instant Answer
Help us make DuckDuckGo the best search engine for programmers!
### What do I need to know?
You'll need to know how to code in **Perl**, **Python**, **Ruby**, or **JavaScript**.

### What am I doing?
You will write a script that scrapes or downloads the data source below, and generates an **output.txt** file containing the parsed documentation. You can learn more about Fatheads and the `output.txt` syntax [**here**](https://docs.duckduckhack.com/resources/fathead-overview.html).
**Data source**: https://developer.apple.com/reference/objectivec
**Instant Answer Page**: *To be created by developer*
<!-- ^^^ FILL THIS IN, AFTER ISSUE IS CLAIMED ^^^ -->
### What is the Goal?
As part of our [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53), we're aiming to reach 100% Instant Answer (IA) coverage for searches related to programming languages by creating new Instant Answers, and improving existing ones.
Here are some Fathead examples:
- Ruby Docs
- [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/ruby) | [Example Query](https://duckduckgo.com/?q=array+bsearch&ia=about)
- MDN CSS
- [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/mdn_css) | [Example Query](https://duckduckgo.com/?q=css+background-position&ia=about)

[See more related Instant Answers](https://duck.co/ia?repo=fathead)
## Get Started
- [ ] 1) Claim this issue by commenting below
- [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-fathead/blob/master/CONTRIBUTING.md)
- [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository
- [ ] 4) Create a new Instant Answer Page: https://duck.co/ia/new_ia (then let us know, here!)
- [ ] 5) Create the Fathead
- [ ] 6) Create a Pull Request
- [ ] 7) Ping @pjhampton for a review
<!-- ^^^ FILL THIS IN ^^^ -->
## Resources
- Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions
- Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics
- Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help | 1.0 | Objective-C: Create new Objective-C Reference Fathead - # Create a new C Fathead Instant Answer
Help us make DuckDuckGo the best search engine for programmers!
### What do I need to know?
You'll need to know how to code in **Perl**, **Python**, **Ruby**, or **JavaScript**.

### What am I doing?
You will write a script that scrapes or downloads the data source below, and generates an **output.txt** file containing the parsed documentation. You can learn more about Fatheads and the `output.txt` syntax [**here**](https://docs.duckduckhack.com/resources/fathead-overview.html).
**Data source**: https://developer.apple.com/reference/objectivec
**Instant Answer Page**: *To be created by developer*
<!-- ^^^ FILL THIS IN, AFTER ISSUE IS CLAIMED ^^^ -->
### What is the Goal?
As part of our [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53), we're aiming to reach 100% Instant Answer (IA) coverage for searches related to programming languages by creating new Instant Answers, and improving existing ones.
Here are some Fathead examples:
- Ruby Docs
- [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/ruby) | [Example Query](https://duckduckgo.com/?q=array+bsearch&ia=about)
- MDN CSS
- [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/mdn_css) | [Example Query](https://duckduckgo.com/?q=css+background-position&ia=about)

[See more related Instant Answers](https://duck.co/ia?repo=fathead)
## Get Started
- [ ] 1) Claim this issue by commenting below
- [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-fathead/blob/master/CONTRIBUTING.md)
- [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository
- [ ] 4) Create a new Instant Answer Page: https://duck.co/ia/new_ia (then let us know, here!)
- [ ] 5) Create the Fathead
- [ ] 6) Create a Pull Request
- [ ] 7) Ping @pjhampton for a review
<!-- ^^^ FILL THIS IN ^^^ -->
## Resources
- Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions
- Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics
- Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help | priority | objective c create new objective c reference fathead create a new c fathead instant answer help us make duckduckgo the best search engine for programmers what do i need to know you ll need to know how to code in perl python ruby or javascript what am i doing you will write a script that scrapes or downloads the data source below and generates an output txt file containing the parsed documentation you can learn more about fatheads and the output txt syntax data source instant answer page to be created by developer what is the goal as part of our we re aiming to reach instant answer ia coverage for searches related to programming languages by creating new instant answers and improving existing ones here are some fathead examples ruby docs mdn css get started claim this issue by commenting below review our and fork this repository create a new instant answer page then let us know here create the fathead create a pull request ping pjhampton for a review resources join to ask questions join the to discuss project planning and instant answer metrics read the for technical help | 1 |
408,132 | 27,645,353,792 | IssuesEvent | 2023-03-10 22:18:31 | flyteorg/flyte | https://api.github.com/repos/flyteorg/flyte | opened | [Docs] Create model deployment and data annotation-specific examples for gate nodes | documentation untriaged | ### Description
https://github.com/flyteorg/flytesnacks/pull/968 tracks work for adding a gate nodes example to flytesnacks.
This issue is to add examples to the Tutorials section of flytesnacks, under two new headers:
- Model Deployment: create an example where gate nodes are used to pick which model to deploy
- Data Annotation: create an example where gate nodes are used to
Reference: see [these commits](https://github.com/flyteorg/flytesnacks/pull/968/commits/016fdcc4a59202e7ac555380b90fc97524f7444d) and [this notebook](https://github.com/flyteorg/flytesnacks/blob/ae1bef45fd536b0576694b4801db73169157cb47/cookbook/Untitled.ipynb) for examples of the above two use cases.
### Are you sure this issue hasn't been raised already?
- [X] Yes
### Have you read the Code of Conduct?
- [X] Yes | 1.0 | [Docs] Create model deployment and data annotation-specific examples for gate nodes - ### Description
https://github.com/flyteorg/flytesnacks/pull/968 tracks work for adding a gate nodes example to flytesnacks.
This issue is to add examples to the Tutorials section of flytesnacks, under two new headers:
- Model Deployment: create an example where gate nodes are used to pick which model to deploy
- Data Annotation: create an example where gate nodes are used to
Reference: see [these commits](https://github.com/flyteorg/flytesnacks/pull/968/commits/016fdcc4a59202e7ac555380b90fc97524f7444d) and [this notebook](https://github.com/flyteorg/flytesnacks/blob/ae1bef45fd536b0576694b4801db73169157cb47/cookbook/Untitled.ipynb) for examples of the above two use cases.
### Are you sure this issue hasn't been raised already?
- [X] Yes
### Have you read the Code of Conduct?
- [X] Yes | non_priority | create model deployment and data annotation specific examples for gate nodes description tracks work for adding a gate nodes example to flytesnacks this issue is to add examples to the tutorials section of flytesnacks under two new headers model deployment create an example where gate nodes are used to pick which model to deploy data annotation create an example where gate nodes are used to reference see and for examples of the above two use cases are you sure this issue hasn t been raised already yes have you read the code of conduct yes | 0 |
536,999 | 15,721,184,789 | IssuesEvent | 2021-03-29 02:24:09 | wso2/product-microgateway | https://api.github.com/repos/wso2/product-microgateway | opened | Support Basic Auth Endpoint security | Priority/Normal Type/New Feature | ### Describe your problem(s)
Currently any of endpoint security mechanism is not supported in mgw. Basic auth support should be included
### Describe your solution
Adding security details to API object and parse to enforcer in order to add basic auth credentials at runtime
### How will you implement it
JWTAuthenticator flow will be added new header appending basic auth credentials checking endpoint keytype
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| 1.0 | Support Basic Auth Endpoint security - ### Describe your problem(s)
Currently any of endpoint security mechanism is not supported in mgw. Basic auth support should be included
### Describe your solution
Adding security details to API object and parse to enforcer in order to add basic auth credentials at runtime
### How will you implement it
JWTAuthenticator flow will be added new header appending basic auth credentials checking endpoint keytype
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| priority | support basic auth endpoint security describe your problem s currently any of endpoint security mechanism is not supported in mgw basic auth support should be included describe your solution adding security details to api object and parse to enforcer in order to add basic auth credentials at runtime how will you implement it jwtauthenticator flow will be added new header appending basic auth credentials checking endpoint keytype optional fields related issues suggested labels suggested assignees | 1 |
97,363 | 28,214,053,785 | IssuesEvent | 2023-04-05 07:36:51 | microsoft/appcenter | https://api.github.com/repos/microsoft/appcenter | reopened | iOS and Android Build Fail when Test on Real Device is Toggled on | bug build test reviewed-DRI | **What App Center service does this affect?**
React Native SDK, and Build.
**Describe the bug**
Locally Xcode is building normally, but it fails in appcenter with Test on Real Device Toggled on. When it's toggled off, the build is successful.
**Expected behavior**
Build successful.
Additional context
##[error]Error: /usr/local/bin/appcenter failed with return code: 3
##[section]Finishing: Run generated tests on real devices | 1.0 | iOS and Android Build Fail when Test on Real Device is Toggled on - **What App Center service does this affect?**
React Native SDK, and Build.
**Describe the bug**
Locally Xcode is building normally, but it fails in appcenter with Test on Real Device Toggled on. When it's toggled off, the build is successful.
**Expected behavior**
Build successful.
Additional context
##[error]Error: /usr/local/bin/appcenter failed with return code: 3
##[section]Finishing: Run generated tests on real devices | non_priority | ios and android build fail when test on real device is toggled on what app center service does this affect react native sdk and build describe the bug locally xcode is building normally but it fails in appcenter with test on real device toggled on when it s toggled off the build is successful expected behavior build successful additional context error usr local bin appcenter failed with return code finishing run generated tests on real devices | 0 |
235,213 | 7,735,453,222 | IssuesEvent | 2018-05-27 15:17:40 | ream/ream | https://api.github.com/repos/ream/ream | closed | Prefetch data for route components | contribution welcome enhancement priority: high | Currently we have [getInitialData](https://github.com/ream/ream/blob/master/docs/guides/preloading-data.md) but it does not inject data or props to relevant route component, hopefully we can implement a _real_ `getInitialData` which injects resolved object to component data or `getInitialProps` which injects props instead.
`getInitialData`:
```js
export default {
async getInitialData() {
return { count: 0 }
},
render() {
return <button>{this.count /* this.$data.count */}</button>
}
}
```
`getInitialProps`:
```js
export default {
async getInitialProps() {
return { count: 0 }
},
props: ['count']
render() {
return <button>{this.count /* this.$props.count */}</button>
}
}
```
If it's possible to have `getInitialProps` then we don't need to implement `getInitialData`
---
Refs:
- [getInitialProps](https://github.com/zeit/next.js/#fetching-data-and-component-lifecycle) in Next.js
- [asyncData](https://nuxtjs.org/api/) in Nuxt.js | 1.0 | Prefetch data for route components - Currently we have [getInitialData](https://github.com/ream/ream/blob/master/docs/guides/preloading-data.md) but it does not inject data or props to relevant route component, hopefully we can implement a _real_ `getInitialData` which injects resolved object to component data or `getInitialProps` which injects props instead.
`getInitialData`:
```js
export default {
async getInitialData() {
return { count: 0 }
},
render() {
return <button>{this.count /* this.$data.count */}</button>
}
}
```
`getInitialProps`:
```js
export default {
async getInitialProps() {
return { count: 0 }
},
props: ['count']
render() {
return <button>{this.count /* this.$props.count */}</button>
}
}
```
If it's possible to have `getInitialProps` then we don't need to implement `getInitialData`
---
Refs:
- [getInitialProps](https://github.com/zeit/next.js/#fetching-data-and-component-lifecycle) in Next.js
- [asyncData](https://nuxtjs.org/api/) in Nuxt.js | priority | prefetch data for route components currently we have but it does not inject data or props to relevant route component hopefully we can implement a real getinitialdata which injects resolved object to component data or getinitialprops which injects props instead getinitialdata js export default async getinitialdata return count render return this count this data count getinitialprops js export default async getinitialprops return count props render return this count this props count if it s possible to have getinitialprops then we don t need to implement getinitialdata refs in next js in nuxt js | 1 |
305,847 | 9,377,715,858 | IssuesEvent | 2019-04-04 11:02:07 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | m.facebook.com - see bug description | browser-firefox-mobile priority-critical | <!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://m.facebook.com/story.php?story_fbid=2260305700712631&id=1160243874052158&anchor_composer=false
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: I use messenger on a tablet, but i am having issues with message notifications clearing from all of my devices, unless I load the desktop site on my phone and go to the inbox from there. then the notification will clear out on all devices that I am viewing the site from.
**Steps to Reproduce**:
I am using the silk browser, and fb messenger on the tablet. Firefox on the phone, and that's how I clear them. The desktop site is broken as well. Only one day this last year or two has it displayed like the desktop site looks on a PC. The day that it displayed correctly was the recent day that Facebook was broken for the full day. When I could get it to load that day, it looked identical to the desktop site. Every other day, it's just a bunch of text and links.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190218131312</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=2314269647&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2c&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=1710212709&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2e&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | m.facebook.com - see bug description - <!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://m.facebook.com/story.php?story_fbid=2260305700712631&id=1160243874052158&anchor_composer=false
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: I use messenger on a tablet, but i am having issues with message notifications clearing from all of my devices, unless I load the desktop site on my phone and go to the inbox from there. then the notification will clear out on all devices that I am viewing the site from.
**Steps to Reproduce**:
I am using the silk browser, and fb messenger on the tablet. Firefox on the phone, and that's how I clear them. The desktop site is broken as well. Only one day this last year or two has it displayed like the desktop site looks on a PC. The day that it displayed correctly was the recent day that Facebook was broken for the full day. When I could get it to load that day, it looked identical to the desktop site. Every other day, it's just a bunch of text and links.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190218131312</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=2314269647&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2c&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=1710212709&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2e&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | m facebook com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description i use messenger on a tablet but i am having issues with message notifications clearing from all of my devices unless i load the desktop site on my phone and go to the inbox from there then the notification will clear out on all devices that i am viewing the site from steps to reproduce i am using the silk browser and fb messenger on the tablet firefox on the phone and that s how i clear them the desktop site is broken as well only one day this last year or two has it displayed like the desktop site looks on a pc the day that it displayed correctly was the recent day that facebook was broken for the full day when i could get it to load that day it looked identical to the desktop site every other day it s just a bunch of text and links browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u u u u from with ❤️ | 1 |
414,483 | 12,103,836,536 | IssuesEvent | 2020-04-20 19:07:17 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | media.interieur.gouv.fr - see bug description | browser-fenix engine-gecko ml-needsdiagnosis-false priority-normal | <!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51919 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://media.interieur.gouv.fr/deplacement-covid-19/
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: PDF download is blocked
**Steps to Reproduce**:
generated pdf cannot be downloaded, error message.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | media.interieur.gouv.fr - see bug description - <!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51919 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://media.interieur.gouv.fr/deplacement-covid-19/
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: PDF download is blocked
**Steps to Reproduce**:
generated pdf cannot be downloaded, error message.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | media interieur gouv fr see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description pdf download is blocked steps to reproduce generated pdf cannot be downloaded error message browser configuration none from with ❤️ | 1 |
266,464 | 8,367,974,659 | IssuesEvent | 2018-10-04 13:42:16 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Improve program directory related experience | Component/Composer Imported Priority/High Type/Improvement | <a href="https://github.com/kaviththiranga"><img src="https://avatars3.githubusercontent.com/u/1505855?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [kaviththiranga](https://github.com/kaviththiranga)**
_Tuesday Oct 31, 2017 at 03:12 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/4306_
----
- When a user tries to open a packaged file within a program dir, ask whether user wants to open the program dir too and open upon confirmation.
- When a user adds a package to a file for the first time,
- If it's a new file which not yet persisted, upon save, in save dialog, show a check box with message that we need to create the sub folder for package structure.
- if it's file which is already saved, display an alert with same message and provide a btn to create folder structure and move file there
- when a user changes the package of a opened file, ask and move file to correct directory
- provide a wizard to create a file within program dir and a package
| 1.0 | Improve program directory related experience - <a href="https://github.com/kaviththiranga"><img src="https://avatars3.githubusercontent.com/u/1505855?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [kaviththiranga](https://github.com/kaviththiranga)**
_Tuesday Oct 31, 2017 at 03:12 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/4306_
----
- When a user tries to open a packaged file within a program dir, ask whether user wants to open the program dir too and open upon confirmation.
- When a user adds a package to a file for the first time,
- If it's a new file which not yet persisted, upon save, in save dialog, show a check box with message that we need to create the sub folder for package structure.
- if it's file which is already saved, display an alert with same message and provide a btn to create folder structure and move file there
- when a user changes the package of a opened file, ask and move file to correct directory
- provide a wizard to create a file within program dir and a package
| priority | improve program directory related experience issue by tuesday oct at gmt originally opened as when a user tries to open a packaged file within a program dir ask whether user wants to open the program dir too and open upon confirmation when a user adds a package to a file for the first time if it s a new file which not yet persisted upon save in save dialog show a check box with message that we need to create the sub folder for package structure if it s file which is already saved display an alert with same message and provide a btn to create folder structure and move file there when a user changes the package of a opened file ask and move file to correct directory provide a wizard to create a file within program dir and a package | 1 |
611,988 | 18,988,281,137 | IssuesEvent | 2021-11-22 01:44:23 | boostcampwm-2021/iOS05-Escaper | https://api.github.com/repos/boostcampwm-2021/iOS05-Escaper | closed | [E8 S2 T1] 랭킹 View를 구현한다. | feature High Priority | ### Epic - Story - Task
Epic : 상세 페이지
Story : 방에 대한 랭킹을 보낼 수 있다
Task : 랭킹 Cell을 구현한다.
| 1.0 | [E8 S2 T1] 랭킹 View를 구현한다. - ### Epic - Story - Task
Epic : 상세 페이지
Story : 방에 대한 랭킹을 보낼 수 있다
Task : 랭킹 Cell을 구현한다.
| priority | 랭킹 view를 구현한다 epic story task epic 상세 페이지 story 방에 대한 랭킹을 보낼 수 있다 task 랭킹 cell을 구현한다 | 1 |
99,441 | 20,966,290,711 | IssuesEvent | 2022-03-28 07:06:35 | Validator2/MesaBox | https://api.github.com/repos/Validator2/MesaBox | opened | Mining Laser, man-portable/emplacement | code models | Original idea by "sirro"
Powered by U-235, its a modern, finalized successor to the Gluon Gun. Instead of completely vaporizing matter, it has several functions;
tbd | 1.0 | Mining Laser, man-portable/emplacement - Original idea by "sirro"
Powered by U-235, its a modern, finalized successor to the Gluon Gun. Instead of completely vaporizing matter, it has several functions;
tbd | non_priority | mining laser man portable emplacement original idea by sirro powered by u its a modern finalized successor to the gluon gun instead of completely vaporizing matter it has several functions tbd | 0 |
1,604 | 6,445,176,676 | IssuesEvent | 2017-08-12 23:23:56 | p4lang/p4-spec | https://api.github.com/repos/p4lang/p4-spec | closed | [PSA] count() operation on counters | portable switch architecture | For a 'bytes' or 'packets_and_bytes' counter type, count() having a
second parameter 'increment' that specifies how much to add to the
byte counter is a very good thing. It is best if the P4 program has
the flexibility to choose the length in bytes it wants to use for the
packet, e.g. in case it is increasing or decreasing the received
packet's size before transmitting it.
For counters that have type 'packets', presumably every call to
'count' will add exactly 1, yes? In that case, having a separate
method call that only takes an 'index' parameter would be best, since
the 'increment' parameter value would be ignored, anyway. | 1.0 | [PSA] count() operation on counters - For a 'bytes' or 'packets_and_bytes' counter type, count() having a
second parameter 'increment' that specifies how much to add to the
byte counter is a very good thing. It is best if the P4 program has
the flexibility to choose the length in bytes it wants to use for the
packet, e.g. in case it is increasing or decreasing the received
packet's size before transmitting it.
For counters that have type 'packets', presumably every call to
'count' will add exactly 1, yes? In that case, having a separate
method call that only takes an 'index' parameter would be best, since
the 'increment' parameter value would be ignored, anyway. | non_priority | count operation on counters for a bytes or packets and bytes counter type count having a second parameter increment that specifies how much to add to the byte counter is a very good thing it is best if the program has the flexibility to choose the length in bytes it wants to use for the packet e g in case it is increasing or decreasing the received packet s size before transmitting it for counters that have type packets presumably every call to count will add exactly yes in that case having a separate method call that only takes an index parameter would be best since the increment parameter value would be ignored anyway | 0 |
320,160 | 9,777,097,699 | IssuesEvent | 2019-06-07 08:08:15 | DCRGraphsNet/DCROpenCaseManager | https://api.github.com/repos/DCRGraphsNet/DCROpenCaseManager | reopened | Will it be possible to add the child's ages as a guard of an activity? | Priority 3 Udviklingforslag | eg. when the young person turns 16 or 16, the activity becomes pending
So - we need the Age of the child in the process. Or date of birth - not cpr. So we can calculate age. Right now we do not have this capability. You could enter Age in a form - and solve it this way. But we cannot execute "Robot" events that take data from system and set into a graph. But - probably a future requirement | 1.0 | Will it be possible to add the child's ages as a guard of an activity? - eg. when the young person turns 16 or 16, the activity becomes pending
So - we need the Age of the child in the process. Or date of birth - not cpr. So we can calculate age. Right now we do not have this capability. You could enter Age in a form - and solve it this way. But we cannot execute "Robot" events that take data from system and set into a graph. But - probably a future requirement | priority | will it be possible to add the child s ages as a guard of an activity eg when the young person turns or the activity becomes pending so we need the age of the child in the process or date of birth not cpr so we can calculate age right now we do not have this capability you could enter age in a form and solve it this way but we cannot execute robot events that take data from system and set into a graph but probably a future requirement | 1 |
368,791 | 25,808,835,568 | IssuesEvent | 2022-12-11 17:04:26 | 12rambau/sepal_ui | https://api.github.com/repos/12rambau/sepal_ui | closed | refactor the graph in the contributor section | documentation | We are curently suggesting to use the following branching system:

It is not the one we currently use in the repository so it should be updated with a custom one (maybe using graphviz ?)
| 1.0 | refactor the graph in the contributor section - We are curently suggesting to use the following branching system:

It is not the one we currently use in the repository so it should be updated with a custom one (maybe using graphviz ?)
| non_priority | refactor the graph in the contributor section we are curently suggesting to use the following branching system it is not the one we currently use in the repository so it should be updated with a custom one maybe using graphviz | 0 |
593,496 | 18,009,412,923 | IssuesEvent | 2021-09-16 06:40:55 | GC-spigot/AdvancedEnchantments | https://api.github.com/repos/GC-spigot/AdvancedEnchantments | closed | Effect TNT not reduced by Blast Protection or DECREASE_DAMAGE | Priority: Low Bug: Confirmed Resolution: Accepted | ## Details
**Describe the bug**
The effect "TNT" cannot be reduced by vanilla Blast Protection or AE's DECREASE_DAMAGE effect (when activated by EXPLOSION).
**To Reproduce** <!-- !IMPORTANT! -->
1. Create an armor enchantment with a damaging TNT effect, activated by FALL_DAMAGE.
2. Create an armor enchantment that decreases EXPLOSION type damage by 75%.
3. Take fall damage from the same height
a) First without anything.
b) Then with the TNT effect.
c) Finally with the EXPLOSION damage reduction effect.
4. Damage will be the same in all three
**Screenshots / Video**
## Server Information
- "/ae plinfo" link: https://paste.md-5.net/sotureroni
- Server log: No messages from or related to AE.
| 1.0 | Effect TNT not reduced by Blast Protection or DECREASE_DAMAGE - ## Details
**Describe the bug**
The effect "TNT" cannot be reduced by vanilla Blast Protection or AE's DECREASE_DAMAGE effect (when activated by EXPLOSION).
**To Reproduce** <!-- !IMPORTANT! -->
1. Create an armor enchantment with a damaging TNT effect, activated by FALL_DAMAGE.
2. Create an armor enchantment that decreases EXPLOSION type damage by 75%.
3. Take fall damage from the same height
a) First without anything.
b) Then with the TNT effect.
c) Finally with the EXPLOSION damage reduction effect.
4. Damage will be the same in all three
**Screenshots / Video**
## Server Information
- "/ae plinfo" link: https://paste.md-5.net/sotureroni
- Server log: No messages from or related to AE.
| priority | effect tnt not reduced by blast protection or decrease damage details describe the bug the effect tnt cannot be reduced by vanilla blast protection or ae s decrease damage effect when activated by explosion to reproduce create an armor enchantment with a damaging tnt effect activated by fall damage create an armor enchantment that decreases explosion type damage by take fall damage from the same height a first without anything b then with the tnt effect c finally with the explosion damage reduction effect damage will be the same in all three screenshots video server information ae plinfo link server log no messages from or related to ae | 1 |
50,021 | 26,433,021,846 | IssuesEvent | 2023-01-15 02:40:05 | GraphiteEditor/Graphite | https://api.github.com/repos/GraphiteEditor/Graphite | closed | Debounce widget inputs to minimize backend spam | Feature Web P-Medium Performance | Send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently. | True | Debounce widget inputs to minimize backend spam - Send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently. | non_priority | debounce widget inputs to minimize backend spam send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently | 0 |
447,286 | 12,887,563,861 | IssuesEvent | 2020-07-13 11:28:31 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1634] Coroutine couldn't be started | Category: Tech Priority: Medium Status: Fixed Week Task | When you go to main menu:
```
Coroutine couldn't be started because the the game object 'News(Clone)' is inactive!
News.NewsRenderer:DoRendering(News)
News.NewsController:SetNews(List`1)
News.<SetNews>d__4:MoveNext()
UnityEngine.SetupCoroutine:InvokeMoveNext(IEnumerator, IntPtr)
``` | 1.0 | [0.9.0 staging-1634] Coroutine couldn't be started - When you go to main menu:
```
Coroutine couldn't be started because the the game object 'News(Clone)' is inactive!
News.NewsRenderer:DoRendering(News)
News.NewsController:SetNews(List`1)
News.<SetNews>d__4:MoveNext()
UnityEngine.SetupCoroutine:InvokeMoveNext(IEnumerator, IntPtr)
``` | priority | coroutine couldn t be started when you go to main menu coroutine couldn t be started because the the game object news clone is inactive news newsrenderer dorendering news news newscontroller setnews list news d movenext unityengine setupcoroutine invokemovenext ienumerator intptr | 1 |
144,381 | 11,614,148,682 | IssuesEvent | 2020-02-26 12:03:47 | pingcap/tidb-operator | https://api.github.com/repos/pingcap/tidb-operator | closed | e2e: "[Feature: AdvancedStatefulSet] Scaling tidb cluster with advanced statefulset " is flaky | test/e2e | ## Bug Report
https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/operator_ghpr_e2e_test_kind/detail/operator_ghpr_e2e_test_kind/2273/tests
```
Stacktrace
/home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:139
Jan 10 10:04:16.182: Unexpected error:
<*meta.NoKindMatchError | 0xc0010bb400>: {
GroupKind: {Group: "pingcap.com", Kind: "TidbCluster"},
SearchedVersions: ["v1alpha1"],
}
no matches for kind "TidbCluster" in version "pingcap.com/v1alpha1"
occurred
/home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:241
```
k8s version: v1.12.10 | 1.0 | e2e: "[Feature: AdvancedStatefulSet] Scaling tidb cluster with advanced statefulset " is flaky - ## Bug Report
https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/operator_ghpr_e2e_test_kind/detail/operator_ghpr_e2e_test_kind/2273/tests
```
Stacktrace
/home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:139
Jan 10 10:04:16.182: Unexpected error:
<*meta.NoKindMatchError | 0xc0010bb400>: {
GroupKind: {Group: "pingcap.com", Kind: "TidbCluster"},
SearchedVersions: ["v1alpha1"],
}
no matches for kind "TidbCluster" in version "pingcap.com/v1alpha1"
occurred
/home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:241
```
k8s version: v1.12.10 | non_priority | scaling tidb cluster with advanced statefulset is flaky bug report stacktrace home jenkins agent workspace operator ghpr test kind go src github com pingcap tidb operator tests tidbcluster serial go jan unexpected error groupkind group pingcap com kind tidbcluster searchedversions no matches for kind tidbcluster in version pingcap com occurred home jenkins agent workspace operator ghpr test kind go src github com pingcap tidb operator tests tidbcluster serial go version | 0 |
53,463 | 13,167,592,090 | IssuesEvent | 2020-08-11 10:32:49 | wellcomecollection/platform | https://api.github.com/repos/wellcomecollection/platform | closed | Associate CI policy for Buildkite agents IAM Role | :recycle: Builds and CI | Buildkite agents (as EC2 instances) created using the Buildkite AWS elastic stack (https://github.com/buildkite/elastic-ci-stack-for-aws) have associated IAM roles which should have policy documents mirroring our existing CI permissions. | 1.0 | Associate CI policy for Buildkite agents IAM Role - Buildkite agents (as EC2 instances) created using the Buildkite AWS elastic stack (https://github.com/buildkite/elastic-ci-stack-for-aws) have associated IAM roles which should have policy documents mirroring our existing CI permissions. | non_priority | associate ci policy for buildkite agents iam role buildkite agents as instances created using the buildkite aws elastic stack have associated iam roles which should have policy documents mirroring our existing ci permissions | 0 |
802,865 | 29,047,860,883 | IssuesEvent | 2023-05-13 20:06:58 | vdjagilev/nmap-formatter | https://api.github.com/repos/vdjagilev/nmap-formatter | closed | Update Node.js 12 -> 16 for pipelines | priority/medium type/other prop/pipeline | ```
Node.js 12 actions are deprecated. Please update the following actions to use Node.js 16: actions/checkout@v2. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/
```
Need to upgrade github actions to version 16 | 1.0 | Update Node.js 12 -> 16 for pipelines - ```
Node.js 12 actions are deprecated. Please update the following actions to use Node.js 16: actions/checkout@v2. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/
```
Need to upgrade github actions to version 16 | priority | update node js for pipelines node js actions are deprecated please update the following actions to use node js actions checkout for more information see need to upgrade github actions to version | 1 |
661,913 | 22,095,763,009 | IssuesEvent | 2022-06-01 09:57:29 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Allow deleting qcow2 disk snapshot using blockcommit | Category: Drivers - Storage Type: Feature Sponsored Status: Accepted Priority: Normal | **Description**
Currently deleting disk snapshots when using qcow2 DS is very limited. Users can not delete the active snapshots and also snapshots with children, so the only way how to delete the snapshot is to revert back.
**Use case**
e.g: when users get charged in terms of snapshot numbers, they need to delete them
**Interface Changes**
Deleting non-active snapshots with children could be achieved using `blockcommit` op.
https://wiki.libvirt.org/page/Live-disk-backup-with-active-blockcommit
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Allow deleting qcow2 disk snapshot using blockcommit - **Description**
Currently deleting disk snapshots when using qcow2 DS is very limited. Users can not delete the active snapshots and also snapshots with children, so the only way how to delete the snapshot is to revert back.
**Use case**
e.g: when users get charged in terms of snapshot numbers, they need to delete them
**Interface Changes**
Deleting non-active snapshots with children could be achieved using `blockcommit` op.
https://wiki.libvirt.org/page/Live-disk-backup-with-active-blockcommit
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| priority | allow deleting disk snapshot using blockcommit description currently deleting disk snapshots when using ds is very limited users can not delete the active snapshots and also snapshots with children so the only way how to delete the snapshot is to revert back use case e g when users get charged in terms of snapshot numbers they need to delete them interface changes deleting non active snapshots with children could be achieved using blockcommit op progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 1 |
29,789 | 14,265,763,425 | IssuesEvent | 2020-11-20 17:36:50 | ampproject/amp-toolbox-php | https://api.github.com/repos/ampproject/amp-toolbox-php | opened | New Transformer: BrowserHints | Performance | - [ ] Add preconnect link tag for Google fonts resources
- [ ] Add preconnect link tag to the publisher's own origin
- [ ] Preload images
- [ ] Preload AMP runtime script
- [ ] Prune duplicate resource hints
| True | New Transformer: BrowserHints - - [ ] Add preconnect link tag for Google fonts resources
- [ ] Add preconnect link tag to the publisher's own origin
- [ ] Preload images
- [ ] Preload AMP runtime script
- [ ] Prune duplicate resource hints
| non_priority | new transformer browserhints add preconnect link tag for google fonts resources add preconnect link tag to the publisher s own origin preload images preload amp runtime script prune duplicate resource hints | 0 |
88,181 | 15,800,747,050 | IssuesEvent | 2021-04-03 01:06:24 | rammatzkvosky/jenkins | https://api.github.com/repos/rammatzkvosky/jenkins | opened | CVE-2021-21348 (High) detected in xstream-1.4.15.jar | security vulnerability | ## CVE-2021-21348 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.15.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: jenkins/test/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,canner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,jenkins/war/target/jenkins/WEB-INF/lib/xstream-1.4.15.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.15.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348>CVE-2021-21348</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq">https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.15","packageFilePaths":["/test/pom.xml","/war/pom.xml","/core/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21348","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-21348 (High) detected in xstream-1.4.15.jar - ## CVE-2021-21348 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.15.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: jenkins/test/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,canner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,jenkins/war/target/jenkins/WEB-INF/lib/xstream-1.4.15.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.15.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348>CVE-2021-21348</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq">https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.15","packageFilePaths":["/test/pom.xml","/war/pom.xml","/core/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21348","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar library home page a href path to dependency file jenkins test pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar canner repository com thoughtworks xstream xstream xstream jar jenkins war target jenkins web inf lib xstream jar dependency hierarchy x xstream jar vulnerable library found in base branch master vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum cpu time and will never return no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum cpu time and will never return no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl | 0 |
586,874 | 17,599,170,794 | IssuesEvent | 2021-08-17 09:37:18 | googleapis/nodejs-video-intelligence | https://api.github.com/repos/googleapis/nodejs-video-intelligence | opened | analyze samples: should track objects in a GCS file failed | type: bug priority: p1 flakybot: issue | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2f7fe652af0d15621b89fc80cc22f9b2d0a2f209
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/24a5eecc-b508-4ba3-8f13-09342425c56a), [Sponge](http://sponge2/24a5eecc-b508-4ba3-8f13-09342425c56a)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4
analyze.js track-objects-gcs <gcsUri>
Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API.
Options:
--version Show version number [boolean]
--help Show help [boolean]
Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)
at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78
at processTicksAndRejections (internal/process/task_queues.js:79:11) {
code: 16,
details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.',
metadata: Metadata {
internalRepr: Map {
'google.rpc.errorinfo-bin' => [Array],
'grpc-status-details-bin' => [Array],
'www-authenticate' => [Array],
'method' => [Array],
'service' => [Array],
'email' => [Array]
},
options: {}
},
note: 'Exception occurred in retry method that was not classified as transient',
statusDetails: [
ErrorInfo {
metadata: [Object],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
Error: Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4
analyze.js track-objects-gcs <gcsUri>
Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API.
Options:
--version Show version number [boolean]
--help Show help [boolean]
Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)
at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78
at processTicksAndRejections (internal/process/task_queues.js:79:11) {
code: 16,
details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.',
metadata: Metadata {
internalRepr: Map {
'google.rpc.errorinfo-bin' => [Array],
'grpc-status-details-bin' => [Array],
'www-authenticate' => [Array],
'method' => [Array],
'service' => [Array],
'email' => [Array]
},
options: {}
},
note: 'Exception occurred in retry method that was not classified as transient',
statusDetails: [
ErrorInfo {
metadata: [Object],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (system-test/analyze.test.js:23:28)
at Context.<anonymous> (system-test/analyze.test.js:48:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | 1.0 | analyze samples: should track objects in a GCS file failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2f7fe652af0d15621b89fc80cc22f9b2d0a2f209
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/24a5eecc-b508-4ba3-8f13-09342425c56a), [Sponge](http://sponge2/24a5eecc-b508-4ba3-8f13-09342425c56a)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4
analyze.js track-objects-gcs <gcsUri>
Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API.
Options:
--version Show version number [boolean]
--help Show help [boolean]
Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)
at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78
at processTicksAndRejections (internal/process/task_queues.js:79:11) {
code: 16,
details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.',
metadata: Metadata {
internalRepr: Map {
'google.rpc.errorinfo-bin' => [Array],
'grpc-status-details-bin' => [Array],
'www-authenticate' => [Array],
'method' => [Array],
'service' => [Array],
'email' => [Array]
},
options: {}
},
note: 'Exception occurred in retry method that was not classified as transient',
statusDetails: [
ErrorInfo {
metadata: [Object],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
Error: Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4
analyze.js track-objects-gcs <gcsUri>
Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API.
Options:
--version Show version number [boolean]
--help Show help [boolean]
Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)
at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)
at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78
at processTicksAndRejections (internal/process/task_queues.js:79:11) {
code: 16,
details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.',
metadata: Metadata {
internalRepr: Map {
'google.rpc.errorinfo-bin' => [Array],
'grpc-status-details-bin' => [Array],
'www-authenticate' => [Array],
'method' => [Array],
'service' => [Array],
'email' => [Array]
},
options: {}
},
note: 'Exception occurred in retry method that was not classified as transient',
statusDetails: [
ErrorInfo {
metadata: [Object],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
],
reason: 'ACCOUNT_STATE_INVALID',
domain: 'googleapis.com'
}
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (system-test/analyze.test.js:23:28)
at Context.<anonymous> (system-test/analyze.test.js:48:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | priority | analyze samples should track objects in a gcs file failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node analyze js track objects gcs gs cloud samples data video cat analyze js track objects gcs analyzes objects in a video stored in google cloud storage using the cloud video intelligence api options version show version number help show help error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map google rpc errorinfo bin grpc status details bin www authenticate method service email options note exception occurred in retry method that was not classified as transient statusdetails errorinfo metadata reason account state invalid domain googleapis com reason account state invalid domain googleapis com error command failed node analyze js track objects gcs gs cloud samples data video cat analyze js track objects gcs analyzes objects in a video stored in google cloud storage using the cloud video intelligence api options version show version number help show help error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map google rpc errorinfo bin grpc status details bin www authenticate method service email options note exception occurred in retry method that was not classified as transient statusdetails errorinfo metadata reason account state invalid domain googleapis com reason account state invalid domain googleapis com at checkexecsyncerror child process js at object execsync child process js at execsync system test analyze test js at context system test analyze test js at processimmediate internal timers js | 1 |
286,054 | 8,783,367,864 | IssuesEvent | 2018-12-20 05:32:54 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | closed | Project Clone does not copy allowed user list if private | Low Priority bug | Cloning a project keeps the private setting, but does not keep the user list. | 1.0 | Project Clone does not copy allowed user list if private - Cloning a project keeps the private setting, but does not keep the user list. | priority | project clone does not copy allowed user list if private cloning a project keeps the private setting but does not keep the user list | 1 |
37,490 | 5,117,329,407 | IssuesEvent | 2017-01-07 15:41:46 | ngageoint/hootenanny-ui | https://api.github.com/repos/ngageoint/hootenanny-ui | closed | Port refactored schema switcher code | Status: In Progress Status: Ready for Test | This feature was refactored in isolation https://github.com/brianhatchl/iD/pull/1
So may need some patching around the edges to pull out the old stuff and shoehorn the new. | 1.0 | Port refactored schema switcher code - This feature was refactored in isolation https://github.com/brianhatchl/iD/pull/1
So may need some patching around the edges to pull out the old stuff and shoehorn the new. | non_priority | port refactored schema switcher code this feature was refactored in isolation so may need some patching around the edges to pull out the old stuff and shoehorn the new | 0 |
89,076 | 17,783,607,561 | IssuesEvent | 2021-08-31 08:26:18 | zeek/spicy | https://api.github.com/repos/zeek/spicy | closed | Non-public enum type gets optimized out even though needed | Bug Codegen | If a unit depends on a non-public enum type it sometimes gets optimized out, even though it is needed.
```
$ cat x.spicy
module x;
type E = enum { a, b, c };
public type U = unit {
var e: E;
};
```
```
$ spicyc -j x.spicy -D global-optimizer
[debug/global-optimizer] disabling feature 'supports_filters' of type 'x::U' since it is not used
[debug/global-optimizer] disabling feature 'supports_sinks' of type 'x::U' since it is not used
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_init with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_error with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_done with default value
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_init
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_done
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_error
[debug/global-optimizer] removing field for unused method x::U::__str__
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_finally
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_gap
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_overlap
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_skipped
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_undelivered
[debug/global-optimizer] removing declaration for unused function hilti::print
[debug/global-optimizer] removing declaration for unused function hilti::printValues
[debug/global-optimizer] removing declaration for unused function hilti::debug
[debug/global-optimizer] removing declaration for unused function hilti::current_time
[debug/global-optimizer] removing declaration for unused function hilti::mktime
[debug/global-optimizer] removing declaration for unused function hilti::abort
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_connect
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward_eod
[debug/global-optimizer] removing declaration for unused function spicy_rt::createContext
[debug/global-optimizer] removing declaration for unused function spicy_rt::setContext
[debug/global-optimizer] removing declaration for unused function spicy_rt::printParserState
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::atEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::unit_find
[debug/global-optimizer] removing declaration for unused function spicy_rt::backtrack
[debug/global-optimizer] removing declaration for unused function spicy_rt::extractBits
[debug/global-optimizer] removing unused type 'x::E'
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:24:12: error: no type named 'E' in namespace '__hlt::x'
x::E e;
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x'
inline U(std::optional<x::E> e);
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x'
inline U(std::optional<x::E> e);
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:55:45: error: unknown type name 'E'
inline auto operator<<(std::ostream& o, E x) -> std::ostream&;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: use of undeclared identifier 'x'; did you mean '__hlt::x'?
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
__hlt::x
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:21:18: note: '__hlt::x' declared here
namespace __hlt::x {
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:17: error: redefinition of 'to_string' as different kind of symbol
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/usr/local/Cellar/spicy/HEAD-1f9a82a/include/hilti/rt/extension-points.h:17:20: note: previous definition is here
extern std::string to_string();
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: expected ')'
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:26: note: to match this '('
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:37: error: no member named 'E' in namespace '__hlt::x'
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:51: error: expected ';' after top level declarator
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
;
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:78:16: error: no member named 'E' in namespace '__hlt::x'
e = x::E::Undef;
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x'
inline U::U(std::optional<x::E> e) : U() {
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x'
inline U::U(std::optional<x::E> e) : U() {
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:198:51: error: unknown type name 'E'
inline auto __hlt::x::operator<<(std::ostream& o, E x) -> std::ostream& { o << hilti::rt::to_string(x); return o; }
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:200:57: error: no type named 'E' in namespace '__hlt::x'
inline auto hilti::rt::detail::adl::to_string(__hlt::x::E x, adl::tag) -> std::string {
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:202:24: error: no member named 'E' in namespace '__hlt::x'
case __hlt::x::E::c:
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:204:24: error: no member named 'E' in namespace '__hlt::x'
case __hlt::x::E::b:
~~~~~~~~~~^
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
[error] JIT compilation failed
[error] (no error output)
``` | 1.0 | Non-public enum type gets optimized out even though needed - If a unit depends on a non-public enum type it sometimes gets optimized out, even though it is needed.
```
$ cat x.spicy
module x;
type E = enum { a, b, c };
public type U = unit {
var e: E;
};
```
```
$ spicyc -j x.spicy -D global-optimizer
[debug/global-optimizer] disabling feature 'supports_filters' of type 'x::U' since it is not used
[debug/global-optimizer] disabling feature 'supports_sinks' of type 'x::U' since it is not used
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_init with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_error with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value
[debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_done with default value
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_init
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_done
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_error
[debug/global-optimizer] removing field for unused method x::U::__str__
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_finally
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_gap
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_overlap
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_skipped
[debug/global-optimizer] removing field for unused method x::U::__on_0x25_undelivered
[debug/global-optimizer] removing declaration for unused function hilti::print
[debug/global-optimizer] removing declaration for unused function hilti::printValues
[debug/global-optimizer] removing declaration for unused function hilti::debug
[debug/global-optimizer] removing declaration for unused function hilti::current_time
[debug/global-optimizer] removing declaration for unused function hilti::mktime
[debug/global-optimizer] removing declaration for unused function hilti::abort
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_connect
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward
[debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward_eod
[debug/global-optimizer] removing declaration for unused function spicy_rt::createContext
[debug/global-optimizer] removing declaration for unused function spicy_rt::setContext
[debug/global-optimizer] removing declaration for unused function spicy_rt::printParserState
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput
[debug/global-optimizer] removing declaration for unused function spicy_rt::waitForEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::atEod
[debug/global-optimizer] removing declaration for unused function spicy_rt::unit_find
[debug/global-optimizer] removing declaration for unused function spicy_rt::backtrack
[debug/global-optimizer] removing declaration for unused function spicy_rt::extractBits
[debug/global-optimizer] removing unused type 'x::E'
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:24:12: error: no type named 'E' in namespace '__hlt::x'
x::E e;
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x'
inline U(std::optional<x::E> e);
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x'
inline U(std::optional<x::E> e);
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:55:45: error: unknown type name 'E'
inline auto operator<<(std::ostream& o, E x) -> std::ostream&;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: use of undeclared identifier 'x'; did you mean '__hlt::x'?
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
__hlt::x
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:21:18: note: '__hlt::x' declared here
namespace __hlt::x {
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:17: error: redefinition of 'to_string' as different kind of symbol
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/usr/local/Cellar/spicy/HEAD-1f9a82a/include/hilti/rt/extension-points.h:17:20: note: previous definition is here
extern std::string to_string();
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: expected ')'
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:26: note: to match this '('
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:37: error: no member named 'E' in namespace '__hlt::x'
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:51: error: expected ';' after top level declarator
inline auto to_string(__hlt::x::E x, adl::tag) -> std::string;
^
;
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing]
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~~~~
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: note: insert an explicit cast to silence this issue
const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) };
^~~~~~~~~~~~~~~~~~~~~~~~~
static_cast<ptrdiff_t>( )
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof'
#define offsetof(t, d) __builtin_offsetof(t, d)
^~~~~~~~~~~~~~~~~~~~~~~~
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:78:16: error: no member named 'E' in namespace '__hlt::x'
e = x::E::Undef;
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x'
inline U::U(std::optional<x::E> e) : U() {
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x'
inline U::U(std::optional<x::E> e) : U() {
~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:198:51: error: unknown type name 'E'
inline auto __hlt::x::operator<<(std::ostream& o, E x) -> std::ostream& { o << hilti::rt::to_string(x); return o; }
^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:200:57: error: no type named 'E' in namespace '__hlt::x'
inline auto hilti::rt::detail::adl::to_string(__hlt::x::E x, adl::tag) -> std::string {
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:202:24: error: no member named 'E' in namespace '__hlt::x'
case __hlt::x::E::c:
~~~~~~~~~~^
/private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:204:24: error: no member named 'E' in namespace '__hlt::x'
case __hlt::x::E::b:
~~~~~~~~~~^
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
[error] JIT compilation failed
[error] (no error output)
``` | non_priority | non public enum type gets optimized out even though needed if a unit depends on a non public enum type it sometimes gets optimized out even though it is needed cat x spicy module x type e enum a b c public type u unit var e e spicyc j x spicy d global optimizer disabling feature supports filters of type x u since it is not used disabling feature supports sinks of type x u since it is not used replacing call to unimplemented function x u on init with default value replacing call to unimplemented function x u on error with default value replacing call to unimplemented function x u on finally with default value replacing call to unimplemented function x u on finally with default value replacing call to unimplemented function x u on done with default value removing field for unused method x u on init removing field for unused method x u on done removing field for unused method x u on error removing field for unused method x u str removing field for unused method x u on finally removing field for unused method x u on gap removing field for unused method x u on overlap removing field for unused method x u on skipped removing field for unused method x u on undelivered removing declaration for unused function hilti print removing declaration for unused function hilti printvalues removing declaration for unused function hilti debug removing declaration for unused function hilti current time removing declaration for unused function hilti mktime removing declaration for unused function hilti abort removing declaration for unused function spicy rt filter connect removing declaration for unused function spicy rt filter forward removing declaration for unused function spicy rt filter forward eod removing declaration for unused function spicy rt createcontext removing declaration for unused function spicy rt setcontext removing declaration for unused function spicy rt printparserstate removing declaration for unused function spicy rt waitforinputoreod removing declaration for unused function spicy rt waitforinputoreod removing declaration for unused function spicy rt waitforinput removing declaration for unused function spicy rt waitforinput removing declaration for unused function spicy rt waitforeod removing declaration for unused function spicy rt ateod removing declaration for unused function spicy rt unit find removing declaration for unused function spicy rt backtrack removing declaration for unused function spicy rt extractbits removing unused type x e private var folders ht t hilti rt test x cc error no type named e in namespace hlt x x e e private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u std optional e private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u std optional e private var folders ht t hilti rt test x cc error unknown type name e inline auto operator std ostream private var folders ht t hilti rt test x cc error use of undeclared identifier x did you mean hlt x inline auto to string hlt x e x adl tag std string hlt x private var folders ht t hilti rt test x cc note hlt x declared here namespace hlt x private var folders ht t hilti rt test x cc error redefinition of to string as different kind of symbol inline auto to string hlt x e x adl tag std string usr local cellar spicy head include hilti rt extension points h note previous definition is here extern std string to string private var folders ht t hilti rt test x cc error expected inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc note to match this inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error expected after top level declarator inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error no member named e in namespace hlt x e x e undef private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u u std optional e u private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u u std optional e u private var folders ht t hilti rt test x cc error unknown type name e inline auto hlt x operator std ostream o hilti rt to string x return o private var folders ht t hilti rt test x cc error no type named e in namespace hlt x inline auto hilti rt detail adl to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error no member named e in namespace hlt x case hlt x e c private var folders ht t hilti rt test x cc error no member named e in namespace hlt x case hlt x e b fatal error too many errors emitted stopping now errors generated jit compilation failed no error output | 0 |
352,961 | 25,091,913,520 | IssuesEvent | 2022-11-08 07:10:03 | argilla-io/argilla | https://api.github.com/repos/argilla-io/argilla | closed | Docs: Replace one old UI screenshot | bug documentation | **Describe the bug**
A screenshot from Rubrix UI persists on docs .
**To Reproduce**
Steps to reproduce the behavior:
(weak labeling mode view)
https://docs.argilla.io/en/latest/reference/webapp/features.html
**Expected behavior**
Update it with a screenshot from the new UI
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Environment (please complete the following information):**
-
**Additional context**
-
| 1.0 | Docs: Replace one old UI screenshot - **Describe the bug**
A screenshot from Rubrix UI persists on docs .
**To Reproduce**
Steps to reproduce the behavior:
(weak labeling mode view)
https://docs.argilla.io/en/latest/reference/webapp/features.html
**Expected behavior**
Update it with a screenshot from the new UI
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Environment (please complete the following information):**
-
**Additional context**
-
| non_priority | docs replace one old ui screenshot describe the bug a screenshot from rubrix ui persists on docs to reproduce steps to reproduce the behavior weak labeling mode view expected behavior update it with a screenshot from the new ui screenshots if applicable add screenshots to help explain your problem environment please complete the following information additional context | 0 |
207,140 | 16,066,867,139 | IssuesEvent | 2021-04-23 20:40:18 | bounswe/2021SpringGroup12 | https://api.github.com/repos/bounswe/2021SpringGroup12 | closed | Merge Use Case Diagrams | documentation status: Review Request | First Deadline for review 23.04.2021, @18:00
Final Deadline 23.04.2021, @22:00
Others have drawn their parts on Lucid. I will merge them in one page. | 1.0 | Merge Use Case Diagrams - First Deadline for review 23.04.2021, @18:00
Final Deadline 23.04.2021, @22:00
Others have drawn their parts on Lucid. I will merge them in one page. | non_priority | merge use case diagrams first deadline for review final deadline others have drawn their parts on lucid i will merge them in one page | 0 |
721,973 | 24,845,593,829 | IssuesEvent | 2022-10-26 15:38:26 | googleapis/nodejs-pubsub | https://api.github.com/repos/googleapis/nodejs-pubsub | opened | createSubscription method throws if oidcToken is being set | priority: p2 type: bug | Thanks for stopping by to let us know something could be better!
**PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response.
1) Is this a client library issue or a product issue?
Client library issue
2) Did someone already solve this?
This user faced something similar with Python and was solved by changing the Python version.
https://github.com/googleapis/python-pubsub/issues/480
3) Do you have a support contract?
No
If the support paths suggested above still do not result in a resolution, please provide the following details.
#### Environment details
- OS: Mac OS Monterey 12.6 (also happens in firebase cloud functions environment)
- Node.js version: 16.13.1 (also happens with node 14 and node 12)
- npm version: 8.1.2
- `@google-cloud/pubsub` version: 3.2.0
#### Steps to reproduce
1. Create a topic.
`await pubSubClient.createTopic(topicName)`
2. Create a subscription for the topic, setting OIDC token (either in pushConfig or in the root object as admitted by the sdk).
```
await pubSubClient.createSubscription(item.topicName, item.subscriptionName, {
pushConfig: {
pushEndpoint: item.endpoint,
oidcToken: {
serviceAccountEmail,
audience: projectId,
},
},
expirationPolicy: {
ttl: null,
},
retryPolicy: {
minimumBackoff: {
seconds: 10,
},
maximumBackoff: {
seconds: 60,
},
},
});
}));
```
This also fails:
```
await pubSubClient.createSubscription(item.topicName, item.subscriptionName, {
pushEndpoint: item.endpoint,
oidcToken: {
serviceAccountEmail,
audience: projectId,
},
expirationPolicy: {
ttl: null,
},
retryPolicy: {
minimumBackoff: {
seconds: 10,
},
maximumBackoff: {
seconds: 60,
},
},
});
}));
```
This was working properly, we found the first error on Sep 29th 2022.
Making sure to follow these steps will guarantee the quickest resolution possible.
Thanks!
| 1.0 | createSubscription method throws if oidcToken is being set - Thanks for stopping by to let us know something could be better!
**PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response.
1) Is this a client library issue or a product issue?
Client library issue
2) Did someone already solve this?
This user faced something similar with Python and was solved by changing the Python version.
https://github.com/googleapis/python-pubsub/issues/480
3) Do you have a support contract?
No
If the support paths suggested above still do not result in a resolution, please provide the following details.
#### Environment details
- OS: Mac OS Monterey 12.6 (also happens in firebase cloud functions environment)
- Node.js version: 16.13.1 (also happens with node 14 and node 12)
- npm version: 8.1.2
- `@google-cloud/pubsub` version: 3.2.0
#### Steps to reproduce
1. Create a topic.
`await pubSubClient.createTopic(topicName)`
2. Create a subscription for the topic, setting OIDC token (either in pushConfig or in the root object as admitted by the sdk).
```
await pubSubClient.createSubscription(item.topicName, item.subscriptionName, {
pushConfig: {
pushEndpoint: item.endpoint,
oidcToken: {
serviceAccountEmail,
audience: projectId,
},
},
expirationPolicy: {
ttl: null,
},
retryPolicy: {
minimumBackoff: {
seconds: 10,
},
maximumBackoff: {
seconds: 60,
},
},
});
}));
```
This also fails:
```
await pubSubClient.createSubscription(item.topicName, item.subscriptionName, {
pushEndpoint: item.endpoint,
oidcToken: {
serviceAccountEmail,
audience: projectId,
},
expirationPolicy: {
ttl: null,
},
retryPolicy: {
minimumBackoff: {
seconds: 10,
},
maximumBackoff: {
seconds: 60,
},
},
});
}));
```
This was working properly, we found the first error on Sep 29th 2022.
Making sure to follow these steps will guarantee the quickest resolution possible.
Thanks!
| priority | createsubscription method throws if oidctoken is being set thanks for stopping by to let us know something could be better please read if you have a support contract with google please create an issue in the instead of filing on github this will ensure a timely response is this a client library issue or a product issue client library issue did someone already solve this this user faced something similar with python and was solved by changing the python version do you have a support contract no if the support paths suggested above still do not result in a resolution please provide the following details environment details os mac os monterey also happens in firebase cloud functions environment node js version also happens with node and node npm version google cloud pubsub version steps to reproduce create a topic await pubsubclient createtopic topicname create a subscription for the topic setting oidc token either in pushconfig or in the root object as admitted by the sdk await pubsubclient createsubscription item topicname item subscriptionname pushconfig pushendpoint item endpoint oidctoken serviceaccountemail audience projectid expirationpolicy ttl null retrypolicy minimumbackoff seconds maximumbackoff seconds this also fails await pubsubclient createsubscription item topicname item subscriptionname pushendpoint item endpoint oidctoken serviceaccountemail audience projectid expirationpolicy ttl null retrypolicy minimumbackoff seconds maximumbackoff seconds this was working properly we found the first error on sep making sure to follow these steps will guarantee the quickest resolution possible thanks | 1 |
158,288 | 6,025,001,058 | IssuesEvent | 2017-06-08 07:27:09 | VirtoCommerce/vc-platform | https://api.github.com/repos/VirtoCommerce/vc-platform | closed | Theme colors stopped working | bug Priority: High | When changing theme default definition from Blue to Dark, color schema no longer changes, only left filter changes. This is a showcase feature during the demo and needs to work. The bug is current in dev branch. | 1.0 | Theme colors stopped working - When changing theme default definition from Blue to Dark, color schema no longer changes, only left filter changes. This is a showcase feature during the demo and needs to work. The bug is current in dev branch. | priority | theme colors stopped working when changing theme default definition from blue to dark color schema no longer changes only left filter changes this is a showcase feature during the demo and needs to work the bug is current in dev branch | 1 |
26,426 | 26,853,938,009 | IssuesEvent | 2023-02-03 13:15:05 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | opened | Confusing error message: Argument is too big for formatting | usability | ```
SELECT format('{}asdfasd{}', '111')
Query id: 77df41ec-4f04-4017-863f-f7d31b92893d
0 rows in set. Elapsed: 0.031 sec.
Received exception from server (version 22.13.1):
Code: 36. DB::Exception: Received from localhost:9000. DB::Exception: Argument is too big for formatting: While processing format('{}asdfasd{}', 'a'). (BAD_ARGUMENTS)
```
Should be something like 'Not enough arguments provided to fill all placeholders in the format string."'
| True | Confusing error message: Argument is too big for formatting - ```
SELECT format('{}asdfasd{}', '111')
Query id: 77df41ec-4f04-4017-863f-f7d31b92893d
0 rows in set. Elapsed: 0.031 sec.
Received exception from server (version 22.13.1):
Code: 36. DB::Exception: Received from localhost:9000. DB::Exception: Argument is too big for formatting: While processing format('{}asdfasd{}', 'a'). (BAD_ARGUMENTS)
```
Should be something like 'Not enough arguments provided to fill all placeholders in the format string."'
| non_priority | confusing error message argument is too big for formatting select format asdfasd query id rows in set elapsed sec received exception from server version code db exception received from localhost db exception argument is too big for formatting while processing format asdfasd a bad arguments should be something like not enough arguments provided to fill all placeholders in the format string | 0 |
614,857 | 19,191,045,360 | IssuesEvent | 2021-12-06 00:23:20 | apcountryman/picolibrary | https://api.github.com/repos/apcountryman/picolibrary | closed | Remove reverse iterator | priority-normal status-awaiting_review type-refactoring | Remove reverse iterator (`::picolibrary::Reverse_Iterator`). `std::reverse_iterator` will be required instead. | 1.0 | Remove reverse iterator - Remove reverse iterator (`::picolibrary::Reverse_Iterator`). `std::reverse_iterator` will be required instead. | priority | remove reverse iterator remove reverse iterator picolibrary reverse iterator std reverse iterator will be required instead | 1 |
79,561 | 28,375,113,659 | IssuesEvent | 2023-04-12 20:10:30 | JohnAustinDev/xulsword | https://api.github.com/repos/JohnAustinDev/xulsword | closed | Add GUI capability to display the Greek TR words in module KJV version 2.6 lemma markup | Type-Defect Priority-Medium auto-migrated | ```
From kjv.conf
SwordVersionDate=2014-02-15
Version=2.6
History_2.6=Fixed bugs. Added Greek from TR.
Example: (mod2imp)
$$$Matthew 1:1
<w lemma="strong:G976 lemma.TR:βιβλος" morph="robinson:N-NSF" src="1">The
book</w> <w lemma="strong:G1078 lemma.TR:γενεσεως"
morph="robinson:N-GSF" src="2">of the generation</w> <w lemma="strong:G2424
lemma.TR:ιησου" morph="robinson:N-GSM" src="3">of Jesus</w> <w
lemma="strong:G5547 lemma.TR:χριστου" morph="robinson:N-GSM"
src="4">Christ</w>, <w lemma="strong:G5207 lemma.TR:υιου"
morph="robinson:N-GSM" src="5">the son</w> <w lemma="strong:G1138
lemma.TR:δαβιδ" morph="robinson:N-PRI" src="6">of David</w>, <w
lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="7">the
son</w> <w lemma="strong:G11 lemma.TR:αβρααμ" morph="robinson:N-PRI"
src="8">of Abraham</w>.
```
Original issue reported on code.google.com by `DFH...@gmail.com` on 16 Feb 2014 at 3:36
| 1.0 | Add GUI capability to display the Greek TR words in module KJV version 2.6 lemma markup - ```
From kjv.conf
SwordVersionDate=2014-02-15
Version=2.6
History_2.6=Fixed bugs. Added Greek from TR.
Example: (mod2imp)
$$$Matthew 1:1
<w lemma="strong:G976 lemma.TR:βιβλος" morph="robinson:N-NSF" src="1">The
book</w> <w lemma="strong:G1078 lemma.TR:γενεσεως"
morph="robinson:N-GSF" src="2">of the generation</w> <w lemma="strong:G2424
lemma.TR:ιησου" morph="robinson:N-GSM" src="3">of Jesus</w> <w
lemma="strong:G5547 lemma.TR:χριστου" morph="robinson:N-GSM"
src="4">Christ</w>, <w lemma="strong:G5207 lemma.TR:υιου"
morph="robinson:N-GSM" src="5">the son</w> <w lemma="strong:G1138
lemma.TR:δαβιδ" morph="robinson:N-PRI" src="6">of David</w>, <w
lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="7">the
son</w> <w lemma="strong:G11 lemma.TR:αβρααμ" morph="robinson:N-PRI"
src="8">of Abraham</w>.
```
Original issue reported on code.google.com by `DFH...@gmail.com` on 16 Feb 2014 at 3:36
| non_priority | add gui capability to display the greek tr words in module kjv version lemma markup from kjv conf swordversiondate version history fixed bugs added greek from tr example matthew the book w lemma strong lemma tr γενεσεως morph robinson n gsf src of the generation w lemma strong lemma tr ιησου morph robinson n gsm src of jesus w lemma strong lemma tr χριστου morph robinson n gsm src christ w lemma strong lemma tr υιου morph robinson n gsm src the son w lemma strong lemma tr δαβιδ morph robinson n pri src of david w lemma strong lemma tr υιου morph robinson n gsm src the son w lemma strong lemma tr αβρααμ morph robinson n pri src of abraham original issue reported on code google com by dfh gmail com on feb at | 0 |
59,707 | 12,013,373,137 | IssuesEvent | 2020-04-10 08:42:17 | home-assistant/brands | https://api.github.com/repos/home-assistant/brands | closed | Philips TV is missing brand images | domain-missing has-codeowner |
## The problem
The Philips TV integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/philips_js/icon.png`
- `src/philips_js/logo.png`
- `src/philips_js/icon@2x.png`
- `src/philips_js/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: philips.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/philips_js.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @elupus! Mind taking a look at this issue as it is with an integration (philips_js) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/philips_js/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| 1.0 | Philips TV is missing brand images -
## The problem
The Philips TV integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/philips_js/icon.png`
- `src/philips_js/logo.png`
- `src/philips_js/icon@2x.png`
- `src/philips_js/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: philips.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/philips_js.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @elupus! Mind taking a look at this issue as it is with an integration (philips_js) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/philips_js/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| non_priority | philips tv is missing brand images the problem the philips tv integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src philips js icon png src philips js logo png src philips js icon png src philips js logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo philips png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there elupus mind taking a look at this issue as it is with an integration philips js you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already | 0 |
446,697 | 12,876,716,568 | IssuesEvent | 2020-07-11 06:41:57 | luksan47/mars | https://api.github.com/repos/luksan47/mars | closed | Deleting mac addresses is not working | Priority: HIGH bug | on list.blade and and admin.internet.mac_addresses.list.blade also | 1.0 | Deleting mac addresses is not working - on list.blade and and admin.internet.mac_addresses.list.blade also | priority | deleting mac addresses is not working on list blade and and admin internet mac addresses list blade also | 1 |
709,164 | 24,369,190,466 | IssuesEvent | 2022-10-03 17:42:05 | Chatterino/chatterino2 | https://api.github.com/repos/Chatterino/chatterino2 | closed | Migrate /subscribersoff command to Helix API | Platform: Twitch Priority: Medium Deprecation: Twitch IRC Commands hacktoberfest | As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/subscribersoff` command needs to be migrated to use the relevant Helix API endpoint.
Helix API reference: https://dev.twitch.tv/docs/api/reference#update-chat-settings
Split from #3956 | 1.0 | Migrate /subscribersoff command to Helix API - As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/subscribersoff` command needs to be migrated to use the relevant Helix API endpoint.
Helix API reference: https://dev.twitch.tv/docs/api/reference#update-chat-settings
Split from #3956 | priority | migrate subscribersoff command to helix api as part of twitch s announced deprecation of irc based commands the subscribersoff command needs to be migrated to use the relevant helix api endpoint helix api reference split from | 1 |
102,769 | 11,307,004,396 | IssuesEvent | 2020-01-18 17:58:34 | Luceapuce/SEPR-Project | https://api.github.com/repos/Luceapuce/SEPR-Project | closed | willCollide if spawned in Entity | documentation question | Might need to write something to handle the situation of an entity spawning within an entity.
_This is just a reminder to check this later_ | 1.0 | willCollide if spawned in Entity - Might need to write something to handle the situation of an entity spawning within an entity.
_This is just a reminder to check this later_ | non_priority | willcollide if spawned in entity might need to write something to handle the situation of an entity spawning within an entity this is just a reminder to check this later | 0 |
51,739 | 6,195,202,786 | IssuesEvent | 2017-07-05 12:00:24 | mifort-org/mifort-timesheet | https://api.github.com/repos/mifort-org/mifort-timesheet | closed | One week in month is not displayed | bug fixed on test env | When you enter new data in current week and then click on back arrow in the first column, data of previous week isn't displayed. If you reload window data is displayed again. | 1.0 | One week in month is not displayed - When you enter new data in current week and then click on back arrow in the first column, data of previous week isn't displayed. If you reload window data is displayed again. | non_priority | one week in month is not displayed when you enter new data in current week and then click on back arrow in the first column data of previous week isn t displayed if you reload window data is displayed again | 0 |
148,431 | 19,531,026,647 | IssuesEvent | 2021-12-30 16:50:03 | vital-ws/empty | https://api.github.com/repos/vital-ws/empty | closed | CVE-2016-1000236 (Medium) detected in cookie-signature-1.0.3.tgz | security vulnerability | ## CVE-2016-1000236 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookie-signature-1.0.3.tgz</b></p></summary>
<p>Sign and unsign cookies</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz">https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/cookie-signature/package.json</p>
<p>
Dependency Hierarchy:
- express-4.0.0.tgz (Root Library)
- :x: **cookie-signature-1.0.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vital-ws/empty/commit/8b832bf233deeb3f62310d7aa93eae79c2af69ed">8b832bf233deeb3f62310d7aa93eae79c2af69ed</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.
<p>Publish Date: 2019-11-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236>CVE-2016-1000236</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e">https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e</a></p>
<p>Release Date: 2019-11-19</p>
<p>Fix Resolution (cookie-signature): 1.0.6</p>
<p>Direct dependency fix Resolution (express): 4.12.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"express:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.12.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-1000236","vulnerabilityDetails":"Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2016-1000236 (Medium) detected in cookie-signature-1.0.3.tgz - ## CVE-2016-1000236 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookie-signature-1.0.3.tgz</b></p></summary>
<p>Sign and unsign cookies</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz">https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/cookie-signature/package.json</p>
<p>
Dependency Hierarchy:
- express-4.0.0.tgz (Root Library)
- :x: **cookie-signature-1.0.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vital-ws/empty/commit/8b832bf233deeb3f62310d7aa93eae79c2af69ed">8b832bf233deeb3f62310d7aa93eae79c2af69ed</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.
<p>Publish Date: 2019-11-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236>CVE-2016-1000236</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e">https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e</a></p>
<p>Release Date: 2019-11-19</p>
<p>Fix Resolution (cookie-signature): 1.0.6</p>
<p>Direct dependency fix Resolution (express): 4.12.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"express:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.12.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-1000236","vulnerabilityDetails":"Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in cookie signature tgz cve medium severity vulnerability vulnerable library cookie signature tgz sign and unsign cookies library home page a href path to dependency file package json path to vulnerable library node modules cookie signature package json dependency hierarchy express tgz root library x cookie signature tgz vulnerable library found in head commit a href found in base branch main vulnerability details node cookie signature before is affected by a timing attack due to the type of comparison used publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cookie signature direct dependency fix resolution express check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree express isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node cookie signature before is affected by a timing attack due to the type of comparison used vulnerabilityurl | 0 |
174,885 | 27,747,982,185 | IssuesEvent | 2023-03-15 18:24:16 | Olian04/simply-reactive | https://api.github.com/repos/Olian04/simply-reactive | closed | Add createSessionAtom | further design needed | Much like `createQueryAtom` however it stores the value in `sessionStorage` instead of the query string.
```ts
const A = createSessionAtom({
key: 'a',
default: 0,
});
A.set(3);
// Reload page
A.get(); // 3
``` | 1.0 | Add createSessionAtom - Much like `createQueryAtom` however it stores the value in `sessionStorage` instead of the query string.
```ts
const A = createSessionAtom({
key: 'a',
default: 0,
});
A.set(3);
// Reload page
A.get(); // 3
``` | non_priority | add createsessionatom much like createqueryatom however it stores the value in sessionstorage instead of the query string ts const a createsessionatom key a default a set reload page a get | 0 |
56,313 | 15,020,016,603 | IssuesEvent | 2021-02-01 14:14:26 | mozilla-lockwise/lockwise-ios | https://api.github.com/repos/mozilla-lockwise/lockwise-ios | reopened | recently used doesn't work | archived defect | steps to reproduce:
- copy a username from an item
- navigate back to entries list
- sort by recently used
expected: item with copied username appears at the top
actual: not necessarily | 1.0 | recently used doesn't work - steps to reproduce:
- copy a username from an item
- navigate back to entries list
- sort by recently used
expected: item with copied username appears at the top
actual: not necessarily | non_priority | recently used doesn t work steps to reproduce copy a username from an item navigate back to entries list sort by recently used expected item with copied username appears at the top actual not necessarily | 0 |
194,890 | 6,900,397,676 | IssuesEvent | 2017-11-24 18:25:37 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | admin UI: Authentication sources rules description not displayed in listing | Priority: Low Status: For review Type: Feature / Enhancement Type: Nice to have | It would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it. | 1.0 | admin UI: Authentication sources rules description not displayed in listing - It would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it. | priority | admin ui authentication sources rules description not displayed in listing it would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it | 1 |
240,595 | 7,803,312,842 | IssuesEvent | 2018-06-10 22:18:13 | DigitalCampus/django-oppia | https://api.github.com/repos/DigitalCampus/django-oppia | closed | When creating a "ResponseResource" object via the API not all fields are required | invalid medium priority | they should be!
| 1.0 | When creating a "ResponseResource" object via the API not all fields are required - they should be!
| priority | when creating a responseresource object via the api not all fields are required they should be | 1 |
57,749 | 16,039,745,947 | IssuesEvent | 2021-04-22 06:03:14 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | New lightbox mouse wheel zoom too agressive | T-Defect | ### Description
Zooming the image by mouse wheel shows it in maximal/minimal zoom. Single wheel step should do the same thing as (+) and (-) buttons.
### Steps to reproduce
- open the lightbox clicking an image
- scrool mouse wheel one step up
- image is shown in maximal zoom
- scrool mouse wheel one step down
- image is shown in default size
Maximal/minimal zoom is ok when clicking the image by left mouse button, but mouse wheel should behave more fluently.
### Version information
- **Platform**: Element Web 1.7.26-rc.1
- **Browser**: Firefox 89b2
- **OS**: Windows 10
- **URL**: private homeserver
| 1.0 | New lightbox mouse wheel zoom too agressive - ### Description
Zooming the image by mouse wheel shows it in maximal/minimal zoom. Single wheel step should do the same thing as (+) and (-) buttons.
### Steps to reproduce
- open the lightbox clicking an image
- scrool mouse wheel one step up
- image is shown in maximal zoom
- scrool mouse wheel one step down
- image is shown in default size
Maximal/minimal zoom is ok when clicking the image by left mouse button, but mouse wheel should behave more fluently.
### Version information
- **Platform**: Element Web 1.7.26-rc.1
- **Browser**: Firefox 89b2
- **OS**: Windows 10
- **URL**: private homeserver
| non_priority | new lightbox mouse wheel zoom too agressive description zooming the image by mouse wheel shows it in maximal minimal zoom single wheel step should do the same thing as and buttons steps to reproduce open the lightbox clicking an image scrool mouse wheel one step up image is shown in maximal zoom scrool mouse wheel one step down image is shown in default size maximal minimal zoom is ok when clicking the image by left mouse button but mouse wheel should behave more fluently version information platform element web rc browser firefox os windows url private homeserver | 0 |
423,624 | 12,299,239,204 | IssuesEvent | 2020-05-11 12:01:39 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www5.ha.org.hk - site is not usable | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox 76.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52664 -->
**URL**: https://www5.ha.org.hk/rcbts
**Browser / Version**: Firefox 76.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Secure Connection Failed
An error occurred during a connection to www5.ha.org.hk. PR_CONNECT_RESET_ERROR
The site loads fine in Chrome and Chromium Edge
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www5.ha.org.hk - site is not usable - <!-- @browser: Firefox 76.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52664 -->
**URL**: https://www5.ha.org.hk/rcbts
**Browser / Version**: Firefox 76.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Secure Connection Failed
An error occurred during a connection to www5.ha.org.hk. PR_CONNECT_RESET_ERROR
The site loads fine in Chrome and Chromium Edge
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | ha org hk site is not usable url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description page not loading correctly steps to reproduce secure connection failed an error occurred during a connection to ha org hk pr connect reset error the site loads fine in chrome and chromium edge browser configuration none from with ❤️ | 1 |
653,761 | 21,626,002,868 | IssuesEvent | 2022-05-05 02:18:17 | bossbuwi/reality | https://api.github.com/repos/bossbuwi/reality | closed | Create event list view logic | enhancement logic high priority | The event list view must fetch and display data from the server. Clear any hard coded data. | 1.0 | Create event list view logic - The event list view must fetch and display data from the server. Clear any hard coded data. | priority | create event list view logic the event list view must fetch and display data from the server clear any hard coded data | 1 |
307,274 | 9,415,227,394 | IssuesEvent | 2019-04-10 12:09:29 | bio-tools/biotoolsRegistry | https://api.github.com/repos/bio-tools/biotoolsRegistry | closed | Count of search results not being cleared | GUI bug high priority | e.g. when navigating to Tool Card or other page (having done a search) the number of results remains visible in the search box - which looks bad. | 1.0 | Count of search results not being cleared - e.g. when navigating to Tool Card or other page (having done a search) the number of results remains visible in the search box - which looks bad. | priority | count of search results not being cleared e g when navigating to tool card or other page having done a search the number of results remains visible in the search box which looks bad | 1 |
254,930 | 19,277,491,629 | IssuesEvent | 2021-12-10 13:36:06 | PnX-SI/gn_module_monitoring | https://api.github.com/repos/PnX-SI/gn_module_monitoring | closed | Utilisation "filter" dans la configuration json | documentation question | Bonjour,
J'ai configuré un sous-module pour nos suivis par tente malaise et autres pièges fixes.
Je voudrais filtrer une liste de nomenclatures, j'ai vu que c'était prévu dans la doc. Mais (en tous cas avec ma conf) ça ne semble pas fonctionner. La configuration du widget est la suivante :
```
"id_trap_type": {
"type_widget": "datalist",
"attribut_label": "Type de piège",
"api": "nomenclatures/nomenclature/METH_OBS",
"application": "GeoNature",
"keyValue": "id_nomenclature",
"keyLabel": "label_fr",
"data_path": "values",
"type_util": "nomenclature",
"required": true,
"default": {
"cd_nomenclature": "52"
},
"filter" : {
"cd_nomenclature":["52","53","73"]
}
```
Il est bien probable que le soucis vienne de moi, je n'ai pas trouvé d'autres nomenclatures filtrées dans les autres modules déjà partagés. Avez-vous des pistes ? Merci d'avance !
| 1.0 | Utilisation "filter" dans la configuration json - Bonjour,
J'ai configuré un sous-module pour nos suivis par tente malaise et autres pièges fixes.
Je voudrais filtrer une liste de nomenclatures, j'ai vu que c'était prévu dans la doc. Mais (en tous cas avec ma conf) ça ne semble pas fonctionner. La configuration du widget est la suivante :
```
"id_trap_type": {
"type_widget": "datalist",
"attribut_label": "Type de piège",
"api": "nomenclatures/nomenclature/METH_OBS",
"application": "GeoNature",
"keyValue": "id_nomenclature",
"keyLabel": "label_fr",
"data_path": "values",
"type_util": "nomenclature",
"required": true,
"default": {
"cd_nomenclature": "52"
},
"filter" : {
"cd_nomenclature":["52","53","73"]
}
```
Il est bien probable que le soucis vienne de moi, je n'ai pas trouvé d'autres nomenclatures filtrées dans les autres modules déjà partagés. Avez-vous des pistes ? Merci d'avance !
| non_priority | utilisation filter dans la configuration json bonjour j ai configuré un sous module pour nos suivis par tente malaise et autres pièges fixes je voudrais filtrer une liste de nomenclatures j ai vu que c était prévu dans la doc mais en tous cas avec ma conf ça ne semble pas fonctionner la configuration du widget est la suivante id trap type type widget datalist attribut label type de piège api nomenclatures nomenclature meth obs application geonature keyvalue id nomenclature keylabel label fr data path values type util nomenclature required true default cd nomenclature filter cd nomenclature il est bien probable que le soucis vienne de moi je n ai pas trouvé d autres nomenclatures filtrées dans les autres modules déjà partagés avez vous des pistes merci d avance | 0 |
255,443 | 19,302,624,690 | IssuesEvent | 2021-12-13 08:06:32 | it-academyproject/ita-directory | https://api.github.com/repos/it-academyproject/ita-directory | closed | review documentation make install | bug documentation | - [ ] make install doesn't run => `make: *** No rule to make target 'install'. Stop.`
- [ ] make build bug =>
```
rm -f .env
process_begin: CreateProcess(NULL, rm -f .env, ...) failed.
make (e=2): El sistema no puede encontrar el archivo especificado.
make: *** [C:/Users/TESTER/Desktop/it-projecte/ita-directory/scripts/make/build.mk:4: dev] Error 2
```
- [ ] Need to clarify: where is the api-docs URL and how to get there. | 1.0 | review documentation make install - - [ ] make install doesn't run => `make: *** No rule to make target 'install'. Stop.`
- [ ] make build bug =>
```
rm -f .env
process_begin: CreateProcess(NULL, rm -f .env, ...) failed.
make (e=2): El sistema no puede encontrar el archivo especificado.
make: *** [C:/Users/TESTER/Desktop/it-projecte/ita-directory/scripts/make/build.mk:4: dev] Error 2
```
- [ ] Need to clarify: where is the api-docs URL and how to get there. | non_priority | review documentation make install make install doesn t run make no rule to make target install stop make build bug rm f env process begin createprocess null rm f env failed make e el sistema no puede encontrar el archivo especificado make error need to clarify where is the api docs url and how to get there | 0 |
74,260 | 20,101,578,146 | IssuesEvent | 2022-02-07 05:16:34 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | tensorflow.datasets.load() throws an exception | stat:awaiting response type:build/install stalled subtype: ubuntu/linux | <em>Please make sure that this is a bug. As per our
[GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md),
we only address code/doc bugs, performance issues, feature requests and
build/installation issues on GitHub. tag:bug_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): no
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: not a mobile device
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): v2.8.0-rc0-28-g24a4b3b5e58 2.8.0-rc1
- Python version: 3.8
- Bazel version (if compiling from source): 4.2.2
- GCC/Compiler version (if compiling from source): 9.3.0
- CUDA/cuDNN version: 11.2 / 8.1.1.33
- GPU model and memory: NVIDIA GeForce 940MX 1629 MB
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with:
1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"`
2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
After running the script
`import tensorflow as tf
import tensorflow_datasets as tfds
print(tf.__version__)
datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)`
I am getting the following output:
2.8.0-rc1
terminate called after throwing an instance of 'std::system_error'
what(): Invalid argument
Process finished with exit code 134 (interrupted by signal 6: SIGABRT)
**Describe the expected behavior**
No exception is expected
**[Contributing](https://www.tensorflow.org/community/contribute)**
- Do you want to contribute a PR? (yes/no): no
- Briefly describe your candidate solution(if contributing):
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
`import tensorflow as tf
import tensorflow_datasets as tfds
print(tf.__version__)
datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)`
**Other info / logs** Include any logs or source code that would be helpful to
diagnose the problem. If including tracebacks, please include the full
traceback. Large logs and files should be attached.
| 1.0 | tensorflow.datasets.load() throws an exception - <em>Please make sure that this is a bug. As per our
[GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md),
we only address code/doc bugs, performance issues, feature requests and
build/installation issues on GitHub. tag:bug_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): no
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: not a mobile device
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): v2.8.0-rc0-28-g24a4b3b5e58 2.8.0-rc1
- Python version: 3.8
- Bazel version (if compiling from source): 4.2.2
- GCC/Compiler version (if compiling from source): 9.3.0
- CUDA/cuDNN version: 11.2 / 8.1.1.33
- GPU model and memory: NVIDIA GeForce 940MX 1629 MB
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with:
1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"`
2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
After running the script
`import tensorflow as tf
import tensorflow_datasets as tfds
print(tf.__version__)
datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)`
I am getting the following output:
2.8.0-rc1
terminate called after throwing an instance of 'std::system_error'
what(): Invalid argument
Process finished with exit code 134 (interrupted by signal 6: SIGABRT)
**Describe the expected behavior**
No exception is expected
**[Contributing](https://www.tensorflow.org/community/contribute)**
- Do you want to contribute a PR? (yes/no): no
- Briefly describe your candidate solution(if contributing):
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
`import tensorflow as tf
import tensorflow_datasets as tfds
print(tf.__version__)
datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)`
**Other info / logs** Include any logs or source code that would be helpful to
diagnose the problem. If including tracebacks, please include the full
traceback. Large logs and files should be attached.
| non_priority | tensorflow datasets load throws an exception please make sure that this is a bug as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag bug template system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device not a mobile device tensorflow installed from source or binary source tensorflow version use command below python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory nvidia geforce mb you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior after running the script import tensorflow as tf import tensorflow datasets as tfds print tf version datasets info tfds load name mnist with info true as supervised true i am getting the following output terminate called after throwing an instance of std system error what invalid argument process finished with exit code interrupted by signal sigabrt describe the expected behavior no exception is expected do you want to contribute a pr yes no no briefly describe your candidate solution if contributing standalone code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem if possible please share a link to colab jupyter any notebook import tensorflow as tf import tensorflow datasets as tfds print tf version datasets info tfds load name mnist with info true as supervised true other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached | 0 |
315,167 | 27,051,136,237 | IssuesEvent | 2023-02-13 13:19:38 | enonic/app-contentstudio | https://api.github.com/repos/enonic/app-contentstudio | closed | Project wizard - language selector gets disabled after clicking on 'Copy from parent' button | Bug Not in Changelog Test is Failing | 1. Create a layer in Defalt project, do not select a language in the layer
2. Open the layer and select a language
3. Do not click on `Save` button in the wizard page, but click on `Copy from parent` button
**BUG** - Filter input gets disabled in the selector

| 1.0 | Project wizard - language selector gets disabled after clicking on 'Copy from parent' button - 1. Create a layer in Defalt project, do not select a language in the layer
2. Open the layer and select a language
3. Do not click on `Save` button in the wizard page, but click on `Copy from parent` button
**BUG** - Filter input gets disabled in the selector

| non_priority | project wizard language selector gets disabled after clicking on copy from parent button create a layer in defalt project do not select a language in the layer open the layer and select a language do not click on save button in the wizard page but click on copy from parent button bug filter input gets disabled in the selector | 0 |
26,890 | 13,148,611,441 | IssuesEvent | 2020-08-08 22:45:53 | E3SM-Project/E3SM | https://api.github.com/repos/E3SM-Project/E3SM | opened | output of chunk costs (print_cost_p) is expensive for large process counts | Atmosphere bug performance | The runtime of the routine print_cost_p, which writes out the cost of each chunk to the file atm_chunk_costs.txt, is large for large process counts on Cori-KNL. For example, for an F case using the ne256pg2 mesh and 32768 processes, this output took 18 minutes to complete. As print_cost_p is now called by default, this issue needs to be addressed quickly.
Given the severity of the performance problem, I have applied a "bug" label to this issue, even though it is functionally correct.
An alternative implementation has been developed, decreasing the runtime of print_cost_p for the above mentioned case from 18 minutes to less than 6 seconds. A PR with this change will be submitted soon.
| True | output of chunk costs (print_cost_p) is expensive for large process counts - The runtime of the routine print_cost_p, which writes out the cost of each chunk to the file atm_chunk_costs.txt, is large for large process counts on Cori-KNL. For example, for an F case using the ne256pg2 mesh and 32768 processes, this output took 18 minutes to complete. As print_cost_p is now called by default, this issue needs to be addressed quickly.
Given the severity of the performance problem, I have applied a "bug" label to this issue, even though it is functionally correct.
An alternative implementation has been developed, decreasing the runtime of print_cost_p for the above mentioned case from 18 minutes to less than 6 seconds. A PR with this change will be submitted soon.
| non_priority | output of chunk costs print cost p is expensive for large process counts the runtime of the routine print cost p which writes out the cost of each chunk to the file atm chunk costs txt is large for large process counts on cori knl for example for an f case using the mesh and processes this output took minutes to complete as print cost p is now called by default this issue needs to be addressed quickly given the severity of the performance problem i have applied a bug label to this issue even though it is functionally correct an alternative implementation has been developed decreasing the runtime of print cost p for the above mentioned case from minutes to less than seconds a pr with this change will be submitted soon | 0 |
160,215 | 6,084,950,851 | IssuesEvent | 2017-06-17 09:43:21 | climu/openstudyroom | https://api.github.com/repos/climu/openstudyroom | closed | Create a method that return all users one can play with | enhancement help wanted high priority | We need to calculate this players results for each divisions and then to filter the users with whom he didn't play the max_number of games. | 1.0 | Create a method that return all users one can play with - We need to calculate this players results for each divisions and then to filter the users with whom he didn't play the max_number of games. | priority | create a method that return all users one can play with we need to calculate this players results for each divisions and then to filter the users with whom he didn t play the max number of games | 1 |
448,857 | 31,815,653,595 | IssuesEvent | 2023-09-13 20:14:30 | proofcarryingdata/zupass | https://api.github.com/repos/proofcarryingdata/zupass | opened | example 'feed' application that demonstrates how 3rd parties are supposed to issue pcds to pcdpass users | documentation | cc @robknight | 1.0 | example 'feed' application that demonstrates how 3rd parties are supposed to issue pcds to pcdpass users - cc @robknight | non_priority | example feed application that demonstrates how parties are supposed to issue pcds to pcdpass users cc robknight | 0 |
40,782 | 8,847,362,711 | IssuesEvent | 2019-01-08 01:22:07 | pnp/pnpjs | https://api.github.com/repos/pnp/pnpjs | closed | Can't make batching work in sp-taxonomy | area: code status: details needed type: bug | ### Category
- [ ] Enhancement
- [x] Bug
- [x] Question
- [ ] Documentation gap/issue
### Version
Please specify what version of the library you are using: [1.2.7]
Please specify what version(s) of SharePoint you are targeting: [SharePoint Online]
### Expected / Desired Behavior / Question
Batching for `sp-taxonomy` works
### Observed Behavior
I don't see any outgoing requests after `batch.execute`
### Steps to Reproduce
```typescript
taxonomy.setup({
spfxContext: this.context
});
var termStores = await taxonomy.termStores.get();
let returnTerms: IPickerTerm[] = [];
for (let i = 0, len = termStores.length; i < len; i++) {
const pnpTermStore = termStores[i];
const pnpTerms = await pnpTermStore.getTerms({
TermLabel: 'Mar',
StringMatchOption: StringMatchOption.StartsWith,
DefaultLabelOnly: true,
TrimUnavailable: true,
ResultCollectionSize: 10
}).get();
const batch = taxonomy.createBatch();
// in real scenario there should be pnpTerms.forEach
const term = pnpTerms[0];
term.termSet.inBatch(batch).get().then(termSet => {
console.log(termSet);
});
term.labels.inBatch(batch).get().then(labels => {
console.log(labels);
});
await batch.execute();
``` | 1.0 | Can't make batching work in sp-taxonomy - ### Category
- [ ] Enhancement
- [x] Bug
- [x] Question
- [ ] Documentation gap/issue
### Version
Please specify what version of the library you are using: [1.2.7]
Please specify what version(s) of SharePoint you are targeting: [SharePoint Online]
### Expected / Desired Behavior / Question
Batching for `sp-taxonomy` works
### Observed Behavior
I don't see any outgoing requests after `batch.execute`
### Steps to Reproduce
```typescript
taxonomy.setup({
spfxContext: this.context
});
var termStores = await taxonomy.termStores.get();
let returnTerms: IPickerTerm[] = [];
for (let i = 0, len = termStores.length; i < len; i++) {
const pnpTermStore = termStores[i];
const pnpTerms = await pnpTermStore.getTerms({
TermLabel: 'Mar',
StringMatchOption: StringMatchOption.StartsWith,
DefaultLabelOnly: true,
TrimUnavailable: true,
ResultCollectionSize: 10
}).get();
const batch = taxonomy.createBatch();
// in real scenario there should be pnpTerms.forEach
const term = pnpTerms[0];
term.termSet.inBatch(batch).get().then(termSet => {
console.log(termSet);
});
term.labels.inBatch(batch).get().then(labels => {
console.log(labels);
});
await batch.execute();
``` | non_priority | can t make batching work in sp taxonomy category enhancement bug question documentation gap issue version please specify what version of the library you are using please specify what version s of sharepoint you are targeting expected desired behavior question batching for sp taxonomy works observed behavior i don t see any outgoing requests after batch execute steps to reproduce typescript taxonomy setup spfxcontext this context var termstores await taxonomy termstores get let returnterms ipickerterm for let i len termstores length i len i const pnptermstore termstores const pnpterms await pnptermstore getterms termlabel mar stringmatchoption stringmatchoption startswith defaultlabelonly true trimunavailable true resultcollectionsize get const batch taxonomy createbatch in real scenario there should be pnpterms foreach const term pnpterms term termset inbatch batch get then termset console log termset term labels inbatch batch get then labels console log labels await batch execute | 0 |
323,683 | 27,746,189,647 | IssuesEvent | 2023-03-15 17:11:52 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] flaky test: YbAdminSnapshotScheduleUpgradeTestWithYsql.PgsqlTestMigrationFromEarliestSysCatalogSnapshot | kind/bug area/ysql kind/failing-test priority/medium | Jira Link: [DB-5838](https://yugabyte.atlassian.net/browse/DB-5838)
### Description
From https://detective-gcp.dev.yugabyte.com/stability/test?class=YbAdminSnapshotScheduleUpgradeTestWithYsql&name=PgsqlTestMigrationFromEarliestSysCatalogSnapshot
it shows in debug build the test frequently failed with the following error:
```
../../src/yb/tools/yb-admin-snapshot-schedule-test.cc:2598
Failed
Expected to find substring 'Snapshot state and current state have different system catalogs'. Got: 'Subprocess '/net/ip-10-9-10-29.us-west-2.compute.internal/share/jenkins/jenkins-github-yugabyte-db-alma8-master-clang15-debug-aarch64-625/yugabyte-db/build/debug-clang15-dynamic-ninja/tests-tools/../bin/yb-admin' terminated with non-zero exit status 256: Error running restore_snapshot_schedule: Illegal state (yb/master/master_snapshot_coordinator.cc:192): Trying to restore to { days: 19430 time: 19:28:57.612482 } which is earlier than the configured retention. Not allowed. Earliest snapshot that can be used is c2a3ab4c-c14c-40e6-9c49-e27d3f8798d9 and was taken at { days: 19430 time: 19:29:53.222443 }.
```
Under debug build, it took the test longer than 10 minutes and the old snapshots taken 10 minutes ago already expired. We just need to bump the 10 minutes to a larger duration.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-5838]: https://yugabyte.atlassian.net/browse/DB-5838?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YSQL] flaky test: YbAdminSnapshotScheduleUpgradeTestWithYsql.PgsqlTestMigrationFromEarliestSysCatalogSnapshot - Jira Link: [DB-5838](https://yugabyte.atlassian.net/browse/DB-5838)
### Description
From https://detective-gcp.dev.yugabyte.com/stability/test?class=YbAdminSnapshotScheduleUpgradeTestWithYsql&name=PgsqlTestMigrationFromEarliestSysCatalogSnapshot
it shows in debug build the test frequently failed with the following error:
```
../../src/yb/tools/yb-admin-snapshot-schedule-test.cc:2598
Failed
Expected to find substring 'Snapshot state and current state have different system catalogs'. Got: 'Subprocess '/net/ip-10-9-10-29.us-west-2.compute.internal/share/jenkins/jenkins-github-yugabyte-db-alma8-master-clang15-debug-aarch64-625/yugabyte-db/build/debug-clang15-dynamic-ninja/tests-tools/../bin/yb-admin' terminated with non-zero exit status 256: Error running restore_snapshot_schedule: Illegal state (yb/master/master_snapshot_coordinator.cc:192): Trying to restore to { days: 19430 time: 19:28:57.612482 } which is earlier than the configured retention. Not allowed. Earliest snapshot that can be used is c2a3ab4c-c14c-40e6-9c49-e27d3f8798d9 and was taken at { days: 19430 time: 19:29:53.222443 }.
```
Under debug build, it took the test longer than 10 minutes and the old snapshots taken 10 minutes ago already expired. We just need to bump the 10 minutes to a larger duration.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-5838]: https://yugabyte.atlassian.net/browse/DB-5838?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | non_priority | flaky test ybadminsnapshotscheduleupgradetestwithysql pgsqltestmigrationfromearliestsyscatalogsnapshot jira link description from it shows in debug build the test frequently failed with the following error src yb tools yb admin snapshot schedule test cc failed expected to find substring snapshot state and current state have different system catalogs got subprocess net ip us west compute internal share jenkins jenkins github yugabyte db master debug yugabyte db build debug dynamic ninja tests tools bin yb admin terminated with non zero exit status error running restore snapshot schedule illegal state yb master master snapshot coordinator cc trying to restore to days time which is earlier than the configured retention not allowed earliest snapshot that can be used is and was taken at days time under debug build it took the test longer than minutes and the old snapshots taken minutes ago already expired we just need to bump the minutes to a larger duration warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information | 0 |
283,407 | 21,316,554,320 | IssuesEvent | 2022-04-16 11:33:14 | jr-mojito/pe | https://api.github.com/repos/jr-mojito/pe | opened | Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function | severity.Low type.DocumentationBug | Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function

<!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778-->
<!--Version: Web v3.4.2--> | 1.0 | Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function - Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function

<!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778-->
<!--Version: Web v3.4.2--> | non_priority | option field for found on ug but not available for edit function option field for found on ug but not available for edit function | 0 |
85,136 | 3,686,993,306 | IssuesEvent | 2016-02-25 05:13:48 | cs2103jan2016-w13-4j/main | https://api.github.com/repos/cs2103jan2016-w13-4j/main | closed | Create method that allows removing one tag from a task | component.storage priority.medium | Something like `Task.removeTag(int id, String tag)` | 1.0 | Create method that allows removing one tag from a task - Something like `Task.removeTag(int id, String tag)` | priority | create method that allows removing one tag from a task something like task removetag int id string tag | 1 |
216,650 | 16,794,397,425 | IssuesEvent | 2021-06-16 00:04:01 | microsoft/appcenter | https://api.github.com/repos/microsoft/appcenter | closed | Inconclusive and Skipped tests missing in the Test run results on the main page | Portal feature request test | In the log I see
Test Count: 84, Passed: 74, Failed: 8, Warnings: 0, Inconclusive: 0, Skipped: 2
On the main page I see

Will be great to see the real count of tests on the main page. | 1.0 | Inconclusive and Skipped tests missing in the Test run results on the main page - In the log I see
Test Count: 84, Passed: 74, Failed: 8, Warnings: 0, Inconclusive: 0, Skipped: 2
On the main page I see

Will be great to see the real count of tests on the main page. | non_priority | inconclusive and skipped tests missing in the test run results on the main page in the log i see test count passed failed warnings inconclusive skipped on the main page i see will be great to see the real count of tests on the main page | 0 |
484,679 | 13,943,533,823 | IssuesEvent | 2020-10-22 23:22:26 | SkriptLang/Skript | https://api.github.com/repos/SkriptLang/Skript | closed | Can't check if item has NO name applied - Beta 4 | bug completed dev needed priority: low | I used this line of code to check if player had X regular diamonds (that had no special name or lore):
`if arg-player has arg-number of diamonds named "":`
After updating to beta 2, beta 3 i started to get this error when loading the skript:
`the 1st argument of diamond named "" is not an item type (atm_diamond.sk, line 233: if arg-player has arg-number of diamonds named "":')`
After updating to beta 4 the error is gone, **but the syntax is still not working**! I am trying to report this since a month... am I the only fool out there in the need to remove unnamed items from inventory in an efficient way?
### Steps to Reproduce
As simple as this code:

### Expected Behavior
Only UNNAMED items should get removed
### Errors / Screenshots
There are no errors in beta 4
Here is a small video to explain what's happening https://youtu.be/rvMvsi-3akQ
### Server Information
* **Server version/platform:** Paper-199 1.16.3
* **Skript version:** 2.5-beta4
| 1.0 | Can't check if item has NO name applied - Beta 4 - I used this line of code to check if player had X regular diamonds (that had no special name or lore):
`if arg-player has arg-number of diamonds named "":`
After updating to beta 2, beta 3 i started to get this error when loading the skript:
`the 1st argument of diamond named "" is not an item type (atm_diamond.sk, line 233: if arg-player has arg-number of diamonds named "":')`
After updating to beta 4 the error is gone, **but the syntax is still not working**! I am trying to report this since a month... am I the only fool out there in the need to remove unnamed items from inventory in an efficient way?
### Steps to Reproduce
As simple as this code:

### Expected Behavior
Only UNNAMED items should get removed
### Errors / Screenshots
There are no errors in beta 4
Here is a small video to explain what's happening https://youtu.be/rvMvsi-3akQ
### Server Information
* **Server version/platform:** Paper-199 1.16.3
* **Skript version:** 2.5-beta4
| priority | can t check if item has no name applied beta i used this line of code to check if player had x regular diamonds that had no special name or lore if arg player has arg number of diamonds named after updating to beta beta i started to get this error when loading the skript the argument of diamond named is not an item type atm diamond sk line if arg player has arg number of diamonds named after updating to beta the error is gone but the syntax is still not working i am trying to report this since a month am i the only fool out there in the need to remove unnamed items from inventory in an efficient way steps to reproduce as simple as this code expected behavior only unnamed items should get removed errors screenshots there are no errors in beta here is a small video to explain what s happening server information server version platform paper skript version | 1 |
94,251 | 10,817,094,205 | IssuesEvent | 2019-11-08 08:57:30 | influxdata/influxdb-client-java | https://api.github.com/repos/influxdata/influxdb-client-java | closed | Documentation: README.md documents example that should be improved | documentation | The current README.md code documents classes that were renamed.
```
RetentionRule retention = new RetentionRule();
retention.setEverySeconds(3600L);
```
Correct example is:
```
BucketRetentionRules bucketRetentionRules = new BucketRetentionRules();
bucketRetentionRules.setEverySeconds(3600L);
````
It would also be nice to document here the default RP, and how I as a user can set up the autogen policy. | 1.0 | Documentation: README.md documents example that should be improved - The current README.md code documents classes that were renamed.
```
RetentionRule retention = new RetentionRule();
retention.setEverySeconds(3600L);
```
Correct example is:
```
BucketRetentionRules bucketRetentionRules = new BucketRetentionRules();
bucketRetentionRules.setEverySeconds(3600L);
````
It would also be nice to document here the default RP, and how I as a user can set up the autogen policy. | non_priority | documentation readme md documents example that should be improved the current readme md code documents classes that were renamed retentionrule retention new retentionrule retention seteveryseconds correct example is bucketretentionrules bucketretentionrules new bucketretentionrules bucketretentionrules seteveryseconds it would also be nice to document here the default rp and how i as a user can set up the autogen policy | 0 |
831,536 | 32,051,727,476 | IssuesEvent | 2023-09-23 16:33:41 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Action Containers broke spellbooks + exception | Issue: Bug Priority: 2-Before Release Difficulty: 2-Medium | ## Description
<!-- Explain your issue in detail. Issues without proper explanation are liable to be closed by maintainers. -->
An exception is thrown when trying to learn spells from spellbooks. The recent action container PR broke this behavior: #20260
**Reproduction**
<!-- Include the steps to reproduce if applicable. -->
Pick up a spellbook and try to learn it, console will show the exception.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
**Additional context**
<!-- Add any other context about the problem here. Anything you think is related to the issue. -->
> [ERRO] system.actions: Can't resolve "Content.Shared.Actions.ActionsContainerComponent" on entity 1985!
> at Content.Shared.Actions.SharedActionsSystem.GrantActions(EntityUid performer, IEnumerable`1 actions, EntityUid container, ActionsComponent comp, ActionsContainerComponent containerComp) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Actions\SharedActionsSystem.cs:l
> ine 523
> at Content.Shared.Magic.SharedMagicSystem.OnDoAfter(EntityUid uid, SpellbookComponent component, DoAfterEvent args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Magic\SharedMagicSystem.cs:line 85
> at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass47_0`2.<SubscribeLocalEvent>g__EventHandler|0(EntityUid uid, IComponent comp, TEvent& args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 254
> at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass57_0`1.<EntSubscribe>b__0(EntityUid uid, IComponent comp, Unit& ev) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 387
> at Robust.Shared.GameObjects.EntityEventBus.EntDispatch(EntityUid euid, Type eventType, Unit& args, Boolean dispatchByReference) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 541
> at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEventCore(EntityUid uid, Unit& unitRef, Type type, Boolean broadcast, Boolean byRef) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 221
> at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 190
> at Robust.Shared.GameObjects.EntitySystem.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystem.cs:line 155
> at Content.Shared.DoAfter.SharedDoAfterSystem.RaiseDoAfterEvents(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.cs:line 93
> at Content.Shared.DoAfter.SharedDoAfterSystem.TryComplete(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 123
> at Content.Shared.DoAfter.SharedDoAfterSystem.Update(EntityUid uid, ActiveDoAfterComponent active, DoAfterComponent comp, TimeSpan time, EntityQuery`1 xformQuery, EntityQuery`1 handsQuery) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Upd
> ate.cs:line 68
> at Content.Shared.DoAfter.SharedDoAfterSystem.Update(Single frameTime) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 23
> at Robust.Shared.GameObjects.EntitySystemManager.TickUpdate(Single frameTime, Boolean noPredictions) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystemManager.cs:line 306
> at Robust.Shared.GameObjects.EntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityManager.cs:line 278
> at Robust.Server.GameObjects.ServerEntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\GameObjects\ServerEntityManager.cs:line 158
> at Robust.Server.BaseServer.Update(FrameEventArgs frameEventArgs) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 719
> at Robust.Server.BaseServer.<SetupMainLoop>b__66_1(Object sender, FrameEventArgs args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 536
> at Robust.Shared.Timing.GameLoop.Run() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\Timing\GameLoop.cs:line 235
> at Robust.Server.BaseServer.MainLoop() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 563
> at Robust.Server.Program.ParsedMain(CommandLineArgs args, Boolean contentStart, ServerOptions options) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 78
> at Robust.Server.Program.Start(String[] args, ServerOptions options, Boolean contentStart) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 46
> at Robust.Server.ContentStart.Start(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\ContentStart.cs:line 10
> at Content.Server.Program.Main(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Server\Program.cs:line 9
| 1.0 | Action Containers broke spellbooks + exception - ## Description
<!-- Explain your issue in detail. Issues without proper explanation are liable to be closed by maintainers. -->
An exception is thrown when trying to learn spells from spellbooks. The recent action container PR broke this behavior: #20260
**Reproduction**
<!-- Include the steps to reproduce if applicable. -->
Pick up a spellbook and try to learn it, console will show the exception.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
**Additional context**
<!-- Add any other context about the problem here. Anything you think is related to the issue. -->
> [ERRO] system.actions: Can't resolve "Content.Shared.Actions.ActionsContainerComponent" on entity 1985!
> at Content.Shared.Actions.SharedActionsSystem.GrantActions(EntityUid performer, IEnumerable`1 actions, EntityUid container, ActionsComponent comp, ActionsContainerComponent containerComp) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Actions\SharedActionsSystem.cs:l
> ine 523
> at Content.Shared.Magic.SharedMagicSystem.OnDoAfter(EntityUid uid, SpellbookComponent component, DoAfterEvent args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Magic\SharedMagicSystem.cs:line 85
> at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass47_0`2.<SubscribeLocalEvent>g__EventHandler|0(EntityUid uid, IComponent comp, TEvent& args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 254
> at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass57_0`1.<EntSubscribe>b__0(EntityUid uid, IComponent comp, Unit& ev) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 387
> at Robust.Shared.GameObjects.EntityEventBus.EntDispatch(EntityUid euid, Type eventType, Unit& args, Boolean dispatchByReference) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 541
> at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEventCore(EntityUid uid, Unit& unitRef, Type type, Boolean broadcast, Boolean byRef) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 221
> at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 190
> at Robust.Shared.GameObjects.EntitySystem.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystem.cs:line 155
> at Content.Shared.DoAfter.SharedDoAfterSystem.RaiseDoAfterEvents(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.cs:line 93
> at Content.Shared.DoAfter.SharedDoAfterSystem.TryComplete(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 123
> at Content.Shared.DoAfter.SharedDoAfterSystem.Update(EntityUid uid, ActiveDoAfterComponent active, DoAfterComponent comp, TimeSpan time, EntityQuery`1 xformQuery, EntityQuery`1 handsQuery) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Upd
> ate.cs:line 68
> at Content.Shared.DoAfter.SharedDoAfterSystem.Update(Single frameTime) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 23
> at Robust.Shared.GameObjects.EntitySystemManager.TickUpdate(Single frameTime, Boolean noPredictions) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystemManager.cs:line 306
> at Robust.Shared.GameObjects.EntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityManager.cs:line 278
> at Robust.Server.GameObjects.ServerEntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\GameObjects\ServerEntityManager.cs:line 158
> at Robust.Server.BaseServer.Update(FrameEventArgs frameEventArgs) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 719
> at Robust.Server.BaseServer.<SetupMainLoop>b__66_1(Object sender, FrameEventArgs args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 536
> at Robust.Shared.Timing.GameLoop.Run() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\Timing\GameLoop.cs:line 235
> at Robust.Server.BaseServer.MainLoop() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 563
> at Robust.Server.Program.ParsedMain(CommandLineArgs args, Boolean contentStart, ServerOptions options) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 78
> at Robust.Server.Program.Start(String[] args, ServerOptions options, Boolean contentStart) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 46
> at Robust.Server.ContentStart.Start(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\ContentStart.cs:line 10
> at Content.Server.Program.Main(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Server\Program.cs:line 9
| priority | action containers broke spellbooks exception description an exception is thrown when trying to learn spells from spellbooks the recent action container pr broke this behavior reproduction pick up a spellbook and try to learn it console will show the exception screenshots additional context system actions can t resolve content shared actions actionscontainercomponent on entity at content shared actions sharedactionssystem grantactions entityuid performer ienumerable actions entityuid container actionscomponent comp actionscontainercomponent containercomp in j program files space station content shared actions sharedactionssystem cs l ine at content shared magic sharedmagicsystem ondoafter entityuid uid spellbookcomponent component doafterevent args in j program files space station content shared magic sharedmagicsystem cs line at robust shared gameobjects entityeventbus c g eventhandler entityuid uid icomponent comp tevent args in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus c b entityuid uid icomponent comp unit ev in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus entdispatch entityuid euid type eventtype unit args boolean dispatchbyreference in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus raiselocaleventcore entityuid uid unit unitref type type boolean broadcast boolean byref in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus raiselocalevent entityuid uid object args boolean broadcast in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entitysystem raiselocalevent entityuid uid object args boolean broadcast in j program files space station robusttoolbox robust shared gameobjects entitysystem cs line at content shared doafter shareddoaftersystem raisedoafterevents doafter doafter doaftercomponent component in j program files space station content shared doafter shareddoaftersystem cs line at content shared doafter shareddoaftersystem trycomplete doafter doafter doaftercomponent component in j program files space station content shared doafter shareddoaftersystem update cs line at content shared doafter shareddoaftersystem update entityuid uid activedoaftercomponent active doaftercomponent comp timespan time entityquery xformquery entityquery handsquery in j program files space station content shared doafter shareddoaftersystem upd ate cs line at content shared doafter shareddoaftersystem update single frametime in j program files space station content shared doafter shareddoaftersystem update cs line at robust shared gameobjects entitysystemmanager tickupdate single frametime boolean nopredictions in j program files space station robusttoolbox robust shared gameobjects entitysystemmanager cs line at robust shared gameobjects entitymanager tickupdate single frametime boolean nopredictions histogram histogram in j program files space station robusttoolbox robust shared gameobjects entitymanager cs line at robust server gameobjects serverentitymanager tickupdate single frametime boolean nopredictions histogram histogram in j program files space station robusttoolbox robust server gameobjects serverentitymanager cs line at robust server baseserver update frameeventargs frameeventargs in j program files space station robusttoolbox robust server baseserver cs line at robust server baseserver b object sender frameeventargs args in j program files space station robusttoolbox robust server baseserver cs line at robust shared timing gameloop run in j program files space station robusttoolbox robust shared timing gameloop cs line at robust server baseserver mainloop in j program files space station robusttoolbox robust server baseserver cs line at robust server program parsedmain commandlineargs args boolean contentstart serveroptions options in j program files space station robusttoolbox robust server program cs line at robust server program start string args serveroptions options boolean contentstart in j program files space station robusttoolbox robust server program cs line at robust server contentstart start string args in j program files space station robusttoolbox robust server contentstart cs line at content server program main string args in j program files space station content server program cs line | 1 |
163,275 | 20,356,090,147 | IssuesEvent | 2022-02-20 00:37:34 | eclipse/che | https://api.github.com/repos/eclipse/che | closed | REST Service on the Che server-side that will manage user secrets | kind/enhancement severity/P1 lifecycle/stale area/dev-experience area/security | ### Is your enhancement related to a problem? Please describe.
Some che users need to have some secrets on personal workspace. Secrets can be:
- github / SCM token
- files that contain secrets
- Environnment variables
Eclipse CHE provide some function to handle this. All of them consume existings secrets from user namespace on K8s (with some labels/annotations).
The problem is that major of user don't have direct access on kubernetes (or knowledge) to create this secrets.
Moreover, is not really user friendly to know all annotations/labels needed (depend of why we want).
### Describe the solution you'd like
Create user secrets on k8s from chectl (like we can manage workspace) and/or from Eclpise Che UI.
This wrapper must handle the right labels/annotions needed (secret as env, secret as file, secret as github token, etc ...)
To do that, we need some new api to handle secrets from Eclpise che server.
### Describe alternatives you've considered
<!--A clear and concise description of any alternative solutions or features you've considered. -->
### Additional context
<!-- Add any other context or screenshots about the enhancement here. -->
| True | REST Service on the Che server-side that will manage user secrets - ### Is your enhancement related to a problem? Please describe.
Some che users need to have some secrets on personal workspace. Secrets can be:
- github / SCM token
- files that contain secrets
- Environnment variables
Eclipse CHE provide some function to handle this. All of them consume existings secrets from user namespace on K8s (with some labels/annotations).
The problem is that major of user don't have direct access on kubernetes (or knowledge) to create this secrets.
Moreover, is not really user friendly to know all annotations/labels needed (depend of why we want).
### Describe the solution you'd like
Create user secrets on k8s from chectl (like we can manage workspace) and/or from Eclpise Che UI.
This wrapper must handle the right labels/annotions needed (secret as env, secret as file, secret as github token, etc ...)
To do that, we need some new api to handle secrets from Eclpise che server.
### Describe alternatives you've considered
<!--A clear and concise description of any alternative solutions or features you've considered. -->
### Additional context
<!-- Add any other context or screenshots about the enhancement here. -->
| non_priority | rest service on the che server side that will manage user secrets is your enhancement related to a problem please describe some che users need to have some secrets on personal workspace secrets can be github scm token files that contain secrets environnment variables eclipse che provide some function to handle this all of them consume existings secrets from user namespace on with some labels annotations the problem is that major of user don t have direct access on kubernetes or knowledge to create this secrets moreover is not really user friendly to know all annotations labels needed depend of why we want describe the solution you d like create user secrets on from chectl like we can manage workspace and or from eclpise che ui this wrapper must handle the right labels annotions needed secret as env secret as file secret as github token etc to do that we need some new api to handle secrets from eclpise che server describe alternatives you ve considered additional context | 0 |
230,674 | 7,612,865,291 | IssuesEvent | 2018-05-01 19:05:06 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | Reordering layers has no effect on map | bug-type: broken use case bug-type: regression priority: high problem: bug | Tested build: http://fgpv.cloudapp.net/demo/develop/prod/samples/index-fgp-en.html
Imported 2 layers (`happy.json` and then `convert.json`) and reordering them had no effect.
`happy.json` always remained below the other layer even if it was above in the legend stack.
Expected behaviour would be to see `happy.json` layer above `convert.json` | 1.0 | Reordering layers has no effect on map - Tested build: http://fgpv.cloudapp.net/demo/develop/prod/samples/index-fgp-en.html
Imported 2 layers (`happy.json` and then `convert.json`) and reordering them had no effect.
`happy.json` always remained below the other layer even if it was above in the legend stack.
Expected behaviour would be to see `happy.json` layer above `convert.json` | priority | reordering layers has no effect on map tested build imported layers happy json and then convert json and reordering them had no effect happy json always remained below the other layer even if it was above in the legend stack expected behaviour would be to see happy json layer above convert json | 1 |
68,398 | 13,127,364,092 | IssuesEvent | 2020-08-06 10:13:02 | hpi-swa-teaching/Algernon-Launcher | https://api.github.com/repos/hpi-swa-teaching/Algernon-Launcher | closed | Cleanup Git(-hub) | DevOps non-code | Make the Git (-hub) ready for the next team.
This includes:
- [x] Remove Merged/stale/dangling Branches
- [x] Archive/Complete Milestones
- [x] Look through open issues add/remove description/labels/etc if needed | 1.0 | Cleanup Git(-hub) - Make the Git (-hub) ready for the next team.
This includes:
- [x] Remove Merged/stale/dangling Branches
- [x] Archive/Complete Milestones
- [x] Look through open issues add/remove description/labels/etc if needed | non_priority | cleanup git hub make the git hub ready for the next team this includes remove merged stale dangling branches archive complete milestones look through open issues add remove description labels etc if needed | 0 |
422,215 | 12,268,448,667 | IssuesEvent | 2020-05-07 12:32:59 | AxonFramework/AxonFramework | https://api.github.com/repos/AxonFramework/AxonFramework | closed | Missing aggregateIdentifier() method reported by an exception | Priority 4: Would Status: In Progress Type: Bug | When I was writing my tests the other day I happened to encounter an `EventStoreException` with message "You probably want to use aggregateIdentifier() on your fixture".
The solution was to create a separate fixture for each independent unit test, but I could not find the `aggregateIdentifier()` method on any of the fixture classes (even though there is a field on `AggregateTestFixture`).
I suggest to fix the exception message not to point to a missing method or make it clearer if I simply was not looking in the right place. | 1.0 | Missing aggregateIdentifier() method reported by an exception - When I was writing my tests the other day I happened to encounter an `EventStoreException` with message "You probably want to use aggregateIdentifier() on your fixture".
The solution was to create a separate fixture for each independent unit test, but I could not find the `aggregateIdentifier()` method on any of the fixture classes (even though there is a field on `AggregateTestFixture`).
I suggest to fix the exception message not to point to a missing method or make it clearer if I simply was not looking in the right place. | priority | missing aggregateidentifier method reported by an exception when i was writing my tests the other day i happened to encounter an eventstoreexception with message you probably want to use aggregateidentifier on your fixture the solution was to create a separate fixture for each independent unit test but i could not find the aggregateidentifier method on any of the fixture classes even though there is a field on aggregatetestfixture i suggest to fix the exception message not to point to a missing method or make it clearer if i simply was not looking in the right place | 1 |
414,126 | 12,099,647,314 | IssuesEvent | 2020-04-20 12:33:18 | rubyforgood/casa | https://api.github.com/repos/rubyforgood/casa | closed | Add contact medium to new case contact success page | :raised_hands: Volunteer Priority: High Status: Available | relates to epic #3
**What type of user is this for? [volunteer/supervisor/admin/all]**
**Where does/should this occur**
On the page a volunteer views after successfully creating a new `case_contact`.
**Description**
_Contact medium_ should display underneath _Contact type:_
**Screenshots**
<img width="1237" alt="volunteer_bug" src="https://user-images.githubusercontent.com/62810851/79702174-b8a7ed00-8257-11ea-8491-5fa83662ddc7.png">
**Relates to issues...**
dependent on #121, relates to #129 | 1.0 | Add contact medium to new case contact success page - relates to epic #3
**What type of user is this for? [volunteer/supervisor/admin/all]**
**Where does/should this occur**
On the page a volunteer views after successfully creating a new `case_contact`.
**Description**
_Contact medium_ should display underneath _Contact type:_
**Screenshots**
<img width="1237" alt="volunteer_bug" src="https://user-images.githubusercontent.com/62810851/79702174-b8a7ed00-8257-11ea-8491-5fa83662ddc7.png">
**Relates to issues...**
dependent on #121, relates to #129 | priority | add contact medium to new case contact success page relates to epic what type of user is this for where does should this occur on the page a volunteer views after successfully creating a new case contact description contact medium should display underneath contact type screenshots img width alt volunteer bug src relates to issues dependent on relates to | 1 |
98,284 | 12,305,950,794 | IssuesEvent | 2020-05-11 23:58:13 | alice-i-cecile/Fonts-of-Power | https://api.github.com/repos/alice-i-cecile/Fonts-of-Power | closed | Rework attacks of opportunity to fix melee stalemate | bug design | 10:03 PM] Alice: Alright, here we go:
- combat in 5e (and FoP) feels really stale because melee combatants don't try to move
- once you're in melee combat, being the first one to blink just gets you smacked in the face, and nothing else
- this is less bad in FoP, because a) Disengage is a minor action b) if you can force them to Block, there's no penalty to backing away
- however, there's still no benefit to doing so by default. They close the gap and smack you again
[10:04 PM] Alice: Proposal:
- creatures provoke an attack of opportunity whenever they enter or leave a hostile creature's threatened zone
[10:05 PM] Alice: This means that the optimal offensive play is actually to skirmish:
1. hit
2. force a Block
3. attack of opportunity when they close the distance again
| 1.0 | Rework attacks of opportunity to fix melee stalemate - 10:03 PM] Alice: Alright, here we go:
- combat in 5e (and FoP) feels really stale because melee combatants don't try to move
- once you're in melee combat, being the first one to blink just gets you smacked in the face, and nothing else
- this is less bad in FoP, because a) Disengage is a minor action b) if you can force them to Block, there's no penalty to backing away
- however, there's still no benefit to doing so by default. They close the gap and smack you again
[10:04 PM] Alice: Proposal:
- creatures provoke an attack of opportunity whenever they enter or leave a hostile creature's threatened zone
[10:05 PM] Alice: This means that the optimal offensive play is actually to skirmish:
1. hit
2. force a Block
3. attack of opportunity when they close the distance again
| non_priority | rework attacks of opportunity to fix melee stalemate pm alice alright here we go combat in and fop feels really stale because melee combatants don t try to move once you re in melee combat being the first one to blink just gets you smacked in the face and nothing else this is less bad in fop because a disengage is a minor action b if you can force them to block there s no penalty to backing away however there s still no benefit to doing so by default they close the gap and smack you again alice proposal creatures provoke an attack of opportunity whenever they enter or leave a hostile creature s threatened zone alice this means that the optimal offensive play is actually to skirmish hit force a block attack of opportunity when they close the distance again | 0 |
427,085 | 29,795,857,404 | IssuesEvent | 2023-06-16 02:14:19 | inventree/InvenTree | https://api.github.com/repos/inventree/InvenTree | closed | units parameters | question report documentation | ### Body of the issue
is there a way to print the units of the parameters in the labels. behind the value i want the unit to print but don't know how
`<p class="text">
{% if parameters.voltage %}
{{parameters.voltage}}
{% endif %}
</p>` | 1.0 | units parameters - ### Body of the issue
is there a way to print the units of the parameters in the labels. behind the value i want the unit to print but don't know how
`<p class="text">
{% if parameters.voltage %}
{{parameters.voltage}}
{% endif %}
</p>` | non_priority | units parameters body of the issue is there a way to print the units of the parameters in the labels behind the value i want the unit to print but don t know how if parameters voltage parameters voltage endif | 0 |
104,600 | 11,415,273,942 | IssuesEvent | 2020-02-02 09:52:53 | engnogueira/webdjango | https://api.github.com/repos/engnogueira/webdjango | closed | 3.5.3 - Implementando um Breadcrumb | documentation | Nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website.
[Implementando um Breadcrumb](https://www.python.pro.br/modulos/django/topicos/implementando-um-breadcrumb)
Link com documentação do Twitter Bootstrap:
https://getbootstrap.com/docs/4.4/components/breadcrumb/ | 1.0 | 3.5.3 - Implementando um Breadcrumb - Nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website.
[Implementando um Breadcrumb](https://www.python.pro.br/modulos/django/topicos/implementando-um-breadcrumb)
Link com documentação do Twitter Bootstrap:
https://getbootstrap.com/docs/4.4/components/breadcrumb/ | non_priority | implementando um breadcrumb nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website link com documentação do twitter bootstrap | 0 |
490,200 | 14,116,621,813 | IssuesEvent | 2020-11-08 04:19:50 | AY2021S1-CS2103-T16-2/tp | https://api.github.com/repos/AY2021S1-CS2103-T16-2/tp | opened | Check Code Quality for Joven's Features | priority.Medium | Will only close when the documentation for code quality improvements are made. | 1.0 | Check Code Quality for Joven's Features - Will only close when the documentation for code quality improvements are made. | priority | check code quality for joven s features will only close when the documentation for code quality improvements are made | 1 |
163,294 | 13,914,727,812 | IssuesEvent | 2020-10-20 22:47:44 | nexusformat/definitions | https://api.github.com/repos/nexusformat/definitions | closed | DOC: ! LaTeX Error: Too deeply nested. | documentation | When building the documentation, this error is reported by the pdflatex build.
It means that we have requested more than six levels of a latex ``/begin{quote}`` (or other) block. The issue could be with how we:
1. extract the rst from the NXDL files
1. format specific documentation (indentation) in the NXDL files
1. format the rst file for each NXDL
We need better code to flag the first two of these conditions. It is not obvious what code produces this error. It is obvious it is coming from the documentation of more than just any one specific NXDL.
Despite the error, a PDF is generated and looks fine.
There is a note in the `Makefile` acknowledging this has been known for a long time. | 1.0 | DOC: ! LaTeX Error: Too deeply nested. - When building the documentation, this error is reported by the pdflatex build.
It means that we have requested more than six levels of a latex ``/begin{quote}`` (or other) block. The issue could be with how we:
1. extract the rst from the NXDL files
1. format specific documentation (indentation) in the NXDL files
1. format the rst file for each NXDL
We need better code to flag the first two of these conditions. It is not obvious what code produces this error. It is obvious it is coming from the documentation of more than just any one specific NXDL.
Despite the error, a PDF is generated and looks fine.
There is a note in the `Makefile` acknowledging this has been known for a long time. | non_priority | doc latex error too deeply nested when building the documentation this error is reported by the pdflatex build it means that we have requested more than six levels of a latex begin quote or other block the issue could be with how we extract the rst from the nxdl files format specific documentation indentation in the nxdl files format the rst file for each nxdl we need better code to flag the first two of these conditions it is not obvious what code produces this error it is obvious it is coming from the documentation of more than just any one specific nxdl despite the error a pdf is generated and looks fine there is a note in the makefile acknowledging this has been known for a long time | 0 |
65,983 | 19,846,070,635 | IssuesEvent | 2022-01-21 06:31:41 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | opened | Cannot share a video from iOS | T-Defect | ### Steps to reproduce
Steps to reproduce:
1. Have a video present in chat with someone
2. Click on the video and hold
3. Click more in the bottom left corner
4. Select share and select a user to send it to someone
### Outcome
Expected the video to be shared to the other user.
Actual result: The process gets stuck in the sending phase

### Your phone model
iPad 7th Generation
### Operating system version
iOS 14.7.1
### Application version
Element 1.16.12
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Cannot share a video from iOS - ### Steps to reproduce
Steps to reproduce:
1. Have a video present in chat with someone
2. Click on the video and hold
3. Click more in the bottom left corner
4. Select share and select a user to send it to someone
### Outcome
Expected the video to be shared to the other user.
Actual result: The process gets stuck in the sending phase

### Your phone model
iPad 7th Generation
### Operating system version
iOS 14.7.1
### Application version
Element 1.16.12
### Homeserver
_No response_
### Will you send logs?
No | non_priority | cannot share a video from ios steps to reproduce steps to reproduce have a video present in chat with someone click on the video and hold click more in the bottom left corner select share and select a user to send it to someone outcome expected the video to be shared to the other user actual result the process gets stuck in the sending phase your phone model ipad generation operating system version ios application version element homeserver no response will you send logs no | 0 |
12,630 | 4,513,229,528 | IssuesEvent | 2016-09-04 05:27:52 | Jeremy-Barnes/Critters | https://api.github.com/repos/Jeremy-Barnes/Critters | opened | Server: Create Landmark/NPC | Code feature Server | Create endpoints to allow:
Random dialog fetching for distinct pages (NPCs)
Quest management
| 1.0 | Server: Create Landmark/NPC - Create endpoints to allow:
Random dialog fetching for distinct pages (NPCs)
Quest management
| non_priority | server create landmark npc create endpoints to allow random dialog fetching for distinct pages npcs quest management | 0 |
14,164 | 3,807,048,174 | IssuesEvent | 2016-03-25 04:33:06 | stevegrunwell/mcavoy | https://api.github.com/repos/stevegrunwell/mcavoy | closed | Introduce a formal change log | documentation | Once the plugin gets past 0.1.0, a formal change log should be kept based on the [Keep a Changelog standard](http://keepachangelog.com/). | 1.0 | Introduce a formal change log - Once the plugin gets past 0.1.0, a formal change log should be kept based on the [Keep a Changelog standard](http://keepachangelog.com/). | non_priority | introduce a formal change log once the plugin gets past a formal change log should be kept based on the | 0 |
617,407 | 19,349,655,247 | IssuesEvent | 2021-12-15 14:29:24 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | att-yahoo.att.net - see bug description | browser-firefox-mobile priority-normal engine-gecko | <!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/96773 -->
**URL**: https://att-yahoo.att.net/FIM/sps/auth?SPRelayState=https%3A%2F%2Fmail.yahoo.com%2Fd%2F&FedId=uuidd0e66c67-0170-1fb9-80b8-ca3a6bdd2cf8&PartnerId=https%3A%2F%2Flogin.yahoo.com%2Fsaml2&FedName=ATT-Yahoo
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: has timed out for about 8hrs now.
**Steps to Reproduce**:
Keeps saying Timed out. I turned off ad blockers Google search fix updated Ff for Android restarted phone.
I just tried it with Chrome and it did the best thing.
ERR_CONNECTION_TIMED_OUT
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/6d4c6800-5f09-4c07-a171-cd6da09e8aaf.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | att-yahoo.att.net - see bug description - <!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/96773 -->
**URL**: https://att-yahoo.att.net/FIM/sps/auth?SPRelayState=https%3A%2F%2Fmail.yahoo.com%2Fd%2F&FedId=uuidd0e66c67-0170-1fb9-80b8-ca3a6bdd2cf8&PartnerId=https%3A%2F%2Flogin.yahoo.com%2Fsaml2&FedName=ATT-Yahoo
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: has timed out for about 8hrs now.
**Steps to Reproduce**:
Keeps saying Timed out. I turned off ad blockers Google search fix updated Ff for Android restarted phone.
I just tried it with Chrome and it did the best thing.
ERR_CONNECTION_TIMED_OUT
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/6d4c6800-5f09-4c07-a171-cd6da09e8aaf.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | att yahoo att net see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description has timed out for about now steps to reproduce keeps saying timed out i turned off ad blockers google search fix updated ff for android restarted phone i just tried it with chrome and it did the best thing err connection timed out view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
568,720 | 16,987,209,797 | IssuesEvent | 2021-06-30 15:37:39 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | NR Extension Date not updated in NRO | ENTITY OPS Priority1 | ServiceNow incident: INC0098013
Contact information
Staff Name: Debbie Blythe
Staff Email:
Description
NR Extension has been applied in NAMEX but is not updated in NRO. Update Extension Date in NRO to July 31, 2021.
Email from Client:
Patrick can you create a high priority bug for this please that needs to be worked on today.
Kaine Sparks
Director, Digital Operations| BC Registries and Online Services
Service BC
Ministry of Citizens’ Services
T: 250-415-3957 | Web: http://www.servicebc.gov.bc.ca
“Access to government services made easy”
From: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>
Sent: June 25, 2021 10:15 AM
To: SBC IT Operations Support, CITZ <SBC_ITOperationsSupport@gov.bc.ca>
Cc: Sparks, Kaine CITZ:EX <Kaine.Sparks@gov.bc.ca>; McClung, Linda CITZ:EX <Linda.McClung@gov.bc.ca>; Sienna Blumstengel <sienna@freshworks.io>
Subject: FW: Name Requests and Extensions - 5234-7
Importance: High
Client resubmitted the following NR’s and the expiry date in Namex has been extended to July 31, 2021. However NRO is still showing Jun 5, 2021. Can this please get fixed, thanks.
1. Cheslatta Marine Services Limited Partnership - NR 5252553
2. Cheslatta Marine Services GP Ltd. – NR 3664932
3. Cheslatta Contracting GP Ltd. - NR 2528980
4. Cheslatta Contracting Limited Partnership – NR 7383745
Debbie
From: Melanie Starchuk <Melanie_Starchuk@aboriginal-law.com>
Sent: June 22, 2021 1:50 PM
To: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>; BCOLACCOUNTS CITZ:EX <BCOLACCOUNTS@gov.bc.ca>
Cc: Devrin Froese <Devrin_Froese@aboriginal-law.com>
Subject: Name Requests and Extensions - 5234-7
Importance: High
[EXTERNAL] This email came from an external source. Only open attachments or links that you are expecting from a known sender.
Hello,
I submitted the below name requests on March 31st, 2021 and extended the requests on June 4, 2021.
The attached four PDF’s include the receipts of the original requests and the extension receipts. Online it says the extensions expired in June, which shouldn’t be the case because we paid for the name reservation extensions.
1. Cheslatta Marine Services Limited Partnership - NR 5252553
2. Cheslatta Marine Services GP Ltd. – NR 3664932
3. Cheslatta Contracting GP Ltd. - NR 2528980
4. Cheslatta Contracting Limited Partnership – NR 7383745
Our BC OnLine Account number is 862583.
The new extensions should be until July 31, 2021. I did not receive a new name reservation number. Have there been new name reservation numbers created?
Thanks,
Melanie
Melanie Starchuk
Paralegal
Donovan & Company
Barristers and Solicitors
6th Floor, 73 Water Street
Vancouver, B.C. V6B 1A1
Tel: (604) 688-4272
Fax: (604) 688-4282
E-mail: melanie_starchuk@aboriginal-law.com
Website: www.aboriginal-law.com
***IMPORTANT NOTICE***
The contents of this e-mail, including attachments, are strictly confidential and are intended only for the use of the individual or entity to which this e-mail is addressed. If you are not the intended recipient, or an employee or agent responsible for delivering the message to the intended recipient, you are hereby notified that any dissemination, distribution of copying this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by telephone (604-688-4272).
| 1.0 | NR Extension Date not updated in NRO - ServiceNow incident: INC0098013
Contact information
Staff Name: Debbie Blythe
Staff Email:
Description
NR Extension has been applied in NAMEX but is not updated in NRO. Update Extension Date in NRO to July 31, 2021.
Email from Client:
Patrick can you create a high priority bug for this please that needs to be worked on today.
Kaine Sparks
Director, Digital Operations| BC Registries and Online Services
Service BC
Ministry of Citizens’ Services
T: 250-415-3957 | Web: http://www.servicebc.gov.bc.ca
“Access to government services made easy”
From: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>
Sent: June 25, 2021 10:15 AM
To: SBC IT Operations Support, CITZ <SBC_ITOperationsSupport@gov.bc.ca>
Cc: Sparks, Kaine CITZ:EX <Kaine.Sparks@gov.bc.ca>; McClung, Linda CITZ:EX <Linda.McClung@gov.bc.ca>; Sienna Blumstengel <sienna@freshworks.io>
Subject: FW: Name Requests and Extensions - 5234-7
Importance: High
Client resubmitted the following NR’s and the expiry date in Namex has been extended to July 31, 2021. However NRO is still showing Jun 5, 2021. Can this please get fixed, thanks.
1. Cheslatta Marine Services Limited Partnership - NR 5252553
2. Cheslatta Marine Services GP Ltd. – NR 3664932
3. Cheslatta Contracting GP Ltd. - NR 2528980
4. Cheslatta Contracting Limited Partnership – NR 7383745
Debbie
From: Melanie Starchuk <Melanie_Starchuk@aboriginal-law.com>
Sent: June 22, 2021 1:50 PM
To: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>; BCOLACCOUNTS CITZ:EX <BCOLACCOUNTS@gov.bc.ca>
Cc: Devrin Froese <Devrin_Froese@aboriginal-law.com>
Subject: Name Requests and Extensions - 5234-7
Importance: High
[EXTERNAL] This email came from an external source. Only open attachments or links that you are expecting from a known sender.
Hello,
I submitted the below name requests on March 31st, 2021 and extended the requests on June 4, 2021.
The attached four PDF’s include the receipts of the original requests and the extension receipts. Online it says the extensions expired in June, which shouldn’t be the case because we paid for the name reservation extensions.
1. Cheslatta Marine Services Limited Partnership - NR 5252553
2. Cheslatta Marine Services GP Ltd. – NR 3664932
3. Cheslatta Contracting GP Ltd. - NR 2528980
4. Cheslatta Contracting Limited Partnership – NR 7383745
Our BC OnLine Account number is 862583.
The new extensions should be until July 31, 2021. I did not receive a new name reservation number. Have there been new name reservation numbers created?
Thanks,
Melanie
Melanie Starchuk
Paralegal
Donovan & Company
Barristers and Solicitors
6th Floor, 73 Water Street
Vancouver, B.C. V6B 1A1
Tel: (604) 688-4272
Fax: (604) 688-4282
E-mail: melanie_starchuk@aboriginal-law.com
Website: www.aboriginal-law.com
***IMPORTANT NOTICE***
The contents of this e-mail, including attachments, are strictly confidential and are intended only for the use of the individual or entity to which this e-mail is addressed. If you are not the intended recipient, or an employee or agent responsible for delivering the message to the intended recipient, you are hereby notified that any dissemination, distribution of copying this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by telephone (604-688-4272).
| priority | nr extension date not updated in nro servicenow incident contact information staff name debbie blythe staff email description nr extension has been applied in namex but is not updated in nro update extension date in nro to july email from client patrick can you create a high priority bug for this please that needs to be worked on today kaine sparks director digital operations bc registries and online services service bc ministry of citizens’ services t web “access to government services made easy” from bc registry services citz ex sent june am to sbc it operations support citz cc sparks kaine citz ex mcclung linda citz ex sienna blumstengel subject fw name requests and extensions importance high client resubmitted the following nr’s and the expiry date in namex has been extended to july however nro is still showing jun can this please get fixed thanks cheslatta marine services limited partnership nr cheslatta marine services gp ltd – nr cheslatta contracting gp ltd nr cheslatta contracting limited partnership – nr debbie from melanie starchuk sent june pm to bc registry services citz ex bcolaccounts citz ex cc devrin froese subject name requests and extensions importance high this email came from an external source only open attachments or links that you are expecting from a known sender hello i submitted the below name requests on march and extended the requests on june the attached four pdf’s include the receipts of the original requests and the extension receipts online it says the extensions expired in june which shouldn’t be the case because we paid for the name reservation extensions cheslatta marine services limited partnership nr cheslatta marine services gp ltd – nr cheslatta contracting gp ltd nr cheslatta contracting limited partnership – nr our bc online account number is the new extensions should be until july i did not receive a new name reservation number have there been new name reservation numbers created thanks melanie melanie starchuk paralegal donovan company barristers and solicitors floor water street vancouver b c tel fax e mail melanie starchuk aboriginal law com website important notice the contents of this e mail including attachments are strictly confidential and are intended only for the use of the individual or entity to which this e mail is addressed if you are not the intended recipient or an employee or agent responsible for delivering the message to the intended recipient you are hereby notified that any dissemination distribution of copying this communication is strictly prohibited if you have received this communication in error please notify us immediately by telephone | 1 |
820,370 | 30,770,357,905 | IssuesEvent | 2023-07-30 20:43:33 | CueToDo/Free-Vote-SPA | https://api.github.com/repos/CueToDo/Free-Vote-SPA | closed | Remove HTML encoding from website preview meta data description | bug high priority | https://free.vote/Money/merchant-bankers-and-charity
Plus we don't need the url and the actual link | 1.0 | Remove HTML encoding from website preview meta data description - https://free.vote/Money/merchant-bankers-and-charity
Plus we don't need the url and the actual link | priority | remove html encoding from website preview meta data description plus we don t need the url and the actual link | 1 |
37,740 | 8,357,213,237 | IssuesEvent | 2018-10-02 20:51:11 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | opened | Add tests for MethodImplOptions.AggressiveOptimization | area-CodeVersioning | - Depends on https://github.com/dotnet/coreclr/pull/20009
- Add test that verifies that the attribute causes the method to be jitted only once at tier 1
- See https://github.com/dotnet/coreclr/pull/20009#discussion_r221808280
- Add perf test from issue (https://github.com/dotnet/coreclr/issues/19751) to performance repo | 1.0 | Add tests for MethodImplOptions.AggressiveOptimization - - Depends on https://github.com/dotnet/coreclr/pull/20009
- Add test that verifies that the attribute causes the method to be jitted only once at tier 1
- See https://github.com/dotnet/coreclr/pull/20009#discussion_r221808280
- Add perf test from issue (https://github.com/dotnet/coreclr/issues/19751) to performance repo | non_priority | add tests for methodimploptions aggressiveoptimization depends on add test that verifies that the attribute causes the method to be jitted only once at tier see add perf test from issue to performance repo | 0 |
1,478 | 2,514,730,315 | IssuesEvent | 2015-01-15 14:03:16 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | opened | Default Country Page: map | Priority-Medium | Blocked by #2102 and #2103
Annotated (very annotated) design is here:
https://docs.google.com/drawings/d/1qOBKZ7IO7zkEMHh2g3ZiAynh5PiAbO-_-SL4rd9uO_M/edit
Implement the map section | 1.0 | Default Country Page: map - Blocked by #2102 and #2103
Annotated (very annotated) design is here:
https://docs.google.com/drawings/d/1qOBKZ7IO7zkEMHh2g3ZiAynh5PiAbO-_-SL4rd9uO_M/edit
Implement the map section | priority | default country page map blocked by and annotated very annotated design is here implement the map section | 1 |
257,581 | 8,139,280,981 | IssuesEvent | 2018-08-20 17:11:27 | nprapps/elections18-general | https://api.github.com/repos/nprapps/elections18-general | closed | Check in on admin panel | effort:medium priority:high | We'll need to touch up our logic around "who's won the chamber" and handle a few other related aspects, such as whether it can show two Senate races in a single state. | 1.0 | Check in on admin panel - We'll need to touch up our logic around "who's won the chamber" and handle a few other related aspects, such as whether it can show two Senate races in a single state. | priority | check in on admin panel we ll need to touch up our logic around who s won the chamber and handle a few other related aspects such as whether it can show two senate races in a single state | 1 |
612,878 | 19,058,331,235 | IssuesEvent | 2021-11-26 01:40:40 | micronaut-projects/micronaut-kafka | https://api.github.com/repos/micronaut-projects/micronaut-kafka | closed | Micronaut Kafka Listener commits messages at shutdown that should not be commited with offset strategy SYNC | priority: high | With OffsetStrategy set to SYNC or ASYNC the KafkaListner should not commit messages when an Exception occurs. However during shutdown all messages get committed regardless if there was an Exception or not.
### Task List
- [X ] Steps to reproduce provided
- [x] Stacktrace (if present) provided
- [x] Example that reproduces the problem uploaded to Github
- [ X] Full description of the issue provided (see below)
### Steps to Reproduce
1. Have an Micronaut application with a KafkaListener in SYNC mode listening to a Kafka topic.
2. Submit an event to this topic.
3. Start process the event but have an Exception thrown during processing (like a SqlException during a database operation).
4. You now have one message lag on your consumer group as you can validate on Kafka. (as documented)
5. Stop the application
6. You now have a zero message lag on your consumer group as you can validate on Kafka.
### Expected Behaviour
The KafkaListener should not commit messages with offset strategy SYNC (or ASYNC) at shutdown time that could not be processed due to an Exception.
### Actual Behaviour
All events get committed. Only with offset strategy DISABLED this set is skipped.
| 1.0 | Micronaut Kafka Listener commits messages at shutdown that should not be commited with offset strategy SYNC - With OffsetStrategy set to SYNC or ASYNC the KafkaListner should not commit messages when an Exception occurs. However during shutdown all messages get committed regardless if there was an Exception or not.
### Task List
- [X ] Steps to reproduce provided
- [x] Stacktrace (if present) provided
- [x] Example that reproduces the problem uploaded to Github
- [ X] Full description of the issue provided (see below)
### Steps to Reproduce
1. Have an Micronaut application with a KafkaListener in SYNC mode listening to a Kafka topic.
2. Submit an event to this topic.
3. Start process the event but have an Exception thrown during processing (like a SqlException during a database operation).
4. You now have one message lag on your consumer group as you can validate on Kafka. (as documented)
5. Stop the application
6. You now have a zero message lag on your consumer group as you can validate on Kafka.
### Expected Behaviour
The KafkaListener should not commit messages with offset strategy SYNC (or ASYNC) at shutdown time that could not be processed due to an Exception.
### Actual Behaviour
All events get committed. Only with offset strategy DISABLED this set is skipped.
| priority | micronaut kafka listener commits messages at shutdown that should not be commited with offset strategy sync with offsetstrategy set to sync or async the kafkalistner should not commit messages when an exception occurs however during shutdown all messages get committed regardless if there was an exception or not task list steps to reproduce provided stacktrace if present provided example that reproduces the problem uploaded to github full description of the issue provided see below steps to reproduce have an micronaut application with a kafkalistener in sync mode listening to a kafka topic submit an event to this topic start process the event but have an exception thrown during processing like a sqlexception during a database operation you now have one message lag on your consumer group as you can validate on kafka as documented stop the application you now have a zero message lag on your consumer group as you can validate on kafka expected behaviour the kafkalistener should not commit messages with offset strategy sync or async at shutdown time that could not be processed due to an exception actual behaviour all events get committed only with offset strategy disabled this set is skipped | 1 |
58,155 | 3,087,857,656 | IssuesEvent | 2015-08-25 14:05:15 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | Плохо работает разбор вложенных BB-кодов | bug imported Priority-Medium | _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 09, 2013 00:47:35_
Точно не работает
1. Все тэги стиля текста вложенные в тэг цвета
2. Не работает цвет вложенный в другой цвет
3. отображение стилей зависит от порядка применения тэгов
[i][b]болд италик[/b][/i] - отображает только курсив
[b][i]италик болд[/i][/b] - показывает жирный курсив
Кто еще что найдет, дополняйте.
Но лечить надо не указанные ошибки по-отдельности а весь алгоритм разбора который сводится
к разбиению строки на участки и назначении каждому участку правильных атрибутов (я насчитал 5 штук) с учетом вложения тэгов (например с применением какого-то стекового принципа)
**Attachment:** [Fly_r15987_BBcodes.png Fly_r15988_BBcodes.png](http://code.google.com/p/flylinkdc/issues/detail?id=1391)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1391_ | 1.0 | Плохо работает разбор вложенных BB-кодов - _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 09, 2013 00:47:35_
Точно не работает
1. Все тэги стиля текста вложенные в тэг цвета
2. Не работает цвет вложенный в другой цвет
3. отображение стилей зависит от порядка применения тэгов
[i][b]болд италик[/b][/i] - отображает только курсив
[b][i]италик болд[/i][/b] - показывает жирный курсив
Кто еще что найдет, дополняйте.
Но лечить надо не указанные ошибки по-отдельности а весь алгоритм разбора который сводится
к разбиению строки на участки и назначении каждому участку правильных атрибутов (я насчитал 5 штук) с учетом вложения тэгов (например с применением какого-то стекового принципа)
**Attachment:** [Fly_r15987_BBcodes.png Fly_r15988_BBcodes.png](http://code.google.com/p/flylinkdc/issues/detail?id=1391)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1391_ | priority | плохо работает разбор вложенных bb кодов from on november точно не работает все тэги стиля текста вложенные в тэг цвета не работает цвет вложенный в другой цвет отображение стилей зависит от порядка применения тэгов болд италик отображает только курсив италик болд показывает жирный курсив кто еще что найдет дополняйте но лечить надо не указанные ошибки по отдельности а весь алгоритм разбора который сводится к разбиению строки на участки и назначении каждому участку правильных атрибутов я насчитал штук с учетом вложения тэгов например с применением какого то стекового принципа attachment original issue | 1 |
27,666 | 30,056,634,161 | IssuesEvent | 2023-06-28 07:20:44 | Open-EO/openeo-python-client | https://api.github.com/repos/Open-EO/openeo-python-client | closed | RESTJob UX | clean up usability v1.0.0 | make `RESTJob` prettier to work with, e.g.:
- The `REST` prefix is bit unusual and redundant: most other user-facing openeo classes don't have this prefix even though they also are about REST abstractions. Just `Job` or `BatchJob` will look prettier for the general user
- rename `start_job()` (and alike) to just `start()` | True | RESTJob UX - make `RESTJob` prettier to work with, e.g.:
- The `REST` prefix is bit unusual and redundant: most other user-facing openeo classes don't have this prefix even though they also are about REST abstractions. Just `Job` or `BatchJob` will look prettier for the general user
- rename `start_job()` (and alike) to just `start()` | non_priority | restjob ux make restjob prettier to work with e g the rest prefix is bit unusual and redundant most other user facing openeo classes don t have this prefix even though they also are about rest abstractions just job or batchjob will look prettier for the general user rename start job and alike to just start | 0 |
312,338 | 26,858,287,990 | IssuesEvent | 2023-02-03 16:14:09 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix tensor.test_torch_special_long | PyTorch Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
| 1.0 | Fix tensor.test_torch_special_long - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars
2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long(
2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]),
2023-01-22T00:58:45.7935777Z E init_num_positional_args=0,
2023-01-22T00:58:45.7936033Z E method_num_positional_args=0,
2023-01-22T00:58:45.7936278Z E as_variable=[False],
2023-01-22T00:58:45.7936511Z E native_array=[False],
2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'),
2023-01-22T00:58:45.7938047Z E frontend='torch',
2023-01-22T00:58:45.7938250Z E )
2023-01-22T00:58:45.7938414Z E
2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case
</details>
| non_priority | fix tensor test torch special long tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 0 |
279,904 | 8,675,466,654 | IssuesEvent | 2018-11-30 10:57:58 | supergiant/control | https://api.github.com/repos/supergiant/control | opened | Current UI do not allow to leave SubnetID empty to let SG create Subnet | High Priority Needs Standup Review aws | When provision cluster on AWS there is no way to create security group for cluster.
New security group gets created only when SubnetID is empty, but it is prohibited on the UI to have this
field empty. This forces user to use either default or existing subnet which is bad for deletion of resources - dependencies between those resources are uncontrolled. | 1.0 | Current UI do not allow to leave SubnetID empty to let SG create Subnet - When provision cluster on AWS there is no way to create security group for cluster.
New security group gets created only when SubnetID is empty, but it is prohibited on the UI to have this
field empty. This forces user to use either default or existing subnet which is bad for deletion of resources - dependencies between those resources are uncontrolled. | priority | current ui do not allow to leave subnetid empty to let sg create subnet when provision cluster on aws there is no way to create security group for cluster new security group gets created only when subnetid is empty but it is prohibited on the ui to have this field empty this forces user to use either default or existing subnet which is bad for deletion of resources dependencies between those resources are uncontrolled | 1 |
133,273 | 18,289,857,560 | IssuesEvent | 2021-10-05 14:13:20 | finos/symphony-wdk | https://api.github.com/repos/finos/symphony-wdk | opened | CVE-2021-35516 (High) detected in commons-compress-1.20.jar | security vulnerability | ## CVE-2021-35516 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary>
<p>Apache Commons Compress software defines an API for working with
compression and archive formats. These include: bzip2, gzip, pack200,
lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4,
Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p>
<p>Path to dependency file: symphony-wdk/workflow-bot-app/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p>
<p>
Dependency Hierarchy:
- camunda-bpm-spring-boot-starter-7.15.0.jar (Root Library)
- spring-boot-loader-tools-2.5.4.jar
- :x: **commons-compress-1.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/finos/symphony-wdk/commit/004d6ff32b56fa8739f47862c8544f6270c29183">004d6ff32b56fa8739f47862c8544f6270c29183</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' sevenz package.
<p>Publish Date: 2021-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516>CVE-2021-35516</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p>
<p>Release Date: 2021-07-13</p>
<p>Fix Resolution: org.apache.commons:commons-compress:1.21</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.commons","packageName":"commons-compress","packageVersion":"1.20","packageFilePaths":["/workflow-bot-app/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.camunda.bpm.springboot:camunda-bpm-spring-boot-starter:7.15.0;org.springframework.boot:spring-boot-loader-tools:2.5.4;org.apache.commons:commons-compress:1.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-compress:1.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-35516","vulnerabilityDetails":"When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress\u0027 sevenz package.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-35516 (High) detected in commons-compress-1.20.jar - ## CVE-2021-35516 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary>
<p>Apache Commons Compress software defines an API for working with
compression and archive formats. These include: bzip2, gzip, pack200,
lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4,
Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p>
<p>Path to dependency file: symphony-wdk/workflow-bot-app/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p>
<p>
Dependency Hierarchy:
- camunda-bpm-spring-boot-starter-7.15.0.jar (Root Library)
- spring-boot-loader-tools-2.5.4.jar
- :x: **commons-compress-1.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/finos/symphony-wdk/commit/004d6ff32b56fa8739f47862c8544f6270c29183">004d6ff32b56fa8739f47862c8544f6270c29183</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' sevenz package.
<p>Publish Date: 2021-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516>CVE-2021-35516</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p>
<p>Release Date: 2021-07-13</p>
<p>Fix Resolution: org.apache.commons:commons-compress:1.21</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.commons","packageName":"commons-compress","packageVersion":"1.20","packageFilePaths":["/workflow-bot-app/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.camunda.bpm.springboot:camunda-bpm-spring-boot-starter:7.15.0;org.springframework.boot:spring-boot-loader-tools:2.5.4;org.apache.commons:commons-compress:1.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-compress:1.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-35516","vulnerabilityDetails":"When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress\u0027 sevenz package.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in commons compress jar cve high severity vulnerability vulnerable library commons compress jar apache commons compress software defines an api for working with compression and archive formats these include gzip lzma xz snappy traditional unix compress deflate brotli zstandard and ar cpio jar tar zip dump arj library home page a href path to dependency file symphony wdk workflow bot app build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons compress commons compress jar dependency hierarchy camunda bpm spring boot starter jar root library spring boot loader tools jar x commons compress jar vulnerable library found in head commit a href found in base branch master vulnerability details when reading a specially crafted archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress sevenz package publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons compress isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org camunda bpm springboot camunda bpm spring boot starter org springframework boot spring boot loader tools org apache commons commons compress isminimumfixversionavailable true minimumfixversion org apache commons commons compress basebranches vulnerabilityidentifier cve vulnerabilitydetails when reading a specially crafted archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress sevenz package vulnerabilityurl | 0 |
286,109 | 24,719,918,872 | IssuesEvent | 2022-10-20 09:52:12 | callstack/react-native-testing-library | https://api.github.com/repos/callstack/react-native-testing-library | closed | Supports accessibilityStates in ByRole queries | compat: testing-library a11y | ## Describe the Feature
Like in `@testing-library/dom` (see [docs](https://testing-library.com/docs/queries/byrole#api)), we should support accessibility states when querying by role.
The accessibilityStates existing are
- `disabled`: Indicates whether the element is disabled or not
- `selected`: Indicates whether a selectable element is currently selected or not
- `checked`: Indicates the state of a checkable element.
- `busy`: Indicates whether an element is currently busy or not
- `expanded`: Indicates whether an expandable element is currently expanded or collapsed
## Possible Implementations
We could like in DTL check which states correspond to which role and throw. But I'd treat that as a bonus and not a necessity.
## Related Issues
#1133
| 1.0 | Supports accessibilityStates in ByRole queries - ## Describe the Feature
Like in `@testing-library/dom` (see [docs](https://testing-library.com/docs/queries/byrole#api)), we should support accessibility states when querying by role.
The accessibilityStates existing are
- `disabled`: Indicates whether the element is disabled or not
- `selected`: Indicates whether a selectable element is currently selected or not
- `checked`: Indicates the state of a checkable element.
- `busy`: Indicates whether an element is currently busy or not
- `expanded`: Indicates whether an expandable element is currently expanded or collapsed
## Possible Implementations
We could like in DTL check which states correspond to which role and throw. But I'd treat that as a bonus and not a necessity.
## Related Issues
#1133
| non_priority | supports accessibilitystates in byrole queries describe the feature like in testing library dom see we should support accessibility states when querying by role the accessibilitystates existing are disabled indicates whether the element is disabled or not selected indicates whether a selectable element is currently selected or not checked indicates the state of a checkable element busy indicates whether an element is currently busy or not expanded indicates whether an expandable element is currently expanded or collapsed possible implementations we could like in dtl check which states correspond to which role and throw but i d treat that as a bonus and not a necessity related issues | 0 |
111,332 | 11,729,693,114 | IssuesEvent | 2020-03-10 19:52:44 | kubernetes-sigs/kubebuilder | https://api.github.com/repos/kubernetes-sigs/kubebuilder | closed | Kubebuilder book: Why create a custom metadata field "controller" in job instead of just ownerReferences | kind/documentation triage/support | I was going through the tutorial in the book for Cronjobs. On [this page](https://book.kubebuilder.io/cronjob-tutorial/controller-implementation.html#2-list-all-active-jobs-and-update-the-status) the concept of a `jobOwnerKey` field is introduced:
```golang
var (
jobOwnerKey = ".metadata.controller"
```
I don't completely understand why we cannot just rely on the build-in ownerReferences type? Is this because it is not indexable by the cache FieldIndexer and we use the jobOwnerKey solely as an indexable id for it?
**Proposals if this is the case:**
- In the book documentation, explain the function of the jobOwnerKey (If you explain it here, I can also PR it)
- If lists of objects like ownerReferences are not indexable, perhaps start a proposal to make them so?
P.S: I didn't know we could add custom metadata to resources? I assume the jobOwnerKey is only used by the local cache then. | 1.0 | Kubebuilder book: Why create a custom metadata field "controller" in job instead of just ownerReferences - I was going through the tutorial in the book for Cronjobs. On [this page](https://book.kubebuilder.io/cronjob-tutorial/controller-implementation.html#2-list-all-active-jobs-and-update-the-status) the concept of a `jobOwnerKey` field is introduced:
```golang
var (
jobOwnerKey = ".metadata.controller"
```
I don't completely understand why we cannot just rely on the build-in ownerReferences type? Is this because it is not indexable by the cache FieldIndexer and we use the jobOwnerKey solely as an indexable id for it?
**Proposals if this is the case:**
- In the book documentation, explain the function of the jobOwnerKey (If you explain it here, I can also PR it)
- If lists of objects like ownerReferences are not indexable, perhaps start a proposal to make them so?
P.S: I didn't know we could add custom metadata to resources? I assume the jobOwnerKey is only used by the local cache then. | non_priority | kubebuilder book why create a custom metadata field controller in job instead of just ownerreferences i was going through the tutorial in the book for cronjobs on the concept of a jobownerkey field is introduced golang var jobownerkey metadata controller i don t completely understand why we cannot just rely on the build in ownerreferences type is this because it is not indexable by the cache fieldindexer and we use the jobownerkey solely as an indexable id for it proposals if this is the case in the book documentation explain the function of the jobownerkey if you explain it here i can also pr it if lists of objects like ownerreferences are not indexable perhaps start a proposal to make them so p s i didn t know we could add custom metadata to resources i assume the jobownerkey is only used by the local cache then | 0 |
175,905 | 21,365,609,718 | IssuesEvent | 2022-04-20 01:01:26 | improbable/k8s-test-infra | https://api.github.com/repos/improbable/k8s-test-infra | closed | moment-2.24.0.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-24785](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | moment-2.24.0.tgz | Direct | 2.29.2 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-24785</summary>
### Vulnerable Library - <b>moment-2.24.0.tgz</b></p>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment/package.json</p>
<p>
Dependency Hierarchy:
- :x: **moment-2.24.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>improbable</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.
<p>Publish Date: 2022-04-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785>CVE-2022-24785</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p>
<p>Release Date: 2022-04-04</p>
<p>Fix Resolution: 2.29.2</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"moment","packageVersion":"2.24.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"moment:2.24.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.29.2","isBinary":false}],"baseBranches":["improbable"],"vulnerabilityIdentifier":"CVE-2022-24785","vulnerabilityDetails":"Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}]</REMEDIATE> --> | True | moment-2.24.0.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-24785](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | moment-2.24.0.tgz | Direct | 2.29.2 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-24785</summary>
### Vulnerable Library - <b>moment-2.24.0.tgz</b></p>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment/package.json</p>
<p>
Dependency Hierarchy:
- :x: **moment-2.24.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>improbable</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.
<p>Publish Date: 2022-04-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785>CVE-2022-24785</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p>
<p>Release Date: 2022-04-04</p>
<p>Fix Resolution: 2.29.2</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"moment","packageVersion":"2.24.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"moment:2.24.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.29.2","isBinary":false}],"baseBranches":["improbable"],"vulnerabilityIdentifier":"CVE-2022-24785","vulnerabilityDetails":"Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}]</REMEDIATE> --> | non_priority | moment tgz vulnerabilities highest severity is autoclosed vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file package json path to vulnerable library node modules moment package json vulnerabilities cve severity cvss dependency type fixed in remediation available high moment tgz direct details cve vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file package json path to vulnerable library node modules moment package json dependency hierarchy x moment tgz vulnerable library found in base branch improbable vulnerability details moment js is a javascript date library for parsing validating manipulating and formatting dates a path traversal vulnerability impacts npm server users of moment js between versions and especially if a user provided locale string is directly used to switch moment locale this problem is patched in and the patch can be applied to all affected versions as a workaround sanitize the user provided locale name before passing it to moment js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue istransitivedependency false dependencytree moment isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails moment js is a javascript date library for parsing validating manipulating and formatting dates a path traversal vulnerability impacts npm server users of moment js between versions and especially if a user provided locale string is directly used to switch moment locale this problem is patched in and the patch can be applied to all affected versions as a workaround sanitize the user provided locale name before passing it to moment js vulnerabilityurl | 0 |
611,953 | 18,985,572,246 | IssuesEvent | 2021-11-21 17:01:28 | dehy/foodcoop-mobile-app | https://api.github.com/repos/dehy/foodcoop-mobile-app | closed | Migrer la BDD legacy vers TypeORM | Priority: Medium Status: In Progress Type: Refactoring | La bdd sqlite inventaire (legacy, fait maison) est différente de la bdd sqlite réception de marchandises (TypeORM, managée).
Migrer le schéma et les données existantes vers TypeORM. | 1.0 | Migrer la BDD legacy vers TypeORM - La bdd sqlite inventaire (legacy, fait maison) est différente de la bdd sqlite réception de marchandises (TypeORM, managée).
Migrer le schéma et les données existantes vers TypeORM. | priority | migrer la bdd legacy vers typeorm la bdd sqlite inventaire legacy fait maison est différente de la bdd sqlite réception de marchandises typeorm managée migrer le schéma et les données existantes vers typeorm | 1 |
42,646 | 5,513,890,540 | IssuesEvent | 2017-03-17 13:51:47 | Esri/military-tools-geoprocessing-toolbox | https://api.github.com/repos/Esri/military-tools-geoprocessing-toolbox | closed | Using the same Input Area gives different results between Find Local Peaks and Highest Points | B - As Designed effort-medium F - Visibility priority - normal V - 10.3.1 | ## Expected Behavior
Unless the tools need further documentation of the differences between the two, it is expected that when you run Find Local Peaks, the maximum peak value would be the same value returned by Highest Points if you are using the same input feature class.
## Current Behavior
Currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is.
## Possible Solution
This seems to be because there is a difference in the logic of how the tools clip out the elevation layer. Local Peaks seems to only look for points that are entirely within the input polygon, where as it seems that Highest Points also takes into account points are on the edge of the input polygon.
## Steps to Reproduce (for bugs)
1. Run the Local Peaks with the AreaofInterest feature class as the input area
2. Run Highest Points with the AreaofInterest feature class as the input area
3. Note the maximum value of the output of the Local Peaks (1080)
4. Note the maximum value of the output of Highest Points (1123, also observe that it is on the extreme edge of the polygon)
5. Repeat steps using AreaofInterest2
AreaofInterest2 is attached as well as screen captures illustrating problem
[AddVizTestData.zip](https://github.com/Esri/military-tools-geoprocessing-toolbox/files/595797/AddVizTestData.zip)




| 1.0 | Using the same Input Area gives different results between Find Local Peaks and Highest Points - ## Expected Behavior
Unless the tools need further documentation of the differences between the two, it is expected that when you run Find Local Peaks, the maximum peak value would be the same value returned by Highest Points if you are using the same input feature class.
## Current Behavior
Currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is.
## Possible Solution
This seems to be because there is a difference in the logic of how the tools clip out the elevation layer. Local Peaks seems to only look for points that are entirely within the input polygon, where as it seems that Highest Points also takes into account points are on the edge of the input polygon.
## Steps to Reproduce (for bugs)
1. Run the Local Peaks with the AreaofInterest feature class as the input area
2. Run Highest Points with the AreaofInterest feature class as the input area
3. Note the maximum value of the output of the Local Peaks (1080)
4. Note the maximum value of the output of Highest Points (1123, also observe that it is on the extreme edge of the polygon)
5. Repeat steps using AreaofInterest2
AreaofInterest2 is attached as well as screen captures illustrating problem
[AddVizTestData.zip](https://github.com/Esri/military-tools-geoprocessing-toolbox/files/595797/AddVizTestData.zip)




| non_priority | using the same input area gives different results between find local peaks and highest points expected behavior unless the tools need further documentation of the differences between the two it is expected that when you run find local peaks the maximum peak value would be the same value returned by highest points if you are using the same input feature class current behavior currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is possible solution this seems to be because there is a difference in the logic of how the tools clip out the elevation layer local peaks seems to only look for points that are entirely within the input polygon where as it seems that highest points also takes into account points are on the edge of the input polygon steps to reproduce for bugs run the local peaks with the areaofinterest feature class as the input area run highest points with the areaofinterest feature class as the input area note the maximum value of the output of the local peaks note the maximum value of the output of highest points also observe that it is on the extreme edge of the polygon repeat steps using is attached as well as screen captures illustrating problem | 0 |
16,772 | 12,138,443,320 | IssuesEvent | 2020-04-23 17:15:03 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | opened | aspnetcore-tooling migration | area-infrastructure | - [ ] execute phase 1 of https://github.com/dotnet/aspnetcore/blob/master/docs/tooling-consolidation.md
- [ ] plan for phase 2
Phase 1 is ready. Phase 2 needs further planning. | 1.0 | aspnetcore-tooling migration - - [ ] execute phase 1 of https://github.com/dotnet/aspnetcore/blob/master/docs/tooling-consolidation.md
- [ ] plan for phase 2
Phase 1 is ready. Phase 2 needs further planning. | non_priority | aspnetcore tooling migration execute phase of plan for phase phase is ready phase needs further planning | 0 |
304,830 | 9,336,206,263 | IssuesEvent | 2019-03-28 20:36:32 | fecgov/fec-cms | https://api.github.com/repos/fecgov/fec-cms | opened | Events and deadlines do not populate on homepage | Bug High priority | Events and deadlines no longer populate on our homepage. See below screenshot:

| 1.0 | Events and deadlines do not populate on homepage - Events and deadlines no longer populate on our homepage. See below screenshot:

| priority | events and deadlines do not populate on homepage events and deadlines no longer populate on our homepage see below screenshot | 1 |
589,357 | 17,695,160,368 | IssuesEvent | 2021-08-24 14:33:59 | eclipse/dirigible | https://api.github.com/repos/eclipse/dirigible | opened | [IDE] Editors - CSVIM Editor doesn't support # quote character | bug enhancement component-ide usability priority-high efforts-low | **Describe the bug**
Unable to select # as a quote character in CSVIM Editor.
**To Reproduce**
Steps to reproduce the behavior:
1. Open csvim file
2. Select the quote character dropdown
3. See issue
**Expected behavior**
CSVIM Editor should support # as a quote character.
**Desktop:**
- OS: macOS 11.5
- Browser: Firefox 91
- Version: Dirigible 5.12.13
| 1.0 | [IDE] Editors - CSVIM Editor doesn't support # quote character - **Describe the bug**
Unable to select # as a quote character in CSVIM Editor.
**To Reproduce**
Steps to reproduce the behavior:
1. Open csvim file
2. Select the quote character dropdown
3. See issue
**Expected behavior**
CSVIM Editor should support # as a quote character.
**Desktop:**
- OS: macOS 11.5
- Browser: Firefox 91
- Version: Dirigible 5.12.13
| priority | editors csvim editor doesn t support quote character describe the bug unable to select as a quote character in csvim editor to reproduce steps to reproduce the behavior open csvim file select the quote character dropdown see issue expected behavior csvim editor should support as a quote character desktop os macos browser firefox version dirigible | 1 |
150,367 | 5,765,562,281 | IssuesEvent | 2017-04-27 03:01:08 | sussol/mobile | https://api.github.com/repos/sussol/mobile | closed | Item usage not aggregated across batches correctly | Bug Priority: High ~in review | Build Number: 30
Description: It looks like item.dailyUsage() is summing the dailyUsage for each batch, which would over-estimate usage when there are multiple batches. If daily usage for batch 1 is 15, and daily usage for batch 2 is also 15, it looks like item.dailyUsage() would return 30. The spreadsheet example linked in issue #405 also includes something about this. It seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used, then do the division in Item.
Comments: Happy to take a crack at this ticket or #405, but I don't have a server I could test on at the moment..
| 1.0 | Item usage not aggregated across batches correctly - Build Number: 30
Description: It looks like item.dailyUsage() is summing the dailyUsage for each batch, which would over-estimate usage when there are multiple batches. If daily usage for batch 1 is 15, and daily usage for batch 2 is also 15, it looks like item.dailyUsage() would return 30. The spreadsheet example linked in issue #405 also includes something about this. It seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used, then do the division in Item.
Comments: Happy to take a crack at this ticket or #405, but I don't have a server I could test on at the moment..
| priority | item usage not aggregated across batches correctly build number description it looks like item dailyusage is summing the dailyusage for each batch which would over estimate usage when there are multiple batches if daily usage for batch is and daily usage for batch is also it looks like item dailyusage would return the spreadsheet example linked in issue also includes something about this it seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used then do the division in item comments happy to take a crack at this ticket or but i don t have a server i could test on at the moment | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.