Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
143,366
11,545,408,089
IssuesEvent
2020-02-18 13:23:40
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled
Module: IMap Source: Internal Team: Core Type: Test-Failure
http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/4021/testReport/junit/com.hazelcast.map.impl.mapstore.writebehind/WriteBehindEntryStoreQueueReplicationTest/queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up/ ``` java.lang.AssertionError: Expected 'expirationTime' to be between 1572011647548 and 1572011651548, but was 1572011652000 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertBetween(HazelcastTestSupport.java:1461) at com.hazelcast.map.impl.mapstore.TestEntryStore.assertRecordStored(TestEntryStore.java:171) at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.lambda$queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up$2(WriteBehindEntryStoreQueueReplicationTest.java:116) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1318) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1420) at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up(WriteBehindEntryStoreQueueReplicationTest.java:114) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ```
1.0
WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled - http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/4021/testReport/junit/com.hazelcast.map.impl.mapstore.writebehind/WriteBehindEntryStoreQueueReplicationTest/queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up/ ``` java.lang.AssertionError: Expected 'expirationTime' to be between 1572011647548 and 1572011651548, but was 1572011652000 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertBetween(HazelcastTestSupport.java:1461) at com.hazelcast.map.impl.mapstore.TestEntryStore.assertRecordStored(TestEntryStore.java:171) at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.lambda$queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up$2(WriteBehindEntryStoreQueueReplicationTest.java:116) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1318) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1420) at com.hazelcast.map.impl.mapstore.writebehind.WriteBehindEntryStoreQueueReplicationTest.queued_entries_with_expirationTimes_are_replicated_when_cluster_scaled_up(WriteBehindEntryStoreQueueReplicationTest.java:114) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ```
non_priority
writebehindentrystorequeuereplicationtest queued entries with expirationtimes are replicated when cluster scaled java lang assertionerror expected expirationtime to be between and but was at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertbetween hazelcasttestsupport java at com hazelcast map impl mapstore testentrystore assertrecordstored testentrystore java at com hazelcast map impl mapstore writebehind writebehindentrystorequeuereplicationtest lambda queued entries with expirationtimes are replicated when cluster scaled up writebehindentrystorequeuereplicationtest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast map impl mapstore writebehind writebehindentrystorequeuereplicationtest queued entries with expirationtimes are replicated when cluster scaled up writebehindentrystorequeuereplicationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java
0
206,961
16,062,439,546
IssuesEvent
2021-04-23 14:18:05
poliastro/poliastro
https://api.github.com/repos/poliastro/poliastro
closed
Review documentation classification, take 2
documentation
Now that I work at a documentation company 🤓 I have been thinking more and more about poliastro docs, and in particular within the Diátaxis framework https://diataxis.fr/ (formerly known as "The Documentation System"). ![diataxis](https://user-images.githubusercontent.com/316517/113677223-a10e8200-96bd-11eb-9993-4ac6319aea66.png) Even though we did restructure the documentation as part of #975, I would like to revisit this with some ideas: - Add a "Getting started" category that includes the document currently called "Getting started" renamed to "Installation", the "User guide" renamed to "Quickstart guide", and the contributor documentation - Rename "Tutorials" to "Examples" (in the spirit of How-to Guides) - Add a new "Background" category that includes the more theoretical examples, like https://docs.poliastro.space/en/latest/examples/Comparing%20Hohmann%20and%20bielliptic%20transfers.html, https://docs.poliastro.space/en/latest/examples/Propagation%20using%20Cowells%20formulation.html, https://docs.poliastro.space/en/latest/examples/Revisiting%20Lamberts%20problem%20in%20Python.html, https://docs.poliastro.space/en/latest/examples/Studying%20Hohmann%20transfers.html, and https://docs.poliastro.space/en/latest/examples/Multirevolutions%20solution%20in%20Lamberts%20problem.html - Add two TOCs in https://docs.poliastro.space/en/latest/autoapi/index.html, which is something that we had in older versions and I think it's worth bringing back. Not sure how easy it is to do with autoapi.
1.0
Review documentation classification, take 2 - Now that I work at a documentation company 🤓 I have been thinking more and more about poliastro docs, and in particular within the Diátaxis framework https://diataxis.fr/ (formerly known as "The Documentation System"). ![diataxis](https://user-images.githubusercontent.com/316517/113677223-a10e8200-96bd-11eb-9993-4ac6319aea66.png) Even though we did restructure the documentation as part of #975, I would like to revisit this with some ideas: - Add a "Getting started" category that includes the document currently called "Getting started" renamed to "Installation", the "User guide" renamed to "Quickstart guide", and the contributor documentation - Rename "Tutorials" to "Examples" (in the spirit of How-to Guides) - Add a new "Background" category that includes the more theoretical examples, like https://docs.poliastro.space/en/latest/examples/Comparing%20Hohmann%20and%20bielliptic%20transfers.html, https://docs.poliastro.space/en/latest/examples/Propagation%20using%20Cowells%20formulation.html, https://docs.poliastro.space/en/latest/examples/Revisiting%20Lamberts%20problem%20in%20Python.html, https://docs.poliastro.space/en/latest/examples/Studying%20Hohmann%20transfers.html, and https://docs.poliastro.space/en/latest/examples/Multirevolutions%20solution%20in%20Lamberts%20problem.html - Add two TOCs in https://docs.poliastro.space/en/latest/autoapi/index.html, which is something that we had in older versions and I think it's worth bringing back. Not sure how easy it is to do with autoapi.
non_priority
review documentation classification take now that i work at a documentation company 🤓 i have been thinking more and more about poliastro docs and in particular within the diátaxis framework formerly known as the documentation system even though we did restructure the documentation as part of i would like to revisit this with some ideas add a getting started category that includes the document currently called getting started renamed to installation the user guide renamed to quickstart guide and the contributor documentation rename tutorials to examples in the spirit of how to guides add a new background category that includes the more theoretical examples like and add two tocs in which is something that we had in older versions and i think it s worth bringing back not sure how easy it is to do with autoapi
0
553,649
16,376,097,025
IssuesEvent
2021-05-16 05:36:16
KShewengger/benefit-management
https://api.github.com/repos/KShewengger/benefit-management
opened
Setup Vouchers Module, Initial Functionality and Route
Priority: Medium Type: API Type: Functionality
Common: - Entity - Model - Type Providers: - Resolver - Service Core: - Controller - Module Routes: `/vouchers`
1.0
Setup Vouchers Module, Initial Functionality and Route - Common: - Entity - Model - Type Providers: - Resolver - Service Core: - Controller - Module Routes: `/vouchers`
priority
setup vouchers module initial functionality and route common entity model type providers resolver service core controller module routes vouchers
1
21,938
7,098,106,501
IssuesEvent
2018-01-15 02:25:32
hadithhouse/hadithhouse
https://api.github.com/repos/hadithhouse/hadithhouse
opened
Use tslint in build
build & deploy
To make sure that code not styled correctly doesn't get into the repository, I would like to add `tslint` to the build process so build fails if there are style warnings.
1.0
Use tslint in build - To make sure that code not styled correctly doesn't get into the repository, I would like to add `tslint` to the build process so build fails if there are style warnings.
non_priority
use tslint in build to make sure that code not styled correctly doesn t get into the repository i would like to add tslint to the build process so build fails if there are style warnings
0
134,014
5,218,783,475
IssuesEvent
2017-01-26 17:17:47
duckduckgo/zeroclickinfo-fathead
https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead
opened
Objective-C: Create new Objective-C Reference Fathead
Difficulty: High Mission: Programming New Instant Answer Priority: High Status: Needs a Developer Topic: Objective-C Topic: Reference
# Create a new C Fathead Instant Answer Help us make DuckDuckGo the best search engine for programmers! ### What do I need to know? You'll need to know how to code in **Perl**, **Python**, **Ruby**, or **JavaScript**. ![fathead languages](https://cloud.githubusercontent.com/assets/873785/19787916/57b4c31a-9c73-11e6-9bd9-f85c8893ec93.jpg) ### What am I doing? You will write a script that scrapes or downloads the data source below, and generates an **output.txt** file containing the parsed documentation. You can learn more about Fatheads and the `output.txt` syntax [**here**](https://docs.duckduckhack.com/resources/fathead-overview.html). **Data source**: https://developer.apple.com/reference/objectivec **Instant Answer Page**: *To be created by developer* <!-- ^^^ FILL THIS IN, AFTER ISSUE IS CLAIMED ^^^ --> ### What is the Goal? As part of our [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53), we're aiming to reach 100% Instant Answer (IA) coverage for searches related to programming languages by creating new Instant Answers, and improving existing ones. Here are some Fathead examples: - Ruby Docs - [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/ruby) | [Example Query](https://duckduckgo.com/?q=array+bsearch&ia=about) - MDN CSS - [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/mdn_css) | [Example Query](https://duckduckgo.com/?q=css+background-position&ia=about) ![css background-position search](https://cloud.githubusercontent.com/assets/873785/19787980/cdf51566-9c73-11e6-9ef2-ac44dede62c7.png) [See more related Instant Answers](https://duck.co/ia?repo=fathead) ## Get Started - [ ] 1) Claim this issue by commenting below - [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-fathead/blob/master/CONTRIBUTING.md) - [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository - [ ] 4) Create a new Instant Answer Page: https://duck.co/ia/new_ia (then let us know, here!) - [ ] 5) Create the Fathead - [ ] 6) Create a Pull Request - [ ] 7) Ping @pjhampton for a review <!-- ^^^ FILL THIS IN ^^^ --> ## Resources - Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions - Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics - Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help
1.0
Objective-C: Create new Objective-C Reference Fathead - # Create a new C Fathead Instant Answer Help us make DuckDuckGo the best search engine for programmers! ### What do I need to know? You'll need to know how to code in **Perl**, **Python**, **Ruby**, or **JavaScript**. ![fathead languages](https://cloud.githubusercontent.com/assets/873785/19787916/57b4c31a-9c73-11e6-9bd9-f85c8893ec93.jpg) ### What am I doing? You will write a script that scrapes or downloads the data source below, and generates an **output.txt** file containing the parsed documentation. You can learn more about Fatheads and the `output.txt` syntax [**here**](https://docs.duckduckhack.com/resources/fathead-overview.html). **Data source**: https://developer.apple.com/reference/objectivec **Instant Answer Page**: *To be created by developer* <!-- ^^^ FILL THIS IN, AFTER ISSUE IS CLAIMED ^^^ --> ### What is the Goal? As part of our [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53), we're aiming to reach 100% Instant Answer (IA) coverage for searches related to programming languages by creating new Instant Answers, and improving existing ones. Here are some Fathead examples: - Ruby Docs - [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/ruby) | [Example Query](https://duckduckgo.com/?q=array+bsearch&ia=about) - MDN CSS - [Code](https://github.com/duckduckgo/zeroclickinfo-fathead/tree/master/lib/fathead/mdn_css) | [Example Query](https://duckduckgo.com/?q=css+background-position&ia=about) ![css background-position search](https://cloud.githubusercontent.com/assets/873785/19787980/cdf51566-9c73-11e6-9ef2-ac44dede62c7.png) [See more related Instant Answers](https://duck.co/ia?repo=fathead) ## Get Started - [ ] 1) Claim this issue by commenting below - [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-fathead/blob/master/CONTRIBUTING.md) - [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository - [ ] 4) Create a new Instant Answer Page: https://duck.co/ia/new_ia (then let us know, here!) - [ ] 5) Create the Fathead - [ ] 6) Create a Pull Request - [ ] 7) Ping @pjhampton for a review <!-- ^^^ FILL THIS IN ^^^ --> ## Resources - Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions - Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics - Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help
priority
objective c create new objective c reference fathead create a new c fathead instant answer help us make duckduckgo the best search engine for programmers what do i need to know you ll need to know how to code in perl python ruby or javascript what am i doing you will write a script that scrapes or downloads the data source below and generates an output txt file containing the parsed documentation you can learn more about fatheads and the output txt syntax data source instant answer page to be created by developer what is the goal as part of our we re aiming to reach instant answer ia coverage for searches related to programming languages by creating new instant answers and improving existing ones here are some fathead examples ruby docs mdn css get started claim this issue by commenting below review our and fork this repository create a new instant answer page then let us know here create the fathead create a pull request ping pjhampton for a review resources join to ask questions join the to discuss project planning and instant answer metrics read the for technical help
1
408,132
27,645,353,792
IssuesEvent
2023-03-10 22:18:31
flyteorg/flyte
https://api.github.com/repos/flyteorg/flyte
opened
[Docs] Create model deployment and data annotation-specific examples for gate nodes
documentation untriaged
### Description https://github.com/flyteorg/flytesnacks/pull/968 tracks work for adding a gate nodes example to flytesnacks. This issue is to add examples to the Tutorials section of flytesnacks, under two new headers: - Model Deployment: create an example where gate nodes are used to pick which model to deploy - Data Annotation: create an example where gate nodes are used to Reference: see [these commits](https://github.com/flyteorg/flytesnacks/pull/968/commits/016fdcc4a59202e7ac555380b90fc97524f7444d) and [this notebook](https://github.com/flyteorg/flytesnacks/blob/ae1bef45fd536b0576694b4801db73169157cb47/cookbook/Untitled.ipynb) for examples of the above two use cases. ### Are you sure this issue hasn't been raised already? - [X] Yes ### Have you read the Code of Conduct? - [X] Yes
1.0
[Docs] Create model deployment and data annotation-specific examples for gate nodes - ### Description https://github.com/flyteorg/flytesnacks/pull/968 tracks work for adding a gate nodes example to flytesnacks. This issue is to add examples to the Tutorials section of flytesnacks, under two new headers: - Model Deployment: create an example where gate nodes are used to pick which model to deploy - Data Annotation: create an example where gate nodes are used to Reference: see [these commits](https://github.com/flyteorg/flytesnacks/pull/968/commits/016fdcc4a59202e7ac555380b90fc97524f7444d) and [this notebook](https://github.com/flyteorg/flytesnacks/blob/ae1bef45fd536b0576694b4801db73169157cb47/cookbook/Untitled.ipynb) for examples of the above two use cases. ### Are you sure this issue hasn't been raised already? - [X] Yes ### Have you read the Code of Conduct? - [X] Yes
non_priority
create model deployment and data annotation specific examples for gate nodes description tracks work for adding a gate nodes example to flytesnacks this issue is to add examples to the tutorials section of flytesnacks under two new headers model deployment create an example where gate nodes are used to pick which model to deploy data annotation create an example where gate nodes are used to reference see and for examples of the above two use cases are you sure this issue hasn t been raised already yes have you read the code of conduct yes
0
536,999
15,721,184,789
IssuesEvent
2021-03-29 02:24:09
wso2/product-microgateway
https://api.github.com/repos/wso2/product-microgateway
opened
Support Basic Auth Endpoint security
Priority/Normal Type/New Feature
### Describe your problem(s) Currently any of endpoint security mechanism is not supported in mgw. Basic auth support should be included ### Describe your solution Adding security details to API object and parse to enforcer in order to add basic auth credentials at runtime ### How will you implement it JWTAuthenticator flow will be added new header appending basic auth credentials checking endpoint keytype --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
Support Basic Auth Endpoint security - ### Describe your problem(s) Currently any of endpoint security mechanism is not supported in mgw. Basic auth support should be included ### Describe your solution Adding security details to API object and parse to enforcer in order to add basic auth credentials at runtime ### How will you implement it JWTAuthenticator flow will be added new header appending basic auth credentials checking endpoint keytype --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
priority
support basic auth endpoint security describe your problem s currently any of endpoint security mechanism is not supported in mgw basic auth support should be included describe your solution adding security details to api object and parse to enforcer in order to add basic auth credentials at runtime how will you implement it jwtauthenticator flow will be added new header appending basic auth credentials checking endpoint keytype optional fields related issues suggested labels suggested assignees
1
97,363
28,214,053,785
IssuesEvent
2023-04-05 07:36:51
microsoft/appcenter
https://api.github.com/repos/microsoft/appcenter
reopened
iOS and Android Build Fail when Test on Real Device is Toggled on
bug build test reviewed-DRI
**What App Center service does this affect?** React Native SDK, and Build. **Describe the bug** Locally Xcode is building normally, but it fails in appcenter with Test on Real Device Toggled on. When it's toggled off, the build is successful. **Expected behavior** Build successful. Additional context ##[error]Error: /usr/local/bin/appcenter failed with return code: 3 ##[section]Finishing: Run generated tests on real devices
1.0
iOS and Android Build Fail when Test on Real Device is Toggled on - **What App Center service does this affect?** React Native SDK, and Build. **Describe the bug** Locally Xcode is building normally, but it fails in appcenter with Test on Real Device Toggled on. When it's toggled off, the build is successful. **Expected behavior** Build successful. Additional context ##[error]Error: /usr/local/bin/appcenter failed with return code: 3 ##[section]Finishing: Run generated tests on real devices
non_priority
ios and android build fail when test on real device is toggled on what app center service does this affect react native sdk and build describe the bug locally xcode is building normally but it fails in appcenter with test on real device toggled on when it s toggled off the build is successful expected behavior build successful additional context error usr local bin appcenter failed with return code finishing run generated tests on real devices
0
235,213
7,735,453,222
IssuesEvent
2018-05-27 15:17:40
ream/ream
https://api.github.com/repos/ream/ream
closed
Prefetch data for route components
contribution welcome enhancement priority: high
Currently we have [getInitialData](https://github.com/ream/ream/blob/master/docs/guides/preloading-data.md) but it does not inject data or props to relevant route component, hopefully we can implement a _real_ `getInitialData` which injects resolved object to component data or `getInitialProps` which injects props instead. `getInitialData`: ```js export default { async getInitialData() { return { count: 0 } }, render() { return <button>{this.count /* this.$data.count */}</button> } } ``` `getInitialProps`: ```js export default { async getInitialProps() { return { count: 0 } }, props: ['count'] render() { return <button>{this.count /* this.$props.count */}</button> } } ``` If it's possible to have `getInitialProps` then we don't need to implement `getInitialData` --- Refs: - [getInitialProps](https://github.com/zeit/next.js/#fetching-data-and-component-lifecycle) in Next.js - [asyncData](https://nuxtjs.org/api/) in Nuxt.js
1.0
Prefetch data for route components - Currently we have [getInitialData](https://github.com/ream/ream/blob/master/docs/guides/preloading-data.md) but it does not inject data or props to relevant route component, hopefully we can implement a _real_ `getInitialData` which injects resolved object to component data or `getInitialProps` which injects props instead. `getInitialData`: ```js export default { async getInitialData() { return { count: 0 } }, render() { return <button>{this.count /* this.$data.count */}</button> } } ``` `getInitialProps`: ```js export default { async getInitialProps() { return { count: 0 } }, props: ['count'] render() { return <button>{this.count /* this.$props.count */}</button> } } ``` If it's possible to have `getInitialProps` then we don't need to implement `getInitialData` --- Refs: - [getInitialProps](https://github.com/zeit/next.js/#fetching-data-and-component-lifecycle) in Next.js - [asyncData](https://nuxtjs.org/api/) in Nuxt.js
priority
prefetch data for route components currently we have but it does not inject data or props to relevant route component hopefully we can implement a real getinitialdata which injects resolved object to component data or getinitialprops which injects props instead getinitialdata js export default async getinitialdata return count render return this count this data count getinitialprops js export default async getinitialprops return count props render return this count this props count if it s possible to have getinitialprops then we don t need to implement getinitialdata refs in next js in nuxt js
1
305,847
9,377,715,858
IssuesEvent
2019-04-04 11:02:07
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
m.facebook.com - see bug description
browser-firefox-mobile priority-critical
<!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://m.facebook.com/story.php?story_fbid=2260305700712631&id=1160243874052158&anchor_composer=false **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 7.1.1 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: I use messenger on a tablet, but i am having issues with message notifications clearing from all of my devices, unless I load the desktop site on my phone and go to the inbox from there. then the notification will clear out on all devices that I am viewing the site from. **Steps to Reproduce**: I am using the silk browser, and fb messenger on the tablet. Firefox on the phone, and that's how I clear them. The desktop site is broken as well. Only one day this last year or two has it displayed like the desktop site looks on a PC. The day that it displayed correctly was the recent day that Facebook was broken for the full day. When I could get it to load that day, it looked identical to the desktop site. Every other day, it's just a bunch of text and links. <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190218131312</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=2314269647&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2c&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=1710212709&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2e&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
m.facebook.com - see bug description - <!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://m.facebook.com/story.php?story_fbid=2260305700712631&id=1160243874052158&anchor_composer=false **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 7.1.1 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: I use messenger on a tablet, but i am having issues with message notifications clearing from all of my devices, unless I load the desktop site on my phone and go to the inbox from there. then the notification will clear out on all devices that I am viewing the site from. **Steps to Reproduce**: I am using the silk browser, and fb messenger on the tablet. Firefox on the phone, and that's how I clear them. The desktop site is broken as well. Only one day this last year or two has it displayed like the desktop site looks on a PC. The day that it displayed correctly was the recent day that Facebook was broken for the full day. When I could get it to load that day, it looked identical to the desktop site. Every other day, it's just a bunch of text and links. <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190218131312</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=2314269647&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2c&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Error: "Firefox can\'t establish a connection to the server at wss://edge-chat.facebook.com/chat?region=ash." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://edge-chat.facebook.com/pull?channel=p_1405823680&seq=2&clientid=239974c8-1358-463a-a54d-917081a5db0e&profile=mobile&partition=-2&sticky_token=397&msgs_recv=2&qp=y&cb=1710212709&sticky_pool=rsgp0c01_chatproxy-regional&uid=1405823680&viewer_uid=1405823680&m_sess=&__dyn=1KQdAmmcwgVU-4UpwGzWAgy79pkdxu6Erz8C2K2i5UfXwNwTwKwpUuKew8i5orx64o5K58O8wmEcoe8hw8C223G0JoeoK2O1gCwSxu0BU7W1KxO1Cxq78gwJwWwnElzawlo168WUS2K2zwp8eVE5yUmwk8lyEC0NE&__req=2e&__ajax__=AYl8zWQBK1iZcFmPpvUYlkjzDCAvtEX-wq7yl_1CdKY6fg7eYhqxrKTLdjkgS-hIC4Easp3i9MQ75Ik8ZGEU4uYvjP9olMSe0OjiVTSmPR5dmw&__user=1405823680. (Reason: CORS request did not succeed)."]', u'[JavaScript Error: "The connection to wss://edge-chat.facebook.com/chat?region=ash was interrupted while the page was loading." {file: "https://static.xx.fbcdn.net/rsrc.php/v3/yZ/r/BXbUHaqhN0v.js" line: 18}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
m facebook com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description i use messenger on a tablet but i am having issues with message notifications clearing from all of my devices unless i load the desktop site on my phone and go to the inbox from there then the notification will clear out on all devices that i am viewing the site from steps to reproduce i am using the silk browser and fb messenger on the tablet firefox on the phone and that s how i clear them the desktop site is broken as well only one day this last year or two has it displayed like the desktop site looks on a pc the day that it displayed correctly was the recent day that facebook was broken for the full day when i could get it to load that day it looked identical to the desktop site every other day it s just a bunch of text and links browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u u u u from with ❤️
1
414,483
12,103,836,536
IssuesEvent
2020-04-20 19:07:17
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
media.interieur.gouv.fr - see bug description
browser-fenix engine-gecko ml-needsdiagnosis-false priority-normal
<!-- @browser: Firefox Mobile 75.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51919 --> <!-- @extra_labels: browser-fenix --> **URL**: https://media.interieur.gouv.fr/deplacement-covid-19/ **Browser / Version**: Firefox Mobile 75.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: PDF download is blocked **Steps to Reproduce**: generated pdf cannot be downloaded, error message. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
media.interieur.gouv.fr - see bug description - <!-- @browser: Firefox Mobile 75.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51919 --> <!-- @extra_labels: browser-fenix --> **URL**: https://media.interieur.gouv.fr/deplacement-covid-19/ **Browser / Version**: Firefox Mobile 75.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: PDF download is blocked **Steps to Reproduce**: generated pdf cannot be downloaded, error message. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
media interieur gouv fr see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description pdf download is blocked steps to reproduce generated pdf cannot be downloaded error message browser configuration none from with ❤️
1
266,464
8,367,974,659
IssuesEvent
2018-10-04 13:42:16
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Improve program directory related experience
Component/Composer Imported Priority/High Type/Improvement
<a href="https://github.com/kaviththiranga"><img src="https://avatars3.githubusercontent.com/u/1505855?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [kaviththiranga](https://github.com/kaviththiranga)** _Tuesday Oct 31, 2017 at 03:12 GMT_ _Originally opened as https://github.com/ballerina-lang/composer/issues/4306_ ---- - When a user tries to open a packaged file within a program dir, ask whether user wants to open the program dir too and open upon confirmation. - When a user adds a package to a file for the first time, - If it's a new file which not yet persisted, upon save, in save dialog, show a check box with message that we need to create the sub folder for package structure. - if it's file which is already saved, display an alert with same message and provide a btn to create folder structure and move file there - when a user changes the package of a opened file, ask and move file to correct directory - provide a wizard to create a file within program dir and a package
1.0
Improve program directory related experience - <a href="https://github.com/kaviththiranga"><img src="https://avatars3.githubusercontent.com/u/1505855?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [kaviththiranga](https://github.com/kaviththiranga)** _Tuesday Oct 31, 2017 at 03:12 GMT_ _Originally opened as https://github.com/ballerina-lang/composer/issues/4306_ ---- - When a user tries to open a packaged file within a program dir, ask whether user wants to open the program dir too and open upon confirmation. - When a user adds a package to a file for the first time, - If it's a new file which not yet persisted, upon save, in save dialog, show a check box with message that we need to create the sub folder for package structure. - if it's file which is already saved, display an alert with same message and provide a btn to create folder structure and move file there - when a user changes the package of a opened file, ask and move file to correct directory - provide a wizard to create a file within program dir and a package
priority
improve program directory related experience issue by tuesday oct at gmt originally opened as when a user tries to open a packaged file within a program dir ask whether user wants to open the program dir too and open upon confirmation when a user adds a package to a file for the first time if it s a new file which not yet persisted upon save in save dialog show a check box with message that we need to create the sub folder for package structure if it s file which is already saved display an alert with same message and provide a btn to create folder structure and move file there when a user changes the package of a opened file ask and move file to correct directory provide a wizard to create a file within program dir and a package
1
611,988
18,988,281,137
IssuesEvent
2021-11-22 01:44:23
boostcampwm-2021/iOS05-Escaper
https://api.github.com/repos/boostcampwm-2021/iOS05-Escaper
closed
[E8 S2 T1] 랭킹 View를 구현한다.
feature High Priority
### Epic - Story - Task Epic : 상세 페이지 Story : 방에 대한 랭킹을 보낼 수 있다 Task : 랭킹 Cell을 구현한다.
1.0
[E8 S2 T1] 랭킹 View를 구현한다. - ### Epic - Story - Task Epic : 상세 페이지 Story : 방에 대한 랭킹을 보낼 수 있다 Task : 랭킹 Cell을 구현한다.
priority
랭킹 view를 구현한다 epic story task epic 상세 페이지 story 방에 대한 랭킹을 보낼 수 있다 task 랭킹 cell을 구현한다
1
99,441
20,966,290,711
IssuesEvent
2022-03-28 07:06:35
Validator2/MesaBox
https://api.github.com/repos/Validator2/MesaBox
opened
Mining Laser, man-portable/emplacement
code models
Original idea by "sirro" Powered by U-235, its a modern, finalized successor to the Gluon Gun. Instead of completely vaporizing matter, it has several functions; tbd
1.0
Mining Laser, man-portable/emplacement - Original idea by "sirro" Powered by U-235, its a modern, finalized successor to the Gluon Gun. Instead of completely vaporizing matter, it has several functions; tbd
non_priority
mining laser man portable emplacement original idea by sirro powered by u its a modern finalized successor to the gluon gun instead of completely vaporizing matter it has several functions tbd
0
1,604
6,445,176,676
IssuesEvent
2017-08-12 23:23:56
p4lang/p4-spec
https://api.github.com/repos/p4lang/p4-spec
closed
[PSA] count() operation on counters
portable switch architecture
For a 'bytes' or 'packets_and_bytes' counter type, count() having a second parameter 'increment' that specifies how much to add to the byte counter is a very good thing. It is best if the P4 program has the flexibility to choose the length in bytes it wants to use for the packet, e.g. in case it is increasing or decreasing the received packet's size before transmitting it. For counters that have type 'packets', presumably every call to 'count' will add exactly 1, yes? In that case, having a separate method call that only takes an 'index' parameter would be best, since the 'increment' parameter value would be ignored, anyway.
1.0
[PSA] count() operation on counters - For a 'bytes' or 'packets_and_bytes' counter type, count() having a second parameter 'increment' that specifies how much to add to the byte counter is a very good thing. It is best if the P4 program has the flexibility to choose the length in bytes it wants to use for the packet, e.g. in case it is increasing or decreasing the received packet's size before transmitting it. For counters that have type 'packets', presumably every call to 'count' will add exactly 1, yes? In that case, having a separate method call that only takes an 'index' parameter would be best, since the 'increment' parameter value would be ignored, anyway.
non_priority
count operation on counters for a bytes or packets and bytes counter type count having a second parameter increment that specifies how much to add to the byte counter is a very good thing it is best if the program has the flexibility to choose the length in bytes it wants to use for the packet e g in case it is increasing or decreasing the received packet s size before transmitting it for counters that have type packets presumably every call to count will add exactly yes in that case having a separate method call that only takes an index parameter would be best since the increment parameter value would be ignored anyway
0
320,160
9,777,097,699
IssuesEvent
2019-06-07 08:08:15
DCRGraphsNet/DCROpenCaseManager
https://api.github.com/repos/DCRGraphsNet/DCROpenCaseManager
reopened
Will it be possible to add the child's ages as a guard of an activity?
Priority 3 Udviklingforslag
eg. when the young person turns 16 or 16, the activity becomes pending So - we need the Age of the child in the process. Or date of birth - not cpr. So we can calculate age. Right now we do not have this capability. You could enter Age in a form - and solve it this way. But we cannot execute "Robot" events that take data from system and set into a graph. But - probably a future requirement
1.0
Will it be possible to add the child's ages as a guard of an activity? - eg. when the young person turns 16 or 16, the activity becomes pending So - we need the Age of the child in the process. Or date of birth - not cpr. So we can calculate age. Right now we do not have this capability. You could enter Age in a form - and solve it this way. But we cannot execute "Robot" events that take data from system and set into a graph. But - probably a future requirement
priority
will it be possible to add the child s ages as a guard of an activity eg when the young person turns or the activity becomes pending so we need the age of the child in the process or date of birth not cpr so we can calculate age right now we do not have this capability you could enter age in a form and solve it this way but we cannot execute robot events that take data from system and set into a graph but probably a future requirement
1
368,791
25,808,835,568
IssuesEvent
2022-12-11 17:04:26
12rambau/sepal_ui
https://api.github.com/repos/12rambau/sepal_ui
closed
refactor the graph in the contributor section
documentation
We are curently suggesting to use the following branching system: ![](https://nvie.com/img/git-model@2x.png) It is not the one we currently use in the repository so it should be updated with a custom one (maybe using graphviz ?)
1.0
refactor the graph in the contributor section - We are curently suggesting to use the following branching system: ![](https://nvie.com/img/git-model@2x.png) It is not the one we currently use in the repository so it should be updated with a custom one (maybe using graphviz ?)
non_priority
refactor the graph in the contributor section we are curently suggesting to use the following branching system it is not the one we currently use in the repository so it should be updated with a custom one maybe using graphviz
0
593,496
18,009,412,923
IssuesEvent
2021-09-16 06:40:55
GC-spigot/AdvancedEnchantments
https://api.github.com/repos/GC-spigot/AdvancedEnchantments
closed
Effect TNT not reduced by Blast Protection or DECREASE_DAMAGE
Priority: Low Bug: Confirmed Resolution: Accepted
## Details **Describe the bug** The effect "TNT" cannot be reduced by vanilla Blast Protection or AE's DECREASE_DAMAGE effect (when activated by EXPLOSION). **To Reproduce** <!-- !IMPORTANT! --> 1. Create an armor enchantment with a damaging TNT effect, activated by FALL_DAMAGE. 2. Create an armor enchantment that decreases EXPLOSION type damage by 75%. 3. Take fall damage from the same height a) First without anything. b) Then with the TNT effect. c) Finally with the EXPLOSION damage reduction effect. 4. Damage will be the same in all three **Screenshots / Video** ## Server Information - "/ae plinfo" link: https://paste.md-5.net/sotureroni - Server log: No messages from or related to AE.
1.0
Effect TNT not reduced by Blast Protection or DECREASE_DAMAGE - ## Details **Describe the bug** The effect "TNT" cannot be reduced by vanilla Blast Protection or AE's DECREASE_DAMAGE effect (when activated by EXPLOSION). **To Reproduce** <!-- !IMPORTANT! --> 1. Create an armor enchantment with a damaging TNT effect, activated by FALL_DAMAGE. 2. Create an armor enchantment that decreases EXPLOSION type damage by 75%. 3. Take fall damage from the same height a) First without anything. b) Then with the TNT effect. c) Finally with the EXPLOSION damage reduction effect. 4. Damage will be the same in all three **Screenshots / Video** ## Server Information - "/ae plinfo" link: https://paste.md-5.net/sotureroni - Server log: No messages from or related to AE.
priority
effect tnt not reduced by blast protection or decrease damage details describe the bug the effect tnt cannot be reduced by vanilla blast protection or ae s decrease damage effect when activated by explosion to reproduce create an armor enchantment with a damaging tnt effect activated by fall damage create an armor enchantment that decreases explosion type damage by take fall damage from the same height a first without anything b then with the tnt effect c finally with the explosion damage reduction effect damage will be the same in all three screenshots video server information ae plinfo link server log no messages from or related to ae
1
50,021
26,433,021,846
IssuesEvent
2023-01-15 02:40:05
GraphiteEditor/Graphite
https://api.github.com/repos/GraphiteEditor/Graphite
closed
Debounce widget inputs to minimize backend spam
Feature Web P-Medium Performance
Send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently.
True
Debounce widget inputs to minimize backend spam - Send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently.
non_priority
debounce widget inputs to minimize backend spam send fewer messages to the backend for rapidly dragged input widgets so they update the backend a little less frequently
0
447,286
12,887,563,861
IssuesEvent
2020-07-13 11:28:31
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1634] Coroutine couldn't be started
Category: Tech Priority: Medium Status: Fixed Week Task
When you go to main menu: ``` Coroutine couldn't be started because the the game object 'News(Clone)' is inactive! News.NewsRenderer:DoRendering(News) News.NewsController:SetNews(List`1) News.<SetNews>d__4:MoveNext() UnityEngine.SetupCoroutine:InvokeMoveNext(IEnumerator, IntPtr) ```
1.0
[0.9.0 staging-1634] Coroutine couldn't be started - When you go to main menu: ``` Coroutine couldn't be started because the the game object 'News(Clone)' is inactive! News.NewsRenderer:DoRendering(News) News.NewsController:SetNews(List`1) News.<SetNews>d__4:MoveNext() UnityEngine.SetupCoroutine:InvokeMoveNext(IEnumerator, IntPtr) ```
priority
coroutine couldn t be started when you go to main menu coroutine couldn t be started because the the game object news clone is inactive news newsrenderer dorendering news news newscontroller setnews list news d movenext unityengine setupcoroutine invokemovenext ienumerator intptr
1
144,381
11,614,148,682
IssuesEvent
2020-02-26 12:03:47
pingcap/tidb-operator
https://api.github.com/repos/pingcap/tidb-operator
closed
e2e: "[Feature: AdvancedStatefulSet] Scaling tidb cluster with advanced statefulset " is flaky
test/e2e
## Bug Report https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/operator_ghpr_e2e_test_kind/detail/operator_ghpr_e2e_test_kind/2273/tests ``` Stacktrace /home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:139 Jan 10 10:04:16.182: Unexpected error: <*meta.NoKindMatchError | 0xc0010bb400>: { GroupKind: {Group: "pingcap.com", Kind: "TidbCluster"}, SearchedVersions: ["v1alpha1"], } no matches for kind "TidbCluster" in version "pingcap.com/v1alpha1" occurred /home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:241 ``` k8s version: v1.12.10
1.0
e2e: "[Feature: AdvancedStatefulSet] Scaling tidb cluster with advanced statefulset " is flaky - ## Bug Report https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/operator_ghpr_e2e_test_kind/detail/operator_ghpr_e2e_test_kind/2273/tests ``` Stacktrace /home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:139 Jan 10 10:04:16.182: Unexpected error: <*meta.NoKindMatchError | 0xc0010bb400>: { GroupKind: {Group: "pingcap.com", Kind: "TidbCluster"}, SearchedVersions: ["v1alpha1"], } no matches for kind "TidbCluster" in version "pingcap.com/v1alpha1" occurred /home/jenkins/agent/workspace/operator_ghpr_e2e_test_kind/go/src/github.com/pingcap/tidb-operator/tests/e2e/tidbcluster/serial.go:241 ``` k8s version: v1.12.10
non_priority
scaling tidb cluster with advanced statefulset is flaky bug report stacktrace home jenkins agent workspace operator ghpr test kind go src github com pingcap tidb operator tests tidbcluster serial go jan unexpected error groupkind group pingcap com kind tidbcluster searchedversions no matches for kind tidbcluster in version pingcap com occurred home jenkins agent workspace operator ghpr test kind go src github com pingcap tidb operator tests tidbcluster serial go version
0
53,463
13,167,592,090
IssuesEvent
2020-08-11 10:32:49
wellcomecollection/platform
https://api.github.com/repos/wellcomecollection/platform
closed
Associate CI policy for Buildkite agents IAM Role
:recycle: Builds and CI
Buildkite agents (as EC2 instances) created using the Buildkite AWS elastic stack (https://github.com/buildkite/elastic-ci-stack-for-aws) have associated IAM roles which should have policy documents mirroring our existing CI permissions.
1.0
Associate CI policy for Buildkite agents IAM Role - Buildkite agents (as EC2 instances) created using the Buildkite AWS elastic stack (https://github.com/buildkite/elastic-ci-stack-for-aws) have associated IAM roles which should have policy documents mirroring our existing CI permissions.
non_priority
associate ci policy for buildkite agents iam role buildkite agents as instances created using the buildkite aws elastic stack have associated iam roles which should have policy documents mirroring our existing ci permissions
0
802,865
29,047,860,883
IssuesEvent
2023-05-13 20:06:58
vdjagilev/nmap-formatter
https://api.github.com/repos/vdjagilev/nmap-formatter
closed
Update Node.js 12 -> 16 for pipelines
priority/medium type/other prop/pipeline
``` Node.js 12 actions are deprecated. Please update the following actions to use Node.js 16: actions/checkout@v2. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/ ``` Need to upgrade github actions to version 16
1.0
Update Node.js 12 -> 16 for pipelines - ``` Node.js 12 actions are deprecated. Please update the following actions to use Node.js 16: actions/checkout@v2. For more information see: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/ ``` Need to upgrade github actions to version 16
priority
update node js for pipelines node js actions are deprecated please update the following actions to use node js actions checkout for more information see need to upgrade github actions to version
1
661,913
22,095,763,009
IssuesEvent
2022-06-01 09:57:29
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Allow deleting qcow2 disk snapshot using blockcommit
Category: Drivers - Storage Type: Feature Sponsored Status: Accepted Priority: Normal
**Description** Currently deleting disk snapshots when using qcow2 DS is very limited. Users can not delete the active snapshots and also snapshots with children, so the only way how to delete the snapshot is to revert back. **Use case** e.g: when users get charged in terms of snapshot numbers, they need to delete them **Interface Changes** Deleting non-active snapshots with children could be achieved using `blockcommit` op. https://wiki.libvirt.org/page/Live-disk-backup-with-active-blockcommit <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Allow deleting qcow2 disk snapshot using blockcommit - **Description** Currently deleting disk snapshots when using qcow2 DS is very limited. Users can not delete the active snapshots and also snapshots with children, so the only way how to delete the snapshot is to revert back. **Use case** e.g: when users get charged in terms of snapshot numbers, they need to delete them **Interface Changes** Deleting non-active snapshots with children could be achieved using `blockcommit` op. https://wiki.libvirt.org/page/Live-disk-backup-with-active-blockcommit <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
allow deleting disk snapshot using blockcommit description currently deleting disk snapshots when using ds is very limited users can not delete the active snapshots and also snapshots with children so the only way how to delete the snapshot is to revert back use case e g when users get charged in terms of snapshot numbers they need to delete them interface changes deleting non active snapshots with children could be achieved using blockcommit op progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
29,789
14,265,763,425
IssuesEvent
2020-11-20 17:36:50
ampproject/amp-toolbox-php
https://api.github.com/repos/ampproject/amp-toolbox-php
opened
New Transformer: BrowserHints
Performance
- [ ] Add preconnect link tag for Google fonts resources - [ ] Add preconnect link tag to the publisher's own origin - [ ] Preload images - [ ] Preload AMP runtime script - [ ] Prune duplicate resource hints
True
New Transformer: BrowserHints - - [ ] Add preconnect link tag for Google fonts resources - [ ] Add preconnect link tag to the publisher's own origin - [ ] Preload images - [ ] Preload AMP runtime script - [ ] Prune duplicate resource hints
non_priority
new transformer browserhints add preconnect link tag for google fonts resources add preconnect link tag to the publisher s own origin preload images preload amp runtime script prune duplicate resource hints
0
88,181
15,800,747,050
IssuesEvent
2021-04-03 01:06:24
rammatzkvosky/jenkins
https://api.github.com/repos/rammatzkvosky/jenkins
opened
CVE-2021-21348 (High) detected in xstream-1.4.15.jar
security vulnerability
## CVE-2021-21348 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.15.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: jenkins/test/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,canner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,jenkins/war/target/jenkins/WEB-INF/lib/xstream-1.4.15.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.15.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348>CVE-2021-21348</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq">https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.15","packageFilePaths":["/test/pom.xml","/war/pom.xml","/core/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21348","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-21348 (High) detected in xstream-1.4.15.jar - ## CVE-2021-21348 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.15.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: jenkins/test/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,canner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.15/xstream-1.4.15.jar,jenkins/war/target/jenkins/WEB-INF/lib/xstream-1.4.15.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.15.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348>CVE-2021-21348</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq">https://github.com/x-stream/xstream/security/advisories/GHSA-56p8-3fh9-4cvq</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.15","packageFilePaths":["/test/pom.xml","/war/pom.xml","/core/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21348","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum CPU time and will never return. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21348","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar library home page a href path to dependency file jenkins test pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar canner repository com thoughtworks xstream xstream xstream jar jenkins war target jenkins web inf lib xstream jar dependency hierarchy x xstream jar vulnerable library found in base branch master vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum cpu time and will never return no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to occupy a thread that consumes maximum cpu time and will never return no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl
0
586,874
17,599,170,794
IssuesEvent
2021-08-17 09:37:18
googleapis/nodejs-video-intelligence
https://api.github.com/repos/googleapis/nodejs-video-intelligence
opened
analyze samples: should track objects in a GCS file failed
type: bug priority: p1 flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2f7fe652af0d15621b89fc80cc22f9b2d0a2f209 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/24a5eecc-b508-4ba3-8f13-09342425c56a), [Sponge](http://sponge2/24a5eecc-b508-4ba3-8f13-09342425c56a) status: failed <details><summary>Test output</summary><br><pre>Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4 analyze.js track-objects-gcs <gcsUri> Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API. Options: --version Show version number [boolean] --help Show help [boolean] Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'google.rpc.errorinfo-bin' => [Array], 'grpc-status-details-bin' => [Array], 'www-authenticate' => [Array], 'method' => [Array], 'service' => [Array], 'email' => [Array] }, options: {} }, note: 'Exception occurred in retry method that was not classified as transient', statusDetails: [ ErrorInfo { metadata: [Object], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } ], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } Error: Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4 analyze.js track-objects-gcs <gcsUri> Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API. Options: --version Show version number [boolean] --help Show help [boolean] Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'google.rpc.errorinfo-bin' => [Array], 'grpc-status-details-bin' => [Array], 'www-authenticate' => [Array], 'method' => [Array], 'service' => [Array], 'email' => [Array] }, options: {} }, note: 'Exception occurred in retry method that was not classified as transient', statusDetails: [ ErrorInfo { metadata: [Object], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } ], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (system-test/analyze.test.js:23:28) at Context.<anonymous> (system-test/analyze.test.js:48:20) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
analyze samples: should track objects in a GCS file failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2f7fe652af0d15621b89fc80cc22f9b2d0a2f209 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/24a5eecc-b508-4ba3-8f13-09342425c56a), [Sponge](http://sponge2/24a5eecc-b508-4ba3-8f13-09342425c56a) status: failed <details><summary>Test output</summary><br><pre>Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4 analyze.js track-objects-gcs <gcsUri> Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API. Options: --version Show version number [boolean] --help Show help [boolean] Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'google.rpc.errorinfo-bin' => [Array], 'grpc-status-details-bin' => [Array], 'www-authenticate' => [Array], 'method' => [Array], 'service' => [Array], 'email' => [Array] }, options: {} }, note: 'Exception occurred in retry method that was not classified as transient', statusDetails: [ ErrorInfo { metadata: [Object], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } ], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } Error: Command failed: node analyze.js track-objects-gcs gs://cloud-samples-data/video/cat.mp4 analyze.js track-objects-gcs <gcsUri> Analyzes objects in a video stored in Google Cloud Storage using the Cloud Video Intelligence API. Options: --version Show version number [boolean] --help Show help [boolean] Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:26) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:179:52) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'google.rpc.errorinfo-bin' => [Array], 'grpc-status-details-bin' => [Array], 'www-authenticate' => [Array], 'method' => [Array], 'service' => [Array], 'email' => [Array] }, options: {} }, note: 'Exception occurred in retry method that was not classified as transient', statusDetails: [ ErrorInfo { metadata: [Object], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } ], reason: 'ACCOUNT_STATE_INVALID', domain: 'googleapis.com' } at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (system-test/analyze.test.js:23:28) at Context.<anonymous> (system-test/analyze.test.js:48:20) at processImmediate (internal/timers.js:461:21)</pre></details>
priority
analyze samples should track objects in a gcs file failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node analyze js track objects gcs gs cloud samples data video cat analyze js track objects gcs analyzes objects in a video stored in google cloud storage using the cloud video intelligence api options version show version number help show help error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map google rpc errorinfo bin grpc status details bin www authenticate method service email options note exception occurred in retry method that was not classified as transient statusdetails errorinfo metadata reason account state invalid domain googleapis com reason account state invalid domain googleapis com error command failed node analyze js track objects gcs gs cloud samples data video cat analyze js track objects gcs analyzes objects in a video stored in google cloud storage using the cloud video intelligence api options version show version number help show help error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map google rpc errorinfo bin grpc status details bin www authenticate method service email options note exception occurred in retry method that was not classified as transient statusdetails errorinfo metadata reason account state invalid domain googleapis com reason account state invalid domain googleapis com at checkexecsyncerror child process js at object execsync child process js at execsync system test analyze test js at context system test analyze test js at processimmediate internal timers js
1
286,054
8,783,367,864
IssuesEvent
2018-12-20 05:32:54
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
closed
Project Clone does not copy allowed user list if private
Low Priority bug
Cloning a project keeps the private setting, but does not keep the user list.
1.0
Project Clone does not copy allowed user list if private - Cloning a project keeps the private setting, but does not keep the user list.
priority
project clone does not copy allowed user list if private cloning a project keeps the private setting but does not keep the user list
1
37,490
5,117,329,407
IssuesEvent
2017-01-07 15:41:46
ngageoint/hootenanny-ui
https://api.github.com/repos/ngageoint/hootenanny-ui
closed
Port refactored schema switcher code
Status: In Progress Status: Ready for Test
This feature was refactored in isolation https://github.com/brianhatchl/iD/pull/1 So may need some patching around the edges to pull out the old stuff and shoehorn the new.
1.0
Port refactored schema switcher code - This feature was refactored in isolation https://github.com/brianhatchl/iD/pull/1 So may need some patching around the edges to pull out the old stuff and shoehorn the new.
non_priority
port refactored schema switcher code this feature was refactored in isolation so may need some patching around the edges to pull out the old stuff and shoehorn the new
0
89,076
17,783,607,561
IssuesEvent
2021-08-31 08:26:18
zeek/spicy
https://api.github.com/repos/zeek/spicy
closed
Non-public enum type gets optimized out even though needed
Bug Codegen
If a unit depends on a non-public enum type it sometimes gets optimized out, even though it is needed. ``` $ cat x.spicy module x; type E = enum { a, b, c }; public type U = unit { var e: E; }; ``` ``` $ spicyc -j x.spicy -D global-optimizer [debug/global-optimizer] disabling feature 'supports_filters' of type 'x::U' since it is not used [debug/global-optimizer] disabling feature 'supports_sinks' of type 'x::U' since it is not used [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_init with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_error with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_done with default value [debug/global-optimizer] removing field for unused method x::U::__on_0x25_init [debug/global-optimizer] removing field for unused method x::U::__on_0x25_done [debug/global-optimizer] removing field for unused method x::U::__on_0x25_error [debug/global-optimizer] removing field for unused method x::U::__str__ [debug/global-optimizer] removing field for unused method x::U::__on_0x25_finally [debug/global-optimizer] removing field for unused method x::U::__on_0x25_gap [debug/global-optimizer] removing field for unused method x::U::__on_0x25_overlap [debug/global-optimizer] removing field for unused method x::U::__on_0x25_skipped [debug/global-optimizer] removing field for unused method x::U::__on_0x25_undelivered [debug/global-optimizer] removing declaration for unused function hilti::print [debug/global-optimizer] removing declaration for unused function hilti::printValues [debug/global-optimizer] removing declaration for unused function hilti::debug [debug/global-optimizer] removing declaration for unused function hilti::current_time [debug/global-optimizer] removing declaration for unused function hilti::mktime [debug/global-optimizer] removing declaration for unused function hilti::abort [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_connect [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward_eod [debug/global-optimizer] removing declaration for unused function spicy_rt::createContext [debug/global-optimizer] removing declaration for unused function spicy_rt::setContext [debug/global-optimizer] removing declaration for unused function spicy_rt::printParserState [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForEod [debug/global-optimizer] removing declaration for unused function spicy_rt::atEod [debug/global-optimizer] removing declaration for unused function spicy_rt::unit_find [debug/global-optimizer] removing declaration for unused function spicy_rt::backtrack [debug/global-optimizer] removing declaration for unused function spicy_rt::extractBits [debug/global-optimizer] removing unused type 'x::E' /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:24:12: error: no type named 'E' in namespace '__hlt::x' x::E e; ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x' inline U(std::optional<x::E> e); ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x' inline U(std::optional<x::E> e); ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:55:45: error: unknown type name 'E' inline auto operator<<(std::ostream& o, E x) -> std::ostream&; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: use of undeclared identifier 'x'; did you mean '__hlt::x'? inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ __hlt::x /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:21:18: note: '__hlt::x' declared here namespace __hlt::x { ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:17: error: redefinition of 'to_string' as different kind of symbol inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /usr/local/Cellar/spicy/HEAD-1f9a82a/include/hilti/rt/extension-points.h:17:20: note: previous definition is here extern std::string to_string(); ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: expected ')' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:26: note: to match this '(' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:37: error: no member named 'E' in namespace '__hlt::x' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:51: error: expected ';' after top level declarator inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ ; /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:78:16: error: no member named 'E' in namespace '__hlt::x' e = x::E::Undef; ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x' inline U::U(std::optional<x::E> e) : U() { ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x' inline U::U(std::optional<x::E> e) : U() { ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:198:51: error: unknown type name 'E' inline auto __hlt::x::operator<<(std::ostream& o, E x) -> std::ostream& { o << hilti::rt::to_string(x); return o; } ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:200:57: error: no type named 'E' in namespace '__hlt::x' inline auto hilti::rt::detail::adl::to_string(__hlt::x::E x, adl::tag) -> std::string { ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:202:24: error: no member named 'E' in namespace '__hlt::x' case __hlt::x::E::c: ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:204:24: error: no member named 'E' in namespace '__hlt::x' case __hlt::x::E::b: ~~~~~~~~~~^ fatal error: too many errors emitted, stopping now [-ferror-limit=] 20 errors generated. [error] JIT compilation failed [error] (no error output) ```
1.0
Non-public enum type gets optimized out even though needed - If a unit depends on a non-public enum type it sometimes gets optimized out, even though it is needed. ``` $ cat x.spicy module x; type E = enum { a, b, c }; public type U = unit { var e: E; }; ``` ``` $ spicyc -j x.spicy -D global-optimizer [debug/global-optimizer] disabling feature 'supports_filters' of type 'x::U' since it is not used [debug/global-optimizer] disabling feature 'supports_sinks' of type 'x::U' since it is not used [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_init with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_error with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_finally with default value [debug/global-optimizer] replacing call to unimplemented function x::U::__on_0x25_done with default value [debug/global-optimizer] removing field for unused method x::U::__on_0x25_init [debug/global-optimizer] removing field for unused method x::U::__on_0x25_done [debug/global-optimizer] removing field for unused method x::U::__on_0x25_error [debug/global-optimizer] removing field for unused method x::U::__str__ [debug/global-optimizer] removing field for unused method x::U::__on_0x25_finally [debug/global-optimizer] removing field for unused method x::U::__on_0x25_gap [debug/global-optimizer] removing field for unused method x::U::__on_0x25_overlap [debug/global-optimizer] removing field for unused method x::U::__on_0x25_skipped [debug/global-optimizer] removing field for unused method x::U::__on_0x25_undelivered [debug/global-optimizer] removing declaration for unused function hilti::print [debug/global-optimizer] removing declaration for unused function hilti::printValues [debug/global-optimizer] removing declaration for unused function hilti::debug [debug/global-optimizer] removing declaration for unused function hilti::current_time [debug/global-optimizer] removing declaration for unused function hilti::mktime [debug/global-optimizer] removing declaration for unused function hilti::abort [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_connect [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward [debug/global-optimizer] removing declaration for unused function spicy_rt::filter_forward_eod [debug/global-optimizer] removing declaration for unused function spicy_rt::createContext [debug/global-optimizer] removing declaration for unused function spicy_rt::setContext [debug/global-optimizer] removing declaration for unused function spicy_rt::printParserState [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInputOrEod [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForInput [debug/global-optimizer] removing declaration for unused function spicy_rt::waitForEod [debug/global-optimizer] removing declaration for unused function spicy_rt::atEod [debug/global-optimizer] removing declaration for unused function spicy_rt::unit_find [debug/global-optimizer] removing declaration for unused function spicy_rt::backtrack [debug/global-optimizer] removing declaration for unused function spicy_rt::extractBits [debug/global-optimizer] removing unused type 'x::E' /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:24:12: error: no type named 'E' in namespace '__hlt::x' x::E e; ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x' inline U(std::optional<x::E> e); ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:38:35: error: no member named 'E' in namespace '__hlt::x' inline U(std::optional<x::E> e); ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:55:45: error: unknown type name 'E' inline auto operator<<(std::ostream& o, E x) -> std::ostream&; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: use of undeclared identifier 'x'; did you mean '__hlt::x'? inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ __hlt::x /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:21:18: note: '__hlt::x' declared here namespace __hlt::x { ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:17: error: redefinition of 'to_string' as different kind of symbol inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /usr/local/Cellar/spicy/HEAD-1f9a82a/include/hilti/rt/extension-points.h:17:20: note: previous definition is here extern std::string to_string(); ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:39: error: expected ')' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:26: note: to match this '(' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:37: error: no member named 'E' in namespace '__hlt::x' inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:59:51: error: expected ';' after top level declarator inline auto to_string(__hlt::x::E x, adl::tag) -> std::string; ^ ; /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:209: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:322: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: error: non-constant-expression cannot be narrowed from type 'unsigned long' to 'std::ptrdiff_t' (aka 'long') in initializer list [-Wc++11-narrowing] const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~~~~ /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:73:453: note: insert an explicit cast to silence this issue const hilti::rt::TypeInfo __ti_x_U = { "x::U", "x::U", new hilti::rt::type_info::Struct(std::vector<hilti::rt::type_info::struct_::Field>({hilti::rt::type_info::struct_::Field{ "e", &type_info::__ti_x_E, offsetof(x::U, e), false }, hilti::rt::type_info::struct_::Field{ "__sink", &type_info::__ti_spicy_rt_SinkState, offsetof(x::U, __sink), true }, hilti::rt::type_info::struct_::Field{ "__filters", &type_info::__ti_strong_refx30spicy_rt_Filters, offsetof(x::U, __filters), true }})) }; ^~~~~~~~~~~~~~~~~~~~~~~~~ static_cast<ptrdiff_t>( ) /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/12.0.0/include/stddef.h:104:24: note: expanded from macro 'offsetof' #define offsetof(t, d) __builtin_offsetof(t, d) ^~~~~~~~~~~~~~~~~~~~~~~~ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:78:16: error: no member named 'E' in namespace '__hlt::x' e = x::E::Undef; ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x' inline U::U(std::optional<x::E> e) : U() { ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:81:34: error: no member named 'E' in namespace '__hlt::x' inline U::U(std::optional<x::E> e) : U() { ~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:198:51: error: unknown type name 'E' inline auto __hlt::x::operator<<(std::ostream& o, E x) -> std::ostream& { o << hilti::rt::to_string(x); return o; } ^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:200:57: error: no type named 'E' in namespace '__hlt::x' inline auto hilti::rt::detail::adl::to_string(__hlt::x::E x, adl::tag) -> std::string { ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:202:24: error: no member named 'E' in namespace '__hlt::x' case __hlt::x::E::c: ~~~~~~~~~~^ /private/var/folders/ht/vtb0rkbd4ws15vx02xdb84zw0000gn/T/hilti-rt-test-blY2IR/x.cc:204:24: error: no member named 'E' in namespace '__hlt::x' case __hlt::x::E::b: ~~~~~~~~~~^ fatal error: too many errors emitted, stopping now [-ferror-limit=] 20 errors generated. [error] JIT compilation failed [error] (no error output) ```
non_priority
non public enum type gets optimized out even though needed if a unit depends on a non public enum type it sometimes gets optimized out even though it is needed cat x spicy module x type e enum a b c public type u unit var e e spicyc j x spicy d global optimizer disabling feature supports filters of type x u since it is not used disabling feature supports sinks of type x u since it is not used replacing call to unimplemented function x u on init with default value replacing call to unimplemented function x u on error with default value replacing call to unimplemented function x u on finally with default value replacing call to unimplemented function x u on finally with default value replacing call to unimplemented function x u on done with default value removing field for unused method x u on init removing field for unused method x u on done removing field for unused method x u on error removing field for unused method x u str removing field for unused method x u on finally removing field for unused method x u on gap removing field for unused method x u on overlap removing field for unused method x u on skipped removing field for unused method x u on undelivered removing declaration for unused function hilti print removing declaration for unused function hilti printvalues removing declaration for unused function hilti debug removing declaration for unused function hilti current time removing declaration for unused function hilti mktime removing declaration for unused function hilti abort removing declaration for unused function spicy rt filter connect removing declaration for unused function spicy rt filter forward removing declaration for unused function spicy rt filter forward eod removing declaration for unused function spicy rt createcontext removing declaration for unused function spicy rt setcontext removing declaration for unused function spicy rt printparserstate removing declaration for unused function spicy rt waitforinputoreod removing declaration for unused function spicy rt waitforinputoreod removing declaration for unused function spicy rt waitforinput removing declaration for unused function spicy rt waitforinput removing declaration for unused function spicy rt waitforeod removing declaration for unused function spicy rt ateod removing declaration for unused function spicy rt unit find removing declaration for unused function spicy rt backtrack removing declaration for unused function spicy rt extractbits removing unused type x e private var folders ht t hilti rt test x cc error no type named e in namespace hlt x x e e private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u std optional e private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u std optional e private var folders ht t hilti rt test x cc error unknown type name e inline auto operator std ostream private var folders ht t hilti rt test x cc error use of undeclared identifier x did you mean hlt x inline auto to string hlt x e x adl tag std string hlt x private var folders ht t hilti rt test x cc note hlt x declared here namespace hlt x private var folders ht t hilti rt test x cc error redefinition of to string as different kind of symbol inline auto to string hlt x e x adl tag std string usr local cellar spicy head include hilti rt extension points h note previous definition is here extern std string to string private var folders ht t hilti rt test x cc error expected inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc note to match this inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error expected after top level declarator inline auto to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error non constant expression cannot be narrowed from type unsigned long to std ptrdiff t aka long in initializer list const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc note insert an explicit cast to silence this issue const hilti rt typeinfo ti x u x u x u new hilti rt type info struct std vector hilti rt type info struct field e type info ti x e offsetof x u e false hilti rt type info struct field sink type info ti spicy rt sinkstate offsetof x u sink true hilti rt type info struct field filters type info ti strong rt filters offsetof x u filters true static cast applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include stddef h note expanded from macro offsetof define offsetof t d builtin offsetof t d private var folders ht t hilti rt test x cc error no member named e in namespace hlt x e x e undef private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u u std optional e u private var folders ht t hilti rt test x cc error no member named e in namespace hlt x inline u u std optional e u private var folders ht t hilti rt test x cc error unknown type name e inline auto hlt x operator std ostream o hilti rt to string x return o private var folders ht t hilti rt test x cc error no type named e in namespace hlt x inline auto hilti rt detail adl to string hlt x e x adl tag std string private var folders ht t hilti rt test x cc error no member named e in namespace hlt x case hlt x e c private var folders ht t hilti rt test x cc error no member named e in namespace hlt x case hlt x e b fatal error too many errors emitted stopping now errors generated jit compilation failed no error output
0
352,961
25,091,913,520
IssuesEvent
2022-11-08 07:10:03
argilla-io/argilla
https://api.github.com/repos/argilla-io/argilla
closed
Docs: Replace one old UI screenshot
bug documentation
**Describe the bug** A screenshot from Rubrix UI persists on docs . **To Reproduce** Steps to reproduce the behavior: (weak labeling mode view) https://docs.argilla.io/en/latest/reference/webapp/features.html **Expected behavior** Update it with a screenshot from the new UI **Screenshots** If applicable, add screenshots to help explain your problem. **Environment (please complete the following information):** - **Additional context** -
1.0
Docs: Replace one old UI screenshot - **Describe the bug** A screenshot from Rubrix UI persists on docs . **To Reproduce** Steps to reproduce the behavior: (weak labeling mode view) https://docs.argilla.io/en/latest/reference/webapp/features.html **Expected behavior** Update it with a screenshot from the new UI **Screenshots** If applicable, add screenshots to help explain your problem. **Environment (please complete the following information):** - **Additional context** -
non_priority
docs replace one old ui screenshot describe the bug a screenshot from rubrix ui persists on docs to reproduce steps to reproduce the behavior weak labeling mode view expected behavior update it with a screenshot from the new ui screenshots if applicable add screenshots to help explain your problem environment please complete the following information additional context
0
207,140
16,066,867,139
IssuesEvent
2021-04-23 20:40:18
bounswe/2021SpringGroup12
https://api.github.com/repos/bounswe/2021SpringGroup12
closed
Merge Use Case Diagrams
documentation status: Review Request
First Deadline for review 23.04.2021, @18:00 Final Deadline 23.04.2021, @22:00 Others have drawn their parts on Lucid. I will merge them in one page.
1.0
Merge Use Case Diagrams - First Deadline for review 23.04.2021, @18:00 Final Deadline 23.04.2021, @22:00 Others have drawn their parts on Lucid. I will merge them in one page.
non_priority
merge use case diagrams first deadline for review final deadline others have drawn their parts on lucid i will merge them in one page
0
721,973
24,845,593,829
IssuesEvent
2022-10-26 15:38:26
googleapis/nodejs-pubsub
https://api.github.com/repos/googleapis/nodejs-pubsub
opened
createSubscription method throws if oidcToken is being set
priority: p2 type: bug
Thanks for stopping by to let us know something could be better! **PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response. 1) Is this a client library issue or a product issue? Client library issue 2) Did someone already solve this? This user faced something similar with Python and was solved by changing the Python version. https://github.com/googleapis/python-pubsub/issues/480 3) Do you have a support contract? No If the support paths suggested above still do not result in a resolution, please provide the following details. #### Environment details - OS: Mac OS Monterey 12.6 (also happens in firebase cloud functions environment) - Node.js version: 16.13.1 (also happens with node 14 and node 12) - npm version: 8.1.2 - `@google-cloud/pubsub` version: 3.2.0 #### Steps to reproduce 1. Create a topic. `await pubSubClient.createTopic(topicName)` 2. Create a subscription for the topic, setting OIDC token (either in pushConfig or in the root object as admitted by the sdk). ``` await pubSubClient.createSubscription(item.topicName, item.subscriptionName, { pushConfig: { pushEndpoint: item.endpoint, oidcToken: { serviceAccountEmail, audience: projectId, }, }, expirationPolicy: { ttl: null, }, retryPolicy: { minimumBackoff: { seconds: 10, }, maximumBackoff: { seconds: 60, }, }, }); })); ``` This also fails: ``` await pubSubClient.createSubscription(item.topicName, item.subscriptionName, { pushEndpoint: item.endpoint, oidcToken: { serviceAccountEmail, audience: projectId, }, expirationPolicy: { ttl: null, }, retryPolicy: { minimumBackoff: { seconds: 10, }, maximumBackoff: { seconds: 60, }, }, }); })); ``` This was working properly, we found the first error on Sep 29th 2022. Making sure to follow these steps will guarantee the quickest resolution possible. Thanks!
1.0
createSubscription method throws if oidcToken is being set - Thanks for stopping by to let us know something could be better! **PLEASE READ**: If you have a support contract with Google, please create an issue in the [support console](https://cloud.google.com/support/) instead of filing on GitHub. This will ensure a timely response. 1) Is this a client library issue or a product issue? Client library issue 2) Did someone already solve this? This user faced something similar with Python and was solved by changing the Python version. https://github.com/googleapis/python-pubsub/issues/480 3) Do you have a support contract? No If the support paths suggested above still do not result in a resolution, please provide the following details. #### Environment details - OS: Mac OS Monterey 12.6 (also happens in firebase cloud functions environment) - Node.js version: 16.13.1 (also happens with node 14 and node 12) - npm version: 8.1.2 - `@google-cloud/pubsub` version: 3.2.0 #### Steps to reproduce 1. Create a topic. `await pubSubClient.createTopic(topicName)` 2. Create a subscription for the topic, setting OIDC token (either in pushConfig or in the root object as admitted by the sdk). ``` await pubSubClient.createSubscription(item.topicName, item.subscriptionName, { pushConfig: { pushEndpoint: item.endpoint, oidcToken: { serviceAccountEmail, audience: projectId, }, }, expirationPolicy: { ttl: null, }, retryPolicy: { minimumBackoff: { seconds: 10, }, maximumBackoff: { seconds: 60, }, }, }); })); ``` This also fails: ``` await pubSubClient.createSubscription(item.topicName, item.subscriptionName, { pushEndpoint: item.endpoint, oidcToken: { serviceAccountEmail, audience: projectId, }, expirationPolicy: { ttl: null, }, retryPolicy: { minimumBackoff: { seconds: 10, }, maximumBackoff: { seconds: 60, }, }, }); })); ``` This was working properly, we found the first error on Sep 29th 2022. Making sure to follow these steps will guarantee the quickest resolution possible. Thanks!
priority
createsubscription method throws if oidctoken is being set thanks for stopping by to let us know something could be better please read if you have a support contract with google please create an issue in the instead of filing on github this will ensure a timely response is this a client library issue or a product issue client library issue did someone already solve this this user faced something similar with python and was solved by changing the python version do you have a support contract no if the support paths suggested above still do not result in a resolution please provide the following details environment details os mac os monterey also happens in firebase cloud functions environment node js version also happens with node and node npm version google cloud pubsub version steps to reproduce create a topic await pubsubclient createtopic topicname create a subscription for the topic setting oidc token either in pushconfig or in the root object as admitted by the sdk await pubsubclient createsubscription item topicname item subscriptionname pushconfig pushendpoint item endpoint oidctoken serviceaccountemail audience projectid expirationpolicy ttl null retrypolicy minimumbackoff seconds maximumbackoff seconds this also fails await pubsubclient createsubscription item topicname item subscriptionname pushendpoint item endpoint oidctoken serviceaccountemail audience projectid expirationpolicy ttl null retrypolicy minimumbackoff seconds maximumbackoff seconds this was working properly we found the first error on sep making sure to follow these steps will guarantee the quickest resolution possible thanks
1
158,288
6,025,001,058
IssuesEvent
2017-06-08 07:27:09
VirtoCommerce/vc-platform
https://api.github.com/repos/VirtoCommerce/vc-platform
closed
Theme colors stopped working
bug Priority: High
When changing theme default definition from Blue to Dark, color schema no longer changes, only left filter changes. This is a showcase feature during the demo and needs to work. The bug is current in dev branch.
1.0
Theme colors stopped working - When changing theme default definition from Blue to Dark, color schema no longer changes, only left filter changes. This is a showcase feature during the demo and needs to work. The bug is current in dev branch.
priority
theme colors stopped working when changing theme default definition from blue to dark color schema no longer changes only left filter changes this is a showcase feature during the demo and needs to work the bug is current in dev branch
1
26,426
26,853,938,009
IssuesEvent
2023-02-03 13:15:05
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
Confusing error message: Argument is too big for formatting
usability
``` SELECT format('{}asdfasd{}', '111') Query id: 77df41ec-4f04-4017-863f-f7d31b92893d 0 rows in set. Elapsed: 0.031 sec. Received exception from server (version 22.13.1): Code: 36. DB::Exception: Received from localhost:9000. DB::Exception: Argument is too big for formatting: While processing format('{}asdfasd{}', 'a'). (BAD_ARGUMENTS) ``` Should be something like 'Not enough arguments provided to fill all placeholders in the format string."'
True
Confusing error message: Argument is too big for formatting - ``` SELECT format('{}asdfasd{}', '111') Query id: 77df41ec-4f04-4017-863f-f7d31b92893d 0 rows in set. Elapsed: 0.031 sec. Received exception from server (version 22.13.1): Code: 36. DB::Exception: Received from localhost:9000. DB::Exception: Argument is too big for formatting: While processing format('{}asdfasd{}', 'a'). (BAD_ARGUMENTS) ``` Should be something like 'Not enough arguments provided to fill all placeholders in the format string."'
non_priority
confusing error message argument is too big for formatting select format asdfasd query id rows in set elapsed sec received exception from server version code db exception received from localhost db exception argument is too big for formatting while processing format asdfasd a bad arguments should be something like not enough arguments provided to fill all placeholders in the format string
0
614,857
19,191,045,360
IssuesEvent
2021-12-06 00:23:20
apcountryman/picolibrary
https://api.github.com/repos/apcountryman/picolibrary
closed
Remove reverse iterator
priority-normal status-awaiting_review type-refactoring
Remove reverse iterator (`::picolibrary::Reverse_Iterator`). `std::reverse_iterator` will be required instead.
1.0
Remove reverse iterator - Remove reverse iterator (`::picolibrary::Reverse_Iterator`). `std::reverse_iterator` will be required instead.
priority
remove reverse iterator remove reverse iterator picolibrary reverse iterator std reverse iterator will be required instead
1
79,561
28,375,113,659
IssuesEvent
2023-04-12 20:10:30
JohnAustinDev/xulsword
https://api.github.com/repos/JohnAustinDev/xulsword
closed
Add GUI capability to display the Greek TR words in module KJV version 2.6 lemma markup
Type-Defect Priority-Medium auto-migrated
``` From kjv.conf SwordVersionDate=2014-02-15 Version=2.6 History_2.6=Fixed bugs. Added Greek from TR. Example: (mod2imp) $$$Matthew 1:1 <w lemma="strong:G976 lemma.TR:βιβλος" morph="robinson:N-NSF" src="1">The book</w> <w lemma="strong:G1078 lemma.TR:γενεσεως" morph="robinson:N-GSF" src="2">of the generation</w> <w lemma="strong:G2424 lemma.TR:ιησου" morph="robinson:N-GSM" src="3">of Jesus</w> <w lemma="strong:G5547 lemma.TR:χριστου" morph="robinson:N-GSM" src="4">Christ</w>, <w lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="5">the son</w> <w lemma="strong:G1138 lemma.TR:δαβιδ" morph="robinson:N-PRI" src="6">of David</w>, <w lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="7">the son</w> <w lemma="strong:G11 lemma.TR:αβρααμ" morph="robinson:N-PRI" src="8">of Abraham</w>. ``` Original issue reported on code.google.com by `DFH...@gmail.com` on 16 Feb 2014 at 3:36
1.0
Add GUI capability to display the Greek TR words in module KJV version 2.6 lemma markup - ``` From kjv.conf SwordVersionDate=2014-02-15 Version=2.6 History_2.6=Fixed bugs. Added Greek from TR. Example: (mod2imp) $$$Matthew 1:1 <w lemma="strong:G976 lemma.TR:βιβλος" morph="robinson:N-NSF" src="1">The book</w> <w lemma="strong:G1078 lemma.TR:γενεσεως" morph="robinson:N-GSF" src="2">of the generation</w> <w lemma="strong:G2424 lemma.TR:ιησου" morph="robinson:N-GSM" src="3">of Jesus</w> <w lemma="strong:G5547 lemma.TR:χριστου" morph="robinson:N-GSM" src="4">Christ</w>, <w lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="5">the son</w> <w lemma="strong:G1138 lemma.TR:δαβιδ" morph="robinson:N-PRI" src="6">of David</w>, <w lemma="strong:G5207 lemma.TR:υιου" morph="robinson:N-GSM" src="7">the son</w> <w lemma="strong:G11 lemma.TR:αβρααμ" morph="robinson:N-PRI" src="8">of Abraham</w>. ``` Original issue reported on code.google.com by `DFH...@gmail.com` on 16 Feb 2014 at 3:36
non_priority
add gui capability to display the greek tr words in module kjv version lemma markup from kjv conf swordversiondate version history fixed bugs added greek from tr example matthew the book w lemma strong lemma tr γενεσεως morph robinson n gsf src of the generation w lemma strong lemma tr ιησου morph robinson n gsm src of jesus w lemma strong lemma tr χριστου morph robinson n gsm src christ w lemma strong lemma tr υιου morph robinson n gsm src the son w lemma strong lemma tr δαβιδ morph robinson n pri src of david w lemma strong lemma tr υιου morph robinson n gsm src the son w lemma strong lemma tr αβρααμ morph robinson n pri src of abraham original issue reported on code google com by dfh gmail com on feb at
0
59,707
12,013,373,137
IssuesEvent
2020-04-10 08:42:17
home-assistant/brands
https://api.github.com/repos/home-assistant/brands
closed
Philips TV is missing brand images
domain-missing has-codeowner
## The problem The Philips TV integration does not have brand images in this repository. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/philips_js/icon.png` - `src/philips_js/logo.png` - `src/philips_js/icon@2x.png` - `src/philips_js/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Updating the documentation repository Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository. If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: philips.png` line from this file: <https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/philips_js.markdown> **Note**: The documentation PR needs to be opened against the `current` branch. **Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently. ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @elupus! Mind taking a look at this issue as it is with an integration (philips_js) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/philips_js/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
1.0
Philips TV is missing brand images - ## The problem The Philips TV integration does not have brand images in this repository. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/philips_js/icon.png` - `src/philips_js/logo.png` - `src/philips_js/icon@2x.png` - `src/philips_js/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Updating the documentation repository Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository. If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: philips.png` line from this file: <https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/philips_js.markdown> **Note**: The documentation PR needs to be opened against the `current` branch. **Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently. ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @elupus! Mind taking a look at this issue as it is with an integration (philips_js) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/philips_js/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
non_priority
philips tv is missing brand images the problem the philips tv integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src philips js icon png src philips js logo png src philips js icon png src philips js logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo philips png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there elupus mind taking a look at this issue as it is with an integration philips js you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
0
446,697
12,876,716,568
IssuesEvent
2020-07-11 06:41:57
luksan47/mars
https://api.github.com/repos/luksan47/mars
closed
Deleting mac addresses is not working
Priority: HIGH bug
on list.blade and and admin.internet.mac_addresses.list.blade also
1.0
Deleting mac addresses is not working - on list.blade and and admin.internet.mac_addresses.list.blade also
priority
deleting mac addresses is not working on list blade and and admin internet mac addresses list blade also
1
709,164
24,369,190,466
IssuesEvent
2022-10-03 17:42:05
Chatterino/chatterino2
https://api.github.com/repos/Chatterino/chatterino2
closed
Migrate /subscribersoff command to Helix API
Platform: Twitch Priority: Medium Deprecation: Twitch IRC Commands hacktoberfest
As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/subscribersoff` command needs to be migrated to use the relevant Helix API endpoint. Helix API reference: https://dev.twitch.tv/docs/api/reference#update-chat-settings Split from #3956
1.0
Migrate /subscribersoff command to Helix API - As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/subscribersoff` command needs to be migrated to use the relevant Helix API endpoint. Helix API reference: https://dev.twitch.tv/docs/api/reference#update-chat-settings Split from #3956
priority
migrate subscribersoff command to helix api as part of twitch s announced deprecation of irc based commands the subscribersoff command needs to be migrated to use the relevant helix api endpoint helix api reference split from
1
102,769
11,307,004,396
IssuesEvent
2020-01-18 17:58:34
Luceapuce/SEPR-Project
https://api.github.com/repos/Luceapuce/SEPR-Project
closed
willCollide if spawned in Entity
documentation question
Might need to write something to handle the situation of an entity spawning within an entity. _This is just a reminder to check this later_
1.0
willCollide if spawned in Entity - Might need to write something to handle the situation of an entity spawning within an entity. _This is just a reminder to check this later_
non_priority
willcollide if spawned in entity might need to write something to handle the situation of an entity spawning within an entity this is just a reminder to check this later
0
51,739
6,195,202,786
IssuesEvent
2017-07-05 12:00:24
mifort-org/mifort-timesheet
https://api.github.com/repos/mifort-org/mifort-timesheet
closed
One week in month is not displayed
bug fixed on test env
When you enter new data in current week and then click on back arrow in the first column, data of previous week isn't displayed. If you reload window data is displayed again.
1.0
One week in month is not displayed - When you enter new data in current week and then click on back arrow in the first column, data of previous week isn't displayed. If you reload window data is displayed again.
non_priority
one week in month is not displayed when you enter new data in current week and then click on back arrow in the first column data of previous week isn t displayed if you reload window data is displayed again
0
148,431
19,531,026,647
IssuesEvent
2021-12-30 16:50:03
vital-ws/empty
https://api.github.com/repos/vital-ws/empty
closed
CVE-2016-1000236 (Medium) detected in cookie-signature-1.0.3.tgz
security vulnerability
## CVE-2016-1000236 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookie-signature-1.0.3.tgz</b></p></summary> <p>Sign and unsign cookies</p> <p>Library home page: <a href="https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz">https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/cookie-signature/package.json</p> <p> Dependency Hierarchy: - express-4.0.0.tgz (Root Library) - :x: **cookie-signature-1.0.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/empty/commit/8b832bf233deeb3f62310d7aa93eae79c2af69ed">8b832bf233deeb3f62310d7aa93eae79c2af69ed</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used. <p>Publish Date: 2019-11-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236>CVE-2016-1000236</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e">https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e</a></p> <p>Release Date: 2019-11-19</p> <p>Fix Resolution (cookie-signature): 1.0.6</p> <p>Direct dependency fix Resolution (express): 4.12.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"express:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.12.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-1000236","vulnerabilityDetails":"Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-1000236 (Medium) detected in cookie-signature-1.0.3.tgz - ## CVE-2016-1000236 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookie-signature-1.0.3.tgz</b></p></summary> <p>Sign and unsign cookies</p> <p>Library home page: <a href="https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz">https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/cookie-signature/package.json</p> <p> Dependency Hierarchy: - express-4.0.0.tgz (Root Library) - :x: **cookie-signature-1.0.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/empty/commit/8b832bf233deeb3f62310d7aa93eae79c2af69ed">8b832bf233deeb3f62310d7aa93eae79c2af69ed</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used. <p>Publish Date: 2019-11-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236>CVE-2016-1000236</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e">https://github.com/tj/node-cookie-signature/commit/39791081692e9e14aa62855369e1c7f80fbfd50e</a></p> <p>Release Date: 2019-11-19</p> <p>Fix Resolution (cookie-signature): 1.0.6</p> <p>Direct dependency fix Resolution (express): 4.12.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"express:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.12.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-1000236","vulnerabilityDetails":"Node-cookie-signature before 1.0.6 is affected by a timing attack due to the type of comparison used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000236","cvss3Severity":"medium","cvss3Score":"4.4","cvss3Metrics":{"A":"None","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in cookie signature tgz cve medium severity vulnerability vulnerable library cookie signature tgz sign and unsign cookies library home page a href path to dependency file package json path to vulnerable library node modules cookie signature package json dependency hierarchy express tgz root library x cookie signature tgz vulnerable library found in head commit a href found in base branch main vulnerability details node cookie signature before is affected by a timing attack due to the type of comparison used publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cookie signature direct dependency fix resolution express check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree express isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node cookie signature before is affected by a timing attack due to the type of comparison used vulnerabilityurl
0
174,885
27,747,982,185
IssuesEvent
2023-03-15 18:24:16
Olian04/simply-reactive
https://api.github.com/repos/Olian04/simply-reactive
closed
Add createSessionAtom
further design needed
Much like `createQueryAtom` however it stores the value in `sessionStorage` instead of the query string. ```ts const A = createSessionAtom({ key: 'a', default: 0, }); A.set(3); // Reload page A.get(); // 3 ```
1.0
Add createSessionAtom - Much like `createQueryAtom` however it stores the value in `sessionStorage` instead of the query string. ```ts const A = createSessionAtom({ key: 'a', default: 0, }); A.set(3); // Reload page A.get(); // 3 ```
non_priority
add createsessionatom much like createqueryatom however it stores the value in sessionstorage instead of the query string ts const a createsessionatom key a default a set reload page a get
0
56,313
15,020,016,603
IssuesEvent
2021-02-01 14:14:26
mozilla-lockwise/lockwise-ios
https://api.github.com/repos/mozilla-lockwise/lockwise-ios
reopened
recently used doesn't work
archived defect
steps to reproduce: - copy a username from an item - navigate back to entries list - sort by recently used expected: item with copied username appears at the top actual: not necessarily
1.0
recently used doesn't work - steps to reproduce: - copy a username from an item - navigate back to entries list - sort by recently used expected: item with copied username appears at the top actual: not necessarily
non_priority
recently used doesn t work steps to reproduce copy a username from an item navigate back to entries list sort by recently used expected item with copied username appears at the top actual not necessarily
0
194,890
6,900,397,676
IssuesEvent
2017-11-24 18:25:37
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
admin UI: Authentication sources rules description not displayed in listing
Priority: Low Status: For review Type: Feature / Enhancement Type: Nice to have
It would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it.
1.0
admin UI: Authentication sources rules description not displayed in listing - It would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it.
priority
admin ui authentication sources rules description not displayed in listing it would be better if we could see the description of the rules in authentication sources in the listing of the actual rules rather than having to click on each of them to see it
1
240,595
7,803,312,842
IssuesEvent
2018-06-10 22:18:13
DigitalCampus/django-oppia
https://api.github.com/repos/DigitalCampus/django-oppia
closed
When creating a "ResponseResource" object via the API not all fields are required
invalid medium priority
they should be!
1.0
When creating a "ResponseResource" object via the API not all fields are required - they should be!
priority
when creating a responseresource object via the api not all fields are required they should be
1
57,749
16,039,745,947
IssuesEvent
2021-04-22 06:03:14
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
New lightbox mouse wheel zoom too agressive
T-Defect
### Description Zooming the image by mouse wheel shows it in maximal/minimal zoom. Single wheel step should do the same thing as (+) and (-) buttons. ### Steps to reproduce - open the lightbox clicking an image - scrool mouse wheel one step up - image is shown in maximal zoom - scrool mouse wheel one step down - image is shown in default size Maximal/minimal zoom is ok when clicking the image by left mouse button, but mouse wheel should behave more fluently. ### Version information - **Platform**: Element Web 1.7.26-rc.1 - **Browser**: Firefox 89b2 - **OS**: Windows 10 - **URL**: private homeserver
1.0
New lightbox mouse wheel zoom too agressive - ### Description Zooming the image by mouse wheel shows it in maximal/minimal zoom. Single wheel step should do the same thing as (+) and (-) buttons. ### Steps to reproduce - open the lightbox clicking an image - scrool mouse wheel one step up - image is shown in maximal zoom - scrool mouse wheel one step down - image is shown in default size Maximal/minimal zoom is ok when clicking the image by left mouse button, but mouse wheel should behave more fluently. ### Version information - **Platform**: Element Web 1.7.26-rc.1 - **Browser**: Firefox 89b2 - **OS**: Windows 10 - **URL**: private homeserver
non_priority
new lightbox mouse wheel zoom too agressive description zooming the image by mouse wheel shows it in maximal minimal zoom single wheel step should do the same thing as and buttons steps to reproduce open the lightbox clicking an image scrool mouse wheel one step up image is shown in maximal zoom scrool mouse wheel one step down image is shown in default size maximal minimal zoom is ok when clicking the image by left mouse button but mouse wheel should behave more fluently version information platform element web rc browser firefox os windows url private homeserver
0
423,624
12,299,239,204
IssuesEvent
2020-05-11 12:01:39
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www5.ha.org.hk - site is not usable
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 76.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52664 --> **URL**: https://www5.ha.org.hk/rcbts **Browser / Version**: Firefox 76.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Secure Connection Failed An error occurred during a connection to www5.ha.org.hk. PR_CONNECT_RESET_ERROR The site loads fine in Chrome and Chromium Edge <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www5.ha.org.hk - site is not usable - <!-- @browser: Firefox 76.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52664 --> **URL**: https://www5.ha.org.hk/rcbts **Browser / Version**: Firefox 76.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Secure Connection Failed An error occurred during a connection to www5.ha.org.hk. PR_CONNECT_RESET_ERROR The site loads fine in Chrome and Chromium Edge <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
ha org hk site is not usable url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description page not loading correctly steps to reproduce secure connection failed an error occurred during a connection to ha org hk pr connect reset error the site loads fine in chrome and chromium edge browser configuration none from with ❤️
1
653,761
21,626,002,868
IssuesEvent
2022-05-05 02:18:17
bossbuwi/reality
https://api.github.com/repos/bossbuwi/reality
closed
Create event list view logic
enhancement logic high priority
The event list view must fetch and display data from the server. Clear any hard coded data.
1.0
Create event list view logic - The event list view must fetch and display data from the server. Clear any hard coded data.
priority
create event list view logic the event list view must fetch and display data from the server clear any hard coded data
1
307,274
9,415,227,394
IssuesEvent
2019-04-10 12:09:29
bio-tools/biotoolsRegistry
https://api.github.com/repos/bio-tools/biotoolsRegistry
closed
Count of search results not being cleared
GUI bug high priority
e.g. when navigating to Tool Card or other page (having done a search) the number of results remains visible in the search box - which looks bad.
1.0
Count of search results not being cleared - e.g. when navigating to Tool Card or other page (having done a search) the number of results remains visible in the search box - which looks bad.
priority
count of search results not being cleared e g when navigating to tool card or other page having done a search the number of results remains visible in the search box which looks bad
1
254,930
19,277,491,629
IssuesEvent
2021-12-10 13:36:06
PnX-SI/gn_module_monitoring
https://api.github.com/repos/PnX-SI/gn_module_monitoring
closed
Utilisation "filter" dans la configuration json
documentation question
Bonjour, J'ai configuré un sous-module pour nos suivis par tente malaise et autres pièges fixes. Je voudrais filtrer une liste de nomenclatures, j'ai vu que c'était prévu dans la doc. Mais (en tous cas avec ma conf) ça ne semble pas fonctionner. La configuration du widget est la suivante : ``` "id_trap_type": { "type_widget": "datalist", "attribut_label": "Type de piège", "api": "nomenclatures/nomenclature/METH_OBS", "application": "GeoNature", "keyValue": "id_nomenclature", "keyLabel": "label_fr", "data_path": "values", "type_util": "nomenclature", "required": true, "default": { "cd_nomenclature": "52" }, "filter" : { "cd_nomenclature":["52","53","73"] } ``` Il est bien probable que le soucis vienne de moi, je n'ai pas trouvé d'autres nomenclatures filtrées dans les autres modules déjà partagés. Avez-vous des pistes ? Merci d'avance !
1.0
Utilisation "filter" dans la configuration json - Bonjour, J'ai configuré un sous-module pour nos suivis par tente malaise et autres pièges fixes. Je voudrais filtrer une liste de nomenclatures, j'ai vu que c'était prévu dans la doc. Mais (en tous cas avec ma conf) ça ne semble pas fonctionner. La configuration du widget est la suivante : ``` "id_trap_type": { "type_widget": "datalist", "attribut_label": "Type de piège", "api": "nomenclatures/nomenclature/METH_OBS", "application": "GeoNature", "keyValue": "id_nomenclature", "keyLabel": "label_fr", "data_path": "values", "type_util": "nomenclature", "required": true, "default": { "cd_nomenclature": "52" }, "filter" : { "cd_nomenclature":["52","53","73"] } ``` Il est bien probable que le soucis vienne de moi, je n'ai pas trouvé d'autres nomenclatures filtrées dans les autres modules déjà partagés. Avez-vous des pistes ? Merci d'avance !
non_priority
utilisation filter dans la configuration json bonjour j ai configuré un sous module pour nos suivis par tente malaise et autres pièges fixes je voudrais filtrer une liste de nomenclatures j ai vu que c était prévu dans la doc mais en tous cas avec ma conf ça ne semble pas fonctionner la configuration du widget est la suivante id trap type type widget datalist attribut label type de piège api nomenclatures nomenclature meth obs application geonature keyvalue id nomenclature keylabel label fr data path values type util nomenclature required true default cd nomenclature filter cd nomenclature il est bien probable que le soucis vienne de moi je n ai pas trouvé d autres nomenclatures filtrées dans les autres modules déjà partagés avez vous des pistes merci d avance
0
255,443
19,302,624,690
IssuesEvent
2021-12-13 08:06:32
it-academyproject/ita-directory
https://api.github.com/repos/it-academyproject/ita-directory
closed
review documentation make install
bug documentation
- [ ] make install doesn't run => `make: *** No rule to make target 'install'. Stop.` - [ ] make build bug => ``` rm -f .env process_begin: CreateProcess(NULL, rm -f .env, ...) failed. make (e=2): El sistema no puede encontrar el archivo especificado. make: *** [C:/Users/TESTER/Desktop/it-projecte/ita-directory/scripts/make/build.mk:4: dev] Error 2 ``` - [ ] Need to clarify: where is the api-docs URL and how to get there.
1.0
review documentation make install - - [ ] make install doesn't run => `make: *** No rule to make target 'install'. Stop.` - [ ] make build bug => ``` rm -f .env process_begin: CreateProcess(NULL, rm -f .env, ...) failed. make (e=2): El sistema no puede encontrar el archivo especificado. make: *** [C:/Users/TESTER/Desktop/it-projecte/ita-directory/scripts/make/build.mk:4: dev] Error 2 ``` - [ ] Need to clarify: where is the api-docs URL and how to get there.
non_priority
review documentation make install make install doesn t run make no rule to make target install stop make build bug rm f env process begin createprocess null rm f env failed make e el sistema no puede encontrar el archivo especificado make error need to clarify where is the api docs url and how to get there
0
74,260
20,101,578,146
IssuesEvent
2022-02-07 05:16:34
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
tensorflow.datasets.load() throws an exception
stat:awaiting response type:build/install stalled subtype: ubuntu/linux
<em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): no - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: not a mobile device - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): v2.8.0-rc0-28-g24a4b3b5e58 2.8.0-rc1 - Python version: 3.8 - Bazel version (if compiling from source): 4.2.2 - GCC/Compiler version (if compiling from source): 9.3.0 - CUDA/cuDNN version: 11.2 / 8.1.1.33 - GPU model and memory: NVIDIA GeForce 940MX 1629 MB You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** After running the script `import tensorflow as tf import tensorflow_datasets as tfds print(tf.__version__) datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)` I am getting the following output: 2.8.0-rc1 terminate called after throwing an instance of 'std::system_error' what(): Invalid argument Process finished with exit code 134 (interrupted by signal 6: SIGABRT) **Describe the expected behavior** No exception is expected **[Contributing](https://www.tensorflow.org/community/contribute)** - Do you want to contribute a PR? (yes/no): no - Briefly describe your candidate solution(if contributing): **Standalone code to reproduce the issue** Provide a reproducible test case that is the bare minimum necessary to generate the problem. If possible, please share a link to Colab/Jupyter/any notebook. `import tensorflow as tf import tensorflow_datasets as tfds print(tf.__version__) datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)` **Other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
1.0
tensorflow.datasets.load() throws an exception - <em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): no - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: not a mobile device - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): v2.8.0-rc0-28-g24a4b3b5e58 2.8.0-rc1 - Python version: 3.8 - Bazel version (if compiling from source): 4.2.2 - GCC/Compiler version (if compiling from source): 9.3.0 - CUDA/cuDNN version: 11.2 / 8.1.1.33 - GPU model and memory: NVIDIA GeForce 940MX 1629 MB You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** After running the script `import tensorflow as tf import tensorflow_datasets as tfds print(tf.__version__) datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)` I am getting the following output: 2.8.0-rc1 terminate called after throwing an instance of 'std::system_error' what(): Invalid argument Process finished with exit code 134 (interrupted by signal 6: SIGABRT) **Describe the expected behavior** No exception is expected **[Contributing](https://www.tensorflow.org/community/contribute)** - Do you want to contribute a PR? (yes/no): no - Briefly describe your candidate solution(if contributing): **Standalone code to reproduce the issue** Provide a reproducible test case that is the bare minimum necessary to generate the problem. If possible, please share a link to Colab/Jupyter/any notebook. `import tensorflow as tf import tensorflow_datasets as tfds print(tf.__version__) datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True)` **Other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
non_priority
tensorflow datasets load throws an exception please make sure that this is a bug as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag bug template system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device not a mobile device tensorflow installed from source or binary source tensorflow version use command below python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory nvidia geforce mb you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior after running the script import tensorflow as tf import tensorflow datasets as tfds print tf version datasets info tfds load name mnist with info true as supervised true i am getting the following output terminate called after throwing an instance of std system error what invalid argument process finished with exit code interrupted by signal sigabrt describe the expected behavior no exception is expected do you want to contribute a pr yes no no briefly describe your candidate solution if contributing standalone code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem if possible please share a link to colab jupyter any notebook import tensorflow as tf import tensorflow datasets as tfds print tf version datasets info tfds load name mnist with info true as supervised true other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
0
315,167
27,051,136,237
IssuesEvent
2023-02-13 13:19:38
enonic/app-contentstudio
https://api.github.com/repos/enonic/app-contentstudio
closed
Project wizard - language selector gets disabled after clicking on 'Copy from parent' button
Bug Not in Changelog Test is Failing
1. Create a layer in Defalt project, do not select a language in the layer 2. Open the layer and select a language 3. Do not click on `Save` button in the wizard page, but click on `Copy from parent` button **BUG** - Filter input gets disabled in the selector ![image](https://user-images.githubusercontent.com/3728712/216944335-e66ac1c1-7d51-453c-9292-c7348a9df6b8.png)
1.0
Project wizard - language selector gets disabled after clicking on 'Copy from parent' button - 1. Create a layer in Defalt project, do not select a language in the layer 2. Open the layer and select a language 3. Do not click on `Save` button in the wizard page, but click on `Copy from parent` button **BUG** - Filter input gets disabled in the selector ![image](https://user-images.githubusercontent.com/3728712/216944335-e66ac1c1-7d51-453c-9292-c7348a9df6b8.png)
non_priority
project wizard language selector gets disabled after clicking on copy from parent button create a layer in defalt project do not select a language in the layer open the layer and select a language do not click on save button in the wizard page but click on copy from parent button bug filter input gets disabled in the selector
0
26,890
13,148,611,441
IssuesEvent
2020-08-08 22:45:53
E3SM-Project/E3SM
https://api.github.com/repos/E3SM-Project/E3SM
opened
output of chunk costs (print_cost_p) is expensive for large process counts
Atmosphere bug performance
The runtime of the routine print_cost_p, which writes out the cost of each chunk to the file atm_chunk_costs.txt, is large for large process counts on Cori-KNL. For example, for an F case using the ne256pg2 mesh and 32768 processes, this output took 18 minutes to complete. As print_cost_p is now called by default, this issue needs to be addressed quickly. Given the severity of the performance problem, I have applied a "bug" label to this issue, even though it is functionally correct. An alternative implementation has been developed, decreasing the runtime of print_cost_p for the above mentioned case from 18 minutes to less than 6 seconds. A PR with this change will be submitted soon.
True
output of chunk costs (print_cost_p) is expensive for large process counts - The runtime of the routine print_cost_p, which writes out the cost of each chunk to the file atm_chunk_costs.txt, is large for large process counts on Cori-KNL. For example, for an F case using the ne256pg2 mesh and 32768 processes, this output took 18 minutes to complete. As print_cost_p is now called by default, this issue needs to be addressed quickly. Given the severity of the performance problem, I have applied a "bug" label to this issue, even though it is functionally correct. An alternative implementation has been developed, decreasing the runtime of print_cost_p for the above mentioned case from 18 minutes to less than 6 seconds. A PR with this change will be submitted soon.
non_priority
output of chunk costs print cost p is expensive for large process counts the runtime of the routine print cost p which writes out the cost of each chunk to the file atm chunk costs txt is large for large process counts on cori knl for example for an f case using the mesh and processes this output took minutes to complete as print cost p is now called by default this issue needs to be addressed quickly given the severity of the performance problem i have applied a bug label to this issue even though it is functionally correct an alternative implementation has been developed decreasing the runtime of print cost p for the above mentioned case from minutes to less than seconds a pr with this change will be submitted soon
0
160,215
6,084,950,851
IssuesEvent
2017-06-17 09:43:21
climu/openstudyroom
https://api.github.com/repos/climu/openstudyroom
closed
Create a method that return all users one can play with
enhancement help wanted high priority
We need to calculate this players results for each divisions and then to filter the users with whom he didn't play the max_number of games.
1.0
Create a method that return all users one can play with - We need to calculate this players results for each divisions and then to filter the users with whom he didn't play the max_number of games.
priority
create a method that return all users one can play with we need to calculate this players results for each divisions and then to filter the users with whom he didn t play the max number of games
1
448,857
31,815,653,595
IssuesEvent
2023-09-13 20:14:30
proofcarryingdata/zupass
https://api.github.com/repos/proofcarryingdata/zupass
opened
example 'feed' application that demonstrates how 3rd parties are supposed to issue pcds to pcdpass users
documentation
cc @robknight
1.0
example 'feed' application that demonstrates how 3rd parties are supposed to issue pcds to pcdpass users - cc @robknight
non_priority
example feed application that demonstrates how parties are supposed to issue pcds to pcdpass users cc robknight
0
40,782
8,847,362,711
IssuesEvent
2019-01-08 01:22:07
pnp/pnpjs
https://api.github.com/repos/pnp/pnpjs
closed
Can't make batching work in sp-taxonomy
area: code status: details needed type: bug
### Category - [ ] Enhancement - [x] Bug - [x] Question - [ ] Documentation gap/issue ### Version Please specify what version of the library you are using: [1.2.7] Please specify what version(s) of SharePoint you are targeting: [SharePoint Online] ### Expected / Desired Behavior / Question Batching for `sp-taxonomy` works ### Observed Behavior I don't see any outgoing requests after `batch.execute` ### Steps to Reproduce ```typescript taxonomy.setup({ spfxContext: this.context }); var termStores = await taxonomy.termStores.get(); let returnTerms: IPickerTerm[] = []; for (let i = 0, len = termStores.length; i < len; i++) { const pnpTermStore = termStores[i]; const pnpTerms = await pnpTermStore.getTerms({ TermLabel: 'Mar', StringMatchOption: StringMatchOption.StartsWith, DefaultLabelOnly: true, TrimUnavailable: true, ResultCollectionSize: 10 }).get(); const batch = taxonomy.createBatch(); // in real scenario there should be pnpTerms.forEach const term = pnpTerms[0]; term.termSet.inBatch(batch).get().then(termSet => { console.log(termSet); }); term.labels.inBatch(batch).get().then(labels => { console.log(labels); }); await batch.execute(); ```
1.0
Can't make batching work in sp-taxonomy - ### Category - [ ] Enhancement - [x] Bug - [x] Question - [ ] Documentation gap/issue ### Version Please specify what version of the library you are using: [1.2.7] Please specify what version(s) of SharePoint you are targeting: [SharePoint Online] ### Expected / Desired Behavior / Question Batching for `sp-taxonomy` works ### Observed Behavior I don't see any outgoing requests after `batch.execute` ### Steps to Reproduce ```typescript taxonomy.setup({ spfxContext: this.context }); var termStores = await taxonomy.termStores.get(); let returnTerms: IPickerTerm[] = []; for (let i = 0, len = termStores.length; i < len; i++) { const pnpTermStore = termStores[i]; const pnpTerms = await pnpTermStore.getTerms({ TermLabel: 'Mar', StringMatchOption: StringMatchOption.StartsWith, DefaultLabelOnly: true, TrimUnavailable: true, ResultCollectionSize: 10 }).get(); const batch = taxonomy.createBatch(); // in real scenario there should be pnpTerms.forEach const term = pnpTerms[0]; term.termSet.inBatch(batch).get().then(termSet => { console.log(termSet); }); term.labels.inBatch(batch).get().then(labels => { console.log(labels); }); await batch.execute(); ```
non_priority
can t make batching work in sp taxonomy category enhancement bug question documentation gap issue version please specify what version of the library you are using please specify what version s of sharepoint you are targeting expected desired behavior question batching for sp taxonomy works observed behavior i don t see any outgoing requests after batch execute steps to reproduce typescript taxonomy setup spfxcontext this context var termstores await taxonomy termstores get let returnterms ipickerterm for let i len termstores length i len i const pnptermstore termstores const pnpterms await pnptermstore getterms termlabel mar stringmatchoption stringmatchoption startswith defaultlabelonly true trimunavailable true resultcollectionsize get const batch taxonomy createbatch in real scenario there should be pnpterms foreach const term pnpterms term termset inbatch batch get then termset console log termset term labels inbatch batch get then labels console log labels await batch execute
0
323,683
27,746,189,647
IssuesEvent
2023-03-15 17:11:52
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] flaky test: YbAdminSnapshotScheduleUpgradeTestWithYsql.PgsqlTestMigrationFromEarliestSysCatalogSnapshot
kind/bug area/ysql kind/failing-test priority/medium
Jira Link: [DB-5838](https://yugabyte.atlassian.net/browse/DB-5838) ### Description From https://detective-gcp.dev.yugabyte.com/stability/test?class=YbAdminSnapshotScheduleUpgradeTestWithYsql&name=PgsqlTestMigrationFromEarliestSysCatalogSnapshot it shows in debug build the test frequently failed with the following error: ``` ../../src/yb/tools/yb-admin-snapshot-schedule-test.cc:2598 Failed Expected to find substring 'Snapshot state and current state have different system catalogs'. Got: 'Subprocess '/net/ip-10-9-10-29.us-west-2.compute.internal/share/jenkins/jenkins-github-yugabyte-db-alma8-master-clang15-debug-aarch64-625/yugabyte-db/build/debug-clang15-dynamic-ninja/tests-tools/../bin/yb-admin' terminated with non-zero exit status 256: Error running restore_snapshot_schedule: Illegal state (yb/master/master_snapshot_coordinator.cc:192): Trying to restore to { days: 19430 time: 19:28:57.612482 } which is earlier than the configured retention. Not allowed. Earliest snapshot that can be used is c2a3ab4c-c14c-40e6-9c49-e27d3f8798d9 and was taken at { days: 19430 time: 19:29:53.222443 }. ``` Under debug build, it took the test longer than 10 minutes and the old snapshots taken 10 minutes ago already expired. We just need to bump the 10 minutes to a larger duration. ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information. [DB-5838]: https://yugabyte.atlassian.net/browse/DB-5838?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
1.0
[YSQL] flaky test: YbAdminSnapshotScheduleUpgradeTestWithYsql.PgsqlTestMigrationFromEarliestSysCatalogSnapshot - Jira Link: [DB-5838](https://yugabyte.atlassian.net/browse/DB-5838) ### Description From https://detective-gcp.dev.yugabyte.com/stability/test?class=YbAdminSnapshotScheduleUpgradeTestWithYsql&name=PgsqlTestMigrationFromEarliestSysCatalogSnapshot it shows in debug build the test frequently failed with the following error: ``` ../../src/yb/tools/yb-admin-snapshot-schedule-test.cc:2598 Failed Expected to find substring 'Snapshot state and current state have different system catalogs'. Got: 'Subprocess '/net/ip-10-9-10-29.us-west-2.compute.internal/share/jenkins/jenkins-github-yugabyte-db-alma8-master-clang15-debug-aarch64-625/yugabyte-db/build/debug-clang15-dynamic-ninja/tests-tools/../bin/yb-admin' terminated with non-zero exit status 256: Error running restore_snapshot_schedule: Illegal state (yb/master/master_snapshot_coordinator.cc:192): Trying to restore to { days: 19430 time: 19:28:57.612482 } which is earlier than the configured retention. Not allowed. Earliest snapshot that can be used is c2a3ab4c-c14c-40e6-9c49-e27d3f8798d9 and was taken at { days: 19430 time: 19:29:53.222443 }. ``` Under debug build, it took the test longer than 10 minutes and the old snapshots taken 10 minutes ago already expired. We just need to bump the 10 minutes to a larger duration. ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information. [DB-5838]: https://yugabyte.atlassian.net/browse/DB-5838?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
non_priority
flaky test ybadminsnapshotscheduleupgradetestwithysql pgsqltestmigrationfromearliestsyscatalogsnapshot jira link description from it shows in debug build the test frequently failed with the following error src yb tools yb admin snapshot schedule test cc failed expected to find substring snapshot state and current state have different system catalogs got subprocess net ip us west compute internal share jenkins jenkins github yugabyte db master debug yugabyte db build debug dynamic ninja tests tools bin yb admin terminated with non zero exit status error running restore snapshot schedule illegal state yb master master snapshot coordinator cc trying to restore to days time which is earlier than the configured retention not allowed earliest snapshot that can be used is and was taken at days time under debug build it took the test longer than minutes and the old snapshots taken minutes ago already expired we just need to bump the minutes to a larger duration warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information
0
283,407
21,316,554,320
IssuesEvent
2022-04-16 11:33:14
jr-mojito/pe
https://api.github.com/repos/jr-mojito/pe
opened
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function
severity.Low type.DocumentationBug
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function ![Screenshot 2022-04-16 at 7.14.40 PM.png](https://raw.githubusercontent.com/jr-mojito/pe/main/files/5b1a0ed8-3f15-45e5-960a-dfa13c61cdd8.png) <!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778--> <!--Version: Web v3.4.2-->
1.0
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function - Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function ![Screenshot 2022-04-16 at 7.14.40 PM.png](https://raw.githubusercontent.com/jr-mojito/pe/main/files/5b1a0ed8-3f15-45e5-960a-dfa13c61cdd8.png) <!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778--> <!--Version: Web v3.4.2-->
non_priority
option field for found on ug but not available for edit function option field for found on ug but not available for edit function
0
85,136
3,686,993,306
IssuesEvent
2016-02-25 05:13:48
cs2103jan2016-w13-4j/main
https://api.github.com/repos/cs2103jan2016-w13-4j/main
closed
Create method that allows removing one tag from a task
component.storage priority.medium
Something like `Task.removeTag(int id, String tag)`
1.0
Create method that allows removing one tag from a task - Something like `Task.removeTag(int id, String tag)`
priority
create method that allows removing one tag from a task something like task removetag int id string tag
1
216,650
16,794,397,425
IssuesEvent
2021-06-16 00:04:01
microsoft/appcenter
https://api.github.com/repos/microsoft/appcenter
closed
Inconclusive and Skipped tests missing in the Test run results on the main page
Portal feature request test
In the log I see Test Count: 84, Passed: 74, Failed: 8, Warnings: 0, Inconclusive: 0, Skipped: 2 On the main page I see ![image](https://user-images.githubusercontent.com/45621056/111629280-ccf2c200-87f9-11eb-9039-44ed8e40e1d7.png) Will be great to see the real count of tests on the main page.
1.0
Inconclusive and Skipped tests missing in the Test run results on the main page - In the log I see Test Count: 84, Passed: 74, Failed: 8, Warnings: 0, Inconclusive: 0, Skipped: 2 On the main page I see ![image](https://user-images.githubusercontent.com/45621056/111629280-ccf2c200-87f9-11eb-9039-44ed8e40e1d7.png) Will be great to see the real count of tests on the main page.
non_priority
inconclusive and skipped tests missing in the test run results on the main page in the log i see test count passed failed warnings inconclusive skipped on the main page i see will be great to see the real count of tests on the main page
0
484,679
13,943,533,823
IssuesEvent
2020-10-22 23:22:26
SkriptLang/Skript
https://api.github.com/repos/SkriptLang/Skript
closed
Can't check if item has NO name applied - Beta 4
bug completed dev needed priority: low
I used this line of code to check if player had X regular diamonds (that had no special name or lore): `if arg-player has arg-number of diamonds named "":` After updating to beta 2, beta 3 i started to get this error when loading the skript: `the 1st argument of diamond named "" is not an item type (atm_diamond.sk, line 233: if arg-player has arg-number of diamonds named "":')` After updating to beta 4 the error is gone, **but the syntax is still not working**! I am trying to report this since a month... am I the only fool out there in the need to remove unnamed items from inventory in an efficient way? ### Steps to Reproduce As simple as this code: ![image](https://user-images.githubusercontent.com/25549272/94337468-c756b500-ffea-11ea-9b1f-733332bf93d2.png) ### Expected Behavior Only UNNAMED items should get removed ### Errors / Screenshots There are no errors in beta 4 Here is a small video to explain what's happening https://youtu.be/rvMvsi-3akQ ### Server Information * **Server version/platform:** Paper-199 1.16.3 * **Skript version:** 2.5-beta4
1.0
Can't check if item has NO name applied - Beta 4 - I used this line of code to check if player had X regular diamonds (that had no special name or lore): `if arg-player has arg-number of diamonds named "":` After updating to beta 2, beta 3 i started to get this error when loading the skript: `the 1st argument of diamond named "" is not an item type (atm_diamond.sk, line 233: if arg-player has arg-number of diamonds named "":')` After updating to beta 4 the error is gone, **but the syntax is still not working**! I am trying to report this since a month... am I the only fool out there in the need to remove unnamed items from inventory in an efficient way? ### Steps to Reproduce As simple as this code: ![image](https://user-images.githubusercontent.com/25549272/94337468-c756b500-ffea-11ea-9b1f-733332bf93d2.png) ### Expected Behavior Only UNNAMED items should get removed ### Errors / Screenshots There are no errors in beta 4 Here is a small video to explain what's happening https://youtu.be/rvMvsi-3akQ ### Server Information * **Server version/platform:** Paper-199 1.16.3 * **Skript version:** 2.5-beta4
priority
can t check if item has no name applied beta i used this line of code to check if player had x regular diamonds that had no special name or lore if arg player has arg number of diamonds named after updating to beta beta i started to get this error when loading the skript the argument of diamond named is not an item type atm diamond sk line if arg player has arg number of diamonds named after updating to beta the error is gone but the syntax is still not working i am trying to report this since a month am i the only fool out there in the need to remove unnamed items from inventory in an efficient way steps to reproduce as simple as this code expected behavior only unnamed items should get removed errors screenshots there are no errors in beta here is a small video to explain what s happening server information server version platform paper skript version
1
94,251
10,817,094,205
IssuesEvent
2019-11-08 08:57:30
influxdata/influxdb-client-java
https://api.github.com/repos/influxdata/influxdb-client-java
closed
Documentation: README.md documents example that should be improved
documentation
The current README.md code documents classes that were renamed. ``` RetentionRule retention = new RetentionRule(); retention.setEverySeconds(3600L); ``` Correct example is: ``` BucketRetentionRules bucketRetentionRules = new BucketRetentionRules(); bucketRetentionRules.setEverySeconds(3600L); ```` It would also be nice to document here the default RP, and how I as a user can set up the autogen policy.
1.0
Documentation: README.md documents example that should be improved - The current README.md code documents classes that were renamed. ``` RetentionRule retention = new RetentionRule(); retention.setEverySeconds(3600L); ``` Correct example is: ``` BucketRetentionRules bucketRetentionRules = new BucketRetentionRules(); bucketRetentionRules.setEverySeconds(3600L); ```` It would also be nice to document here the default RP, and how I as a user can set up the autogen policy.
non_priority
documentation readme md documents example that should be improved the current readme md code documents classes that were renamed retentionrule retention new retentionrule retention seteveryseconds correct example is bucketretentionrules bucketretentionrules new bucketretentionrules bucketretentionrules seteveryseconds it would also be nice to document here the default rp and how i as a user can set up the autogen policy
0
831,536
32,051,727,476
IssuesEvent
2023-09-23 16:33:41
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Action Containers broke spellbooks + exception
Issue: Bug Priority: 2-Before Release Difficulty: 2-Medium
## Description <!-- Explain your issue in detail. Issues without proper explanation are liable to be closed by maintainers. --> An exception is thrown when trying to learn spells from spellbooks. The recent action container PR broke this behavior: #20260 **Reproduction** <!-- Include the steps to reproduce if applicable. --> Pick up a spellbook and try to learn it, console will show the exception. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> **Additional context** <!-- Add any other context about the problem here. Anything you think is related to the issue. --> > [ERRO] system.actions: Can't resolve "Content.Shared.Actions.ActionsContainerComponent" on entity 1985! > at Content.Shared.Actions.SharedActionsSystem.GrantActions(EntityUid performer, IEnumerable`1 actions, EntityUid container, ActionsComponent comp, ActionsContainerComponent containerComp) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Actions\SharedActionsSystem.cs:l > ine 523 > at Content.Shared.Magic.SharedMagicSystem.OnDoAfter(EntityUid uid, SpellbookComponent component, DoAfterEvent args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Magic\SharedMagicSystem.cs:line 85 > at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass47_0`2.<SubscribeLocalEvent>g__EventHandler|0(EntityUid uid, IComponent comp, TEvent& args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 254 > at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass57_0`1.<EntSubscribe>b__0(EntityUid uid, IComponent comp, Unit& ev) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 387 > at Robust.Shared.GameObjects.EntityEventBus.EntDispatch(EntityUid euid, Type eventType, Unit& args, Boolean dispatchByReference) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 541 > at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEventCore(EntityUid uid, Unit& unitRef, Type type, Boolean broadcast, Boolean byRef) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 221 > at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 190 > at Robust.Shared.GameObjects.EntitySystem.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystem.cs:line 155 > at Content.Shared.DoAfter.SharedDoAfterSystem.RaiseDoAfterEvents(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.cs:line 93 > at Content.Shared.DoAfter.SharedDoAfterSystem.TryComplete(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 123 > at Content.Shared.DoAfter.SharedDoAfterSystem.Update(EntityUid uid, ActiveDoAfterComponent active, DoAfterComponent comp, TimeSpan time, EntityQuery`1 xformQuery, EntityQuery`1 handsQuery) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Upd > ate.cs:line 68 > at Content.Shared.DoAfter.SharedDoAfterSystem.Update(Single frameTime) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 23 > at Robust.Shared.GameObjects.EntitySystemManager.TickUpdate(Single frameTime, Boolean noPredictions) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystemManager.cs:line 306 > at Robust.Shared.GameObjects.EntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityManager.cs:line 278 > at Robust.Server.GameObjects.ServerEntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\GameObjects\ServerEntityManager.cs:line 158 > at Robust.Server.BaseServer.Update(FrameEventArgs frameEventArgs) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 719 > at Robust.Server.BaseServer.<SetupMainLoop>b__66_1(Object sender, FrameEventArgs args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 536 > at Robust.Shared.Timing.GameLoop.Run() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\Timing\GameLoop.cs:line 235 > at Robust.Server.BaseServer.MainLoop() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 563 > at Robust.Server.Program.ParsedMain(CommandLineArgs args, Boolean contentStart, ServerOptions options) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 78 > at Robust.Server.Program.Start(String[] args, ServerOptions options, Boolean contentStart) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 46 > at Robust.Server.ContentStart.Start(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\ContentStart.cs:line 10 > at Content.Server.Program.Main(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Server\Program.cs:line 9
1.0
Action Containers broke spellbooks + exception - ## Description <!-- Explain your issue in detail. Issues without proper explanation are liable to be closed by maintainers. --> An exception is thrown when trying to learn spells from spellbooks. The recent action container PR broke this behavior: #20260 **Reproduction** <!-- Include the steps to reproduce if applicable. --> Pick up a spellbook and try to learn it, console will show the exception. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> **Additional context** <!-- Add any other context about the problem here. Anything you think is related to the issue. --> > [ERRO] system.actions: Can't resolve "Content.Shared.Actions.ActionsContainerComponent" on entity 1985! > at Content.Shared.Actions.SharedActionsSystem.GrantActions(EntityUid performer, IEnumerable`1 actions, EntityUid container, ActionsComponent comp, ActionsContainerComponent containerComp) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Actions\SharedActionsSystem.cs:l > ine 523 > at Content.Shared.Magic.SharedMagicSystem.OnDoAfter(EntityUid uid, SpellbookComponent component, DoAfterEvent args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\Magic\SharedMagicSystem.cs:line 85 > at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass47_0`2.<SubscribeLocalEvent>g__EventHandler|0(EntityUid uid, IComponent comp, TEvent& args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 254 > at Robust.Shared.GameObjects.EntityEventBus.<>c__DisplayClass57_0`1.<EntSubscribe>b__0(EntityUid uid, IComponent comp, Unit& ev) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 387 > at Robust.Shared.GameObjects.EntityEventBus.EntDispatch(EntityUid euid, Type eventType, Unit& args, Boolean dispatchByReference) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 541 > at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEventCore(EntityUid uid, Unit& unitRef, Type type, Boolean broadcast, Boolean byRef) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 221 > at Robust.Shared.GameObjects.EntityEventBus.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityEventBus.Directed.cs:line 190 > at Robust.Shared.GameObjects.EntitySystem.RaiseLocalEvent(EntityUid uid, Object args, Boolean broadcast) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystem.cs:line 155 > at Content.Shared.DoAfter.SharedDoAfterSystem.RaiseDoAfterEvents(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.cs:line 93 > at Content.Shared.DoAfter.SharedDoAfterSystem.TryComplete(DoAfter doAfter, DoAfterComponent component) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 123 > at Content.Shared.DoAfter.SharedDoAfterSystem.Update(EntityUid uid, ActiveDoAfterComponent active, DoAfterComponent comp, TimeSpan time, EntityQuery`1 xformQuery, EntityQuery`1 handsQuery) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Upd > ate.cs:line 68 > at Content.Shared.DoAfter.SharedDoAfterSystem.Update(Single frameTime) in J:\Program Files\SS14CloneNew\space-station-14\Content.Shared\DoAfter\SharedDoAfterSystem.Update.cs:line 23 > at Robust.Shared.GameObjects.EntitySystemManager.TickUpdate(Single frameTime, Boolean noPredictions) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntitySystemManager.cs:line 306 > at Robust.Shared.GameObjects.EntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\GameObjects\EntityManager.cs:line 278 > at Robust.Server.GameObjects.ServerEntityManager.TickUpdate(Single frameTime, Boolean noPredictions, Histogram histogram) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\GameObjects\ServerEntityManager.cs:line 158 > at Robust.Server.BaseServer.Update(FrameEventArgs frameEventArgs) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 719 > at Robust.Server.BaseServer.<SetupMainLoop>b__66_1(Object sender, FrameEventArgs args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 536 > at Robust.Shared.Timing.GameLoop.Run() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Shared\Timing\GameLoop.cs:line 235 > at Robust.Server.BaseServer.MainLoop() in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\BaseServer.cs:line 563 > at Robust.Server.Program.ParsedMain(CommandLineArgs args, Boolean contentStart, ServerOptions options) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 78 > at Robust.Server.Program.Start(String[] args, ServerOptions options, Boolean contentStart) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\Program.cs:line 46 > at Robust.Server.ContentStart.Start(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\RobustToolbox\Robust.Server\ContentStart.cs:line 10 > at Content.Server.Program.Main(String[] args) in J:\Program Files\SS14CloneNew\space-station-14\Content.Server\Program.cs:line 9
priority
action containers broke spellbooks exception description an exception is thrown when trying to learn spells from spellbooks the recent action container pr broke this behavior reproduction pick up a spellbook and try to learn it console will show the exception screenshots additional context system actions can t resolve content shared actions actionscontainercomponent on entity at content shared actions sharedactionssystem grantactions entityuid performer ienumerable actions entityuid container actionscomponent comp actionscontainercomponent containercomp in j program files space station content shared actions sharedactionssystem cs l ine at content shared magic sharedmagicsystem ondoafter entityuid uid spellbookcomponent component doafterevent args in j program files space station content shared magic sharedmagicsystem cs line at robust shared gameobjects entityeventbus c g eventhandler entityuid uid icomponent comp tevent args in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus c b entityuid uid icomponent comp unit ev in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus entdispatch entityuid euid type eventtype unit args boolean dispatchbyreference in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus raiselocaleventcore entityuid uid unit unitref type type boolean broadcast boolean byref in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entityeventbus raiselocalevent entityuid uid object args boolean broadcast in j program files space station robusttoolbox robust shared gameobjects entityeventbus directed cs line at robust shared gameobjects entitysystem raiselocalevent entityuid uid object args boolean broadcast in j program files space station robusttoolbox robust shared gameobjects entitysystem cs line at content shared doafter shareddoaftersystem raisedoafterevents doafter doafter doaftercomponent component in j program files space station content shared doafter shareddoaftersystem cs line at content shared doafter shareddoaftersystem trycomplete doafter doafter doaftercomponent component in j program files space station content shared doafter shareddoaftersystem update cs line at content shared doafter shareddoaftersystem update entityuid uid activedoaftercomponent active doaftercomponent comp timespan time entityquery xformquery entityquery handsquery in j program files space station content shared doafter shareddoaftersystem upd ate cs line at content shared doafter shareddoaftersystem update single frametime in j program files space station content shared doafter shareddoaftersystem update cs line at robust shared gameobjects entitysystemmanager tickupdate single frametime boolean nopredictions in j program files space station robusttoolbox robust shared gameobjects entitysystemmanager cs line at robust shared gameobjects entitymanager tickupdate single frametime boolean nopredictions histogram histogram in j program files space station robusttoolbox robust shared gameobjects entitymanager cs line at robust server gameobjects serverentitymanager tickupdate single frametime boolean nopredictions histogram histogram in j program files space station robusttoolbox robust server gameobjects serverentitymanager cs line at robust server baseserver update frameeventargs frameeventargs in j program files space station robusttoolbox robust server baseserver cs line at robust server baseserver b object sender frameeventargs args in j program files space station robusttoolbox robust server baseserver cs line at robust shared timing gameloop run in j program files space station robusttoolbox robust shared timing gameloop cs line at robust server baseserver mainloop in j program files space station robusttoolbox robust server baseserver cs line at robust server program parsedmain commandlineargs args boolean contentstart serveroptions options in j program files space station robusttoolbox robust server program cs line at robust server program start string args serveroptions options boolean contentstart in j program files space station robusttoolbox robust server program cs line at robust server contentstart start string args in j program files space station robusttoolbox robust server contentstart cs line at content server program main string args in j program files space station content server program cs line
1
163,275
20,356,090,147
IssuesEvent
2022-02-20 00:37:34
eclipse/che
https://api.github.com/repos/eclipse/che
closed
REST Service on the Che server-side that will manage user secrets
kind/enhancement severity/P1 lifecycle/stale area/dev-experience area/security
### Is your enhancement related to a problem? Please describe. Some che users need to have some secrets on personal workspace. Secrets can be: - github / SCM token - files that contain secrets - Environnment variables Eclipse CHE provide some function to handle this. All of them consume existings secrets from user namespace on K8s (with some labels/annotations). The problem is that major of user don't have direct access on kubernetes (or knowledge) to create this secrets. Moreover, is not really user friendly to know all annotations/labels needed (depend of why we want). ### Describe the solution you'd like Create user secrets on k8s from chectl (like we can manage workspace) and/or from Eclpise Che UI. This wrapper must handle the right labels/annotions needed (secret as env, secret as file, secret as github token, etc ...) To do that, we need some new api to handle secrets from Eclpise che server. ### Describe alternatives you've considered <!--A clear and concise description of any alternative solutions or features you've considered. --> ### Additional context <!-- Add any other context or screenshots about the enhancement here. -->
True
REST Service on the Che server-side that will manage user secrets - ### Is your enhancement related to a problem? Please describe. Some che users need to have some secrets on personal workspace. Secrets can be: - github / SCM token - files that contain secrets - Environnment variables Eclipse CHE provide some function to handle this. All of them consume existings secrets from user namespace on K8s (with some labels/annotations). The problem is that major of user don't have direct access on kubernetes (or knowledge) to create this secrets. Moreover, is not really user friendly to know all annotations/labels needed (depend of why we want). ### Describe the solution you'd like Create user secrets on k8s from chectl (like we can manage workspace) and/or from Eclpise Che UI. This wrapper must handle the right labels/annotions needed (secret as env, secret as file, secret as github token, etc ...) To do that, we need some new api to handle secrets from Eclpise che server. ### Describe alternatives you've considered <!--A clear and concise description of any alternative solutions or features you've considered. --> ### Additional context <!-- Add any other context or screenshots about the enhancement here. -->
non_priority
rest service on the che server side that will manage user secrets is your enhancement related to a problem please describe some che users need to have some secrets on personal workspace secrets can be github scm token files that contain secrets environnment variables eclipse che provide some function to handle this all of them consume existings secrets from user namespace on with some labels annotations the problem is that major of user don t have direct access on kubernetes or knowledge to create this secrets moreover is not really user friendly to know all annotations labels needed depend of why we want describe the solution you d like create user secrets on from chectl like we can manage workspace and or from eclpise che ui this wrapper must handle the right labels annotions needed secret as env secret as file secret as github token etc to do that we need some new api to handle secrets from eclpise che server describe alternatives you ve considered additional context
0
230,674
7,612,865,291
IssuesEvent
2018-05-01 19:05:06
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
closed
Reordering layers has no effect on map
bug-type: broken use case bug-type: regression priority: high problem: bug
Tested build: http://fgpv.cloudapp.net/demo/develop/prod/samples/index-fgp-en.html Imported 2 layers (`happy.json` and then `convert.json`) and reordering them had no effect. `happy.json` always remained below the other layer even if it was above in the legend stack. Expected behaviour would be to see `happy.json` layer above `convert.json`
1.0
Reordering layers has no effect on map - Tested build: http://fgpv.cloudapp.net/demo/develop/prod/samples/index-fgp-en.html Imported 2 layers (`happy.json` and then `convert.json`) and reordering them had no effect. `happy.json` always remained below the other layer even if it was above in the legend stack. Expected behaviour would be to see `happy.json` layer above `convert.json`
priority
reordering layers has no effect on map tested build imported layers happy json and then convert json and reordering them had no effect happy json always remained below the other layer even if it was above in the legend stack expected behaviour would be to see happy json layer above convert json
1
68,398
13,127,364,092
IssuesEvent
2020-08-06 10:13:02
hpi-swa-teaching/Algernon-Launcher
https://api.github.com/repos/hpi-swa-teaching/Algernon-Launcher
closed
Cleanup Git(-hub)
DevOps non-code
Make the Git (-hub) ready for the next team. This includes: - [x] Remove Merged/stale/dangling Branches - [x] Archive/Complete Milestones - [x] Look through open issues add/remove description/labels/etc if needed
1.0
Cleanup Git(-hub) - Make the Git (-hub) ready for the next team. This includes: - [x] Remove Merged/stale/dangling Branches - [x] Archive/Complete Milestones - [x] Look through open issues add/remove description/labels/etc if needed
non_priority
cleanup git hub make the git hub ready for the next team this includes remove merged stale dangling branches archive complete milestones look through open issues add remove description labels etc if needed
0
422,215
12,268,448,667
IssuesEvent
2020-05-07 12:32:59
AxonFramework/AxonFramework
https://api.github.com/repos/AxonFramework/AxonFramework
closed
Missing aggregateIdentifier() method reported by an exception
Priority 4: Would Status: In Progress Type: Bug
When I was writing my tests the other day I happened to encounter an `EventStoreException` with message "You probably want to use aggregateIdentifier() on your fixture". The solution was to create a separate fixture for each independent unit test, but I could not find the `aggregateIdentifier()` method on any of the fixture classes (even though there is a field on `AggregateTestFixture`). I suggest to fix the exception message not to point to a missing method or make it clearer if I simply was not looking in the right place.
1.0
Missing aggregateIdentifier() method reported by an exception - When I was writing my tests the other day I happened to encounter an `EventStoreException` with message "You probably want to use aggregateIdentifier() on your fixture". The solution was to create a separate fixture for each independent unit test, but I could not find the `aggregateIdentifier()` method on any of the fixture classes (even though there is a field on `AggregateTestFixture`). I suggest to fix the exception message not to point to a missing method or make it clearer if I simply was not looking in the right place.
priority
missing aggregateidentifier method reported by an exception when i was writing my tests the other day i happened to encounter an eventstoreexception with message you probably want to use aggregateidentifier on your fixture the solution was to create a separate fixture for each independent unit test but i could not find the aggregateidentifier method on any of the fixture classes even though there is a field on aggregatetestfixture i suggest to fix the exception message not to point to a missing method or make it clearer if i simply was not looking in the right place
1
414,126
12,099,647,314
IssuesEvent
2020-04-20 12:33:18
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Add contact medium to new case contact success page
:raised_hands: Volunteer Priority: High Status: Available
relates to epic #3 **What type of user is this for? [volunteer/supervisor/admin/all]** **Where does/should this occur** On the page a volunteer views after successfully creating a new `case_contact`. **Description** _Contact medium_ should display underneath _Contact type:_ **Screenshots** <img width="1237" alt="volunteer_bug" src="https://user-images.githubusercontent.com/62810851/79702174-b8a7ed00-8257-11ea-8491-5fa83662ddc7.png"> **Relates to issues...** dependent on #121, relates to #129
1.0
Add contact medium to new case contact success page - relates to epic #3 **What type of user is this for? [volunteer/supervisor/admin/all]** **Where does/should this occur** On the page a volunteer views after successfully creating a new `case_contact`. **Description** _Contact medium_ should display underneath _Contact type:_ **Screenshots** <img width="1237" alt="volunteer_bug" src="https://user-images.githubusercontent.com/62810851/79702174-b8a7ed00-8257-11ea-8491-5fa83662ddc7.png"> **Relates to issues...** dependent on #121, relates to #129
priority
add contact medium to new case contact success page relates to epic what type of user is this for where does should this occur on the page a volunteer views after successfully creating a new case contact description contact medium should display underneath contact type screenshots img width alt volunteer bug src relates to issues dependent on relates to
1
98,284
12,305,950,794
IssuesEvent
2020-05-11 23:58:13
alice-i-cecile/Fonts-of-Power
https://api.github.com/repos/alice-i-cecile/Fonts-of-Power
closed
Rework attacks of opportunity to fix melee stalemate
bug design
10:03 PM] Alice: Alright, here we go: - combat in 5e (and FoP) feels really stale because melee combatants don't try to move - once you're in melee combat, being the first one to blink just gets you smacked in the face, and nothing else - this is less bad in FoP, because a) Disengage is a minor action b) if you can force them to Block, there's no penalty to backing away - however, there's still no benefit to doing so by default. They close the gap and smack you again [10:04 PM] Alice: Proposal: - creatures provoke an attack of opportunity whenever they enter or leave a hostile creature's threatened zone [10:05 PM] Alice: This means that the optimal offensive play is actually to skirmish: 1. hit 2. force a Block 3. attack of opportunity when they close the distance again
1.0
Rework attacks of opportunity to fix melee stalemate - 10:03 PM] Alice: Alright, here we go: - combat in 5e (and FoP) feels really stale because melee combatants don't try to move - once you're in melee combat, being the first one to blink just gets you smacked in the face, and nothing else - this is less bad in FoP, because a) Disengage is a minor action b) if you can force them to Block, there's no penalty to backing away - however, there's still no benefit to doing so by default. They close the gap and smack you again [10:04 PM] Alice: Proposal: - creatures provoke an attack of opportunity whenever they enter or leave a hostile creature's threatened zone [10:05 PM] Alice: This means that the optimal offensive play is actually to skirmish: 1. hit 2. force a Block 3. attack of opportunity when they close the distance again
non_priority
rework attacks of opportunity to fix melee stalemate pm alice alright here we go combat in and fop feels really stale because melee combatants don t try to move once you re in melee combat being the first one to blink just gets you smacked in the face and nothing else this is less bad in fop because a disengage is a minor action b if you can force them to block there s no penalty to backing away however there s still no benefit to doing so by default they close the gap and smack you again alice proposal creatures provoke an attack of opportunity whenever they enter or leave a hostile creature s threatened zone alice this means that the optimal offensive play is actually to skirmish hit force a block attack of opportunity when they close the distance again
0
427,085
29,795,857,404
IssuesEvent
2023-06-16 02:14:19
inventree/InvenTree
https://api.github.com/repos/inventree/InvenTree
closed
units parameters
question report documentation
### Body of the issue is there a way to print the units of the parameters in the labels. behind the value i want the unit to print but don't know how `<p class="text"> {% if parameters.voltage %} {{parameters.voltage}} {% endif %} </p>`
1.0
units parameters - ### Body of the issue is there a way to print the units of the parameters in the labels. behind the value i want the unit to print but don't know how `<p class="text"> {% if parameters.voltage %} {{parameters.voltage}} {% endif %} </p>`
non_priority
units parameters body of the issue is there a way to print the units of the parameters in the labels behind the value i want the unit to print but don t know how if parameters voltage parameters voltage endif
0
104,600
11,415,273,942
IssuesEvent
2020-02-02 09:52:53
engnogueira/webdjango
https://api.github.com/repos/engnogueira/webdjango
closed
3.5.3 - Implementando um Breadcrumb
documentation
Nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website. [Implementando um Breadcrumb](https://www.python.pro.br/modulos/django/topicos/implementando-um-breadcrumb) Link com documentação do Twitter Bootstrap: https://getbootstrap.com/docs/4.4/components/breadcrumb/
1.0
3.5.3 - Implementando um Breadcrumb - Nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website. [Implementando um Breadcrumb](https://www.python.pro.br/modulos/django/topicos/implementando-um-breadcrumb) Link com documentação do Twitter Bootstrap: https://getbootstrap.com/docs/4.4/components/breadcrumb/
non_priority
implementando um breadcrumb nessa aula implementamos a funcionalidade de breadcrumb para melhorar a navegalibilidade das aulas no website link com documentação do twitter bootstrap
0
490,200
14,116,621,813
IssuesEvent
2020-11-08 04:19:50
AY2021S1-CS2103-T16-2/tp
https://api.github.com/repos/AY2021S1-CS2103-T16-2/tp
opened
Check Code Quality for Joven's Features
priority.Medium
Will only close when the documentation for code quality improvements are made.
1.0
Check Code Quality for Joven's Features - Will only close when the documentation for code quality improvements are made.
priority
check code quality for joven s features will only close when the documentation for code quality improvements are made
1
163,294
13,914,727,812
IssuesEvent
2020-10-20 22:47:44
nexusformat/definitions
https://api.github.com/repos/nexusformat/definitions
closed
DOC: ! LaTeX Error: Too deeply nested.
documentation
When building the documentation, this error is reported by the pdflatex build. It means that we have requested more than six levels of a latex ``/begin{quote}`` (or other) block. The issue could be with how we: 1. extract the rst from the NXDL files 1. format specific documentation (indentation) in the NXDL files 1. format the rst file for each NXDL We need better code to flag the first two of these conditions. It is not obvious what code produces this error. It is obvious it is coming from the documentation of more than just any one specific NXDL. Despite the error, a PDF is generated and looks fine. There is a note in the `Makefile` acknowledging this has been known for a long time.
1.0
DOC: ! LaTeX Error: Too deeply nested. - When building the documentation, this error is reported by the pdflatex build. It means that we have requested more than six levels of a latex ``/begin{quote}`` (or other) block. The issue could be with how we: 1. extract the rst from the NXDL files 1. format specific documentation (indentation) in the NXDL files 1. format the rst file for each NXDL We need better code to flag the first two of these conditions. It is not obvious what code produces this error. It is obvious it is coming from the documentation of more than just any one specific NXDL. Despite the error, a PDF is generated and looks fine. There is a note in the `Makefile` acknowledging this has been known for a long time.
non_priority
doc latex error too deeply nested when building the documentation this error is reported by the pdflatex build it means that we have requested more than six levels of a latex begin quote or other block the issue could be with how we extract the rst from the nxdl files format specific documentation indentation in the nxdl files format the rst file for each nxdl we need better code to flag the first two of these conditions it is not obvious what code produces this error it is obvious it is coming from the documentation of more than just any one specific nxdl despite the error a pdf is generated and looks fine there is a note in the makefile acknowledging this has been known for a long time
0
65,983
19,846,070,635
IssuesEvent
2022-01-21 06:31:41
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
Cannot share a video from iOS
T-Defect
### Steps to reproduce Steps to reproduce: 1. Have a video present in chat with someone 2. Click on the video and hold 3. Click more in the bottom left corner 4. Select share and select a user to send it to someone ### Outcome Expected the video to be shared to the other user. Actual result: The process gets stuck in the sending phase ![Screenshot 2022-01-21 at 11 47 54 AM](https://user-images.githubusercontent.com/29797823/150477900-04b1eda5-1c1a-49e3-8ff9-8feae950b518.png) ### Your phone model iPad 7th Generation ### Operating system version iOS 14.7.1 ### Application version Element 1.16.12 ### Homeserver _No response_ ### Will you send logs? No
1.0
Cannot share a video from iOS - ### Steps to reproduce Steps to reproduce: 1. Have a video present in chat with someone 2. Click on the video and hold 3. Click more in the bottom left corner 4. Select share and select a user to send it to someone ### Outcome Expected the video to be shared to the other user. Actual result: The process gets stuck in the sending phase ![Screenshot 2022-01-21 at 11 47 54 AM](https://user-images.githubusercontent.com/29797823/150477900-04b1eda5-1c1a-49e3-8ff9-8feae950b518.png) ### Your phone model iPad 7th Generation ### Operating system version iOS 14.7.1 ### Application version Element 1.16.12 ### Homeserver _No response_ ### Will you send logs? No
non_priority
cannot share a video from ios steps to reproduce steps to reproduce have a video present in chat with someone click on the video and hold click more in the bottom left corner select share and select a user to send it to someone outcome expected the video to be shared to the other user actual result the process gets stuck in the sending phase your phone model ipad generation operating system version ios application version element homeserver no response will you send logs no
0
12,630
4,513,229,528
IssuesEvent
2016-09-04 05:27:52
Jeremy-Barnes/Critters
https://api.github.com/repos/Jeremy-Barnes/Critters
opened
Server: Create Landmark/NPC
Code feature Server
Create endpoints to allow: Random dialog fetching for distinct pages (NPCs) Quest management
1.0
Server: Create Landmark/NPC - Create endpoints to allow: Random dialog fetching for distinct pages (NPCs) Quest management
non_priority
server create landmark npc create endpoints to allow random dialog fetching for distinct pages npcs quest management
0
14,164
3,807,048,174
IssuesEvent
2016-03-25 04:33:06
stevegrunwell/mcavoy
https://api.github.com/repos/stevegrunwell/mcavoy
closed
Introduce a formal change log
documentation
Once the plugin gets past 0.1.0, a formal change log should be kept based on the [Keep a Changelog standard](http://keepachangelog.com/).
1.0
Introduce a formal change log - Once the plugin gets past 0.1.0, a formal change log should be kept based on the [Keep a Changelog standard](http://keepachangelog.com/).
non_priority
introduce a formal change log once the plugin gets past a formal change log should be kept based on the
0
617,407
19,349,655,247
IssuesEvent
2021-12-15 14:29:24
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
att-yahoo.att.net - see bug description
browser-firefox-mobile priority-normal engine-gecko
<!-- @browser: Firefox Mobile 96.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/96773 --> **URL**: https://att-yahoo.att.net/FIM/sps/auth?SPRelayState=https%3A%2F%2Fmail.yahoo.com%2Fd%2F&FedId=uuidd0e66c67-0170-1fb9-80b8-ca3a6bdd2cf8&PartnerId=https%3A%2F%2Flogin.yahoo.com%2Fsaml2&FedName=ATT-Yahoo **Browser / Version**: Firefox Mobile 96.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: has timed out for about 8hrs now. **Steps to Reproduce**: Keeps saying Timed out. I turned off ad blockers Google search fix updated Ff for Android restarted phone. I just tried it with Chrome and it did the best thing. ERR_CONNECTION_TIMED_OUT <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/6d4c6800-5f09-4c07-a171-cd6da09e8aaf.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
att-yahoo.att.net - see bug description - <!-- @browser: Firefox Mobile 96.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/96773 --> **URL**: https://att-yahoo.att.net/FIM/sps/auth?SPRelayState=https%3A%2F%2Fmail.yahoo.com%2Fd%2F&FedId=uuidd0e66c67-0170-1fb9-80b8-ca3a6bdd2cf8&PartnerId=https%3A%2F%2Flogin.yahoo.com%2Fsaml2&FedName=ATT-Yahoo **Browser / Version**: Firefox Mobile 96.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: has timed out for about 8hrs now. **Steps to Reproduce**: Keeps saying Timed out. I turned off ad blockers Google search fix updated Ff for Android restarted phone. I just tried it with Chrome and it did the best thing. ERR_CONNECTION_TIMED_OUT <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/6d4c6800-5f09-4c07-a171-cd6da09e8aaf.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
att yahoo att net see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description has timed out for about now steps to reproduce keeps saying timed out i turned off ad blockers google search fix updated ff for android restarted phone i just tried it with chrome and it did the best thing err connection timed out view the screenshot img alt screenshot src browser configuration none from with ❤️
1
568,720
16,987,209,797
IssuesEvent
2021-06-30 15:37:39
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
NR Extension Date not updated in NRO
ENTITY OPS Priority1
ServiceNow incident: INC0098013 Contact information Staff Name: Debbie Blythe Staff Email: Description NR Extension has been applied in NAMEX but is not updated in NRO. Update Extension Date in NRO to July 31, 2021. Email from Client: Patrick can you create a high priority bug for this please that needs to be worked on today. Kaine Sparks Director, Digital Operations| BC Registries and Online Services Service BC Ministry of Citizens’ Services T: 250-415-3957 | Web: http://www.servicebc.gov.bc.ca “Access to government services made easy” From: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca> Sent: June 25, 2021 10:15 AM To: SBC IT Operations Support, CITZ <SBC_ITOperationsSupport@gov.bc.ca> Cc: Sparks, Kaine CITZ:EX <Kaine.Sparks@gov.bc.ca>; McClung, Linda CITZ:EX <Linda.McClung@gov.bc.ca>; Sienna Blumstengel <sienna@freshworks.io> Subject: FW: Name Requests and Extensions - 5234-7 Importance: High Client resubmitted the following NR’s and the expiry date in Namex has been extended to July 31, 2021. However NRO is still showing Jun 5, 2021. Can this please get fixed, thanks. 1. Cheslatta Marine Services Limited Partnership - NR 5252553 2. Cheslatta Marine Services GP Ltd. – NR 3664932 3. Cheslatta Contracting GP Ltd. - NR 2528980 4. Cheslatta Contracting Limited Partnership – NR 7383745 Debbie From: Melanie Starchuk <Melanie_Starchuk@aboriginal-law.com> Sent: June 22, 2021 1:50 PM To: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>; BCOLACCOUNTS CITZ:EX <BCOLACCOUNTS@gov.bc.ca> Cc: Devrin Froese <Devrin_Froese@aboriginal-law.com> Subject: Name Requests and Extensions - 5234-7 Importance: High [EXTERNAL] This email came from an external source. Only open attachments or links that you are expecting from a known sender. Hello, I submitted the below name requests on March 31st, 2021 and extended the requests on June 4, 2021. The attached four PDF’s include the receipts of the original requests and the extension receipts. Online it says the extensions expired in June, which shouldn’t be the case because we paid for the name reservation extensions. 1. Cheslatta Marine Services Limited Partnership - NR 5252553 2. Cheslatta Marine Services GP Ltd. – NR 3664932 3. Cheslatta Contracting GP Ltd. - NR 2528980 4. Cheslatta Contracting Limited Partnership – NR 7383745 Our BC OnLine Account number is 862583. The new extensions should be until July 31, 2021. I did not receive a new name reservation number. Have there been new name reservation numbers created? Thanks, Melanie Melanie Starchuk Paralegal Donovan & Company Barristers and Solicitors 6th Floor, 73 Water Street Vancouver, B.C. V6B 1A1 Tel: (604) 688-4272 Fax: (604) 688-4282 E-mail: melanie_starchuk@aboriginal-law.com Website: www.aboriginal-law.com ***IMPORTANT NOTICE*** The contents of this e-mail, including attachments, are strictly confidential and are intended only for the use of the individual or entity to which this e-mail is addressed. If you are not the intended recipient, or an employee or agent responsible for delivering the message to the intended recipient, you are hereby notified that any dissemination, distribution of copying this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by telephone (604-688-4272).
1.0
NR Extension Date not updated in NRO - ServiceNow incident: INC0098013 Contact information Staff Name: Debbie Blythe Staff Email: Description NR Extension has been applied in NAMEX but is not updated in NRO. Update Extension Date in NRO to July 31, 2021. Email from Client: Patrick can you create a high priority bug for this please that needs to be worked on today. Kaine Sparks Director, Digital Operations| BC Registries and Online Services Service BC Ministry of Citizens’ Services T: 250-415-3957 | Web: http://www.servicebc.gov.bc.ca “Access to government services made easy” From: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca> Sent: June 25, 2021 10:15 AM To: SBC IT Operations Support, CITZ <SBC_ITOperationsSupport@gov.bc.ca> Cc: Sparks, Kaine CITZ:EX <Kaine.Sparks@gov.bc.ca>; McClung, Linda CITZ:EX <Linda.McClung@gov.bc.ca>; Sienna Blumstengel <sienna@freshworks.io> Subject: FW: Name Requests and Extensions - 5234-7 Importance: High Client resubmitted the following NR’s and the expiry date in Namex has been extended to July 31, 2021. However NRO is still showing Jun 5, 2021. Can this please get fixed, thanks. 1. Cheslatta Marine Services Limited Partnership - NR 5252553 2. Cheslatta Marine Services GP Ltd. – NR 3664932 3. Cheslatta Contracting GP Ltd. - NR 2528980 4. Cheslatta Contracting Limited Partnership – NR 7383745 Debbie From: Melanie Starchuk <Melanie_Starchuk@aboriginal-law.com> Sent: June 22, 2021 1:50 PM To: BC Registry Services CITZ:EX <bcregistries@gov.bc.ca>; BCOLACCOUNTS CITZ:EX <BCOLACCOUNTS@gov.bc.ca> Cc: Devrin Froese <Devrin_Froese@aboriginal-law.com> Subject: Name Requests and Extensions - 5234-7 Importance: High [EXTERNAL] This email came from an external source. Only open attachments or links that you are expecting from a known sender. Hello, I submitted the below name requests on March 31st, 2021 and extended the requests on June 4, 2021. The attached four PDF’s include the receipts of the original requests and the extension receipts. Online it says the extensions expired in June, which shouldn’t be the case because we paid for the name reservation extensions. 1. Cheslatta Marine Services Limited Partnership - NR 5252553 2. Cheslatta Marine Services GP Ltd. – NR 3664932 3. Cheslatta Contracting GP Ltd. - NR 2528980 4. Cheslatta Contracting Limited Partnership – NR 7383745 Our BC OnLine Account number is 862583. The new extensions should be until July 31, 2021. I did not receive a new name reservation number. Have there been new name reservation numbers created? Thanks, Melanie Melanie Starchuk Paralegal Donovan & Company Barristers and Solicitors 6th Floor, 73 Water Street Vancouver, B.C. V6B 1A1 Tel: (604) 688-4272 Fax: (604) 688-4282 E-mail: melanie_starchuk@aboriginal-law.com Website: www.aboriginal-law.com ***IMPORTANT NOTICE*** The contents of this e-mail, including attachments, are strictly confidential and are intended only for the use of the individual or entity to which this e-mail is addressed. If you are not the intended recipient, or an employee or agent responsible for delivering the message to the intended recipient, you are hereby notified that any dissemination, distribution of copying this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by telephone (604-688-4272).
priority
nr extension date not updated in nro servicenow incident contact information staff name debbie blythe staff email description nr extension has been applied in namex but is not updated in nro update extension date in nro to july email from client patrick can you create a high priority bug for this please that needs to be worked on today kaine sparks director digital operations bc registries and online services service bc ministry of citizens’ services t web “access to government services made easy” from bc registry services citz ex sent june am to sbc it operations support citz cc sparks kaine citz ex mcclung linda citz ex sienna blumstengel subject fw name requests and extensions importance high client resubmitted the following nr’s and the expiry date in namex has been extended to july however nro is still showing jun can this please get fixed thanks cheslatta marine services limited partnership nr cheslatta marine services gp ltd – nr cheslatta contracting gp ltd nr cheslatta contracting limited partnership – nr debbie from melanie starchuk sent june pm to bc registry services citz ex bcolaccounts citz ex cc devrin froese subject name requests and extensions importance high this email came from an external source only open attachments or links that you are expecting from a known sender hello i submitted the below name requests on march and extended the requests on june the attached four pdf’s include the receipts of the original requests and the extension receipts online it says the extensions expired in june which shouldn’t be the case because we paid for the name reservation extensions cheslatta marine services limited partnership nr cheslatta marine services gp ltd – nr cheslatta contracting gp ltd nr cheslatta contracting limited partnership – nr our bc online account number is the new extensions should be until july i did not receive a new name reservation number have there been new name reservation numbers created thanks melanie melanie starchuk paralegal donovan company barristers and solicitors floor water street vancouver b c tel fax e mail melanie starchuk aboriginal law com website important notice the contents of this e mail including attachments are strictly confidential and are intended only for the use of the individual or entity to which this e mail is addressed if you are not the intended recipient or an employee or agent responsible for delivering the message to the intended recipient you are hereby notified that any dissemination distribution of copying this communication is strictly prohibited if you have received this communication in error please notify us immediately by telephone
1
820,370
30,770,357,905
IssuesEvent
2023-07-30 20:43:33
CueToDo/Free-Vote-SPA
https://api.github.com/repos/CueToDo/Free-Vote-SPA
closed
Remove HTML encoding from website preview meta data description
bug high priority
https://free.vote/Money/merchant-bankers-and-charity Plus we don't need the url and the actual link
1.0
Remove HTML encoding from website preview meta data description - https://free.vote/Money/merchant-bankers-and-charity Plus we don't need the url and the actual link
priority
remove html encoding from website preview meta data description plus we don t need the url and the actual link
1
37,740
8,357,213,237
IssuesEvent
2018-10-02 20:51:11
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
Add tests for MethodImplOptions.AggressiveOptimization
area-CodeVersioning
- Depends on https://github.com/dotnet/coreclr/pull/20009 - Add test that verifies that the attribute causes the method to be jitted only once at tier 1 - See https://github.com/dotnet/coreclr/pull/20009#discussion_r221808280 - Add perf test from issue (https://github.com/dotnet/coreclr/issues/19751) to performance repo
1.0
Add tests for MethodImplOptions.AggressiveOptimization - - Depends on https://github.com/dotnet/coreclr/pull/20009 - Add test that verifies that the attribute causes the method to be jitted only once at tier 1 - See https://github.com/dotnet/coreclr/pull/20009#discussion_r221808280 - Add perf test from issue (https://github.com/dotnet/coreclr/issues/19751) to performance repo
non_priority
add tests for methodimploptions aggressiveoptimization depends on add test that verifies that the attribute causes the method to be jitted only once at tier see add perf test from issue to performance repo
0
1,478
2,514,730,315
IssuesEvent
2015-01-15 14:03:16
OCHA-DAP/hdx-ckan
https://api.github.com/repos/OCHA-DAP/hdx-ckan
opened
Default Country Page: map
Priority-Medium
Blocked by #2102 and #2103 Annotated (very annotated) design is here: https://docs.google.com/drawings/d/1qOBKZ7IO7zkEMHh2g3ZiAynh5PiAbO-_-SL4rd9uO_M/edit Implement the map section
1.0
Default Country Page: map - Blocked by #2102 and #2103 Annotated (very annotated) design is here: https://docs.google.com/drawings/d/1qOBKZ7IO7zkEMHh2g3ZiAynh5PiAbO-_-SL4rd9uO_M/edit Implement the map section
priority
default country page map blocked by and annotated very annotated design is here implement the map section
1
257,581
8,139,280,981
IssuesEvent
2018-08-20 17:11:27
nprapps/elections18-general
https://api.github.com/repos/nprapps/elections18-general
closed
Check in on admin panel
effort:medium priority:high
We'll need to touch up our logic around "who's won the chamber" and handle a few other related aspects, such as whether it can show two Senate races in a single state.
1.0
Check in on admin panel - We'll need to touch up our logic around "who's won the chamber" and handle a few other related aspects, such as whether it can show two Senate races in a single state.
priority
check in on admin panel we ll need to touch up our logic around who s won the chamber and handle a few other related aspects such as whether it can show two senate races in a single state
1
612,878
19,058,331,235
IssuesEvent
2021-11-26 01:40:40
micronaut-projects/micronaut-kafka
https://api.github.com/repos/micronaut-projects/micronaut-kafka
closed
Micronaut Kafka Listener commits messages at shutdown that should not be commited with offset strategy SYNC
priority: high
With OffsetStrategy set to SYNC or ASYNC the KafkaListner should not commit messages when an Exception occurs. However during shutdown all messages get committed regardless if there was an Exception or not. ### Task List - [X ] Steps to reproduce provided - [x] Stacktrace (if present) provided - [x] Example that reproduces the problem uploaded to Github - [ X] Full description of the issue provided (see below) ### Steps to Reproduce 1. Have an Micronaut application with a KafkaListener in SYNC mode listening to a Kafka topic. 2. Submit an event to this topic. 3. Start process the event but have an Exception thrown during processing (like a SqlException during a database operation). 4. You now have one message lag on your consumer group as you can validate on Kafka. (as documented) 5. Stop the application 6. You now have a zero message lag on your consumer group as you can validate on Kafka. ### Expected Behaviour The KafkaListener should not commit messages with offset strategy SYNC (or ASYNC) at shutdown time that could not be processed due to an Exception. ### Actual Behaviour All events get committed. Only with offset strategy DISABLED this set is skipped.
1.0
Micronaut Kafka Listener commits messages at shutdown that should not be commited with offset strategy SYNC - With OffsetStrategy set to SYNC or ASYNC the KafkaListner should not commit messages when an Exception occurs. However during shutdown all messages get committed regardless if there was an Exception or not. ### Task List - [X ] Steps to reproduce provided - [x] Stacktrace (if present) provided - [x] Example that reproduces the problem uploaded to Github - [ X] Full description of the issue provided (see below) ### Steps to Reproduce 1. Have an Micronaut application with a KafkaListener in SYNC mode listening to a Kafka topic. 2. Submit an event to this topic. 3. Start process the event but have an Exception thrown during processing (like a SqlException during a database operation). 4. You now have one message lag on your consumer group as you can validate on Kafka. (as documented) 5. Stop the application 6. You now have a zero message lag on your consumer group as you can validate on Kafka. ### Expected Behaviour The KafkaListener should not commit messages with offset strategy SYNC (or ASYNC) at shutdown time that could not be processed due to an Exception. ### Actual Behaviour All events get committed. Only with offset strategy DISABLED this set is skipped.
priority
micronaut kafka listener commits messages at shutdown that should not be commited with offset strategy sync with offsetstrategy set to sync or async the kafkalistner should not commit messages when an exception occurs however during shutdown all messages get committed regardless if there was an exception or not task list steps to reproduce provided stacktrace if present provided example that reproduces the problem uploaded to github full description of the issue provided see below steps to reproduce have an micronaut application with a kafkalistener in sync mode listening to a kafka topic submit an event to this topic start process the event but have an exception thrown during processing like a sqlexception during a database operation you now have one message lag on your consumer group as you can validate on kafka as documented stop the application you now have a zero message lag on your consumer group as you can validate on kafka expected behaviour the kafkalistener should not commit messages with offset strategy sync or async at shutdown time that could not be processed due to an exception actual behaviour all events get committed only with offset strategy disabled this set is skipped
1
58,155
3,087,857,656
IssuesEvent
2015-08-25 14:05:15
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
Плохо работает разбор вложенных BB-кодов
bug imported Priority-Medium
_From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 09, 2013 00:47:35_ Точно не работает 1. Все тэги стиля текста вложенные в тэг цвета 2. Не работает цвет вложенный в другой цвет 3. отображение стилей зависит от порядка применения тэгов [i][b]болд италик[/b][/i] - отображает только курсив [b][i]италик болд[/i][/b] - показывает жирный курсив Кто еще что найдет, дополняйте. Но лечить надо не указанные ошибки по-отдельности а весь алгоритм разбора который сводится к разбиению строки на участки и назначении каждому участку правильных атрибутов (я насчитал 5 штук) с учетом вложения тэгов (например с применением какого-то стекового принципа) **Attachment:** [Fly_r15987_BBcodes.png Fly_r15988_BBcodes.png](http://code.google.com/p/flylinkdc/issues/detail?id=1391) _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1391_
1.0
Плохо работает разбор вложенных BB-кодов - _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 09, 2013 00:47:35_ Точно не работает 1. Все тэги стиля текста вложенные в тэг цвета 2. Не работает цвет вложенный в другой цвет 3. отображение стилей зависит от порядка применения тэгов [i][b]болд италик[/b][/i] - отображает только курсив [b][i]италик болд[/i][/b] - показывает жирный курсив Кто еще что найдет, дополняйте. Но лечить надо не указанные ошибки по-отдельности а весь алгоритм разбора который сводится к разбиению строки на участки и назначении каждому участку правильных атрибутов (я насчитал 5 штук) с учетом вложения тэгов (например с применением какого-то стекового принципа) **Attachment:** [Fly_r15987_BBcodes.png Fly_r15988_BBcodes.png](http://code.google.com/p/flylinkdc/issues/detail?id=1391) _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1391_
priority
плохо работает разбор вложенных bb кодов from on november точно не работает все тэги стиля текста вложенные в тэг цвета не работает цвет вложенный в другой цвет отображение стилей зависит от порядка применения тэгов болд италик отображает только курсив италик болд показывает жирный курсив кто еще что найдет дополняйте но лечить надо не указанные ошибки по отдельности а весь алгоритм разбора который сводится к разбиению строки на участки и назначении каждому участку правильных атрибутов я насчитал штук с учетом вложения тэгов например с применением какого то стекового принципа attachment original issue
1
27,666
30,056,634,161
IssuesEvent
2023-06-28 07:20:44
Open-EO/openeo-python-client
https://api.github.com/repos/Open-EO/openeo-python-client
closed
RESTJob UX
clean up usability v1.0.0
make `RESTJob` prettier to work with, e.g.: - The `REST` prefix is bit unusual and redundant: most other user-facing openeo classes don't have this prefix even though they also are about REST abstractions. Just `Job` or `BatchJob` will look prettier for the general user - rename `start_job()` (and alike) to just `start()`
True
RESTJob UX - make `RESTJob` prettier to work with, e.g.: - The `REST` prefix is bit unusual and redundant: most other user-facing openeo classes don't have this prefix even though they also are about REST abstractions. Just `Job` or `BatchJob` will look prettier for the general user - rename `start_job()` (and alike) to just `start()`
non_priority
restjob ux make restjob prettier to work with e g the rest prefix is bit unusual and redundant most other user facing openeo classes don t have this prefix even though they also are about rest abstractions just job or batchjob will look prettier for the general user rename start job and alike to just start
0
312,338
26,858,287,990
IssuesEvent
2023-02-03 16:14:09
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix tensor.test_torch_special_long
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details>
1.0
Fix tensor.test_torch_special_long - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3977337296/jobs/6818403602" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_tensor.py::test_torch_special_long[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-01-22T00:58:45.7934743Z E ValueError: only one element tensors can be converted to Python scalars 2023-01-22T00:58:45.7935075Z E Falsifying example: test_torch_special_long( 2023-01-22T00:58:45.7935485Z E dtype_and_x=(['uint8'], [array([0, 0], dtype=uint8)]), 2023-01-22T00:58:45.7935777Z E init_num_positional_args=0, 2023-01-22T00:58:45.7936033Z E method_num_positional_args=0, 2023-01-22T00:58:45.7936278Z E as_variable=[False], 2023-01-22T00:58:45.7936511Z E native_array=[False], 2023-01-22T00:58:45.7937360Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.torch' from '/ivy/ivy/functional/frontends/torch/__init__.py'>, framework_init_module=<module 'torch' from '/usr/local/lib/python3.8/dist-packages/torch/__init__.py'>, init_name='tensor', method_name='__long__'), 2023-01-22T00:58:45.7938047Z E frontend='torch', 2023-01-22T00:58:45.7938250Z E ) 2023-01-22T00:58:45.7938414Z E 2023-01-22T00:58:45.7938914Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BkYGRAAAAAKQAD') as a decorator on your test case </details>
non_priority
fix tensor test torch special long tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test torch test tensor py test torch special long e valueerror only one element tensors can be converted to python scalars e falsifying example test torch special long e dtype and x dtype e init num positional args e method num positional args e as variable e native array e frontend method data frontendmethoddata ivy init module framework init module init name tensor method name long e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
0
279,904
8,675,466,654
IssuesEvent
2018-11-30 10:57:58
supergiant/control
https://api.github.com/repos/supergiant/control
opened
Current UI do not allow to leave SubnetID empty to let SG create Subnet
High Priority Needs Standup Review aws
When provision cluster on AWS there is no way to create security group for cluster. New security group gets created only when SubnetID is empty, but it is prohibited on the UI to have this field empty. This forces user to use either default or existing subnet which is bad for deletion of resources - dependencies between those resources are uncontrolled.
1.0
Current UI do not allow to leave SubnetID empty to let SG create Subnet - When provision cluster on AWS there is no way to create security group for cluster. New security group gets created only when SubnetID is empty, but it is prohibited on the UI to have this field empty. This forces user to use either default or existing subnet which is bad for deletion of resources - dependencies between those resources are uncontrolled.
priority
current ui do not allow to leave subnetid empty to let sg create subnet when provision cluster on aws there is no way to create security group for cluster new security group gets created only when subnetid is empty but it is prohibited on the ui to have this field empty this forces user to use either default or existing subnet which is bad for deletion of resources dependencies between those resources are uncontrolled
1
133,273
18,289,857,560
IssuesEvent
2021-10-05 14:13:20
finos/symphony-wdk
https://api.github.com/repos/finos/symphony-wdk
opened
CVE-2021-35516 (High) detected in commons-compress-1.20.jar
security vulnerability
## CVE-2021-35516 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4, Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p> <p>Path to dependency file: symphony-wdk/workflow-bot-app/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p> <p> Dependency Hierarchy: - camunda-bpm-spring-boot-starter-7.15.0.jar (Root Library) - spring-boot-loader-tools-2.5.4.jar - :x: **commons-compress-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/finos/symphony-wdk/commit/004d6ff32b56fa8739f47862c8544f6270c29183">004d6ff32b56fa8739f47862c8544f6270c29183</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' sevenz package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516>CVE-2021-35516</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.commons","packageName":"commons-compress","packageVersion":"1.20","packageFilePaths":["/workflow-bot-app/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.camunda.bpm.springboot:camunda-bpm-spring-boot-starter:7.15.0;org.springframework.boot:spring-boot-loader-tools:2.5.4;org.apache.commons:commons-compress:1.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-compress:1.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-35516","vulnerabilityDetails":"When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress\u0027 sevenz package.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-35516 (High) detected in commons-compress-1.20.jar - ## CVE-2021-35516 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4, Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p> <p>Path to dependency file: symphony-wdk/workflow-bot-app/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p> <p> Dependency Hierarchy: - camunda-bpm-spring-boot-starter-7.15.0.jar (Root Library) - spring-boot-loader-tools-2.5.4.jar - :x: **commons-compress-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/finos/symphony-wdk/commit/004d6ff32b56fa8739f47862c8544f6270c29183">004d6ff32b56fa8739f47862c8544f6270c29183</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' sevenz package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516>CVE-2021-35516</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.commons","packageName":"commons-compress","packageVersion":"1.20","packageFilePaths":["/workflow-bot-app/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.camunda.bpm.springboot:camunda-bpm-spring-boot-starter:7.15.0;org.springframework.boot:spring-boot-loader-tools:2.5.4;org.apache.commons:commons-compress:1.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.commons:commons-compress:1.21"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-35516","vulnerabilityDetails":"When reading a specially crafted 7Z archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress\u0027 sevenz package.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35516","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in commons compress jar cve high severity vulnerability vulnerable library commons compress jar apache commons compress software defines an api for working with compression and archive formats these include gzip lzma xz snappy traditional unix compress deflate brotli zstandard and ar cpio jar tar zip dump arj library home page a href path to dependency file symphony wdk workflow bot app build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons compress commons compress jar dependency hierarchy camunda bpm spring boot starter jar root library spring boot loader tools jar x commons compress jar vulnerable library found in head commit a href found in base branch master vulnerability details when reading a specially crafted archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress sevenz package publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons compress isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org camunda bpm springboot camunda bpm spring boot starter org springframework boot spring boot loader tools org apache commons commons compress isminimumfixversionavailable true minimumfixversion org apache commons commons compress basebranches vulnerabilityidentifier cve vulnerabilitydetails when reading a specially crafted archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress sevenz package vulnerabilityurl
0
286,109
24,719,918,872
IssuesEvent
2022-10-20 09:52:12
callstack/react-native-testing-library
https://api.github.com/repos/callstack/react-native-testing-library
closed
Supports accessibilityStates in ByRole queries
compat: testing-library a11y
## Describe the Feature Like in `@testing-library/dom` (see [docs](https://testing-library.com/docs/queries/byrole#api)), we should support accessibility states when querying by role. The accessibilityStates existing are - `disabled`: Indicates whether the element is disabled or not - `selected`: Indicates whether a selectable element is currently selected or not - `checked`: Indicates the state of a checkable element. - `busy`: Indicates whether an element is currently busy or not - `expanded`: Indicates whether an expandable element is currently expanded or collapsed ## Possible Implementations We could like in DTL check which states correspond to which role and throw. But I'd treat that as a bonus and not a necessity. ## Related Issues #1133
1.0
Supports accessibilityStates in ByRole queries - ## Describe the Feature Like in `@testing-library/dom` (see [docs](https://testing-library.com/docs/queries/byrole#api)), we should support accessibility states when querying by role. The accessibilityStates existing are - `disabled`: Indicates whether the element is disabled or not - `selected`: Indicates whether a selectable element is currently selected or not - `checked`: Indicates the state of a checkable element. - `busy`: Indicates whether an element is currently busy or not - `expanded`: Indicates whether an expandable element is currently expanded or collapsed ## Possible Implementations We could like in DTL check which states correspond to which role and throw. But I'd treat that as a bonus and not a necessity. ## Related Issues #1133
non_priority
supports accessibilitystates in byrole queries describe the feature like in testing library dom see we should support accessibility states when querying by role the accessibilitystates existing are disabled indicates whether the element is disabled or not selected indicates whether a selectable element is currently selected or not checked indicates the state of a checkable element busy indicates whether an element is currently busy or not expanded indicates whether an expandable element is currently expanded or collapsed possible implementations we could like in dtl check which states correspond to which role and throw but i d treat that as a bonus and not a necessity related issues
0
111,332
11,729,693,114
IssuesEvent
2020-03-10 19:52:44
kubernetes-sigs/kubebuilder
https://api.github.com/repos/kubernetes-sigs/kubebuilder
closed
Kubebuilder book: Why create a custom metadata field "controller" in job instead of just ownerReferences
kind/documentation triage/support
I was going through the tutorial in the book for Cronjobs. On [this page](https://book.kubebuilder.io/cronjob-tutorial/controller-implementation.html#2-list-all-active-jobs-and-update-the-status) the concept of a `jobOwnerKey` field is introduced: ```golang var ( jobOwnerKey = ".metadata.controller" ``` I don't completely understand why we cannot just rely on the build-in ownerReferences type? Is this because it is not indexable by the cache FieldIndexer and we use the jobOwnerKey solely as an indexable id for it? **Proposals if this is the case:** - In the book documentation, explain the function of the jobOwnerKey (If you explain it here, I can also PR it) - If lists of objects like ownerReferences are not indexable, perhaps start a proposal to make them so? P.S: I didn't know we could add custom metadata to resources? I assume the jobOwnerKey is only used by the local cache then.
1.0
Kubebuilder book: Why create a custom metadata field "controller" in job instead of just ownerReferences - I was going through the tutorial in the book for Cronjobs. On [this page](https://book.kubebuilder.io/cronjob-tutorial/controller-implementation.html#2-list-all-active-jobs-and-update-the-status) the concept of a `jobOwnerKey` field is introduced: ```golang var ( jobOwnerKey = ".metadata.controller" ``` I don't completely understand why we cannot just rely on the build-in ownerReferences type? Is this because it is not indexable by the cache FieldIndexer and we use the jobOwnerKey solely as an indexable id for it? **Proposals if this is the case:** - In the book documentation, explain the function of the jobOwnerKey (If you explain it here, I can also PR it) - If lists of objects like ownerReferences are not indexable, perhaps start a proposal to make them so? P.S: I didn't know we could add custom metadata to resources? I assume the jobOwnerKey is only used by the local cache then.
non_priority
kubebuilder book why create a custom metadata field controller in job instead of just ownerreferences i was going through the tutorial in the book for cronjobs on the concept of a jobownerkey field is introduced golang var jobownerkey metadata controller i don t completely understand why we cannot just rely on the build in ownerreferences type is this because it is not indexable by the cache fieldindexer and we use the jobownerkey solely as an indexable id for it proposals if this is the case in the book documentation explain the function of the jobownerkey if you explain it here i can also pr it if lists of objects like ownerreferences are not indexable perhaps start a proposal to make them so p s i didn t know we could add custom metadata to resources i assume the jobownerkey is only used by the local cache then
0
175,905
21,365,609,718
IssuesEvent
2022-04-20 01:01:26
improbable/k8s-test-infra
https://api.github.com/repos/improbable/k8s-test-infra
closed
moment-2.24.0.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/moment/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-24785](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | moment-2.24.0.tgz | Direct | 2.29.2 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-24785</summary> ### Vulnerable Library - <b>moment-2.24.0.tgz</b></p> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/moment/package.json</p> <p> Dependency Hierarchy: - :x: **moment-2.24.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>improbable</b></p> </p> <p></p> ### Vulnerability Details <p> Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js. <p>Publish Date: 2022-04-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785>CVE-2022-24785</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p> <p>Release Date: 2022-04-04</p> <p>Fix Resolution: 2.29.2</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"moment","packageVersion":"2.24.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"moment:2.24.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.29.2","isBinary":false}],"baseBranches":["improbable"],"vulnerabilityIdentifier":"CVE-2022-24785","vulnerabilityDetails":"Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}]</REMEDIATE> -->
True
moment-2.24.0.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/moment/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-24785](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | moment-2.24.0.tgz | Direct | 2.29.2 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-24785</summary> ### Vulnerable Library - <b>moment-2.24.0.tgz</b></p> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/moment/package.json</p> <p> Dependency Hierarchy: - :x: **moment-2.24.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>improbable</b></p> </p> <p></p> ### Vulnerability Details <p> Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js. <p>Publish Date: 2022-04-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785>CVE-2022-24785</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p> <p>Release Date: 2022-04-04</p> <p>Fix Resolution: 2.29.2</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"moment","packageVersion":"2.24.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"moment:2.24.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.29.2","isBinary":false}],"baseBranches":["improbable"],"vulnerabilityIdentifier":"CVE-2022-24785","vulnerabilityDetails":"Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24785","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}]</REMEDIATE> -->
non_priority
moment tgz vulnerabilities highest severity is autoclosed vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file package json path to vulnerable library node modules moment package json vulnerabilities cve severity cvss dependency type fixed in remediation available high moment tgz direct details cve vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file package json path to vulnerable library node modules moment package json dependency hierarchy x moment tgz vulnerable library found in base branch improbable vulnerability details moment js is a javascript date library for parsing validating manipulating and formatting dates a path traversal vulnerability impacts npm server users of moment js between versions and especially if a user provided locale string is directly used to switch moment locale this problem is patched in and the patch can be applied to all affected versions as a workaround sanitize the user provided locale name before passing it to moment js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue istransitivedependency false dependencytree moment isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails moment js is a javascript date library for parsing validating manipulating and formatting dates a path traversal vulnerability impacts npm server users of moment js between versions and especially if a user provided locale string is directly used to switch moment locale this problem is patched in and the patch can be applied to all affected versions as a workaround sanitize the user provided locale name before passing it to moment js vulnerabilityurl
0
611,953
18,985,572,246
IssuesEvent
2021-11-21 17:01:28
dehy/foodcoop-mobile-app
https://api.github.com/repos/dehy/foodcoop-mobile-app
closed
Migrer la BDD legacy vers TypeORM
Priority: Medium Status: In Progress Type: Refactoring
La bdd sqlite inventaire (legacy, fait maison) est différente de la bdd sqlite réception de marchandises (TypeORM, managée). Migrer le schéma et les données existantes vers TypeORM.
1.0
Migrer la BDD legacy vers TypeORM - La bdd sqlite inventaire (legacy, fait maison) est différente de la bdd sqlite réception de marchandises (TypeORM, managée). Migrer le schéma et les données existantes vers TypeORM.
priority
migrer la bdd legacy vers typeorm la bdd sqlite inventaire legacy fait maison est différente de la bdd sqlite réception de marchandises typeorm managée migrer le schéma et les données existantes vers typeorm
1
42,646
5,513,890,540
IssuesEvent
2017-03-17 13:51:47
Esri/military-tools-geoprocessing-toolbox
https://api.github.com/repos/Esri/military-tools-geoprocessing-toolbox
closed
Using the same Input Area gives different results between Find Local Peaks and Highest Points
B - As Designed effort-medium F - Visibility priority - normal V - 10.3.1
## Expected Behavior Unless the tools need further documentation of the differences between the two, it is expected that when you run Find Local Peaks, the maximum peak value would be the same value returned by Highest Points if you are using the same input feature class. ## Current Behavior Currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is. ## Possible Solution This seems to be because there is a difference in the logic of how the tools clip out the elevation layer. Local Peaks seems to only look for points that are entirely within the input polygon, where as it seems that Highest Points also takes into account points are on the edge of the input polygon. ## Steps to Reproduce (for bugs) 1. Run the Local Peaks with the AreaofInterest feature class as the input area 2. Run Highest Points with the AreaofInterest feature class as the input area 3. Note the maximum value of the output of the Local Peaks (1080) 4. Note the maximum value of the output of Highest Points (1123, also observe that it is on the extreme edge of the polygon) 5. Repeat steps using AreaofInterest2 AreaofInterest2 is attached as well as screen captures illustrating problem [AddVizTestData.zip](https://github.com/Esri/military-tools-geoprocessing-toolbox/files/595797/AddVizTestData.zip) ![highestpointsao 1](https://cloud.githubusercontent.com/assets/5322669/20363929/03934ea4-abf6-11e6-8f21-c2f1df2ed9aa.PNG) ![localpeaksaoi1](https://cloud.githubusercontent.com/assets/5322669/20363928/0392d0dc-abf6-11e6-8fac-f2017052602d.PNG) ![highestpointsao 2](https://cloud.githubusercontent.com/assets/5322669/20363938/0aa8b166-abf6-11e6-985c-5e7314c56f49.PNG) ![localpeaksaoi2](https://cloud.githubusercontent.com/assets/5322669/20363937/0aa88646-abf6-11e6-87cf-5982ea8a8c9e.PNG)
1.0
Using the same Input Area gives different results between Find Local Peaks and Highest Points - ## Expected Behavior Unless the tools need further documentation of the differences between the two, it is expected that when you run Find Local Peaks, the maximum peak value would be the same value returned by Highest Points if you are using the same input feature class. ## Current Behavior Currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is. ## Possible Solution This seems to be because there is a difference in the logic of how the tools clip out the elevation layer. Local Peaks seems to only look for points that are entirely within the input polygon, where as it seems that Highest Points also takes into account points are on the edge of the input polygon. ## Steps to Reproduce (for bugs) 1. Run the Local Peaks with the AreaofInterest feature class as the input area 2. Run Highest Points with the AreaofInterest feature class as the input area 3. Note the maximum value of the output of the Local Peaks (1080) 4. Note the maximum value of the output of Highest Points (1123, also observe that it is on the extreme edge of the polygon) 5. Repeat steps using AreaofInterest2 AreaofInterest2 is attached as well as screen captures illustrating problem [AddVizTestData.zip](https://github.com/Esri/military-tools-geoprocessing-toolbox/files/595797/AddVizTestData.zip) ![highestpointsao 1](https://cloud.githubusercontent.com/assets/5322669/20363929/03934ea4-abf6-11e6-8f21-c2f1df2ed9aa.PNG) ![localpeaksaoi1](https://cloud.githubusercontent.com/assets/5322669/20363928/0392d0dc-abf6-11e6-8fac-f2017052602d.PNG) ![highestpointsao 2](https://cloud.githubusercontent.com/assets/5322669/20363938/0aa8b166-abf6-11e6-985c-5e7314c56f49.PNG) ![localpeaksaoi2](https://cloud.githubusercontent.com/assets/5322669/20363937/0aa88646-abf6-11e6-87cf-5982ea8a8c9e.PNG)
non_priority
using the same input area gives different results between find local peaks and highest points expected behavior unless the tools need further documentation of the differences between the two it is expected that when you run find local peaks the maximum peak value would be the same value returned by highest points if you are using the same input feature class current behavior currently there have been instances found where running both tools give different results for what the absolute highest peak in the area is possible solution this seems to be because there is a difference in the logic of how the tools clip out the elevation layer local peaks seems to only look for points that are entirely within the input polygon where as it seems that highest points also takes into account points are on the edge of the input polygon steps to reproduce for bugs run the local peaks with the areaofinterest feature class as the input area run highest points with the areaofinterest feature class as the input area note the maximum value of the output of the local peaks note the maximum value of the output of highest points also observe that it is on the extreme edge of the polygon repeat steps using is attached as well as screen captures illustrating problem
0
16,772
12,138,443,320
IssuesEvent
2020-04-23 17:15:03
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
opened
aspnetcore-tooling migration
area-infrastructure
- [ ] execute phase 1 of https://github.com/dotnet/aspnetcore/blob/master/docs/tooling-consolidation.md - [ ] plan for phase 2 Phase 1 is ready. Phase 2 needs further planning.
1.0
aspnetcore-tooling migration - - [ ] execute phase 1 of https://github.com/dotnet/aspnetcore/blob/master/docs/tooling-consolidation.md - [ ] plan for phase 2 Phase 1 is ready. Phase 2 needs further planning.
non_priority
aspnetcore tooling migration execute phase of plan for phase phase is ready phase needs further planning
0
304,830
9,336,206,263
IssuesEvent
2019-03-28 20:36:32
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
opened
Events and deadlines do not populate on homepage
Bug High priority
Events and deadlines no longer populate on our homepage. See below screenshot: ![image (2)](https://user-images.githubusercontent.com/12799132/55191076-9e483080-5177-11e9-8b38-1ca46fe2ede1.png)
1.0
Events and deadlines do not populate on homepage - Events and deadlines no longer populate on our homepage. See below screenshot: ![image (2)](https://user-images.githubusercontent.com/12799132/55191076-9e483080-5177-11e9-8b38-1ca46fe2ede1.png)
priority
events and deadlines do not populate on homepage events and deadlines no longer populate on our homepage see below screenshot
1
589,357
17,695,160,368
IssuesEvent
2021-08-24 14:33:59
eclipse/dirigible
https://api.github.com/repos/eclipse/dirigible
opened
[IDE] Editors - CSVIM Editor doesn't support # quote character
bug enhancement component-ide usability priority-high efforts-low
**Describe the bug** Unable to select # as a quote character in CSVIM Editor. **To Reproduce** Steps to reproduce the behavior: 1. Open csvim file 2. Select the quote character dropdown 3. See issue **Expected behavior** CSVIM Editor should support # as a quote character. **Desktop:** - OS: macOS 11.5 - Browser: Firefox 91 - Version: Dirigible 5.12.13
1.0
[IDE] Editors - CSVIM Editor doesn't support # quote character - **Describe the bug** Unable to select # as a quote character in CSVIM Editor. **To Reproduce** Steps to reproduce the behavior: 1. Open csvim file 2. Select the quote character dropdown 3. See issue **Expected behavior** CSVIM Editor should support # as a quote character. **Desktop:** - OS: macOS 11.5 - Browser: Firefox 91 - Version: Dirigible 5.12.13
priority
editors csvim editor doesn t support quote character describe the bug unable to select as a quote character in csvim editor to reproduce steps to reproduce the behavior open csvim file select the quote character dropdown see issue expected behavior csvim editor should support as a quote character desktop os macos browser firefox version dirigible
1
150,367
5,765,562,281
IssuesEvent
2017-04-27 03:01:08
sussol/mobile
https://api.github.com/repos/sussol/mobile
closed
Item usage not aggregated across batches correctly
Bug Priority: High ~in review
Build Number: 30 Description: It looks like item.dailyUsage() is summing the dailyUsage for each batch, which would over-estimate usage when there are multiple batches. If daily usage for batch 1 is 15, and daily usage for batch 2 is also 15, it looks like item.dailyUsage() would return 30. The spreadsheet example linked in issue #405 also includes something about this. It seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used, then do the division in Item. Comments: Happy to take a crack at this ticket or #405, but I don't have a server I could test on at the moment..
1.0
Item usage not aggregated across batches correctly - Build Number: 30 Description: It looks like item.dailyUsage() is summing the dailyUsage for each batch, which would over-estimate usage when there are multiple batches. If daily usage for batch 1 is 15, and daily usage for batch 2 is also 15, it looks like item.dailyUsage() would return 30. The spreadsheet example linked in issue #405 also includes something about this. It seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used, then do the division in Item. Comments: Happy to take a crack at this ticket or #405, but I don't have a server I could test on at the moment..
priority
item usage not aggregated across batches correctly build number description it looks like item dailyusage is summing the dailyusage for each batch which would over estimate usage when there are multiple batches if daily usage for batch is and daily usage for batch is also it looks like item dailyusage would return the spreadsheet example linked in issue also includes something about this it seems like the correct approach to get the total daily usage for an item across batches would be to sum up the total usage for all batches and the number of days each batch was used then do the division in item comments happy to take a crack at this ticket or but i don t have a server i could test on at the moment
1