Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
569,371 | 17,012,873,184 | IssuesEvent | 2021-07-02 07:57:20 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Documents - Preview taking time to load the activity page and other pages where preview is coming. | component: document feature: enhancement priority: medium | **Describe the bug**
Due to the load of the WordPress in `preview.php` file it's slowing the actvity and other pages where document is loading.
**To Reproduce**
Steps to reproduce the behavior:
1. Add to many documents which have a preview image like add too many PDF
2. Go to the activity page and it will slow the activity load.
**Expected behavior**
It should load the instantantly as others like media.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| 1.0 | Documents - Preview taking time to load the activity page and other pages where preview is coming. - **Describe the bug**
Due to the load of the WordPress in `preview.php` file it's slowing the actvity and other pages where document is loading.
**To Reproduce**
Steps to reproduce the behavior:
1. Add to many documents which have a preview image like add too many PDF
2. Go to the activity page and it will slow the activity load.
**Expected behavior**
It should load the instantantly as others like media.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| priority | documents preview taking time to load the activity page and other pages where preview is coming describe the bug due to the load of the wordpress in preview php file it s slowing the actvity and other pages where document is loading to reproduce steps to reproduce the behavior add to many documents which have a preview image like add too many pdf go to the activity page and it will slow the activity load expected behavior it should load the instantantly as others like media screenshots if applicable add screenshots to help explain your problem support ticket links if applicable add helpscout link or ticket number where the issue was originally reported | 1 |
248,138 | 7,927,798,801 | IssuesEvent | 2018-07-06 09:18:42 | minio/minio-java | https://api.github.com/repos/minio/minio-java | closed | Guava version used in this code base has known vulnerabilities. | community priority: medium triage | Our client (Ernst & Young) is using MINIO on one of it's projects. Snyk, our code scanning tool reports the following A9 (known vulnerability) in your code:
Deserialization of Untrusted Data
Vulnerable module: com.google.guava:guava
Introduced through: io.minio:minio@4.0.1
Detailed paths
Introduced through: com.ey:crane_app@0.0.1 › io.minio:minio@4.0.1 › com.google.guava:guava@18.0
Overview
com.google.guava:guava is a set of core libraries that includes new collection types (such as multimap and multiset,immutable collections, a graph library, functional types, an in-memory cache and more.
Affected versions of this package are vulnerable to Deserialization of Untrusted Data.
During deserialization, two Guava classes accept a caller-specified size parameter and eagerly allocate an array of that size:
AtomicDoubleArray (when serialized with Java serialization)
CompoundOrdering (when serialized with GWT serialization)
An attacker may be able to send a specially crafted request which with then cause the server to allocate all it's memory, without validation whether the data size is reasonable.
More information can be found here: https://snyk.io/vuln/SNYK-JAVA-COMGOOGLEGUAVA-32236
The version of Guava that you are using is VERY old and should be upgraded to fix this issue and others that I see have been reported. | 1.0 | Guava version used in this code base has known vulnerabilities. - Our client (Ernst & Young) is using MINIO on one of it's projects. Snyk, our code scanning tool reports the following A9 (known vulnerability) in your code:
Deserialization of Untrusted Data
Vulnerable module: com.google.guava:guava
Introduced through: io.minio:minio@4.0.1
Detailed paths
Introduced through: com.ey:crane_app@0.0.1 › io.minio:minio@4.0.1 › com.google.guava:guava@18.0
Overview
com.google.guava:guava is a set of core libraries that includes new collection types (such as multimap and multiset,immutable collections, a graph library, functional types, an in-memory cache and more.
Affected versions of this package are vulnerable to Deserialization of Untrusted Data.
During deserialization, two Guava classes accept a caller-specified size parameter and eagerly allocate an array of that size:
AtomicDoubleArray (when serialized with Java serialization)
CompoundOrdering (when serialized with GWT serialization)
An attacker may be able to send a specially crafted request which with then cause the server to allocate all it's memory, without validation whether the data size is reasonable.
More information can be found here: https://snyk.io/vuln/SNYK-JAVA-COMGOOGLEGUAVA-32236
The version of Guava that you are using is VERY old and should be upgraded to fix this issue and others that I see have been reported. | priority | guava version used in this code base has known vulnerabilities our client ernst young is using minio on one of it s projects snyk our code scanning tool reports the following known vulnerability in your code deserialization of untrusted data vulnerable module com google guava guava introduced through io minio minio detailed paths introduced through com ey crane app › io minio minio › com google guava guava overview com google guava guava is a set of core libraries that includes new collection types such as multimap and multiset immutable collections a graph library functional types an in memory cache and more affected versions of this package are vulnerable to deserialization of untrusted data during deserialization two guava classes accept a caller specified size parameter and eagerly allocate an array of that size atomicdoublearray when serialized with java serialization compoundordering when serialized with gwt serialization an attacker may be able to send a specially crafted request which with then cause the server to allocate all it s memory without validation whether the data size is reasonable more information can be found here the version of guava that you are using is very old and should be upgraded to fix this issue and others that i see have been reported | 1 |
94,146 | 3,921,903,609 | IssuesEvent | 2016-04-22 02:12:32 | TranslationWMcs435/TranslationWMcs435 | https://api.github.com/repos/TranslationWMcs435/TranslationWMcs435 | closed | UiAutomator launch app before testing | Medium Priority | Add functionality so that the app is open before the tests are performed. | 1.0 | UiAutomator launch app before testing - Add functionality so that the app is open before the tests are performed. | priority | uiautomator launch app before testing add functionality so that the app is open before the tests are performed | 1 |
353,273 | 10,550,971,683 | IssuesEvent | 2019-10-03 12:22:31 | nmstate/kubernetes-nmstate | https://api.github.com/repos/nmstate/kubernetes-nmstate | closed | Stabilize vlan script | priority:medium type:enhancement | The vlan script is a weak point in the configuration. However, there are few things we may do to make it more stable and predictable.
1. Wait for the bridge to become available before we start configuring.
2. Use nmstate rollback mechanism to revert configuration in case the VLAN script fails. | 1.0 | Stabilize vlan script - The vlan script is a weak point in the configuration. However, there are few things we may do to make it more stable and predictable.
1. Wait for the bridge to become available before we start configuring.
2. Use nmstate rollback mechanism to revert configuration in case the VLAN script fails. | priority | stabilize vlan script the vlan script is a weak point in the configuration however there are few things we may do to make it more stable and predictable wait for the bridge to become available before we start configuring use nmstate rollback mechanism to revert configuration in case the vlan script fails | 1 |
196,231 | 6,926,087,595 | IssuesEvent | 2017-11-30 17:55:59 | STB1019/JAM2018 | https://api.github.com/repos/STB1019/JAM2018 | opened | Add Animations To The Enemy | medium priority | We need to add some working animations to the enemies, that react to their AI. | 1.0 | Add Animations To The Enemy - We need to add some working animations to the enemies, that react to their AI. | priority | add animations to the enemy we need to add some working animations to the enemies that react to their ai | 1 |
26,935 | 2,688,969,402 | IssuesEvent | 2015-03-31 06:28:42 | cs2103jan2015-t09-4j/main | https://api.github.com/repos/cs2103jan2015-t09-4j/main | closed | sort: keep both the gui task list and Task list in order according to Date | priority.medium Task | If got end date use end date. If not use start date.
| 1.0 | sort: keep both the gui task list and Task list in order according to Date - If got end date use end date. If not use start date.
| priority | sort keep both the gui task list and task list in order according to date if got end date use end date if not use start date | 1 |
531,136 | 15,441,630,424 | IssuesEvent | 2021-03-08 06:17:50 | code4moldova/voluntar-web | https://api.github.com/repos/code4moldova/voluntar-web | opened | [demand] add information about cluster and volunteer for in_progress, canceled and resolved demands | Priority: Medium | - add link to the cluster to the tabs or in the edit forms
- add volunteer information like name and phone number to the edit form or the tabs | 1.0 | [demand] add information about cluster and volunteer for in_progress, canceled and resolved demands - - add link to the cluster to the tabs or in the edit forms
- add volunteer information like name and phone number to the edit form or the tabs | priority | add information about cluster and volunteer for in progress canceled and resolved demands add link to the cluster to the tabs or in the edit forms add volunteer information like name and phone number to the edit form or the tabs | 1 |
254,218 | 8,071,421,325 | IssuesEvent | 2018-08-06 13:09:14 | andgein/SIStema | https://api.github.com/repos/andgein/SIStema | opened | Вики: для сохранения изменений приходится долго проматывать вниз | group:ux group:wiki priority:2:medium type:feature | В идеале, сохранить изменения должно быть можно из любого положения скрола | 1.0 | Вики: для сохранения изменений приходится долго проматывать вниз - В идеале, сохранить изменения должно быть можно из любого положения скрола | priority | вики для сохранения изменений приходится долго проматывать вниз в идеале сохранить изменения должно быть можно из любого положения скрола | 1 |
502,380 | 14,545,153,942 | IssuesEvent | 2020-12-15 19:13:46 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | [KT] DMSP/MEL Support pack create process to request addition of new guidelines | Priority - Medium Type - Enhancement | Issue description

| 1.0 | [KT] DMSP/MEL Support pack create process to request addition of new guidelines - Issue description

| priority | dmsp mel support pack create process to request addition of new guidelines issue description | 1 |
39,973 | 2,862,013,980 | IssuesEvent | 2015-06-04 00:17:00 | dart-lang/smoke | https://api.github.com/repos/dart-lang/smoke | opened | support generic type parameters in smoke/polymer build | bug PolymerMilestone-Next Priority-Medium | <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as https://github.com/dart-lang/sdk/issues/18491_
----
Smoke should not record <T> for the type of a field, it should use 'dynamic' instead.
see issue #18481 for an example bug.
| 1.0 | support generic type parameters in smoke/polymer build - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as https://github.com/dart-lang/sdk/issues/18491_
----
Smoke should not record <T> for the type of a field, it should use 'dynamic' instead.
see issue #18481 for an example bug.
| priority | support generic type parameters in smoke polymer build issue by originally opened as smoke should not record lt t gt for the type of a field it should use dynamic instead see issue for an example bug | 1 |
240,554 | 7,802,890,763 | IssuesEvent | 2018-06-10 17:30:36 | OperationCode/operationcode_frontend | https://api.github.com/repos/OperationCode/operationcode_frontend | closed | Update David Molina's title on landing page | Priority: Medium Type: Content Change | <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?

David Molina is no longer the Executive Director, he is now the Founder / Board Chair.
## What should your feature do?
Update the text on the landing page to reflect his correct role.
| 1.0 | Update David Molina's title on landing page - <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?

David Molina is no longer the Executive Director, he is now the Founder / Board Chair.
## What should your feature do?
Update the text on the landing page to reflect his correct role.
| priority | update david molina s title on landing page feature why is this feature being added david molina is no longer the executive director he is now the founder board chair what should your feature do update the text on the landing page to reflect his correct role | 1 |
632,146 | 20,174,778,443 | IssuesEvent | 2022-02-10 13:38:14 | SAP/xsk | https://api.github.com/repos/SAP/xsk | closed | [XSOData] Create entity returns 500 but is created successfully | bug priority-medium | **Describe the bug**
Create entity returns 500 but the record is created in HANA Cloud
> What version of the XSK are you using?
latest
**To Reproduce**
Steps to reproduce the behavior:
1. Import `products` sample
2. `POST http://localhost:8080/services/v4/web/products/odata/Products.xsodata/Orders` with body
```json
{
"Id": "123",
}
```
3. See the following error returned
Status Code: 500
Body:
```xml
<?xml version='1.0' encoding='UTF-8'?>
<error xmlns="http://schemas.microsoft.com/ado/2007/08/dataservices/metadata">
<code/>
<message xml:lang="en">The metadata do not allow a null value for property 'CustomerName'.</message>
</error>
```
4. Go to the HANA Cloud db explorer and see that the record is created successfully
**Expected behavior**
As per [documentation](https://help.sap.com/viewer/52715f71adba4aaeb480d946c742d1f6/2.0.03/en-US/78746b112d794f459978eb55b20a8a6f.html#loio78746b112d794f459978eb55b20a8a6f__section_dgb_vyj_fn) - if neither `null` nor `not null` are specified the default value is `null` which means that `NULL` values are allowed. So I should not be getting an error
**Additional context**
The following exception is logged
```
org.apache.olingo.odata2.core.ep.EntityProviderProducerException: The metadata do not allow a null value for property 'CustomerSurname'.
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.appendProperty(XmlPropertyEntityProducer.java:223) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.append(XmlPropertyEntityProducer.java:79) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.appendPropertyNameValue(AtomEntryEntityProducer.java:659) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.appendProperties(AtomEntryEntityProducer.java:636) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.append(AtomEntryEntityProducer.java:135) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.AtomEntityProvider.writeEntry(AtomEntityProvider.java:170) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.ProviderFacadeImpl.writeEntry(ProviderFacadeImpl.java:129) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.api.ep.EntityProvider.writeEntry(EntityProvider.java:608) ~[olingo-odata2-api-2.0.11.jar:2.0.11]
at org.eclipse.dirigible.engine.odata2.sql.processor.AbstractSQLProcessor.createEntity(AbstractSQLProcessor.java:497) ~[dirigible-odata-core-6.1.7.jar:na]
at org.apache.olingo.odata2.core.Dispatcher.dispatch(Dispatcher.java:79) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ODataRequestHandler.handle(ODataRequestHandler.java:131) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.rest.ODataSubLocator.handle(ODataSubLocator.java:164) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.rest.ODataSubLocator.handlePost(ODataSubLocator.java:86) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:201) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:285) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:104) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:265) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:225) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:298) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:217) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) ~[servlet-api.jar:na]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:273) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) ~[tomcat-websocket.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.invoke(ApplicationDispatcher.java:728) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.processRequest(ApplicationDispatcher.java:470) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.doForward(ApplicationDispatcher.java:395) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.forward(ApplicationDispatcher.java:316) ~[catalina.jar:8.5.43]
at com.sap.xsk.xsodata.ds.filter.XSODataForwardFilter.doFilter(XSODataForwardFilter.java:47) ~[xsk-modules-engines-xsodata-0.12.0-SNAPSHOT.jar:na]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at com.sap.xsk.xsaccess.ds.filter.XSKSecurityFilter.doFilter(XSKSecurityFilter.java:138) ~[xsk-modules-engines-xssecurity-0.12.0-SNAPSHOT.jar:na]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.filters.CorsFilter.handleNonCORS(CorsFilter.java:364) ~[catalina.jar:8.5.43]
at org.apache.catalina.filters.CorsFilter.doFilter(CorsFilter.java:170) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.eclipse.jetty.servlets.DoSFilter.doFilterChain(DoSFilter.java:482) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.eclipse.jetty.servlets.DoSFilter.doFilter(DoSFilter.java:327) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.eclipse.jetty.servlets.DoSFilter.doFilter(DoSFilter.java:297) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.eclipse.jetty.servlets.QoSFilter.doFilter(QoSFilter.java:203) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:199) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) ~[catalina.jar:8.5.43]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:610) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:137) ~[catalina.jar:8.5.43]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) ~[catalina.jar:8.5.43]
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:660) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) ~[catalina.jar:8.5.43]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) ~[catalina.jar:8.5.43]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:798) ~[tomcat-coyote.jar:8.5.43]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) ~[tomcat-coyote.jar:8.5.43]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:808) ~[tomcat-coyote.jar:8.5.43]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1498) ~[tomcat-coyote.jar:8.5.43]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) ~[tomcat-coyote.jar:8.5.43]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) ~[tomcat-util.jar:8.5.43]
at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na]
Caused by: org.apache.olingo.odata2.api.edm.EdmSimpleTypeException: The metadata do not allow a null value.
at org.apache.olingo.odata2.core.edm.AbstractSimpleType.valueToString(AbstractSimpleType.java:104) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.appendProperty(XmlPropertyEntityProducer.java:220) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
... 78 common frames omitted
```
Caused by #825 | 1.0 | [XSOData] Create entity returns 500 but is created successfully - **Describe the bug**
Create entity returns 500 but the record is created in HANA Cloud
> What version of the XSK are you using?
latest
**To Reproduce**
Steps to reproduce the behavior:
1. Import `products` sample
2. `POST http://localhost:8080/services/v4/web/products/odata/Products.xsodata/Orders` with body
```json
{
"Id": "123",
}
```
3. See the following error returned
Status Code: 500
Body:
```xml
<?xml version='1.0' encoding='UTF-8'?>
<error xmlns="http://schemas.microsoft.com/ado/2007/08/dataservices/metadata">
<code/>
<message xml:lang="en">The metadata do not allow a null value for property 'CustomerName'.</message>
</error>
```
4. Go to the HANA Cloud db explorer and see that the record is created successfully
**Expected behavior**
As per [documentation](https://help.sap.com/viewer/52715f71adba4aaeb480d946c742d1f6/2.0.03/en-US/78746b112d794f459978eb55b20a8a6f.html#loio78746b112d794f459978eb55b20a8a6f__section_dgb_vyj_fn) - if neither `null` nor `not null` are specified the default value is `null` which means that `NULL` values are allowed. So I should not be getting an error
**Additional context**
The following exception is logged
```
org.apache.olingo.odata2.core.ep.EntityProviderProducerException: The metadata do not allow a null value for property 'CustomerSurname'.
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.appendProperty(XmlPropertyEntityProducer.java:223) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.append(XmlPropertyEntityProducer.java:79) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.appendPropertyNameValue(AtomEntryEntityProducer.java:659) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.appendProperties(AtomEntryEntityProducer.java:636) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.AtomEntryEntityProducer.append(AtomEntryEntityProducer.java:135) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.AtomEntityProvider.writeEntry(AtomEntityProvider.java:170) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.ProviderFacadeImpl.writeEntry(ProviderFacadeImpl.java:129) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.api.ep.EntityProvider.writeEntry(EntityProvider.java:608) ~[olingo-odata2-api-2.0.11.jar:2.0.11]
at org.eclipse.dirigible.engine.odata2.sql.processor.AbstractSQLProcessor.createEntity(AbstractSQLProcessor.java:497) ~[dirigible-odata-core-6.1.7.jar:na]
at org.apache.olingo.odata2.core.Dispatcher.dispatch(Dispatcher.java:79) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ODataRequestHandler.handle(ODataRequestHandler.java:131) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.rest.ODataSubLocator.handle(ODataSubLocator.java:164) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.rest.ODataSubLocator.handlePost(ODataSubLocator.java:86) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:201) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:285) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:104) ~[cxf-rt-frontend-jaxrs-3.4.2.jar:3.4.2]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) ~[cxf-core-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:265) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:225) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:298) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:217) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) ~[servlet-api.jar:na]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:273) ~[cxf-rt-transports-http-3.4.2.jar:3.4.2]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) ~[tomcat-websocket.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.invoke(ApplicationDispatcher.java:728) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.processRequest(ApplicationDispatcher.java:470) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.doForward(ApplicationDispatcher.java:395) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationDispatcher.forward(ApplicationDispatcher.java:316) ~[catalina.jar:8.5.43]
at com.sap.xsk.xsodata.ds.filter.XSODataForwardFilter.doFilter(XSODataForwardFilter.java:47) ~[xsk-modules-engines-xsodata-0.12.0-SNAPSHOT.jar:na]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at com.sap.xsk.xsaccess.ds.filter.XSKSecurityFilter.doFilter(XSKSecurityFilter.java:138) ~[xsk-modules-engines-xssecurity-0.12.0-SNAPSHOT.jar:na]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.filters.CorsFilter.handleNonCORS(CorsFilter.java:364) ~[catalina.jar:8.5.43]
at org.apache.catalina.filters.CorsFilter.doFilter(CorsFilter.java:170) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.eclipse.jetty.servlets.DoSFilter.doFilterChain(DoSFilter.java:482) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.eclipse.jetty.servlets.DoSFilter.doFilter(DoSFilter.java:327) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.eclipse.jetty.servlets.DoSFilter.doFilter(DoSFilter.java:297) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.eclipse.jetty.servlets.QoSFilter.doFilter(QoSFilter.java:203) ~[jetty-servlets-9.4.12.v20180830.jar:9.4.12.v20180830]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:199) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) ~[catalina.jar:8.5.43]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:610) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:137) ~[catalina.jar:8.5.43]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) ~[catalina.jar:8.5.43]
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:660) ~[catalina.jar:8.5.43]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) ~[catalina.jar:8.5.43]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) ~[catalina.jar:8.5.43]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:798) ~[tomcat-coyote.jar:8.5.43]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) ~[tomcat-coyote.jar:8.5.43]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:808) ~[tomcat-coyote.jar:8.5.43]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1498) ~[tomcat-coyote.jar:8.5.43]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) ~[tomcat-coyote.jar:8.5.43]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) ~[tomcat-util.jar:8.5.43]
at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na]
Caused by: org.apache.olingo.odata2.api.edm.EdmSimpleTypeException: The metadata do not allow a null value.
at org.apache.olingo.odata2.core.edm.AbstractSimpleType.valueToString(AbstractSimpleType.java:104) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
at org.apache.olingo.odata2.core.ep.producer.XmlPropertyEntityProducer.appendProperty(XmlPropertyEntityProducer.java:220) ~[olingo-odata2-core-2.0.11.jar:2.0.11]
... 78 common frames omitted
```
Caused by #825 | priority | create entity returns but is created successfully describe the bug create entity returns but the record is created in hana cloud what version of the xsk are you using latest to reproduce steps to reproduce the behavior import products sample post with body json id see the following error returned status code body xml error xmlns the metadata do not allow a null value for property customername go to the hana cloud db explorer and see that the record is created successfully expected behavior as per if neither null nor not null are specified the default value is null which means that null values are allowed so i should not be getting an error additional context the following exception is logged org apache olingo core ep entityproviderproducerexception the metadata do not allow a null value for property customersurname at org apache olingo core ep producer xmlpropertyentityproducer appendproperty xmlpropertyentityproducer java at org apache olingo core ep producer xmlpropertyentityproducer append xmlpropertyentityproducer java at org apache olingo core ep producer atomentryentityproducer appendpropertynamevalue atomentryentityproducer java at org apache olingo core ep producer atomentryentityproducer appendproperties atomentryentityproducer java at org apache olingo core ep producer atomentryentityproducer append atomentryentityproducer java at org apache olingo core ep atomentityprovider writeentry atomentityprovider java at org apache olingo core ep providerfacadeimpl writeentry providerfacadeimpl java at org apache olingo api ep entityprovider writeentry entityprovider java at org eclipse dirigible engine sql processor abstractsqlprocessor createentity abstractsqlprocessor java at org apache olingo core dispatcher dispatch dispatcher java at org apache olingo core odatarequesthandler handle odatarequesthandler java at org apache olingo core rest odatasublocator handle odatasublocator java at org apache olingo core rest odatasublocator handlepost odatasublocator java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org apache cxf service invoker abstractinvoker performinvocation abstractinvoker java at org apache cxf service invoker abstractinvoker invoke abstractinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf interceptor serviceinvokerinterceptor run serviceinvokerinterceptor java at org apache cxf interceptor serviceinvokerinterceptor handlemessage serviceinvokerinterceptor java at org apache cxf phase phaseinterceptorchain dointercept phaseinterceptorchain java at org apache cxf transport chaininitiationobserver onmessage chaininitiationobserver java at org apache cxf transport http abstracthttpdestination invoke abstracthttpdestination java at org apache cxf transport servlet servletcontroller invokedestination servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet cxfnonspringservlet invoke cxfnonspringservlet java at org apache cxf transport servlet abstracthttpservlet handlerequest abstracthttpservlet java at org apache cxf transport servlet abstracthttpservlet dopost abstracthttpservlet java at javax servlet http httpservlet service httpservlet java at org apache cxf transport servlet abstracthttpservlet service abstracthttpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core applicationdispatcher invoke applicationdispatcher java at org apache catalina core applicationdispatcher processrequest applicationdispatcher java at org apache catalina core applicationdispatcher doforward applicationdispatcher java at org apache catalina core applicationdispatcher forward applicationdispatcher java at com sap xsk xsodata ds filter xsodataforwardfilter dofilter xsodataforwardfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at com sap xsk xsaccess ds filter xsksecurityfilter dofilter xsksecurityfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina filters corsfilter handlenoncors corsfilter java at org apache catalina filters corsfilter dofilter corsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org eclipse jetty servlets dosfilter dofilterchain dosfilter java at org eclipse jetty servlets dosfilter dofilter dosfilter java at org eclipse jetty servlets dosfilter dofilter dosfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org eclipse jetty servlets qosfilter dofilter qosfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java base java lang thread run thread java caused by org apache olingo api edm edmsimpletypeexception the metadata do not allow a null value at org apache olingo core edm abstractsimpletype valuetostring abstractsimpletype java at org apache olingo core ep producer xmlpropertyentityproducer appendproperty xmlpropertyentityproducer java common frames omitted caused by | 1 |
108,738 | 4,349,845,397 | IssuesEvent | 2016-07-30 21:09:10 | CovertJaguar/Railcraft | https://api.github.com/repos/CovertJaguar/Railcraft | closed | Increase Turbine Steam Usage/Output to match Boiler Output | enhancement priority-medium | The Steam Turbine currently uses 320 mB of Steam per tick.
A 36LP Boiler outputs 360 mB per tick.
A 18HP Boiler outputs 360 mB per tick.
A 36HP Boiler outputs 720 mB per tick (360 x 2).
It would make sense for the Steam Turbine to use 360 mB of Steam per tick in order to match the boiler outputs.
This would result in about a 1.125x increase of usage and output. Going from 200 EU/t to 225 EU/t.
Comments. | 1.0 | Increase Turbine Steam Usage/Output to match Boiler Output - The Steam Turbine currently uses 320 mB of Steam per tick.
A 36LP Boiler outputs 360 mB per tick.
A 18HP Boiler outputs 360 mB per tick.
A 36HP Boiler outputs 720 mB per tick (360 x 2).
It would make sense for the Steam Turbine to use 360 mB of Steam per tick in order to match the boiler outputs.
This would result in about a 1.125x increase of usage and output. Going from 200 EU/t to 225 EU/t.
Comments. | priority | increase turbine steam usage output to match boiler output the steam turbine currently uses mb of steam per tick a boiler outputs mb per tick a boiler outputs mb per tick a boiler outputs mb per tick x it would make sense for the steam turbine to use mb of steam per tick in order to match the boiler outputs this would result in about a increase of usage and output going from eu t to eu t comments | 1 |
49,643 | 3,003,799,126 | IssuesEvent | 2015-07-25 08:25:52 | jayway/powermock | https://api.github.com/repos/jayway/powermock | opened | javaagent/java 7 and custom javassist not working together: stack shape inconsistent | bug imported Priority-Medium | _From [gdel...@gmail.com](https://code.google.com/u/108881632138955056358/) on December 02, 2014 18:23:18_
What steps will reproduce the problem? 1. Write an empty test with a before class method
2. add as content of the beforeClass method this:
ClassPool cp = ClassPool.getDefault();
CtClass cc = cp.get("java.lang.String");
ClassFile cf = cc.getClassFile();
List<MethodInfo> methods = cf.getMethods();
for (MethodInfo m : methods) {
System.out.println(m.getName());
}
3. Add the javaagent to the VM argument : -javaagent:powermock-module-javaagent-1.6.0.jar, in the run classpath, set the javaagent jar first
4. Make sure to set the compliance of the project to 1.7 (using a JRE 7)
4. Launch test with the standard JUnit4 runner What is the expected output? What do you see instead? I expect the test to pass as with Java 6, instead I get:
java.lang.VerifyError: JVMVRFY012 stack shape inconsistent; class=TestPSC, method=beforeClass()V, pc=29
at java.lang.J9VMInternals.prepareClassImpl(Native Method)
at java.lang.J9VMInternals.prepare(J9VMInternals.java:430)
at java.lang.Class.getMethod(Class.java:1061)
at org.junit.internal.builders.SuiteMethodBuilder.hasSuiteMethod(SuiteMethodBuilder.java:18)
at org.junit.internal.builders.SuiteMethodBuilder.runnerForClass(SuiteMethodBuilder.java:10)
at org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java:59)
at org.junit.internal.builders.AllDefaultPossibilitiesBuilder.runnerForClass(AllDefaultPossibilitiesBuilder.java:26)
at org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java:59)
at org.junit.internal.requests.ClassRequest.getRunner(ClassRequest.java:26)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.<init>(JUnit4TestReference.java:33)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestClassReference.<init>(JUnit4TestClassReference.java:25)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestLoader.createTest(JUnit4TestLoader.java:48)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestLoader.loadTests(JUnit4TestLoader.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:444)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) What version of the product are you using? On what operating system? JUnit 4.11, Powermock 1.6.0, javassist 3.18.2-GA Please provide any additional information below. Launching with a compliance 1.6 works perfectly fine, launching with 1.7 fails. I have also tried to add the option -XX:-UseSplitVerifier. And the javaagent jar is first in the classpath.
Any ideas?
Thank you!
Guillaume
The class is:
@Rule
public PowerMockRule rule = new PowerMockRule ();
static { PowerMockAgent .initializeIfNeeded();
}
@BeforeClass
public static void beforeClass() throws NotFoundException, CannotCompileException {
ClassPool cp = ClassPool.getDefault();
CtClass cc = cp.get("java.lang.String");
ClassFile cf = cc.getClassFile();
List<MethodInfo> methods = cf.getMethods();
for (MethodInfo m : methods) {
System.out.println(m.getName());
}
}
@Test
public void test() throws Exception {
}
_Original issue: http://code.google.com/p/powermock/issues/detail?id=529_ | 1.0 | javaagent/java 7 and custom javassist not working together: stack shape inconsistent - _From [gdel...@gmail.com](https://code.google.com/u/108881632138955056358/) on December 02, 2014 18:23:18_
What steps will reproduce the problem? 1. Write an empty test with a before class method
2. add as content of the beforeClass method this:
ClassPool cp = ClassPool.getDefault();
CtClass cc = cp.get("java.lang.String");
ClassFile cf = cc.getClassFile();
List<MethodInfo> methods = cf.getMethods();
for (MethodInfo m : methods) {
System.out.println(m.getName());
}
3. Add the javaagent to the VM argument : -javaagent:powermock-module-javaagent-1.6.0.jar, in the run classpath, set the javaagent jar first
4. Make sure to set the compliance of the project to 1.7 (using a JRE 7)
4. Launch test with the standard JUnit4 runner What is the expected output? What do you see instead? I expect the test to pass as with Java 6, instead I get:
java.lang.VerifyError: JVMVRFY012 stack shape inconsistent; class=TestPSC, method=beforeClass()V, pc=29
at java.lang.J9VMInternals.prepareClassImpl(Native Method)
at java.lang.J9VMInternals.prepare(J9VMInternals.java:430)
at java.lang.Class.getMethod(Class.java:1061)
at org.junit.internal.builders.SuiteMethodBuilder.hasSuiteMethod(SuiteMethodBuilder.java:18)
at org.junit.internal.builders.SuiteMethodBuilder.runnerForClass(SuiteMethodBuilder.java:10)
at org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java:59)
at org.junit.internal.builders.AllDefaultPossibilitiesBuilder.runnerForClass(AllDefaultPossibilitiesBuilder.java:26)
at org.junit.runners.model.RunnerBuilder.safeRunnerForClass(RunnerBuilder.java:59)
at org.junit.internal.requests.ClassRequest.getRunner(ClassRequest.java:26)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.<init>(JUnit4TestReference.java:33)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestClassReference.<init>(JUnit4TestClassReference.java:25)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestLoader.createTest(JUnit4TestLoader.java:48)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestLoader.loadTests(JUnit4TestLoader.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:444)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) What version of the product are you using? On what operating system? JUnit 4.11, Powermock 1.6.0, javassist 3.18.2-GA Please provide any additional information below. Launching with a compliance 1.6 works perfectly fine, launching with 1.7 fails. I have also tried to add the option -XX:-UseSplitVerifier. And the javaagent jar is first in the classpath.
Any ideas?
Thank you!
Guillaume
The class is:
@Rule
public PowerMockRule rule = new PowerMockRule ();
static { PowerMockAgent .initializeIfNeeded();
}
@BeforeClass
public static void beforeClass() throws NotFoundException, CannotCompileException {
ClassPool cp = ClassPool.getDefault();
CtClass cc = cp.get("java.lang.String");
ClassFile cf = cc.getClassFile();
List<MethodInfo> methods = cf.getMethods();
for (MethodInfo m : methods) {
System.out.println(m.getName());
}
}
@Test
public void test() throws Exception {
}
_Original issue: http://code.google.com/p/powermock/issues/detail?id=529_ | priority | javaagent java and custom javassist not working together stack shape inconsistent from on december what steps will reproduce the problem write an empty test with a before class method add as content of the beforeclass method this classpool cp classpool getdefault ctclass cc cp get java lang string classfile cf cc getclassfile list methods cf getmethods for methodinfo m methods system out println m getname add the javaagent to the vm argument javaagent powermock module javaagent jar in the run classpath set the javaagent jar first make sure to set the compliance of the project to using a jre launch test with the standard runner what is the expected output what do you see instead i expect the test to pass as with java instead i get java lang verifyerror stack shape inconsistent class testpsc method beforeclass v pc at java lang prepareclassimpl native method at java lang prepare java at java lang class getmethod class java at org junit internal builders suitemethodbuilder hassuitemethod suitemethodbuilder java at org junit internal builders suitemethodbuilder runnerforclass suitemethodbuilder java at org junit runners model runnerbuilder saferunnerforclass runnerbuilder java at org junit internal builders alldefaultpossibilitiesbuilder runnerforclass alldefaultpossibilitiesbuilder java at org junit runners model runnerbuilder saferunnerforclass runnerbuilder java at org junit internal requests classrequest getrunner classrequest java at org eclipse jdt internal runner java at org eclipse jdt internal runner java at org eclipse jdt internal runner createtest java at org eclipse jdt internal runner loadtests java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner run remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner main remotetestrunner java what version of the product are you using on what operating system junit powermock javassist ga please provide any additional information below launching with a compliance works perfectly fine launching with fails i have also tried to add the option xx usesplitverifier and the javaagent jar is first in the classpath any ideas thank you guillaume the class is rule public powermockrule rule new powermockrule static powermockagent initializeifneeded beforeclass public static void beforeclass throws notfoundexception cannotcompileexception classpool cp classpool getdefault ctclass cc cp get java lang string classfile cf cc getclassfile list methods cf getmethods for methodinfo m methods system out println m getname test public void test throws exception original issue | 1 |
172,955 | 6,518,308,684 | IssuesEvent | 2017-08-28 07:25:51 | vmware/admiral | https://api.github.com/repos/vmware/admiral | closed | The way to add host is still existing in the management portal of vic 1.2 | kind/bug priority/medium | <img width="698" alt="2017-08-08 12 57 53" src="https://user-images.githubusercontent.com/5753287/29056964-70e07d24-7c39-11e7-901a-487e4c687d07.png">
Create a volume;
Click the disk icon at the left of the footer of above create volume;
A '+Add Host' button is located at the left-top | 1.0 | The way to add host is still existing in the management portal of vic 1.2 - <img width="698" alt="2017-08-08 12 57 53" src="https://user-images.githubusercontent.com/5753287/29056964-70e07d24-7c39-11e7-901a-487e4c687d07.png">
Create a volume;
Click the disk icon at the left of the footer of above create volume;
A '+Add Host' button is located at the left-top | priority | the way to add host is still existing in the management portal of vic img width alt src create a volume click the disk icon at the left of the footer of above create volume a add host button is located at the left top | 1 |
676,729 | 23,135,813,061 | IssuesEvent | 2022-07-28 14:14:51 | owncloud/web | https://api.github.com/repos/owncloud/web | closed | Sharing Panel: Show label instead of description | Type:Bug Priority:p3-medium GA-Blocker | Should show the label "Viewer" instead of whole description:
 | 1.0 | Sharing Panel: Show label instead of description - Should show the label "Viewer" instead of whole description:
 | priority | sharing panel show label instead of description should show the label viewer instead of whole description | 1 |
78,037 | 3,508,887,491 | IssuesEvent | 2016-01-08 19:58:06 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Dataset - Edit Files "Please select..." Message Missing | Component: File Upload & Handling Component: UX & Upgrade Priority: Medium Status: QA | When you click the "Edit Files" button on the dataset pg, you are provided a dropdown with the options to Delete, Metadata, Restrict, Unrestrict, Tags. If you have no files selected, and choose Delete and Metadata, you get a message "Please select a file or files..." If you choose Restrict, Unrestrict, Tags, there is no such message -- for Unrestrict it even displays a success message.
There needs to be a "Please select a file or files..." popup and message for these workflows.
- Dataset pg Files tab > Edit Files button > Restirct
- Dataset pg Files tab > Edit Files button > Unrestirct
- Dataset pg Files tab > Edit Files button > Tags | 1.0 | Dataset - Edit Files "Please select..." Message Missing - When you click the "Edit Files" button on the dataset pg, you are provided a dropdown with the options to Delete, Metadata, Restrict, Unrestrict, Tags. If you have no files selected, and choose Delete and Metadata, you get a message "Please select a file or files..." If you choose Restrict, Unrestrict, Tags, there is no such message -- for Unrestrict it even displays a success message.
There needs to be a "Please select a file or files..." popup and message for these workflows.
- Dataset pg Files tab > Edit Files button > Restirct
- Dataset pg Files tab > Edit Files button > Unrestirct
- Dataset pg Files tab > Edit Files button > Tags | priority | dataset edit files please select message missing when you click the edit files button on the dataset pg you are provided a dropdown with the options to delete metadata restrict unrestrict tags if you have no files selected and choose delete and metadata you get a message please select a file or files if you choose restrict unrestrict tags there is no such message for unrestrict it even displays a success message there needs to be a please select a file or files popup and message for these workflows dataset pg files tab edit files button restirct dataset pg files tab edit files button unrestirct dataset pg files tab edit files button tags | 1 |
1,461 | 2,514,592,603 | IssuesEvent | 2015-01-15 12:51:56 | eclipsesource/tabris-js | https://api.github.com/repos/eclipsesource/tabris-js | opened | API to set tintColor for Actions | enhancement priority: medium | Currently images used in Actions are tinted blue by default.
Images (PNGs) with transparency are used as alpha masks.
It would make sense to provide API to disable or change the tintColor.
The current default behavior prevents us from using a gravatar image as Action. | 1.0 | API to set tintColor for Actions - Currently images used in Actions are tinted blue by default.
Images (PNGs) with transparency are used as alpha masks.
It would make sense to provide API to disable or change the tintColor.
The current default behavior prevents us from using a gravatar image as Action. | priority | api to set tintcolor for actions currently images used in actions are tinted blue by default images pngs with transparency are used as alpha masks it would make sense to provide api to disable or change the tintcolor the current default behavior prevents us from using a gravatar image as action | 1 |
22,162 | 2,645,696,278 | IssuesEvent | 2015-03-13 01:12:37 | prikhi/evoluspencil | https://api.github.com/repos/prikhi/evoluspencil | opened | starting the names of the pages with "*" will not save png | 1 star bug imported Priority-Medium | _From [pvojni...@gmail.com](https://code.google.com/u/100848826838861863508/) on September 14, 2008 16:47:44_
What steps will reproduce the problem? 1. create a page
2. name it starting with "*". like "*page1"
3. click to export document as PNGs..
4. chose the folder and click save What is the expected output? What do you see instead? I would like to get a PNG of my page. What version of the product are you using? On what operating system? win xp, pencil 1.0 Please provide any additional information below. Windows does not handle files containing some chars: /\:* etc.. you will
have to replace them..
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=53_ | 1.0 | starting the names of the pages with "*" will not save png - _From [pvojni...@gmail.com](https://code.google.com/u/100848826838861863508/) on September 14, 2008 16:47:44_
What steps will reproduce the problem? 1. create a page
2. name it starting with "*". like "*page1"
3. click to export document as PNGs..
4. chose the folder and click save What is the expected output? What do you see instead? I would like to get a PNG of my page. What version of the product are you using? On what operating system? win xp, pencil 1.0 Please provide any additional information below. Windows does not handle files containing some chars: /\:* etc.. you will
have to replace them..
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=53_ | priority | starting the names of the pages with will not save png from on september what steps will reproduce the problem create a page name it starting with like click to export document as pngs chose the folder and click save what is the expected output what do you see instead i would like to get a png of my page what version of the product are you using on what operating system win xp pencil please provide any additional information below windows does not handle files containing some chars etc you will have to replace them original issue | 1 |
398,491 | 11,741,570,193 | IssuesEvent | 2020-03-11 22:04:43 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Aanmelden als Thaliaan via site | board feature priority: medium | In GitLab by lennartjansen on Oct 7, 2016, 15:25
From: [registrations/README.md](https://gitlab.science.ru.nl/thalia/concrexit/blob/feature/registrations/website/registrations/README.md)
Registrations
=====
This document explains how the registrations module behaviour is defined.
The behaviour of upgrading an existing 'year' membership to a 'study' membership (until graduation) is taken from the HR. If the HR ever changes this behaviour should be changed to reflect those changes.
_Note that registrations and renewals for supporters are implemented in the models, there are simply no views providing this functionality. If we ever want to implement this then it would be best to create a complete new form just for supporter registrations._
## New member registration
### Frontend
- User enters information
- User accepts privacy policy
- System validates info
- Correct address
- Valid and unique email address
- Checked against existing users
- Privacy policy accepted
- If the selected member type is 'member':
- valid and unique student number
- selected programme
- cohort
- Registration model created (status: Awaiting email confirmation)
- Email address confirmation sent
- User confirms email address
- Registration model status changed (status: Ready for review)
### Backend
1. Admin accepts registration
- System checks if username is unique
- If it's not unique a username can be entered manually
- If it's still not unique the registration cannot be accepted
- If it's unique the generated username will be added to the registration
- Payment model is created (processed: False)
- Amount is calculated based on the selected length ('study' or 'year')
- Values are located in thaliawebsite.settings
- Email is sent as acceptance confirmation containg instructions for [payment](#payment-processing)
2. Admin rejects registration
- Email is sent as rejection message
## Existing user membership renewal
### Frontend
- User enters information (length, type)
- If latest membership has not ended yet: always allow 'study' length
- If latest membership has ended or ends within 1 month: also allow 'year' length
- If latest membership is 'study' and did not end: do not allow renewal
- Renewal model created (status: Ready for review)
### Backend
1. Admin accepts renewal
- Payment model is created (processed: False)
- Amount is calculated based on selected length ('study' or 'year')
- Values are located in thaliawebsite.settings
- If the current membership has not ended yet and an until date is present for that membership and
the selected length is 'study' the amount will be price['study'] - price['year']
- Email is sent as acceptance confirmation containg instructions for [payment](#payment-processing)
2. Admin rejects renewal
- Email is sent as rejection message
## Payment processing
### Backend
- Admin (or the system, if automated using e.g. iDeal) processes payment
- If this is a Registration model then User and Member models are created
- If this is a Renewal model then the Member is retrieved
- A membership is added to the provided Member model based on the provided length
- If the __latest__ (_not current, since there may have been some time between asking for the upgrade and accepting it_) membership has an until date and
the selected length is 'study' that membership will be updated to have None as until date. No new membership will be created.
- During a lecture year the until date will be the 31 August of the lecture year + 1. Thus is you process payments in November 2016 that means the memberships will end on 31 August 2017
- For payments processed in August the lecture year will be increased by 1. So if you process payments in August 2017 that means the memberships will end on 31 August 2018.
- Payment confirmation sent (if this is a Renewal model) | 1.0 | Aanmelden als Thaliaan via site - In GitLab by lennartjansen on Oct 7, 2016, 15:25
From: [registrations/README.md](https://gitlab.science.ru.nl/thalia/concrexit/blob/feature/registrations/website/registrations/README.md)
Registrations
=====
This document explains how the registrations module behaviour is defined.
The behaviour of upgrading an existing 'year' membership to a 'study' membership (until graduation) is taken from the HR. If the HR ever changes this behaviour should be changed to reflect those changes.
_Note that registrations and renewals for supporters are implemented in the models, there are simply no views providing this functionality. If we ever want to implement this then it would be best to create a complete new form just for supporter registrations._
## New member registration
### Frontend
- User enters information
- User accepts privacy policy
- System validates info
- Correct address
- Valid and unique email address
- Checked against existing users
- Privacy policy accepted
- If the selected member type is 'member':
- valid and unique student number
- selected programme
- cohort
- Registration model created (status: Awaiting email confirmation)
- Email address confirmation sent
- User confirms email address
- Registration model status changed (status: Ready for review)
### Backend
1. Admin accepts registration
- System checks if username is unique
- If it's not unique a username can be entered manually
- If it's still not unique the registration cannot be accepted
- If it's unique the generated username will be added to the registration
- Payment model is created (processed: False)
- Amount is calculated based on the selected length ('study' or 'year')
- Values are located in thaliawebsite.settings
- Email is sent as acceptance confirmation containg instructions for [payment](#payment-processing)
2. Admin rejects registration
- Email is sent as rejection message
## Existing user membership renewal
### Frontend
- User enters information (length, type)
- If latest membership has not ended yet: always allow 'study' length
- If latest membership has ended or ends within 1 month: also allow 'year' length
- If latest membership is 'study' and did not end: do not allow renewal
- Renewal model created (status: Ready for review)
### Backend
1. Admin accepts renewal
- Payment model is created (processed: False)
- Amount is calculated based on selected length ('study' or 'year')
- Values are located in thaliawebsite.settings
- If the current membership has not ended yet and an until date is present for that membership and
the selected length is 'study' the amount will be price['study'] - price['year']
- Email is sent as acceptance confirmation containg instructions for [payment](#payment-processing)
2. Admin rejects renewal
- Email is sent as rejection message
## Payment processing
### Backend
- Admin (or the system, if automated using e.g. iDeal) processes payment
- If this is a Registration model then User and Member models are created
- If this is a Renewal model then the Member is retrieved
- A membership is added to the provided Member model based on the provided length
- If the __latest__ (_not current, since there may have been some time between asking for the upgrade and accepting it_) membership has an until date and
the selected length is 'study' that membership will be updated to have None as until date. No new membership will be created.
- During a lecture year the until date will be the 31 August of the lecture year + 1. Thus is you process payments in November 2016 that means the memberships will end on 31 August 2017
- For payments processed in August the lecture year will be increased by 1. So if you process payments in August 2017 that means the memberships will end on 31 August 2018.
- Payment confirmation sent (if this is a Renewal model) | priority | aanmelden als thaliaan via site in gitlab by lennartjansen on oct from registrations this document explains how the registrations module behaviour is defined the behaviour of upgrading an existing year membership to a study membership until graduation is taken from the hr if the hr ever changes this behaviour should be changed to reflect those changes note that registrations and renewals for supporters are implemented in the models there are simply no views providing this functionality if we ever want to implement this then it would be best to create a complete new form just for supporter registrations new member registration frontend user enters information user accepts privacy policy system validates info correct address valid and unique email address checked against existing users privacy policy accepted if the selected member type is member valid and unique student number selected programme cohort registration model created status awaiting email confirmation email address confirmation sent user confirms email address registration model status changed status ready for review backend admin accepts registration system checks if username is unique if it s not unique a username can be entered manually if it s still not unique the registration cannot be accepted if it s unique the generated username will be added to the registration payment model is created processed false amount is calculated based on the selected length study or year values are located in thaliawebsite settings email is sent as acceptance confirmation containg instructions for payment processing admin rejects registration email is sent as rejection message existing user membership renewal frontend user enters information length type if latest membership has not ended yet always allow study length if latest membership has ended or ends within month also allow year length if latest membership is study and did not end do not allow renewal renewal model created status ready for review backend admin accepts renewal payment model is created processed false amount is calculated based on selected length study or year values are located in thaliawebsite settings if the current membership has not ended yet and an until date is present for that membership and the selected length is study the amount will be price price email is sent as acceptance confirmation containg instructions for payment processing admin rejects renewal email is sent as rejection message payment processing backend admin or the system if automated using e g ideal processes payment if this is a registration model then user and member models are created if this is a renewal model then the member is retrieved a membership is added to the provided member model based on the provided length if the latest not current since there may have been some time between asking for the upgrade and accepting it membership has an until date and the selected length is study that membership will be updated to have none as until date no new membership will be created during a lecture year the until date will be the august of the lecture year thus is you process payments in november that means the memberships will end on august for payments processed in august the lecture year will be increased by so if you process payments in august that means the memberships will end on august payment confirmation sent if this is a renewal model | 1 |
538,102 | 15,762,633,821 | IssuesEvent | 2021-03-31 11:18:40 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | [BUG] Alarm not switched off unless we refresh page | bug mgmt-navigation-team-bugs priority/medium | <!---
If you are a member of the Netdata organization, add the label 'internal submit'.
-->
**Describe the bug**
In case we have concurrently in a room:
- An alarm getting switched off (from either state, critical or warning)
- No other active alarms
In the top bar we continue seeing the alarm being active even if it is no longer the case. If we refresh the page, we see the correct information of 0 alarms.
**To Reproduce**
Steps to reproduce the behavior:
1. Trigger a single alarm (to an agent claimed to cloud in a room with no other alarms).
2. Switch off the alarm.
3. In the war room where the node is claimed, in the top bar where alarms are presented, the alarm will remain visible until we reload the page.
**Expected behavior**
We should see no active alarms in the top bar (see atached screenshot)
**Screenshots**
If applicable, add screenshots to help explain your problem.

| 1.0 | [BUG] Alarm not switched off unless we refresh page - <!---
If you are a member of the Netdata organization, add the label 'internal submit'.
-->
**Describe the bug**
In case we have concurrently in a room:
- An alarm getting switched off (from either state, critical or warning)
- No other active alarms
In the top bar we continue seeing the alarm being active even if it is no longer the case. If we refresh the page, we see the correct information of 0 alarms.
**To Reproduce**
Steps to reproduce the behavior:
1. Trigger a single alarm (to an agent claimed to cloud in a room with no other alarms).
2. Switch off the alarm.
3. In the war room where the node is claimed, in the top bar where alarms are presented, the alarm will remain visible until we reload the page.
**Expected behavior**
We should see no active alarms in the top bar (see atached screenshot)
**Screenshots**
If applicable, add screenshots to help explain your problem.

| priority | alarm not switched off unless we refresh page if you are a member of the netdata organization add the label internal submit describe the bug in case we have concurrently in a room an alarm getting switched off from either state critical or warning no other active alarms in the top bar we continue seeing the alarm being active even if it is no longer the case if we refresh the page we see the correct information of alarms to reproduce steps to reproduce the behavior trigger a single alarm to an agent claimed to cloud in a room with no other alarms switch off the alarm in the war room where the node is claimed in the top bar where alarms are presented the alarm will remain visible until we reload the page expected behavior we should see no active alarms in the top bar see atached screenshot screenshots if applicable add screenshots to help explain your problem | 1 |
561,192 | 16,613,041,307 | IssuesEvent | 2021-06-02 13:45:34 | Zubiik/pokedex | https://api.github.com/repos/Zubiik/pokedex | opened | feat: Header | feat priority medium | - [ ] Create a arrow button for display search container
- [ ] Return the name of the Pokemon with the API info request | 1.0 | feat: Header - - [ ] Create a arrow button for display search container
- [ ] Return the name of the Pokemon with the API info request | priority | feat header create a arrow button for display search container return the name of the pokemon with the api info request | 1 |
657,588 | 21,797,348,494 | IssuesEvent | 2022-05-15 20:34:41 | Blackoutburst/Wally | https://api.github.com/repos/Blackoutburst/Wally | closed | Create the leaderboard manager | feature medium priority | Create a leaderboard manager class using a singleton patern, it must contain everything needed to create quick leaderboard
Create multiple static list for each leaderboard and sort them before Wally login speed > ram is neccesary here | 1.0 | Create the leaderboard manager - Create a leaderboard manager class using a singleton patern, it must contain everything needed to create quick leaderboard
Create multiple static list for each leaderboard and sort them before Wally login speed > ram is neccesary here | priority | create the leaderboard manager create a leaderboard manager class using a singleton patern it must contain everything needed to create quick leaderboard create multiple static list for each leaderboard and sort them before wally login speed ram is neccesary here | 1 |
782,589 | 27,500,512,235 | IssuesEvent | 2023-03-05 16:42:37 | clt313/SuperballVR | https://api.github.com/repos/clt313/SuperballVR | closed | Superissue for loading screen bugs | priority: medium bug | During beta testing, several loading screen bugs were found in some edge cases. I'll tackle all of these so going to track them all in one issue.
## Bugs List
- [x] When exiting just before the end of a match, time freezes
- [x] Can move outside of loading screen
- Idea: prevent player movement during this time
- [x] Calling fadeOut() while fadeIn() is running causes visual glitch
- Idea: wait for the other function to finish before running the current (have a bool to "lock" the loading screen) | 1.0 | Superissue for loading screen bugs - During beta testing, several loading screen bugs were found in some edge cases. I'll tackle all of these so going to track them all in one issue.
## Bugs List
- [x] When exiting just before the end of a match, time freezes
- [x] Can move outside of loading screen
- Idea: prevent player movement during this time
- [x] Calling fadeOut() while fadeIn() is running causes visual glitch
- Idea: wait for the other function to finish before running the current (have a bool to "lock" the loading screen) | priority | superissue for loading screen bugs during beta testing several loading screen bugs were found in some edge cases i ll tackle all of these so going to track them all in one issue bugs list when exiting just before the end of a match time freezes can move outside of loading screen idea prevent player movement during this time calling fadeout while fadein is running causes visual glitch idea wait for the other function to finish before running the current have a bool to lock the loading screen | 1 |
822,114 | 30,853,657,416 | IssuesEvent | 2023-08-02 18:44:23 | DDMAL/CantusDB | https://api.github.com/repos/DDMAL/CantusDB | closed | Admin Chant Edit page: we should make "title" field longer | priority: medium simple fix | In an email from Debra:
> https://cantusdatabase.org/admin/main_app/source/123611/change/
For a page like this one, could I please have a longer box for the source title? The width of the "provenance note" field would be about right, I think, so that I can see the whole title without having to scroll. | 1.0 | Admin Chant Edit page: we should make "title" field longer - In an email from Debra:
> https://cantusdatabase.org/admin/main_app/source/123611/change/
For a page like this one, could I please have a longer box for the source title? The width of the "provenance note" field would be about right, I think, so that I can see the whole title without having to scroll. | priority | admin chant edit page we should make title field longer in an email from debra for a page like this one could i please have a longer box for the source title the width of the provenance note field would be about right i think so that i can see the whole title without having to scroll | 1 |
174,640 | 6,541,982,299 | IssuesEvent | 2017-09-01 23:21:06 | vmware/vic-product | https://api.github.com/repos/vmware/vic-product | closed | Provisioning of dch-photon fails with Invalid image format | priority/medium status/need-info | When trying to provision containers based on the `default-project/vmware/dch-photon` image, I get the error `Invalid image format: 10.160.94.153:443/default-project/vmware/dch-photon`. The image name should be changed to \<namespace\>/<repo_name>:\<tag\> or \<project\>/<repo_name>:\<tag\> in https://github.com/vmware/vic-product/blob/524510a28da216ba95c86c70e3f7717371794079/installer/packer/scripts/provision_harbor.sh#L56 | 1.0 | Provisioning of dch-photon fails with Invalid image format - When trying to provision containers based on the `default-project/vmware/dch-photon` image, I get the error `Invalid image format: 10.160.94.153:443/default-project/vmware/dch-photon`. The image name should be changed to \<namespace\>/<repo_name>:\<tag\> or \<project\>/<repo_name>:\<tag\> in https://github.com/vmware/vic-product/blob/524510a28da216ba95c86c70e3f7717371794079/installer/packer/scripts/provision_harbor.sh#L56 | priority | provisioning of dch photon fails with invalid image format when trying to provision containers based on the default project vmware dch photon image i get the error invalid image format default project vmware dch photon the image name should be changed to or in | 1 |
258,823 | 8,179,968,667 | IssuesEvent | 2018-08-28 17:59:42 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Law for all plants and all animals | Feature Laws Medium Priority Needs Triage | Version: 0.7.6.3 beta
Steps to Reproduce:
Login to the game and hit L to open up the law window
Try and specify a law for all animals or all plants
Expected behavior:
To write a law that would cover all plants or all animals.
Actual behavior:
Well the actual behavior right now is to do a law for each one of the species
of animals separately it makes it for a novel of a law trust me I did it.
An example of this is:
When attempting to Harvest where (species is Bison)
If in district Hunting District #1
Then allow
Otherwise If in district Hunting District #2
Then allow
Otherwise If in district Hunting District #3
Then allow
Otherwise prevent
Ok so basically what this law does is if the species is a Bison and you are in one
of the three hunting districts which are already setup on this server then
allow the hunting to occur but if you are outside of these three hunting zones
prevent the player from killing the bison.
Now I had to specify each species separately because there
was no way to say "all animals" in the law
So you end up doing a law for Bison, Elk, Wolves, Turkey, Fox, Tortoise, Hare
all separately the same way as above all in one big long rule.
It would really be easier if we could specify either "all animals" or "all plants"
and would shorten the law a lot .
Make it so that all animals or plants are included so that you can just
do this once instead of having to repeat it over and over again.
An example of what the law may look like may be:
For an animal:
When attempting to Harvest where (species is Animal)
If in district Hunting District #1
Then allow
Otherwise If in district Hunting District #2
Then allow
Otherwise If in district Hunting District #3
Then allow
Otherwise prevent
And one for plants:
When attempting to Harvest where (species is Plant)
If in district Gathering District #1
Then allow
Otherwise If in district Gathering District #2
Then allow
Otherwise If in district Gathering District #3
Then allow
Otherwise prevent
It would make things so much easier and simpler can you please
fix the laws this would be a great fix if you can do this because it will
shorten down the laws a lot when writing laws like this!
My law for animals is massive right now so big I couldn't even get the whole thing
in the screenshot. you only see a couple of the laws for the animals.

| 1.0 | Law for all plants and all animals - Version: 0.7.6.3 beta
Steps to Reproduce:
Login to the game and hit L to open up the law window
Try and specify a law for all animals or all plants
Expected behavior:
To write a law that would cover all plants or all animals.
Actual behavior:
Well the actual behavior right now is to do a law for each one of the species
of animals separately it makes it for a novel of a law trust me I did it.
An example of this is:
When attempting to Harvest where (species is Bison)
If in district Hunting District #1
Then allow
Otherwise If in district Hunting District #2
Then allow
Otherwise If in district Hunting District #3
Then allow
Otherwise prevent
Ok so basically what this law does is if the species is a Bison and you are in one
of the three hunting districts which are already setup on this server then
allow the hunting to occur but if you are outside of these three hunting zones
prevent the player from killing the bison.
Now I had to specify each species separately because there
was no way to say "all animals" in the law
So you end up doing a law for Bison, Elk, Wolves, Turkey, Fox, Tortoise, Hare
all separately the same way as above all in one big long rule.
It would really be easier if we could specify either "all animals" or "all plants"
and would shorten the law a lot .
Make it so that all animals or plants are included so that you can just
do this once instead of having to repeat it over and over again.
An example of what the law may look like may be:
For an animal:
When attempting to Harvest where (species is Animal)
If in district Hunting District #1
Then allow
Otherwise If in district Hunting District #2
Then allow
Otherwise If in district Hunting District #3
Then allow
Otherwise prevent
And one for plants:
When attempting to Harvest where (species is Plant)
If in district Gathering District #1
Then allow
Otherwise If in district Gathering District #2
Then allow
Otherwise If in district Gathering District #3
Then allow
Otherwise prevent
It would make things so much easier and simpler can you please
fix the laws this would be a great fix if you can do this because it will
shorten down the laws a lot when writing laws like this!
My law for animals is massive right now so big I couldn't even get the whole thing
in the screenshot. you only see a couple of the laws for the animals.

| priority | law for all plants and all animals version beta steps to reproduce login to the game and hit l to open up the law window try and specify a law for all animals or all plants expected behavior to write a law that would cover all plants or all animals actual behavior well the actual behavior right now is to do a law for each one of the species of animals separately it makes it for a novel of a law trust me i did it an example of this is when attempting to harvest where species is bison if in district hunting district then allow otherwise if in district hunting district then allow otherwise if in district hunting district then allow otherwise prevent ok so basically what this law does is if the species is a bison and you are in one of the three hunting districts which are already setup on this server then allow the hunting to occur but if you are outside of these three hunting zones prevent the player from killing the bison now i had to specify each species separately because there was no way to say all animals in the law so you end up doing a law for bison elk wolves turkey fox tortoise hare all separately the same way as above all in one big long rule it would really be easier if we could specify either all animals or all plants and would shorten the law a lot make it so that all animals or plants are included so that you can just do this once instead of having to repeat it over and over again an example of what the law may look like may be for an animal when attempting to harvest where species is animal if in district hunting district then allow otherwise if in district hunting district then allow otherwise if in district hunting district then allow otherwise prevent and one for plants when attempting to harvest where species is plant if in district gathering district then allow otherwise if in district gathering district then allow otherwise if in district gathering district then allow otherwise prevent it would make things so much easier and simpler can you please fix the laws this would be a great fix if you can do this because it will shorten down the laws a lot when writing laws like this my law for animals is massive right now so big i couldn t even get the whole thing in the screenshot you only see a couple of the laws for the animals | 1 |
540,354 | 15,806,556,888 | IssuesEvent | 2021-04-04 05:56:25 | AY2021S2-CS2103T-W14-1/tp | https://api.github.com/repos/AY2021S2-CS2103T-W14-1/tp | closed | [PE-D] Adding images that are not square do not crop properly in the UI | priority.Medium severity.Low type.Enhancement | ### What is the issue:
Adding an image that is wider than it is longer (i.e rectangular, landscape) does not crop to the center
### How to replicate:
1. Start FriendDex
2. Add an image for any user
3. Observe that the cropping is on the extreme left of the image rather than the centre
### Expected behaviour
Images should crop to the centre
### Why this is an issue:
The subject of most profile photos would be somewhere in the center, and not towards the sides
### Screenshot:

### Images used:


<!--session: 1617429934695-f5d57f1e-5986-44e2-ab31-d20e9d0aba5c-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: DrWala/ped#3 | 1.0 | [PE-D] Adding images that are not square do not crop properly in the UI - ### What is the issue:
Adding an image that is wider than it is longer (i.e rectangular, landscape) does not crop to the center
### How to replicate:
1. Start FriendDex
2. Add an image for any user
3. Observe that the cropping is on the extreme left of the image rather than the centre
### Expected behaviour
Images should crop to the centre
### Why this is an issue:
The subject of most profile photos would be somewhere in the center, and not towards the sides
### Screenshot:

### Images used:


<!--session: 1617429934695-f5d57f1e-5986-44e2-ab31-d20e9d0aba5c-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: DrWala/ped#3 | priority | adding images that are not square do not crop properly in the ui what is the issue adding an image that is wider than it is longer i e rectangular landscape does not crop to the center how to replicate start frienddex add an image for any user observe that the cropping is on the extreme left of the image rather than the centre expected behaviour images should crop to the centre why this is an issue the subject of most profile photos would be somewhere in the center and not towards the sides screenshot images used labels severity medium type functionalitybug original drwala ped | 1 |
136,212 | 5,276,648,576 | IssuesEvent | 2017-02-07 00:04:17 | rm-code/On-The-Roadside | https://api.github.com/repos/rm-code/On-The-Roadside | closed | Movement plotting in two steps | Priority: Medium Status: Accepted Type: Feature | This is how it used to work in earlier versions already:
1. Click on tile to plot a path
2. Click on the same tile to confirm
This ties into #108. | 1.0 | Movement plotting in two steps - This is how it used to work in earlier versions already:
1. Click on tile to plot a path
2. Click on the same tile to confirm
This ties into #108. | priority | movement plotting in two steps this is how it used to work in earlier versions already click on tile to plot a path click on the same tile to confirm this ties into | 1 |
538,308 | 15,766,616,712 | IssuesEvent | 2021-03-31 15:14:31 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | opened | Gotify is sending one single message with all notifications from watchtower | Priority: Medium Status: Available Type: Bug | **Describe the bug**
Before updating from v1.1.5 to v.1.2.1 watchtower was sending each notification to gotify separately, so I was receiving a message from gotify for each message from watchtower. See below an example

but now the notifications coming in just one message:

**Expected behavior**
I am expecting to see one message per one notification. It is hard to read which docker has been updated when all notifications are in one message in gotify
**Environment**
- Kernel 4.14.0.x
- Linux x86_64
- Docker version 19.03.8 (API: 1.40)
| 1.0 | Gotify is sending one single message with all notifications from watchtower - **Describe the bug**
Before updating from v1.1.5 to v.1.2.1 watchtower was sending each notification to gotify separately, so I was receiving a message from gotify for each message from watchtower. See below an example

but now the notifications coming in just one message:

**Expected behavior**
I am expecting to see one message per one notification. It is hard to read which docker has been updated when all notifications are in one message in gotify
**Environment**
- Kernel 4.14.0.x
- Linux x86_64
- Docker version 19.03.8 (API: 1.40)
| priority | gotify is sending one single message with all notifications from watchtower describe the bug before updating from to v watchtower was sending each notification to gotify separately so i was receiving a message from gotify for each message from watchtower see below an example but now the notifications coming in just one message expected behavior i am expecting to see one message per one notification it is hard to read which docker has been updated when all notifications are in one message in gotify environment kernel x linux docker version api | 1 |
567,174 | 16,849,525,214 | IssuesEvent | 2021-06-20 07:57:49 | buttercup/buttercup-desktop | https://api.github.com/repos/buttercup/buttercup-desktop | opened | Better vault avatars | Effort: Medium Priority: Medium Status: Available Type: Enhancement | The current "vault type" avatars are a bit boring and don't do well to indicate different vaults that a user might have added.
Something like [Boring Avatars](https://boringavatars.com/) would be great to use for this, but we'd need some reproducible hash produced in the core library first. | 1.0 | Better vault avatars - The current "vault type" avatars are a bit boring and don't do well to indicate different vaults that a user might have added.
Something like [Boring Avatars](https://boringavatars.com/) would be great to use for this, but we'd need some reproducible hash produced in the core library first. | priority | better vault avatars the current vault type avatars are a bit boring and don t do well to indicate different vaults that a user might have added something like would be great to use for this but we d need some reproducible hash produced in the core library first | 1 |
449,518 | 12,970,013,877 | IssuesEvent | 2020-07-21 08:41:30 | enso-org/ide | https://api.github.com/repos/enso-org/ide | closed | Handling Timeouts in RPC Calls | Category: IDE Change: Non-Breaking Difficulty: Core Contributor Priority: Medium | ### Summary
All our clients for Enso Engine Services should properly handle timeouts for remote calls.
### Value
IDE can properly react when due to connectivity issue or remote issue the expected answer does not income.
### Specification
Each RPC method in our clients should behave that the returned Future will complete with a timeout error after elapsing some defined timeout period. I'd expect this to be implemented as part of json-rpc crate `Handler` type.
Likely the implementation can be just wrapping the current future into `select` with async `sleep` for a defined time.
The timeout should be settable on the macro-generated clients.
### Acceptance Criteria & Test Cases
Code should be accompanied by the test cases.
| 1.0 | Handling Timeouts in RPC Calls - ### Summary
All our clients for Enso Engine Services should properly handle timeouts for remote calls.
### Value
IDE can properly react when due to connectivity issue or remote issue the expected answer does not income.
### Specification
Each RPC method in our clients should behave that the returned Future will complete with a timeout error after elapsing some defined timeout period. I'd expect this to be implemented as part of json-rpc crate `Handler` type.
Likely the implementation can be just wrapping the current future into `select` with async `sleep` for a defined time.
The timeout should be settable on the macro-generated clients.
### Acceptance Criteria & Test Cases
Code should be accompanied by the test cases.
| priority | handling timeouts in rpc calls summary all our clients for enso engine services should properly handle timeouts for remote calls value ide can properly react when due to connectivity issue or remote issue the expected answer does not income specification each rpc method in our clients should behave that the returned future will complete with a timeout error after elapsing some defined timeout period i d expect this to be implemented as part of json rpc crate handler type likely the implementation can be just wrapping the current future into select with async sleep for a defined time the timeout should be settable on the macro generated clients acceptance criteria test cases code should be accompanied by the test cases | 1 |
343,724 | 10,335,405,746 | IssuesEvent | 2019-09-03 10:30:05 | robotframework/robotframework | https://api.github.com/repos/robotframework/robotframework | closed | Support any file extension when explicitly running file and when using `--extension` | alpha 1 enhancement priority: medium | RF 3.2 parses only `.robot` files (see #2820 for discussion) by default and requires using `--extension` if test data files use some other extension. Earlier the valid extensions were limited to `.robot`, `.txt`, `.tsv`, `.html` (not supported at all anymore), `.rest` and `.rst`, but we don't need to have that restriction anymore. We can simply parse all files that match whatever is configured with `--extension` and only `.robot` files if that options isn't used at all.
When execution a simple file, we can parse it regardless the extension. Just need to make sure we use reST parser when execution `.rest` or `.rst` files. When executing a directory, `--extension` is needed. | 1.0 | Support any file extension when explicitly running file and when using `--extension` - RF 3.2 parses only `.robot` files (see #2820 for discussion) by default and requires using `--extension` if test data files use some other extension. Earlier the valid extensions were limited to `.robot`, `.txt`, `.tsv`, `.html` (not supported at all anymore), `.rest` and `.rst`, but we don't need to have that restriction anymore. We can simply parse all files that match whatever is configured with `--extension` and only `.robot` files if that options isn't used at all.
When execution a simple file, we can parse it regardless the extension. Just need to make sure we use reST parser when execution `.rest` or `.rst` files. When executing a directory, `--extension` is needed. | priority | support any file extension when explicitly running file and when using extension rf parses only robot files see for discussion by default and requires using extension if test data files use some other extension earlier the valid extensions were limited to robot txt tsv html not supported at all anymore rest and rst but we don t need to have that restriction anymore we can simply parse all files that match whatever is configured with extension and only robot files if that options isn t used at all when execution a simple file we can parse it regardless the extension just need to make sure we use rest parser when execution rest or rst files when executing a directory extension is needed | 1 |
797,022 | 28,135,455,544 | IssuesEvent | 2023-04-01 10:21:29 | bounswe/bounswe2023group5 | https://api.github.com/repos/bounswe/bounswe2023group5 | closed | Adding Notes of Meeting #5 | Priority: Medium Type: Discussion Status: Assigned | ### Description
I was the note taker of the meeting 5 (30.03.2023). I will upload the notes of the meeting to the wiki page and clarify what we decided and did. [Meeting note template](https://github.com/bounswe/bounswe2023group5/wiki/Meeting-Note-Template) will be used to be consistent.
### 👮♀️ Reviewer
Bilal Atım
### ⏰ Deadline
03.04.2023-Monday | 1.0 | Adding Notes of Meeting #5 - ### Description
I was the note taker of the meeting 5 (30.03.2023). I will upload the notes of the meeting to the wiki page and clarify what we decided and did. [Meeting note template](https://github.com/bounswe/bounswe2023group5/wiki/Meeting-Note-Template) will be used to be consistent.
### 👮♀️ Reviewer
Bilal Atım
### ⏰ Deadline
03.04.2023-Monday | priority | adding notes of meeting description i was the note taker of the meeting i will upload the notes of the meeting to the wiki page and clarify what we decided and did will be used to be consistent 👮♀️ reviewer bilal atım ⏰ deadline monday | 1 |
691,471 | 23,697,747,679 | IssuesEvent | 2022-08-29 16:01:32 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Upgrade react-intl | enhancement priority: medium validate | ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
There are several of improvements on the latest versions of `react-intl` both in functionality and maintainability.
### Describe the solution you'd like
- Upgrade [`react-intl`](https://formatjs.io/docs/react-intl), and the translation maintenance scripts.
- Remove `react-intl-translations-manager` | 1.0 | [studio-ui] Upgrade react-intl - ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
There are several of improvements on the latest versions of `react-intl` both in functionality and maintainability.
### Describe the solution you'd like
- Upgrade [`react-intl`](https://formatjs.io/docs/react-intl), and the translation maintenance scripts.
- Remove `react-intl-translations-manager` | priority | upgrade react intl duplicates i have searched the existing issues is your feature request related to a problem please describe there are several of improvements on the latest versions of react intl both in functionality and maintainability describe the solution you d like upgrade and the translation maintenance scripts remove react intl translations manager | 1 |
322,410 | 9,817,360,792 | IssuesEvent | 2019-06-13 16:32:07 | milnel2/blocks4alliOS | https://api.github.com/repos/milnel2/blocks4alliOS | closed | delete vars in drop-down menu code | delete easy medium priority | In block class (around lines 21-24), check and see if the vars for the old drop-down menu are used through BlocksMenu.plist and delete the ones not used
vars include:
editable, options, optionsLabels, and pickedOption | 1.0 | delete vars in drop-down menu code - In block class (around lines 21-24), check and see if the vars for the old drop-down menu are used through BlocksMenu.plist and delete the ones not used
vars include:
editable, options, optionsLabels, and pickedOption | priority | delete vars in drop down menu code in block class around lines check and see if the vars for the old drop down menu are used through blocksmenu plist and delete the ones not used vars include editable options optionslabels and pickedoption | 1 |
830,003 | 31,933,744,277 | IssuesEvent | 2023-09-19 09:07:36 | pybamm-team/PyBaMM | https://api.github.com/repos/pybamm-team/PyBaMM | closed | Improve memory usage for experiments with `start_time` | difficulty: medium priority:medium in-progress | At the moment, two steps with the same operating conditions but different `start_time` are treated as different. This means that the model needs to be processed twice, which can end up with a big memory consumption (and longer running times) for very long experiments.
A way to improve this would be to read the `start_time` from the unprocessed step, this way we can have a single processed step for all of them. | 1.0 | Improve memory usage for experiments with `start_time` - At the moment, two steps with the same operating conditions but different `start_time` are treated as different. This means that the model needs to be processed twice, which can end up with a big memory consumption (and longer running times) for very long experiments.
A way to improve this would be to read the `start_time` from the unprocessed step, this way we can have a single processed step for all of them. | priority | improve memory usage for experiments with start time at the moment two steps with the same operating conditions but different start time are treated as different this means that the model needs to be processed twice which can end up with a big memory consumption and longer running times for very long experiments a way to improve this would be to read the start time from the unprocessed step this way we can have a single processed step for all of them | 1 |
88,271 | 3,775,680,029 | IssuesEvent | 2016-03-17 14:24:37 | WeAreAthlon/silla.io | https://api.github.com/repos/WeAreAthlon/silla.io | closed | Lack of Password Policy | feature medium priority security | The system does not enforce a strong password policy.
Enforce a password policy, which will include at least the following guidelines:
- [ ] Passwords will expire after no more than 90 days.
- [ ] Passwords will not repeat themselves for 4 generations.
- [ ] Force the users to change their password upon first logon to the system, and after any time it was reset by an administrator.
- [ ] The password length will be at least 8 characters long, and will contain at least three of the following group of characters: * Capital letters. * Lowercase letters * Digits * Special characters.
| 1.0 | Lack of Password Policy - The system does not enforce a strong password policy.
Enforce a password policy, which will include at least the following guidelines:
- [ ] Passwords will expire after no more than 90 days.
- [ ] Passwords will not repeat themselves for 4 generations.
- [ ] Force the users to change their password upon first logon to the system, and after any time it was reset by an administrator.
- [ ] The password length will be at least 8 characters long, and will contain at least three of the following group of characters: * Capital letters. * Lowercase letters * Digits * Special characters.
| priority | lack of password policy the system does not enforce a strong password policy enforce a password policy which will include at least the following guidelines passwords will expire after no more than days passwords will not repeat themselves for generations force the users to change their password upon first logon to the system and after any time it was reset by an administrator the password length will be at least characters long and will contain at least three of the following group of characters capital letters lowercase letters digits special characters | 1 |
816,491 | 30,600,469,197 | IssuesEvent | 2023-07-22 10:04:36 | Joemwa/Coursework-Planner | https://api.github.com/repos/Joemwa/Coursework-Planner | opened | [TECH ED] Mailing list API | 🏕 Priority Mandatory 🐂 Size Medium 📅 Node 🔑 Priority Key | From Module-Node created by [Dedekind561](https://github.com/Dedekind561): CodeYourFuture/Module-Node#24
### Link to the coursework
https://github.com/CodeYourFuture/Module-Node/blob/main/mailing-list-api/README.md
### Why are we doing this?
In this challenge you will take all your knowledge of APIs and backend thus far to build an API that can process a full range of different HTTP requests.
### Maximum time in hours
12
### How to get help
Share your blockers in your class channel
https://syllabus.codeyourfuture.io/guides/asking-questions
### How to submit
1. Fork to your Github account.
2. Make a branch for this project.
3. Make regular small commits in this branch with clear messages.
4. When you are ready, open a PR to the CYF repo, following the instructions in the PR template.
```mermaid
gitGraph
commit id: "start"
branch feature/mailing-list-api
commit id: "skeleton page code"
commit id: "Fonts and colours"
commit id: "mobile layout"
commit id: "lighthouse audit revisions mobile"
commit id: "desktop layout"
commit id: "lighthouse audit revisions desktop"
checkout main
merge feature/mailing-list-api
``` | 2.0 | [TECH ED] Mailing list API - From Module-Node created by [Dedekind561](https://github.com/Dedekind561): CodeYourFuture/Module-Node#24
### Link to the coursework
https://github.com/CodeYourFuture/Module-Node/blob/main/mailing-list-api/README.md
### Why are we doing this?
In this challenge you will take all your knowledge of APIs and backend thus far to build an API that can process a full range of different HTTP requests.
### Maximum time in hours
12
### How to get help
Share your blockers in your class channel
https://syllabus.codeyourfuture.io/guides/asking-questions
### How to submit
1. Fork to your Github account.
2. Make a branch for this project.
3. Make regular small commits in this branch with clear messages.
4. When you are ready, open a PR to the CYF repo, following the instructions in the PR template.
```mermaid
gitGraph
commit id: "start"
branch feature/mailing-list-api
commit id: "skeleton page code"
commit id: "Fonts and colours"
commit id: "mobile layout"
commit id: "lighthouse audit revisions mobile"
commit id: "desktop layout"
commit id: "lighthouse audit revisions desktop"
checkout main
merge feature/mailing-list-api
``` | priority | mailing list api from module node created by codeyourfuture module node link to the coursework why are we doing this in this challenge you will take all your knowledge of apis and backend thus far to build an api that can process a full range of different http requests maximum time in hours how to get help share your blockers in your class channel how to submit fork to your github account make a branch for this project make regular small commits in this branch with clear messages when you are ready open a pr to the cyf repo following the instructions in the pr template mermaid gitgraph commit id start branch feature mailing list api commit id skeleton page code commit id fonts and colours commit id mobile layout commit id lighthouse audit revisions mobile commit id desktop layout commit id lighthouse audit revisions desktop checkout main merge feature mailing list api | 1 |
121,260 | 4,807,118,760 | IssuesEvent | 2016-11-02 20:30:55 | PovertyAction/high-frequency-checks | https://api.github.com/repos/PovertyAction/high-frequency-checks | opened | Have template run some checks on daily data and others on full data? And/or two diff outputs? | enhancement Medium Priority | The output on some of the checks can get unwieldy. Some teams have split up their data so that certain checks just run on that day's data while others run on the full running dataset. Another idea is to run the template twice with two different outputs: one daily output; one running output.
Should brainstorm if any modifications of these make sense. One concern with doing some (like specify or outliers) on the daily data is losing the ability to see patterns (e.g. an "other" response that comes up a lot and should get coded into the survey as an answer option).
A column that allows the managing RA to mark which observations have been checked ( #24 ) could help. So all flagged observations would show, but easier to quickly scan to the ones still needing verification. | 1.0 | Have template run some checks on daily data and others on full data? And/or two diff outputs? - The output on some of the checks can get unwieldy. Some teams have split up their data so that certain checks just run on that day's data while others run on the full running dataset. Another idea is to run the template twice with two different outputs: one daily output; one running output.
Should brainstorm if any modifications of these make sense. One concern with doing some (like specify or outliers) on the daily data is losing the ability to see patterns (e.g. an "other" response that comes up a lot and should get coded into the survey as an answer option).
A column that allows the managing RA to mark which observations have been checked ( #24 ) could help. So all flagged observations would show, but easier to quickly scan to the ones still needing verification. | priority | have template run some checks on daily data and others on full data and or two diff outputs the output on some of the checks can get unwieldy some teams have split up their data so that certain checks just run on that day s data while others run on the full running dataset another idea is to run the template twice with two different outputs one daily output one running output should brainstorm if any modifications of these make sense one concern with doing some like specify or outliers on the daily data is losing the ability to see patterns e g an other response that comes up a lot and should get coded into the survey as an answer option a column that allows the managing ra to mark which observations have been checked could help so all flagged observations would show but easier to quickly scan to the ones still needing verification | 1 |
77,470 | 3,506,392,364 | IssuesEvent | 2016-01-08 06:25:05 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Server crashes (BB #515) | duplicate migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** danqu
**Original Date:** 01.03.2014 10:25:41 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** duplicate
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/515
<hr>
When I was in the BlackTemple called Ge tower Gurtogg Bloodboil BOSS, server crashes
Revision: OregonCore Rev: Archive Hash: c0cf03df5 (Win32,little-endian)
Date 1:3:2014. Time 17:33
//=====================================================
*** Hardware ***
Processor: Intel(R) Core(TM) i5-3470 CPU @ 3.20GHz
Number Of Processors: 4
Physical Memory: 4194303 KB (Available: 4194303 KB)
Commit Charge Limit: 4194303 KB
*** Operation System ***
Windows Vista or Windows Server 2008 Server 4.0, Enterprise Edition (Version 6.1, Build 7600)
//=====================================================
Exception code: C0000005 ACCESS_VIOLATION
Fault address: 014A9A53 01:000D8A53 D:\243\oregon-core.exe
Registers:
EAX:0C6C6A90
EBX:0000008C
ECX:00000000
EDX:00000000
ESI:057E8078
EDI:0C6C6A90
CS:EIP:0023:014A9A53
SS:ESP:002B:0E0CF808 EBP:0E0CF808
DS:002B ES:002B FS:0053 GS:002B
Flags:00010216
Call stack:
Address Frame Function SourceFile
014A9A53 00000000 0001:000D8A53 D:\243\oregon-core.exe
0164FE68 00000000 0001:0027EE68 D:\243\oregon-core.exe
014D4957 00000000 0001:00103957 D:\243\oregon-core.exe
0149B5CD 00000000 0001:000CA5CD D:\243\oregon-core.exe
0149F556 00000000 0001:000CE556 D:\243\oregon-core.exe
014A0932 00000000 0001:000CF932 D:\243\oregon-core.exe
014A22FE 00000000 0001:000D12FE D:\243\oregon-core.exe
014A2890 00000000 0001:000D1890 D:\243\oregon-core.exe
0159D176 00000000 0001:001CC176 D:\243\oregon-core.exe
0159CEA1 00000000 0001:001CBEA1 D:\243\oregon-core.exe
6FA8AC5E 00000000 ?svc_run@ACE_Task_Base@@SAKPAX@Z+2E
6FA8B03B 00000000 ?invoke_i@ACE_Thread_Adapter@@EAEKXZ+6B
6FA8B183 00000000 ?invoke@ACE_Thread_Adapter@@UAEKXZ+83
71C9C6DE 00000000 _endthreadex+3A
71C9C788 00000000 _endthreadex+E4
75A33677 00000000 BaseThreadInitThunk+12
77959D72 00000000 RtlInitializeExceptionChain+63
77959D45 00000000 RtlInitializeExceptionChain+36
========================
Local Variables And Parameters
Call stack:
Address Frame Function SourceFile
014A9A53 00000000 0001:000D8A53 D:\243\oregon-core.exe
0164FE68 00000000 0001:0027EE68 D:\243\oregon-core.exe
014D4957 00000000 0001:00103957 D:\243\oregon-core.exe
0149B5CD 00000000 0001:000CA5CD D:\243\oregon-core.exe
0149F556 00000000 0001:000CE556 D:\243\oregon-core.exe
014A0932 00000000 0001:000CF932 D:\243\oregon-core.exe
014A22FE 00000000 0001:000D12FE D:\243\oregon-core.exe
014A2890 00000000 0001:000D1890 D:\243\oregon-core.exe
0159D176 00000000 0001:001CC176 D:\243\oregon-core.exe
0159CEA1 00000000 0001:001CBEA1 D:\243\oregon-core.exe
6FA8AC5E 00000000 ?svc_run@ACE_Task_Base@@SAKPAX@Z+2E
6FA8B03B 00000000 ?invoke_i@ACE_Thread_Adapter@@EAEKXZ+6B
6FA8B183 00000000 ?invoke@ACE_Thread_Adapter@@UAEKXZ+83
71C9C6DE 00000000 _endthreadex+3A
71C9C788 00000000 _endthreadex+E4
75A33677 00000000 BaseThreadInitThunk+12
77959D72 00000000 RtlInitializeExceptionChain+63
77959D45 00000000 RtlInitializeExceptionChain+36
========================
Global Variables
<user defined> '??4_Init_locks@std@@QAEAAV01@ABV01@@Z'
<user defined> '??0_Mutex@std@@QAE@W4_Uninitialized@1@@Z'
| 1.0 | Server crashes (BB #515) - This issue was migrated from bitbucket.
**Original Reporter:** danqu
**Original Date:** 01.03.2014 10:25:41 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** duplicate
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/515
<hr>
When I was in the BlackTemple called Ge tower Gurtogg Bloodboil BOSS, server crashes
Revision: OregonCore Rev: Archive Hash: c0cf03df5 (Win32,little-endian)
Date 1:3:2014. Time 17:33
//=====================================================
*** Hardware ***
Processor: Intel(R) Core(TM) i5-3470 CPU @ 3.20GHz
Number Of Processors: 4
Physical Memory: 4194303 KB (Available: 4194303 KB)
Commit Charge Limit: 4194303 KB
*** Operation System ***
Windows Vista or Windows Server 2008 Server 4.0, Enterprise Edition (Version 6.1, Build 7600)
//=====================================================
Exception code: C0000005 ACCESS_VIOLATION
Fault address: 014A9A53 01:000D8A53 D:\243\oregon-core.exe
Registers:
EAX:0C6C6A90
EBX:0000008C
ECX:00000000
EDX:00000000
ESI:057E8078
EDI:0C6C6A90
CS:EIP:0023:014A9A53
SS:ESP:002B:0E0CF808 EBP:0E0CF808
DS:002B ES:002B FS:0053 GS:002B
Flags:00010216
Call stack:
Address Frame Function SourceFile
014A9A53 00000000 0001:000D8A53 D:\243\oregon-core.exe
0164FE68 00000000 0001:0027EE68 D:\243\oregon-core.exe
014D4957 00000000 0001:00103957 D:\243\oregon-core.exe
0149B5CD 00000000 0001:000CA5CD D:\243\oregon-core.exe
0149F556 00000000 0001:000CE556 D:\243\oregon-core.exe
014A0932 00000000 0001:000CF932 D:\243\oregon-core.exe
014A22FE 00000000 0001:000D12FE D:\243\oregon-core.exe
014A2890 00000000 0001:000D1890 D:\243\oregon-core.exe
0159D176 00000000 0001:001CC176 D:\243\oregon-core.exe
0159CEA1 00000000 0001:001CBEA1 D:\243\oregon-core.exe
6FA8AC5E 00000000 ?svc_run@ACE_Task_Base@@SAKPAX@Z+2E
6FA8B03B 00000000 ?invoke_i@ACE_Thread_Adapter@@EAEKXZ+6B
6FA8B183 00000000 ?invoke@ACE_Thread_Adapter@@UAEKXZ+83
71C9C6DE 00000000 _endthreadex+3A
71C9C788 00000000 _endthreadex+E4
75A33677 00000000 BaseThreadInitThunk+12
77959D72 00000000 RtlInitializeExceptionChain+63
77959D45 00000000 RtlInitializeExceptionChain+36
========================
Local Variables And Parameters
Call stack:
Address Frame Function SourceFile
014A9A53 00000000 0001:000D8A53 D:\243\oregon-core.exe
0164FE68 00000000 0001:0027EE68 D:\243\oregon-core.exe
014D4957 00000000 0001:00103957 D:\243\oregon-core.exe
0149B5CD 00000000 0001:000CA5CD D:\243\oregon-core.exe
0149F556 00000000 0001:000CE556 D:\243\oregon-core.exe
014A0932 00000000 0001:000CF932 D:\243\oregon-core.exe
014A22FE 00000000 0001:000D12FE D:\243\oregon-core.exe
014A2890 00000000 0001:000D1890 D:\243\oregon-core.exe
0159D176 00000000 0001:001CC176 D:\243\oregon-core.exe
0159CEA1 00000000 0001:001CBEA1 D:\243\oregon-core.exe
6FA8AC5E 00000000 ?svc_run@ACE_Task_Base@@SAKPAX@Z+2E
6FA8B03B 00000000 ?invoke_i@ACE_Thread_Adapter@@EAEKXZ+6B
6FA8B183 00000000 ?invoke@ACE_Thread_Adapter@@UAEKXZ+83
71C9C6DE 00000000 _endthreadex+3A
71C9C788 00000000 _endthreadex+E4
75A33677 00000000 BaseThreadInitThunk+12
77959D72 00000000 RtlInitializeExceptionChain+63
77959D45 00000000 RtlInitializeExceptionChain+36
========================
Global Variables
<user defined> '??4_Init_locks@std@@QAEAAV01@ABV01@@Z'
<user defined> '??0_Mutex@std@@QAE@W4_Uninitialized@1@@Z'
| priority | server crashes bb this issue was migrated from bitbucket original reporter danqu original date gmt original priority major original type bug original state duplicate direct link when i was in the blacktemple called ge tower gurtogg bloodboil boss server crashes revision oregoncore rev archive hash little endian date time hardware processor intel r core tm cpu number of processors physical memory kb available kb commit charge limit kb operation system windows vista or windows server server enterprise edition version build exception code access violation fault address d oregon core exe registers eax ebx ecx edx esi edi cs eip ss esp ebp ds es fs gs flags call stack address frame function sourcefile d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe svc run ace task base sakpax z invoke i ace thread adapter eaekxz invoke ace thread adapter uaekxz endthreadex endthreadex basethreadinitthunk rtlinitializeexceptionchain rtlinitializeexceptionchain local variables and parameters call stack address frame function sourcefile d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe d oregon core exe svc run ace task base sakpax z invoke i ace thread adapter eaekxz invoke ace thread adapter uaekxz endthreadex endthreadex basethreadinitthunk rtlinitializeexceptionchain rtlinitializeexceptionchain global variables init locks std z mutex std qae uninitialized z | 1 |
747,422 | 26,083,445,675 | IssuesEvent | 2022-12-25 18:53:58 | bounswe/bounswe2022group7 | https://api.github.com/repos/bounswe/bounswe2022group7 | closed | [Mobile] Reporting ArtItem | Status: Pending Review Priority: Medium Difficulty: Medium Type: Implementation Target: Mobile | I will implement the reporting copyright infringing art item feature to the mobile app.
- [x] new branch
- [x] implement the end point connection
- [x] create a ui as a pop up
- [x] connect end point to pop up
- [x] create PR
Due Date: 25.12.2022
Reviewer: @canatakan | 1.0 | [Mobile] Reporting ArtItem - I will implement the reporting copyright infringing art item feature to the mobile app.
- [x] new branch
- [x] implement the end point connection
- [x] create a ui as a pop up
- [x] connect end point to pop up
- [x] create PR
Due Date: 25.12.2022
Reviewer: @canatakan | priority | reporting artitem i will implement the reporting copyright infringing art item feature to the mobile app new branch implement the end point connection create a ui as a pop up connect end point to pop up create pr due date reviewer canatakan | 1 |
676,203 | 23,119,276,381 | IssuesEvent | 2022-07-27 19:39:08 | codbex/codbex-kronos | https://api.github.com/repos/codbex/codbex-kronos | opened | Copy HDB artifacts behavior | enhancement priority-medium IDE investigation / discussion | From xsk created by [dpanayotov](https://github.com/dpanayotov): SAP/xsk#640
Same as #639 but for copy file functionality
**Describe the solution you'd like**
Calculate changes and document behavior when copying HDB artifacts in another package/project.
**Describe alternatives you've considered**
In XSK, copying HDB artifacts may be disabled as this may result in errors - probably db conflict or referencing a missing artifact.
| 1.0 | Copy HDB artifacts behavior - From xsk created by [dpanayotov](https://github.com/dpanayotov): SAP/xsk#640
Same as #639 but for copy file functionality
**Describe the solution you'd like**
Calculate changes and document behavior when copying HDB artifacts in another package/project.
**Describe alternatives you've considered**
In XSK, copying HDB artifacts may be disabled as this may result in errors - probably db conflict or referencing a missing artifact.
| priority | copy hdb artifacts behavior from xsk created by sap xsk same as but for copy file functionality describe the solution you d like calculate changes and document behavior when copying hdb artifacts in another package project describe alternatives you ve considered in xsk copying hdb artifacts may be disabled as this may result in errors probably db conflict or referencing a missing artifact | 1 |
787,552 | 27,722,172,226 | IssuesEvent | 2023-03-14 21:38:41 | docker-mailserver/docker-mailserver | https://api.github.com/repos/docker-mailserver/docker-mailserver | opened | [BUG] | kind/bug meta/needs triage priority/medium | ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
fail2ban
### What happened and when does this occur?
As mentioned in https://github.com/docker-mailserver/docker-mailserver/issues/1810, DMS in its default configuration keeps being constantly hammered by zombie servers, script kiddies and other bad actors. These actions make up for the bulk of DMS request-response cycles on my server. Having postscreen in play and not using it for anything else than DNSBL is a security vulnerability, IMO. These kinds of attacks should clearly be stopped, as all necessary components are in place (postscreen/fail2ban).
Log samples with a few attack types:
```
Mar 14 21:58:05 mail postfix/postscreen[12729]: CONNECT from [141.98.11.52]:60710 to [redacted]:25
Mar 14 21:58:05 mail postfix/postscreen[12729]: PASS OLD [141.98.11.52]:60710
...
Mar 14 21:58:05 mail postfix/smtpd[12730]: connect from unknown[141.98.11.52]
Mar 14 21:58:06 mail dovecot: auth: passwd-file(info,141.98.11.52): unknown user (SHA1 of given password: 34dec8)
Mar 14 21:58:08 mail postfix/smtpd[12730]: warning: unknown[141.98.11.52]: SASL LOGIN authentication failed: UGFzc3dvcmQ6
...
Mar 14 21:40:06 mail postfix/postscreen[7478]: PREGREET 11 after 0.02 from [79.110.62.240]:54115: EHLO User\r\n
Mar 14 20:50:32 mail postfix/postscreen[32395]: PREGREET 11 after 0.02 from [79.110.62.240]:57517: EHLO User\r\n
Mar 14 20:22:16 mail postfix/postscreen[27498]: PREGREET 1 after 0 from [43.158.217.180]:53256: \002
Mar 14 20:22:16 mail postfix/postscreen[27498]: PREGREET 611 after 0 from [43.158.217.180]:53112: o\372\300\276\270\300\244\311\211\242\302\217\203\257\221\227\276\315\271\317\254\233\260\253\240\26
Mar 14 20:22:15 mail postfix/postscreen[27498]: PREGREET 1023 after 0 from [43.158.217.180]:52998: \377\242\377\000\377=\377\255\377\000\377\321\377\330\361\255\377\276H\373_5\377\255\361\rl\345\372p
```
Failures also often include broken/malformed TLS communication attempts which also could be considered a security problem:
```
Mar 14 21:58:55 mail postfix/smtps/smtpd[13069]: warning: TLS library problem: error:1408F10B:SSL routines:ssl3_get_record:wrong version number:../ssl/record/ssl3_record.c:331: |
Mar 14 21:58:55 mail postfix/smtps/smtpd[13069]: SSL_accept error from unknown[194.165.16.78]: -1 |
```
These attacks can use arbitrary intervals, therefore checks which take only 10-minute sample window into consideration are completely useless (a small excerpt from today):
<img width="940" alt="image" src="https://user-images.githubusercontent.com/3662386/225138559-2dc4c152-b640-43ba-8693-612ea30b219f.png">
Which is clearly not caught by F2B (using default configuration):
```
root@mail:/etc/fail2ban# fail2ban-client status dovecot
Status for the jail: dovecot
|- Filter
| |- Currently failed: 33
| |- Total failed: 1233
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 1
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status postfix
Status for the jail: postfix
|- Filter
| |- Currently failed: 3
| |- Total failed: 3
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 0
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status postfix-sasl
Status for the jail: postfix-sasl
|- Filter
| |- Currently failed: 33
| |- Total failed: 1233
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 1
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status custom
Status for the jail: custom
|- Filter
| |- Currently failed: 0
| |- Total failed: 0
| `- File list:
`- Actions
|- Currently banned: 8
|- Total banned: 8
`- Banned IP list: 107.170.243.25 159.203.224.10 162.142.125.213 167.94.138.62 180.214.239.18 193.42.33.76 93.174.95.106 163.123.143.10
```
While I understand fail2ban is a software package from another maintainer, so is every other part of DMS. We should do what we can to minimize attack vectors. This is a low hanging fruit, as the author of linked issue already provided a solution for this security vulnerability here: https://github.com/docker-mailserver/docker-mailserver/issues/1810#issuecomment-836593632.
### What did you expect to happen?
The following should be expected:
- postscreen should get a proper jail config which would catch these "usual suspects"
- the sampling window of F2B to be of a meaningful value (not 10 minutes)
- typical low-hanging fruit attack vectors to be banned
- especially dangerous attack vectors like user/password brute-forcing (repeated failed auth attempts) to be banned either for for some meaningful unit of time, not an hour, or using incremental ban times
### How do we replicate the issue?
1. Use DMS in its default config and wait for a few hours. You'll get plenty of samples.
### DMS version
v11.3.1
### What operating system is DMS running on?
Linux
### Which operating system version?
4.4.180+ SMP Tue Jan 31 23:19:16 CST 2023 x86_64 GNU/Linux synology_apollolake_918+
### What instruction set architecture is DMS running on?
AMD64 / x86_64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yml
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:11.3.1
container_name: mailserver
env_file: mailserver.env
volumes:
- ./data/mail-data/:/var/mail/
- ./data/mail-state/:/var/mail-state/
- ./data/mail-logs/:/var/log/mail/
- ./data/config/:/tmp/docker-mailserver/
- /etc/localtime:/etc/localtime:ro
- /...redacted.../traefik/config/acme.json:/etc/letsencrypt/acme.json:ro
restart: unless-stopped
stop_grace_period: 1m
cap_add:
- NET_ADMIN
healthcheck:
test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1"
timeout: 3s
retries: 0
network_mode: host
```
### Relevant log output
_No response_
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [X] I am rather experienced with docker
- [ ] I am inexperienced with mail servers
- [X] I am rather experienced with mail servers
- [ ] I am uncomfortable with the CLI
- [X] I am rather comfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | 1.0 | [BUG] - ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
fail2ban
### What happened and when does this occur?
As mentioned in https://github.com/docker-mailserver/docker-mailserver/issues/1810, DMS in its default configuration keeps being constantly hammered by zombie servers, script kiddies and other bad actors. These actions make up for the bulk of DMS request-response cycles on my server. Having postscreen in play and not using it for anything else than DNSBL is a security vulnerability, IMO. These kinds of attacks should clearly be stopped, as all necessary components are in place (postscreen/fail2ban).
Log samples with a few attack types:
```
Mar 14 21:58:05 mail postfix/postscreen[12729]: CONNECT from [141.98.11.52]:60710 to [redacted]:25
Mar 14 21:58:05 mail postfix/postscreen[12729]: PASS OLD [141.98.11.52]:60710
...
Mar 14 21:58:05 mail postfix/smtpd[12730]: connect from unknown[141.98.11.52]
Mar 14 21:58:06 mail dovecot: auth: passwd-file(info,141.98.11.52): unknown user (SHA1 of given password: 34dec8)
Mar 14 21:58:08 mail postfix/smtpd[12730]: warning: unknown[141.98.11.52]: SASL LOGIN authentication failed: UGFzc3dvcmQ6
...
Mar 14 21:40:06 mail postfix/postscreen[7478]: PREGREET 11 after 0.02 from [79.110.62.240]:54115: EHLO User\r\n
Mar 14 20:50:32 mail postfix/postscreen[32395]: PREGREET 11 after 0.02 from [79.110.62.240]:57517: EHLO User\r\n
Mar 14 20:22:16 mail postfix/postscreen[27498]: PREGREET 1 after 0 from [43.158.217.180]:53256: \002
Mar 14 20:22:16 mail postfix/postscreen[27498]: PREGREET 611 after 0 from [43.158.217.180]:53112: o\372\300\276\270\300\244\311\211\242\302\217\203\257\221\227\276\315\271\317\254\233\260\253\240\26
Mar 14 20:22:15 mail postfix/postscreen[27498]: PREGREET 1023 after 0 from [43.158.217.180]:52998: \377\242\377\000\377=\377\255\377\000\377\321\377\330\361\255\377\276H\373_5\377\255\361\rl\345\372p
```
Failures also often include broken/malformed TLS communication attempts which also could be considered a security problem:
```
Mar 14 21:58:55 mail postfix/smtps/smtpd[13069]: warning: TLS library problem: error:1408F10B:SSL routines:ssl3_get_record:wrong version number:../ssl/record/ssl3_record.c:331: |
Mar 14 21:58:55 mail postfix/smtps/smtpd[13069]: SSL_accept error from unknown[194.165.16.78]: -1 |
```
These attacks can use arbitrary intervals, therefore checks which take only 10-minute sample window into consideration are completely useless (a small excerpt from today):
<img width="940" alt="image" src="https://user-images.githubusercontent.com/3662386/225138559-2dc4c152-b640-43ba-8693-612ea30b219f.png">
Which is clearly not caught by F2B (using default configuration):
```
root@mail:/etc/fail2ban# fail2ban-client status dovecot
Status for the jail: dovecot
|- Filter
| |- Currently failed: 33
| |- Total failed: 1233
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 1
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status postfix
Status for the jail: postfix
|- Filter
| |- Currently failed: 3
| |- Total failed: 3
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 0
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status postfix-sasl
Status for the jail: postfix-sasl
|- Filter
| |- Currently failed: 33
| |- Total failed: 1233
| `- File list: /var/log/mail.log
`- Actions
|- Currently banned: 0
|- Total banned: 1
`- Banned IP list:
root@mail:/etc/fail2ban# fail2ban-client status custom
Status for the jail: custom
|- Filter
| |- Currently failed: 0
| |- Total failed: 0
| `- File list:
`- Actions
|- Currently banned: 8
|- Total banned: 8
`- Banned IP list: 107.170.243.25 159.203.224.10 162.142.125.213 167.94.138.62 180.214.239.18 193.42.33.76 93.174.95.106 163.123.143.10
```
While I understand fail2ban is a software package from another maintainer, so is every other part of DMS. We should do what we can to minimize attack vectors. This is a low hanging fruit, as the author of linked issue already provided a solution for this security vulnerability here: https://github.com/docker-mailserver/docker-mailserver/issues/1810#issuecomment-836593632.
### What did you expect to happen?
The following should be expected:
- postscreen should get a proper jail config which would catch these "usual suspects"
- the sampling window of F2B to be of a meaningful value (not 10 minutes)
- typical low-hanging fruit attack vectors to be banned
- especially dangerous attack vectors like user/password brute-forcing (repeated failed auth attempts) to be banned either for for some meaningful unit of time, not an hour, or using incremental ban times
### How do we replicate the issue?
1. Use DMS in its default config and wait for a few hours. You'll get plenty of samples.
### DMS version
v11.3.1
### What operating system is DMS running on?
Linux
### Which operating system version?
4.4.180+ SMP Tue Jan 31 23:19:16 CST 2023 x86_64 GNU/Linux synology_apollolake_918+
### What instruction set architecture is DMS running on?
AMD64 / x86_64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yml
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:11.3.1
container_name: mailserver
env_file: mailserver.env
volumes:
- ./data/mail-data/:/var/mail/
- ./data/mail-state/:/var/mail-state/
- ./data/mail-logs/:/var/log/mail/
- ./data/config/:/tmp/docker-mailserver/
- /etc/localtime:/etc/localtime:ro
- /...redacted.../traefik/config/acme.json:/etc/letsencrypt/acme.json:ro
restart: unless-stopped
stop_grace_period: 1m
cap_add:
- NET_ADMIN
healthcheck:
test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1"
timeout: 3s
retries: 0
network_mode: host
```
### Relevant log output
_No response_
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [X] I am rather experienced with docker
- [ ] I am inexperienced with mail servers
- [X] I am rather experienced with mail servers
- [ ] I am uncomfortable with the CLI
- [X] I am rather comfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | priority | miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s what happened and when does this occur as mentioned in dms in its default configuration keeps being constantly hammered by zombie servers script kiddies and other bad actors these actions make up for the bulk of dms request response cycles on my server having postscreen in play and not using it for anything else than dnsbl is a security vulnerability imo these kinds of attacks should clearly be stopped as all necessary components are in place postscreen log samples with a few attack types mar mail postfix postscreen connect from to mar mail postfix postscreen pass old mar mail postfix smtpd connect from unknown mar mail dovecot auth passwd file info unknown user of given password mar mail postfix smtpd warning unknown sasl login authentication failed mar mail postfix postscreen pregreet after from ehlo user r n mar mail postfix postscreen pregreet after from ehlo user r n mar mail postfix postscreen pregreet after from mar mail postfix postscreen pregreet after from o mar mail postfix postscreen pregreet after from rl failures also often include broken malformed tls communication attempts which also could be considered a security problem mar mail postfix smtps smtpd warning tls library problem error ssl routines get record wrong version number ssl record record c mar mail postfix smtps smtpd ssl accept error from unknown these attacks can use arbitrary intervals therefore checks which take only minute sample window into consideration are completely useless a small excerpt from today img width alt image src which is clearly not caught by using default configuration root mail etc client status dovecot status for the jail dovecot filter currently failed total failed file list var log mail log actions currently banned total banned banned ip list root mail etc client status postfix status for the jail postfix filter currently failed total failed file list var log mail log actions currently banned total banned banned ip list root mail etc client status postfix sasl status for the jail postfix sasl filter currently failed total failed file list var log mail log actions currently banned total banned banned ip list root mail etc client status custom status for the jail custom filter currently failed total failed file list actions currently banned total banned banned ip list while i understand is a software package from another maintainer so is every other part of dms we should do what we can to minimize attack vectors this is a low hanging fruit as the author of linked issue already provided a solution for this security vulnerability here what did you expect to happen the following should be expected postscreen should get a proper jail config which would catch these usual suspects the sampling window of to be of a meaningful value not minutes typical low hanging fruit attack vectors to be banned especially dangerous attack vectors like user password brute forcing repeated failed auth attempts to be banned either for for some meaningful unit of time not an hour or using incremental ban times how do we replicate the issue use dms in its default config and wait for a few hours you ll get plenty of samples dms version what operating system is dms running on linux which operating system version smp tue jan cst gnu linux synology apollolake what instruction set architecture is dms running on what container orchestration tool are you using docker compose docker compose yml yml services mailserver image docker io mailserver docker mailserver container name mailserver env file mailserver env volumes data mail data var mail data mail state var mail state data mail logs var log mail data config tmp docker mailserver etc localtime etc localtime ro redacted traefik config acme json etc letsencrypt acme json ro restart unless stopped stop grace period cap add net admin healthcheck test ss listening tcp grep p listen smtp exit timeout retries network mode host relevant log output no response other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am rather experienced with docker i am inexperienced with mail servers i am rather experienced with mail servers i am uncomfortable with the cli i am rather comfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response | 1 |
371,053 | 10,960,844,599 | IssuesEvent | 2019-11-27 14:20:47 | CN-UPB/tng-sdk-benchmark | https://api.github.com/repos/CN-UPB/tng-sdk-benchmark | closed | Move osmclient outside of tngsdk folder | bug priority: medium | It was a technical mistake to add the osmclient files under the tngsdk folder.
For correctness and to make pytests run correctly as well, we must move the folder outside the tngsdk folder and put it in the same directory tree level as the tng sdk folder. | 1.0 | Move osmclient outside of tngsdk folder - It was a technical mistake to add the osmclient files under the tngsdk folder.
For correctness and to make pytests run correctly as well, we must move the folder outside the tngsdk folder and put it in the same directory tree level as the tng sdk folder. | priority | move osmclient outside of tngsdk folder it was a technical mistake to add the osmclient files under the tngsdk folder for correctness and to make pytests run correctly as well we must move the folder outside the tngsdk folder and put it in the same directory tree level as the tng sdk folder | 1 |
587,543 | 17,619,019,443 | IssuesEvent | 2021-08-18 13:21:02 | erlang/otp | https://api.github.com/repos/erlang/otp | closed | Distribution handshake error ( Connection attempt from node .. rejected. Invalid challenge reply ) | team:VM in progress not a bug priority:medium | **Describe the bug**
Two nodes with different cookies trying to connect to each other after both nodes have previously
use set_cookie(OtherNode, OtherCookie)
**To Reproduce**
1. Start two nodes with different cookies.
2. On each node call erlang:set_cookie( Node, Cookie ), replacing Node & Cookie with the Other Nodes details
3. attempt connecting the nodes from either node ( i.e net_adm:ping )
4. Error appears "Connection attempt from node .. rejected. Invalid challenge reply "
**Expected behavior**
I expect the nodes to connect successfully.
**Affected versions**
23.3.3 is what I have tested on
**Additional context**
I created a Pull request with Tests and Fix.
( https://github.com/erlang/otp/pull/5062 )
| 1.0 | Distribution handshake error ( Connection attempt from node .. rejected. Invalid challenge reply ) - **Describe the bug**
Two nodes with different cookies trying to connect to each other after both nodes have previously
use set_cookie(OtherNode, OtherCookie)
**To Reproduce**
1. Start two nodes with different cookies.
2. On each node call erlang:set_cookie( Node, Cookie ), replacing Node & Cookie with the Other Nodes details
3. attempt connecting the nodes from either node ( i.e net_adm:ping )
4. Error appears "Connection attempt from node .. rejected. Invalid challenge reply "
**Expected behavior**
I expect the nodes to connect successfully.
**Affected versions**
23.3.3 is what I have tested on
**Additional context**
I created a Pull request with Tests and Fix.
( https://github.com/erlang/otp/pull/5062 )
| priority | distribution handshake error connection attempt from node rejected invalid challenge reply describe the bug two nodes with different cookies trying to connect to each other after both nodes have previously use set cookie othernode othercookie to reproduce start two nodes with different cookies on each node call erlang set cookie node cookie replacing node cookie with the other nodes details attempt connecting the nodes from either node i e net adm ping error appears connection attempt from node rejected invalid challenge reply expected behavior i expect the nodes to connect successfully affected versions is what i have tested on additional context i created a pull request with tests and fix | 1 |
99,655 | 4,058,686,911 | IssuesEvent | 2016-05-25 06:13:31 | quantopian/pyfolio | https://api.github.com/repos/quantopian/pyfolio | closed | Add net exposure line to the Long/Short Exposure plot | enhancement help wanted medium priority | Just subtract Short from Long and plot the difference as a timeseries.
Also a couple of minor formatting changes.
Long exposure = _Green_
Short exposure = _Red_
Difference line = _Black dashed_
Make Long exposure line positive number, and Short exposure line all negative numbers:
(input from @Jstauth) rather than showing both as positive (abs) we should show the long as positive and short as negative, then we could overlay the net exposure as a line)

| 1.0 | Add net exposure line to the Long/Short Exposure plot - Just subtract Short from Long and plot the difference as a timeseries.
Also a couple of minor formatting changes.
Long exposure = _Green_
Short exposure = _Red_
Difference line = _Black dashed_
Make Long exposure line positive number, and Short exposure line all negative numbers:
(input from @Jstauth) rather than showing both as positive (abs) we should show the long as positive and short as negative, then we could overlay the net exposure as a line)

| priority | add net exposure line to the long short exposure plot just subtract short from long and plot the difference as a timeseries also a couple of minor formatting changes long exposure green short exposure red difference line black dashed make long exposure line positive number and short exposure line all negative numbers input from jstauth rather than showing both as positive abs we should show the long as positive and short as negative then we could overlay the net exposure as a line | 1 |
770,794 | 27,056,888,129 | IssuesEvent | 2023-02-13 16:45:26 | noctuelles/42-ft_transcendance | https://api.github.com/repos/noctuelles/42-ft_transcendance | closed | Liste des utilisateurs ne s'update pas en temps reel | medium priority | La liste des utilisateurs a droite du chat ne s'update pas quand un utilisateur rentre dans le channel | 1.0 | Liste des utilisateurs ne s'update pas en temps reel - La liste des utilisateurs a droite du chat ne s'update pas quand un utilisateur rentre dans le channel | priority | liste des utilisateurs ne s update pas en temps reel la liste des utilisateurs a droite du chat ne s update pas quand un utilisateur rentre dans le channel | 1 |
202,881 | 7,055,944,927 | IssuesEvent | 2018-01-04 10:34:33 | briqInstitute/interalpy | https://api.github.com/repos/briqInstitute/interalpy | closed | Random Initialization file with None as bounds .... | pb-estimation priority-medium size-S | Right now, all random files have both bounds for each parameter. | 1.0 | Random Initialization file with None as bounds .... - Right now, all random files have both bounds for each parameter. | priority | random initialization file with none as bounds right now all random files have both bounds for each parameter | 1 |
257,162 | 8,133,468,952 | IssuesEvent | 2018-08-19 02:13:13 | minio/minio | https://api.github.com/repos/minio/minio | closed | High memory usage in http/rpc framework | priority: medium | <!--- Provide a general summary of the issue in the Title above -->
** not using the template on purpose **
After using Minio from the latest master and doing some performance testing while reading 32MiB objects with 4096 workers, we see a significant increase in memory. Erasure coding is not using a lot of memory here instead it seems that memory is significantly higher due to gob.NewDecoder using bytes.Buffer.
The relevant graph attached here... [mem.pdf](https://github.com/minio/minio/files/2264057/mem.pdf)
| 1.0 | High memory usage in http/rpc framework - <!--- Provide a general summary of the issue in the Title above -->
** not using the template on purpose **
After using Minio from the latest master and doing some performance testing while reading 32MiB objects with 4096 workers, we see a significant increase in memory. Erasure coding is not using a lot of memory here instead it seems that memory is significantly higher due to gob.NewDecoder using bytes.Buffer.
The relevant graph attached here... [mem.pdf](https://github.com/minio/minio/files/2264057/mem.pdf)
| priority | high memory usage in http rpc framework not using the template on purpose after using minio from the latest master and doing some performance testing while reading objects with workers we see a significant increase in memory erasure coding is not using a lot of memory here instead it seems that memory is significantly higher due to gob newdecoder using bytes buffer the relevant graph attached here | 1 |
365,635 | 10,790,149,592 | IssuesEvent | 2019-11-05 13:52:40 | bounswe/bounswe2019group6 | https://api.github.com/repos/bounswe/bounswe2019group6 | opened | [Frontend] Implement Trading Equipment Interface | priority:medium related:frontend status:in-progress type:new-feature type:project-task | An interface for showing a list of trading equipments needs to be implemented. | 1.0 | [Frontend] Implement Trading Equipment Interface - An interface for showing a list of trading equipments needs to be implemented. | priority | implement trading equipment interface an interface for showing a list of trading equipments needs to be implemented | 1 |
25,556 | 2,683,841,479 | IssuesEvent | 2015-03-28 11:23:45 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | ConEmu.Maximus5.091016: теряется длинный вывод предпоследней команды | 1 star bug imported Priority-Medium | _From [yurivk...@gmail.com](https://code.google.com/u/109564459582005085765/) on October 22, 2009 07:34:38_
Версия ОС: Windows XP SP3
Версия FAR: 2.0 build 1086 x86
При просмотре консольного вывода в редакторе и просмотрщике отображается не
вся история, которая могла бы вместиться в заданное количество строк.
To reproduce:
[x] MultiCon
[x] Long console output [1000]
cd "C:\Program Files\FAR"
dir /b /s
dir /s
F11 | ConEmu support | Edit console output
Expected:
в редакторе вывод обеих команд dir, обрезанный до 1000 строк.
Observed:
в редакторе вывод предпоследней команды, обрезанный до одного экрана, затем
полный вывод последней команды.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=115_ | 1.0 | ConEmu.Maximus5.091016: теряется длинный вывод предпоследней команды - _From [yurivk...@gmail.com](https://code.google.com/u/109564459582005085765/) on October 22, 2009 07:34:38_
Версия ОС: Windows XP SP3
Версия FAR: 2.0 build 1086 x86
При просмотре консольного вывода в редакторе и просмотрщике отображается не
вся история, которая могла бы вместиться в заданное количество строк.
To reproduce:
[x] MultiCon
[x] Long console output [1000]
cd "C:\Program Files\FAR"
dir /b /s
dir /s
F11 | ConEmu support | Edit console output
Expected:
в редакторе вывод обеих команд dir, обрезанный до 1000 строк.
Observed:
в редакторе вывод предпоследней команды, обрезанный до одного экрана, затем
полный вывод последней команды.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=115_ | priority | conemu теряется длинный вывод предпоследней команды from on october версия ос windows xp версия far build при просмотре консольного вывода в редакторе и просмотрщике отображается не вся история которая могла бы вместиться в заданное количество строк to reproduce multicon long console output cd c program files far dir b s dir s conemu support edit console output expected в редакторе вывод обеих команд dir обрезанный до строк observed в редакторе вывод предпоследней команды обрезанный до одного экрана затем полный вывод последней команды original issue | 1 |
437,189 | 12,564,786,335 | IssuesEvent | 2020-06-08 08:38:30 | kir-dev/pek-next | https://api.github.com/repos/kir-dev/pek-next | opened | Add a delete button to notifications | UI / UX enhancement priority: medium | A magyar leírás lent található
## English description
A request came that it'd be nice to be able to delete the notifications. There should be an 'X' next to every notification item that deletes it from the list.
---
## Magyar leírás
Jött egy kérés, hogy lehessen értesítéseket egyesével törölni. Ehhez kéne egy "X" gomb minden értesítéshez a felugró menüben és az aloldalon is.
| 1.0 | Add a delete button to notifications - A magyar leírás lent található
## English description
A request came that it'd be nice to be able to delete the notifications. There should be an 'X' next to every notification item that deletes it from the list.
---
## Magyar leírás
Jött egy kérés, hogy lehessen értesítéseket egyesével törölni. Ehhez kéne egy "X" gomb minden értesítéshez a felugró menüben és az aloldalon is.
| priority | add a delete button to notifications a magyar leírás lent található english description a request came that it d be nice to be able to delete the notifications there should be an x next to every notification item that deletes it from the list magyar leírás jött egy kérés hogy lehessen értesítéseket egyesével törölni ehhez kéne egy x gomb minden értesítéshez a felugró menüben és az aloldalon is | 1 |
80,621 | 3,568,341,471 | IssuesEvent | 2016-01-26 04:49:56 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | closed | Add a "New Simulation" button to the parameters view to create a new simulation with the current parameters | > Enhancement Affects Usability Concerns Interface Concerns Simulation OS All Priority Medium Version Git | In order to leverage the work that has been done on the support of multiple simulations running side by side inside one experiment, I think a good idea would be to add a button in the Parameters view which would launch a new simulation with the current parameters, in order for people to quickly test the effect of parameters without having to modify the model. | 1.0 | Add a "New Simulation" button to the parameters view to create a new simulation with the current parameters - In order to leverage the work that has been done on the support of multiple simulations running side by side inside one experiment, I think a good idea would be to add a button in the Parameters view which would launch a new simulation with the current parameters, in order for people to quickly test the effect of parameters without having to modify the model. | priority | add a new simulation button to the parameters view to create a new simulation with the current parameters in order to leverage the work that has been done on the support of multiple simulations running side by side inside one experiment i think a good idea would be to add a button in the parameters view which would launch a new simulation with the current parameters in order for people to quickly test the effect of parameters without having to modify the model | 1 |
85,311 | 3,689,365,186 | IssuesEvent | 2016-02-25 16:13:24 | DigitalCampus/django-oppia | https://api.github.com/repos/DigitalCampus/django-oppia | closed | Add default report for no badges (course completion rates) | medium priority time: 2 hours | Required for Instrat Nigeria work - but will also be useful for most (all?) implementations
Under the reports app - add a report that will allow admins/staff to select a course, then show a listing of all the users who have completed the course (ie. have been awarded the badge) and all those who have started the course but not yet completed | 1.0 | Add default report for no badges (course completion rates) - Required for Instrat Nigeria work - but will also be useful for most (all?) implementations
Under the reports app - add a report that will allow admins/staff to select a course, then show a listing of all the users who have completed the course (ie. have been awarded the badge) and all those who have started the course but not yet completed | priority | add default report for no badges course completion rates required for instrat nigeria work but will also be useful for most all implementations under the reports app add a report that will allow admins staff to select a course then show a listing of all the users who have completed the course ie have been awarded the badge and all those who have started the course but not yet completed | 1 |
256,687 | 8,128,470,632 | IssuesEvent | 2018-08-17 11:57:11 | MARKETProtocol/dApp | https://api.github.com/repos/MARKETProtocol/dApp | closed | [Design] Mock Trading UI | Priority: Medium Status: Completed Type: Enhancement | ### User Story
As a user, I would like to see one screen that encompasses all needed information for me to be able to make a trade using MARKET Protocol
### Why Is this Needed?
*Summary*: Better UI/UX
### Description
*Type*: Feature
### Solution
*Summary*: @nitinrgupta and @auregimon to mock up trading UI
### Definition of complete
- [x] Overall trading UI
### Additional Information
Reference for a nice example - https://app.radarrelay.com/ZRX/WETH
| 1.0 | [Design] Mock Trading UI - ### User Story
As a user, I would like to see one screen that encompasses all needed information for me to be able to make a trade using MARKET Protocol
### Why Is this Needed?
*Summary*: Better UI/UX
### Description
*Type*: Feature
### Solution
*Summary*: @nitinrgupta and @auregimon to mock up trading UI
### Definition of complete
- [x] Overall trading UI
### Additional Information
Reference for a nice example - https://app.radarrelay.com/ZRX/WETH
| priority | mock trading ui user story as a user i would like to see one screen that encompasses all needed information for me to be able to make a trade using market protocol why is this needed summary better ui ux description type feature solution summary nitinrgupta and auregimon to mock up trading ui definition of complete overall trading ui additional information reference for a nice example | 1 |
717,529 | 24,678,958,031 | IssuesEvent | 2022-10-18 19:29:30 | AY2223S1-CS2103T-T12-4/tp | https://api.github.com/repos/AY2223S1-CS2103T-T12-4/tp | closed | Delete multiple patients or tasks by a common name | type.Story priority.Medium | As a private nurse with many patients and tasks I want to search patients or tasks by name and delete them so that I can delete multiple at once with one command. | 1.0 | Delete multiple patients or tasks by a common name - As a private nurse with many patients and tasks I want to search patients or tasks by name and delete them so that I can delete multiple at once with one command. | priority | delete multiple patients or tasks by a common name as a private nurse with many patients and tasks i want to search patients or tasks by name and delete them so that i can delete multiple at once with one command | 1 |
363,845 | 10,755,995,722 | IssuesEvent | 2019-10-31 10:17:07 | seternate/lanpartymanager | https://api.github.com/repos/seternate/lanpartymanager | opened | Game server consoles dont show output | App: Client Priority: Medium Status: Accepted Type: Bug Type: Unexpected Behavior | Chivalery console does not show its output, redirect the pipe from java to the window. | 1.0 | Game server consoles dont show output - Chivalery console does not show its output, redirect the pipe from java to the window. | priority | game server consoles dont show output chivalery console does not show its output redirect the pipe from java to the window | 1 |
533,128 | 15,577,365,992 | IssuesEvent | 2021-03-17 13:27:50 | nhost/cli | https://api.github.com/repos/nhost/cli | closed | `nhost dev` fails if project is initiated but no .nhost folder exists. | Priority: Medium Status: Available Type: Bug | ## Reproduce:
`nhost init`
remove `.nhost`.
`nhost dev`.
## Fix
`nhost dev` should first make sure the `.nhost` exists. If `.nhost` does not exists, create it.
| 1.0 | `nhost dev` fails if project is initiated but no .nhost folder exists. - ## Reproduce:
`nhost init`
remove `.nhost`.
`nhost dev`.
## Fix
`nhost dev` should first make sure the `.nhost` exists. If `.nhost` does not exists, create it.
| priority | nhost dev fails if project is initiated but no nhost folder exists reproduce nhost init remove nhost nhost dev fix nhost dev should first make sure the nhost exists if nhost does not exists create it | 1 |
552,391 | 16,239,872,654 | IssuesEvent | 2021-05-07 08:11:05 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | opened | 3.9 Alpha 5 and 6: Accessing website from IE11 while delay js is enabled isnot redirecting to the uncached version of the page | module: delay JS priority: medium severity: major type: bug | **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => Alpha 6
- Used the search feature to ensure that the bug hasn’t been reported before => Yes
**Describe the bug**
Accessing the website from IE11 while new delayjs is on, not redirecting to the uncached version as mentioned in the doc but using the cached version (Using the theme as Sydney, the user will stuck in white page with loader and nothing happen when user make any action)
**To Reproduce**
Steps to reproduce the behavior:
1. From the dashboard make sure that delayjs is on
2. Visit the home page from IE11
3. No redirection to the uncached version of page
**Expected behavior**
Redirect to the uncached version of the page as IE11 is not compatible with this feature as mentioned here https://www.notion.so/wpmedia/Delay-JS-Docs-49e40c0b46ae424eb9e2aeb04833ce4c
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | 3.9 Alpha 5 and 6: Accessing website from IE11 while delay js is enabled isnot redirecting to the uncached version of the page - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => Alpha 6
- Used the search feature to ensure that the bug hasn’t been reported before => Yes
**Describe the bug**
Accessing the website from IE11 while new delayjs is on, not redirecting to the uncached version as mentioned in the doc but using the cached version (Using the theme as Sydney, the user will stuck in white page with loader and nothing happen when user make any action)
**To Reproduce**
Steps to reproduce the behavior:
1. From the dashboard make sure that delayjs is on
2. Visit the home page from IE11
3. No redirection to the uncached version of page
**Expected behavior**
Redirect to the uncached version of the page as IE11 is not compatible with this feature as mentioned here https://www.notion.so/wpmedia/Delay-JS-Docs-49e40c0b46ae424eb9e2aeb04833ce4c
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| priority | alpha and accessing website from while delay js is enabled isnot redirecting to the uncached version of the page before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version alpha used the search feature to ensure that the bug hasn’t been reported before yes describe the bug accessing the website from while new delayjs is on not redirecting to the uncached version as mentioned in the doc but using the cached version using the theme as sydney the user will stuck in white page with loader and nothing happen when user make any action to reproduce steps to reproduce the behavior from the dashboard make sure that delayjs is on visit the home page from no redirection to the uncached version of page expected behavior redirect to the uncached version of the page as is not compatible with this feature as mentioned here backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
249,139 | 7,953,891,972 | IssuesEvent | 2018-07-12 04:35:02 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Icons of user modded items missing or broken in preview texts | Medium Priority | Version: 0.7.3.3
**Steps to Reproduce:**
Look at the textual description of a community modded item
**Expected behavior:**
looks like vanilla items, has an icon on the left side of the item name
**Actual behavior:**
doesn't show an icon or displays text instead of an icon
__
As far as I can tell, items that are from community made mods are missing their icons in the textual bits of the game. They show up as items in the inventory, craft windows etc. just fine.
I've observed 2 different behaviours.
- If the icon is from a single sprite, it simply doesn't show up.
- If it belongs to a sprite sheet, the link to the right sprite is shown where the icon should be, making it hard to read any item information or receipe.
I included pictures of the two behaviours.
Tested this with my own as well as other mods.
example of a missing icon

example of text replacing the missing icon if from a sprite sheet
 | 1.0 | Icons of user modded items missing or broken in preview texts - Version: 0.7.3.3
**Steps to Reproduce:**
Look at the textual description of a community modded item
**Expected behavior:**
looks like vanilla items, has an icon on the left side of the item name
**Actual behavior:**
doesn't show an icon or displays text instead of an icon
__
As far as I can tell, items that are from community made mods are missing their icons in the textual bits of the game. They show up as items in the inventory, craft windows etc. just fine.
I've observed 2 different behaviours.
- If the icon is from a single sprite, it simply doesn't show up.
- If it belongs to a sprite sheet, the link to the right sprite is shown where the icon should be, making it hard to read any item information or receipe.
I included pictures of the two behaviours.
Tested this with my own as well as other mods.
example of a missing icon

example of text replacing the missing icon if from a sprite sheet
 | priority | icons of user modded items missing or broken in preview texts version steps to reproduce look at the textual description of a community modded item expected behavior looks like vanilla items has an icon on the left side of the item name actual behavior doesn t show an icon or displays text instead of an icon as far as i can tell items that are from community made mods are missing their icons in the textual bits of the game they show up as items in the inventory craft windows etc just fine i ve observed different behaviours if the icon is from a single sprite it simply doesn t show up if it belongs to a sprite sheet the link to the right sprite is shown where the icon should be making it hard to read any item information or receipe i included pictures of the two behaviours tested this with my own as well as other mods example of a missing icon example of text replacing the missing icon if from a sprite sheet | 1 |
392,575 | 11,593,190,295 | IssuesEvent | 2020-02-24 13:08:48 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | opened | [LB] (CLARISA) Project Page | Priority - Medium Type -Task | External Partnership
- [ ] Get
- [ ] Post
- [ ] Delete
- [ ] Put
- [ ] Validator
**Deliverable:** CLARISA funtionality
**Move to Review when:** Funtionality will be implemente on CLARISA DeV.
**Move to Closed when:** Funtionality will be implemente on Doppler.
| 1.0 | [LB] (CLARISA) Project Page - External Partnership
- [ ] Get
- [ ] Post
- [ ] Delete
- [ ] Put
- [ ] Validator
**Deliverable:** CLARISA funtionality
**Move to Review when:** Funtionality will be implemente on CLARISA DeV.
**Move to Closed when:** Funtionality will be implemente on Doppler.
| priority | clarisa project page external partnership get post delete put validator deliverable clarisa funtionality move to review when funtionality will be implemente on clarisa dev move to closed when funtionality will be implemente on doppler | 1 |
795,157 | 28,064,070,651 | IssuesEvent | 2023-03-29 14:20:45 | zowe/vscode-extension-for-zowe | https://api.github.com/repos/zowe/vscode-extension-for-zowe | closed | Can't connect to mainframe in Theia | bug docs Theia On-Hold priority-medium severity-high | <!--
Before opening a new issue, please search our existing issues: https://github.com/zowe/vscode-extension-for-zowe/issues
-->
**Describe the bug**
In Theia 1.33.0, attempts to connect to mainframe with ZE result in this error: "Failed to establish a socket connection to proxies"
**To Reproduce**
1. Start Theia in a Docker container: `docker run -it --init --name theia -p 3000:3000 t1m0thyj/theia-alpine`
2. Copy ZE VSIX into the container: `docker cp dist/vscode-extension-for-zowe-2.6.0-SNAPSHOT.vsix theia:/home/theia/plugins`
3. Go to Theia (http://localhost:3000) and install ZE from VSIX
4. Open Zowe Explorer panel and add a new profile
5. Try to search for data sets and see the error in the screenshot below
**Expected behavior**
Should be able to search for data sets as long as the profile contains valid connection info.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Theia 1.33.0
- Zowe Explorer Version: 2.6.0-SNAPSHOT
**Additional context**
This is likely caused by a bug in Theia: https://github.com/eclipse-theia/theia/issues/11911
Perhaps we can comment on that issue to provide additional info.
Also, a workaround may be to downgrade to an older version of Theia.
Interestingly, Zowe CLI successfully connects to the mainframe in the Theia terminal with these commands:
```
mkdir test && cd test && npm init -y && npm install @zowe/cli
npx zowe files ls ds SYS1.PARMLIB
``` | 1.0 | Can't connect to mainframe in Theia - <!--
Before opening a new issue, please search our existing issues: https://github.com/zowe/vscode-extension-for-zowe/issues
-->
**Describe the bug**
In Theia 1.33.0, attempts to connect to mainframe with ZE result in this error: "Failed to establish a socket connection to proxies"
**To Reproduce**
1. Start Theia in a Docker container: `docker run -it --init --name theia -p 3000:3000 t1m0thyj/theia-alpine`
2. Copy ZE VSIX into the container: `docker cp dist/vscode-extension-for-zowe-2.6.0-SNAPSHOT.vsix theia:/home/theia/plugins`
3. Go to Theia (http://localhost:3000) and install ZE from VSIX
4. Open Zowe Explorer panel and add a new profile
5. Try to search for data sets and see the error in the screenshot below
**Expected behavior**
Should be able to search for data sets as long as the profile contains valid connection info.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Theia 1.33.0
- Zowe Explorer Version: 2.6.0-SNAPSHOT
**Additional context**
This is likely caused by a bug in Theia: https://github.com/eclipse-theia/theia/issues/11911
Perhaps we can comment on that issue to provide additional info.
Also, a workaround may be to downgrade to an older version of Theia.
Interestingly, Zowe CLI successfully connects to the mainframe in the Theia terminal with these commands:
```
mkdir test && cd test && npm init -y && npm install @zowe/cli
npx zowe files ls ds SYS1.PARMLIB
``` | priority | can t connect to mainframe in theia before opening a new issue please search our existing issues describe the bug in theia attempts to connect to mainframe with ze result in this error failed to establish a socket connection to proxies to reproduce start theia in a docker container docker run it init name theia p theia alpine copy ze vsix into the container docker cp dist vscode extension for zowe snapshot vsix theia home theia plugins go to theia and install ze from vsix open zowe explorer panel and add a new profile try to search for data sets and see the error in the screenshot below expected behavior should be able to search for data sets as long as the profile contains valid connection info screenshots desktop please complete the following information os theia zowe explorer version snapshot additional context this is likely caused by a bug in theia perhaps we can comment on that issue to provide additional info also a workaround may be to downgrade to an older version of theia interestingly zowe cli successfully connects to the mainframe in the theia terminal with these commands mkdir test cd test npm init y npm install zowe cli npx zowe files ls ds parmlib | 1 |
27,424 | 2,692,700,928 | IssuesEvent | 2015-04-01 09:52:46 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | opened | Custom org page: Admin - visualization type | bug Custom org page Priority-Medium | When editing an organization there are 2 values for visualization type: 3w and wfp. If we save wfp as value, the value is saved but in edit form is not shown this value. | 1.0 | Custom org page: Admin - visualization type - When editing an organization there are 2 values for visualization type: 3w and wfp. If we save wfp as value, the value is saved but in edit form is not shown this value. | priority | custom org page admin visualization type when editing an organization there are values for visualization type and wfp if we save wfp as value the value is saved but in edit form is not shown this value | 1 |
607,310 | 18,779,543,442 | IssuesEvent | 2021-11-08 03:39:33 | matrixorigin/matrixcube | https://api.github.com/repos/matrixorigin/matrixcube | opened | shard object has to be copied from the SM when sending raft messages | component/raftstore priority/medium kind/refactoring | See link below
https://github.com/matrixorigin/matrixcube/blob/5e47a03cc25d1ad70f8f0e037eb826b6ac73d31f/raftstore/replica_event_raft_ready.go#L183
Another related question is whether it is possible to avoid including the Epoch value in the raft message.
| 1.0 | shard object has to be copied from the SM when sending raft messages - See link below
https://github.com/matrixorigin/matrixcube/blob/5e47a03cc25d1ad70f8f0e037eb826b6ac73d31f/raftstore/replica_event_raft_ready.go#L183
Another related question is whether it is possible to avoid including the Epoch value in the raft message.
| priority | shard object has to be copied from the sm when sending raft messages see link below another related question is whether it is possible to avoid including the epoch value in the raft message | 1 |
395,473 | 11,686,829,466 | IssuesEvent | 2020-03-05 11:37:36 | threefoldtech/3Bot_connect | https://api.github.com/repos/threefoldtech/3Bot_connect | opened | The support page does not scale correctly when the keyboard is opened. ( Both IOS and Android ) | priority_medium type_bug | **Repro steps**
1) Go to support page in the 3bot Connect app
2) Attempt to type a message
**Expected Result**
The layout scales correctly, the input field is fully visible and the last messages remain visible.
**Actual Result**

**System Info**
Type: Staging
Git user: Mathias De Weerdt
Git branch: v1.5.1
Git hash: accfd96
Time: 17:46:59 04.03.2020 | 1.0 | The support page does not scale correctly when the keyboard is opened. ( Both IOS and Android ) - **Repro steps**
1) Go to support page in the 3bot Connect app
2) Attempt to type a message
**Expected Result**
The layout scales correctly, the input field is fully visible and the last messages remain visible.
**Actual Result**

**System Info**
Type: Staging
Git user: Mathias De Weerdt
Git branch: v1.5.1
Git hash: accfd96
Time: 17:46:59 04.03.2020 | priority | the support page does not scale correctly when the keyboard is opened both ios and android repro steps go to support page in the connect app attempt to type a message expected result the layout scales correctly the input field is fully visible and the last messages remain visible actual result system info type staging git user mathias de weerdt git branch git hash time | 1 |
829,999 | 31,933,613,068 | IssuesEvent | 2023-09-19 09:02:52 | vscentrum/vsc-software-stack | https://api.github.com/repos/vscentrum/vsc-software-stack | opened | Columbus | difficulty: medium Fortran new priority: medium site:ugent sources-only | * link to support ticket: [#2023090860001081](https://otrsdict.ugent.be/otrs/index.pl?Action=AgentTicketZoom;TicketID=130122)
* website: https://columbus-program-system.gitlab.io/columbus/
* installation docs: https://columbus-program-system.gitlab.io/columbus/installation.html
* toolchain: `...`
* easyblock to use: `...`
* required dependencies:
* [ ] Fortran90 compiler
* [ ] C compiler
* [ ] [Perl](http://www.cpan.org/) , essential add-on modules are contained in the $COLUMBUS/../TOOLS/CPAN directory
* [ ] UNIX utilities: gmake, grep, tar, gunzip
* [ ] optimized BLAS and LAPACK libraries such as the [MKL ](http://software.intel.com/en-us/intel-mkl)or [ACML](http://developer.amd.com/cpu/libraries/acml)libraries; alternative is an optimized BLAS library (e.g. [GOTOBLAS ](http://www.tacc.utexas.edu/tacc-projects/gotoblas2/)plus a fortran based LAPACK library; the fortran sources for [BLAS and LAPACK ](https://columbus-program-system.gitlab.io/columbus/www.netlib.org/lapack)can be found in the $COLUMBUS/../TOOLS/LAPACK directory
* [ ] for parallel operation: installed MPI1 or MPI2 (e.g. the [MPICH ](http://www.mcs.anl.gov/mpi)implementation) plus installed [Global Array package](http://www.emsl.pnl.gov/docs/global/ga.html) (4.2 and higher)
* [ ] for [OpenMolcas](https://gitlab.com/Molcas/OpenMolcas/) / [Molcas](http://www.molcas.org/) support a serial installation of the MOLCAS package ( 7.3 and higher)
* [ ] The molecular visualization packages [MOLDEN](http://www.caos.kun.nl/~schaft/molden/molden.html) is supported. A separate installation of this software is assumed.
* optional dependencies:
* [ ] ...
* notes:
* ...
* effort: *(TBD)*
* other install methods
* conda: no
* container image: no
* pre-built binaries (RHEL8 Linux x86_64): [yes?](https://gitlab.com/columbus-program-system/columbus/-/releases/) / no
* easyconfig outside EasyBuild: no
| 1.0 | Columbus - * link to support ticket: [#2023090860001081](https://otrsdict.ugent.be/otrs/index.pl?Action=AgentTicketZoom;TicketID=130122)
* website: https://columbus-program-system.gitlab.io/columbus/
* installation docs: https://columbus-program-system.gitlab.io/columbus/installation.html
* toolchain: `...`
* easyblock to use: `...`
* required dependencies:
* [ ] Fortran90 compiler
* [ ] C compiler
* [ ] [Perl](http://www.cpan.org/) , essential add-on modules are contained in the $COLUMBUS/../TOOLS/CPAN directory
* [ ] UNIX utilities: gmake, grep, tar, gunzip
* [ ] optimized BLAS and LAPACK libraries such as the [MKL ](http://software.intel.com/en-us/intel-mkl)or [ACML](http://developer.amd.com/cpu/libraries/acml)libraries; alternative is an optimized BLAS library (e.g. [GOTOBLAS ](http://www.tacc.utexas.edu/tacc-projects/gotoblas2/)plus a fortran based LAPACK library; the fortran sources for [BLAS and LAPACK ](https://columbus-program-system.gitlab.io/columbus/www.netlib.org/lapack)can be found in the $COLUMBUS/../TOOLS/LAPACK directory
* [ ] for parallel operation: installed MPI1 or MPI2 (e.g. the [MPICH ](http://www.mcs.anl.gov/mpi)implementation) plus installed [Global Array package](http://www.emsl.pnl.gov/docs/global/ga.html) (4.2 and higher)
* [ ] for [OpenMolcas](https://gitlab.com/Molcas/OpenMolcas/) / [Molcas](http://www.molcas.org/) support a serial installation of the MOLCAS package ( 7.3 and higher)
* [ ] The molecular visualization packages [MOLDEN](http://www.caos.kun.nl/~schaft/molden/molden.html) is supported. A separate installation of this software is assumed.
* optional dependencies:
* [ ] ...
* notes:
* ...
* effort: *(TBD)*
* other install methods
* conda: no
* container image: no
* pre-built binaries (RHEL8 Linux x86_64): [yes?](https://gitlab.com/columbus-program-system/columbus/-/releases/) / no
* easyconfig outside EasyBuild: no
| priority | columbus link to support ticket website installation docs toolchain easyblock to use required dependencies compiler c compiler essential add on modules are contained in the columbus tools cpan directory unix utilities gmake grep tar gunzip optimized blas and lapack libraries such as the alternative is an optimized blas library e g a fortran based lapack library the fortran sources for be found in the columbus tools lapack directory for parallel operation installed or e g the plus installed and higher for support a serial installation of the molcas package and higher the molecular visualization packages is supported a separate installation of this software is assumed optional dependencies notes effort tbd other install methods conda no container image no pre built binaries linux no easyconfig outside easybuild no | 1 |
6,519 | 2,588,964,153 | IssuesEvent | 2015-02-18 08:28:18 | calblueprint/PHC | https://api.github.com/repos/calblueprint/PHC | opened | fully implement security key | feature medium priority | should check for actual security key in the onCreate of all activities.
notify Kate about this too. | 1.0 | fully implement security key - should check for actual security key in the onCreate of all activities.
notify Kate about this too. | priority | fully implement security key should check for actual security key in the oncreate of all activities notify kate about this too | 1 |
136,789 | 5,288,846,653 | IssuesEvent | 2017-02-08 16:03:23 | rd-alliance/metadata-catalog-dev | https://api.github.com/repos/rd-alliance/metadata-catalog-dev | opened | Search by data type | GUI medium priority | Responding to a search query of a data type, display a list of matching records. | 1.0 | Search by data type - Responding to a search query of a data type, display a list of matching records. | priority | search by data type responding to a search query of a data type display a list of matching records | 1 |
401,026 | 11,784,266,378 | IssuesEvent | 2020-03-17 07:58:49 | HE-Arc/CSRuby | https://api.github.com/repos/HE-Arc/CSRuby | closed | Mettre en place le déploiement continu | CI deploy medium priority | Si possible avec les github actions. Sinon trouver une autre solution ([semaphore](https://semaphoreci.com/), [capistrano)](https://capistranorb.com/documentation/overview/what-is-capistrano/) ?) | 1.0 | Mettre en place le déploiement continu - Si possible avec les github actions. Sinon trouver une autre solution ([semaphore](https://semaphoreci.com/), [capistrano)](https://capistranorb.com/documentation/overview/what-is-capistrano/) ?) | priority | mettre en place le déploiement continu si possible avec les github actions sinon trouver une autre solution | 1 |
715,467 | 24,600,239,290 | IssuesEvent | 2022-10-14 11:52:34 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | opened | Revising the Diagrams based on Signup&Login | priority-medium status-new diagrams | ### Issue Description
We have a detailed discussion about our requirements in our last weekly meeting. We have determined the requirements to be updated and how to do updates. Now, we will revise and re-organize the diagrams according to the new requirement updates. Since I am the responsible person for updating the signup&login requirements, I will also update the diagrams related to those parts. You can see the details of the requirement update issue [here](https://github.com/bounswe/bounswe2022group2/issues/331).
### Step Details
Steps that will be performed:
- [ ] Control the class diagram and detect the points that need to be updated (if there are any) with the update of requirements
- [ ] Control the sequence and use-case diagrams and make the necessary updates (if any) with the update of requirements
### Final Actions
- [ ] The final version of the diagrams will be available to be seen in their corresponding documents:
[Class Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Class-Diagram)
[Use-case Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Use-Case-Diagram)
[Sequence Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Sequence-Diagram)
- [ ] Share the changes under this issue as a comment. So that, the other teammates will be informed.
### Deadline of the Issue
15.10.2022 - Saturday - 23:59
### Reviewer
Ecenur Sezer
### Deadline for the Review
16.10.2022 - Sunday - 23:59 | 1.0 | Revising the Diagrams based on Signup&Login - ### Issue Description
We have a detailed discussion about our requirements in our last weekly meeting. We have determined the requirements to be updated and how to do updates. Now, we will revise and re-organize the diagrams according to the new requirement updates. Since I am the responsible person for updating the signup&login requirements, I will also update the diagrams related to those parts. You can see the details of the requirement update issue [here](https://github.com/bounswe/bounswe2022group2/issues/331).
### Step Details
Steps that will be performed:
- [ ] Control the class diagram and detect the points that need to be updated (if there are any) with the update of requirements
- [ ] Control the sequence and use-case diagrams and make the necessary updates (if any) with the update of requirements
### Final Actions
- [ ] The final version of the diagrams will be available to be seen in their corresponding documents:
[Class Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Class-Diagram)
[Use-case Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Use-Case-Diagram)
[Sequence Diagram](https://github.com/bounswe/bounswe2022group2/wiki/Sequence-Diagram)
- [ ] Share the changes under this issue as a comment. So that, the other teammates will be informed.
### Deadline of the Issue
15.10.2022 - Saturday - 23:59
### Reviewer
Ecenur Sezer
### Deadline for the Review
16.10.2022 - Sunday - 23:59 | priority | revising the diagrams based on signup login issue description we have a detailed discussion about our requirements in our last weekly meeting we have determined the requirements to be updated and how to do updates now we will revise and re organize the diagrams according to the new requirement updates since i am the responsible person for updating the signup login requirements i will also update the diagrams related to those parts you can see the details of the requirement update issue step details steps that will be performed control the class diagram and detect the points that need to be updated if there are any with the update of requirements control the sequence and use case diagrams and make the necessary updates if any with the update of requirements final actions the final version of the diagrams will be available to be seen in their corresponding documents share the changes under this issue as a comment so that the other teammates will be informed deadline of the issue saturday reviewer ecenur sezer deadline for the review sunday | 1 |
100,048 | 4,075,823,403 | IssuesEvent | 2016-05-29 13:44:02 | EsotericSoftware/jsonbeans | https://api.github.com/repos/EsotericSoftware/jsonbeans | closed | BSON Support | bug imported Priority-Medium | _From [harry.c...@cloudshards.com](https://code.google.com/u/103605517436890058774/) on May 02, 2013 10:36:58_
Looking to use this with MongoDB. Possible to add BSON support? Thanks.
_Original issue: http://code.google.com/p/jsonbeans/issues/detail?id=3_ | 1.0 | BSON Support - _From [harry.c...@cloudshards.com](https://code.google.com/u/103605517436890058774/) on May 02, 2013 10:36:58_
Looking to use this with MongoDB. Possible to add BSON support? Thanks.
_Original issue: http://code.google.com/p/jsonbeans/issues/detail?id=3_ | priority | bson support from on may looking to use this with mongodb possible to add bson support thanks original issue | 1 |
673,449 | 22,969,996,691 | IssuesEvent | 2022-07-20 01:32:05 | MineInAbyss/Chatty | https://api.github.com/repos/MineInAbyss/Chatty | opened | Make release contain Paper and Velocity Jars | difficulty:medium priority:medium type:enhancement | Currently the jar in https://github.com/MineInAbyss/Chatty/releases is either only paper or velocity is manually added. | 1.0 | Make release contain Paper and Velocity Jars - Currently the jar in https://github.com/MineInAbyss/Chatty/releases is either only paper or velocity is manually added. | priority | make release contain paper and velocity jars currently the jar in is either only paper or velocity is manually added | 1 |
317,097 | 9,660,840,990 | IssuesEvent | 2019-05-20 16:23:58 | geosolutions-it/tdipisa | https://api.github.com/repos/geosolutions-it/tdipisa | closed | COMUNE FIRENZE: Mockup Catalog AND MS2 update (Priority 3) | Priority: Medium task todo | Sviluppi e aggiornamento mapstore2: attendiamo il mockup, poi ne riparliamo. AL momento pensiamo di aggiornare mapstore2 successivamente agli sviluppi | 1.0 | COMUNE FIRENZE: Mockup Catalog AND MS2 update (Priority 3) - Sviluppi e aggiornamento mapstore2: attendiamo il mockup, poi ne riparliamo. AL momento pensiamo di aggiornare mapstore2 successivamente agli sviluppi | priority | comune firenze mockup catalog and update priority sviluppi e aggiornamento attendiamo il mockup poi ne riparliamo al momento pensiamo di aggiornare successivamente agli sviluppi | 1 |
780,530 | 27,398,914,434 | IssuesEvent | 2023-02-28 22:14:31 | SETI/pds-data-projects | https://api.github.com/repos/SETI/pds-data-projects | closed | COISS_100x EARTH_RECEIVED_START/STOP_TIME are sometimes backwards | Wontfix Effort 2 Medium Priority 3 Important A-Bug | From pds-webserver created by [rfrenchseti](https://github.com/rfrenchseti): SETI/pds-webserver#2
From pds-opus created by [rfrenchseti](https://github.com/rfrenchseti) : SETI/pds-opus#564
```
mysql> select volume_id, opus_id, ert1, ert2 from obs_mission_cassini where ert1 > ert2;
+------------+--------------------+---------------+---------------+
| volume_id | opus_id | ert1 | ert2 |
+------------+--------------------+---------------+---------------+
| COISS_1001 | co-iss-n1349081927 | 23957818.153 | 23957006.639 |
| COISS_1001 | co-iss-n1349087639 | 23958093.216 | 23957268.716 |
| COISS_1001 | co-iss-n1349087709 | 23958149.519 | 23957337.486 |
| COISS_1001 | co-iss-n1349369175 | 24119298.375 | 24118479.149 |
| COISS_1001 | co-iss-n1349375027 | 24119605.215 | 24118785.653 |
| COISS_1001 | co-iss-n1349518442 | 24388803.394 | 24387978.906 |
| COISS_1001 | co-iss-n1349518626 | 24388938.077 | 24388126.111 |
| COISS_1001 | co-iss-n1349805794 | 24550378.995 | 24549558.758 |
| COISS_1001 | co-iss-n1349805908 | 24550457.652 | 24549638.799 |
| COISS_1001 | co-iss-w1349949286 | 24819620.349 | 24818811.102 |
| COISS_1001 | co-iss-n1350230678 | 24980803.252 | 24979984.757 |
| COISS_1001 | co-iss-n1350236633 | 24981181.735 | 24980362.302 |
| COISS_1001 | co-iss-n1350380089 | 25250402.168 | 25249576.829 |
| COISS_1001 | co-iss-n1351092421 | 25842473.925 | 25841655.725 |
| COISS_1001 | co-iss-n1351235654 | 26111543.337 | 26110718.667 |
| COISS_1001 | co-iss-n1351235695 | 26111569.701 | 26110745.332 |
| COISS_1001 | co-iss-n1351235760 | 26111615.937 | 26110805.667 |
| COISS_1001 | co-iss-n1351235805 | 26111656.88 | 26110848.006 |
| COISS_1002 | co-iss-n1352103365 | 26973743.071 | 26972877.895 |
| COISS_1002 | co-iss-w1352534103 | 27404540.266 | 27403732.059 |
| COISS_1002 | co-iss-w1352821385 | 27565905.805 | 27565088.735 |
| COISS_1002 | co-iss-w1352821450 | 27565967.928 | 27565150.793 |
| COISS_1002 | co-iss-n1352959142 | 27691835.991 | 27691020.614 |
| COISS_1002 | co-iss-w1352959250 | 27691949.284 | 27691133.049 |
| COISS_1002 | co-iss-w1352959604 | 27692140.761 | 27691325.654 |
| COISS_1002 | co-iss-n1353103023 | 27835932.817 | 27835089.134 |
| COISS_1002 | co-iss-n1353246939 | 27997432.803 | 27996612.476 |
| COISS_1002 | co-iss-w1353389896 | 28122619.158 | 28121808.323 |
| COISS_1002 | co-iss-n1353390109 | 28122811.371 | 28121979.005 |
| COISS_1002 | co-iss-w1353534699 | 28267166.167 | 28266301.865 |
| COISS_1003 | co-iss-w1354108491 | 28859039.841 | 28858218.133 |
| COISS_1003 | co-iss-n1354108602 | 28859097.564 | 28858263.538 |
| COISS_1003 | co-iss-n1354251550 | 28984275.917 | 28983462.481 |
| COISS_1003 | co-iss-w1354251658 | 28984393.59 | 28983577.447 |
| COISS_1003 | co-iss-n1354396007 | 29128726.534 | 29127883.944 |
| COISS_1003 | co-iss-w1354539405 | 29289944.366 | 29289122.259 |
| COISS_1003 | co-iss-n1354539714 | 29290009.316 | 29289188.938 |
| COISS_1003 | co-iss-n1354539825 | 29290058.459 | 29289225.574 |
| COISS_1003 | co-iss-n1354540018 | 29290075.479 | 29289255.267 |
| COISS_1003 | co-iss-w1354825928 | 29559115.881 | 29558270.098 |
| COISS_1004 | co-iss-n1356864735 | 31567469.092 | 31566648.944 |
| COISS_1004 | co-iss-n1356864777 | 31567506.48 | 31566702.471 |
| COISS_1004 | co-iss-n1356975912 | 31679022.099 | 31678229.579 |
| COISS_1004 | co-iss-w1357431545 | 32109327.67 | 32108502.242 |
| COISS_1004 | co-iss-n1357431653 | 32109393.818 | 32108575.733 |
| COISS_1004 | co-iss-w1357431786 | 32109505.133 | 32108678.177 |
| COISS_1005 | co-iss-n1357863387 | 32564818.116 | 32564001.064 |
| COISS_1005 | co-iss-w1357971535 | 32661833.307 | 32661048.098 |
| COISS_1005 | co-iss-n1358245168 | 33002132.033 | 33001347.082 |
| COISS_1005 | co-iss-w1358663405 | 33363658.981 | 33362851.696 |
| COISS_1005 | co-iss-n1358663623 | 33363788.801 | 33362981.883 |
| COISS_1005 | co-iss-n1358927780 | 33662603.336 | 33661821.067 |
| COISS_1005 | co-iss-n1358927964 | 33662771.458 | 33661974.278 |
| COISS_1005 | co-iss-n1359070782 | 33806348.667 | 33805549.494 |
| COISS_1005 | co-iss-n1359071373 | 33806525.275 | 33805733.94 |
| COISS_1006 | co-iss-w1359502060 | 34237307.461 | 34236508.77 |
| COISS_1006 | co-iss-w1360363027 | 35098798.924 | 35097997.179 |
| COISS_1006 | co-iss-n1360937901 | 35673953.939 | 35673153.317 |
| COISS_1006 | co-iss-n1361226923 | 35959092.187 | 35958289.342 |
| COISS_1006 | co-iss-n1361228653 | 35959334.752 | 35958532.527 |
| COISS_1006 | co-iss-n1361368738 | 36104841.363 | 36104042.417 |
| COISS_1006 | co-iss-n1361655926 | 36389658.751 | 36388856.819 |
| COISS_1006 | co-iss-n1361657689 | 36389937.778 | 36389135.955 |
| COISS_1007 | co-iss-n1364541189 | 39209911.435 | 39209105.791 |
| COISS_1007 | co-iss-n1364541321 | 39210006.646 | 39209200.988 |
| COISS_1007 | co-iss-n1364541420 | 39210071.276 | 39209259.072 |
| COISS_1007 | co-iss-n1364541584 | 39210164.402 | 39209359.088 |
| COISS_1007 | co-iss-n1364541716 | 39210261.13 | 39209455.666 |
| COISS_1007 | co-iss-n1364541749 | 39210285.335 | 39209479.712 |
| COISS_1007 | co-iss-n1364541979 | 39210419.813 | 39209614.705 |
| COISS_1007 | co-iss-n1404856096 | 79539950.668 | 79539318.142 |
| COISS_1008 | co-iss-n1412619196 | 87379169.361 | 87378397.73 |
| COISS_1008 | co-iss-n1412906634 | 87721828.069 | 87721071.006 |
| COISS_1008 | co-iss-n1412908730 | 87722239.197 | 87721495.891 |
| COISS_1008 | co-iss-n1412909266 | 87722377.528 | 87721628.333 |
| COISS_1008 | co-iss-n1412909802 | 87722510.979 | 87721760.81 |
| COISS_1008 | co-iss-n1413902027 | 88583979.532 | 88583249.968 |
| COISS_1008 | co-iss-n1413902105 | 88584085.201 | 88583373.068 |
| COISS_1008 | co-iss-n1421686131 | 96442078.492 | 96441309.157 |
| COISS_1008 | co-iss-n1421838870 | 96548946.531 | 96548179.089 |
| COISS_1008 | co-iss-w1429505174 | 104435821.548 | 104287270.587 |
+------------+--------------------+---------------+---------------+
```
| 1.0 | COISS_100x EARTH_RECEIVED_START/STOP_TIME are sometimes backwards - From pds-webserver created by [rfrenchseti](https://github.com/rfrenchseti): SETI/pds-webserver#2
From pds-opus created by [rfrenchseti](https://github.com/rfrenchseti) : SETI/pds-opus#564
```
mysql> select volume_id, opus_id, ert1, ert2 from obs_mission_cassini where ert1 > ert2;
+------------+--------------------+---------------+---------------+
| volume_id | opus_id | ert1 | ert2 |
+------------+--------------------+---------------+---------------+
| COISS_1001 | co-iss-n1349081927 | 23957818.153 | 23957006.639 |
| COISS_1001 | co-iss-n1349087639 | 23958093.216 | 23957268.716 |
| COISS_1001 | co-iss-n1349087709 | 23958149.519 | 23957337.486 |
| COISS_1001 | co-iss-n1349369175 | 24119298.375 | 24118479.149 |
| COISS_1001 | co-iss-n1349375027 | 24119605.215 | 24118785.653 |
| COISS_1001 | co-iss-n1349518442 | 24388803.394 | 24387978.906 |
| COISS_1001 | co-iss-n1349518626 | 24388938.077 | 24388126.111 |
| COISS_1001 | co-iss-n1349805794 | 24550378.995 | 24549558.758 |
| COISS_1001 | co-iss-n1349805908 | 24550457.652 | 24549638.799 |
| COISS_1001 | co-iss-w1349949286 | 24819620.349 | 24818811.102 |
| COISS_1001 | co-iss-n1350230678 | 24980803.252 | 24979984.757 |
| COISS_1001 | co-iss-n1350236633 | 24981181.735 | 24980362.302 |
| COISS_1001 | co-iss-n1350380089 | 25250402.168 | 25249576.829 |
| COISS_1001 | co-iss-n1351092421 | 25842473.925 | 25841655.725 |
| COISS_1001 | co-iss-n1351235654 | 26111543.337 | 26110718.667 |
| COISS_1001 | co-iss-n1351235695 | 26111569.701 | 26110745.332 |
| COISS_1001 | co-iss-n1351235760 | 26111615.937 | 26110805.667 |
| COISS_1001 | co-iss-n1351235805 | 26111656.88 | 26110848.006 |
| COISS_1002 | co-iss-n1352103365 | 26973743.071 | 26972877.895 |
| COISS_1002 | co-iss-w1352534103 | 27404540.266 | 27403732.059 |
| COISS_1002 | co-iss-w1352821385 | 27565905.805 | 27565088.735 |
| COISS_1002 | co-iss-w1352821450 | 27565967.928 | 27565150.793 |
| COISS_1002 | co-iss-n1352959142 | 27691835.991 | 27691020.614 |
| COISS_1002 | co-iss-w1352959250 | 27691949.284 | 27691133.049 |
| COISS_1002 | co-iss-w1352959604 | 27692140.761 | 27691325.654 |
| COISS_1002 | co-iss-n1353103023 | 27835932.817 | 27835089.134 |
| COISS_1002 | co-iss-n1353246939 | 27997432.803 | 27996612.476 |
| COISS_1002 | co-iss-w1353389896 | 28122619.158 | 28121808.323 |
| COISS_1002 | co-iss-n1353390109 | 28122811.371 | 28121979.005 |
| COISS_1002 | co-iss-w1353534699 | 28267166.167 | 28266301.865 |
| COISS_1003 | co-iss-w1354108491 | 28859039.841 | 28858218.133 |
| COISS_1003 | co-iss-n1354108602 | 28859097.564 | 28858263.538 |
| COISS_1003 | co-iss-n1354251550 | 28984275.917 | 28983462.481 |
| COISS_1003 | co-iss-w1354251658 | 28984393.59 | 28983577.447 |
| COISS_1003 | co-iss-n1354396007 | 29128726.534 | 29127883.944 |
| COISS_1003 | co-iss-w1354539405 | 29289944.366 | 29289122.259 |
| COISS_1003 | co-iss-n1354539714 | 29290009.316 | 29289188.938 |
| COISS_1003 | co-iss-n1354539825 | 29290058.459 | 29289225.574 |
| COISS_1003 | co-iss-n1354540018 | 29290075.479 | 29289255.267 |
| COISS_1003 | co-iss-w1354825928 | 29559115.881 | 29558270.098 |
| COISS_1004 | co-iss-n1356864735 | 31567469.092 | 31566648.944 |
| COISS_1004 | co-iss-n1356864777 | 31567506.48 | 31566702.471 |
| COISS_1004 | co-iss-n1356975912 | 31679022.099 | 31678229.579 |
| COISS_1004 | co-iss-w1357431545 | 32109327.67 | 32108502.242 |
| COISS_1004 | co-iss-n1357431653 | 32109393.818 | 32108575.733 |
| COISS_1004 | co-iss-w1357431786 | 32109505.133 | 32108678.177 |
| COISS_1005 | co-iss-n1357863387 | 32564818.116 | 32564001.064 |
| COISS_1005 | co-iss-w1357971535 | 32661833.307 | 32661048.098 |
| COISS_1005 | co-iss-n1358245168 | 33002132.033 | 33001347.082 |
| COISS_1005 | co-iss-w1358663405 | 33363658.981 | 33362851.696 |
| COISS_1005 | co-iss-n1358663623 | 33363788.801 | 33362981.883 |
| COISS_1005 | co-iss-n1358927780 | 33662603.336 | 33661821.067 |
| COISS_1005 | co-iss-n1358927964 | 33662771.458 | 33661974.278 |
| COISS_1005 | co-iss-n1359070782 | 33806348.667 | 33805549.494 |
| COISS_1005 | co-iss-n1359071373 | 33806525.275 | 33805733.94 |
| COISS_1006 | co-iss-w1359502060 | 34237307.461 | 34236508.77 |
| COISS_1006 | co-iss-w1360363027 | 35098798.924 | 35097997.179 |
| COISS_1006 | co-iss-n1360937901 | 35673953.939 | 35673153.317 |
| COISS_1006 | co-iss-n1361226923 | 35959092.187 | 35958289.342 |
| COISS_1006 | co-iss-n1361228653 | 35959334.752 | 35958532.527 |
| COISS_1006 | co-iss-n1361368738 | 36104841.363 | 36104042.417 |
| COISS_1006 | co-iss-n1361655926 | 36389658.751 | 36388856.819 |
| COISS_1006 | co-iss-n1361657689 | 36389937.778 | 36389135.955 |
| COISS_1007 | co-iss-n1364541189 | 39209911.435 | 39209105.791 |
| COISS_1007 | co-iss-n1364541321 | 39210006.646 | 39209200.988 |
| COISS_1007 | co-iss-n1364541420 | 39210071.276 | 39209259.072 |
| COISS_1007 | co-iss-n1364541584 | 39210164.402 | 39209359.088 |
| COISS_1007 | co-iss-n1364541716 | 39210261.13 | 39209455.666 |
| COISS_1007 | co-iss-n1364541749 | 39210285.335 | 39209479.712 |
| COISS_1007 | co-iss-n1364541979 | 39210419.813 | 39209614.705 |
| COISS_1007 | co-iss-n1404856096 | 79539950.668 | 79539318.142 |
| COISS_1008 | co-iss-n1412619196 | 87379169.361 | 87378397.73 |
| COISS_1008 | co-iss-n1412906634 | 87721828.069 | 87721071.006 |
| COISS_1008 | co-iss-n1412908730 | 87722239.197 | 87721495.891 |
| COISS_1008 | co-iss-n1412909266 | 87722377.528 | 87721628.333 |
| COISS_1008 | co-iss-n1412909802 | 87722510.979 | 87721760.81 |
| COISS_1008 | co-iss-n1413902027 | 88583979.532 | 88583249.968 |
| COISS_1008 | co-iss-n1413902105 | 88584085.201 | 88583373.068 |
| COISS_1008 | co-iss-n1421686131 | 96442078.492 | 96441309.157 |
| COISS_1008 | co-iss-n1421838870 | 96548946.531 | 96548179.089 |
| COISS_1008 | co-iss-w1429505174 | 104435821.548 | 104287270.587 |
+------------+--------------------+---------------+---------------+
```
| priority | coiss earth received start stop time are sometimes backwards from pds webserver created by seti pds webserver from pds opus created by seti pds opus mysql select volume id opus id from obs mission cassini where volume id opus id coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss coiss co iss | 1 |
244,585 | 7,877,023,698 | IssuesEvent | 2018-06-26 04:48:12 | commonality/getting-started-inner-source | https://api.github.com/repos/commonality/getting-started-inner-source | opened | docs(developer-guide): create ERB templates | priority: medium status: accepted status: available type: docs type: feature | ## 1. User story: summary of expected behavior
> ⌦ Describe what you want to accomplish, in what role/capacity,
> ⌦ and why it's important to you.
As a potential contributor,
I want source code development guidelines
In order to increase the likelihood of contributing changes.
## 2. Acceptance criteria
_We'll be done when:_
- [ ] 1. A succinct README provides:
- [ ] 1.1. Recommended skills
- [ ] 1.2. Development software requirements
- [ ] 1.3. Source code retrieval instructions
- [ ] 1.4. Dependency installation and update instructions
- [ ] 1.5. Build instructions
- [ ] 1.6. Test execution instructions
- [ ] 1.7. Source code style guidelines
- [ ] 1.8. DevSecOps overview
- [ ] 2. A debugging doc with set up and execution instructions exists
- [ ] 3. A document with API design, security, and documentation guidelines exists
- [ ] 4. A document exists covering Architecture Decision Records (ADR): process, submission, and maintainer instructions
- [ ] 5. A copy of (or link to) Codacy's _Software metrics: a practical guide for the curious developer_ PDF exists.
## 3. Notes
....
<!-- ⛔️ Do not remove this comment or anything below it. ⛔️ -->
[icon-info-image]: ../docs/img/icons8/icon-info-50.png | 1.0 | docs(developer-guide): create ERB templates - ## 1. User story: summary of expected behavior
> ⌦ Describe what you want to accomplish, in what role/capacity,
> ⌦ and why it's important to you.
As a potential contributor,
I want source code development guidelines
In order to increase the likelihood of contributing changes.
## 2. Acceptance criteria
_We'll be done when:_
- [ ] 1. A succinct README provides:
- [ ] 1.1. Recommended skills
- [ ] 1.2. Development software requirements
- [ ] 1.3. Source code retrieval instructions
- [ ] 1.4. Dependency installation and update instructions
- [ ] 1.5. Build instructions
- [ ] 1.6. Test execution instructions
- [ ] 1.7. Source code style guidelines
- [ ] 1.8. DevSecOps overview
- [ ] 2. A debugging doc with set up and execution instructions exists
- [ ] 3. A document with API design, security, and documentation guidelines exists
- [ ] 4. A document exists covering Architecture Decision Records (ADR): process, submission, and maintainer instructions
- [ ] 5. A copy of (or link to) Codacy's _Software metrics: a practical guide for the curious developer_ PDF exists.
## 3. Notes
....
<!-- ⛔️ Do not remove this comment or anything below it. ⛔️ -->
[icon-info-image]: ../docs/img/icons8/icon-info-50.png | priority | docs developer guide create erb templates user story summary of expected behavior ⌦ describe what you want to accomplish in what role capacity ⌦ and why it s important to you as a potential contributor i want source code development guidelines in order to increase the likelihood of contributing changes acceptance criteria we ll be done when a succinct readme provides recommended skills development software requirements source code retrieval instructions dependency installation and update instructions build instructions test execution instructions source code style guidelines devsecops overview a debugging doc with set up and execution instructions exists a document with api design security and documentation guidelines exists a document exists covering architecture decision records adr process submission and maintainer instructions a copy of or link to codacy s software metrics a practical guide for the curious developer pdf exists notes docs img icon info png | 1 |
175,149 | 6,547,619,233 | IssuesEvent | 2017-09-04 15:36:03 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [craftercms] Windows Bundle: Stopping leaves mysqld.exe running | bug Priority: Medium | Stopping the windows bundle using:
* `.\gradlew.bat stop`
Leaves the mysqld.exe running which causes this error:

| 1.0 | [craftercms] Windows Bundle: Stopping leaves mysqld.exe running - Stopping the windows bundle using:
* `.\gradlew.bat stop`
Leaves the mysqld.exe running which causes this error:

| priority | windows bundle stopping leaves mysqld exe running stopping the windows bundle using gradlew bat stop leaves the mysqld exe running which causes this error | 1 |
32,701 | 2,759,090,446 | IssuesEvent | 2015-04-28 00:02:13 | ai7/sagetv-for-plexmediacenter | https://api.github.com/repos/ai7/sagetv-for-plexmediacenter | closed | Plex Ap | auto-migrated Priority-Medium Type-Task | ```
Research plugin. Can we bundle the scanner and agent as a plugin and make it
available on the plex aps?
```
Original issue reported on code.google.com by `Mreid910` on 6 Jan 2012 at 3:12 | 1.0 | Plex Ap - ```
Research plugin. Can we bundle the scanner and agent as a plugin and make it
available on the plex aps?
```
Original issue reported on code.google.com by `Mreid910` on 6 Jan 2012 at 3:12 | priority | plex ap research plugin can we bundle the scanner and agent as a plugin and make it available on the plex aps original issue reported on code google com by on jan at | 1 |
135,581 | 5,254,850,614 | IssuesEvent | 2017-02-02 14:08:05 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | остатки склада в розничных ценах | Depot Garant Priority-Medium Type-Task | Originally reported on Google Code with ID 1662
```
1. Создать отчет – «остатки склада в розничных ценах по
подразделениям».
2. Создать отчет – «остатки склада в розничных ценах с выбором
подразделения».
3. Создать отчет – «остатки склада в розничных ценах с разбивкой по
товарным группам».
4. Создать отчет – «остатки склада в розничных ценах с выбором
товарной группы».
```
Reported by `alexandra.gsoftware` on 2009-10-09 14:42:22
<hr>
- _Attachment: [Остатки в розничных ценах.doc](https://storage.googleapis.com/google-code-attachments/gedemin/issue-1662/comment-0/Остатки в розничных ценах.doc)_
| 1.0 | остатки склада в розничных ценах - Originally reported on Google Code with ID 1662
```
1. Создать отчет – «остатки склада в розничных ценах по
подразделениям».
2. Создать отчет – «остатки склада в розничных ценах с выбором
подразделения».
3. Создать отчет – «остатки склада в розничных ценах с разбивкой по
товарным группам».
4. Создать отчет – «остатки склада в розничных ценах с выбором
товарной группы».
```
Reported by `alexandra.gsoftware` on 2009-10-09 14:42:22
<hr>
- _Attachment: [Остатки в розничных ценах.doc](https://storage.googleapis.com/google-code-attachments/gedemin/issue-1662/comment-0/Остатки в розничных ценах.doc)_
| priority | остатки склада в розничных ценах originally reported on google code with id создать отчет – «остатки склада в розничных ценах по подразделениям» создать отчет – «остатки склада в розничных ценах с выбором подразделения» создать отчет – «остатки склада в розничных ценах с разбивкой по товарным группам» создать отчет – «остатки склада в розничных ценах с выбором товарной группы» reported by alexandra gsoftware on attachment в розничных ценах doc | 1 |
801,266 | 28,482,051,808 | IssuesEvent | 2023-04-18 04:06:03 | CodeforHawaii/HIERR | https://api.github.com/repos/CodeforHawaii/HIERR | closed | Any web pages accessed without being logged in should redirect the browser to the home page | Medium Priority | I found in local testing that accessing some of the non-home pages in my browser (/zipcode, /survey) are return 200 even without being logged in. This shouldn't be allowed.
Any non-home page should redirect the user to the home page to login first (when accessed with a browser - probably can look at Accept header to determine this).
Acceptance criteria:
* Log out or remove all cookies for HIERR site
* Access /zipcode (and /survey and all other endpoints meant to be viewed with a browser)
* Verify that 302 response is sent with a Location header of / (home page) | 1.0 | Any web pages accessed without being logged in should redirect the browser to the home page - I found in local testing that accessing some of the non-home pages in my browser (/zipcode, /survey) are return 200 even without being logged in. This shouldn't be allowed.
Any non-home page should redirect the user to the home page to login first (when accessed with a browser - probably can look at Accept header to determine this).
Acceptance criteria:
* Log out or remove all cookies for HIERR site
* Access /zipcode (and /survey and all other endpoints meant to be viewed with a browser)
* Verify that 302 response is sent with a Location header of / (home page) | priority | any web pages accessed without being logged in should redirect the browser to the home page i found in local testing that accessing some of the non home pages in my browser zipcode survey are return even without being logged in this shouldn t be allowed any non home page should redirect the user to the home page to login first when accessed with a browser probably can look at accept header to determine this acceptance criteria log out or remove all cookies for hierr site access zipcode and survey and all other endpoints meant to be viewed with a browser verify that response is sent with a location header of home page | 1 |
698,693 | 23,989,312,247 | IssuesEvent | 2022-09-13 22:34:20 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] get_clients.sh should use curl instead of wget | kind/bug area/ysql priority/medium | Jira Link: [DB-613](https://yugabyte.atlassian.net/browse/DB-613)
### Description
To connect to a Yugabyte Cluster using YugabyteDB Client Shell..., the get_clients.sh script requires wget . Not everyone has package manager or xcode, so the wget on Mac OSX fails
@chirag-yb | 1.0 | [YSQL] get_clients.sh should use curl instead of wget - Jira Link: [DB-613](https://yugabyte.atlassian.net/browse/DB-613)
### Description
To connect to a Yugabyte Cluster using YugabyteDB Client Shell..., the get_clients.sh script requires wget . Not everyone has package manager or xcode, so the wget on Mac OSX fails
@chirag-yb | priority | get clients sh should use curl instead of wget jira link description to connect to a yugabyte cluster using yugabytedb client shell the get clients sh script requires wget not everyone has package manager or xcode so the wget on mac osx fails chirag yb | 1 |
582,890 | 17,373,037,792 | IssuesEvent | 2021-07-30 16:28:49 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | the applications deployed under the project are not deleted when the project is deleted | area/app-management kind/bug priority/medium stale | **Describe the Bug**
After the project is deleted, the applications deployed under the project are not deleted
After deleting the project p1, see the application under the project still exists:

/kind bug
/assign @xyz-li
/priority High
/milestone 3.1.0
/area app-management | 1.0 | the applications deployed under the project are not deleted when the project is deleted - **Describe the Bug**
After the project is deleted, the applications deployed under the project are not deleted
After deleting the project p1, see the application under the project still exists:

/kind bug
/assign @xyz-li
/priority High
/milestone 3.1.0
/area app-management | priority | the applications deployed under the project are not deleted when the project is deleted describe the bug after the project is deleted the applications deployed under the project are not deleted after deleting the project see the application under the project still exists kind bug assign xyz li priority high milestone area app management | 1 |
292,947 | 8,971,102,338 | IssuesEvent | 2019-01-29 15:12:01 | xlayers/xlayers | https://api.github.com/repos/xlayers/xlayers | closed | add documentation to editor/viewer/lib | Priority: High effort2: medium (days) type: docs | The core viewer library and the parsers need developer documentation.
- [x] Let's add create a wiki and make sure the API is well documented.
- [x] add a detailed architecture diagram of the core parsing process | 1.0 | add documentation to editor/viewer/lib - The core viewer library and the parsers need developer documentation.
- [x] Let's add create a wiki and make sure the API is well documented.
- [x] add a detailed architecture diagram of the core parsing process | priority | add documentation to editor viewer lib the core viewer library and the parsers need developer documentation let s add create a wiki and make sure the api is well documented add a detailed architecture diagram of the core parsing process | 1 |
826,535 | 31,652,287,541 | IssuesEvent | 2023-09-07 00:05:26 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | opened | Destructible needs moving to shared. | Priority: 3-Not Required Issue: Needs Cleanup Difficulty: 2-Medium | At least for gathering, once we have predicted entity deletion. | 1.0 | Destructible needs moving to shared. - At least for gathering, once we have predicted entity deletion. | priority | destructible needs moving to shared at least for gathering once we have predicted entity deletion | 1 |
487,229 | 14,020,975,777 | IssuesEvent | 2020-10-29 20:31:42 | Novik/ruTorrent | https://api.github.com/repos/Novik/ruTorrent | closed | Feature Request: Change all (or multiple) filters to a different feed | Priority-Medium enhancement imported wontfix | _From [Wonslung@gmail.com](https://code.google.com/u/Wonslung@gmail.com/) on October 16, 2009 05:32:11_
I would find it VERY useful to change filters in bulk to another feed. I tend to switch mine often to
build ratio on different sites, or when i decide a newer site is better for some reason and i often
change them all
or most of them....
_Original issue: http://code.google.com/p/rutorrent/issues/detail?id=127_
| 1.0 | Feature Request: Change all (or multiple) filters to a different feed - _From [Wonslung@gmail.com](https://code.google.com/u/Wonslung@gmail.com/) on October 16, 2009 05:32:11_
I would find it VERY useful to change filters in bulk to another feed. I tend to switch mine often to
build ratio on different sites, or when i decide a newer site is better for some reason and i often
change them all
or most of them....
_Original issue: http://code.google.com/p/rutorrent/issues/detail?id=127_
| priority | feature request change all or multiple filters to a different feed from on october i would find it very useful to change filters in bulk to another feed i tend to switch mine often to build ratio on different sites or when i decide a newer site is better for some reason and i often change them all or most of them original issue | 1 |
757,698 | 26,524,788,827 | IssuesEvent | 2023-01-19 07:44:11 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | Winterfall Runner - Pathing Underground | NPC Pathfinding Priority: Medium Status: Confirmed | [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
There is a Winterfall Runner who is totally underground in Winterspring. You should be able to see it if you track humanoids in Winterspring between
It paths between 40.35, 52.08 and 44.60, 49.13 (possibly goes further than this but I stopped following)
**How to reproduce:**
Go to Winterspring 40, 52 and track Humanoids on a Hunter (I'd imagine this will be the easiest way to find the mob). I've seen it a couple of times just when running down the path so I know its a reccuring issue.
**How it should work:**
Mob should be above ground (or potentially not even there at all)
**Database links:**
https://wotlk-twinhead.twinstar.cz/?npc=10916

| 1.0 | Winterfall Runner - Pathing Underground - [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
There is a Winterfall Runner who is totally underground in Winterspring. You should be able to see it if you track humanoids in Winterspring between
It paths between 40.35, 52.08 and 44.60, 49.13 (possibly goes further than this but I stopped following)
**How to reproduce:**
Go to Winterspring 40, 52 and track Humanoids on a Hunter (I'd imagine this will be the easiest way to find the mob). I've seen it a couple of times just when running down the path so I know its a reccuring issue.
**How it should work:**
Mob should be above ground (or potentially not even there at all)
**Database links:**
https://wotlk-twinhead.twinstar.cz/?npc=10916

| priority | winterfall runner pathing underground rembember add links to things related to the bug using for example cata twinhead twinstar cz description there is a winterfall runner who is totally underground in winterspring you should be able to see it if you track humanoids in winterspring between it paths between and possibly goes further than this but i stopped following how to reproduce go to winterspring and track humanoids on a hunter i d imagine this will be the easiest way to find the mob i ve seen it a couple of times just when running down the path so i know its a reccuring issue how it should work mob should be above ground or potentially not even there at all database links | 1 |
158,666 | 6,033,299,211 | IssuesEvent | 2017-06-09 07:53:25 | Caleydo/taggle | https://api.github.com/repos/Caleydo/taggle | opened | Remove column button is unnecessary when matrix is added | medium priority question | The circled button is not working, it is possible to remove the matrix only in its upper filter panel. Another question is - is it really necessary to have two identical histograms for one matrix in the side panel?

| 1.0 | Remove column button is unnecessary when matrix is added - The circled button is not working, it is possible to remove the matrix only in its upper filter panel. Another question is - is it really necessary to have two identical histograms for one matrix in the side panel?

| priority | remove column button is unnecessary when matrix is added the circled button is not working it is possible to remove the matrix only in its upper filter panel another question is is it really necessary to have two identical histograms for one matrix in the side panel | 1 |
113,280 | 4,545,088,565 | IssuesEvent | 2016-09-11 02:23:03 | 4-20ma/ModbusMaster | https://api.github.com/repos/4-20ma/ModbusMaster | closed | Add continuous integration testing with travis | Priority: Medium Status: In Progress Type: Feature Request | <!----------------------------------------------------------------------------
Title - ensure the issue title is clear & concise
- QUESTIONS - describe the specific question
- BUG REPORTS - describe an activity
- FEATURE REQUESTS - describe an activity
-->
<!----------------------------------------------------------------------------
Provide the following information for all issues. Replace [brackets] and placeholder text with your responses.
(QUESTIONS, BUG REPORTS, FEATURE REQUESTS)
-->
### ModbusMaster version
0.11.0
### Arduino IDE version
1.6
### Arduino Hardware
all
### Platform Details
any
---
<!----------------------------------------------------------------------------
Provide the following for FEATURE REQUESTS. Replace [brackets] and placeholder text with your responses.
Refer to [What's in a Story?](https://dannorth.net/whats-in-a-story/)
-->
### Feature Request
#### Narrative:
<!-- Replace role, feature, benefit. -->
As a project owner
I want all changes to compile without error
So that end users have a working library
#### Acceptance Criteria:
<!--
Present as one or more Scenarios, replacing context, event, outcome.
-->
Scenario 1: Build is successful
Given a user submits a pull request
When it compiles without error on the automated ci tool
Then it is flagged as passing the build test
Scenario 1: Build fails
Given a user submits a pull request
When the build fails on the automated ci tool
Then it is flagged as failing the build test
| 1.0 | Add continuous integration testing with travis - <!----------------------------------------------------------------------------
Title - ensure the issue title is clear & concise
- QUESTIONS - describe the specific question
- BUG REPORTS - describe an activity
- FEATURE REQUESTS - describe an activity
-->
<!----------------------------------------------------------------------------
Provide the following information for all issues. Replace [brackets] and placeholder text with your responses.
(QUESTIONS, BUG REPORTS, FEATURE REQUESTS)
-->
### ModbusMaster version
0.11.0
### Arduino IDE version
1.6
### Arduino Hardware
all
### Platform Details
any
---
<!----------------------------------------------------------------------------
Provide the following for FEATURE REQUESTS. Replace [brackets] and placeholder text with your responses.
Refer to [What's in a Story?](https://dannorth.net/whats-in-a-story/)
-->
### Feature Request
#### Narrative:
<!-- Replace role, feature, benefit. -->
As a project owner
I want all changes to compile without error
So that end users have a working library
#### Acceptance Criteria:
<!--
Present as one or more Scenarios, replacing context, event, outcome.
-->
Scenario 1: Build is successful
Given a user submits a pull request
When it compiles without error on the automated ci tool
Then it is flagged as passing the build test
Scenario 1: Build fails
Given a user submits a pull request
When the build fails on the automated ci tool
Then it is flagged as failing the build test
| priority | add continuous integration testing with travis title ensure the issue title is clear concise questions describe the specific question bug reports describe an activity feature requests describe an activity provide the following information for all issues replace and placeholder text with your responses questions bug reports feature requests modbusmaster version arduino ide version arduino hardware all platform details any provide the following for feature requests replace and placeholder text with your responses refer to feature request narrative as a project owner i want all changes to compile without error so that end users have a working library acceptance criteria present as one or more scenarios replacing context event outcome scenario build is successful given a user submits a pull request when it compiles without error on the automated ci tool then it is flagged as passing the build test scenario build fails given a user submits a pull request when the build fails on the automated ci tool then it is flagged as failing the build test | 1 |
24,379 | 2,667,383,904 | IssuesEvent | 2015-03-22 15:19:40 | NewCreature/EOF | https://api.github.com/repos/NewCreature/EOF | closed | Loading a song without vocals causes crash | bug imported Priority-Medium | _From [xander4j...@yahoo.com](https://code.google.com/u/111302640723734240985/) on June 09, 2010 20:03:50_
I haven't been able to confirm it is related to vocals but the only chart I can get to load is the one that has PART_VOCALS. Just a wild guess but I think it might be the vocals preview that displays in the 3D preview that is causing this issue.
_Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=115_ | 1.0 | Loading a song without vocals causes crash - _From [xander4j...@yahoo.com](https://code.google.com/u/111302640723734240985/) on June 09, 2010 20:03:50_
I haven't been able to confirm it is related to vocals but the only chart I can get to load is the one that has PART_VOCALS. Just a wild guess but I think it might be the vocals preview that displays in the 3D preview that is causing this issue.
_Original issue: http://code.google.com/p/editor-on-fire/issues/detail?id=115_ | priority | loading a song without vocals causes crash from on june i haven t been able to confirm it is related to vocals but the only chart i can get to load is the one that has part vocals just a wild guess but i think it might be the vocals preview that displays in the preview that is causing this issue original issue | 1 |
100,671 | 4,102,545,113 | IssuesEvent | 2016-06-04 02:55:16 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Markdown Cheat Sheet: Update Perl Module | Maintainer Input Requested Priority: Medium | Currently the metadata states that `DDG::Goodie::MarkdownCheatSheet` is the `perl_module`, we should change this to `DDG::Goodie::CheatSheets`.
/cc @moollaza @zekiel
------
IA Page: http://duck.co/ia/view/markdown_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @marianosimone | 1.0 | Markdown Cheat Sheet: Update Perl Module - Currently the metadata states that `DDG::Goodie::MarkdownCheatSheet` is the `perl_module`, we should change this to `DDG::Goodie::CheatSheets`.
/cc @moollaza @zekiel
------
IA Page: http://duck.co/ia/view/markdown_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @marianosimone | priority | markdown cheat sheet update perl module currently the metadata states that ddg goodie markdowncheatsheet is the perl module we should change this to ddg goodie cheatsheets cc moollaza zekiel ia page marianosimone | 1 |
193,762 | 6,887,816,514 | IssuesEvent | 2017-11-22 01:42:16 | minio/minio-go | https://api.github.com/repos/minio/minio-go | opened | gateway-gcs: testRemoveMultipleObjects functional test fails since gcs server cannot handle the requested number of files | priority: medium | When minio-go functional test `testRemoveMultipleObjects` is run against gateway-gcs, it creates 1100 files, each 8 bytes in size, and then tries to remove them. However, gcs server cannot handle the load and fails with `We encountered an internal error, please try again.` error message.
We just simply don't need to push the gcs server to the edge and use a much smaller number of files to test multiple object removal functionality. | 1.0 | gateway-gcs: testRemoveMultipleObjects functional test fails since gcs server cannot handle the requested number of files - When minio-go functional test `testRemoveMultipleObjects` is run against gateway-gcs, it creates 1100 files, each 8 bytes in size, and then tries to remove them. However, gcs server cannot handle the load and fails with `We encountered an internal error, please try again.` error message.
We just simply don't need to push the gcs server to the edge and use a much smaller number of files to test multiple object removal functionality. | priority | gateway gcs testremovemultipleobjects functional test fails since gcs server cannot handle the requested number of files when minio go functional test testremovemultipleobjects is run against gateway gcs it creates files each bytes in size and then tries to remove them however gcs server cannot handle the load and fails with we encountered an internal error please try again error message we just simply don t need to push the gcs server to the edge and use a much smaller number of files to test multiple object removal functionality | 1 |
465,436 | 13,385,760,382 | IssuesEvent | 2020-09-02 13:53:11 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | opened | Litigation Support New 90 Day Letter Task | Priority: Medium Product: caseflow-queue Stakeholder: BVA Team: Echo 🐬 Type: New Development | <!-- The goal of this template is to be a tool to communicate the requirements for a story related task. It is not intended as a mandate, adapt as needed. -->
## User or job story
User story: As a litigation support user I should have access to create admin actions on a 90 day letter task.
## Acceptance criteria
- [ ] Please put this work behind the feature toggle: CAVC_Remand
- [ ] This feature should be accessible to the following user groups: Litigation Support
- [ ] Include screenshot(s) in the Github issue if there are front-end changes
- [ ] Admin Actions is listed in the Actions drop down of a 90 day letter task
- [ ] The following actions are listed on the submit admin actions page:
- [ ] CUE-Related
- [ ] Change of Address
- [ ] Congressional Interest
- [ ] Controlled Correspondence
- [ ] Death Certificate
- [ ] Evidence of Argument
- [ ] Extension Request
- [ ] FOIA Request
- [ ] Hearing Related
- [ ] Motion for Reconsideration
- [ ] Motion to Advance on Docket
- [ ] Motion to Vacate
- [ ] Other Motion
- [ ] Power of Attorney-Related
- [ ] Privacy Act Request
- [ ] Privacy Complaint
- [ ] Returned of Undeliverable Mail
- [ ] Status Inquiry
- [ ] Withdrawal of Appeal
- [ ] Provide Instructions and context for this action appears on the submit admin actions page
## Release notes
<!-- Write what should be included in release notes (Caseflow uses Headway), updated when the story is built, before it's deployed. -->
<!-- The following sections can be deleted if they are not needed -->
### Out of scope
<!-- Clarify what is out of scope if the designs include more or there are many tickets for this chunk of work -->
### Designs
<!-- Include screenshots or links to designs if applicable. -->
### Background/context
<!-- Include as needed, especially for issues that aren't part of epics. Include a value statement - why is this feature being developed? -->
### Technical notes
<!-- Include notes that might help an engineer get started on this more quickly, or potential pitfalls to watch out for. -->
### Other notes
### Resources/other links
<!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
| 1.0 | Litigation Support New 90 Day Letter Task - <!-- The goal of this template is to be a tool to communicate the requirements for a story related task. It is not intended as a mandate, adapt as needed. -->
## User or job story
User story: As a litigation support user I should have access to create admin actions on a 90 day letter task.
## Acceptance criteria
- [ ] Please put this work behind the feature toggle: CAVC_Remand
- [ ] This feature should be accessible to the following user groups: Litigation Support
- [ ] Include screenshot(s) in the Github issue if there are front-end changes
- [ ] Admin Actions is listed in the Actions drop down of a 90 day letter task
- [ ] The following actions are listed on the submit admin actions page:
- [ ] CUE-Related
- [ ] Change of Address
- [ ] Congressional Interest
- [ ] Controlled Correspondence
- [ ] Death Certificate
- [ ] Evidence of Argument
- [ ] Extension Request
- [ ] FOIA Request
- [ ] Hearing Related
- [ ] Motion for Reconsideration
- [ ] Motion to Advance on Docket
- [ ] Motion to Vacate
- [ ] Other Motion
- [ ] Power of Attorney-Related
- [ ] Privacy Act Request
- [ ] Privacy Complaint
- [ ] Returned of Undeliverable Mail
- [ ] Status Inquiry
- [ ] Withdrawal of Appeal
- [ ] Provide Instructions and context for this action appears on the submit admin actions page
## Release notes
<!-- Write what should be included in release notes (Caseflow uses Headway), updated when the story is built, before it's deployed. -->
<!-- The following sections can be deleted if they are not needed -->
### Out of scope
<!-- Clarify what is out of scope if the designs include more or there are many tickets for this chunk of work -->
### Designs
<!-- Include screenshots or links to designs if applicable. -->
### Background/context
<!-- Include as needed, especially for issues that aren't part of epics. Include a value statement - why is this feature being developed? -->
### Technical notes
<!-- Include notes that might help an engineer get started on this more quickly, or potential pitfalls to watch out for. -->
### Other notes
### Resources/other links
<!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
| priority | litigation support new day letter task user or job story user story as a litigation support user i should have access to create admin actions on a day letter task acceptance criteria please put this work behind the feature toggle cavc remand this feature should be accessible to the following user groups litigation support include screenshot s in the github issue if there are front end changes admin actions is listed in the actions drop down of a day letter task the following actions are listed on the submit admin actions page cue related change of address congressional interest controlled correspondence death certificate evidence of argument extension request foia request hearing related motion for reconsideration motion to advance on docket motion to vacate other motion power of attorney related privacy act request privacy complaint returned of undeliverable mail status inquiry withdrawal of appeal provide instructions and context for this action appears on the submit admin actions page release notes out of scope designs background context technical notes other notes resources other links | 1 |
402,882 | 11,826,022,302 | IssuesEvent | 2020-03-21 15:51:15 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | Orphan xmp | bug: pending priority: medium scope: OS support understood: clear | Copy an image file into a folder, import it to dt.
Take it into darkroom and do some trivial editing.
Use selected images to send the image to trash.
Verify in the folder that the dng is gone but the xmp remains. Also, the folder hasn't been deleted because it still contains the xmp.
Expected behaviour: the xmp should be deleted at the same time as the dng, and the folder should then be deleted since it's empty.
dt 3.1.0+547 gb447df7c8
Built via msys64
Windows 10pro build 18362, 19h1 release 190318-1202
| 1.0 | Orphan xmp - Copy an image file into a folder, import it to dt.
Take it into darkroom and do some trivial editing.
Use selected images to send the image to trash.
Verify in the folder that the dng is gone but the xmp remains. Also, the folder hasn't been deleted because it still contains the xmp.
Expected behaviour: the xmp should be deleted at the same time as the dng, and the folder should then be deleted since it's empty.
dt 3.1.0+547 gb447df7c8
Built via msys64
Windows 10pro build 18362, 19h1 release 190318-1202
| priority | orphan xmp copy an image file into a folder import it to dt take it into darkroom and do some trivial editing use selected images to send the image to trash verify in the folder that the dng is gone but the xmp remains also the folder hasn t been deleted because it still contains the xmp expected behaviour the xmp should be deleted at the same time as the dng and the folder should then be deleted since it s empty dt built via windows build release | 1 |
567,548 | 16,870,495,251 | IssuesEvent | 2021-06-22 03:27:54 | microbiomedata/nmdc-server | https://api.github.com/repos/microbiomedata/nmdc-server | closed | Portal Download - bulk | In Progress X LARGE priority: medium | Researchers who are interested in raw data generally want to do bulk downloads (to a server/cloud) for custom analyses. Researchers who are interested in data products may want to download to a server or to their local computer for custom analysis or sending to KBase. Researchers are also interested in downloading the associated sample metadata only.
Priority - medium
Urgency - low | 1.0 | Portal Download - bulk - Researchers who are interested in raw data generally want to do bulk downloads (to a server/cloud) for custom analyses. Researchers who are interested in data products may want to download to a server or to their local computer for custom analysis or sending to KBase. Researchers are also interested in downloading the associated sample metadata only.
Priority - medium
Urgency - low | priority | portal download bulk researchers who are interested in raw data generally want to do bulk downloads to a server cloud for custom analyses researchers who are interested in data products may want to download to a server or to their local computer for custom analysis or sending to kbase researchers are also interested in downloading the associated sample metadata only priority medium urgency low | 1 |
541,977 | 15,836,844,204 | IssuesEvent | 2021-04-06 19:55:20 | light-town/api | https://api.github.com/repos/light-town/api | closed | Add vaults | priority: medium (3) scope: core status: done 👏 | - [x] Make encrypt and decrypt vault with key sets
- [x] Add functional to set vault name
- [x] Add functional to set vault description | 1.0 | Add vaults - - [x] Make encrypt and decrypt vault with key sets
- [x] Add functional to set vault name
- [x] Add functional to set vault description | priority | add vaults make encrypt and decrypt vault with key sets add functional to set vault name add functional to set vault description | 1 |
384,679 | 11,396,438,870 | IssuesEvent | 2020-01-30 13:35:00 | opentargets/platform | https://api.github.com/repos/opentargets/platform | closed | Harmonize ontology mapping process | Kind: Maintenance Priority: Medium Topic: Ontology | In several places, we have subtly different processes for ontology mapping from string (e.g. OnToma vs evidence_datasource_parsers).
These should be harmonised into one place for consistency, and maintainability. | 1.0 | Harmonize ontology mapping process - In several places, we have subtly different processes for ontology mapping from string (e.g. OnToma vs evidence_datasource_parsers).
These should be harmonised into one place for consistency, and maintainability. | priority | harmonize ontology mapping process in several places we have subtly different processes for ontology mapping from string e g ontoma vs evidence datasource parsers these should be harmonised into one place for consistency and maintainability | 1 |
444,349 | 12,810,316,233 | IssuesEvent | 2020-07-03 18:13:20 | crcn/paperclip | https://api.github.com/repos/crcn/paperclip | closed | virt:Text needs AST location | area: rust engine bug estimate: 2+ days priority: medium | This is causing errors when meta + click is pressed on text elements in the preview.
Requires location to also be added to evaluated JS values | 1.0 | virt:Text needs AST location - This is causing errors when meta + click is pressed on text elements in the preview.
Requires location to also be added to evaluated JS values | priority | virt text needs ast location this is causing errors when meta click is pressed on text elements in the preview requires location to also be added to evaluated js values | 1 |
282,489 | 8,706,945,992 | IssuesEvent | 2018-12-06 05:34:33 | qhacks/qhacks-dashboard | https://api.github.com/repos/qhacks/qhacks-dashboard | closed | Add offset/limit pagination to endpoints requiring it | package: server priority: required (medium) | ## Problem
Currently our endpoints that return lists of data don't have pagination functionality, this would be great to add as it is causing a huge performance impact. The pagination on the frontend is purely for user the interface and is very buggy.
@joeytepp @rosslh do you think limit/offset pagination will suffice or should we implement cursor based pagination? Let's discuss here.
| 1.0 | Add offset/limit pagination to endpoints requiring it - ## Problem
Currently our endpoints that return lists of data don't have pagination functionality, this would be great to add as it is causing a huge performance impact. The pagination on the frontend is purely for user the interface and is very buggy.
@joeytepp @rosslh do you think limit/offset pagination will suffice or should we implement cursor based pagination? Let's discuss here.
| priority | add offset limit pagination to endpoints requiring it problem currently our endpoints that return lists of data don t have pagination functionality this would be great to add as it is causing a huge performance impact the pagination on the frontend is purely for user the interface and is very buggy joeytepp rosslh do you think limit offset pagination will suffice or should we implement cursor based pagination let s discuss here | 1 |
17,153 | 2,615,132,248 | IssuesEvent | 2015-03-01 06:02:26 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | MediaHttpUploader.serverErrorCallback() does not work if data has been partially or fully uploaded | auto-migrated Component-Media Milestone-Version1.16.0 Priority-Medium Type-Enhancement | ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
1.14.1-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
All
Describe the problem.
See:
http://javadoc.google-api-java-client.googlecode.com/hg/1.14.1-beta/com/google/a
pi/client/googleapis/media/MediaHttpUploader.html#serverErrorCallback()
I added a test in
MediaHttpUploaderTest.testUpload_ResumableIOExceptionWithIOExceptionHandler to
demonstrate what happens if data has been partially or fully uploaded despite
the I/O exception.
I can actually reproduce this running against the production server. I
modified the drive-cmdline-sample by adding an HttpRequestInitializer that sets
request.setReadTimeout(100) on the first upload call. It actually uploads the
whole thing, but it still times out and throws an I/O exception. It then
either throws an IllegalArgumentException or IllegalStateException depending on
whether we know the content length.
While you are at it, we also need to deal with the possibility of the server
error callback returning an OK response with the JSON metadata. At least the
documentation claims it is possible, but we should try to verify by testing it
against the production server.
How would you expect it to be fixed?
This is potentially non-trivial because we also need to be sensitive to the
chunk size. My understanding is that the Google server requires the chunk size
to be a certain multiple of 256KB. Alternatively, we may want to respect the
chunk size that the developer requested. There is also the case of fully
uploaded, in which case we need to just continue with the next chunk entirely.
```
Original issue reported on code.google.com by `yan...@google.com` on 23 Apr 2013 at 2:58 | 1.0 | MediaHttpUploader.serverErrorCallback() does not work if data has been partially or fully uploaded - ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
1.14.1-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
All
Describe the problem.
See:
http://javadoc.google-api-java-client.googlecode.com/hg/1.14.1-beta/com/google/a
pi/client/googleapis/media/MediaHttpUploader.html#serverErrorCallback()
I added a test in
MediaHttpUploaderTest.testUpload_ResumableIOExceptionWithIOExceptionHandler to
demonstrate what happens if data has been partially or fully uploaded despite
the I/O exception.
I can actually reproduce this running against the production server. I
modified the drive-cmdline-sample by adding an HttpRequestInitializer that sets
request.setReadTimeout(100) on the first upload call. It actually uploads the
whole thing, but it still times out and throws an I/O exception. It then
either throws an IllegalArgumentException or IllegalStateException depending on
whether we know the content length.
While you are at it, we also need to deal with the possibility of the server
error callback returning an OK response with the JSON metadata. At least the
documentation claims it is possible, but we should try to verify by testing it
against the production server.
How would you expect it to be fixed?
This is potentially non-trivial because we also need to be sensitive to the
chunk size. My understanding is that the Google server requires the chunk size
to be a certain multiple of 256KB. Alternatively, we may want to respect the
chunk size that the developer requested. There is also the case of fully
uploaded, in which case we need to just continue with the next chunk entirely.
```
Original issue reported on code.google.com by `yan...@google.com` on 23 Apr 2013 at 2:58 | priority | mediahttpuploader servererrorcallback does not work if data has been partially or fully uploaded version of google api java client e g beta beta java environment e g java android app engine all describe the problem see pi client googleapis media mediahttpuploader html servererrorcallback i added a test in mediahttpuploadertest testupload resumableioexceptionwithioexceptionhandler to demonstrate what happens if data has been partially or fully uploaded despite the i o exception i can actually reproduce this running against the production server i modified the drive cmdline sample by adding an httprequestinitializer that sets request setreadtimeout on the first upload call it actually uploads the whole thing but it still times out and throws an i o exception it then either throws an illegalargumentexception or illegalstateexception depending on whether we know the content length while you are at it we also need to deal with the possibility of the server error callback returning an ok response with the json metadata at least the documentation claims it is possible but we should try to verify by testing it against the production server how would you expect it to be fixed this is potentially non trivial because we also need to be sensitive to the chunk size my understanding is that the google server requires the chunk size to be a certain multiple of alternatively we may want to respect the chunk size that the developer requested there is also the case of fully uploaded in which case we need to just continue with the next chunk entirely original issue reported on code google com by yan google com on apr at | 1 |
118,549 | 4,750,613,343 | IssuesEvent | 2016-10-22 12:46:50 | CS2103AUG2016-T09-C3/main | https://api.github.com/repos/CS2103AUG2016-T09-C3/main | closed | JUnit tests terminate after the help command test | priority.medium type.bug | Tests do not fail, but terminate after the help JUnit test. Please check. | 1.0 | JUnit tests terminate after the help command test - Tests do not fail, but terminate after the help JUnit test. Please check. | priority | junit tests terminate after the help command test tests do not fail but terminate after the help junit test please check | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.