Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
502,585 | 14,549,716,918 | IssuesEvent | 2020-12-16 04:34:55 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Tenant wise publisher config loading returns the resource type's files instead of a specific resource's file | Priority/Normal Severity/Minor bug | **Describe the issue:**
When loading tenant publisher configuration from ConfigStore, in `loadTenantPublisherConfigurationFromConfigStore` method, resource.getFiles return all the files related to **resource type** instead of a particular **resource's files**.
https://github.com/wso2-extensions/identity-governance/blob/9a04f338d343c412d3dae000f6502825a253f32e/components/org.wso2.carbon.identity.tenant.resource.manager/src/main/java/org/wso2/carbon/identity/tenant/resource/manager/TenantAwareAxis2ConfigurationContextObserver.java#L106
**Cause for the issue:**
Setting files to the resource is erroneous. This function sets the entire resource type's files.
https://github.com/wso2/carbon-identity-framework/blob/b0e78d597482c922503156a6834c753ab2f3f2f7/components/configuration-mgt/org.wso2.carbon.identity.configuration.mgt.core/src/main/java/org/wso2/carbon/identity/configuration/mgt/core/dao/impl/ConfigurationDAOImpl.java#L1531
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: IS 5.11.0
- OS: Linux
- Database: H2
- Userstore: LDAP
| 1.0 | Tenant wise publisher config loading returns the resource type's files instead of a specific resource's file - **Describe the issue:**
When loading tenant publisher configuration from ConfigStore, in `loadTenantPublisherConfigurationFromConfigStore` method, resource.getFiles return all the files related to **resource type** instead of a particular **resource's files**.
https://github.com/wso2-extensions/identity-governance/blob/9a04f338d343c412d3dae000f6502825a253f32e/components/org.wso2.carbon.identity.tenant.resource.manager/src/main/java/org/wso2/carbon/identity/tenant/resource/manager/TenantAwareAxis2ConfigurationContextObserver.java#L106
**Cause for the issue:**
Setting files to the resource is erroneous. This function sets the entire resource type's files.
https://github.com/wso2/carbon-identity-framework/blob/b0e78d597482c922503156a6834c753ab2f3f2f7/components/configuration-mgt/org.wso2.carbon.identity.configuration.mgt.core/src/main/java/org/wso2/carbon/identity/configuration/mgt/core/dao/impl/ConfigurationDAOImpl.java#L1531
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: IS 5.11.0
- OS: Linux
- Database: H2
- Userstore: LDAP
| priority | tenant wise publisher config loading returns the resource type s files instead of a specific resource s file describe the issue when loading tenant publisher configuration from configstore in loadtenantpublisherconfigurationfromconfigstore method resource getfiles return all the files related to resource type instead of a particular resource s files cause for the issue setting files to the resource is erroneous this function sets the entire resource type s files environment information please complete the following information remove any unnecessary fields product version is os linux database userstore ldap | 1 |
165,223 | 26,125,504,747 | IssuesEvent | 2022-12-28 17:55:53 | KitWallace/FO | https://api.github.com/repos/KitWallace/FO | closed | Profile parameters sb on sliders | redesign | Would need to hide/show these which could hide show the visualised profile too | 1.0 | Profile parameters sb on sliders - Would need to hide/show these which could hide show the visualised profile too | non_priority | profile parameters sb on sliders would need to hide show these which could hide show the visualised profile too | 0 |
797,575 | 28,148,744,726 | IssuesEvent | 2023-04-02 19:46:45 | Greenstand/treetracker-admin-client | https://api.github.com/repos/Greenstand/treetracker-admin-client | opened | Species not working in Captures or Capture Details | type: bug good first issue priority size: small tool: Captures | Related to #1017
This is a quick fix to get species information back into the Captures page, filter and Capture Detail dialog.
Not working:
* Species filter in _Captures_ tool
* Species column in _Captures_ tool
* Species field in Capture Detail dialog
The new Query API expects a UUID for the species, available as `species.uuid` in the legacy Species data.
In the future, species information will be part of a dedicated schema, indexed by UUID. | 1.0 | Species not working in Captures or Capture Details - Related to #1017
This is a quick fix to get species information back into the Captures page, filter and Capture Detail dialog.
Not working:
* Species filter in _Captures_ tool
* Species column in _Captures_ tool
* Species field in Capture Detail dialog
The new Query API expects a UUID for the species, available as `species.uuid` in the legacy Species data.
In the future, species information will be part of a dedicated schema, indexed by UUID. | priority | species not working in captures or capture details related to this is a quick fix to get species information back into the captures page filter and capture detail dialog not working species filter in captures tool species column in captures tool species field in capture detail dialog the new query api expects a uuid for the species available as species uuid in the legacy species data in the future species information will be part of a dedicated schema indexed by uuid | 1 |
74,663 | 7,435,012,627 | IssuesEvent | 2018-03-26 13:01:47 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | closed | InvocationTargetException with -Xshareclasses from Spring App | test failure userRaised | Running the petclinic sample spring app (https://github.com/spring-projects/spring-petclinic) with `-Xscmx50M -Xshareclasses` causes `Exception in thread "main" java.lang.reflect.InvocationTargetException` on second run.
Output:
`2018-02-22 12:46:31.942 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : Starting PetClinicApplication v2.0.0 on 85de4731ae7f with PID 1 (/tmp/spring-petclinic-2.0.0.jar started by root in /)
2018-02-22 12:46:31.957 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : No active profile set, falling back to default profiles: default
2018-02-22 12:46:32.082 INFO 1 --- [ main] ConfigServletWebServerApplicationContext : Refreshing org.springframework.boot.web.servlet.context.AnnotationConfigServletWebServerApplicationContext@44fd0272: startup date [Thu Feb 22 12:46:32 UTC 2018]; root of context hierarchy
2018-02-22 12:46:35.192 INFO 1 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration$$EnhancerBySpringCGLIB$$1cb4efe6] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
Exception in thread "main" java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:87)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:50)
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:51)
Caused by: java.lang.NoClassDefFoundError: ch.qos.logback.classic.spi.ThrowableProxy
at ch.qos.logback.classic.spi.LoggingEvent.<init>(LoggingEvent.java:119)
at ch.qos.logback.classic.Logger.buildLoggingEventAndAppend(Logger.java:419)
at ch.qos.logback.classic.Logger.filterAndLog_0_Or3Plus(Logger.java:383)
at ch.qos.logback.classic.Logger.log(Logger.java:765)
at org.apache.logging.slf4j.SLF4JLogger.logMessage(SLF4JLogger.java:232)
at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2163)
at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2118)
at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2101)
at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:1995)
at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1967)
at org.apache.commons.logging.LogFactory$Log4jLog.log(LogFactory.java:302)
at org.apache.commons.logging.LogFactory$Log4jLog.error(LogFactory.java:254)
at org.springframework.boot.SpringApplication.reportFailure(SpringApplication.java:833)
at org.springframework.boot.SpringApplication.handleRunFailure(SpringApplication.java:807)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:338)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1246)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1234)
at org.springframework.samples.petclinic.PetClinicApplication.main(PetClinicApplication.java:32)
... 8 more
Caused by: java.lang.ClassNotFoundException: ch.qos.logback.classic.spi.ThrowableProxy
at java.net.URLClassLoader.findClass(URLClassLoader.java:588)
at java.lang.ClassLoader.loadClassHelper(ClassLoader.java:933)
at java.lang.ClassLoader.loadClass(ClassLoader.java:878)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
at java.lang.ClassLoader.loadClass(ClassLoader.java:861)
... 26 more
`
| 1.0 | InvocationTargetException with -Xshareclasses from Spring App - Running the petclinic sample spring app (https://github.com/spring-projects/spring-petclinic) with `-Xscmx50M -Xshareclasses` causes `Exception in thread "main" java.lang.reflect.InvocationTargetException` on second run.
Output:
`2018-02-22 12:46:31.942 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : Starting PetClinicApplication v2.0.0 on 85de4731ae7f with PID 1 (/tmp/spring-petclinic-2.0.0.jar started by root in /)
2018-02-22 12:46:31.957 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : No active profile set, falling back to default profiles: default
2018-02-22 12:46:32.082 INFO 1 --- [ main] ConfigServletWebServerApplicationContext : Refreshing org.springframework.boot.web.servlet.context.AnnotationConfigServletWebServerApplicationContext@44fd0272: startup date [Thu Feb 22 12:46:32 UTC 2018]; root of context hierarchy
2018-02-22 12:46:35.192 INFO 1 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration$$EnhancerBySpringCGLIB$$1cb4efe6] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
Exception in thread "main" java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:87)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:50)
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:51)
Caused by: java.lang.NoClassDefFoundError: ch.qos.logback.classic.spi.ThrowableProxy
at ch.qos.logback.classic.spi.LoggingEvent.<init>(LoggingEvent.java:119)
at ch.qos.logback.classic.Logger.buildLoggingEventAndAppend(Logger.java:419)
at ch.qos.logback.classic.Logger.filterAndLog_0_Or3Plus(Logger.java:383)
at ch.qos.logback.classic.Logger.log(Logger.java:765)
at org.apache.logging.slf4j.SLF4JLogger.logMessage(SLF4JLogger.java:232)
at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2163)
at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2118)
at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2101)
at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:1995)
at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1967)
at org.apache.commons.logging.LogFactory$Log4jLog.log(LogFactory.java:302)
at org.apache.commons.logging.LogFactory$Log4jLog.error(LogFactory.java:254)
at org.springframework.boot.SpringApplication.reportFailure(SpringApplication.java:833)
at org.springframework.boot.SpringApplication.handleRunFailure(SpringApplication.java:807)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:338)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1246)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1234)
at org.springframework.samples.petclinic.PetClinicApplication.main(PetClinicApplication.java:32)
... 8 more
Caused by: java.lang.ClassNotFoundException: ch.qos.logback.classic.spi.ThrowableProxy
at java.net.URLClassLoader.findClass(URLClassLoader.java:588)
at java.lang.ClassLoader.loadClassHelper(ClassLoader.java:933)
at java.lang.ClassLoader.loadClass(ClassLoader.java:878)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
at java.lang.ClassLoader.loadClass(ClassLoader.java:861)
... 26 more
`
| non_priority | invocationtargetexception with xshareclasses from spring app running the petclinic sample spring app with xshareclasses causes exception in thread main java lang reflect invocationtargetexception on second run output info o s s petclinic petclinicapplication starting petclinicapplication on with pid tmp spring petclinic jar started by root in info o s s petclinic petclinicapplication no active profile set falling back to default profiles default info configservletwebserverapplicationcontext refreshing org springframework boot web servlet context annotationconfigservletwebserverapplicationcontext startup date root of context hierarchy info trationdelegate beanpostprocessorchecker bean org springframework transaction annotation proxytransactionmanagementconfiguration of type is not eligible for getting processed by all beanpostprocessors for example not eligible for auto proxying exception in thread main java lang reflect invocationtargetexception at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org springframework boot loader mainmethodrunner run mainmethodrunner java at org springframework boot loader launcher launch launcher java at org springframework boot loader launcher launch launcher java at org springframework boot loader jarlauncher main jarlauncher java caused by java lang noclassdeffounderror ch qos logback classic spi throwableproxy at ch qos logback classic spi loggingevent loggingevent java at ch qos logback classic logger buildloggingeventandappend logger java at ch qos logback classic logger filterandlog logger java at ch qos logback classic logger log logger java at org apache logging logmessage java at org apache logging spi abstractlogger trylogmessage abstractlogger java at org apache logging spi abstractlogger logmessagetrackrecursion abstractlogger java at org apache logging spi abstractlogger logmessagesafely abstractlogger java at org apache logging spi abstractlogger logmessage abstractlogger java at org apache logging spi abstractlogger logifenabled abstractlogger java at org apache commons logging logfactory log logfactory java at org apache commons logging logfactory error logfactory java at org springframework boot springapplication reportfailure springapplication java at org springframework boot springapplication handlerunfailure springapplication java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org springframework samples petclinic petclinicapplication main petclinicapplication java more caused by java lang classnotfoundexception ch qos logback classic spi throwableproxy at java net urlclassloader findclass urlclassloader java at java lang classloader loadclasshelper classloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java more | 0 |
73,501 | 14,085,472,318 | IssuesEvent | 2020-11-05 01:05:14 | microsoft/vscode-python | https://api.github.com/repos/microsoft/vscode-python | opened | GHA: Detect a new Virtual Environment fails with timeout | area-internal classify type-code health | ```
1) GlobalVirtualEnvironment Locator
Detect a new Virtual Environment:
Error: Timeout of 25000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (d:\a\vscode-python\vscode-python\path with spaces\out\test\pythonEnvironments\discovery\locators\globalVirtualEnvironmentLocator.testvirtualenvs.js)
at listOnTimeout (internal/timers.js:531:17)
at processTimers (internal/timers.js:475:7)
``` | 1.0 | GHA: Detect a new Virtual Environment fails with timeout - ```
1) GlobalVirtualEnvironment Locator
Detect a new Virtual Environment:
Error: Timeout of 25000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (d:\a\vscode-python\vscode-python\path with spaces\out\test\pythonEnvironments\discovery\locators\globalVirtualEnvironmentLocator.testvirtualenvs.js)
at listOnTimeout (internal/timers.js:531:17)
at processTimers (internal/timers.js:475:7)
``` | non_priority | gha detect a new virtual environment fails with timeout globalvirtualenvironment locator detect a new virtual environment error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves d a vscode python vscode python path with spaces out test pythonenvironments discovery locators globalvirtualenvironmentlocator testvirtualenvs js at listontimeout internal timers js at processtimers internal timers js | 0 |
65,913 | 3,248,157,952 | IssuesEvent | 2015-10-17 02:39:05 | UniVR/GolfVR | https://api.github.com/repos/UniVR/GolfVR | closed | Hole ! panel at the beginning of Hole 3 ! | priority:normal type:bug | There is a panel saying "Hole !" at the begining of the hole 3 ! | 1.0 | Hole ! panel at the beginning of Hole 3 ! - There is a panel saying "Hole !" at the begining of the hole 3 ! | priority | hole panel at the beginning of hole there is a panel saying hole at the begining of the hole | 1 |
788,760 | 27,764,945,961 | IssuesEvent | 2023-03-16 10:49:48 | googleapis/java-firestore | https://api.github.com/repos/googleapis/java-firestore | closed | firestore.it.ITQueryWatchTest: emptyResults_newDocument_ADDED failed | priority: p2 type: bug api: firestore flakybot: issue | Note: #1214 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: a3cc6febfe9edc289db0968ee583c39143190ca5
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/aefd2208-2139-4aa2-8c39-894aff3f8750), [Sponge](http://sponge2/aefd2208-2139-4aa2-8c39-894aff3f8750)
status: failed
<details><summary>Test output</summary><br><pre>events[
event{error=null,value={documents[],documentChanges[]}},
]
value of : size()
expected to be in range: [2..2]
but was : 1
at com.google.cloud.firestore.it.ITQueryWatchTest$QuerySnapshotEventListener$ListenerAssertions.eventCountIsAnyOf(ITQueryWatchTest.java:809)
at com.google.cloud.firestore.it.ITQueryWatchTest.emptyResults_newDocument_ADDED(ITQueryWatchTest.java:243)
</pre></details> | 1.0 | firestore.it.ITQueryWatchTest: emptyResults_newDocument_ADDED failed - Note: #1214 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: a3cc6febfe9edc289db0968ee583c39143190ca5
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/aefd2208-2139-4aa2-8c39-894aff3f8750), [Sponge](http://sponge2/aefd2208-2139-4aa2-8c39-894aff3f8750)
status: failed
<details><summary>Test output</summary><br><pre>events[
event{error=null,value={documents[],documentChanges[]}},
]
value of : size()
expected to be in range: [2..2]
but was : 1
at com.google.cloud.firestore.it.ITQueryWatchTest$QuerySnapshotEventListener$ListenerAssertions.eventCountIsAnyOf(ITQueryWatchTest.java:809)
at com.google.cloud.firestore.it.ITQueryWatchTest.emptyResults_newDocument_ADDED(ITQueryWatchTest.java:243)
</pre></details> | priority | firestore it itquerywatchtest emptyresults newdocument added failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output events event error null value documents documentchanges value of size expected to be in range but was at com google cloud firestore it itquerywatchtest querysnapshoteventlistener listenerassertions eventcountisanyof itquerywatchtest java at com google cloud firestore it itquerywatchtest emptyresults newdocument added itquerywatchtest java | 1 |
6,105 | 13,735,166,857 | IssuesEvent | 2020-10-05 09:45:18 | onelocom/onelo | https://api.github.com/repos/onelocom/onelo | opened | Logging | architecture | ## What
We want to record every relevant action that happens during each phase
## Why
We want to be able to debug, report to the user, etc.
## Ideas
1. Use a `log` table that stores events as blobs (e.g. json blobs).
2. Use a set of tables specialised by type of log, e.g. `sourcing_entry_log` to capture what actions happened for each file during the sourcing phase. | 1.0 | Logging - ## What
We want to record every relevant action that happens during each phase
## Why
We want to be able to debug, report to the user, etc.
## Ideas
1. Use a `log` table that stores events as blobs (e.g. json blobs).
2. Use a set of tables specialised by type of log, e.g. `sourcing_entry_log` to capture what actions happened for each file during the sourcing phase. | non_priority | logging what we want to record every relevant action that happens during each phase why we want to be able to debug report to the user etc ideas use a log table that stores events as blobs e g json blobs use a set of tables specialised by type of log e g sourcing entry log to capture what actions happened for each file during the sourcing phase | 0 |
3,982 | 2,544,328,980 | IssuesEvent | 2015-01-29 08:57:05 | integram/cleverbus | https://api.github.com/repos/integram/cleverbus | closed | Handled requests/responses (body of exchange) will be dumped as a String | bug Core Priority major | In RequestSendingEventNotifier and ResponseReceiveEventNotifier it is necessary to store dumped requests/responses as a String, e.g.:
* String (for example XML) is dumped as a String
* byte array is dumped as HEX String
* Object as instance of HumanReadable is dumped by **toHumanString()** method
* others are dumped by by **toString()** method of object | 1.0 | Handled requests/responses (body of exchange) will be dumped as a String - In RequestSendingEventNotifier and ResponseReceiveEventNotifier it is necessary to store dumped requests/responses as a String, e.g.:
* String (for example XML) is dumped as a String
* byte array is dumped as HEX String
* Object as instance of HumanReadable is dumped by **toHumanString()** method
* others are dumped by by **toString()** method of object | priority | handled requests responses body of exchange will be dumped as a string in requestsendingeventnotifier and responsereceiveeventnotifier it is necessary to store dumped requests responses as a string e g string for example xml is dumped as a string byte array is dumped as hex string object as instance of humanreadable is dumped by tohumanstring method others are dumped by by tostring method of object | 1 |
159,068 | 6,040,198,996 | IssuesEvent | 2017-06-10 11:57:09 | OWASP/BLT | https://api.github.com/repos/OWASP/BLT | closed | Have badges for different types of bugs you've found | enhancement fullstack in-progress priority: high | - See the error types on the homepage, there are 6 | 1.0 | Have badges for different types of bugs you've found - - See the error types on the homepage, there are 6 | priority | have badges for different types of bugs you ve found see the error types on the homepage there are | 1 |
306,426 | 23,159,799,114 | IssuesEvent | 2022-07-29 16:25:19 | MaxBQb/InversionFilterManager | https://api.github.com/repos/MaxBQb/InversionFilterManager | closed | Where is the Readme file? | documentation good first issue | No visitors -> author not interested in investing his own time to write readme file
->
No readme file -> random visitors don't understand what's going on
->
Visitors not interested in project -> no visitors
| 1.0 | Where is the Readme file? - No visitors -> author not interested in investing his own time to write readme file
->
No readme file -> random visitors don't understand what's going on
->
Visitors not interested in project -> no visitors
| non_priority | where is the readme file no visitors author not interested in investing his own time to write readme file no readme file random visitors don t understand what s going on visitors not interested in project no visitors | 0 |
12,464 | 3,274,653,272 | IssuesEvent | 2015-10-26 12:09:01 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Cannot locate the impl of the trait from nested type parameter. | A-traits A-typesystem E-needstest I-ICE | When I used bound::typeparam as type, thread `rustc` panicked.
```rust
use rustc_serialize::{Encodable, Decodable};
use std::fmt::Display;
pub trait Entity : Decodable + Encodable + Sized {
type Key: Clone + Decodable + Encodable + ToString + Display + Eq + Ord + Sized;
fn id(&self) -> Self::Key;
fn find_by_id(id: Self::Key) -> Option<Self>;
}
pub struct DbRef<E: Entity> {
pub id: E::Key,
}
impl<E> DbRef<E>
where E: Entity {
fn get(self) -> Option<E> {
E::find_by_id(self.id)
}
}
```
Source code: https://github.com/ceram1/appcore/blob/master/src/db/mod.rs
Build log: https://travis-ci.org/ceram1/appcore/builds/60641835
| 1.0 | Cannot locate the impl of the trait from nested type parameter. - When I used bound::typeparam as type, thread `rustc` panicked.
```rust
use rustc_serialize::{Encodable, Decodable};
use std::fmt::Display;
pub trait Entity : Decodable + Encodable + Sized {
type Key: Clone + Decodable + Encodable + ToString + Display + Eq + Ord + Sized;
fn id(&self) -> Self::Key;
fn find_by_id(id: Self::Key) -> Option<Self>;
}
pub struct DbRef<E: Entity> {
pub id: E::Key,
}
impl<E> DbRef<E>
where E: Entity {
fn get(self) -> Option<E> {
E::find_by_id(self.id)
}
}
```
Source code: https://github.com/ceram1/appcore/blob/master/src/db/mod.rs
Build log: https://travis-ci.org/ceram1/appcore/builds/60641835
| non_priority | cannot locate the impl of the trait from nested type parameter when i used bound typeparam as type thread rustc panicked rust use rustc serialize encodable decodable use std fmt display pub trait entity decodable encodable sized type key clone decodable encodable tostring display eq ord sized fn id self self key fn find by id id self key option pub struct dbref pub id e key impl dbref where e entity fn get self option e find by id self id source code build log | 0 |
100,775 | 4,103,371,359 | IssuesEvent | 2016-06-04 16:43:36 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | closed | full screen 100% full not 99% | > Enhancement >> Stand-by Display OpenGL OS All Priority Low Version Git | You will say that I am fussy but this is important.
As I want to project a GAMA model on a table in some case if table is black I put a black background in GAMA and then only my model will appear on the table. I ve just notice that there is a little white frame on the fullscreen mode (see screenshot)
<img width="1920" alt="screen shot 2016-05-26 at 10 40 31" src="https://cloud.githubusercontent.com/assets/3928502/15578552/e00cf318-232e-11e6-87b3-4c5d042bb20f.png">
is there a way to disable this white frame? maybe a simple option of the fullscreen mode? | 1.0 | full screen 100% full not 99% - You will say that I am fussy but this is important.
As I want to project a GAMA model on a table in some case if table is black I put a black background in GAMA and then only my model will appear on the table. I ve just notice that there is a little white frame on the fullscreen mode (see screenshot)
<img width="1920" alt="screen shot 2016-05-26 at 10 40 31" src="https://cloud.githubusercontent.com/assets/3928502/15578552/e00cf318-232e-11e6-87b3-4c5d042bb20f.png">
is there a way to disable this white frame? maybe a simple option of the fullscreen mode? | priority | full screen full not you will say that i am fussy but this is important as i want to project a gama model on a table in some case if table is black i put a black background in gama and then only my model will appear on the table i ve just notice that there is a little white frame on the fullscreen mode see screenshot img width alt screen shot at src is there a way to disable this white frame maybe a simple option of the fullscreen mode | 1 |
44,658 | 2,910,368,497 | IssuesEvent | 2015-06-21 17:47:51 | PredictiveEcology/SpaDES | https://api.github.com/repos/PredictiveEcology/SpaDES | closed | simtime accessors should convert to `simTimeStepUnit` | core SpaDES High priority resolved in branch | continuation of #151
*E.g.*, calling `simCurrentTime(mySim)` in a module that uses time step of "weeks" should return the currunt time in weeks. | 1.0 | simtime accessors should convert to `simTimeStepUnit` - continuation of #151
*E.g.*, calling `simCurrentTime(mySim)` in a module that uses time step of "weeks" should return the currunt time in weeks. | priority | simtime accessors should convert to simtimestepunit continuation of e g calling simcurrenttime mysim in a module that uses time step of weeks should return the currunt time in weeks | 1 |
56,422 | 6,979,321,023 | IssuesEvent | 2017-12-12 20:36:31 | PnX-SI/GeoNature | https://api.github.com/repos/PnX-SI/GeoNature | opened | Revoir HOME et customisation | design frontend généricité | - Message d'intro modifiable
- Supprimer galerie photo
- Pouvoir masquer la carte et les graphs | 1.0 | Revoir HOME et customisation - - Message d'intro modifiable
- Supprimer galerie photo
- Pouvoir masquer la carte et les graphs | non_priority | revoir home et customisation message d intro modifiable supprimer galerie photo pouvoir masquer la carte et les graphs | 0 |
358,982 | 25,211,521,386 | IssuesEvent | 2022-11-14 04:37:44 | SigNoz/signoz-website | https://api.github.com/repos/SigNoz/signoz-website | closed | Linking internal pages | documentation | Will this type of linking work?
<img width="841" alt="Screenshot 2022-10-28 at 4 52 39 PM" src="https://user-images.githubusercontent.com/83692067/198575728-69341b32-1786-43ab-92e7-00b31b269ef5.png">
Currently, I am using the entire link like shown below. The below opens in a new page. I think for docs section opening in the same tab makes more sense just like our current behaviour.
<img width="942" alt="Screenshot 2022-10-28 at 4 54 12 PM" src="https://user-images.githubusercontent.com/83692067/198575916-e64924b6-ab07-47b2-94f0-8770e717e90c.png">
| 1.0 | Linking internal pages - Will this type of linking work?
<img width="841" alt="Screenshot 2022-10-28 at 4 52 39 PM" src="https://user-images.githubusercontent.com/83692067/198575728-69341b32-1786-43ab-92e7-00b31b269ef5.png">
Currently, I am using the entire link like shown below. The below opens in a new page. I think for docs section opening in the same tab makes more sense just like our current behaviour.
<img width="942" alt="Screenshot 2022-10-28 at 4 54 12 PM" src="https://user-images.githubusercontent.com/83692067/198575916-e64924b6-ab07-47b2-94f0-8770e717e90c.png">
| non_priority | linking internal pages will this type of linking work img width alt screenshot at pm src currently i am using the entire link like shown below the below opens in a new page i think for docs section opening in the same tab makes more sense just like our current behaviour img width alt screenshot at pm src | 0 |
61,115 | 17,023,608,868 | IssuesEvent | 2021-07-03 02:54:13 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Merkaartor should show all the available decimal places in coordinates for the node Properties pane | Component: merkaartor Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.16am, Monday, 21st June 2010]**
Currently, Merkaartor only shows 8 digits for the whole coordinates including the integral part. The schema allows for 7 decimal places so Merkaartor should show that level of precision, at least in the Properties pane of nodes.
| 1.0 | Merkaartor should show all the available decimal places in coordinates for the node Properties pane - **[Submitted to the original trac issue database at 9.16am, Monday, 21st June 2010]**
Currently, Merkaartor only shows 8 digits for the whole coordinates including the integral part. The schema allows for 7 decimal places so Merkaartor should show that level of precision, at least in the Properties pane of nodes.
| non_priority | merkaartor should show all the available decimal places in coordinates for the node properties pane currently merkaartor only shows digits for the whole coordinates including the integral part the schema allows for decimal places so merkaartor should show that level of precision at least in the properties pane of nodes | 0 |
400,079 | 27,267,644,070 | IssuesEvent | 2023-02-22 19:24:28 | Iron-E/nvim-libmodal | https://api.github.com/repos/Iron-E/nvim-libmodal | closed | feline archieved switch to lualine? | documentation enhancement | Hi, I wanted to try out your plugin but didn't found an example how to show the mods in lualine, then I saw that you use feline and saw that it's archived by the owner, do you plan to switch to some other plugin like lualine?
If yes, I would love to see your configuration for nvim-libmodal for it. | 1.0 | feline archieved switch to lualine? - Hi, I wanted to try out your plugin but didn't found an example how to show the mods in lualine, then I saw that you use feline and saw that it's archived by the owner, do you plan to switch to some other plugin like lualine?
If yes, I would love to see your configuration for nvim-libmodal for it. | non_priority | feline archieved switch to lualine hi i wanted to try out your plugin but didn t found an example how to show the mods in lualine then i saw that you use feline and saw that it s archived by the owner do you plan to switch to some other plugin like lualine if yes i would love to see your configuration for nvim libmodal for it | 0 |
317,135 | 9,661,303,496 | IssuesEvent | 2019-05-20 17:40:00 | sot/proseco | https://api.github.com/repos/sot/proseco | opened | Remove or address all items that would require manual / SAUSAGE edits | enhancement priority-low | Right now I'm thinking of:
* Fid edits (already in #208)
* Monitor window converted to guide star (placed in slot 7 with 8x8)
* Monitor window DTS
I don't recall if we had anything else that we expect to require SAUSAGE edits. I'm not seeing that we'd need ER monitor windows anytime soon, for example. | 1.0 | Remove or address all items that would require manual / SAUSAGE edits - Right now I'm thinking of:
* Fid edits (already in #208)
* Monitor window converted to guide star (placed in slot 7 with 8x8)
* Monitor window DTS
I don't recall if we had anything else that we expect to require SAUSAGE edits. I'm not seeing that we'd need ER monitor windows anytime soon, for example. | priority | remove or address all items that would require manual sausage edits right now i m thinking of fid edits already in monitor window converted to guide star placed in slot with monitor window dts i don t recall if we had anything else that we expect to require sausage edits i m not seeing that we d need er monitor windows anytime soon for example | 1 |
26,511 | 12,414,630,491 | IssuesEvent | 2020-05-22 14:52:34 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | Error accessing gitlab.com repo | bug team/core-services | - Go to https://gitlab.com/sourcegraph/codeintellify/-/blob/master/src/hoverifier.ts
- See no code intel, error in console:
```
Could not resolve file info for code view Error: RepoLookup for {Repo:gitlab.com/sourcegraph/codeintellify} failed with http status 500: syncer.syncsubset.store.upsert-repos: insert: pq: could not serialize access due to concurrent update
``` | 1.0 | Error accessing gitlab.com repo - - Go to https://gitlab.com/sourcegraph/codeintellify/-/blob/master/src/hoverifier.ts
- See no code intel, error in console:
```
Could not resolve file info for code view Error: RepoLookup for {Repo:gitlab.com/sourcegraph/codeintellify} failed with http status 500: syncer.syncsubset.store.upsert-repos: insert: pq: could not serialize access due to concurrent update
``` | non_priority | error accessing gitlab com repo go to see no code intel error in console could not resolve file info for code view error repolookup for repo gitlab com sourcegraph codeintellify failed with http status syncer syncsubset store upsert repos insert pq could not serialize access due to concurrent update | 0 |
12,479 | 7,887,590,017 | IssuesEvent | 2018-06-27 18:59:30 | iodide-project/pyodide | https://api.github.com/repos/iodide-project/pyodide | closed | Fix and/or document issues around finalization of objects | C/C++ Javascript Python/C API WASM performance | From the embind documentation:
Embind has support for binding most C++ constructs, including those introduced in C++11 and C++14. Its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics.
This means that Python objects will "leak" under the current framework. Should investigate what, if anything, can be done there. | True | Fix and/or document issues around finalization of objects - From the embind documentation:
Embind has support for binding most C++ constructs, including those introduced in C++11 and C++14. Its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics.
This means that Python objects will "leak" under the current framework. Should investigate what, if anything, can be done there. | non_priority | fix and or document issues around finalization of objects from the embind documentation embind has support for binding most c constructs including those introduced in c and c its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics this means that python objects will leak under the current framework should investigate what if anything can be done there | 0 |
61,211 | 7,451,244,690 | IssuesEvent | 2018-03-29 01:46:09 | phetsims/gas-properties | https://api.github.com/repos/phetsims/gas-properties | opened | Allowing toggle between degrees C and K | design:general type:user-feedback | A twitter user mentioned it might be nice to choose between degrees C and degrees K.
Since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well. | 1.0 | Allowing toggle between degrees C and K - A twitter user mentioned it might be nice to choose between degrees C and degrees K.
Since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well. | non_priority | allowing toggle between degrees c and k a twitter user mentioned it might be nice to choose between degrees c and degrees k since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well | 0 |
16,909 | 9,547,573,621 | IssuesEvent | 2019-05-02 00:05:02 | tfussell/xlnt | https://api.github.com/repos/tfussell/xlnt | closed | Optimize compound document streambufs | enhancement performance | Only a single sector should be read/written at a time. Consecutive read/writes to the same sector should not cause the sector to be read/written again. Extract read/write to separate methods. | True | Optimize compound document streambufs - Only a single sector should be read/written at a time. Consecutive read/writes to the same sector should not cause the sector to be read/written again. Extract read/write to separate methods. | non_priority | optimize compound document streambufs only a single sector should be read written at a time consecutive read writes to the same sector should not cause the sector to be read written again extract read write to separate methods | 0 |
214,737 | 7,276,372,061 | IssuesEvent | 2018-02-21 16:12:11 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Status page: we are stripping the username + realm for local accounts | Priority: High Status: In progress Type: Bug | For example if you would use an email address that is a local account on PF, that would strip the username and prevent authentication to be correct.
Potential fix would be to have the local source declare it doesn't want a stripped username.
Not sure if it would work but it technically should | 1.0 | Status page: we are stripping the username + realm for local accounts - For example if you would use an email address that is a local account on PF, that would strip the username and prevent authentication to be correct.
Potential fix would be to have the local source declare it doesn't want a stripped username.
Not sure if it would work but it technically should | priority | status page we are stripping the username realm for local accounts for example if you would use an email address that is a local account on pf that would strip the username and prevent authentication to be correct potential fix would be to have the local source declare it doesn t want a stripped username not sure if it would work but it technically should | 1 |
768,482 | 26,965,108,836 | IssuesEvent | 2023-02-08 21:36:04 | googleapis/java-storage | https://api.github.com/repos/googleapis/java-storage | closed | gRPC: use of `userProject` options results in error from gcs when used with credentials which define quota_project_id | type: bug api: storage priority: p3 | Affected version(s): 2.14.0+
#### Steps to reproduce
1. Set project via gcloud config set project $projectId
2. Create adc following instructions from https://google.aip.dev/auth/4113
3. <details>
<summary>JUnit tests to repro failure</summary>
```java
import static com.google.common.truth.Truth.assertThat;
import com.google.cloud.storage.Bucket;
import com.google.cloud.storage.BucketFixture;
import com.google.cloud.storage.BucketInfo;
import com.google.cloud.storage.Storage;
import com.google.cloud.storage.Storage.BucketField;
import com.google.cloud.storage.Storage.BucketGetOption;
import com.google.cloud.storage.Storage.BucketSourceOption;
import com.google.cloud.storage.Storage.BucketTargetOption;
import com.google.cloud.storage.StorageFixture;
import com.google.cloud.storage.conformance.retry.CleanupStrategy;
import org.junit.AfterClass;
import org.junit.BeforeClass;
import org.junit.ClassRule;
import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.slf4j.Marker;
import org.slf4j.MarkerFactory;
import org.slf4j.bridge.SLF4JBridgeHandler;
public final class ITRequesterPaysTest {
static {
SLF4JBridgeHandler.removeHandlersForRootLogger();
SLF4JBridgeHandler.install();
}
private static final Logger LOGGER = LoggerFactory.getLogger(ITRequesterPaysTest.class);
private static final Marker TRACE_ENTER = MarkerFactory.getMarker("enter");
private static final Marker TRACE_EXIT = MarkerFactory.getMarker("exit");
@ClassRule
public static final StorageFixture sfGadc = StorageFixture.defaultGrpc();
private static final BucketFixture bf =
BucketFixture.newBuilder()
.setCleanupStrategy(CleanupStrategy.NEVER)
.setBucketNameFmtString("java-storage-grpc-%s")
.setHandle(sfJadc::getInstance)
.build();
private static final String bucketName = bf.newBucketName();
@BeforeClass
public static void beforeClass() {
LOGGER.trace(TRACE_ENTER, "beforeClass()");
try {
BucketInfo info1 = BucketInfo.newBuilder(bucketName).setRequesterPays(true).build();
Bucket bucket = sfJadc.getInstance().create(info1, BucketTargetOption.projection("noAcl"));
assertThat(bucket.requesterPays()).isTrue();
} finally {
LOGGER.trace(TRACE_EXIT, "beforeClass()");
}
}
@AfterClass
public static void afterClass() {
LOGGER.trace(TRACE_ENTER, "afterClass()");
try {
String projectId = sfJadc.getInstance().getOptions().getProjectId();
sfJadc.getInstance().delete(bucketName, BucketSourceOption.userProject(projectId));
} finally {
LOGGER.trace(TRACE_EXIT, "afterClass()");
}
}
@Test
public void grpc_adc() {
LOGGER.trace(TRACE_ENTER, "grpc_adc()");
try {
doTest(sfGadc.getInstance());
} finally {
LOGGER.trace(TRACE_EXIT, "grpc_adc()");
}
}
private static void doTest(Storage storage) {
LOGGER.trace(TRACE_ENTER, "doTest(storage : {})", storage);
try {
String projectId = storage.getOptions().getProjectId();
doGet(storage, projectId);
// doList(storage, projectId);
} finally {
LOGGER.trace(TRACE_EXIT, "doTest(storage : {})", storage);
}
}
private static void doGet(Storage storage, String projectId) {
Bucket bucket =
storage
.get(
bucketName,
BucketGetOption.fields(BucketField.ID, BucketField.BILLING),
BucketGetOption.userProject(projectId));
assertThat(bucket.requesterPays()).isTrue();
}
}
```
</details>
#### Stack trace
```
com.google.cloud.storage.StorageException: INVALID_ARGUMENT: User project specified in the request is invalid.
at com.google.cloud.storage.StorageException.asStorageException(StorageException.java:144)
at com.google.cloud.storage.StorageException.coalesce(StorageException.java:117)
at com.google.cloud.storage.Retrying.run(Retrying.java:96)
at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376)
at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154)
at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid.
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:92)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:98)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:563)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:744)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:723)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
... 1 more
Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed
at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57)
at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:100)
at com.google.cloud.storage.GrpcStorageImpl.lambda$get$2(GrpcStorageImpl.java:379)
at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103)
at com.google.cloud.RetryHelper.run(RetryHelper.java:76)
at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50)
at com.google.cloud.storage.Retrying.run(Retrying.java:93)
at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376)
at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154)
at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
... 1 more
Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid.
at io.grpc.Status.asRuntimeException(Status.java:535)
... 14 more
```
#### Current Workaround
Decorate default credentials to hide quota_project_id
```java
Storage s = StorageOptions.grpc()
.setCredentials(new QuotaProjectIdHidingCredentials(GoogleCredentials.getApplicationDefault()))
.build()
.getService();
```
Internal reference bug: b/248605515 | 1.0 | gRPC: use of `userProject` options results in error from gcs when used with credentials which define quota_project_id - Affected version(s): 2.14.0+
#### Steps to reproduce
1. Set project via gcloud config set project $projectId
2. Create adc following instructions from https://google.aip.dev/auth/4113
3. <details>
<summary>JUnit tests to repro failure</summary>
```java
import static com.google.common.truth.Truth.assertThat;
import com.google.cloud.storage.Bucket;
import com.google.cloud.storage.BucketFixture;
import com.google.cloud.storage.BucketInfo;
import com.google.cloud.storage.Storage;
import com.google.cloud.storage.Storage.BucketField;
import com.google.cloud.storage.Storage.BucketGetOption;
import com.google.cloud.storage.Storage.BucketSourceOption;
import com.google.cloud.storage.Storage.BucketTargetOption;
import com.google.cloud.storage.StorageFixture;
import com.google.cloud.storage.conformance.retry.CleanupStrategy;
import org.junit.AfterClass;
import org.junit.BeforeClass;
import org.junit.ClassRule;
import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.slf4j.Marker;
import org.slf4j.MarkerFactory;
import org.slf4j.bridge.SLF4JBridgeHandler;
public final class ITRequesterPaysTest {
static {
SLF4JBridgeHandler.removeHandlersForRootLogger();
SLF4JBridgeHandler.install();
}
private static final Logger LOGGER = LoggerFactory.getLogger(ITRequesterPaysTest.class);
private static final Marker TRACE_ENTER = MarkerFactory.getMarker("enter");
private static final Marker TRACE_EXIT = MarkerFactory.getMarker("exit");
@ClassRule
public static final StorageFixture sfGadc = StorageFixture.defaultGrpc();
private static final BucketFixture bf =
BucketFixture.newBuilder()
.setCleanupStrategy(CleanupStrategy.NEVER)
.setBucketNameFmtString("java-storage-grpc-%s")
.setHandle(sfJadc::getInstance)
.build();
private static final String bucketName = bf.newBucketName();
@BeforeClass
public static void beforeClass() {
LOGGER.trace(TRACE_ENTER, "beforeClass()");
try {
BucketInfo info1 = BucketInfo.newBuilder(bucketName).setRequesterPays(true).build();
Bucket bucket = sfJadc.getInstance().create(info1, BucketTargetOption.projection("noAcl"));
assertThat(bucket.requesterPays()).isTrue();
} finally {
LOGGER.trace(TRACE_EXIT, "beforeClass()");
}
}
@AfterClass
public static void afterClass() {
LOGGER.trace(TRACE_ENTER, "afterClass()");
try {
String projectId = sfJadc.getInstance().getOptions().getProjectId();
sfJadc.getInstance().delete(bucketName, BucketSourceOption.userProject(projectId));
} finally {
LOGGER.trace(TRACE_EXIT, "afterClass()");
}
}
@Test
public void grpc_adc() {
LOGGER.trace(TRACE_ENTER, "grpc_adc()");
try {
doTest(sfGadc.getInstance());
} finally {
LOGGER.trace(TRACE_EXIT, "grpc_adc()");
}
}
private static void doTest(Storage storage) {
LOGGER.trace(TRACE_ENTER, "doTest(storage : {})", storage);
try {
String projectId = storage.getOptions().getProjectId();
doGet(storage, projectId);
// doList(storage, projectId);
} finally {
LOGGER.trace(TRACE_EXIT, "doTest(storage : {})", storage);
}
}
private static void doGet(Storage storage, String projectId) {
Bucket bucket =
storage
.get(
bucketName,
BucketGetOption.fields(BucketField.ID, BucketField.BILLING),
BucketGetOption.userProject(projectId));
assertThat(bucket.requesterPays()).isTrue();
}
}
```
</details>
#### Stack trace
```
com.google.cloud.storage.StorageException: INVALID_ARGUMENT: User project specified in the request is invalid.
at com.google.cloud.storage.StorageException.asStorageException(StorageException.java:144)
at com.google.cloud.storage.StorageException.coalesce(StorageException.java:117)
at com.google.cloud.storage.Retrying.run(Retrying.java:96)
at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376)
at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154)
at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid.
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:92)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:98)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:563)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:744)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:723)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
... 1 more
Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed
at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57)
at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:100)
at com.google.cloud.storage.GrpcStorageImpl.lambda$get$2(GrpcStorageImpl.java:379)
at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103)
at com.google.cloud.RetryHelper.run(RetryHelper.java:76)
at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50)
at com.google.cloud.storage.Retrying.run(Retrying.java:93)
at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376)
at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154)
at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
... 1 more
Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid.
at io.grpc.Status.asRuntimeException(Status.java:535)
... 14 more
```
#### Current Workaround
Decorate default credentials to hide quota_project_id
```java
Storage s = StorageOptions.grpc()
.setCredentials(new QuotaProjectIdHidingCredentials(GoogleCredentials.getApplicationDefault()))
.build()
.getService();
```
Internal reference bug: b/248605515 | priority | grpc use of userproject options results in error from gcs when used with credentials which define quota project id affected version s steps to reproduce set project via gcloud config set project projectid create adc following instructions from junit tests to repro failure java import static com google common truth truth assertthat import com google cloud storage bucket import com google cloud storage bucketfixture import com google cloud storage bucketinfo import com google cloud storage storage import com google cloud storage storage bucketfield import com google cloud storage storage bucketgetoption import com google cloud storage storage bucketsourceoption import com google cloud storage storage buckettargetoption import com google cloud storage storagefixture import com google cloud storage conformance retry cleanupstrategy import org junit afterclass import org junit beforeclass import org junit classrule import org junit test import org logger import org loggerfactory import org marker import org markerfactory import org bridge public final class itrequesterpaystest static removehandlersforrootlogger install private static final logger logger loggerfactory getlogger itrequesterpaystest class private static final marker trace enter markerfactory getmarker enter private static final marker trace exit markerfactory getmarker exit classrule public static final storagefixture sfgadc storagefixture defaultgrpc private static final bucketfixture bf bucketfixture newbuilder setcleanupstrategy cleanupstrategy never setbucketnamefmtstring java storage grpc s sethandle sfjadc getinstance build private static final string bucketname bf newbucketname beforeclass public static void beforeclass logger trace trace enter beforeclass try bucketinfo bucketinfo newbuilder bucketname setrequesterpays true build bucket bucket sfjadc getinstance create buckettargetoption projection noacl assertthat bucket requesterpays istrue finally logger trace trace exit beforeclass afterclass public static void afterclass logger trace trace enter afterclass try string projectid sfjadc getinstance getoptions getprojectid sfjadc getinstance delete bucketname bucketsourceoption userproject projectid finally logger trace trace exit afterclass test public void grpc adc logger trace trace enter grpc adc try dotest sfgadc getinstance finally logger trace trace exit grpc adc private static void dotest storage storage logger trace trace enter dotest storage storage try string projectid storage getoptions getprojectid doget storage projectid dolist storage projectid finally logger trace trace exit dotest storage storage private static void doget storage storage string projectid bucket bucket storage get bucketname bucketgetoption fields bucketfield id bucketfield billing bucketgetoption userproject projectid assertthat bucket requesterpays istrue stack trace com google cloud storage storageexception invalid argument user project specified in the request is invalid at com google cloud storage storageexception asstorageexception storageexception java at com google cloud storage storageexception coalesce storageexception java at com google cloud storage retrying run retrying java at com google cloud storage grpcstorageimpl get grpcstorageimpl java at com google cloud storage it itobjecttest unsetrequesterpays itobjecttest java at com google cloud storage it itobjecttest testlistblobrequesterpays itobjecttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument user project specified in the request is invalid at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more suppressed com google api gax rpc asynctaskexception asynchronous task failed at com google api gax rpc apiexceptions callandtranslateapiexception apiexceptions java at com google api gax rpc unarycallable call unarycallable java at com google cloud storage grpcstorageimpl lambda get grpcstorageimpl java at com google api gax retrying directretryingexecutor submit directretryingexecutor java at com google cloud retryhelper run retryhelper java at com google cloud retryhelper runwithretries retryhelper java at com google cloud storage retrying run retrying java at com google cloud storage grpcstorageimpl get grpcstorageimpl java at com google cloud storage it itobjecttest unsetrequesterpays itobjecttest java at com google cloud storage it itobjecttest testlistblobrequesterpays itobjecttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java util concurrent futuretask run futuretask java more caused by io grpc statusruntimeexception invalid argument user project specified in the request is invalid at io grpc status asruntimeexception status java more current workaround decorate default credentials to hide quota project id java storage s storageoptions grpc setcredentials new quotaprojectidhidingcredentials googlecredentials getapplicationdefault build getservice internal reference bug b | 1 |
393,733 | 11,624,079,444 | IssuesEvent | 2020-02-27 10:07:10 | nimblehq/rails-templates | https://api.github.com/repos/nimblehq/rails-templates | closed | Upgrade Node.js version. | priority : normal type : chore | In the Dockerfile, the Node.js version still set to version 8. I think we should upgrade the version number to the latest version of the LTS release. (Now the latest LTS version is 10.16.3) | 1.0 | Upgrade Node.js version. - In the Dockerfile, the Node.js version still set to version 8. I think we should upgrade the version number to the latest version of the LTS release. (Now the latest LTS version is 10.16.3) | priority | upgrade node js version in the dockerfile the node js version still set to version i think we should upgrade the version number to the latest version of the lts release now the latest lts version is | 1 |
67,163 | 20,923,492,350 | IssuesEvent | 2022-03-24 19:51:19 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Group chat icon on HomePage | T-Defect | ### Steps to reproduce
1. Open the home page
### Outcome
#### What did you expect?
The group icon should be displayed.
#### What happened instead?
It is blank. I guess the code to display the icon has been removed because it included `community` in its file name.

### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Group chat icon on HomePage - ### Steps to reproduce
1. Open the home page
### Outcome
#### What did you expect?
The group icon should be displayed.
#### What happened instead?
It is blank. I guess the code to display the icon has been removed because it included `community` in its file name.

### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No | non_priority | group chat icon on homepage steps to reproduce open the home page outcome what did you expect the group icon should be displayed what happened instead it is blank i guess the code to display the icon has been removed because it included community in its file name operating system no response browser information no response url for webapp localhost application version develop branch homeserver no response will you send logs no | 0 |
194,779 | 6,898,967,563 | IssuesEvent | 2017-11-24 11:41:23 | minio/mint | https://api.github.com/repos/minio/mint | closed | Gateway: Awscli head-object test fails | priority: medium | Awscli headobject test fails on all gateways.
Here is the failure entry
```
{"name": "awscli", "duration": 7574, "function": "aws --endpoint-url http://192.168.86.102:9000 s3api head-object --bucket awscli-mint-test-bucket-11375 --key prefix/directory/\n", "status": "FAIL", "error": "An error occurred (404) when calling the HeadObject operation: Not Found"}
```
This call succeeds on regular server mode. | 1.0 | Gateway: Awscli head-object test fails - Awscli headobject test fails on all gateways.
Here is the failure entry
```
{"name": "awscli", "duration": 7574, "function": "aws --endpoint-url http://192.168.86.102:9000 s3api head-object --bucket awscli-mint-test-bucket-11375 --key prefix/directory/\n", "status": "FAIL", "error": "An error occurred (404) when calling the HeadObject operation: Not Found"}
```
This call succeeds on regular server mode. | priority | gateway awscli head object test fails awscli headobject test fails on all gateways here is the failure entry name awscli duration function aws endpoint url head object bucket awscli mint test bucket key prefix directory n status fail error an error occurred when calling the headobject operation not found this call succeeds on regular server mode | 1 |
183,539 | 14,236,465,042 | IssuesEvent | 2020-11-18 16:01:17 | serverless/serverless | https://api.github.com/repos/serverless/serverless | opened | Tests: Refactor lib/plugins/aws/package/lib/mergeIamTemplates.test.js | good first issue help wanted refactor tests | _Most of the tests that currently cover core functionalities depend heavily on (and sometimes test) internal implementation characteristics. While they should be testing whether given implementation produces desired outcome (treating its implementation more as a black box)._
_Current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce, as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities ._
_At some point we've introduced a new (black box based) way of testing the internals. It's through `runServerless` utlity, which allows us create natural (as in real world) `serverless` instance, and inspect the produced outcome for chosen command. More details here: https://github.com/serverless/serverless/tree/master/test#unit-tests_
---
This is issue is about refactoring [lib/plugins/aws/package/lib/mergeIamTemplates.test.js](https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js) to `runServerless` based variant.
It's needed, so we cleanly move forward with https://github.com/serverless/serverless/issues/8396
To make refactor relatively easy:
- Spec for new tests is fully defined and is already committed in. It's in scope of this _describe_ block: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L774
- Each test slot indicates which test it is about to replace
- It is indicated where (and how many) exactly `runServerless` runs are expected to happen, against which fixture and command they should be based
- Necessary hints are provided
For every refactored test, the old test should be removed
Refactor can be submitted with one PR, but partial refactors that cover some of below listed parts are also very welcome:
- [ ] __No default role__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L776-L793
- [ ] __Default role with defaults__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L808-L828
- [ ] __Provider properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L839-L865
- [ ] __Function properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L890-L893
| 1.0 | Tests: Refactor lib/plugins/aws/package/lib/mergeIamTemplates.test.js - _Most of the tests that currently cover core functionalities depend heavily on (and sometimes test) internal implementation characteristics. While they should be testing whether given implementation produces desired outcome (treating its implementation more as a black box)._
_Current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce, as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities ._
_At some point we've introduced a new (black box based) way of testing the internals. It's through `runServerless` utlity, which allows us create natural (as in real world) `serverless` instance, and inspect the produced outcome for chosen command. More details here: https://github.com/serverless/serverless/tree/master/test#unit-tests_
---
This is issue is about refactoring [lib/plugins/aws/package/lib/mergeIamTemplates.test.js](https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js) to `runServerless` based variant.
It's needed, so we cleanly move forward with https://github.com/serverless/serverless/issues/8396
To make refactor relatively easy:
- Spec for new tests is fully defined and is already committed in. It's in scope of this _describe_ block: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L774
- Each test slot indicates which test it is about to replace
- It is indicated where (and how many) exactly `runServerless` runs are expected to happen, against which fixture and command they should be based
- Necessary hints are provided
For every refactored test, the old test should be removed
Refactor can be submitted with one PR, but partial refactors that cover some of below listed parts are also very welcome:
- [ ] __No default role__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L776-L793
- [ ] __Default role with defaults__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L808-L828
- [ ] __Provider properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L839-L865
- [ ] __Function properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L890-L893
| non_priority | tests refactor lib plugins aws package lib mergeiamtemplates test js most of the tests that currently cover core functionalities depend heavily on and sometimes test internal implementation characteristics while they should be testing whether given implementation produces desired outcome treating its implementation more as a black box current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities at some point we ve introduced a new black box based way of testing the internals it s through runserverless utlity which allows us create natural as in real world serverless instance and inspect the produced outcome for chosen command more details here this is issue is about refactoring to runserverless based variant it s needed so we cleanly move forward with to make refactor relatively easy spec for new tests is fully defined and is already committed in it s in scope of this describe block each test slot indicates which test it is about to replace it is indicated where and how many exactly runserverless runs are expected to happen against which fixture and command they should be based necessary hints are provided for every refactored test the old test should be removed refactor can be submitted with one pr but partial refactors that cover some of below listed parts are also very welcome no default role default role with defaults provider properties function properties | 0 |
322,075 | 23,888,322,592 | IssuesEvent | 2022-09-08 09:26:51 | k0sproject/k0s | https://api.github.com/repos/k0sproject/k0s | closed | Improve dual-stack documentation | documentation enhancement | ### Is your feature request related to a problem? Please describe.
* The template shown at https://docs.k0sproject.io/v1.24.3+k0s.0/dual-stack/ should specify `spec.network.provider = calico` for it to work. Otherwise the cluster will be using kube-proxy.
* The template mentioned above says "Use the following k0s.yaml as a template", but it's only listing the `spec` object. Since k0s seems to accept partial configuration (see #2090), replace the example with an actually valid k0s.yaml file (with apiVersion, kind fields, etc.)
* In the configuration docs, there is no mention of `spec.network.calico.mode` accepting the option `bird`. Only `vxlan` and `ipip` are shown to be acceptable.
### Describe the solution you would like
_No response_
### Describe alternatives you've considered
_No response_
### Additional context
_No response_ | 1.0 | Improve dual-stack documentation - ### Is your feature request related to a problem? Please describe.
* The template shown at https://docs.k0sproject.io/v1.24.3+k0s.0/dual-stack/ should specify `spec.network.provider = calico` for it to work. Otherwise the cluster will be using kube-proxy.
* The template mentioned above says "Use the following k0s.yaml as a template", but it's only listing the `spec` object. Since k0s seems to accept partial configuration (see #2090), replace the example with an actually valid k0s.yaml file (with apiVersion, kind fields, etc.)
* In the configuration docs, there is no mention of `spec.network.calico.mode` accepting the option `bird`. Only `vxlan` and `ipip` are shown to be acceptable.
### Describe the solution you would like
_No response_
### Describe alternatives you've considered
_No response_
### Additional context
_No response_ | non_priority | improve dual stack documentation is your feature request related to a problem please describe the template shown at should specify spec network provider calico for it to work otherwise the cluster will be using kube proxy the template mentioned above says use the following yaml as a template but it s only listing the spec object since seems to accept partial configuration see replace the example with an actually valid yaml file with apiversion kind fields etc in the configuration docs there is no mention of spec network calico mode accepting the option bird only vxlan and ipip are shown to be acceptable describe the solution you would like no response describe alternatives you ve considered no response additional context no response | 0 |
200,879 | 7,017,881,552 | IssuesEvent | 2017-12-21 11:22:06 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | opened | Duplicated NIC when save a template | Category: CLI Priority: High Status: Accepted Type: Bug | <!--////////////////////////////////////////////-->
<!-- COMPLETE ONLY ONE SECTION -->
<!-- BUG REPORT *OR* ENHANCEMENT REQUEST -->
<!--////////////////////////////////////////////-->
<!--////////////////////////////////////////////-->
<!-- COMPLETE THIS SECTION FOR BUG REPORTS -->
<!--////////////////////////////////////////////-->
# Bug Report
## Version of OpenNebula
<!--Mark the relevant versions affected with [X] -->
- [ ] 5.2.2
- [X] 5.4.0
- [X] 5.4.1
- [X] 5.4.2
- [X] 5.4.3
- [ ] Development build
## Component
<!-- Mark the relevant versions affected with [X] -->
- [ ] Authorization (LDAP, x509 certs...)
- [X] Command Line Interface (CLI)
- [ ] Contextualization
- [ ] Documentation
- [ ] Federation and HA
- [ ] Host, Clusters and Monitorization
- [ ] KVM
- [ ] Networking
- [ ] Orchestration (OpenNebula Flow)
- [ ] Packages
- [ ] Scheduler
- [ ] Storage & Images
- [ ] Sunstone
- [ ] Upgrades
- [ ] User, Groups, VDCs and ACL
- [ ] vCenter
## Description
<!-- Brief description of your problem -->
When you save a virtual Machine with two or more nics associated, the template generated has double the nics of the original machine.
### Expected Behavior
### Actual Behavior
## How to reproduce
<!-- Steps to reproduce the issue -->
Save a Virtual Machine with two NICs associated.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
# Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Duplicated NIC when save a template - <!--////////////////////////////////////////////-->
<!-- COMPLETE ONLY ONE SECTION -->
<!-- BUG REPORT *OR* ENHANCEMENT REQUEST -->
<!--////////////////////////////////////////////-->
<!--////////////////////////////////////////////-->
<!-- COMPLETE THIS SECTION FOR BUG REPORTS -->
<!--////////////////////////////////////////////-->
# Bug Report
## Version of OpenNebula
<!--Mark the relevant versions affected with [X] -->
- [ ] 5.2.2
- [X] 5.4.0
- [X] 5.4.1
- [X] 5.4.2
- [X] 5.4.3
- [ ] Development build
## Component
<!-- Mark the relevant versions affected with [X] -->
- [ ] Authorization (LDAP, x509 certs...)
- [X] Command Line Interface (CLI)
- [ ] Contextualization
- [ ] Documentation
- [ ] Federation and HA
- [ ] Host, Clusters and Monitorization
- [ ] KVM
- [ ] Networking
- [ ] Orchestration (OpenNebula Flow)
- [ ] Packages
- [ ] Scheduler
- [ ] Storage & Images
- [ ] Sunstone
- [ ] Upgrades
- [ ] User, Groups, VDCs and ACL
- [ ] vCenter
## Description
<!-- Brief description of your problem -->
When you save a virtual Machine with two or more nics associated, the template generated has double the nics of the original machine.
### Expected Behavior
### Actual Behavior
## How to reproduce
<!-- Steps to reproduce the issue -->
Save a Virtual Machine with two NICs associated.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
# Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| priority | duplicated nic when save a template bug report version of opennebula development build component authorization ldap certs command line interface cli contextualization documentation federation and ha host clusters and monitorization kvm networking orchestration opennebula flow packages scheduler storage images sunstone upgrades user groups vdcs and acl vcenter description when you save a virtual machine with two or more nics associated the template generated has double the nics of the original machine expected behavior actual behavior how to reproduce save a virtual machine with two nics associated progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 1 |
451,449 | 32,028,893,581 | IssuesEvent | 2023-09-22 10:49:18 | airalab/robonomics-wiki | https://api.github.com/repos/airalab/robonomics-wiki | closed | [Deprecation]: Article "Robonomics hass integration" is deprecated due to a release "Robonomics Integration 1.6.0" in "homeassistant-robonomics-integration". | documentation deprecation | ### Issue description
Article ["Robonomics hass integration"](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) has been automatically marked as deprecated due to a recent release ["Robonomics Integration 1.6.0"](https://github.com/airalab/homeassistant-robonomics-integration/releases/tag/1.6.0) in ["homeassistant-robonomics-integration"](https://github.com/airalab/homeassistant-robonomics-integration).
An action has been requested from the article contributors:
- @LoSk-p
- @nakata5321
- @Fingerling42
Deprecation reference id: e4ad529468ba0f2bb19b8aa6ee47922c
### Doc Page
[Robonomics hass integration](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md)
### Note
This issue has been automatically created by the [Wiki deprecation bot](https://github.com/airalab/robonomics-wiki-deprecation-notifier). If this is a false alarm and the article is not actually deprecated - feel free to close this issue. It will not be created again unless a new release comes out in [homeassistant-robonomics-integration](https://github.com/airalab/homeassistant-robonomics-integration). | 1.0 | [Deprecation]: Article "Robonomics hass integration" is deprecated due to a release "Robonomics Integration 1.6.0" in "homeassistant-robonomics-integration". - ### Issue description
Article ["Robonomics hass integration"](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) has been automatically marked as deprecated due to a recent release ["Robonomics Integration 1.6.0"](https://github.com/airalab/homeassistant-robonomics-integration/releases/tag/1.6.0) in ["homeassistant-robonomics-integration"](https://github.com/airalab/homeassistant-robonomics-integration).
An action has been requested from the article contributors:
- @LoSk-p
- @nakata5321
- @Fingerling42
Deprecation reference id: e4ad529468ba0f2bb19b8aa6ee47922c
### Doc Page
[Robonomics hass integration](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md)
### Note
This issue has been automatically created by the [Wiki deprecation bot](https://github.com/airalab/robonomics-wiki-deprecation-notifier). If this is a false alarm and the article is not actually deprecated - feel free to close this issue. It will not be created again unless a new release comes out in [homeassistant-robonomics-integration](https://github.com/airalab/homeassistant-robonomics-integration). | non_priority | article robonomics hass integration is deprecated due to a release robonomics integration in homeassistant robonomics integration issue description article has been automatically marked as deprecated due to a recent release in an action has been requested from the article contributors losk p deprecation reference id doc page note this issue has been automatically created by the if this is a false alarm and the article is not actually deprecated feel free to close this issue it will not be created again unless a new release comes out in | 0 |
216,372 | 7,307,313,093 | IssuesEvent | 2018-02-28 02:06:21 | tgstation/tgstation-server | https://api.github.com/repos/tgstation/tgstation-server | closed | SSL issues after the PR attempts to merge (Test merge) | Bug Priority: High Review | https://i.imgur.com/fjRL2oQ.png
This is after updating to the latest TGS3 and updating the JSON. | 1.0 | SSL issues after the PR attempts to merge (Test merge) - https://i.imgur.com/fjRL2oQ.png
This is after updating to the latest TGS3 and updating the JSON. | priority | ssl issues after the pr attempts to merge test merge this is after updating to the latest and updating the json | 1 |
667,341 | 22,466,486,712 | IssuesEvent | 2022-06-22 02:33:43 | ml4ai/tomcat | https://api.github.com/repos/ml4ai/tomcat | closed | Connect audioStreamer to ASR agent | bug Priority: High | Test session: April 12, 2020
Before starting the trial, speechAgent connected to 2 client maps. During the trial, the speechAgent still only connected to 2 client maps, and did not connect to the 3rd client map (on Tiger).
We need a consistent procedure for Clientmaps to connect to speechAgent. | 1.0 | Connect audioStreamer to ASR agent - Test session: April 12, 2020
Before starting the trial, speechAgent connected to 2 client maps. During the trial, the speechAgent still only connected to 2 client maps, and did not connect to the 3rd client map (on Tiger).
We need a consistent procedure for Clientmaps to connect to speechAgent. | priority | connect audiostreamer to asr agent test session april before starting the trial speechagent connected to client maps during the trial the speechagent still only connected to client maps and did not connect to the client map on tiger we need a consistent procedure for clientmaps to connect to speechagent | 1 |
59,589 | 3,114,532,498 | IssuesEvent | 2015-09-03 09:21:12 | HubTurbo/HubTurbo | https://api.github.com/repos/HubTurbo/HubTurbo | closed | Ensure that globalConfigTest passes | priority.medium | Or at least most of the time, it seems to fail ~half the time now?
```
guitests.UseGlobalConfigsTest > globalConfigTest FAILED
org.loadui.testfx.exceptions.NoNodesVisibleException: Matching nodes were found, but none of them were visible. Screenshot saved as /home/travis/build/HubTurbo/HubTurbo/screenshot1441084988176.png.
at org.loadui.testfx.GuiTest.getVisibleNodes(GuiTest.java:238)
at org.loadui.testfx.GuiTest.findAll(GuiTest.java:251)
at org.loadui.testfx.GuiTest.findByCssSelector(GuiTest.java:432)
at org.loadui.testfx.GuiTest.find(GuiTest.java:300)
at org.loadui.testfx.GuiTest.pointFor(GuiTest.java:1159)
at org.loadui.testfx.GuiTest.move(GuiTest.java:849)
at org.loadui.testfx.GuiTest.click(GuiTest.java:627)
at guitests.UseGlobalConfigsTest.globalConfigTest(UseGlobalConfigsTest.java:68)
``` | 1.0 | Ensure that globalConfigTest passes - Or at least most of the time, it seems to fail ~half the time now?
```
guitests.UseGlobalConfigsTest > globalConfigTest FAILED
org.loadui.testfx.exceptions.NoNodesVisibleException: Matching nodes were found, but none of them were visible. Screenshot saved as /home/travis/build/HubTurbo/HubTurbo/screenshot1441084988176.png.
at org.loadui.testfx.GuiTest.getVisibleNodes(GuiTest.java:238)
at org.loadui.testfx.GuiTest.findAll(GuiTest.java:251)
at org.loadui.testfx.GuiTest.findByCssSelector(GuiTest.java:432)
at org.loadui.testfx.GuiTest.find(GuiTest.java:300)
at org.loadui.testfx.GuiTest.pointFor(GuiTest.java:1159)
at org.loadui.testfx.GuiTest.move(GuiTest.java:849)
at org.loadui.testfx.GuiTest.click(GuiTest.java:627)
at guitests.UseGlobalConfigsTest.globalConfigTest(UseGlobalConfigsTest.java:68)
``` | priority | ensure that globalconfigtest passes or at least most of the time it seems to fail half the time now guitests useglobalconfigstest globalconfigtest failed org loadui testfx exceptions nonodesvisibleexception matching nodes were found but none of them were visible screenshot saved as home travis build hubturbo hubturbo png at org loadui testfx guitest getvisiblenodes guitest java at org loadui testfx guitest findall guitest java at org loadui testfx guitest findbycssselector guitest java at org loadui testfx guitest find guitest java at org loadui testfx guitest pointfor guitest java at org loadui testfx guitest move guitest java at org loadui testfx guitest click guitest java at guitests useglobalconfigstest globalconfigtest useglobalconfigstest java | 1 |
292,350 | 8,956,651,425 | IssuesEvent | 2019-01-26 19:20:25 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256 | medium priority | ## 🐛 Bug
torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256
## To Reproduce
```
import torch
batch_size = 256*256
transform_parameters = torch.cuda.FloatTensor([[1,0,0],[0,1,0]])
transform_parameters = torch.stack([transform_parameters] * batch_size, 0).contiguous()
resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2)))
print(resampling_grids.size())
```
crashes with the following error:
```
Traceback (most recent call last):
File "bug.py", line 5, in <module>
resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2)))
File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/functional.py", line 2615, in affine_grid
return vision.affine_grid_generator(theta, size)
File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/_functions/vision.py", line 10, in affine_grid_generator
ret = torch.cudnn_affine_grid_generator(theta, N, C, H, W)
RuntimeError: cuDNN error: CUDNN_STATUS_EXECUTION_FAILED
```
If I set batch_size to 256*256 - 1 or do operation on a CPU everything works fine.
Behavior is very similar to the torch.inverse bug #13276 , but with a different function and a different error message.
## Expected behavior
The code should print `torch.Size([65536, 2, 2, 2])`
## Environment
PyTorch version: 1.0.0 (checked with pytorch-nightly of 24.01.2019)
Is debug build: No
CUDA used to build PyTorch: 10.0.130
OS: Ubuntu 18.04.1 LTS
GCC version: (Ubuntu 7.3.0-27ubuntu1~18.04) 7.3.0
CMake version: Could not collect
Python version: 3.7
Is CUDA available: Yes
CUDA runtime version: 10.0.130
GPU models and configuration: GPU 0: GeForce GTX 1080 Ti
Nvidia driver version: 410.79
cuDNN version: 7401 (installed by pytorch itself)
Versions of relevant libraries:
[pip3] maskrcnn-benchmark (0.1, /media/aosokin/kingston2tb/software/pytorch/maskrcnn-benchmark)
[pip3] numpy (1.14.0)
[pip3] numpydoc (0.7.0)
[pip3] torch (1.0.0a0+db5d313)
[pip3] torchfile (0.1.0)
[pip3] torchvision (0.2.1)
[conda] blas 1.0 mkl
[conda] cuda100 1.0 0 pytorch
[conda] mkl 2019.1 144
[conda] mkl_fft 1.0.6 py37hd81dba3_0
[conda] mkl_random 1.0.2 py37hd81dba3_0
[conda] pytorch 1.0.0 py3.7_cuda10.0.130_cudnn7.4.1_1 [cuda100] pytorch
[conda] torchvision 0.2.1 py_2 pytorch
| 1.0 | torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256 - ## 🐛 Bug
torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256
## To Reproduce
```
import torch
batch_size = 256*256
transform_parameters = torch.cuda.FloatTensor([[1,0,0],[0,1,0]])
transform_parameters = torch.stack([transform_parameters] * batch_size, 0).contiguous()
resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2)))
print(resampling_grids.size())
```
crashes with the following error:
```
Traceback (most recent call last):
File "bug.py", line 5, in <module>
resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2)))
File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/functional.py", line 2615, in affine_grid
return vision.affine_grid_generator(theta, size)
File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/_functions/vision.py", line 10, in affine_grid_generator
ret = torch.cudnn_affine_grid_generator(theta, N, C, H, W)
RuntimeError: cuDNN error: CUDNN_STATUS_EXECUTION_FAILED
```
If I set batch_size to 256*256 - 1 or do operation on a CPU everything works fine.
Behavior is very similar to the torch.inverse bug #13276 , but with a different function and a different error message.
## Expected behavior
The code should print `torch.Size([65536, 2, 2, 2])`
## Environment
PyTorch version: 1.0.0 (checked with pytorch-nightly of 24.01.2019)
Is debug build: No
CUDA used to build PyTorch: 10.0.130
OS: Ubuntu 18.04.1 LTS
GCC version: (Ubuntu 7.3.0-27ubuntu1~18.04) 7.3.0
CMake version: Could not collect
Python version: 3.7
Is CUDA available: Yes
CUDA runtime version: 10.0.130
GPU models and configuration: GPU 0: GeForce GTX 1080 Ti
Nvidia driver version: 410.79
cuDNN version: 7401 (installed by pytorch itself)
Versions of relevant libraries:
[pip3] maskrcnn-benchmark (0.1, /media/aosokin/kingston2tb/software/pytorch/maskrcnn-benchmark)
[pip3] numpy (1.14.0)
[pip3] numpydoc (0.7.0)
[pip3] torch (1.0.0a0+db5d313)
[pip3] torchfile (0.1.0)
[pip3] torchvision (0.2.1)
[conda] blas 1.0 mkl
[conda] cuda100 1.0 0 pytorch
[conda] mkl 2019.1 144
[conda] mkl_fft 1.0.6 py37hd81dba3_0
[conda] mkl_random 1.0.2 py37hd81dba3_0
[conda] pytorch 1.0.0 py3.7_cuda10.0.130_cudnn7.4.1_1 [cuda100] pytorch
[conda] torchvision 0.2.1 py_2 pytorch
| priority | torch nn functional affine grid on gpu crashes when batch size 🐛 bug torch nn functional affine grid on gpu crashes when batch size to reproduce import torch batch size transform parameters torch cuda floattensor transform parameters torch stack batch size contiguous resampling grids torch nn functional affine grid transform parameters torch size batch size print resampling grids size crashes with the following error traceback most recent call last file bug py line in resampling grids torch nn functional affine grid transform parameters torch size batch size file home aosokin local software envs pytorch lib site packages torch nn functional py line in affine grid return vision affine grid generator theta size file home aosokin local software envs pytorch lib site packages torch nn functions vision py line in affine grid generator ret torch cudnn affine grid generator theta n c h w runtimeerror cudnn error cudnn status execution failed if i set batch size to or do operation on a cpu everything works fine behavior is very similar to the torch inverse bug but with a different function and a different error message expected behavior the code should print torch size environment pytorch version checked with pytorch nightly of is debug build no cuda used to build pytorch os ubuntu lts gcc version ubuntu cmake version could not collect python version is cuda available yes cuda runtime version gpu models and configuration gpu geforce gtx ti nvidia driver version cudnn version installed by pytorch itself versions of relevant libraries maskrcnn benchmark media aosokin software pytorch maskrcnn benchmark numpy numpydoc torch torchfile torchvision blas mkl pytorch mkl mkl fft mkl random pytorch pytorch torchvision py pytorch | 1 |
165,456 | 14,002,622,331 | IssuesEvent | 2020-10-28 14:59:12 | tezos-reward-distributor-organization/tezos-reward-distributor | https://api.github.com/repos/tezos-reward-distributor-organization/tezos-reward-distributor | closed | Exception: Lock file present. Another process is running... | documentation | I keep getting when I run the script:
```
2020-10-10 01:28:33,474 - MainThread - --------------------------------------------
Traceback (most recent call last):
File "src/main.py", line 225, in <module>
main(args)
File "src/main.py", line 146, in main
life_cycle.start(not dry_run)
File "/Users/-/tezos-reward-distributor/src/util/process_life_cycle.py", line 20, in start
self.lock_file.lock()
File "/Users/-/tezos-reward-distributor/src/util/lock_file.py", line 10, in lock
self.tryLock()
File "/Users/-tezos-reward-distributor/src/util/lock_file.py", line 20, in tryLock
raise Exception("Lock file present. Another process is running...")
Exception: Lock file present. Another process is running...
```
Not sure how this happened, but I have to imagine the last script didn't end neatly. How do I go about reseting this? Can't seem to find a command.
Thanks! | 1.0 | Exception: Lock file present. Another process is running... - I keep getting when I run the script:
```
2020-10-10 01:28:33,474 - MainThread - --------------------------------------------
Traceback (most recent call last):
File "src/main.py", line 225, in <module>
main(args)
File "src/main.py", line 146, in main
life_cycle.start(not dry_run)
File "/Users/-/tezos-reward-distributor/src/util/process_life_cycle.py", line 20, in start
self.lock_file.lock()
File "/Users/-/tezos-reward-distributor/src/util/lock_file.py", line 10, in lock
self.tryLock()
File "/Users/-tezos-reward-distributor/src/util/lock_file.py", line 20, in tryLock
raise Exception("Lock file present. Another process is running...")
Exception: Lock file present. Another process is running...
```
Not sure how this happened, but I have to imagine the last script didn't end neatly. How do I go about reseting this? Can't seem to find a command.
Thanks! | non_priority | exception lock file present another process is running i keep getting when i run the script mainthread traceback most recent call last file src main py line in main args file src main py line in main life cycle start not dry run file users tezos reward distributor src util process life cycle py line in start self lock file lock file users tezos reward distributor src util lock file py line in lock self trylock file users tezos reward distributor src util lock file py line in trylock raise exception lock file present another process is running exception lock file present another process is running not sure how this happened but i have to imagine the last script didn t end neatly how do i go about reseting this can t seem to find a command thanks | 0 |
85,248 | 15,736,660,755 | IssuesEvent | 2021-03-30 01:08:59 | jgeraigery/k8-secrets-as-hyperledger-fabric-wallet | https://api.github.com/repos/jgeraigery/k8-secrets-as-hyperledger-fabric-wallet | opened | CVE-2020-13935 (High) detected in tomcat-embed-websocket-9.0.27.jar | security vulnerability | ## CVE-2020-13935 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-websocket-9.0.27.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: k8-secrets-as-hyperledger-fabric-wallet/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-websocket/9.0.27/tomcat-embed-websocket-9.0.27.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.1.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.2.1.RELEASE.jar
- :x: **tomcat-embed-websocket-9.0.27.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.
<p>Publish Date: 2020-07-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935>CVE-2020-13935</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E</a></p>
<p>Release Date: 2020-07-14</p>
<p>Fix Resolution: org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-websocket","packageVersion":"9.0.27","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.2.1.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.2.1.RELEASE;org.apache.tomcat.embed:tomcat-embed-websocket:9.0.27","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13935","vulnerabilityDetails":"The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-13935 (High) detected in tomcat-embed-websocket-9.0.27.jar - ## CVE-2020-13935 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-websocket-9.0.27.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: k8-secrets-as-hyperledger-fabric-wallet/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-websocket/9.0.27/tomcat-embed-websocket-9.0.27.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.1.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.2.1.RELEASE.jar
- :x: **tomcat-embed-websocket-9.0.27.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.
<p>Publish Date: 2020-07-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935>CVE-2020-13935</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E</a></p>
<p>Release Date: 2020-07-14</p>
<p>Fix Resolution: org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-websocket","packageVersion":"9.0.27","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.2.1.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.2.1.RELEASE;org.apache.tomcat.embed:tomcat-embed-websocket:9.0.27","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13935","vulnerabilityDetails":"The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in tomcat embed websocket jar cve high severity vulnerability vulnerable library tomcat embed websocket jar core tomcat implementation path to dependency file secrets as hyperledger fabric wallet pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed websocket tomcat embed websocket jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed websocket jar vulnerable library found in base branch master vulnerability details the payload length in a websocket frame was not correctly validated in apache tomcat to to to and to invalid payload lengths could trigger an infinite loop multiple requests with invalid payload lengths could lead to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat websocket org apache tomcat embed tomcat embed websocket isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed websocket isminimumfixversionavailable true minimumfixversion org apache tomcat tomcat websocket org apache tomcat embed tomcat embed websocket basebranches vulnerabilityidentifier cve vulnerabilitydetails the payload length in a websocket frame was not correctly validated in apache tomcat to to to and to invalid payload lengths could trigger an infinite loop multiple requests with invalid payload lengths could lead to a denial of service vulnerabilityurl | 0 |
75,444 | 3,462,478,282 | IssuesEvent | 2015-12-20 23:37:50 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | Unable to echo unsigned ints at compile-time | Medium Priority Stdlib | ```nim
import macros
macro foo(): stmt =
var x = 8'u8
echo(x)
foo()
```
```
c:\users\dominik\appdata\local\temp\aporia\a14.nim(7, 3) Info: template/generic instantiation from here
c:\users\dominik\git\nimrod\lib\system\repr.nim(25, 10) Error: cannot 'importc' variable at compile time
```
If anything, at least the error should be better. | 1.0 | Unable to echo unsigned ints at compile-time - ```nim
import macros
macro foo(): stmt =
var x = 8'u8
echo(x)
foo()
```
```
c:\users\dominik\appdata\local\temp\aporia\a14.nim(7, 3) Info: template/generic instantiation from here
c:\users\dominik\git\nimrod\lib\system\repr.nim(25, 10) Error: cannot 'importc' variable at compile time
```
If anything, at least the error should be better. | priority | unable to echo unsigned ints at compile time nim import macros macro foo stmt var x echo x foo c users dominik appdata local temp aporia nim info template generic instantiation from here c users dominik git nimrod lib system repr nim error cannot importc variable at compile time if anything at least the error should be better | 1 |
49,565 | 7,522,142,282 | IssuesEvent | 2018-04-12 19:26:47 | patternplate/patternplate | https://api.github.com/repos/patternplate/patternplate | closed | Terminal GIF out of sync with documentation | documentation help wanted | The GIF in the documentation here:
https://patternplate.github.io/doc/docs/advanced/deploy?guides-enabled=true&advanced-enabled=true#build-a-static-interface
Is out of sync with the written documentation.
How was the GIF created? Happy to fix myself. | 1.0 | Terminal GIF out of sync with documentation - The GIF in the documentation here:
https://patternplate.github.io/doc/docs/advanced/deploy?guides-enabled=true&advanced-enabled=true#build-a-static-interface
Is out of sync with the written documentation.
How was the GIF created? Happy to fix myself. | non_priority | terminal gif out of sync with documentation the gif in the documentation here is out of sync with the written documentation how was the gif created happy to fix myself | 0 |
322,347 | 23,903,634,165 | IssuesEvent | 2022-09-08 21:22:01 | neurobagel/concept_query | https://api.github.com/repos/neurobagel/concept_query | opened | Pushing Docker image to Docker hub | documentation deployment | Once the Dockerfile is in our repo and builds a local Docker image, let's
- [] Push the image to Dockerhub
- [] Write documentation for users of the repo to use `docker pull` the image instead of having to build it from the Dockerfile
- [] In building this image, ensure that versions of base image, python, and required packages are locked in for consistent behavior | 1.0 | Pushing Docker image to Docker hub - Once the Dockerfile is in our repo and builds a local Docker image, let's
- [] Push the image to Dockerhub
- [] Write documentation for users of the repo to use `docker pull` the image instead of having to build it from the Dockerfile
- [] In building this image, ensure that versions of base image, python, and required packages are locked in for consistent behavior | non_priority | pushing docker image to docker hub once the dockerfile is in our repo and builds a local docker image let s push the image to dockerhub write documentation for users of the repo to use docker pull the image instead of having to build it from the dockerfile in building this image ensure that versions of base image python and required packages are locked in for consistent behavior | 0 |
384,426 | 11,392,342,154 | IssuesEvent | 2020-01-30 02:10:42 | michaelsanford/bittray | https://api.github.com/repos/michaelsanford/bittray | closed | Fix rsrc syso inclusion | bug ci high-priority | It seems the existing way of doing that no longer works with go 1.13.6.
Try with `rsrc`
- https://github.com/akavel/rsrc
- https://github.com/lxn/walk/issues/28 (good reference 👍 )
- https://www.programming-books.io/essential/go/a-first-window-app-e87631a7f8a94e8f92f9114e8e6baf0a
| 1.0 | Fix rsrc syso inclusion - It seems the existing way of doing that no longer works with go 1.13.6.
Try with `rsrc`
- https://github.com/akavel/rsrc
- https://github.com/lxn/walk/issues/28 (good reference 👍 )
- https://www.programming-books.io/essential/go/a-first-window-app-e87631a7f8a94e8f92f9114e8e6baf0a
| priority | fix rsrc syso inclusion it seems the existing way of doing that no longer works with go try with rsrc good reference 👍 | 1 |
55,865 | 6,926,283,297 | IssuesEvent | 2017-11-30 18:37:54 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Streak Bonus snack notification does not include Gold icon | POST-REDESIGN priority: medium section: Task Page status: issue: in progress | Bonus Streak snacks do not contain the Gold icon. The icon should be identical to the earned/lost snack sizing.
 | 1.0 | Streak Bonus snack notification does not include Gold icon - Bonus Streak snacks do not contain the Gold icon. The icon should be identical to the earned/lost snack sizing.
 | non_priority | streak bonus snack notification does not include gold icon bonus streak snacks do not contain the gold icon the icon should be identical to the earned lost snack sizing | 0 |
361,949 | 10,721,605,323 | IssuesEvent | 2019-10-27 04:21:14 | AY1920S1-CS2103T-W11-2/main | https://api.github.com/repos/AY1920S1-CS2103T-W11-2/main | closed | Amend List command to display either a list of activities or contacts | priority.High type.Task | Amend the existing List command to allow a user to select between displaying the list of contacts and list of activities. | 1.0 | Amend List command to display either a list of activities or contacts - Amend the existing List command to allow a user to select between displaying the list of contacts and list of activities. | priority | amend list command to display either a list of activities or contacts amend the existing list command to allow a user to select between displaying the list of contacts and list of activities | 1 |
809,705 | 30,206,385,112 | IssuesEvent | 2023-07-05 09:42:52 | informalsystems/quint | https://api.github.com/repos/informalsystems/quint | closed | Topological sort of definitions in phase 2 | W8 Fresolver (phase 2) impact-medium effort-hard product-priority-general | The language manual lets the user to write definitions out of order. For instance:
```scala
module OutOfOrder {
val n1 = e1
val n2 = n3 + 1
val n3 = n1 * 2
}
```
Currently, the parser (phase 1) and the name resolver (phase 2) keep the order of definitions. Moreover, the name resolver does not check for cyclic definitions.
As a result, when writing a later phase (e.g., the simulator) with `IRVisitor`, I have to think about the out-of-order definitions. It looks like we have enough information to make a topological sort of definitions in phase 2. If we do so, using `IRVisitor` in later phases would be simpler, as we can assume that all names are defined before they are used (in the traversal order of `IRVisitor`).
Apalache is doing a topological sort of definitions already, perhaps not optimally, so we can get some inspiration from [DeclarationSorter.scala](https://github.com/informalsystems/apalache/blob/unstable/tlair/src/main/scala/at/forsyte/apalache/tla/lir/transformations/standard/DeclarationSorter.scala).
Once we have topological sort, we can detect cyclic references in phase 2. | 1.0 | Topological sort of definitions in phase 2 - The language manual lets the user to write definitions out of order. For instance:
```scala
module OutOfOrder {
val n1 = e1
val n2 = n3 + 1
val n3 = n1 * 2
}
```
Currently, the parser (phase 1) and the name resolver (phase 2) keep the order of definitions. Moreover, the name resolver does not check for cyclic definitions.
As a result, when writing a later phase (e.g., the simulator) with `IRVisitor`, I have to think about the out-of-order definitions. It looks like we have enough information to make a topological sort of definitions in phase 2. If we do so, using `IRVisitor` in later phases would be simpler, as we can assume that all names are defined before they are used (in the traversal order of `IRVisitor`).
Apalache is doing a topological sort of definitions already, perhaps not optimally, so we can get some inspiration from [DeclarationSorter.scala](https://github.com/informalsystems/apalache/blob/unstable/tlair/src/main/scala/at/forsyte/apalache/tla/lir/transformations/standard/DeclarationSorter.scala).
Once we have topological sort, we can detect cyclic references in phase 2. | priority | topological sort of definitions in phase the language manual lets the user to write definitions out of order for instance scala module outoforder val val val currently the parser phase and the name resolver phase keep the order of definitions moreover the name resolver does not check for cyclic definitions as a result when writing a later phase e g the simulator with irvisitor i have to think about the out of order definitions it looks like we have enough information to make a topological sort of definitions in phase if we do so using irvisitor in later phases would be simpler as we can assume that all names are defined before they are used in the traversal order of irvisitor apalache is doing a topological sort of definitions already perhaps not optimally so we can get some inspiration from once we have topological sort we can detect cyclic references in phase | 1 |
103,703 | 12,963,534,326 | IssuesEvent | 2020-07-20 18:56:59 | microsoft/WSL | https://api.github.com/repos/microsoft/WSL | closed | generateResolvConf = false did not work in wsl.conf | bydesign | <!--
🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
I ACKNOWLEDGE THE FOLLOWING BEFORE PROCEEDING:
1. If I delete this entire template and go my own path, the core team may close my issue without further explanation or engagement.
2. If I list multiple bugs/concerns in this one issue, the core team may close my issue without further explanation or engagement.
3. If I write an issue that has many duplicates, the core team may close my issue without further explanation or engagement (and without necessarily spending time to find the exact duplicate ID number).
4. If I leave the title incomplete when filing the issue, the core team may close my issue without further explanation or engagement.
5. If I file something completely blank in the body, the core team may close my issue without further explanation or engagement.
All good? Then proceed!
-->
I have WSL2 with debian -all good.
I created a /etc/wsl.conf file
it contains
```
[network]
generateResolvConf = false
```
as per https://docs.microsoft.com/en-us/windows/wsl/wsl-config#network
if i delete /etc/resolv.conf (or edit and save) on start of the next debia WSL2 session it has been regenerated again.
running WSL2 in windows 10 build 20170.rs_prerelease.200710-1419
<!--
This bug tracker is monitored by Windows Subsystem for Linux development team and other technical folks.
Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues.
Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. Ideally, please configure your machine to capture minidumps, repro the issue, and send the minidump from "C:\Windows\minidump\".
You can find instructions to do that here: https://support.microsoft.com/en-us/help/315263/how-to-read-the-small-memory-dump-file-that-is-created-by-windows-if-a
If this is a console issue (a problem with layout, rendering, colors, etc.), please post the issue to the Terminal tracker: https://github.com/microsoft/terminal/issues
For documentation improvements, please post to the documentation tracker: https://github.com/MicrosoftDocs/WSL/issues
For any other questions on contributing please see our contribution guidelines: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md
Please fill out the items below.
-->
# Environment
```none
Windows build number: Microsoft Windows [Version 10.0.20170.1000]
Your Distribution version:
PRETTY_NAME="Debian GNU/Linux 10 (buster)"
NAME="Debian GNU/Linux"
VERSION_ID="10"
VERSION="10 (buster)"
VERSION_CODENAME=buster
ID=debian
WSL 2
```
# Steps to reproduce
<!-- What you're doing and what's happening. Copy&paste the full set of specific command-line steps necessary to reproduce the behavior, and their output. Include screenshots if that helps demonstrate the problem. -->
<!--
If you'd like to provide logs you can provide an `strace(1)` log of the failing command (if `some_command` is failing, then run `strace -o some_command.strace -f some_command some_args`, and link the contents of `some_command.strace` in a gist.
More info on `strace` can be found here: https://www.man7.org/linux/man-pages/man1/strace.1.html
You can use Github gists to share the output: https://gist.github.com/
Additionally, For WSL launch issues, please collect detailed logs, instructions here: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md#8-detailed-logs
-->
# Expected behavior
<!-- A description of what you're expecting, possibly containing screenshots or reference material. -->
# Actual behavior
<!-- What's actually happening? -->
| 1.0 | generateResolvConf = false did not work in wsl.conf - <!--
🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨
I ACKNOWLEDGE THE FOLLOWING BEFORE PROCEEDING:
1. If I delete this entire template and go my own path, the core team may close my issue without further explanation or engagement.
2. If I list multiple bugs/concerns in this one issue, the core team may close my issue without further explanation or engagement.
3. If I write an issue that has many duplicates, the core team may close my issue without further explanation or engagement (and without necessarily spending time to find the exact duplicate ID number).
4. If I leave the title incomplete when filing the issue, the core team may close my issue without further explanation or engagement.
5. If I file something completely blank in the body, the core team may close my issue without further explanation or engagement.
All good? Then proceed!
-->
I have WSL2 with debian -all good.
I created a /etc/wsl.conf file
it contains
```
[network]
generateResolvConf = false
```
as per https://docs.microsoft.com/en-us/windows/wsl/wsl-config#network
if i delete /etc/resolv.conf (or edit and save) on start of the next debia WSL2 session it has been regenerated again.
running WSL2 in windows 10 build 20170.rs_prerelease.200710-1419
<!--
This bug tracker is monitored by Windows Subsystem for Linux development team and other technical folks.
Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues.
Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. Ideally, please configure your machine to capture minidumps, repro the issue, and send the minidump from "C:\Windows\minidump\".
You can find instructions to do that here: https://support.microsoft.com/en-us/help/315263/how-to-read-the-small-memory-dump-file-that-is-created-by-windows-if-a
If this is a console issue (a problem with layout, rendering, colors, etc.), please post the issue to the Terminal tracker: https://github.com/microsoft/terminal/issues
For documentation improvements, please post to the documentation tracker: https://github.com/MicrosoftDocs/WSL/issues
For any other questions on contributing please see our contribution guidelines: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md
Please fill out the items below.
-->
# Environment
```none
Windows build number: Microsoft Windows [Version 10.0.20170.1000]
Your Distribution version:
PRETTY_NAME="Debian GNU/Linux 10 (buster)"
NAME="Debian GNU/Linux"
VERSION_ID="10"
VERSION="10 (buster)"
VERSION_CODENAME=buster
ID=debian
WSL 2
```
# Steps to reproduce
<!-- What you're doing and what's happening. Copy&paste the full set of specific command-line steps necessary to reproduce the behavior, and their output. Include screenshots if that helps demonstrate the problem. -->
<!--
If you'd like to provide logs you can provide an `strace(1)` log of the failing command (if `some_command` is failing, then run `strace -o some_command.strace -f some_command some_args`, and link the contents of `some_command.strace` in a gist.
More info on `strace` can be found here: https://www.man7.org/linux/man-pages/man1/strace.1.html
You can use Github gists to share the output: https://gist.github.com/
Additionally, For WSL launch issues, please collect detailed logs, instructions here: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md#8-detailed-logs
-->
# Expected behavior
<!-- A description of what you're expecting, possibly containing screenshots or reference material. -->
# Actual behavior
<!-- What's actually happening? -->
| non_priority | generateresolvconf false did not work in wsl conf 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 i acknowledge the following before proceeding if i delete this entire template and go my own path the core team may close my issue without further explanation or engagement if i list multiple bugs concerns in this one issue the core team may close my issue without further explanation or engagement if i write an issue that has many duplicates the core team may close my issue without further explanation or engagement and without necessarily spending time to find the exact duplicate id number if i leave the title incomplete when filing the issue the core team may close my issue without further explanation or engagement if i file something completely blank in the body the core team may close my issue without further explanation or engagement all good then proceed i have with debian all good i created a etc wsl conf file it contains generateresolvconf false as per if i delete etc resolv conf or edit and save on start of the next debia session it has been regenerated again running in windows build rs prerelease this bug tracker is monitored by windows subsystem for linux development team and other technical folks important when reporting bsods or security issues do not attach memory dumps logs or traces to github issues instead send dumps traces to secure microsoft com referencing this github issue ideally please configure your machine to capture minidumps repro the issue and send the minidump from c windows minidump you can find instructions to do that here if this is a console issue a problem with layout rendering colors etc please post the issue to the terminal tracker for documentation improvements please post to the documentation tracker for any other questions on contributing please see our contribution guidelines please fill out the items below environment none windows build number microsoft windows your distribution version pretty name debian gnu linux buster name debian gnu linux version id version buster version codename buster id debian wsl steps to reproduce if you d like to provide logs you can provide an strace log of the failing command if some command is failing then run strace o some command strace f some command some args and link the contents of some command strace in a gist more info on strace can be found here you can use github gists to share the output additionally for wsl launch issues please collect detailed logs instructions here expected behavior actual behavior | 0 |
790,635 | 27,830,950,291 | IssuesEvent | 2023-03-20 04:49:49 | AY2223S2-CS2103T-W14-1/tp | https://api.github.com/repos/AY2223S2-CS2103T-W14-1/tp | closed | Refactor UI package to be more extensible | priority.med type.refactor severity.Medium | UI package is a messy platter of java files with little to no organisation. | 1.0 | Refactor UI package to be more extensible - UI package is a messy platter of java files with little to no organisation. | priority | refactor ui package to be more extensible ui package is a messy platter of java files with little to no organisation | 1 |
75,802 | 3,476,131,524 | IssuesEvent | 2015-12-26 14:23:50 | Stephane-D/SGDK | https://api.github.com/repos/Stephane-D/SGDK | closed | recomp sprite collision, wrong label name | Priority-Medium | Labels name are generated multiple time:
see
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1247
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1259
using this .res file:
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-res | 1.0 | recomp sprite collision, wrong label name - Labels name are generated multiple time:
see
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1247
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1259
using this .res file:
https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-res | priority | recomp sprite collision wrong label name labels name are generated multiple time see using this res file | 1 |
351,054 | 25,011,721,862 | IssuesEvent | 2022-11-03 15:41:27 | deforum-art/deforum-for-automatic1111-webui | https://api.github.com/repos/deforum-art/deforum-for-automatic1111-webui | closed | Add tooltips | documentation enhancement good first issue help wanted | Tooltips would definitely help the newcomers, but the list is pretty large, so if you have some time to spare, I'd welcome your contributions. The list of tooltips is a simple JSON file in a form `"label":"tooltip"`. The names of the widgets lie in https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum.py and https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum/args.py.
Examples from the main webui
```
titles = {
"Sampling steps": "How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results",
"Sampling method": "Which algorithm to use to produce the image",
"GFPGAN": "Restore low quality faces using GFPGAN neural network",
"Euler a": "Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help",
"DDIM": "Denoising Diffusion Implicit Models - best at inpainting",
}
``` | 1.0 | Add tooltips - Tooltips would definitely help the newcomers, but the list is pretty large, so if you have some time to spare, I'd welcome your contributions. The list of tooltips is a simple JSON file in a form `"label":"tooltip"`. The names of the widgets lie in https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum.py and https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum/args.py.
Examples from the main webui
```
titles = {
"Sampling steps": "How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results",
"Sampling method": "Which algorithm to use to produce the image",
"GFPGAN": "Restore low quality faces using GFPGAN neural network",
"Euler a": "Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help",
"DDIM": "Denoising Diffusion Implicit Models - best at inpainting",
}
``` | non_priority | add tooltips tooltips would definitely help the newcomers but the list is pretty large so if you have some time to spare i d welcome your contributions the list of tooltips is a simple json file in a form label tooltip the names of the widgets lie in and examples from the main webui titles sampling steps how many times to improve the generated image iteratively higher values take longer very low values can produce bad results sampling method which algorithm to use to produce the image gfpgan restore low quality faces using gfpgan neural network euler a euler ancestral very creative each can get a completely different picture depending on step count setting steps to higher than does not help ddim denoising diffusion implicit models best at inpainting | 0 |
355,778 | 10,584,634,504 | IssuesEvent | 2019-10-08 15:48:53 | angular/angular-cli | https://api.github.com/repos/angular/angular-cli | closed | Unable to spyOn Math.floor | priority: 2 (required) type: bug/fix | ### Versions
```
@angular/cli: 1.4.9
node: 6.9.3
os: win32 x64
@angular/animations: 4.4.6
@angular/cdk: 2.0.0-beta.12
@angular/common: 4.4.6
@angular/compiler: 4.4.6
@angular/core: 4.4.6
@angular/forms: 4.4.6
@angular/flex-layout: 2.0.0-beta.10-4905443
@angular/http: 4.4.6
@angular/material: 2.0.0-beta.12
@angular/platform-browser: 4.4.6
@angular/platform-browser-dynamic: 4.4.6
@angular/router: 4.4.6
@angular/cli: 1.4.9
@angular/compiler-cli: 4.4.6
@angular/language-service: 4.4.6
typescript: 2.5.3
```
### Repro steps
Create a simple test and attempt to do `spyOn(Math, 'floor').and.returnValue(10);
Run ng test
### Observed behavior
```
RangeError: Maximum call stack size exceeded
at c (node_modules/source-map-support/browser-source-map-support.js:53:483)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
```
### Desired behavior
I should be able to spyOn Math.floor similar to other Math functions
### Mention any other details that might be useful (optional)
Spying on Math.ceil works as expected and running without source maps `ng test --sm=false` also works.
| 1.0 | Unable to spyOn Math.floor - ### Versions
```
@angular/cli: 1.4.9
node: 6.9.3
os: win32 x64
@angular/animations: 4.4.6
@angular/cdk: 2.0.0-beta.12
@angular/common: 4.4.6
@angular/compiler: 4.4.6
@angular/core: 4.4.6
@angular/forms: 4.4.6
@angular/flex-layout: 2.0.0-beta.10-4905443
@angular/http: 4.4.6
@angular/material: 2.0.0-beta.12
@angular/platform-browser: 4.4.6
@angular/platform-browser-dynamic: 4.4.6
@angular/router: 4.4.6
@angular/cli: 1.4.9
@angular/compiler-cli: 4.4.6
@angular/language-service: 4.4.6
typescript: 2.5.3
```
### Repro steps
Create a simple test and attempt to do `spyOn(Math, 'floor').and.returnValue(10);
Run ng test
### Observed behavior
```
RangeError: Maximum call stack size exceeded
at c (node_modules/source-map-support/browser-source-map-support.js:53:483)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
at c (node_modules/source-map-support/browser-source-map-support.js:54:101)
```
### Desired behavior
I should be able to spyOn Math.floor similar to other Math functions
### Mention any other details that might be useful (optional)
Spying on Math.ceil works as expected and running without source maps `ng test --sm=false` also works.
| priority | unable to spyon math floor versions angular cli node os angular animations angular cdk beta angular common angular compiler angular core angular forms angular flex layout beta angular http angular material beta angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli angular language service typescript repro steps create a simple test and attempt to do spyon math floor and returnvalue run ng test observed behavior rangeerror maximum call stack size exceeded at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js desired behavior i should be able to spyon math floor similar to other math functions mention any other details that might be useful optional spying on math ceil works as expected and running without source maps ng test sm false also works | 1 |
2,093 | 4,381,721,899 | IssuesEvent | 2016-08-06 12:14:02 | support-project/knowledge | https://api.github.com/repos/support-project/knowledge | closed | ファイルサーバへのリンク | [Status] 3.merged [Type] 1.requirement | Markdown本文の中で、ファイルサーバーへのリンクを記載したい。
以下のようなネットワークディレクトリの表記を行うと、Markdownのパース処理か
XSS対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという
報告があった。
\\server_host\hoge\fuga\sample.txt
file://server_host/hoge/fuga/sample.txt
| 1.0 | ファイルサーバへのリンク - Markdown本文の中で、ファイルサーバーへのリンクを記載したい。
以下のようなネットワークディレクトリの表記を行うと、Markdownのパース処理か
XSS対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという
報告があった。
\\server_host\hoge\fuga\sample.txt
file://server_host/hoge/fuga/sample.txt
| non_priority | ファイルサーバへのリンク markdown本文の中で、ファイルサーバーへのリンクを記載したい。 以下のようなネットワークディレクトリの表記を行うと、markdownのパース処理か xss対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという 報告があった。 server host hoge fuga sample txt file server host hoge fuga sample txt | 0 |
777,049 | 27,266,788,975 | IssuesEvent | 2023-02-22 18:41:57 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | MSRuleCleaner potentially not cleaning unlocked input dataset rules | BUG High Priority MSRuleCleaner | **Impact of the bug**
MSRuleCleaner
**Describe the bug**
A possible issue with MSRuleCleaner has been reported in the past week, where it leaves Rucio rules behind for some of the input datasets. Amanda initially provided these 3 RelVal input datasets (blocks):
```
/RelValZMM_14/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#621aab39-d57d-4308-b38c-fa8bfeb65fdb
/RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun4_realistic_v8_2026D88noPU-v1/GEN-SIM-DIGI-RAW#fef08bee-3d76-485d-b4b4-68ed0b181b70
/RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#4c88cda3-eebb-4e7f-9860-ef201108d9d8
```
which are still locked under the account `wmcore_transferor`, even though all the relevant workflows processing it are no longer active in the system (they have been archived).
A non-relval input dataset example is:
https://cmsweb.cern.ch/reqmgr2/data/request?inputdataset=/BTohh_hToMuNu_BsBdMixture_modHadLifetime_TuneCP5_13TeV-pythia8-evtgen/RunIISummer20UL17RECO-106X_mc2017_realistic_v6-v3/AODSIM
which has been archived in December 2021, leaving these two rucio rules behind:
```
c55dcaa65f2b443b9824fff1983228e1
61fa02b8b4694f27b80f91adec4d1591
```
**How to reproduce it**
No clue
**Expected behavior**
From a quick debugging that I have done so far, MSRuleCleaner seem to be properly deleting rules for the input dataset. So there might be a specific scenario that triggers this ill behaviour.
On what concerns this issue, we should:
* improve logging, such that we can explicitly see whenever an input data is looked up, and under which classification (primary, parent or secondary);
* ensure that rules allowed for deletion are actually removed, especially for input dataset
**Additional context and error message**
None
| 1.0 | MSRuleCleaner potentially not cleaning unlocked input dataset rules - **Impact of the bug**
MSRuleCleaner
**Describe the bug**
A possible issue with MSRuleCleaner has been reported in the past week, where it leaves Rucio rules behind for some of the input datasets. Amanda initially provided these 3 RelVal input datasets (blocks):
```
/RelValZMM_14/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#621aab39-d57d-4308-b38c-fa8bfeb65fdb
/RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun4_realistic_v8_2026D88noPU-v1/GEN-SIM-DIGI-RAW#fef08bee-3d76-485d-b4b4-68ed0b181b70
/RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#4c88cda3-eebb-4e7f-9860-ef201108d9d8
```
which are still locked under the account `wmcore_transferor`, even though all the relevant workflows processing it are no longer active in the system (they have been archived).
A non-relval input dataset example is:
https://cmsweb.cern.ch/reqmgr2/data/request?inputdataset=/BTohh_hToMuNu_BsBdMixture_modHadLifetime_TuneCP5_13TeV-pythia8-evtgen/RunIISummer20UL17RECO-106X_mc2017_realistic_v6-v3/AODSIM
which has been archived in December 2021, leaving these two rucio rules behind:
```
c55dcaa65f2b443b9824fff1983228e1
61fa02b8b4694f27b80f91adec4d1591
```
**How to reproduce it**
No clue
**Expected behavior**
From a quick debugging that I have done so far, MSRuleCleaner seem to be properly deleting rules for the input dataset. So there might be a specific scenario that triggers this ill behaviour.
On what concerns this issue, we should:
* improve logging, such that we can explicitly see whenever an input data is looked up, and under which classification (primary, parent or secondary);
* ensure that rules allowed for deletion are actually removed, especially for input dataset
**Additional context and error message**
None
| priority | msrulecleaner potentially not cleaning unlocked input dataset rules impact of the bug msrulecleaner describe the bug a possible issue with msrulecleaner has been reported in the past week where it leaves rucio rules behind for some of the input datasets amanda initially provided these relval input datasets blocks relvalzmm cmssw realistic gen sim digi raw relvalttbar cmssw realistic gen sim digi raw relvalttbar cmssw realistic gen sim digi raw eebb which are still locked under the account wmcore transferor even though all the relevant workflows processing it are no longer active in the system they have been archived a non relval input dataset example is which has been archived in december leaving these two rucio rules behind how to reproduce it no clue expected behavior from a quick debugging that i have done so far msrulecleaner seem to be properly deleting rules for the input dataset so there might be a specific scenario that triggers this ill behaviour on what concerns this issue we should improve logging such that we can explicitly see whenever an input data is looked up and under which classification primary parent or secondary ensure that rules allowed for deletion are actually removed especially for input dataset additional context and error message none | 1 |
47,977 | 25,288,583,756 | IssuesEvent | 2022-11-16 21:36:37 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | opened | optimize_move_to_prewhere does not work with Engine=merge -> distributed -> mergeTree | performance | ```sql
drop table if exists foo_merge;
drop table if exists foo_merge_distributed;
drop table if exists foo_distributed;
drop table if exists foo_distributed_merge;
drop table if exists foo;
create table foo(A Int64, V Int64, S String) Engine=MergeTree order by A;
insert into foo select number, sipHash64(number), toString(arrayMap(x->sipHash64(number*x), range(10))) from numbers(10000000);
-- merge -> distributed -> mergeTree
create table foo_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo);
create table foo_distributed_merge as foo ENGINE=Merge(currentDatabase(), '^foo_distributed$');
-- No optimize_move_to_prewhere :(
select * from foo_distributed_merge where V = 666;
0 rows in set. Elapsed: 0.530 sec. Processed 10.00 million rows, 2.31 GB (18.85 million rows/s., 4.35 GB/s.)
-- manual prewhere
select * from foo_distributed_merge prewhere V = 666;
0 rows in set. Elapsed: 0.016 sec. Processed 10.00 million rows, 80.00 MB (617.71 million rows/s., 4.94 GB/s.)
-- distributed -> merge -> mergeTree
create table foo_merge as foo ENGINE=Merge(currentDatabase(), '^foo$');
create table foo_merge_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo_merge);
-- optimize_move_to_prewhere :)
select * from foo_merge_distributed where V = 666;
0 rows in set. Elapsed: 0.024 sec. Processed 10.00 million rows, 80.00 MB (411.88 million rows/s., 3.30 GB/s.)
```
| True | optimize_move_to_prewhere does not work with Engine=merge -> distributed -> mergeTree - ```sql
drop table if exists foo_merge;
drop table if exists foo_merge_distributed;
drop table if exists foo_distributed;
drop table if exists foo_distributed_merge;
drop table if exists foo;
create table foo(A Int64, V Int64, S String) Engine=MergeTree order by A;
insert into foo select number, sipHash64(number), toString(arrayMap(x->sipHash64(number*x), range(10))) from numbers(10000000);
-- merge -> distributed -> mergeTree
create table foo_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo);
create table foo_distributed_merge as foo ENGINE=Merge(currentDatabase(), '^foo_distributed$');
-- No optimize_move_to_prewhere :(
select * from foo_distributed_merge where V = 666;
0 rows in set. Elapsed: 0.530 sec. Processed 10.00 million rows, 2.31 GB (18.85 million rows/s., 4.35 GB/s.)
-- manual prewhere
select * from foo_distributed_merge prewhere V = 666;
0 rows in set. Elapsed: 0.016 sec. Processed 10.00 million rows, 80.00 MB (617.71 million rows/s., 4.94 GB/s.)
-- distributed -> merge -> mergeTree
create table foo_merge as foo ENGINE=Merge(currentDatabase(), '^foo$');
create table foo_merge_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo_merge);
-- optimize_move_to_prewhere :)
select * from foo_merge_distributed where V = 666;
0 rows in set. Elapsed: 0.024 sec. Processed 10.00 million rows, 80.00 MB (411.88 million rows/s., 3.30 GB/s.)
```
| non_priority | optimize move to prewhere does not work with engine merge distributed mergetree sql drop table if exists foo merge drop table if exists foo merge distributed drop table if exists foo distributed drop table if exists foo distributed merge drop table if exists foo create table foo a v s string engine mergetree order by a insert into foo select number number tostring arraymap x number x range from numbers merge distributed mergetree create table foo distributed as foo engine distributed test shard localhost currentdatabase foo create table foo distributed merge as foo engine merge currentdatabase foo distributed no optimize move to prewhere select from foo distributed merge where v rows in set elapsed sec processed million rows gb million rows s gb s manual prewhere select from foo distributed merge prewhere v rows in set elapsed sec processed million rows mb million rows s gb s distributed merge mergetree create table foo merge as foo engine merge currentdatabase foo create table foo merge distributed as foo engine distributed test shard localhost currentdatabase foo merge optimize move to prewhere select from foo merge distributed where v rows in set elapsed sec processed million rows mb million rows s gb s | 0 |
257,649 | 8,140,033,770 | IssuesEvent | 2018-08-20 19:42:26 | nprapps/elections18-graphics | https://api.github.com/repos/nprapps/elections18-graphics | closed | Run an informational member-station webinar | effort:medium priority:high | We'll be working with Janeen and Franklin to coordinate a webinar with member station staff. At this webinar, we'll present our existing design work, and show how it could be embedded on their site as a standalone page or widget. We'll also solicit feedback from the member stations, and incorporate these into the issue matrix (priority vs effort).
cc @constancebmiller | 1.0 | Run an informational member-station webinar - We'll be working with Janeen and Franklin to coordinate a webinar with member station staff. At this webinar, we'll present our existing design work, and show how it could be embedded on their site as a standalone page or widget. We'll also solicit feedback from the member stations, and incorporate these into the issue matrix (priority vs effort).
cc @constancebmiller | priority | run an informational member station webinar we ll be working with janeen and franklin to coordinate a webinar with member station staff at this webinar we ll present our existing design work and show how it could be embedded on their site as a standalone page or widget we ll also solicit feedback from the member stations and incorporate these into the issue matrix priority vs effort cc constancebmiller | 1 |
502,887 | 14,569,265,032 | IssuesEvent | 2020-12-17 12:51:02 | staxrip/staxrip | https://api.github.com/repos/staxrip/staxrip | closed | AOMENC Better defaults with explanations. | feature request fixed priority medium | Change:
Basic > Mode: Change from One-pass to Two-pass
Reason(s):
1. first pass is usually always very fast
2. Aomenc does not use alt ref frames when performing a single pass, therefore it is generally recommended to always perform a 2-pass encode, no matter what Rate Mode is chosen.
------------
Change:
Basic > Rate Mode:
1. Rename "Q" to "Q (Constant Quality)"
2. Rename "CQ" to "CQ (Constrained Quality)"
Reason(s):
1. This will help eliminate confusion because users coming from x265 might choose "CQ", expecting it to behave similar to x265's CRF (CQ), but it will result in a terrible quality encode due to it producing a constrained quality video.
------------
Change:
Basic > Rate Mode: Change from VBR to Q
Reason(s):
1. Aomenc is less likely to be used for targeting precise filesizes, and more likely to be used for targeting quality at imprecise, yet small filesizes.
------------
Change:
Basic > Bit Depth: Change from 8 to 10
Reason(s):
1. 10bit is basically the new norm as, there are almost no new devices/software that do not support 10bit decoding
2. 10bit almost always produces a better quality encode with smoother gradients in skies.
3. It's safer to do 10bit final encode than 8bit, even if converting an 8bit source video into a 10bit result. There will almost always be fewer anomalies going with a 10bit final encode.
------------
Change:
Basic > CPU Used: Change from 8 to 4 (Medium)
Reason(s):
1. 8 Will produce a very low quality encode with poor VMAF scores in a lot of cases.
2. 4 is a good balance in terms of speed and quality and produces good VMAF scores.
3. Charts show that cpu speeds 0-4 all have very close VMAF scores, but cpu speeds 5-8 begin producing much worse VMAF scores, therefore CPU 4 should be the default (in my opinion)
------------
Change:
Basic > CQ Level: Change from 0 to 25
Reason(s):
1. 0 is lossless, and lossless should never be the default for a lossy encoder. This shouldn't require a huge explanation, as the reasons should be obvious.
2. 25 is a good default quality level since it produces a high quality encode at a small filesize. AV1 excels at higher CQ levels compared to x265, and AV1 is primarily used to create low filesize encodes at a target quality. AV1 isn't well optimized for low CQ levels due to quantizer related denoising anomalies (especially on grainy content). The goal of AV1 is to denoise / remove granular details without becoming a blocky, swimming mess, and it is better able to achieve that goal at higher CQ levels, until some kinks have been ironed out. Believe me, I would LOVE to use a lower CQ level as a default, but CQ 30 just, across the board, produces more consistent results.
3. CQ Level 25 provides good quality for both 8bit and native 10bit content. According to user BlueSwordM, native 10bit content benefits more from a CQ level of 25 compared with higher CQ levels. 25 is going to be a safe starting point for content of either bit depth.
------------
Change:
Analysis > Tile Columns: Change from 0 to 2
Analysis > Tile Rows: Change from 0 to 1
Reason(s):
1. Assists with multi-threaded encoding by better utilizing cores on your cpu, with a downside of a small hit to encoding compression. However, many AV1 users agree that the speed gained is worth the tradeoff.
3. Reddit user said this:
"By keeping tiles as "square" as possible, you have the lowest edge/area ratio. Edges cause a small loss in coding efficiency.
So for a 16:9 video, 2x1, 2x2 or 4x2 is recommended depending on resolution. Most times 2x1 is good for 1080p, 2x2 for 1440p and 4x2 for 2160p and higher."
4. Since most content being encoded is still 1080p, then 2x1 is probably a safe default.
------------
Change:
Slice Decision > Min GOP Size: Rename to Keyframe Min Distance
Slice Decision > Max GOP Size: Rename to Keyframe Max Distance
Reason(s):
1. This will better reflect the parameter name since AV1 refers to this as keyframe distance, rather than "Group of Picture" size
2. Easier to understand without inexperienced users having to resort to google to figure out what GOP means.
------------
Change:
Slice Decision > Max GOP Size (i.e. Keyframe Max Distance): Change from 0 to 240
Reason(s):
1. I heard AV1 users at one point discussing this number being recommended by the AV1 developers, and also I believe this is a recommended number even for VP9.
2. It represents approximately maximum of 10 seconds between keyframes (if your framerate is approximately 24fps), otherwise keyframes could be placed at "who knows". Better to specify than to leave it unspecified, since the AV1 parameter doesn't say what the default value is.
------------
Change:
Slice Decision > Lag in Frames: Change from 0 to 25
Reason(s):
1. --lag-in-frames parameter defines an upper limit on the number of frames into the future that the encoder can look
2. I have seen people preferring 25 for AV1.
3. For VP9, a value of 16 was recommended. I have no idea what the recommended or the max value is in AV1, but it seems safe to default this to 25.
------------
Change:
Performance > Add a checkbox called "Automatically determine thread count". Enabling this would gray out/disallow user input for "Threads", and the Thread count would be automatically determined based on the CPU's multi-threading capability.
Reason(s):
1. 0 doesn't actually cause aomenc to auto multi thread. 0 disables multi-threading.
2. Sloppy Workaround: You could try setting default thread count to 8 since that's a safe number in 2020, since most people doing video conversion in 2020 have at least a quad core with hyper threading, however it would be best if StaxRip could automatically determine the cpu's maximum thread capability and automatically set this number accordingly.
------------
Change:
Misc 2 > Auto Alt Ref: Change from 0 to 1
Reason(s):
1. It is better to have "auto alt ref" Enabled because according to some AV1 users in discord, having it disabled will result in a "completely shit encode"
2. If you're doing a single pass encode, auto-alt-ref is actually disabled by default, but you can force it to be enabled by forcibly specifying it.
3. alt-ref frames are used when using a 2-pass encode by default if it is not specified. Based on my limited understanding, you are generally better off just forcibly enabling this setting for most case scenarios. | 1.0 | AOMENC Better defaults with explanations. - Change:
Basic > Mode: Change from One-pass to Two-pass
Reason(s):
1. first pass is usually always very fast
2. Aomenc does not use alt ref frames when performing a single pass, therefore it is generally recommended to always perform a 2-pass encode, no matter what Rate Mode is chosen.
------------
Change:
Basic > Rate Mode:
1. Rename "Q" to "Q (Constant Quality)"
2. Rename "CQ" to "CQ (Constrained Quality)"
Reason(s):
1. This will help eliminate confusion because users coming from x265 might choose "CQ", expecting it to behave similar to x265's CRF (CQ), but it will result in a terrible quality encode due to it producing a constrained quality video.
------------
Change:
Basic > Rate Mode: Change from VBR to Q
Reason(s):
1. Aomenc is less likely to be used for targeting precise filesizes, and more likely to be used for targeting quality at imprecise, yet small filesizes.
------------
Change:
Basic > Bit Depth: Change from 8 to 10
Reason(s):
1. 10bit is basically the new norm as, there are almost no new devices/software that do not support 10bit decoding
2. 10bit almost always produces a better quality encode with smoother gradients in skies.
3. It's safer to do 10bit final encode than 8bit, even if converting an 8bit source video into a 10bit result. There will almost always be fewer anomalies going with a 10bit final encode.
------------
Change:
Basic > CPU Used: Change from 8 to 4 (Medium)
Reason(s):
1. 8 Will produce a very low quality encode with poor VMAF scores in a lot of cases.
2. 4 is a good balance in terms of speed and quality and produces good VMAF scores.
3. Charts show that cpu speeds 0-4 all have very close VMAF scores, but cpu speeds 5-8 begin producing much worse VMAF scores, therefore CPU 4 should be the default (in my opinion)
------------
Change:
Basic > CQ Level: Change from 0 to 25
Reason(s):
1. 0 is lossless, and lossless should never be the default for a lossy encoder. This shouldn't require a huge explanation, as the reasons should be obvious.
2. 25 is a good default quality level since it produces a high quality encode at a small filesize. AV1 excels at higher CQ levels compared to x265, and AV1 is primarily used to create low filesize encodes at a target quality. AV1 isn't well optimized for low CQ levels due to quantizer related denoising anomalies (especially on grainy content). The goal of AV1 is to denoise / remove granular details without becoming a blocky, swimming mess, and it is better able to achieve that goal at higher CQ levels, until some kinks have been ironed out. Believe me, I would LOVE to use a lower CQ level as a default, but CQ 30 just, across the board, produces more consistent results.
3. CQ Level 25 provides good quality for both 8bit and native 10bit content. According to user BlueSwordM, native 10bit content benefits more from a CQ level of 25 compared with higher CQ levels. 25 is going to be a safe starting point for content of either bit depth.
------------
Change:
Analysis > Tile Columns: Change from 0 to 2
Analysis > Tile Rows: Change from 0 to 1
Reason(s):
1. Assists with multi-threaded encoding by better utilizing cores on your cpu, with a downside of a small hit to encoding compression. However, many AV1 users agree that the speed gained is worth the tradeoff.
3. Reddit user said this:
"By keeping tiles as "square" as possible, you have the lowest edge/area ratio. Edges cause a small loss in coding efficiency.
So for a 16:9 video, 2x1, 2x2 or 4x2 is recommended depending on resolution. Most times 2x1 is good for 1080p, 2x2 for 1440p and 4x2 for 2160p and higher."
4. Since most content being encoded is still 1080p, then 2x1 is probably a safe default.
------------
Change:
Slice Decision > Min GOP Size: Rename to Keyframe Min Distance
Slice Decision > Max GOP Size: Rename to Keyframe Max Distance
Reason(s):
1. This will better reflect the parameter name since AV1 refers to this as keyframe distance, rather than "Group of Picture" size
2. Easier to understand without inexperienced users having to resort to google to figure out what GOP means.
------------
Change:
Slice Decision > Max GOP Size (i.e. Keyframe Max Distance): Change from 0 to 240
Reason(s):
1. I heard AV1 users at one point discussing this number being recommended by the AV1 developers, and also I believe this is a recommended number even for VP9.
2. It represents approximately maximum of 10 seconds between keyframes (if your framerate is approximately 24fps), otherwise keyframes could be placed at "who knows". Better to specify than to leave it unspecified, since the AV1 parameter doesn't say what the default value is.
------------
Change:
Slice Decision > Lag in Frames: Change from 0 to 25
Reason(s):
1. --lag-in-frames parameter defines an upper limit on the number of frames into the future that the encoder can look
2. I have seen people preferring 25 for AV1.
3. For VP9, a value of 16 was recommended. I have no idea what the recommended or the max value is in AV1, but it seems safe to default this to 25.
------------
Change:
Performance > Add a checkbox called "Automatically determine thread count". Enabling this would gray out/disallow user input for "Threads", and the Thread count would be automatically determined based on the CPU's multi-threading capability.
Reason(s):
1. 0 doesn't actually cause aomenc to auto multi thread. 0 disables multi-threading.
2. Sloppy Workaround: You could try setting default thread count to 8 since that's a safe number in 2020, since most people doing video conversion in 2020 have at least a quad core with hyper threading, however it would be best if StaxRip could automatically determine the cpu's maximum thread capability and automatically set this number accordingly.
------------
Change:
Misc 2 > Auto Alt Ref: Change from 0 to 1
Reason(s):
1. It is better to have "auto alt ref" Enabled because according to some AV1 users in discord, having it disabled will result in a "completely shit encode"
2. If you're doing a single pass encode, auto-alt-ref is actually disabled by default, but you can force it to be enabled by forcibly specifying it.
3. alt-ref frames are used when using a 2-pass encode by default if it is not specified. Based on my limited understanding, you are generally better off just forcibly enabling this setting for most case scenarios. | priority | aomenc better defaults with explanations change basic mode change from one pass to two pass reason s first pass is usually always very fast aomenc does not use alt ref frames when performing a single pass therefore it is generally recommended to always perform a pass encode no matter what rate mode is chosen change basic rate mode rename q to q constant quality rename cq to cq constrained quality reason s this will help eliminate confusion because users coming from might choose cq expecting it to behave similar to s crf cq but it will result in a terrible quality encode due to it producing a constrained quality video change basic rate mode change from vbr to q reason s aomenc is less likely to be used for targeting precise filesizes and more likely to be used for targeting quality at imprecise yet small filesizes change basic bit depth change from to reason s is basically the new norm as there are almost no new devices software that do not support decoding almost always produces a better quality encode with smoother gradients in skies it s safer to do final encode than even if converting an source video into a result there will almost always be fewer anomalies going with a final encode change basic cpu used change from to medium reason s will produce a very low quality encode with poor vmaf scores in a lot of cases is a good balance in terms of speed and quality and produces good vmaf scores charts show that cpu speeds all have very close vmaf scores but cpu speeds begin producing much worse vmaf scores therefore cpu should be the default in my opinion change basic cq level change from to reason s is lossless and lossless should never be the default for a lossy encoder this shouldn t require a huge explanation as the reasons should be obvious is a good default quality level since it produces a high quality encode at a small filesize excels at higher cq levels compared to and is primarily used to create low filesize encodes at a target quality isn t well optimized for low cq levels due to quantizer related denoising anomalies especially on grainy content the goal of is to denoise remove granular details without becoming a blocky swimming mess and it is better able to achieve that goal at higher cq levels until some kinks have been ironed out believe me i would love to use a lower cq level as a default but cq just across the board produces more consistent results cq level provides good quality for both and native content according to user blueswordm native content benefits more from a cq level of compared with higher cq levels is going to be a safe starting point for content of either bit depth change analysis tile columns change from to analysis tile rows change from to reason s assists with multi threaded encoding by better utilizing cores on your cpu with a downside of a small hit to encoding compression however many users agree that the speed gained is worth the tradeoff reddit user said this by keeping tiles as square as possible you have the lowest edge area ratio edges cause a small loss in coding efficiency so for a video or is recommended depending on resolution most times is good for for and for and higher since most content being encoded is still then is probably a safe default change slice decision min gop size rename to keyframe min distance slice decision max gop size rename to keyframe max distance reason s this will better reflect the parameter name since refers to this as keyframe distance rather than group of picture size easier to understand without inexperienced users having to resort to google to figure out what gop means change slice decision max gop size i e keyframe max distance change from to reason s i heard users at one point discussing this number being recommended by the developers and also i believe this is a recommended number even for it represents approximately maximum of seconds between keyframes if your framerate is approximately otherwise keyframes could be placed at who knows better to specify than to leave it unspecified since the parameter doesn t say what the default value is change slice decision lag in frames change from to reason s lag in frames parameter defines an upper limit on the number of frames into the future that the encoder can look i have seen people preferring for for a value of was recommended i have no idea what the recommended or the max value is in but it seems safe to default this to change performance add a checkbox called automatically determine thread count enabling this would gray out disallow user input for threads and the thread count would be automatically determined based on the cpu s multi threading capability reason s doesn t actually cause aomenc to auto multi thread disables multi threading sloppy workaround you could try setting default thread count to since that s a safe number in since most people doing video conversion in have at least a quad core with hyper threading however it would be best if staxrip could automatically determine the cpu s maximum thread capability and automatically set this number accordingly change misc auto alt ref change from to reason s it is better to have auto alt ref enabled because according to some users in discord having it disabled will result in a completely shit encode if you re doing a single pass encode auto alt ref is actually disabled by default but you can force it to be enabled by forcibly specifying it alt ref frames are used when using a pass encode by default if it is not specified based on my limited understanding you are generally better off just forcibly enabling this setting for most case scenarios | 1 |
24,949 | 24,512,381,144 | IssuesEvent | 2022-10-10 23:25:25 | tailscale/tailscale | https://api.github.com/repos/tailscale/tailscale | closed | Raspberry Pi Libreelec does not login after reboot | L1 Very few P1 Nuisance T5 Usability bug | ### What is the issue?
I have installed Tailscale on Libreelec manually and I don't know if it's working correctly. To install I downloaded the arm64 static binary, unpacked it and made this service to automatically start tailscale on startup:
```
[Unit]
Description=Tailscale node agent
Documentation=https://tailscale.com/kb/
Wants=network-pre.target
After=network-pre.target
StartLimitIntervalSec=0
StartLimitBurst=0
[Service]
EnvironmentFile=/storage/tailscale/tailscaled.defaults
ExecStartPre=/storage/tailscale/tailscaled --cleanup
ExecStart=/storage/tailscale/tailscaled --state=/var/lib/tailscale/tailscaled.state --socket=/run/tailscale/tailscaled.sock --port $PORT $FLAGS
ExecStopPost=/storage/tailscale/tailscaled --cleanup
Restart=on-failure
RuntimeDirectory=tailscale
RuntimeDirectoryMode=0777
StateDirectory=tailscale
StateDirectoryMode=0777
CacheDirectory=tailscale
CacheDirectoryMode=0777
Type=notify
[Install]
WantedBy=multi-user.target
```
Now when I try to login for the first time with `tailscale up` and authorize in the browser it is successful and there seems to be something safed to `/var/lib/tailscale/` as a new folder "files", tailscaled.state and two empty log files are created.
However, when the Pi is rebooted the service runs fine but it seems like it is unable to login as the Pi is not connected. Running `tailscale up` again needs reauthentication and the IP changes which I try to avoid because I want to use a DNS. I tried to authorize with a timeout and a reuseable key at start, this works but it is also always a new IP. What I hope is that it can behave like on a laptop where I can disconnect and connect again without reauthorization, keeping the same IP.
This is the log after rebooting:
[log.txt](https://github.com/tailscale/tailscale/files/9717796/log.txt)
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
Linux
### OS version
Libreelec 10.2
### Tailscale version
1.30.2
### Bug report
BUG-a42c55eaa2754c6172171faab53e660ad278bb51e7f4c1ccf001b2343be12370-20221005174348Z-499137b2ec467d14 | True | Raspberry Pi Libreelec does not login after reboot - ### What is the issue?
I have installed Tailscale on Libreelec manually and I don't know if it's working correctly. To install I downloaded the arm64 static binary, unpacked it and made this service to automatically start tailscale on startup:
```
[Unit]
Description=Tailscale node agent
Documentation=https://tailscale.com/kb/
Wants=network-pre.target
After=network-pre.target
StartLimitIntervalSec=0
StartLimitBurst=0
[Service]
EnvironmentFile=/storage/tailscale/tailscaled.defaults
ExecStartPre=/storage/tailscale/tailscaled --cleanup
ExecStart=/storage/tailscale/tailscaled --state=/var/lib/tailscale/tailscaled.state --socket=/run/tailscale/tailscaled.sock --port $PORT $FLAGS
ExecStopPost=/storage/tailscale/tailscaled --cleanup
Restart=on-failure
RuntimeDirectory=tailscale
RuntimeDirectoryMode=0777
StateDirectory=tailscale
StateDirectoryMode=0777
CacheDirectory=tailscale
CacheDirectoryMode=0777
Type=notify
[Install]
WantedBy=multi-user.target
```
Now when I try to login for the first time with `tailscale up` and authorize in the browser it is successful and there seems to be something safed to `/var/lib/tailscale/` as a new folder "files", tailscaled.state and two empty log files are created.
However, when the Pi is rebooted the service runs fine but it seems like it is unable to login as the Pi is not connected. Running `tailscale up` again needs reauthentication and the IP changes which I try to avoid because I want to use a DNS. I tried to authorize with a timeout and a reuseable key at start, this works but it is also always a new IP. What I hope is that it can behave like on a laptop where I can disconnect and connect again without reauthorization, keeping the same IP.
This is the log after rebooting:
[log.txt](https://github.com/tailscale/tailscale/files/9717796/log.txt)
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
Linux
### OS version
Libreelec 10.2
### Tailscale version
1.30.2
### Bug report
BUG-a42c55eaa2754c6172171faab53e660ad278bb51e7f4c1ccf001b2343be12370-20221005174348Z-499137b2ec467d14 | non_priority | raspberry pi libreelec does not login after reboot what is the issue i have installed tailscale on libreelec manually and i don t know if it s working correctly to install i downloaded the static binary unpacked it and made this service to automatically start tailscale on startup description tailscale node agent documentation wants network pre target after network pre target startlimitintervalsec startlimitburst environmentfile storage tailscale tailscaled defaults execstartpre storage tailscale tailscaled cleanup execstart storage tailscale tailscaled state var lib tailscale tailscaled state socket run tailscale tailscaled sock port port flags execstoppost storage tailscale tailscaled cleanup restart on failure runtimedirectory tailscale runtimedirectorymode statedirectory tailscale statedirectorymode cachedirectory tailscale cachedirectorymode type notify wantedby multi user target now when i try to login for the first time with tailscale up and authorize in the browser it is successful and there seems to be something safed to var lib tailscale as a new folder files tailscaled state and two empty log files are created however when the pi is rebooted the service runs fine but it seems like it is unable to login as the pi is not connected running tailscale up again needs reauthentication and the ip changes which i try to avoid because i want to use a dns i tried to authorize with a timeout and a reuseable key at start this works but it is also always a new ip what i hope is that it can behave like on a laptop where i can disconnect and connect again without reauthorization keeping the same ip this is the log after rebooting steps to reproduce no response are there any recent changes that introduced the issue no response os linux os version libreelec tailscale version bug report bug | 0 |
134,855 | 19,399,949,497 | IssuesEvent | 2021-12-19 01:25:12 | hackforla/website | https://api.github.com/repos/hackforla/website | closed | Create template for Reviewing the alt text on all pages | Status: Updated Feature: Accessibility role: design Size: Small | ### Overview
We need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way.
### Action Items
- [x] Create template draft in comment below
- [x] Talk to Matthew to find out if dev is going to make this list
- [x] revise after Matthew conversation
- [x] review with Bonnie
- [x] have Bonnie help you make it into a template on the site
- [x] Make for all relevant issues
- [ ] List the issues under resources with checkboxes
### Resources/Instructions
- [Entire worksheet Google Sheet](https://docs.google.com/spreadsheets/d/1abMlPOX88zHTV29E4BFxAYwahCxgJN28Mn95cXr8jak/edit#gid=308160713)
- Issue we are modeling this template after: https://github.com/hackforla/website/issues/2354 | 1.0 | Create template for Reviewing the alt text on all pages - ### Overview
We need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way.
### Action Items
- [x] Create template draft in comment below
- [x] Talk to Matthew to find out if dev is going to make this list
- [x] revise after Matthew conversation
- [x] review with Bonnie
- [x] have Bonnie help you make it into a template on the site
- [x] Make for all relevant issues
- [ ] List the issues under resources with checkboxes
### Resources/Instructions
- [Entire worksheet Google Sheet](https://docs.google.com/spreadsheets/d/1abMlPOX88zHTV29E4BFxAYwahCxgJN28Mn95cXr8jak/edit#gid=308160713)
- Issue we are modeling this template after: https://github.com/hackforla/website/issues/2354 | non_priority | create template for reviewing the alt text on all pages overview we need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way action items create template draft in comment below talk to matthew to find out if dev is going to make this list revise after matthew conversation review with bonnie have bonnie help you make it into a template on the site make for all relevant issues list the issues under resources with checkboxes resources instructions issue we are modeling this template after | 0 |
74,352 | 25,080,647,581 | IssuesEvent | 2022-11-07 19:00:27 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | opened | Strictness of event dispatch result for ORM checkRules() | defect | ### Description
Error
> Cake\ORM\Table::checkRules(): Return value must be of type bool, null returned
As the signature of that method is:
```php
public function checkRules(...): bool
```
The code inside checkRules() reads
```php
$event = $this->dispatchEvent(
'Model.afterRules',
compact('entity', 'options', 'result', 'operation')
);
if ($event->isStopped()) {
return $event->getResult();
}
```
Maybe, to make this less hard to fail, we should allow null as OK value?
return $event->getResult() ?? true;
What do you think?
### CakePHP Version
4.4.7
### PHP Version
8.1 | 1.0 | Strictness of event dispatch result for ORM checkRules() - ### Description
Error
> Cake\ORM\Table::checkRules(): Return value must be of type bool, null returned
As the signature of that method is:
```php
public function checkRules(...): bool
```
The code inside checkRules() reads
```php
$event = $this->dispatchEvent(
'Model.afterRules',
compact('entity', 'options', 'result', 'operation')
);
if ($event->isStopped()) {
return $event->getResult();
}
```
Maybe, to make this less hard to fail, we should allow null as OK value?
return $event->getResult() ?? true;
What do you think?
### CakePHP Version
4.4.7
### PHP Version
8.1 | non_priority | strictness of event dispatch result for orm checkrules description error cake orm table checkrules return value must be of type bool null returned as the signature of that method is php public function checkrules bool the code inside checkrules reads php event this dispatchevent model afterrules compact entity options result operation if event isstopped return event getresult maybe to make this less hard to fail we should allow null as ok value return event getresult true what do you think cakephp version php version | 0 |
62,875 | 3,193,579,159 | IssuesEvent | 2015-09-30 06:42:50 | fusioninventory/fusioninventory-for-glpi | https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi | closed | Package Files : rename uncompress field | Category: Deploy Component: For junior contributor Priority: Normal Status: Closed Tracker: Feature | ---
Author Name: **alexandre delaunay** (@orthagh)
Original Redmine Issue: 1254, http://forge.fusioninventory.org/issues/1254
Original Date: 2011-10-13
---
None
| 1.0 | Package Files : rename uncompress field - ---
Author Name: **alexandre delaunay** (@orthagh)
Original Redmine Issue: 1254, http://forge.fusioninventory.org/issues/1254
Original Date: 2011-10-13
---
None
| priority | package files rename uncompress field author name alexandre delaunay orthagh original redmine issue original date none | 1 |
679,704 | 23,242,466,059 | IssuesEvent | 2022-08-03 16:46:55 | spiffe/spire | https://api.github.com/repos/spiffe/spire | closed | Upgrade to google.golang.org/protobuf 1.28.1 breaks unit test assertions | priority/backlog | Upgrading from [google.golang.org/protobuf](http://google.golang.org/protobuf) 1.28.0 to 1.28.1 and results in some unit test failures. Something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields.
Example: [these assertions](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache_test.go#L540-L541) in `pkg/agent/manager/storecache:TestUpdateEntries` fail, likely because the [storecache.Record](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache.go#L21) and [cache.StaleEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/cache/cache.go#L126) structs include a *[common.RegistrationEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/proto/spire/common/common.proto#L59-L91) protobuf message field that is not cleanly matching using `reflect.DeepEqual()`. Digging into the failure, it looks like some internal proto field called `atomicMessageInfo` is not matching - one of the objects contains a pointer value for that field, the other a nil value.
It is probably worth evaluating whether protobuf messages should be embedded in Go structs in SPIRE code because of the challenges involved with comparing Go structs containing several fields, where some of the fields are plain Go structs and some are protobuf messages. | 1.0 | Upgrade to google.golang.org/protobuf 1.28.1 breaks unit test assertions - Upgrading from [google.golang.org/protobuf](http://google.golang.org/protobuf) 1.28.0 to 1.28.1 and results in some unit test failures. Something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields.
Example: [these assertions](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache_test.go#L540-L541) in `pkg/agent/manager/storecache:TestUpdateEntries` fail, likely because the [storecache.Record](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache.go#L21) and [cache.StaleEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/cache/cache.go#L126) structs include a *[common.RegistrationEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/proto/spire/common/common.proto#L59-L91) protobuf message field that is not cleanly matching using `reflect.DeepEqual()`. Digging into the failure, it looks like some internal proto field called `atomicMessageInfo` is not matching - one of the objects contains a pointer value for that field, the other a nil value.
It is probably worth evaluating whether protobuf messages should be embedded in Go structs in SPIRE code because of the challenges involved with comparing Go structs containing several fields, where some of the fields are plain Go structs and some are protobuf messages. | priority | upgrade to google golang org protobuf breaks unit test assertions upgrading from to and results in some unit test failures something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields example in pkg agent manager storecache testupdateentries fail likely because the and structs include a protobuf message field that is not cleanly matching using reflect deepequal digging into the failure it looks like some internal proto field called atomicmessageinfo is not matching one of the objects contains a pointer value for that field the other a nil value it is probably worth evaluating whether protobuf messages should be embedded in go structs in spire code because of the challenges involved with comparing go structs containing several fields where some of the fields are plain go structs and some are protobuf messages | 1 |
72,411 | 24,108,144,953 | IssuesEvent | 2022-09-20 09:09:18 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Element desktop frequently crashes | T-Defect | ### Steps to reproduce
1. Open the app, wait around 10 seconds (sometimes a bit more, but not longer than a minute i guess).
2. Experience the crash
This happens very reliably. I cannot use the app for more than a few seconds. I've tried deleting `~/.config/Elemen` and reinstalling the app without success.
### Outcome
#### What did you expect?
Element donesn't crash
#### What happened instead?
A crash
### Operating system
Arch Linux
### Application version
Element version: 1.11.4 Olm version: 3.2.12
### How did you install the app?
From the official Arch repo (community)
### Homeserver
Private homeserver, Synapse 1.67.0 (from nixpkgs)
### Will you send logs?
Yes | 1.0 | Element desktop frequently crashes - ### Steps to reproduce
1. Open the app, wait around 10 seconds (sometimes a bit more, but not longer than a minute i guess).
2. Experience the crash
This happens very reliably. I cannot use the app for more than a few seconds. I've tried deleting `~/.config/Elemen` and reinstalling the app without success.
### Outcome
#### What did you expect?
Element donesn't crash
#### What happened instead?
A crash
### Operating system
Arch Linux
### Application version
Element version: 1.11.4 Olm version: 3.2.12
### How did you install the app?
From the official Arch repo (community)
### Homeserver
Private homeserver, Synapse 1.67.0 (from nixpkgs)
### Will you send logs?
Yes | non_priority | element desktop frequently crashes steps to reproduce open the app wait around seconds sometimes a bit more but not longer than a minute i guess experience the crash this happens very reliably i cannot use the app for more than a few seconds i ve tried deleting config elemen and reinstalling the app without success outcome what did you expect element donesn t crash what happened instead a crash operating system arch linux application version element version olm version how did you install the app from the official arch repo community homeserver private homeserver synapse from nixpkgs will you send logs yes | 0 |
51,192 | 6,150,536,753 | IssuesEvent | 2017-06-27 22:54:06 | SavageCore/new-retro-arcade-neon-attract-screen-tool | https://api.github.com/repos/SavageCore/new-retro-arcade-neon-attract-screen-tool | closed | App fails to load for certain people | bug critical needs-testing | 
Disabling `updateChecker()` resolves the issue for [mrleelee123](http://steamcommunity.com/app/465780/discussions/0/343787283758035953/?tscn=1479142088#c217691032444349428) | 1.0 | App fails to load for certain people - 
Disabling `updateChecker()` resolves the issue for [mrleelee123](http://steamcommunity.com/app/465780/discussions/0/343787283758035953/?tscn=1479142088#c217691032444349428) | non_priority | app fails to load for certain people disabling updatechecker resolves the issue for | 0 |
708,994 | 24,364,483,305 | IssuesEvent | 2022-10-03 14:13:02 | IAmTamal/Milan | https://api.github.com/repos/IAmTamal/Milan | opened | [DOCS] Add setup instruction in the readme file | 📄 aspect: text ✨ goal: improvement 🟨 priority: medium 🛠 status : under development | ### Description
There is no setup instructions in the readme file.
i think we can add both setup steps in the readme file, one with npm and one with docker.
### Screenshots
_No response_
### Additional information
_No response_ | 1.0 | [DOCS] Add setup instruction in the readme file - ### Description
There is no setup instructions in the readme file.
i think we can add both setup steps in the readme file, one with npm and one with docker.
### Screenshots
_No response_
### Additional information
_No response_ | priority | add setup instruction in the readme file description there is no setup instructions in the readme file i think we can add both setup steps in the readme file one with npm and one with docker screenshots no response additional information no response | 1 |
85,063 | 15,731,185,231 | IssuesEvent | 2021-03-29 16:46:12 | wrbejar/bag-of-holding | https://api.github.com/repos/wrbejar/bag-of-holding | opened | CVE-2020-11022 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-2.1.3.min.js</b>, <b>jquery-1.10.2.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p>
<p>Path to vulnerable library: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.3.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html</p>
<p>Path to vulnerable library: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html,bag-of-holding/bower_components/select2-bootstrap-css/_jekyll/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","packageFilePaths":["/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.3","packageFilePaths":["/bower_components/Chart.js/samples/line-customTooltips.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html","/bower_components/select2-bootstrap-css/_jekyll/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-11022 (Medium) detected in multiple libraries - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-2.1.3.min.js</b>, <b>jquery-1.10.2.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p>
<p>Path to vulnerable library: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.3.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.10.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p>
<p>Path to dependency file: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html</p>
<p>Path to vulnerable library: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html,bag-of-holding/bower_components/select2-bootstrap-css/_jekyll/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.10.2.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","packageFilePaths":["/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.3","packageFilePaths":["/bower_components/Chart.js/samples/line-customTooltips.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html","/bower_components/select2-bootstrap-css/_jekyll/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding node modules bower lib node modules redeyed examples browser index html path to vulnerable library bag of holding node modules bower lib node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding bower components chart js samples line customtooltips html path to vulnerable library bag of holding bower components chart js samples line customtooltips html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding bower components bootstrap css docs html path to vulnerable library bag of holding bower components bootstrap css docs html bag of holding bower components bootstrap css jekyll index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery basebranches vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl | 0 |
66,222 | 20,072,280,336 | IssuesEvent | 2022-02-04 08:40:33 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Message search initialisation failed (invalid passphrase) | T-Defect X-Needs-Info S-Major A-E2EE A-Search A-Indexing X-Needs-Investigation O-Uncommon | ### Steps to reproduce
1. Where are you starting? What can you see?
If I try to search messages in a non group chat, I get `Message search initialisation failed, check your settings for more information`.
If I click the settings, I can see: `Message search initialisation failed`. If I click the drop down I see `Error opening the database: DatabaseUnlockError("Invalid passphrase")`.
I have no idea what passphrase this is. When I launch element, I am prompted for a password, which I type in and everything launches ok. I dont understand what passphrase this is that it is missing.
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
Search to work
#### What happened instead?
Search didnt work
### Operating system
Ubuntu 20.04 lts
### Application version
Element Desktop Element version: 1.9.9 Olm version: 3.2.8
### How did you install the app?
debian package
### Homeserver
Non-public, not sure
### Will you send logs?
No | 1.0 | Message search initialisation failed (invalid passphrase) - ### Steps to reproduce
1. Where are you starting? What can you see?
If I try to search messages in a non group chat, I get `Message search initialisation failed, check your settings for more information`.
If I click the settings, I can see: `Message search initialisation failed`. If I click the drop down I see `Error opening the database: DatabaseUnlockError("Invalid passphrase")`.
I have no idea what passphrase this is. When I launch element, I am prompted for a password, which I type in and everything launches ok. I dont understand what passphrase this is that it is missing.
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
Search to work
#### What happened instead?
Search didnt work
### Operating system
Ubuntu 20.04 lts
### Application version
Element Desktop Element version: 1.9.9 Olm version: 3.2.8
### How did you install the app?
debian package
### Homeserver
Non-public, not sure
### Will you send logs?
No | non_priority | message search initialisation failed invalid passphrase steps to reproduce where are you starting what can you see if i try to search messages in a non group chat i get message search initialisation failed check your settings for more information if i click the settings i can see message search initialisation failed if i click the drop down i see error opening the database databaseunlockerror invalid passphrase i have no idea what passphrase this is when i launch element i am prompted for a password which i type in and everything launches ok i dont understand what passphrase this is that it is missing what do you click more steps… outcome what did you expect search to work what happened instead search didnt work operating system ubuntu lts application version element desktop element version olm version how did you install the app debian package homeserver non public not sure will you send logs no | 0 |
236,454 | 18,097,964,142 | IssuesEvent | 2021-09-22 11:13:12 | alphagov/govuk-frontend | https://api.github.com/repos/alphagov/govuk-frontend | opened | Add info about automated tools to accessibility criteria document | documentation accessibility | ## What
Based on user feedback, add info about automated tools to [accessibility criteria doc](https://github.com/alphagov/govuk-frontend/blob/main/docs/contributing/test-components-using-accessibility-acceptance-criteria.md).
## Why
After we got 2i approval to publish our accessibility criteria, we realised it lacked content about automated tools that help detect accessibility issues.
### Points to consider
- according to [GOV.UK's blog on testing with a deliberately inaccessible webpage](https://accessibility.blog.gov.uk/2017/02/24/what-we-found-when-we-tested-tools-on-the-worlds-least-accessible-webpage/), 29% of the accessibility barriers went undetected by any of the 10 automated tools used - for more info, see the section titled 'Lots of the barriers weren’t found by any of the tools'
- however, automated tools (for example, Axe, SiteImprove) are still useful for picking up basic issues
- we'd like to hear from the Design System community about the automated tools they find useful
## Who needs to know about this
Technical Writer, Community Manager, Developers
## Done when
- [ ] Developer and Technical Writer draft update
- [ ] Update receives review from teammate
- [ ] Update passes 2i
- [ ] We publish update | 1.0 | Add info about automated tools to accessibility criteria document - ## What
Based on user feedback, add info about automated tools to [accessibility criteria doc](https://github.com/alphagov/govuk-frontend/blob/main/docs/contributing/test-components-using-accessibility-acceptance-criteria.md).
## Why
After we got 2i approval to publish our accessibility criteria, we realised it lacked content about automated tools that help detect accessibility issues.
### Points to consider
- according to [GOV.UK's blog on testing with a deliberately inaccessible webpage](https://accessibility.blog.gov.uk/2017/02/24/what-we-found-when-we-tested-tools-on-the-worlds-least-accessible-webpage/), 29% of the accessibility barriers went undetected by any of the 10 automated tools used - for more info, see the section titled 'Lots of the barriers weren’t found by any of the tools'
- however, automated tools (for example, Axe, SiteImprove) are still useful for picking up basic issues
- we'd like to hear from the Design System community about the automated tools they find useful
## Who needs to know about this
Technical Writer, Community Manager, Developers
## Done when
- [ ] Developer and Technical Writer draft update
- [ ] Update receives review from teammate
- [ ] Update passes 2i
- [ ] We publish update | non_priority | add info about automated tools to accessibility criteria document what based on user feedback add info about automated tools to why after we got approval to publish our accessibility criteria we realised it lacked content about automated tools that help detect accessibility issues points to consider according to of the accessibility barriers went undetected by any of the automated tools used for more info see the section titled lots of the barriers weren’t found by any of the tools however automated tools for example axe siteimprove are still useful for picking up basic issues we d like to hear from the design system community about the automated tools they find useful who needs to know about this technical writer community manager developers done when developer and technical writer draft update update receives review from teammate update passes we publish update | 0 |
713,018 | 24,515,057,791 | IssuesEvent | 2022-10-11 03:43:15 | matrixorigin/matrixone | https://api.github.com/repos/matrixorigin/matrixone | closed | [Subtask]: remove update related | priority/p0 component/distributed tae kind/subtask | ### Parent Issue
#4628
### Detail of Subtask
Remove update-related code and test and refactor any code that depends on the code.
### Describe implementation you've considered
_No response_
### Additional information
_No response_ | 1.0 | [Subtask]: remove update related - ### Parent Issue
#4628
### Detail of Subtask
Remove update-related code and test and refactor any code that depends on the code.
### Describe implementation you've considered
_No response_
### Additional information
_No response_ | priority | remove update related parent issue detail of subtask remove update related code and test and refactor any code that depends on the code describe implementation you ve considered no response additional information no response | 1 |
357,242 | 10,604,023,074 | IssuesEvent | 2019-10-10 17:15:02 | satcom886/rugl | https://api.github.com/repos/satcom886/rugl | closed | Make the labels a little more original... Maybe? | low-priority organisation | pycracker - What could be wrong?
pylogin - What is this even for?
pyparosci - How broken is it? | 1.0 | Make the labels a little more original... Maybe? - pycracker - What could be wrong?
pylogin - What is this even for?
pyparosci - How broken is it? | priority | make the labels a little more original maybe pycracker what could be wrong pylogin what is this even for pyparosci how broken is it | 1 |
759,408 | 26,593,109,321 | IssuesEvent | 2023-01-23 10:19:09 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | [Performance] Improve Reply message author detail fetch | priority 2: medium E:PerformanceImprovements S:3 messenger | ### Description
Right now, when there is a reply message, we get the details for the author of the replied message using the pubkey only and calling the backend to get JSON. That's very inefficient, as we need to do a back and forth and then parse the data.
Plus, I needed to do a iterator property just to tell the frontend to call the JSON again.
One way I realized that will improve that a lot is to just populate it all in the backend (in the module most likely).
We just need to use the new properties Pascal added here: https://github.com/status-im/status-desktop/pull/9111/files#diff-4f86d004be69cc9b51a70273a5ec7ab80e851a4231773dad079e8c01afa753b5R53-R54
When we detect that the author was updated (the model already has `modelContactUpdateIterator` taht is called in those situations), we just update the details directly instead of relying on the iterator. | 1.0 | [Performance] Improve Reply message author detail fetch - ### Description
Right now, when there is a reply message, we get the details for the author of the replied message using the pubkey only and calling the backend to get JSON. That's very inefficient, as we need to do a back and forth and then parse the data.
Plus, I needed to do a iterator property just to tell the frontend to call the JSON again.
One way I realized that will improve that a lot is to just populate it all in the backend (in the module most likely).
We just need to use the new properties Pascal added here: https://github.com/status-im/status-desktop/pull/9111/files#diff-4f86d004be69cc9b51a70273a5ec7ab80e851a4231773dad079e8c01afa753b5R53-R54
When we detect that the author was updated (the model already has `modelContactUpdateIterator` taht is called in those situations), we just update the details directly instead of relying on the iterator. | priority | improve reply message author detail fetch description right now when there is a reply message we get the details for the author of the replied message using the pubkey only and calling the backend to get json that s very inefficient as we need to do a back and forth and then parse the data plus i needed to do a iterator property just to tell the frontend to call the json again one way i realized that will improve that a lot is to just populate it all in the backend in the module most likely we just need to use the new properties pascal added here when we detect that the author was updated the model already has modelcontactupdateiterator taht is called in those situations we just update the details directly instead of relying on the iterator | 1 |
317,538 | 9,666,447,637 | IssuesEvent | 2019-05-21 10:50:39 | oceanprotocol/ocean | https://api.github.com/repos/oceanprotocol/ocean | opened | Integration of Aquarius with Elastic Search for Commons | L2 Metadata bug commons priority:medium | In order to support better search capabilities, it would be necessary to integrate Aquarius with Elastic Search in the `Duero` and `Nile` environments.
The tasks included are:
- [ ] Deployment of Elastic Search in `Duero` https://github.com/oceanprotocol/atlantic/issues/54
- [ ] Integration of Aquarius with Elastic Search in `Duero` https://github.com/oceanprotocol/aquarius/issues/184
- [ ] Deployment of Elastic Search in `Nile`
- [ ] Integration of Aquarius with Elastic Search in `Nile`
| 1.0 | Integration of Aquarius with Elastic Search for Commons - In order to support better search capabilities, it would be necessary to integrate Aquarius with Elastic Search in the `Duero` and `Nile` environments.
The tasks included are:
- [ ] Deployment of Elastic Search in `Duero` https://github.com/oceanprotocol/atlantic/issues/54
- [ ] Integration of Aquarius with Elastic Search in `Duero` https://github.com/oceanprotocol/aquarius/issues/184
- [ ] Deployment of Elastic Search in `Nile`
- [ ] Integration of Aquarius with Elastic Search in `Nile`
| priority | integration of aquarius with elastic search for commons in order to support better search capabilities it would be necessary to integrate aquarius with elastic search in the duero and nile environments the tasks included are deployment of elastic search in duero integration of aquarius with elastic search in duero deployment of elastic search in nile integration of aquarius with elastic search in nile | 1 |
297,349 | 25,723,568,962 | IssuesEvent | 2022-12-07 15:13:57 | vegaprotocol/frontend-monorepo | https://api.github.com/repos/vegaprotocol/frontend-monorepo | closed | Support starting the front-ends with a long-living token | Trading Testing 🧪 Wallet-Automation common | Blocked by:
- https://github.com/vegaprotocol/frontend-monorepo/issues/2249
---
Long-living tokens are tokens that are pre-generated from the command line.
```
vega wallet api-token generate
```
These tokens allows headless software to query the service without human intervention. It primarily exists to support bots and automated scripts.
It can also be of good use for UI testing. However, since this feature is designed with a slightly different request flow than the "interactive" one (manual approval and all), the front-ends needs to be able to account for that difference.
Interactive workflow is:
1. `client.connect_wallet` -> get a connection token
2. `client.list_keys { "token": token }`
3. etc
Headless workflow is:
1. `client.list_keys { "token": longLivingToken }`
2. etc.
With a long-living token, there is no need to call `client.connect_wallet` since we already have the token. So the front-ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow.
| 1.0 | Support starting the front-ends with a long-living token - Blocked by:
- https://github.com/vegaprotocol/frontend-monorepo/issues/2249
---
Long-living tokens are tokens that are pre-generated from the command line.
```
vega wallet api-token generate
```
These tokens allows headless software to query the service without human intervention. It primarily exists to support bots and automated scripts.
It can also be of good use for UI testing. However, since this feature is designed with a slightly different request flow than the "interactive" one (manual approval and all), the front-ends needs to be able to account for that difference.
Interactive workflow is:
1. `client.connect_wallet` -> get a connection token
2. `client.list_keys { "token": token }`
3. etc
Headless workflow is:
1. `client.list_keys { "token": longLivingToken }`
2. etc.
With a long-living token, there is no need to call `client.connect_wallet` since we already have the token. So the front-ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow.
| non_priority | support starting the front ends with a long living token blocked by long living tokens are tokens that are pre generated from the command line vega wallet api token generate these tokens allows headless software to query the service without human intervention it primarily exists to support bots and automated scripts it can also be of good use for ui testing however since this feature is designed with a slightly different request flow than the interactive one manual approval and all the front ends needs to be able to account for that difference interactive workflow is client connect wallet get a connection token client list keys token token etc headless workflow is client list keys token longlivingtoken etc with a long living token there is no need to call client connect wallet since we already have the token so the front ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow | 0 |
72,507 | 3,386,588,651 | IssuesEvent | 2015-11-27 19:00:37 | YetiForceCompany/YetiForceCRM | https://api.github.com/repos/YetiForceCompany/YetiForceCRM | closed | How to reset backup | Label::Module Priority::#3 High Type::Bug | I tried the backup option in 2.2 and it was hanging again. Even updated to 2.3 but still the same screen:
Any idea how to reset this?

| 1.0 | How to reset backup - I tried the backup option in 2.2 and it was hanging again. Even updated to 2.3 but still the same screen:
Any idea how to reset this?

| priority | how to reset backup i tried the backup option in and it was hanging again even updated to but still the same screen any idea how to reset this | 1 |
523,219 | 15,175,711,590 | IssuesEvent | 2021-02-14 00:40:13 | nischalshrestha/DataTutor | https://api.github.com/repos/nischalshrestha/DataTutor | closed | Stepper: make it work for R too. | priority | Upon detection of language engine, we should have the same functionality of parse/execute/store outputs, and stepping to work with R as well.
It could also be easier to do this because R is easier to introspect and no worries about output rendering as much. | 1.0 | Stepper: make it work for R too. - Upon detection of language engine, we should have the same functionality of parse/execute/store outputs, and stepping to work with R as well.
It could also be easier to do this because R is easier to introspect and no worries about output rendering as much. | priority | stepper make it work for r too upon detection of language engine we should have the same functionality of parse execute store outputs and stepping to work with r as well it could also be easier to do this because r is easier to introspect and no worries about output rendering as much | 1 |
309,516 | 23,297,904,423 | IssuesEvent | 2022-08-06 21:55:31 | MattTheLegoman/RealmsInExile | https://api.github.com/repos/MattTheLegoman/RealmsInExile | opened | Suggestion: More forms of Executions | documentation help wanted localization lore suggestion research | Some people in discussion mentioned recently and i've heard it before that a lot of our executions tends to be by burning at the stakes for most faiths. This is something we likely we can change and do more flavorful. I'll keep this as a suggestion if someone is interested.
I guess this is most likely a localization heavy issue, but research into lore and so on is still needed. | 1.0 | Suggestion: More forms of Executions - Some people in discussion mentioned recently and i've heard it before that a lot of our executions tends to be by burning at the stakes for most faiths. This is something we likely we can change and do more flavorful. I'll keep this as a suggestion if someone is interested.
I guess this is most likely a localization heavy issue, but research into lore and so on is still needed. | non_priority | suggestion more forms of executions some people in discussion mentioned recently and i ve heard it before that a lot of our executions tends to be by burning at the stakes for most faiths this is something we likely we can change and do more flavorful i ll keep this as a suggestion if someone is interested i guess this is most likely a localization heavy issue but research into lore and so on is still needed | 0 |
349,026 | 24,931,147,096 | IssuesEvent | 2022-10-31 11:44:09 | ToolJet/ToolJet | https://api.github.com/repos/ToolJet/ToolJet | closed | [docs]: add image to button group docs | documentation good first issue hacktoberfest up-for-grabs | ### Summary
There is no image of the button group widget.
Add the following image in the button group doc below the description:

### Issue Type
Documentation bug
### The entire URL of the documentation with the issue
https://docs.tooljet.com/docs/widgets/button-group
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
No
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct | 1.0 | [docs]: add image to button group docs - ### Summary
There is no image of the button group widget.
Add the following image in the button group doc below the description:

### Issue Type
Documentation bug
### The entire URL of the documentation with the issue
https://docs.tooljet.com/docs/widgets/button-group
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
No
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct | non_priority | add image to button group docs summary there is no image of the button group widget add the following image in the button group doc below the description issue type documentation bug the entire url of the documentation with the issue steps to reproduce the issue no response additional information no response if the issue is confirmed would you be willing to submit a pull request no code of conduct i agree to follow the tooljet code of conduct | 0 |
211,519 | 16,448,252,862 | IssuesEvent | 2021-05-20 23:02:23 | insertcode/insertcode-website | https://api.github.com/repos/insertcode/insertcode-website | closed | Mejorar el README | documentation good first issue | Sería bueno incluir al menos:
- Instrucciones para desarrollo local
- Stack actual
- Instrucciones para contribuir | 1.0 | Mejorar el README - Sería bueno incluir al menos:
- Instrucciones para desarrollo local
- Stack actual
- Instrucciones para contribuir | non_priority | mejorar el readme sería bueno incluir al menos instrucciones para desarrollo local stack actual instrucciones para contribuir | 0 |
51,554 | 3,013,031,880 | IssuesEvent | 2015-07-29 05:35:17 | OctopusDeploy/Issues | https://api.github.com/repos/OctopusDeploy/Issues | closed | Pagination missing from library packages | bug priority | When you click through to view the packages in the library you only see 30, and there is no pagination to see more. Confirmed on demo which has approx 1700 packages but I can only see 30 of each.
Source: http://help.octopusdeploy.com/discussions/problems/36342 | 1.0 | Pagination missing from library packages - When you click through to view the packages in the library you only see 30, and there is no pagination to see more. Confirmed on demo which has approx 1700 packages but I can only see 30 of each.
Source: http://help.octopusdeploy.com/discussions/problems/36342 | priority | pagination missing from library packages when you click through to view the packages in the library you only see and there is no pagination to see more confirmed on demo which has approx packages but i can only see of each source | 1 |
539,294 | 15,786,562,035 | IssuesEvent | 2021-04-01 17:55:50 | code4romania/de-urgenta-android | https://api.github.com/repos/code4romania/de-urgenta-android | opened | Implement proper error handling mechanism | high-priority :fire: | Implement an easy to use error handling mechanism.
Should support displaying snackbars and dialogs. Check the [design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530) for hints on how to design them, but follow the android material design guidelines for them.
These should be displayed easily in case of validation or api errors.
| 1.0 | Implement proper error handling mechanism - Implement an easy to use error handling mechanism.
Should support displaying snackbars and dialogs. Check the [design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530) for hints on how to design them, but follow the android material design guidelines for them.
These should be displayed easily in case of validation or api errors.
| priority | implement proper error handling mechanism implement an easy to use error handling mechanism should support displaying snackbars and dialogs check the for hints on how to design them but follow the android material design guidelines for them these should be displayed easily in case of validation or api errors | 1 |
238,432 | 26,107,631,719 | IssuesEvent | 2022-12-27 15:10:51 | elastic/cloudbeat | https://api.github.com/repos/elastic/cloudbeat | opened | Launcher wrapper layer and the validator abstraction | Team:Cloud Security Posture | **Motivation**
Today our `cloudbeat` is wrapped by a `launcher` component that manages the reconfigurations and runs the beat itself.
This is different from other beats that do not have this wrapper logic and is also not part of the `beat` development generation code.
The goal of the `launcher` was to be agnostic to the underlying managed `beat` and to eventually be transformed into the `beat` development generation code.
Currently, this wrapper is being used only by our beat and there are no future plans for it to change (not that I'm aware of as of right now).
Part of the `launcher` includes the implementation of an abstraction layer of `validator` that has the purpose of validating configurations for the `beat` instance.
Our current implementation of the `validator` just calls the `config` constructor and doesn't do anything besides that.
We should discuss if we still want to have this whole abstraction layer or if it is redundant.
While working on this PR:
- https://github.com/elastic/cloudbeat/pull/597
I've had multiple commits regarding different ways how to tackle this configuration issue.
We've decided to keep the `validator` and leave the checks inside the `config` for now.
**Definition of done**
What needs to be completed at the end of this task
- [] Decide the fate of the abstraction and act accordingly
- [] If we decide to leave it then create a plan/roadmap to utilize it and design everything accordingly | True | Launcher wrapper layer and the validator abstraction - **Motivation**
Today our `cloudbeat` is wrapped by a `launcher` component that manages the reconfigurations and runs the beat itself.
This is different from other beats that do not have this wrapper logic and is also not part of the `beat` development generation code.
The goal of the `launcher` was to be agnostic to the underlying managed `beat` and to eventually be transformed into the `beat` development generation code.
Currently, this wrapper is being used only by our beat and there are no future plans for it to change (not that I'm aware of as of right now).
Part of the `launcher` includes the implementation of an abstraction layer of `validator` that has the purpose of validating configurations for the `beat` instance.
Our current implementation of the `validator` just calls the `config` constructor and doesn't do anything besides that.
We should discuss if we still want to have this whole abstraction layer or if it is redundant.
While working on this PR:
- https://github.com/elastic/cloudbeat/pull/597
I've had multiple commits regarding different ways how to tackle this configuration issue.
We've decided to keep the `validator` and leave the checks inside the `config` for now.
**Definition of done**
What needs to be completed at the end of this task
- [] Decide the fate of the abstraction and act accordingly
- [] If we decide to leave it then create a plan/roadmap to utilize it and design everything accordingly | non_priority | launcher wrapper layer and the validator abstraction motivation today our cloudbeat is wrapped by a launcher component that manages the reconfigurations and runs the beat itself this is different from other beats that do not have this wrapper logic and is also not part of the beat development generation code the goal of the launcher was to be agnostic to the underlying managed beat and to eventually be transformed into the beat development generation code currently this wrapper is being used only by our beat and there are no future plans for it to change not that i m aware of as of right now part of the launcher includes the implementation of an abstraction layer of validator that has the purpose of validating configurations for the beat instance our current implementation of the validator just calls the config constructor and doesn t do anything besides that we should discuss if we still want to have this whole abstraction layer or if it is redundant while working on this pr i ve had multiple commits regarding different ways how to tackle this configuration issue we ve decided to keep the validator and leave the checks inside the config for now definition of done what needs to be completed at the end of this task decide the fate of the abstraction and act accordingly if we decide to leave it then create a plan roadmap to utilize it and design everything accordingly | 0 |
9,891 | 12,890,275,957 | IssuesEvent | 2020-07-13 15:45:05 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | Token for the current date seems missing | Pri2 devops-cicd-process/tech devops/prod doc-enhancement | Please also include a token for the current day. We already have a token for yyyymmdd. I don't know if its not supported or missing in documentation but we needed a token for just the day (dd) component of the date.
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | Token for the current date seems missing - Please also include a token for the current day. We already have a token for yyyymmdd. I don't know if its not supported or missing in documentation but we needed a token for just the day (dd) component of the date.
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | non_priority | token for the current date seems missing please also include a token for the current day we already have a token for yyyymmdd i don t know if its not supported or missing in documentation but we needed a token for just the day dd component of the date document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 0 |
240,038 | 19,988,375,057 | IssuesEvent | 2022-01-31 00:43:03 | backend-br/vagas | https://api.github.com/repos/backend-br/vagas | closed | [Remoto] Tech Lead Java @ Mollica IT | CLT PJ Java Remoto Especialista Exterior Testes automatizados Alocado Redis SQL Git RabbitMQ Stale | ## Nossa empresa
Mais do que uma consultoria, temos uma solução de Recrutamento & Seleção, sendo uma conexão que une profissionais adequados com diversos clientes no Brasil e exterior.
Somos mais que uma Consultoria, somos seus Parceiros! A Mollica conta com um DNA inovador, somos especialistas no recrutamento de profissionais de tecnologias, especializados em soluções de TI.
OBS: Estamos trabalhando essa posição para a Plus IT com alocação no Banco Digio.
## Descrição da vaga
Quer trabalhar em um banco digital, uma bantech que conta com a solidez de um banco e agilidade de uma fintech?
Então vem com a gente e se liga só nessa oportunidade:
Somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um Tech Lead com foco em Java.
Qual será a sua missão:
Definir e entender requisitos técnicos e funcionais com a equipe;
Planejar e desenhar detalhes de componentes e serviços;
Codificar, testar e publicar serviços;
Participar de code e design review;
## Local
Remoto - aceitamos de candidatos todo o Brasil
## Requisitos
**Obrigatórios:**
- Ensino superior completo (ou em andamento) em tecnologia da informação;
- Mínimo de 7 anos de experiência com Desenvolvimento Web;
- Conhecimentos avançados na Stack JAVA e seus principais frameworks;
- Conhecimentos avançados em SQL Server (estrutura de dados, views e procedures);
- Conhecimentos avançados no controle de versão Git;
- Conhecimento de boas práticas de desenvolvimento de software, incluindo padrões de desenvolvimento, documentação, code reviews, controle de versão e testes ;
**Diferenciais:**
- Inglês intermediário para comunicação técnica com equipes de outros países;
- Práticas de testes automatizados;
- Conhecimento em arquitetura micro-serviços e soluções de cache (Redis) e mensageria (RabbitMQ);
- Habilidade em lidar com problemas ambíguos/indefinidos e pensar de forma abstrata;
## Benefícios
Benefícios para CLT:
✔️ Vale Refeição
✔️ Assistência Médica
✔️ Assistência Odontológica
✔️Seguro de vida
✔️Auxilio Creche
Benefícios para PJ:
✔️ Férias remuneradas
## Contratação
PJ ou CLT (a escolha do candidato)
## Como se candidatar
Por favor envie um e-mail para thais@mollicait.com com seu CV anexado ou candidatar-se através do link: https://jobs.quickin.io/mollicait/jobs/619e85f74af9e9001cde5a57
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 05 dias após cada processo.
E-mail para contato em caso de não haver resposta: thais@mollicait.com
## Labels
<!-- retire os labels que não fazem sentido à vaga -->
#### Alocação
- Alocado
- Remoto
#### Regime
- CLT
- PJ
#### Nível
- Sênior
- Especialista
| 1.0 | [Remoto] Tech Lead Java @ Mollica IT - ## Nossa empresa
Mais do que uma consultoria, temos uma solução de Recrutamento & Seleção, sendo uma conexão que une profissionais adequados com diversos clientes no Brasil e exterior.
Somos mais que uma Consultoria, somos seus Parceiros! A Mollica conta com um DNA inovador, somos especialistas no recrutamento de profissionais de tecnologias, especializados em soluções de TI.
OBS: Estamos trabalhando essa posição para a Plus IT com alocação no Banco Digio.
## Descrição da vaga
Quer trabalhar em um banco digital, uma bantech que conta com a solidez de um banco e agilidade de uma fintech?
Então vem com a gente e se liga só nessa oportunidade:
Somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um Tech Lead com foco em Java.
Qual será a sua missão:
Definir e entender requisitos técnicos e funcionais com a equipe;
Planejar e desenhar detalhes de componentes e serviços;
Codificar, testar e publicar serviços;
Participar de code e design review;
## Local
Remoto - aceitamos de candidatos todo o Brasil
## Requisitos
**Obrigatórios:**
- Ensino superior completo (ou em andamento) em tecnologia da informação;
- Mínimo de 7 anos de experiência com Desenvolvimento Web;
- Conhecimentos avançados na Stack JAVA e seus principais frameworks;
- Conhecimentos avançados em SQL Server (estrutura de dados, views e procedures);
- Conhecimentos avançados no controle de versão Git;
- Conhecimento de boas práticas de desenvolvimento de software, incluindo padrões de desenvolvimento, documentação, code reviews, controle de versão e testes ;
**Diferenciais:**
- Inglês intermediário para comunicação técnica com equipes de outros países;
- Práticas de testes automatizados;
- Conhecimento em arquitetura micro-serviços e soluções de cache (Redis) e mensageria (RabbitMQ);
- Habilidade em lidar com problemas ambíguos/indefinidos e pensar de forma abstrata;
## Benefícios
Benefícios para CLT:
✔️ Vale Refeição
✔️ Assistência Médica
✔️ Assistência Odontológica
✔️Seguro de vida
✔️Auxilio Creche
Benefícios para PJ:
✔️ Férias remuneradas
## Contratação
PJ ou CLT (a escolha do candidato)
## Como se candidatar
Por favor envie um e-mail para thais@mollicait.com com seu CV anexado ou candidatar-se através do link: https://jobs.quickin.io/mollicait/jobs/619e85f74af9e9001cde5a57
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 05 dias após cada processo.
E-mail para contato em caso de não haver resposta: thais@mollicait.com
## Labels
<!-- retire os labels que não fazem sentido à vaga -->
#### Alocação
- Alocado
- Remoto
#### Regime
- CLT
- PJ
#### Nível
- Sênior
- Especialista
| non_priority | tech lead java mollica it nossa empresa mais do que uma consultoria temos uma solução de recrutamento seleção sendo uma conexão que une profissionais adequados com diversos clientes no brasil e exterior somos mais que uma consultoria somos seus parceiros a mollica conta com um dna inovador somos especialistas no recrutamento de profissionais de tecnologias especializados em soluções de ti obs estamos trabalhando essa posição para a plus it com alocação no banco digio descrição da vaga quer trabalhar em um banco digital uma bantech que conta com a solidez de um banco e agilidade de uma fintech então vem com a gente e se liga só nessa oportunidade somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um tech lead com foco em java qual será a sua missão definir e entender requisitos técnicos e funcionais com a equipe planejar e desenhar detalhes de componentes e serviços codificar testar e publicar serviços participar de code e design review local remoto aceitamos de candidatos todo o brasil requisitos obrigatórios ensino superior completo ou em andamento em tecnologia da informação mínimo de anos de experiência com desenvolvimento web conhecimentos avançados na stack java e seus principais frameworks conhecimentos avançados em sql server estrutura de dados views e procedures conhecimentos avançados no controle de versão git conhecimento de boas práticas de desenvolvimento de software incluindo padrões de desenvolvimento documentação code reviews controle de versão e testes diferenciais inglês intermediário para comunicação técnica com equipes de outros países práticas de testes automatizados conhecimento em arquitetura micro serviços e soluções de cache redis e mensageria rabbitmq habilidade em lidar com problemas ambíguos indefinidos e pensar de forma abstrata benefícios benefícios para clt ✔️ vale refeição ✔️ assistência médica ✔️ assistência odontológica ✔️seguro de vida ✔️auxilio creche benefícios para pj ✔️ férias remuneradas contratação pj ou clt a escolha do candidato como se candidatar por favor envie um e mail para thais mollicait com com seu cv anexado ou candidatar se através do link tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta thais mollicait com labels alocação alocado remoto regime clt pj nível sênior especialista | 0 |
101,613 | 11,253,391,442 | IssuesEvent | 2020-01-11 15:58:40 | vuepressjs/vuepress-theme-blog | https://api.github.com/repos/vuepressjs/vuepress-theme-blog | closed | Improving Documentation | documentation | A lot of the options and features are not properly explained in the documentation. It would serve the project better if we had more explicit and detailed documentation. Happy to take thison this issue. | 1.0 | Improving Documentation - A lot of the options and features are not properly explained in the documentation. It would serve the project better if we had more explicit and detailed documentation. Happy to take thison this issue. | non_priority | improving documentation a lot of the options and features are not properly explained in the documentation it would serve the project better if we had more explicit and detailed documentation happy to take thison this issue | 0 |
226,724 | 7,522,532,874 | IssuesEvent | 2018-04-12 20:45:12 | runconduit/conduit | https://api.github.com/repos/runconduit/conduit | closed | Public-api should query Prometheus directly | area/controller area/telemetry priority/P0 | Since the telemetry service is going away, the public-api service should be updated to query Prometheus directly. The metrics available are described in #425 and #428.
The result should be that all data in the Conduit web dashboard and in the Conduit CLI stat command are provided by these new metrics. | 1.0 | Public-api should query Prometheus directly - Since the telemetry service is going away, the public-api service should be updated to query Prometheus directly. The metrics available are described in #425 and #428.
The result should be that all data in the Conduit web dashboard and in the Conduit CLI stat command are provided by these new metrics. | priority | public api should query prometheus directly since the telemetry service is going away the public api service should be updated to query prometheus directly the metrics available are described in and the result should be that all data in the conduit web dashboard and in the conduit cli stat command are provided by these new metrics | 1 |
29,588 | 2,716,608,483 | IssuesEvent | 2015-04-10 20:12:18 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | Error on JSON serialization (used on REST services or client DB transactions) when Class has an internal parameterized type | bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.3 | _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 20, 2014 16:28:25_
Error when tryong to create a JSONSerializer for types like:
public class MyClass{
private MyOtherClass<String> a;
}
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=475_ | 1.0 | Error on JSON serialization (used on REST services or client DB transactions) when Class has an internal parameterized type - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 20, 2014 16:28:25_
Error when tryong to create a JSONSerializer for types like:
public class MyClass{
private MyOtherClass<String> a;
}
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=475_ | priority | error on json serialization used on rest services or client db transactions when class has an internal parameterized type from on august error when tryong to create a jsonserializer for types like public class myclass private myotherclass a original issue | 1 |
130,695 | 5,120,219,969 | IssuesEvent | 2017-01-09 01:32:05 | geekgirlbeta/linkapp | https://api.github.com/repos/geekgirlbeta/linkapp | opened | Deleting Links | Priority: High Status: Needs Research Type: Feature | Currently, there is no way for the user to delete a Link. We really won't want them to do this - we want our URLs to persist indefinitely.
We need to allow the user to delete a post, but in a way that preserves the URLs. | 1.0 | Deleting Links - Currently, there is no way for the user to delete a Link. We really won't want them to do this - we want our URLs to persist indefinitely.
We need to allow the user to delete a post, but in a way that preserves the URLs. | priority | deleting links currently there is no way for the user to delete a link we really won t want them to do this we want our urls to persist indefinitely we need to allow the user to delete a post but in a way that preserves the urls | 1 |
248,216 | 7,928,274,384 | IssuesEvent | 2018-07-06 11:00:59 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Debug hit highlighting should go away when the flow is finish | Component/Composer Imported Priority/High Type/Bug | <a href="https://github.com/yasassri"><img src="https://avatars1.githubusercontent.com/u/7681361?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [yasassri](https://github.com/yasassri)**
_Sunday Nov 05, 2017 at 09:54 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/4464_
----
This happens in split view.

| 1.0 | Debug hit highlighting should go away when the flow is finish - <a href="https://github.com/yasassri"><img src="https://avatars1.githubusercontent.com/u/7681361?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [yasassri](https://github.com/yasassri)**
_Sunday Nov 05, 2017 at 09:54 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/4464_
----
This happens in split view.

| priority | debug hit highlighting should go away when the flow is finish issue by sunday nov at gmt originally opened as this happens in split view | 1 |
667,787 | 22,500,379,733 | IssuesEvent | 2022-06-23 11:16:19 | ita-social-projects/horondi_client_fe | https://api.github.com/repos/ita-social-projects/horondi_client_fe | closed | [Product Details Page] Incorrect vertical alignment of currency icon on Product Details Page | bug UI FrontEnd part priority: medium severity: trivial | **Environment** Windows 10, Chrome Version 102.0.5005.63
**Reproducible:** Always
**Preconditions:**
Go to https://horondi-front-staging.azurewebsites.net/
**Steps to reproduce:**
Open Catalog
Click on any product to open Product Details Page
Pay attention to currency icon
**Actual result:** Currency icon is higher than price

**Expected result:** Currency icon and price should be at the same level
| 1.0 | [Product Details Page] Incorrect vertical alignment of currency icon on Product Details Page - **Environment** Windows 10, Chrome Version 102.0.5005.63
**Reproducible:** Always
**Preconditions:**
Go to https://horondi-front-staging.azurewebsites.net/
**Steps to reproduce:**
Open Catalog
Click on any product to open Product Details Page
Pay attention to currency icon
**Actual result:** Currency icon is higher than price

**Expected result:** Currency icon and price should be at the same level
| priority | incorrect vertical alignment of currency icon on product details page environment windows chrome version reproducible always preconditions go to steps to reproduce open catalog click on any product to open product details page pay attention to currency icon actual result currency icon is higher than price expected result currency icon and price should be at the same level | 1 |
759,985 | 26,621,834,415 | IssuesEvent | 2023-01-24 11:50:41 | ethereum-push-notification-service/push-dapp | https://api.github.com/repos/ethereum-push-notification-service/push-dapp | closed | Delegate should be in CAIP10 format to ensure people understand | enhancement Priority 1 | Rationale: current flow doesn't point out the delegations blockchain and since it's triggered and guided by the individual smart contract on that chain, it needs to be made clear for the user.
<img width="1810" alt="Screenshot 2022-09-04 at 8 47 11 PM" src="https://user-images.githubusercontent.com/22663032/188324401-029b85b8-8127-4564-95f8-7e19498bc5c0.png">
| 1.0 | Delegate should be in CAIP10 format to ensure people understand - Rationale: current flow doesn't point out the delegations blockchain and since it's triggered and guided by the individual smart contract on that chain, it needs to be made clear for the user.
<img width="1810" alt="Screenshot 2022-09-04 at 8 47 11 PM" src="https://user-images.githubusercontent.com/22663032/188324401-029b85b8-8127-4564-95f8-7e19498bc5c0.png">
| priority | delegate should be in format to ensure people understand rationale current flow doesn t point out the delegations blockchain and since it s triggered and guided by the individual smart contract on that chain it needs to be made clear for the user img width alt screenshot at pm src | 1 |
45,620 | 5,722,505,801 | IssuesEvent | 2017-04-20 09:39:29 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | ClientReconnectTest.testReconnectToNewInstanceAtSameAddress | Team: Client Type: Test-Failure | https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/8400/
```
java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds, count left: 1
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:899)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:892)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:884)
at com.hazelcast.client.ClientReconnectTest.testReconnectToNewInstanceAtSameAddress(ClientReconnectTest.java:95)
``` | 1.0 | ClientReconnectTest.testReconnectToNewInstanceAtSameAddress - https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/8400/
```
java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds, count left: 1
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:899)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:892)
at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:884)
at com.hazelcast.client.ClientReconnectTest.testReconnectToNewInstanceAtSameAddress(ClientReconnectTest.java:95)
``` | non_priority | clientreconnecttest testreconnecttonewinstanceatsameaddress java lang assertionerror countdownlatch failed to complete within seconds count left at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast client clientreconnecttest testreconnecttonewinstanceatsameaddress clientreconnecttest java | 0 |
460,485 | 13,210,768,437 | IssuesEvent | 2020-08-15 18:44:49 | radiorabe/klangbecken | https://api.github.com/repos/radiorabe/klangbecken | closed | Counter for skipped tracks is not reset when a track was ok | bug priority/1 | This track should not have been played, because it there was an accepted track in the meantime.
```
[root@vm-0016 ~]# journalctl --unit liquidsoap@klangbecken.service --since "2020-08-14 11:30" | grep -E "track was recently played: |too many skipped tracks"
Aug 14 11:47:01 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 11:47:01 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/db378175-5d46-488a-b1ea-20bc7f56ae32.mp3 (128685 seconds ago)
Aug 14 12:02:26 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:02:26 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/ea7f0b81-8316-4f59-9fa8-d8e5945c6aed.mp3 (27725 seconds ago)
Aug 14 12:09:04 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:09:04 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/9c0e36cc-7798-42c0-883c-6129158506f9.mp3 (25581 seconds ago)
Aug 14 12:12:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:12:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/a6683e6d-6def-45cb-b3b3-8b00be4eea6d.mp3 (238572 seconds ago)
Aug 14 12:26:38 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:38 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/28200f9a-4704-44ba-ba9e-11a76b045b24.mp3 (69360 seconds ago)
Aug 14 12:26:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/4ba09c6a-e922-4b7f-a999-163306f3cd75.mp3 (63754 seconds ago)
Aug 14 12:47:25 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:47:25 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/0c8782db-e9c2-4447-92f6-207510c44194.mp3 (163224 seconds ago)
Aug 14 12:50:16 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:50:16 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/c38972ed-af1f-4928-a266-838b53ef519c.mp3 (172108 seconds ago)
Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 (172703 seconds ago)
Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] too many skipped tracks, playing /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 anyway
``` | 1.0 | Counter for skipped tracks is not reset when a track was ok - This track should not have been played, because it there was an accepted track in the meantime.
```
[root@vm-0016 ~]# journalctl --unit liquidsoap@klangbecken.service --since "2020-08-14 11:30" | grep -E "track was recently played: |too many skipped tracks"
Aug 14 11:47:01 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 11:47:01 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/db378175-5d46-488a-b1ea-20bc7f56ae32.mp3 (128685 seconds ago)
Aug 14 12:02:26 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:02:26 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/ea7f0b81-8316-4f59-9fa8-d8e5945c6aed.mp3 (27725 seconds ago)
Aug 14 12:09:04 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:09:04 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/9c0e36cc-7798-42c0-883c-6129158506f9.mp3 (25581 seconds ago)
Aug 14 12:12:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:12:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/a6683e6d-6def-45cb-b3b3-8b00be4eea6d.mp3 (238572 seconds ago)
Aug 14 12:26:38 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:38 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/28200f9a-4704-44ba-ba9e-11a76b045b24.mp3 (69360 seconds ago)
Aug 14 12:26:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/4ba09c6a-e922-4b7f-a999-163306f3cd75.mp3 (63754 seconds ago)
Aug 14 12:47:25 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:47:25 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/0c8782db-e9c2-4447-92f6-207510c44194.mp3 (163224 seconds ago)
Aug 14 12:50:16 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:50:16 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/c38972ed-af1f-4928-a266-838b53ef519c.mp3 (172108 seconds ago)
Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 (172703 seconds ago)
Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] too many skipped tracks, playing /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 anyway
``` | priority | counter for skipped tracks is not reset when a track was ok this track should not have been played because it there was an accepted track in the meantime journalctl unit liquidsoap klangbecken service since grep e track was recently played too many skipped tracks aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap too many skipped tracks playing var lib klangbecken music anyway | 1 |
68,553 | 9,200,329,705 | IssuesEvent | 2019-03-07 16:49:01 | terraform-providers/terraform-provider-google | https://api.github.com/repos/terraform-providers/terraform-provider-google | closed | google_container_cluster: network_policy_config defaults to disabled | documentation | The docs at [1] imply that network_policy_config defaults to enabled and can be disabled. In fact, it's the opposite.
[1] https://www.terraform.io/docs/providers/google/r/container_cluster.html#network_policy_config
### Terraform Version
```
$ terraform -v
Terraform v0.11.7
+ provider.google v1.15.0
``` | 1.0 | google_container_cluster: network_policy_config defaults to disabled - The docs at [1] imply that network_policy_config defaults to enabled and can be disabled. In fact, it's the opposite.
[1] https://www.terraform.io/docs/providers/google/r/container_cluster.html#network_policy_config
### Terraform Version
```
$ terraform -v
Terraform v0.11.7
+ provider.google v1.15.0
``` | non_priority | google container cluster network policy config defaults to disabled the docs at imply that network policy config defaults to enabled and can be disabled in fact it s the opposite terraform version terraform v terraform provider google | 0 |
188,515 | 6,777,155,107 | IssuesEvent | 2017-10-27 20:50:53 | buttercup/buttercup-mobile | https://api.github.com/repos/buttercup/buttercup-mobile | closed | Add padding to home page right-icon | Effort: Low Priority: High Status: Completed Type: Enhancement | For better touch area, add padding to the left side of the right-icon. | 1.0 | Add padding to home page right-icon - For better touch area, add padding to the left side of the right-icon. | priority | add padding to home page right icon for better touch area add padding to the left side of the right icon | 1 |
92,767 | 8,377,788,214 | IssuesEvent | 2018-10-06 05:58:13 | snext1220/stext | https://api.github.com/repos/snext1220/stext | closed | 「七惑星の欠片」をシナリオ側で自動加算させる機能 | Testing enhancement | 「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+4する事」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。
(『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした)
戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。
(減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています) | 1.0 | 「七惑星の欠片」をシナリオ側で自動加算させる機能 - 「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+4する事」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。
(『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした)
戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。
(減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています) | non_priority | 「七惑星の欠片」をシナリオ側で自動加算させる機能 「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+ 」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。 (『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした) 戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。 (減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています) | 0 |
125,483 | 10,344,625,754 | IssuesEvent | 2019-09-04 11:42:07 | ansgohar/urbanharmony | https://api.github.com/repos/ansgohar/urbanharmony | closed | FrontEnd -In الفائزون في المسابقات , "المرصد الحضاري" " and "صحافة"- The picture is displayed as black although it was uploaded | Needs Re-testing State: Pending Testing bug | ندوات و مسابقات1.
Then
الفائزون في المسابقات
The picture is displayed as black although it was uploaded as below

2.
المرصد الحضاري

3. اخر الاخبار
Then
صحافة

| 2.0 | FrontEnd -In الفائزون في المسابقات , "المرصد الحضاري" " and "صحافة"- The picture is displayed as black although it was uploaded - ندوات و مسابقات1.
Then
الفائزون في المسابقات
The picture is displayed as black although it was uploaded as below

2.
المرصد الحضاري

3. اخر الاخبار
Then
صحافة

| non_priority | frontend in الفائزون في المسابقات المرصد الحضاري and صحافة the picture is displayed as black although it was uploaded ندوات و then الفائزون في المسابقات the picture is displayed as black although it was uploaded as below المرصد الحضاري اخر الاخبار then صحافة | 0 |
283,117 | 24,525,354,088 | IssuesEvent | 2022-10-11 12:45:37 | ns1labs/orb | https://api.github.com/repos/ns1labs/orb | opened | Tests: create automated tests to validate UI filters | test | Agents:
- name [ ]
- tags [ ]
- status [ ]
-policies [ ]
Groups:
- Name [ ]
- Tags [ ]
Policies:
- Name [ ]
- Description [ ]
- Tags [ ]
- Version [ ]
Sinks:
- Name
- Tags
- Status
- Backend
| 1.0 | Tests: create automated tests to validate UI filters - Agents:
- name [ ]
- tags [ ]
- status [ ]
-policies [ ]
Groups:
- Name [ ]
- Tags [ ]
Policies:
- Name [ ]
- Description [ ]
- Tags [ ]
- Version [ ]
Sinks:
- Name
- Tags
- Status
- Backend
| non_priority | tests create automated tests to validate ui filters agents name tags status policies groups name tags policies name description tags version sinks name tags status backend | 0 |
176,206 | 6,557,270,117 | IssuesEvent | 2017-09-06 16:47:44 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | OpenShift hitting external kubernetes for internal OpenShift resources | area/security component/auth priority/P2 | Running against external kube version:
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"3+", GitVersion:"v1.3.0-alpha.5-dirty", GitCommit:"914163247d9a16b46921e83f7dbedb572859b3e4", GitTreeState:"dirty", BuildDate:"2016-06-14T18:43:00Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.5", GitCommit:"b0deb2eb8f4037421077f77cb163dbb4c0a2a9f5", GitTreeState:"clean", BuildDate:"2016-08-11T20:21:58Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
```
SecurityContextConstraint does not exist in Kubernetes API, I think. So it's trying to hit external kube for something that is openshift specific.
| 1.0 | OpenShift hitting external kubernetes for internal OpenShift resources - Running against external kube version:
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"3+", GitVersion:"v1.3.0-alpha.5-dirty", GitCommit:"914163247d9a16b46921e83f7dbedb572859b3e4", GitTreeState:"dirty", BuildDate:"2016-06-14T18:43:00Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.5", GitCommit:"b0deb2eb8f4037421077f77cb163dbb4c0a2a9f5", GitTreeState:"clean", BuildDate:"2016-08-11T20:21:58Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
```
SecurityContextConstraint does not exist in Kubernetes API, I think. So it's trying to hit external kube for something that is openshift specific.
| priority | openshift hitting external kubernetes for internal openshift resources running against external kube version kubectl version client version version info major minor gitversion alpha dirty gitcommit gittreestate dirty builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux securitycontextconstraint does not exist in kubernetes api i think so it s trying to hit external kube for something that is openshift specific | 1 |
373,908 | 11,052,371,223 | IssuesEvent | 2019-12-10 09:16:31 | incognitochain/incognito-chain | https://api.github.com/repos/incognitochain/incognito-chain | opened | [Testnet] Panic by 'fatal error: concurrent map read and map write' on synker.go | Priority: High Type: Bug | fatal error: concurrent map read and map write
goroutine 227 [running]:
runtime.throw(0x13f372d, 0x21)
/usr/local/go/src/runtime/panic.go:617 +0x72 fp=0xc001930b48 sp=0xc001930b18 pc=0x4301a2
runtime.mapaccess2(0x1202220, 0xc01782e270, 0xc002382dd0, 0xc002382dd0, 0x1b0)
/usr/local/go/src/runtime/map.go:472 +0x284 fp=0xc001930b90 sp=0xc001930b48 pc=0x410684
reflect.mapaccess(0x1202220, 0xc01782e270, 0xc002382dd0, 0x13e38b4)
/usr/local/go/src/runtime/map.go:1321 +0x3f fp=0xc001930bc8 sp=0xc001930b90 pc=0x4128df
reflect.Value.MapIndex(0x1202220, 0xc000448eb0, 0x195, 0x119c5a0, 0xc002382dd0, 0x98, 0x119c820, 0xc0077e8d50, 0x8b)
/usr/local/go/src/reflect/value.go:1112 +0x10a fp=0xc001930c28 sp=0xc001930bc8 pc=0x49d28a
encoding/json.mapEncoder.encode(0x1568cc0, 0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100)
/usr/local/go/src/encoding/json/encode.go:702 +0x303 fp=0xc001930d70 sp=0xc001930c28 pc=0x5cfe73
encoding/json.mapEncoder.encode-fm(0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100)
/usr/local/go/src/encoding/json/encode.go:678 +0x64 fp=0xc001930db0 sp=0xc001930d70 pc=0x5db8e4
encoding/json.structEncoder.encode(0xc000584600, 0x17, 0x21, 0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0x5c0100)
/usr/local/go/src/encoding/json/encode.go:660 +0x304 fp=0xc001930e50 sp=0xc001930db0 pc=0x5cf944
encoding/json.structEncoder.encode-fm(0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0xc001df0100)
/usr/local/go/src/encoding/json/encode.go:631 +0x76 fp=0xc001930ea0 sp=0xc001930e50 pc=0x5db866
encoding/json.ptrEncoder.encode(0xc00067afc0, 0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x1170100)
/usr/local/go/src/encoding/json/encode.go:806 +0xb2 fp=0xc001930ee8 sp=0xc001930ea0 pc=0x5d0dc2
encoding/json.ptrEncoder.encode-fm(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0xc001930100)
/usr/local/go/src/encoding/json/encode.go:801 +0x64 fp=0xc001930f28 sp=0xc001930ee8 pc=0x5dba64
encoding/json.(*encodeState).reflectValue(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x2400100)
/usr/local/go/src/encoding/json/encode.go:334 +0x82 fp=0xc001930f60 sp=0xc001930f28 pc=0x5cd112
encoding/json.(*encodeState).marshal(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x400100, 0x0, 0x0)
/usr/local/go/src/encoding/json/encode.go:306 +0xee fp=0xc001930fa8 sp=0xc001930f60 pc=0x5ccbbe
encoding/json.Marshal(0x1170020, 0xc001dfa430, 0x1, 0xf7218f3940a6636e, 0xc001931088, 0x4102df, 0xc001931106)
/usr/local/go/src/encoding/json/encode.go:160 +0x52 fp=0xc001931020 sp=0xc001930fa8 pc=0x5cc1e2
github.com/incognitochain/incognito-chain/blockchain.(*ShardBestState).MarshalJSON(0xc000448dc0, 0xc0002d6870, 0xc001931106, 0xc000427a20, 0xc0077e8d40, 0x0)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/shardbeststate.go:244 +0x60 fp=0xc001931098 sp=0xc001931020 pc=0xc07100
github.com/incognitochain/incognito-chain/blockchain.(*Synker).UpdateStatev2(0xc000136628)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:719 +0x8a1 fp=0xc001931cf0 sp=0xc001931098 pc=0xc3e971
github.com/incognitochain/incognito-chain/blockchain.(*Synker).Start(0xc000136628)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:163 +0xa29 fp=0xc001931fd8 sp=0xc001931cf0 pc=0xc3d2c9
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:1337 +0x1 fp=0xc001931fe0 sp=0xc001931fd8 pc=0x45f3a1
created by main.Server.Start
/root/go/src/github.com/incognitochain/incognito-chain/server.go:733 +0x3a9
goroutine 1 [chan receive, 15 minutes]:
main.mainMaster(0x0, 0x0, 0x0)
/root/go/src/github.com/incognitochain/incognito-chain/incognito.go:132 +0x8eb
main.main()
/root/go/src/github.com/incognitochain/incognito-chain/incognito.go:178 +0x246 | 1.0 | [Testnet] Panic by 'fatal error: concurrent map read and map write' on synker.go - fatal error: concurrent map read and map write
goroutine 227 [running]:
runtime.throw(0x13f372d, 0x21)
/usr/local/go/src/runtime/panic.go:617 +0x72 fp=0xc001930b48 sp=0xc001930b18 pc=0x4301a2
runtime.mapaccess2(0x1202220, 0xc01782e270, 0xc002382dd0, 0xc002382dd0, 0x1b0)
/usr/local/go/src/runtime/map.go:472 +0x284 fp=0xc001930b90 sp=0xc001930b48 pc=0x410684
reflect.mapaccess(0x1202220, 0xc01782e270, 0xc002382dd0, 0x13e38b4)
/usr/local/go/src/runtime/map.go:1321 +0x3f fp=0xc001930bc8 sp=0xc001930b90 pc=0x4128df
reflect.Value.MapIndex(0x1202220, 0xc000448eb0, 0x195, 0x119c5a0, 0xc002382dd0, 0x98, 0x119c820, 0xc0077e8d50, 0x8b)
/usr/local/go/src/reflect/value.go:1112 +0x10a fp=0xc001930c28 sp=0xc001930bc8 pc=0x49d28a
encoding/json.mapEncoder.encode(0x1568cc0, 0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100)
/usr/local/go/src/encoding/json/encode.go:702 +0x303 fp=0xc001930d70 sp=0xc001930c28 pc=0x5cfe73
encoding/json.mapEncoder.encode-fm(0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100)
/usr/local/go/src/encoding/json/encode.go:678 +0x64 fp=0xc001930db0 sp=0xc001930d70 pc=0x5db8e4
encoding/json.structEncoder.encode(0xc000584600, 0x17, 0x21, 0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0x5c0100)
/usr/local/go/src/encoding/json/encode.go:660 +0x304 fp=0xc001930e50 sp=0xc001930db0 pc=0x5cf944
encoding/json.structEncoder.encode-fm(0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0xc001df0100)
/usr/local/go/src/encoding/json/encode.go:631 +0x76 fp=0xc001930ea0 sp=0xc001930e50 pc=0x5db866
encoding/json.ptrEncoder.encode(0xc00067afc0, 0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x1170100)
/usr/local/go/src/encoding/json/encode.go:806 +0xb2 fp=0xc001930ee8 sp=0xc001930ea0 pc=0x5d0dc2
encoding/json.ptrEncoder.encode-fm(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0xc001930100)
/usr/local/go/src/encoding/json/encode.go:801 +0x64 fp=0xc001930f28 sp=0xc001930ee8 pc=0x5dba64
encoding/json.(*encodeState).reflectValue(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x2400100)
/usr/local/go/src/encoding/json/encode.go:334 +0x82 fp=0xc001930f60 sp=0xc001930f28 pc=0x5cd112
encoding/json.(*encodeState).marshal(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x400100, 0x0, 0x0)
/usr/local/go/src/encoding/json/encode.go:306 +0xee fp=0xc001930fa8 sp=0xc001930f60 pc=0x5ccbbe
encoding/json.Marshal(0x1170020, 0xc001dfa430, 0x1, 0xf7218f3940a6636e, 0xc001931088, 0x4102df, 0xc001931106)
/usr/local/go/src/encoding/json/encode.go:160 +0x52 fp=0xc001931020 sp=0xc001930fa8 pc=0x5cc1e2
github.com/incognitochain/incognito-chain/blockchain.(*ShardBestState).MarshalJSON(0xc000448dc0, 0xc0002d6870, 0xc001931106, 0xc000427a20, 0xc0077e8d40, 0x0)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/shardbeststate.go:244 +0x60 fp=0xc001931098 sp=0xc001931020 pc=0xc07100
github.com/incognitochain/incognito-chain/blockchain.(*Synker).UpdateStatev2(0xc000136628)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:719 +0x8a1 fp=0xc001931cf0 sp=0xc001931098 pc=0xc3e971
github.com/incognitochain/incognito-chain/blockchain.(*Synker).Start(0xc000136628)
/root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:163 +0xa29 fp=0xc001931fd8 sp=0xc001931cf0 pc=0xc3d2c9
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:1337 +0x1 fp=0xc001931fe0 sp=0xc001931fd8 pc=0x45f3a1
created by main.Server.Start
/root/go/src/github.com/incognitochain/incognito-chain/server.go:733 +0x3a9
goroutine 1 [chan receive, 15 minutes]:
main.mainMaster(0x0, 0x0, 0x0)
/root/go/src/github.com/incognitochain/incognito-chain/incognito.go:132 +0x8eb
main.main()
/root/go/src/github.com/incognitochain/incognito-chain/incognito.go:178 +0x246 | priority | panic by fatal error concurrent map read and map write on synker go fatal error concurrent map read and map write goroutine runtime throw usr local go src runtime panic go fp sp pc runtime usr local go src runtime map go fp sp pc reflect mapaccess usr local go src runtime map go fp sp pc reflect value mapindex usr local go src reflect value go fp sp pc encoding json mapencoder encode usr local go src encoding json encode go fp sp pc encoding json mapencoder encode fm usr local go src encoding json encode go fp sp pc encoding json structencoder encode usr local go src encoding json encode go fp sp pc encoding json structencoder encode fm usr local go src encoding json encode go fp sp pc encoding json ptrencoder encode usr local go src encoding json encode go fp sp pc encoding json ptrencoder encode fm usr local go src encoding json encode go fp sp pc encoding json encodestate reflectvalue usr local go src encoding json encode go fp sp pc encoding json encodestate marshal usr local go src encoding json encode go fp sp pc encoding json marshal usr local go src encoding json encode go fp sp pc github com incognitochain incognito chain blockchain shardbeststate marshaljson root go src github com incognitochain incognito chain blockchain shardbeststate go fp sp pc github com incognitochain incognito chain blockchain synker root go src github com incognitochain incognito chain blockchain synker go fp sp pc github com incognitochain incognito chain blockchain synker start root go src github com incognitochain incognito chain blockchain synker go fp sp pc runtime goexit usr local go src runtime asm s fp sp pc created by main server start root go src github com incognitochain incognito chain server go goroutine main mainmaster root go src github com incognitochain incognito chain incognito go main main root go src github com incognitochain incognito chain incognito go | 1 |
355,185 | 10,577,268,740 | IssuesEvent | 2019-10-07 19:44:06 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | closed | must gather for OCP fails to run during deploy | Medium Priority | `oc adm must-gather` requires `KUBECONFIG` environment variable to properly run. It is set by `OCP.set_kubeconfig()`, but **only** if path to file exists. This file is created early in `openshift-install create cluster` execution, but earliest `OCP.set_kubeconfig()` call is usually near the end of `deploy` method (through `test_cluster()`).
In other words, errors in `openshift-install` will trigger must gather, but without KUBECONFIG variable set. This causes must gather to fail.
How to reproduce:
1. In config, set `gather_on_deploy_failure` to True and `installer_version` to `4.2.0-0.nightly`
2. Ensure your pull-secret **does not** have `registry.svc.ci.openshift.org` entry (this is sure way to make installer fail)
3. Run `run-ci -m deployment <options> --deploy`
```
$ run-ci -m deployment --ocsci-conf ocs_ci/framework/conf/config.yaml --cluster-conf conf/ocs_basic_install.yml --cluster-name mzalewsk-test --cluster-path /home/mzalewsk/ocs-test/ --deploy
# snip
18:59:49 - MainThread - ocs_ci.utility.utils - INFO - Executing command: /home/mzalewsk/sources/ocs-ci/bin/openshift-install create cluster --dir /home/mzalewsk/ocs-test/ --log-level INFO
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather.
Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker
19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker.
Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
``` | 1.0 | must gather for OCP fails to run during deploy - `oc adm must-gather` requires `KUBECONFIG` environment variable to properly run. It is set by `OCP.set_kubeconfig()`, but **only** if path to file exists. This file is created early in `openshift-install create cluster` execution, but earliest `OCP.set_kubeconfig()` call is usually near the end of `deploy` method (through `test_cluster()`).
In other words, errors in `openshift-install` will trigger must gather, but without KUBECONFIG variable set. This causes must gather to fail.
How to reproduce:
1. In config, set `gather_on_deploy_failure` to True and `installer_version` to `4.2.0-0.nightly`
2. Ensure your pull-secret **does not** have `registry.svc.ci.openshift.org` entry (this is sure way to make installer fail)
3. Run `run-ci -m deployment <options> --deploy`
```
$ run-ci -m deployment --ocsci-conf ocs_ci/framework/conf/config.yaml --cluster-conf conf/ocs_basic_install.yml --cluster-name mzalewsk-test --cluster-path /home/mzalewsk/ocs-test/ --deploy
# snip
18:59:49 - MainThread - ocs_ci.utility.utils - INFO - Executing command: /home/mzalewsk/sources/ocs-ci/bin/openshift-install create cluster --dir /home/mzalewsk/ocs-test/ --log-level INFO
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather.
Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used.
19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather
19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker
19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker.
Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file:
1. Via the command-line flag --config
2. Via the KUBECONFIG environment variable
3. In your home directory as ~/.kube/config
To view or setup config directly use the 'config' command.
``` | priority | must gather for ocp fails to run during deploy oc adm must gather requires kubeconfig environment variable to properly run it is set by ocp set kubeconfig but only if path to file exists this file is created early in openshift install create cluster execution but earliest ocp set kubeconfig call is usually near the end of deploy method through test cluster in other words errors in openshift install will trigger must gather but without kubeconfig variable set this causes must gather to fail how to reproduce in config set gather on deploy failure to true and installer version to nightly ensure your pull secret does not have registry svc ci openshift org entry this is sure way to make installer fail run run ci m deployment deploy run ci m deployment ocsci conf ocs ci framework conf config yaml cluster conf conf ocs basic install yml cluster name mzalewsk test cluster path home mzalewsk ocs test deploy snip mainthread ocs ci utility utils info executing command home mzalewsk sources ocs ci bin openshift install create cluster dir home mzalewsk ocs test log level info mainthread ocs ci ocs utils info must gather image quay io openshift origin must gather will be used mainthread ocs ci ocs utils info ocs logs will be placed in location tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci utility utils info executing command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci ocs utils error failed during must gather logs error error during execution of command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather error is error missing or incomplete configuration info please login or point to an existing complete config file via the command line flag config via the kubeconfig environment variable in your home directory as kube config to view or setup config directly use the config command mainthread ocs ci ocs utils info must gather image quay io openshift origin must gather will be used mainthread ocs ci ocs utils info ocs logs will be placed in location tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci utility utils info executing command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather usr bin gather service logs worker mainthread ocs ci ocs utils error failed during must gather logs error error during execution of command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather usr bin gather service logs worker error is error missing or incomplete configuration info please login or point to an existing complete config file via the command line flag config via the kubeconfig environment variable in your home directory as kube config to view or setup config directly use the config command | 1 |
114,223 | 11,841,098,679 | IssuesEvent | 2020-03-23 20:07:52 | gaugsouza/petshow-backend | https://api.github.com/repos/gaugsouza/petshow-backend | closed | Definir e escrever sobre Gestão de tempo (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.) | documentation | (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.) | 1.0 | Definir e escrever sobre Gestão de tempo (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.) - (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.) | non_priority | definir e escrever sobre gestão de tempo exemplos cronograma diagrama de gantt formato de sprints etc exemplos cronograma diagrama de gantt formato de sprints etc | 0 |
712,545 | 24,498,660,617 | IssuesEvent | 2022-10-10 10:54:17 | AY2223S1-CS2103T-W15-4/tp | https://api.github.com/repos/AY2223S1-CS2103T-W15-4/tp | closed | As a user, I am given descriptive and helpful error messages when I give an invalid or incorrect command | type.Story priority.High | ...so that I can troubleshoot easily without consulting the User Guide every time | 1.0 | As a user, I am given descriptive and helpful error messages when I give an invalid or incorrect command - ...so that I can troubleshoot easily without consulting the User Guide every time | priority | as a user i am given descriptive and helpful error messages when i give an invalid or incorrect command so that i can troubleshoot easily without consulting the user guide every time | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.