Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
502,585
14,549,716,918
IssuesEvent
2020-12-16 04:34:55
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Tenant wise publisher config loading returns the resource type's files instead of a specific resource's file
Priority/Normal Severity/Minor bug
**Describe the issue:** When loading tenant publisher configuration from ConfigStore, in `loadTenantPublisherConfigurationFromConfigStore` method, resource.getFiles return all the files related to **resource type** instead of a particular **resource's files**. https://github.com/wso2-extensions/identity-governance/blob/9a04f338d343c412d3dae000f6502825a253f32e/components/org.wso2.carbon.identity.tenant.resource.manager/src/main/java/org/wso2/carbon/identity/tenant/resource/manager/TenantAwareAxis2ConfigurationContextObserver.java#L106 **Cause for the issue:** Setting files to the resource is erroneous. This function sets the entire resource type's files. https://github.com/wso2/carbon-identity-framework/blob/b0e78d597482c922503156a6834c753ab2f3f2f7/components/configuration-mgt/org.wso2.carbon.identity.configuration.mgt.core/src/main/java/org/wso2/carbon/identity/configuration/mgt/core/dao/impl/ConfigurationDAOImpl.java#L1531 **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: IS 5.11.0 - OS: Linux - Database: H2 - Userstore: LDAP
1.0
Tenant wise publisher config loading returns the resource type's files instead of a specific resource's file - **Describe the issue:** When loading tenant publisher configuration from ConfigStore, in `loadTenantPublisherConfigurationFromConfigStore` method, resource.getFiles return all the files related to **resource type** instead of a particular **resource's files**. https://github.com/wso2-extensions/identity-governance/blob/9a04f338d343c412d3dae000f6502825a253f32e/components/org.wso2.carbon.identity.tenant.resource.manager/src/main/java/org/wso2/carbon/identity/tenant/resource/manager/TenantAwareAxis2ConfigurationContextObserver.java#L106 **Cause for the issue:** Setting files to the resource is erroneous. This function sets the entire resource type's files. https://github.com/wso2/carbon-identity-framework/blob/b0e78d597482c922503156a6834c753ab2f3f2f7/components/configuration-mgt/org.wso2.carbon.identity.configuration.mgt.core/src/main/java/org/wso2/carbon/identity/configuration/mgt/core/dao/impl/ConfigurationDAOImpl.java#L1531 **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: IS 5.11.0 - OS: Linux - Database: H2 - Userstore: LDAP
priority
tenant wise publisher config loading returns the resource type s files instead of a specific resource s file describe the issue when loading tenant publisher configuration from configstore in loadtenantpublisherconfigurationfromconfigstore method resource getfiles return all the files related to resource type instead of a particular resource s files cause for the issue setting files to the resource is erroneous this function sets the entire resource type s files environment information please complete the following information remove any unnecessary fields product version is os linux database userstore ldap
1
165,223
26,125,504,747
IssuesEvent
2022-12-28 17:55:53
KitWallace/FO
https://api.github.com/repos/KitWallace/FO
closed
Profile parameters sb on sliders
redesign
Would need to hide/show these which could hide show the visualised profile too
1.0
Profile parameters sb on sliders - Would need to hide/show these which could hide show the visualised profile too
non_priority
profile parameters sb on sliders would need to hide show these which could hide show the visualised profile too
0
797,575
28,148,744,726
IssuesEvent
2023-04-02 19:46:45
Greenstand/treetracker-admin-client
https://api.github.com/repos/Greenstand/treetracker-admin-client
opened
Species not working in Captures or Capture Details
type: bug good first issue priority size: small tool: Captures
Related to #1017 This is a quick fix to get species information back into the Captures page, filter and Capture Detail dialog. Not working: * Species filter in _Captures_ tool * Species column in _Captures_ tool * Species field in Capture Detail dialog The new Query API expects a UUID for the species, available as `species.uuid` in the legacy Species data. In the future, species information will be part of a dedicated schema, indexed by UUID.
1.0
Species not working in Captures or Capture Details - Related to #1017 This is a quick fix to get species information back into the Captures page, filter and Capture Detail dialog. Not working: * Species filter in _Captures_ tool * Species column in _Captures_ tool * Species field in Capture Detail dialog The new Query API expects a UUID for the species, available as `species.uuid` in the legacy Species data. In the future, species information will be part of a dedicated schema, indexed by UUID.
priority
species not working in captures or capture details related to this is a quick fix to get species information back into the captures page filter and capture detail dialog not working species filter in captures tool species column in captures tool species field in capture detail dialog the new query api expects a uuid for the species available as species uuid in the legacy species data in the future species information will be part of a dedicated schema indexed by uuid
1
74,663
7,435,012,627
IssuesEvent
2018-03-26 13:01:47
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
InvocationTargetException with -Xshareclasses from Spring App
test failure userRaised
Running the petclinic sample spring app (https://github.com/spring-projects/spring-petclinic) with `-Xscmx50M -Xshareclasses` causes `Exception in thread "main" java.lang.reflect.InvocationTargetException` on second run. Output: `2018-02-22 12:46:31.942 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : Starting PetClinicApplication v2.0.0 on 85de4731ae7f with PID 1 (/tmp/spring-petclinic-2.0.0.jar started by root in /) 2018-02-22 12:46:31.957 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : No active profile set, falling back to default profiles: default 2018-02-22 12:46:32.082 INFO 1 --- [ main] ConfigServletWebServerApplicationContext : Refreshing org.springframework.boot.web.servlet.context.AnnotationConfigServletWebServerApplicationContext@44fd0272: startup date [Thu Feb 22 12:46:32 UTC 2018]; root of context hierarchy 2018-02-22 12:46:35.192 INFO 1 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration$$EnhancerBySpringCGLIB$$1cb4efe6] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) Exception in thread "main" java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48) at org.springframework.boot.loader.Launcher.launch(Launcher.java:87) at org.springframework.boot.loader.Launcher.launch(Launcher.java:50) at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:51) Caused by: java.lang.NoClassDefFoundError: ch.qos.logback.classic.spi.ThrowableProxy at ch.qos.logback.classic.spi.LoggingEvent.<init>(LoggingEvent.java:119) at ch.qos.logback.classic.Logger.buildLoggingEventAndAppend(Logger.java:419) at ch.qos.logback.classic.Logger.filterAndLog_0_Or3Plus(Logger.java:383) at ch.qos.logback.classic.Logger.log(Logger.java:765) at org.apache.logging.slf4j.SLF4JLogger.logMessage(SLF4JLogger.java:232) at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2163) at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2118) at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2101) at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:1995) at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1967) at org.apache.commons.logging.LogFactory$Log4jLog.log(LogFactory.java:302) at org.apache.commons.logging.LogFactory$Log4jLog.error(LogFactory.java:254) at org.springframework.boot.SpringApplication.reportFailure(SpringApplication.java:833) at org.springframework.boot.SpringApplication.handleRunFailure(SpringApplication.java:807) at org.springframework.boot.SpringApplication.run(SpringApplication.java:338) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1246) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1234) at org.springframework.samples.petclinic.PetClinicApplication.main(PetClinicApplication.java:32) ... 8 more Caused by: java.lang.ClassNotFoundException: ch.qos.logback.classic.spi.ThrowableProxy at java.net.URLClassLoader.findClass(URLClassLoader.java:588) at java.lang.ClassLoader.loadClassHelper(ClassLoader.java:933) at java.lang.ClassLoader.loadClass(ClassLoader.java:878) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:861) ... 26 more `
1.0
InvocationTargetException with -Xshareclasses from Spring App - Running the petclinic sample spring app (https://github.com/spring-projects/spring-petclinic) with `-Xscmx50M -Xshareclasses` causes `Exception in thread "main" java.lang.reflect.InvocationTargetException` on second run. Output: `2018-02-22 12:46:31.942 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : Starting PetClinicApplication v2.0.0 on 85de4731ae7f with PID 1 (/tmp/spring-petclinic-2.0.0.jar started by root in /) 2018-02-22 12:46:31.957 INFO 1 --- [ main] o.s.s.petclinic.PetClinicApplication : No active profile set, falling back to default profiles: default 2018-02-22 12:46:32.082 INFO 1 --- [ main] ConfigServletWebServerApplicationContext : Refreshing org.springframework.boot.web.servlet.context.AnnotationConfigServletWebServerApplicationContext@44fd0272: startup date [Thu Feb 22 12:46:32 UTC 2018]; root of context hierarchy 2018-02-22 12:46:35.192 INFO 1 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration$$EnhancerBySpringCGLIB$$1cb4efe6] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) Exception in thread "main" java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:48) at org.springframework.boot.loader.Launcher.launch(Launcher.java:87) at org.springframework.boot.loader.Launcher.launch(Launcher.java:50) at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:51) Caused by: java.lang.NoClassDefFoundError: ch.qos.logback.classic.spi.ThrowableProxy at ch.qos.logback.classic.spi.LoggingEvent.<init>(LoggingEvent.java:119) at ch.qos.logback.classic.Logger.buildLoggingEventAndAppend(Logger.java:419) at ch.qos.logback.classic.Logger.filterAndLog_0_Or3Plus(Logger.java:383) at ch.qos.logback.classic.Logger.log(Logger.java:765) at org.apache.logging.slf4j.SLF4JLogger.logMessage(SLF4JLogger.java:232) at org.apache.logging.log4j.spi.AbstractLogger.tryLogMessage(AbstractLogger.java:2163) at org.apache.logging.log4j.spi.AbstractLogger.logMessageTrackRecursion(AbstractLogger.java:2118) at org.apache.logging.log4j.spi.AbstractLogger.logMessageSafely(AbstractLogger.java:2101) at org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:1995) at org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:1967) at org.apache.commons.logging.LogFactory$Log4jLog.log(LogFactory.java:302) at org.apache.commons.logging.LogFactory$Log4jLog.error(LogFactory.java:254) at org.springframework.boot.SpringApplication.reportFailure(SpringApplication.java:833) at org.springframework.boot.SpringApplication.handleRunFailure(SpringApplication.java:807) at org.springframework.boot.SpringApplication.run(SpringApplication.java:338) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1246) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1234) at org.springframework.samples.petclinic.PetClinicApplication.main(PetClinicApplication.java:32) ... 8 more Caused by: java.lang.ClassNotFoundException: ch.qos.logback.classic.spi.ThrowableProxy at java.net.URLClassLoader.findClass(URLClassLoader.java:588) at java.lang.ClassLoader.loadClassHelper(ClassLoader.java:933) at java.lang.ClassLoader.loadClass(ClassLoader.java:878) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:861) ... 26 more `
non_priority
invocationtargetexception with xshareclasses from spring app running the petclinic sample spring app with xshareclasses causes exception in thread main java lang reflect invocationtargetexception on second run output info o s s petclinic petclinicapplication starting petclinicapplication on with pid tmp spring petclinic jar started by root in info o s s petclinic petclinicapplication no active profile set falling back to default profiles default info configservletwebserverapplicationcontext refreshing org springframework boot web servlet context annotationconfigservletwebserverapplicationcontext startup date root of context hierarchy info trationdelegate beanpostprocessorchecker bean org springframework transaction annotation proxytransactionmanagementconfiguration of type is not eligible for getting processed by all beanpostprocessors for example not eligible for auto proxying exception in thread main java lang reflect invocationtargetexception at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org springframework boot loader mainmethodrunner run mainmethodrunner java at org springframework boot loader launcher launch launcher java at org springframework boot loader launcher launch launcher java at org springframework boot loader jarlauncher main jarlauncher java caused by java lang noclassdeffounderror ch qos logback classic spi throwableproxy at ch qos logback classic spi loggingevent loggingevent java at ch qos logback classic logger buildloggingeventandappend logger java at ch qos logback classic logger filterandlog logger java at ch qos logback classic logger log logger java at org apache logging logmessage java at org apache logging spi abstractlogger trylogmessage abstractlogger java at org apache logging spi abstractlogger logmessagetrackrecursion abstractlogger java at org apache logging spi abstractlogger logmessagesafely abstractlogger java at org apache logging spi abstractlogger logmessage abstractlogger java at org apache logging spi abstractlogger logifenabled abstractlogger java at org apache commons logging logfactory log logfactory java at org apache commons logging logfactory error logfactory java at org springframework boot springapplication reportfailure springapplication java at org springframework boot springapplication handlerunfailure springapplication java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org springframework samples petclinic petclinicapplication main petclinicapplication java more caused by java lang classnotfoundexception ch qos logback classic spi throwableproxy at java net urlclassloader findclass urlclassloader java at java lang classloader loadclasshelper classloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java more
0
73,501
14,085,472,318
IssuesEvent
2020-11-05 01:05:14
microsoft/vscode-python
https://api.github.com/repos/microsoft/vscode-python
opened
GHA: Detect a new Virtual Environment fails with timeout
area-internal classify type-code health
``` 1) GlobalVirtualEnvironment Locator Detect a new Virtual Environment: Error: Timeout of 25000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (d:\a\vscode-python\vscode-python\path with spaces\out\test\pythonEnvironments\discovery\locators\globalVirtualEnvironmentLocator.testvirtualenvs.js) at listOnTimeout (internal/timers.js:531:17) at processTimers (internal/timers.js:475:7) ```
1.0
GHA: Detect a new Virtual Environment fails with timeout - ``` 1) GlobalVirtualEnvironment Locator Detect a new Virtual Environment: Error: Timeout of 25000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (d:\a\vscode-python\vscode-python\path with spaces\out\test\pythonEnvironments\discovery\locators\globalVirtualEnvironmentLocator.testvirtualenvs.js) at listOnTimeout (internal/timers.js:531:17) at processTimers (internal/timers.js:475:7) ```
non_priority
gha detect a new virtual environment fails with timeout globalvirtualenvironment locator detect a new virtual environment error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves d a vscode python vscode python path with spaces out test pythonenvironments discovery locators globalvirtualenvironmentlocator testvirtualenvs js at listontimeout internal timers js at processtimers internal timers js
0
65,913
3,248,157,952
IssuesEvent
2015-10-17 02:39:05
UniVR/GolfVR
https://api.github.com/repos/UniVR/GolfVR
closed
Hole ! panel at the beginning of Hole 3 !
priority:normal type:bug
There is a panel saying "Hole !" at the begining of the hole 3 !
1.0
Hole ! panel at the beginning of Hole 3 ! - There is a panel saying "Hole !" at the begining of the hole 3 !
priority
hole panel at the beginning of hole there is a panel saying hole at the begining of the hole
1
788,760
27,764,945,961
IssuesEvent
2023-03-16 10:49:48
googleapis/java-firestore
https://api.github.com/repos/googleapis/java-firestore
closed
firestore.it.ITQueryWatchTest: emptyResults_newDocument_ADDED failed
priority: p2 type: bug api: firestore flakybot: issue
Note: #1214 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: a3cc6febfe9edc289db0968ee583c39143190ca5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/aefd2208-2139-4aa2-8c39-894aff3f8750), [Sponge](http://sponge2/aefd2208-2139-4aa2-8c39-894aff3f8750) status: failed <details><summary>Test output</summary><br><pre>events[ event{error=null,value={documents[],documentChanges[]}}, ] value of : size() expected to be in range: [2..2] but was : 1 at com.google.cloud.firestore.it.ITQueryWatchTest$QuerySnapshotEventListener$ListenerAssertions.eventCountIsAnyOf(ITQueryWatchTest.java:809) at com.google.cloud.firestore.it.ITQueryWatchTest.emptyResults_newDocument_ADDED(ITQueryWatchTest.java:243) </pre></details>
1.0
firestore.it.ITQueryWatchTest: emptyResults_newDocument_ADDED failed - Note: #1214 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: a3cc6febfe9edc289db0968ee583c39143190ca5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/aefd2208-2139-4aa2-8c39-894aff3f8750), [Sponge](http://sponge2/aefd2208-2139-4aa2-8c39-894aff3f8750) status: failed <details><summary>Test output</summary><br><pre>events[ event{error=null,value={documents[],documentChanges[]}}, ] value of : size() expected to be in range: [2..2] but was : 1 at com.google.cloud.firestore.it.ITQueryWatchTest$QuerySnapshotEventListener$ListenerAssertions.eventCountIsAnyOf(ITQueryWatchTest.java:809) at com.google.cloud.firestore.it.ITQueryWatchTest.emptyResults_newDocument_ADDED(ITQueryWatchTest.java:243) </pre></details>
priority
firestore it itquerywatchtest emptyresults newdocument added failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output events event error null value documents documentchanges value of size expected to be in range but was at com google cloud firestore it itquerywatchtest querysnapshoteventlistener listenerassertions eventcountisanyof itquerywatchtest java at com google cloud firestore it itquerywatchtest emptyresults newdocument added itquerywatchtest java
1
6,105
13,735,166,857
IssuesEvent
2020-10-05 09:45:18
onelocom/onelo
https://api.github.com/repos/onelocom/onelo
opened
Logging
architecture
## What We want to record every relevant action that happens during each phase ## Why We want to be able to debug, report to the user, etc. ## Ideas 1. Use a `log` table that stores events as blobs (e.g. json blobs). 2. Use a set of tables specialised by type of log, e.g. `sourcing_entry_log` to capture what actions happened for each file during the sourcing phase.
1.0
Logging - ## What We want to record every relevant action that happens during each phase ## Why We want to be able to debug, report to the user, etc. ## Ideas 1. Use a `log` table that stores events as blobs (e.g. json blobs). 2. Use a set of tables specialised by type of log, e.g. `sourcing_entry_log` to capture what actions happened for each file during the sourcing phase.
non_priority
logging what we want to record every relevant action that happens during each phase why we want to be able to debug report to the user etc ideas use a log table that stores events as blobs e g json blobs use a set of tables specialised by type of log e g sourcing entry log to capture what actions happened for each file during the sourcing phase
0
3,982
2,544,328,980
IssuesEvent
2015-01-29 08:57:05
integram/cleverbus
https://api.github.com/repos/integram/cleverbus
closed
Handled requests/responses (body of exchange) will be dumped as a String
bug Core Priority major
In RequestSendingEventNotifier and ResponseReceiveEventNotifier it is necessary to store dumped requests/responses as a String, e.g.: * String (for example XML) is dumped as a String * byte array is dumped as HEX String * Object as instance of HumanReadable is dumped by **toHumanString()** method * others are dumped by by **toString()** method of object
1.0
Handled requests/responses (body of exchange) will be dumped as a String - In RequestSendingEventNotifier and ResponseReceiveEventNotifier it is necessary to store dumped requests/responses as a String, e.g.: * String (for example XML) is dumped as a String * byte array is dumped as HEX String * Object as instance of HumanReadable is dumped by **toHumanString()** method * others are dumped by by **toString()** method of object
priority
handled requests responses body of exchange will be dumped as a string in requestsendingeventnotifier and responsereceiveeventnotifier it is necessary to store dumped requests responses as a string e g string for example xml is dumped as a string byte array is dumped as hex string object as instance of humanreadable is dumped by tohumanstring method others are dumped by by tostring method of object
1
159,068
6,040,198,996
IssuesEvent
2017-06-10 11:57:09
OWASP/BLT
https://api.github.com/repos/OWASP/BLT
closed
Have badges for different types of bugs you've found
enhancement fullstack in-progress priority: high
- See the error types on the homepage, there are 6
1.0
Have badges for different types of bugs you've found - - See the error types on the homepage, there are 6
priority
have badges for different types of bugs you ve found see the error types on the homepage there are
1
306,426
23,159,799,114
IssuesEvent
2022-07-29 16:25:19
MaxBQb/InversionFilterManager
https://api.github.com/repos/MaxBQb/InversionFilterManager
closed
Where is the Readme file?
documentation good first issue
No visitors -> author not interested in investing his own time to write readme file -> No readme file -> random visitors don't understand what's going on -> Visitors not interested in project -> no visitors
1.0
Where is the Readme file? - No visitors -> author not interested in investing his own time to write readme file -> No readme file -> random visitors don't understand what's going on -> Visitors not interested in project -> no visitors
non_priority
where is the readme file no visitors author not interested in investing his own time to write readme file no readme file random visitors don t understand what s going on visitors not interested in project no visitors
0
12,464
3,274,653,272
IssuesEvent
2015-10-26 12:09:01
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Cannot locate the impl of the trait from nested type parameter.
A-traits A-typesystem E-needstest I-ICE
When I used bound::typeparam as type, thread `rustc` panicked. ```rust use rustc_serialize::{Encodable, Decodable}; use std::fmt::Display; pub trait Entity : Decodable + Encodable + Sized { type Key: Clone + Decodable + Encodable + ToString + Display + Eq + Ord + Sized; fn id(&self) -> Self::Key; fn find_by_id(id: Self::Key) -> Option<Self>; } pub struct DbRef<E: Entity> { pub id: E::Key, } impl<E> DbRef<E> where E: Entity { fn get(self) -> Option<E> { E::find_by_id(self.id) } } ``` Source code: https://github.com/ceram1/appcore/blob/master/src/db/mod.rs Build log: https://travis-ci.org/ceram1/appcore/builds/60641835
1.0
Cannot locate the impl of the trait from nested type parameter. - When I used bound::typeparam as type, thread `rustc` panicked. ```rust use rustc_serialize::{Encodable, Decodable}; use std::fmt::Display; pub trait Entity : Decodable + Encodable + Sized { type Key: Clone + Decodable + Encodable + ToString + Display + Eq + Ord + Sized; fn id(&self) -> Self::Key; fn find_by_id(id: Self::Key) -> Option<Self>; } pub struct DbRef<E: Entity> { pub id: E::Key, } impl<E> DbRef<E> where E: Entity { fn get(self) -> Option<E> { E::find_by_id(self.id) } } ``` Source code: https://github.com/ceram1/appcore/blob/master/src/db/mod.rs Build log: https://travis-ci.org/ceram1/appcore/builds/60641835
non_priority
cannot locate the impl of the trait from nested type parameter when i used bound typeparam as type thread rustc panicked rust use rustc serialize encodable decodable use std fmt display pub trait entity decodable encodable sized type key clone decodable encodable tostring display eq ord sized fn id self self key fn find by id id self key option pub struct dbref pub id e key impl dbref where e entity fn get self option e find by id self id source code build log
0
100,775
4,103,371,359
IssuesEvent
2016-06-04 16:43:36
gama-platform/gama
https://api.github.com/repos/gama-platform/gama
closed
full screen 100% full not 99%
> Enhancement >> Stand-by Display OpenGL OS All Priority Low Version Git
You will say that I am fussy but this is important. As I want to project a GAMA model on a table in some case if table is black I put a black background in GAMA and then only my model will appear on the table. I ve just notice that there is a little white frame on the fullscreen mode (see screenshot) <img width="1920" alt="screen shot 2016-05-26 at 10 40 31" src="https://cloud.githubusercontent.com/assets/3928502/15578552/e00cf318-232e-11e6-87b3-4c5d042bb20f.png"> is there a way to disable this white frame? maybe a simple option of the fullscreen mode?
1.0
full screen 100% full not 99% - You will say that I am fussy but this is important. As I want to project a GAMA model on a table in some case if table is black I put a black background in GAMA and then only my model will appear on the table. I ve just notice that there is a little white frame on the fullscreen mode (see screenshot) <img width="1920" alt="screen shot 2016-05-26 at 10 40 31" src="https://cloud.githubusercontent.com/assets/3928502/15578552/e00cf318-232e-11e6-87b3-4c5d042bb20f.png"> is there a way to disable this white frame? maybe a simple option of the fullscreen mode?
priority
full screen full not you will say that i am fussy but this is important as i want to project a gama model on a table in some case if table is black i put a black background in gama and then only my model will appear on the table i ve just notice that there is a little white frame on the fullscreen mode see screenshot img width alt screen shot at src is there a way to disable this white frame maybe a simple option of the fullscreen mode
1
44,658
2,910,368,497
IssuesEvent
2015-06-21 17:47:51
PredictiveEcology/SpaDES
https://api.github.com/repos/PredictiveEcology/SpaDES
closed
simtime accessors should convert to `simTimeStepUnit`
core SpaDES High priority resolved in branch
continuation of #151 *E.g.*, calling `simCurrentTime(mySim)` in a module that uses time step of "weeks" should return the currunt time in weeks.
1.0
simtime accessors should convert to `simTimeStepUnit` - continuation of #151 *E.g.*, calling `simCurrentTime(mySim)` in a module that uses time step of "weeks" should return the currunt time in weeks.
priority
simtime accessors should convert to simtimestepunit continuation of e g calling simcurrenttime mysim in a module that uses time step of weeks should return the currunt time in weeks
1
56,422
6,979,321,023
IssuesEvent
2017-12-12 20:36:31
PnX-SI/GeoNature
https://api.github.com/repos/PnX-SI/GeoNature
opened
Revoir HOME et customisation
design frontend généricité
- Message d'intro modifiable - Supprimer galerie photo - Pouvoir masquer la carte et les graphs
1.0
Revoir HOME et customisation - - Message d'intro modifiable - Supprimer galerie photo - Pouvoir masquer la carte et les graphs
non_priority
revoir home et customisation message d intro modifiable supprimer galerie photo pouvoir masquer la carte et les graphs
0
358,982
25,211,521,386
IssuesEvent
2022-11-14 04:37:44
SigNoz/signoz-website
https://api.github.com/repos/SigNoz/signoz-website
closed
Linking internal pages
documentation
Will this type of linking work? <img width="841" alt="Screenshot 2022-10-28 at 4 52 39 PM" src="https://user-images.githubusercontent.com/83692067/198575728-69341b32-1786-43ab-92e7-00b31b269ef5.png"> Currently, I am using the entire link like shown below. The below opens in a new page. I think for docs section opening in the same tab makes more sense just like our current behaviour. <img width="942" alt="Screenshot 2022-10-28 at 4 54 12 PM" src="https://user-images.githubusercontent.com/83692067/198575916-e64924b6-ab07-47b2-94f0-8770e717e90c.png">
1.0
Linking internal pages - Will this type of linking work? <img width="841" alt="Screenshot 2022-10-28 at 4 52 39 PM" src="https://user-images.githubusercontent.com/83692067/198575728-69341b32-1786-43ab-92e7-00b31b269ef5.png"> Currently, I am using the entire link like shown below. The below opens in a new page. I think for docs section opening in the same tab makes more sense just like our current behaviour. <img width="942" alt="Screenshot 2022-10-28 at 4 54 12 PM" src="https://user-images.githubusercontent.com/83692067/198575916-e64924b6-ab07-47b2-94f0-8770e717e90c.png">
non_priority
linking internal pages will this type of linking work img width alt screenshot at pm src currently i am using the entire link like shown below the below opens in a new page i think for docs section opening in the same tab makes more sense just like our current behaviour img width alt screenshot at pm src
0
61,115
17,023,608,868
IssuesEvent
2021-07-03 02:54:13
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Merkaartor should show all the available decimal places in coordinates for the node Properties pane
Component: merkaartor Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 9.16am, Monday, 21st June 2010]** Currently, Merkaartor only shows 8 digits for the whole coordinates including the integral part. The schema allows for 7 decimal places so Merkaartor should show that level of precision, at least in the Properties pane of nodes.
1.0
Merkaartor should show all the available decimal places in coordinates for the node Properties pane - **[Submitted to the original trac issue database at 9.16am, Monday, 21st June 2010]** Currently, Merkaartor only shows 8 digits for the whole coordinates including the integral part. The schema allows for 7 decimal places so Merkaartor should show that level of precision, at least in the Properties pane of nodes.
non_priority
merkaartor should show all the available decimal places in coordinates for the node properties pane currently merkaartor only shows digits for the whole coordinates including the integral part the schema allows for decimal places so merkaartor should show that level of precision at least in the properties pane of nodes
0
400,079
27,267,644,070
IssuesEvent
2023-02-22 19:24:28
Iron-E/nvim-libmodal
https://api.github.com/repos/Iron-E/nvim-libmodal
closed
feline archieved switch to lualine?
documentation enhancement
Hi, I wanted to try out your plugin but didn't found an example how to show the mods in lualine, then I saw that you use feline and saw that it's archived by the owner, do you plan to switch to some other plugin like lualine? If yes, I would love to see your configuration for nvim-libmodal for it.
1.0
feline archieved switch to lualine? - Hi, I wanted to try out your plugin but didn't found an example how to show the mods in lualine, then I saw that you use feline and saw that it's archived by the owner, do you plan to switch to some other plugin like lualine? If yes, I would love to see your configuration for nvim-libmodal for it.
non_priority
feline archieved switch to lualine hi i wanted to try out your plugin but didn t found an example how to show the mods in lualine then i saw that you use feline and saw that it s archived by the owner do you plan to switch to some other plugin like lualine if yes i would love to see your configuration for nvim libmodal for it
0
317,135
9,661,303,496
IssuesEvent
2019-05-20 17:40:00
sot/proseco
https://api.github.com/repos/sot/proseco
opened
Remove or address all items that would require manual / SAUSAGE edits
enhancement priority-low
Right now I'm thinking of: * Fid edits (already in #208) * Monitor window converted to guide star (placed in slot 7 with 8x8) * Monitor window DTS I don't recall if we had anything else that we expect to require SAUSAGE edits. I'm not seeing that we'd need ER monitor windows anytime soon, for example.
1.0
Remove or address all items that would require manual / SAUSAGE edits - Right now I'm thinking of: * Fid edits (already in #208) * Monitor window converted to guide star (placed in slot 7 with 8x8) * Monitor window DTS I don't recall if we had anything else that we expect to require SAUSAGE edits. I'm not seeing that we'd need ER monitor windows anytime soon, for example.
priority
remove or address all items that would require manual sausage edits right now i m thinking of fid edits already in monitor window converted to guide star placed in slot with monitor window dts i don t recall if we had anything else that we expect to require sausage edits i m not seeing that we d need er monitor windows anytime soon for example
1
26,511
12,414,630,491
IssuesEvent
2020-05-22 14:52:34
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
Error accessing gitlab.com repo
bug team/core-services
- Go to https://gitlab.com/sourcegraph/codeintellify/-/blob/master/src/hoverifier.ts - See no code intel, error in console: ``` Could not resolve file info for code view Error: RepoLookup for {Repo:gitlab.com/sourcegraph/codeintellify} failed with http status 500: syncer.syncsubset.store.upsert-repos: insert: pq: could not serialize access due to concurrent update ```
1.0
Error accessing gitlab.com repo - - Go to https://gitlab.com/sourcegraph/codeintellify/-/blob/master/src/hoverifier.ts - See no code intel, error in console: ``` Could not resolve file info for code view Error: RepoLookup for {Repo:gitlab.com/sourcegraph/codeintellify} failed with http status 500: syncer.syncsubset.store.upsert-repos: insert: pq: could not serialize access due to concurrent update ```
non_priority
error accessing gitlab com repo go to see no code intel error in console could not resolve file info for code view error repolookup for repo gitlab com sourcegraph codeintellify failed with http status syncer syncsubset store upsert repos insert pq could not serialize access due to concurrent update
0
12,479
7,887,590,017
IssuesEvent
2018-06-27 18:59:30
iodide-project/pyodide
https://api.github.com/repos/iodide-project/pyodide
closed
Fix and/or document issues around finalization of objects
C/C++ Javascript Python/C API WASM performance
From the embind documentation: Embind has support for binding most C++ constructs, including those introduced in C++11 and C++14. Its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics. This means that Python objects will "leak" under the current framework. Should investigate what, if anything, can be done there.
True
Fix and/or document issues around finalization of objects - From the embind documentation: Embind has support for binding most C++ constructs, including those introduced in C++11 and C++14. Its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics. This means that Python objects will "leak" under the current framework. Should investigate what, if anything, can be done there.
non_priority
fix and or document issues around finalization of objects from the embind documentation embind has support for binding most c constructs including those introduced in c and c its only significant limitation is that it does not currently support raw pointers with complicated lifetime semantics this means that python objects will leak under the current framework should investigate what if anything can be done there
0
61,211
7,451,244,690
IssuesEvent
2018-03-29 01:46:09
phetsims/gas-properties
https://api.github.com/repos/phetsims/gas-properties
opened
Allowing toggle between degrees C and K
design:general type:user-feedback
A twitter user mentioned it might be nice to choose between degrees C and degrees K. Since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well.
1.0
Allowing toggle between degrees C and K - A twitter user mentioned it might be nice to choose between degrees C and degrees K. Since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well.
non_priority
allowing toggle between degrees c and k a twitter user mentioned it might be nice to choose between degrees c and degrees k since we allow such a choice in states of matter it seems like it would be useful to have in this sim as well
0
16,909
9,547,573,621
IssuesEvent
2019-05-02 00:05:02
tfussell/xlnt
https://api.github.com/repos/tfussell/xlnt
closed
Optimize compound document streambufs
enhancement performance
Only a single sector should be read/written at a time. Consecutive read/writes to the same sector should not cause the sector to be read/written again. Extract read/write to separate methods.
True
Optimize compound document streambufs - Only a single sector should be read/written at a time. Consecutive read/writes to the same sector should not cause the sector to be read/written again. Extract read/write to separate methods.
non_priority
optimize compound document streambufs only a single sector should be read written at a time consecutive read writes to the same sector should not cause the sector to be read written again extract read write to separate methods
0
214,737
7,276,372,061
IssuesEvent
2018-02-21 16:12:11
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
Status page: we are stripping the username + realm for local accounts
Priority: High Status: In progress Type: Bug
For example if you would use an email address that is a local account on PF, that would strip the username and prevent authentication to be correct. Potential fix would be to have the local source declare it doesn't want a stripped username. Not sure if it would work but it technically should
1.0
Status page: we are stripping the username + realm for local accounts - For example if you would use an email address that is a local account on PF, that would strip the username and prevent authentication to be correct. Potential fix would be to have the local source declare it doesn't want a stripped username. Not sure if it would work but it technically should
priority
status page we are stripping the username realm for local accounts for example if you would use an email address that is a local account on pf that would strip the username and prevent authentication to be correct potential fix would be to have the local source declare it doesn t want a stripped username not sure if it would work but it technically should
1
768,482
26,965,108,836
IssuesEvent
2023-02-08 21:36:04
googleapis/java-storage
https://api.github.com/repos/googleapis/java-storage
closed
gRPC: use of `userProject` options results in error from gcs when used with credentials which define quota_project_id
type: bug api: storage priority: p3
Affected version(s): 2.14.0+ #### Steps to reproduce 1. Set project via gcloud config set project $projectId 2. Create adc following instructions from https://google.aip.dev/auth/4113 3. <details> <summary>JUnit tests to repro failure</summary> ```java import static com.google.common.truth.Truth.assertThat; import com.google.cloud.storage.Bucket; import com.google.cloud.storage.BucketFixture; import com.google.cloud.storage.BucketInfo; import com.google.cloud.storage.Storage; import com.google.cloud.storage.Storage.BucketField; import com.google.cloud.storage.Storage.BucketGetOption; import com.google.cloud.storage.Storage.BucketSourceOption; import com.google.cloud.storage.Storage.BucketTargetOption; import com.google.cloud.storage.StorageFixture; import com.google.cloud.storage.conformance.retry.CleanupStrategy; import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.ClassRule; import org.junit.Test; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.slf4j.Marker; import org.slf4j.MarkerFactory; import org.slf4j.bridge.SLF4JBridgeHandler; public final class ITRequesterPaysTest { static { SLF4JBridgeHandler.removeHandlersForRootLogger(); SLF4JBridgeHandler.install(); } private static final Logger LOGGER = LoggerFactory.getLogger(ITRequesterPaysTest.class); private static final Marker TRACE_ENTER = MarkerFactory.getMarker("enter"); private static final Marker TRACE_EXIT = MarkerFactory.getMarker("exit"); @ClassRule public static final StorageFixture sfGadc = StorageFixture.defaultGrpc(); private static final BucketFixture bf = BucketFixture.newBuilder() .setCleanupStrategy(CleanupStrategy.NEVER) .setBucketNameFmtString("java-storage-grpc-%s") .setHandle(sfJadc::getInstance) .build(); private static final String bucketName = bf.newBucketName(); @BeforeClass public static void beforeClass() { LOGGER.trace(TRACE_ENTER, "beforeClass()"); try { BucketInfo info1 = BucketInfo.newBuilder(bucketName).setRequesterPays(true).build(); Bucket bucket = sfJadc.getInstance().create(info1, BucketTargetOption.projection("noAcl")); assertThat(bucket.requesterPays()).isTrue(); } finally { LOGGER.trace(TRACE_EXIT, "beforeClass()"); } } @AfterClass public static void afterClass() { LOGGER.trace(TRACE_ENTER, "afterClass()"); try { String projectId = sfJadc.getInstance().getOptions().getProjectId(); sfJadc.getInstance().delete(bucketName, BucketSourceOption.userProject(projectId)); } finally { LOGGER.trace(TRACE_EXIT, "afterClass()"); } } @Test public void grpc_adc() { LOGGER.trace(TRACE_ENTER, "grpc_adc()"); try { doTest(sfGadc.getInstance()); } finally { LOGGER.trace(TRACE_EXIT, "grpc_adc()"); } } private static void doTest(Storage storage) { LOGGER.trace(TRACE_ENTER, "doTest(storage : {})", storage); try { String projectId = storage.getOptions().getProjectId(); doGet(storage, projectId); // doList(storage, projectId); } finally { LOGGER.trace(TRACE_EXIT, "doTest(storage : {})", storage); } } private static void doGet(Storage storage, String projectId) { Bucket bucket = storage .get( bucketName, BucketGetOption.fields(BucketField.ID, BucketField.BILLING), BucketGetOption.userProject(projectId)); assertThat(bucket.requesterPays()).isTrue(); } } ``` </details> #### Stack trace ``` com.google.cloud.storage.StorageException: INVALID_ARGUMENT: User project specified in the request is invalid. at com.google.cloud.storage.StorageException.asStorageException(StorageException.java:144) at com.google.cloud.storage.StorageException.coalesce(StorageException.java:117) at com.google.cloud.storage.Retrying.run(Retrying.java:96) at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376) at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154) at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:92) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:98) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:563) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:744) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:723) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:100) at com.google.cloud.storage.GrpcStorageImpl.lambda$get$2(GrpcStorageImpl.java:379) at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103) at com.google.cloud.RetryHelper.run(RetryHelper.java:76) at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50) at com.google.cloud.storage.Retrying.run(Retrying.java:93) at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376) at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154) at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.util.concurrent.FutureTask.run(FutureTask.java:266) ... 1 more Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid. at io.grpc.Status.asRuntimeException(Status.java:535) ... 14 more ``` #### Current Workaround Decorate default credentials to hide quota_project_id ```java Storage s = StorageOptions.grpc() .setCredentials(new QuotaProjectIdHidingCredentials(GoogleCredentials.getApplicationDefault())) .build() .getService(); ``` Internal reference bug: b/248605515
1.0
gRPC: use of `userProject` options results in error from gcs when used with credentials which define quota_project_id - Affected version(s): 2.14.0+ #### Steps to reproduce 1. Set project via gcloud config set project $projectId 2. Create adc following instructions from https://google.aip.dev/auth/4113 3. <details> <summary>JUnit tests to repro failure</summary> ```java import static com.google.common.truth.Truth.assertThat; import com.google.cloud.storage.Bucket; import com.google.cloud.storage.BucketFixture; import com.google.cloud.storage.BucketInfo; import com.google.cloud.storage.Storage; import com.google.cloud.storage.Storage.BucketField; import com.google.cloud.storage.Storage.BucketGetOption; import com.google.cloud.storage.Storage.BucketSourceOption; import com.google.cloud.storage.Storage.BucketTargetOption; import com.google.cloud.storage.StorageFixture; import com.google.cloud.storage.conformance.retry.CleanupStrategy; import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.ClassRule; import org.junit.Test; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.slf4j.Marker; import org.slf4j.MarkerFactory; import org.slf4j.bridge.SLF4JBridgeHandler; public final class ITRequesterPaysTest { static { SLF4JBridgeHandler.removeHandlersForRootLogger(); SLF4JBridgeHandler.install(); } private static final Logger LOGGER = LoggerFactory.getLogger(ITRequesterPaysTest.class); private static final Marker TRACE_ENTER = MarkerFactory.getMarker("enter"); private static final Marker TRACE_EXIT = MarkerFactory.getMarker("exit"); @ClassRule public static final StorageFixture sfGadc = StorageFixture.defaultGrpc(); private static final BucketFixture bf = BucketFixture.newBuilder() .setCleanupStrategy(CleanupStrategy.NEVER) .setBucketNameFmtString("java-storage-grpc-%s") .setHandle(sfJadc::getInstance) .build(); private static final String bucketName = bf.newBucketName(); @BeforeClass public static void beforeClass() { LOGGER.trace(TRACE_ENTER, "beforeClass()"); try { BucketInfo info1 = BucketInfo.newBuilder(bucketName).setRequesterPays(true).build(); Bucket bucket = sfJadc.getInstance().create(info1, BucketTargetOption.projection("noAcl")); assertThat(bucket.requesterPays()).isTrue(); } finally { LOGGER.trace(TRACE_EXIT, "beforeClass()"); } } @AfterClass public static void afterClass() { LOGGER.trace(TRACE_ENTER, "afterClass()"); try { String projectId = sfJadc.getInstance().getOptions().getProjectId(); sfJadc.getInstance().delete(bucketName, BucketSourceOption.userProject(projectId)); } finally { LOGGER.trace(TRACE_EXIT, "afterClass()"); } } @Test public void grpc_adc() { LOGGER.trace(TRACE_ENTER, "grpc_adc()"); try { doTest(sfGadc.getInstance()); } finally { LOGGER.trace(TRACE_EXIT, "grpc_adc()"); } } private static void doTest(Storage storage) { LOGGER.trace(TRACE_ENTER, "doTest(storage : {})", storage); try { String projectId = storage.getOptions().getProjectId(); doGet(storage, projectId); // doList(storage, projectId); } finally { LOGGER.trace(TRACE_EXIT, "doTest(storage : {})", storage); } } private static void doGet(Storage storage, String projectId) { Bucket bucket = storage .get( bucketName, BucketGetOption.fields(BucketField.ID, BucketField.BILLING), BucketGetOption.userProject(projectId)); assertThat(bucket.requesterPays()).isTrue(); } } ``` </details> #### Stack trace ``` com.google.cloud.storage.StorageException: INVALID_ARGUMENT: User project specified in the request is invalid. at com.google.cloud.storage.StorageException.asStorageException(StorageException.java:144) at com.google.cloud.storage.StorageException.coalesce(StorageException.java:117) at com.google.cloud.storage.Retrying.run(Retrying.java:96) at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376) at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154) at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:92) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:98) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:563) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:744) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:723) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:100) at com.google.cloud.storage.GrpcStorageImpl.lambda$get$2(GrpcStorageImpl.java:379) at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103) at com.google.cloud.RetryHelper.run(RetryHelper.java:76) at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50) at com.google.cloud.storage.Retrying.run(Retrying.java:93) at com.google.cloud.storage.GrpcStorageImpl.get(GrpcStorageImpl.java:376) at com.google.cloud.storage.it.ITObjectTest.unsetRequesterPays(ITObjectTest.java:154) at com.google.cloud.storage.it.ITObjectTest.testListBlobRequesterPays(ITObjectTest.java:504) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.util.concurrent.FutureTask.run(FutureTask.java:266) ... 1 more Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: User project specified in the request is invalid. at io.grpc.Status.asRuntimeException(Status.java:535) ... 14 more ``` #### Current Workaround Decorate default credentials to hide quota_project_id ```java Storage s = StorageOptions.grpc() .setCredentials(new QuotaProjectIdHidingCredentials(GoogleCredentials.getApplicationDefault())) .build() .getService(); ``` Internal reference bug: b/248605515
priority
grpc use of userproject options results in error from gcs when used with credentials which define quota project id affected version s steps to reproduce set project via gcloud config set project projectid create adc following instructions from junit tests to repro failure java import static com google common truth truth assertthat import com google cloud storage bucket import com google cloud storage bucketfixture import com google cloud storage bucketinfo import com google cloud storage storage import com google cloud storage storage bucketfield import com google cloud storage storage bucketgetoption import com google cloud storage storage bucketsourceoption import com google cloud storage storage buckettargetoption import com google cloud storage storagefixture import com google cloud storage conformance retry cleanupstrategy import org junit afterclass import org junit beforeclass import org junit classrule import org junit test import org logger import org loggerfactory import org marker import org markerfactory import org bridge public final class itrequesterpaystest static removehandlersforrootlogger install private static final logger logger loggerfactory getlogger itrequesterpaystest class private static final marker trace enter markerfactory getmarker enter private static final marker trace exit markerfactory getmarker exit classrule public static final storagefixture sfgadc storagefixture defaultgrpc private static final bucketfixture bf bucketfixture newbuilder setcleanupstrategy cleanupstrategy never setbucketnamefmtstring java storage grpc s sethandle sfjadc getinstance build private static final string bucketname bf newbucketname beforeclass public static void beforeclass logger trace trace enter beforeclass try bucketinfo bucketinfo newbuilder bucketname setrequesterpays true build bucket bucket sfjadc getinstance create buckettargetoption projection noacl assertthat bucket requesterpays istrue finally logger trace trace exit beforeclass afterclass public static void afterclass logger trace trace enter afterclass try string projectid sfjadc getinstance getoptions getprojectid sfjadc getinstance delete bucketname bucketsourceoption userproject projectid finally logger trace trace exit afterclass test public void grpc adc logger trace trace enter grpc adc try dotest sfgadc getinstance finally logger trace trace exit grpc adc private static void dotest storage storage logger trace trace enter dotest storage storage try string projectid storage getoptions getprojectid doget storage projectid dolist storage projectid finally logger trace trace exit dotest storage storage private static void doget storage storage string projectid bucket bucket storage get bucketname bucketgetoption fields bucketfield id bucketfield billing bucketgetoption userproject projectid assertthat bucket requesterpays istrue stack trace com google cloud storage storageexception invalid argument user project specified in the request is invalid at com google cloud storage storageexception asstorageexception storageexception java at com google cloud storage storageexception coalesce storageexception java at com google cloud storage retrying run retrying java at com google cloud storage grpcstorageimpl get grpcstorageimpl java at com google cloud storage it itobjecttest unsetrequesterpays itobjecttest java at com google cloud storage it itobjecttest testlistblobrequesterpays itobjecttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument user project specified in the request is invalid at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more suppressed com google api gax rpc asynctaskexception asynchronous task failed at com google api gax rpc apiexceptions callandtranslateapiexception apiexceptions java at com google api gax rpc unarycallable call unarycallable java at com google cloud storage grpcstorageimpl lambda get grpcstorageimpl java at com google api gax retrying directretryingexecutor submit directretryingexecutor java at com google cloud retryhelper run retryhelper java at com google cloud retryhelper runwithretries retryhelper java at com google cloud storage retrying run retrying java at com google cloud storage grpcstorageimpl get grpcstorageimpl java at com google cloud storage it itobjecttest unsetrequesterpays itobjecttest java at com google cloud storage it itobjecttest testlistblobrequesterpays itobjecttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements failontimeout callablestatement call failontimeout java at org junit internal runners statements failontimeout callablestatement call failontimeout java at java util concurrent futuretask run futuretask java more caused by io grpc statusruntimeexception invalid argument user project specified in the request is invalid at io grpc status asruntimeexception status java more current workaround decorate default credentials to hide quota project id java storage s storageoptions grpc setcredentials new quotaprojectidhidingcredentials googlecredentials getapplicationdefault build getservice internal reference bug b
1
393,733
11,624,079,444
IssuesEvent
2020-02-27 10:07:10
nimblehq/rails-templates
https://api.github.com/repos/nimblehq/rails-templates
closed
Upgrade Node.js version.
priority : normal type : chore
In the Dockerfile, the Node.js version still set to version 8. I think we should upgrade the version number to the latest version of the LTS release. (Now the latest LTS version is 10.16.3)
1.0
Upgrade Node.js version. - In the Dockerfile, the Node.js version still set to version 8. I think we should upgrade the version number to the latest version of the LTS release. (Now the latest LTS version is 10.16.3)
priority
upgrade node js version in the dockerfile the node js version still set to version i think we should upgrade the version number to the latest version of the lts release now the latest lts version is
1
67,163
20,923,492,350
IssuesEvent
2022-03-24 19:51:19
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Group chat icon on HomePage
T-Defect
### Steps to reproduce 1. Open the home page ### Outcome #### What did you expect? The group icon should be displayed. #### What happened instead? It is blank. I guess the code to display the icon has been removed because it included `community` in its file name. ![before](https://user-images.githubusercontent.com/3362943/159931240-0f6c8bc0-4d9f-4014-9822-457a60d74ff7.png) ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp localhost ### Application version develop branch ### Homeserver _No response_ ### Will you send logs? No
1.0
Group chat icon on HomePage - ### Steps to reproduce 1. Open the home page ### Outcome #### What did you expect? The group icon should be displayed. #### What happened instead? It is blank. I guess the code to display the icon has been removed because it included `community` in its file name. ![before](https://user-images.githubusercontent.com/3362943/159931240-0f6c8bc0-4d9f-4014-9822-457a60d74ff7.png) ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp localhost ### Application version develop branch ### Homeserver _No response_ ### Will you send logs? No
non_priority
group chat icon on homepage steps to reproduce open the home page outcome what did you expect the group icon should be displayed what happened instead it is blank i guess the code to display the icon has been removed because it included community in its file name operating system no response browser information no response url for webapp localhost application version develop branch homeserver no response will you send logs no
0
194,779
6,898,967,563
IssuesEvent
2017-11-24 11:41:23
minio/mint
https://api.github.com/repos/minio/mint
closed
Gateway: Awscli head-object test fails
priority: medium
Awscli headobject test fails on all gateways. Here is the failure entry ``` {"name": "awscli", "duration": 7574, "function": "aws --endpoint-url http://192.168.86.102:9000 s3api head-object --bucket awscli-mint-test-bucket-11375 --key prefix/directory/\n", "status": "FAIL", "error": "An error occurred (404) when calling the HeadObject operation: Not Found"} ``` This call succeeds on regular server mode.
1.0
Gateway: Awscli head-object test fails - Awscli headobject test fails on all gateways. Here is the failure entry ``` {"name": "awscli", "duration": 7574, "function": "aws --endpoint-url http://192.168.86.102:9000 s3api head-object --bucket awscli-mint-test-bucket-11375 --key prefix/directory/\n", "status": "FAIL", "error": "An error occurred (404) when calling the HeadObject operation: Not Found"} ``` This call succeeds on regular server mode.
priority
gateway awscli head object test fails awscli headobject test fails on all gateways here is the failure entry name awscli duration function aws endpoint url head object bucket awscli mint test bucket key prefix directory n status fail error an error occurred when calling the headobject operation not found this call succeeds on regular server mode
1
183,539
14,236,465,042
IssuesEvent
2020-11-18 16:01:17
serverless/serverless
https://api.github.com/repos/serverless/serverless
opened
Tests: Refactor lib/plugins/aws/package/lib/mergeIamTemplates.test.js
good first issue help wanted refactor tests
_Most of the tests that currently cover core functionalities depend heavily on (and sometimes test) internal implementation characteristics. While they should be testing whether given implementation produces desired outcome (treating its implementation more as a black box)._ _Current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce, as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities ._ _At some point we've introduced a new (black box based) way of testing the internals. It's through `runServerless` utlity, which allows us create natural (as in real world) `serverless` instance, and inspect the produced outcome for chosen command. More details here: https://github.com/serverless/serverless/tree/master/test#unit-tests_ --- This is issue is about refactoring [lib/plugins/aws/package/lib/mergeIamTemplates.test.js](https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js) to `runServerless` based variant. It's needed, so we cleanly move forward with https://github.com/serverless/serverless/issues/8396 To make refactor relatively easy: - Spec for new tests is fully defined and is already committed in. It's in scope of this _describe_ block: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L774 - Each test slot indicates which test it is about to replace - It is indicated where (and how many) exactly `runServerless` runs are expected to happen, against which fixture and command they should be based - Necessary hints are provided For every refactored test, the old test should be removed Refactor can be submitted with one PR, but partial refactors that cover some of below listed parts are also very welcome: - [ ] __No default role__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L776-L793 - [ ] __Default role with defaults__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L808-L828 - [ ] __Provider properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L839-L865 - [ ] __Function properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L890-L893
1.0
Tests: Refactor lib/plugins/aws/package/lib/mergeIamTemplates.test.js - _Most of the tests that currently cover core functionalities depend heavily on (and sometimes test) internal implementation characteristics. While they should be testing whether given implementation produces desired outcome (treating its implementation more as a black box)._ _Current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce, as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities ._ _At some point we've introduced a new (black box based) way of testing the internals. It's through `runServerless` utlity, which allows us create natural (as in real world) `serverless` instance, and inspect the produced outcome for chosen command. More details here: https://github.com/serverless/serverless/tree/master/test#unit-tests_ --- This is issue is about refactoring [lib/plugins/aws/package/lib/mergeIamTemplates.test.js](https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js) to `runServerless` based variant. It's needed, so we cleanly move forward with https://github.com/serverless/serverless/issues/8396 To make refactor relatively easy: - Spec for new tests is fully defined and is already committed in. It's in scope of this _describe_ block: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L774 - Each test slot indicates which test it is about to replace - It is indicated where (and how many) exactly `runServerless` runs are expected to happen, against which fixture and command they should be based - Necessary hints are provided For every refactored test, the old test should be removed Refactor can be submitted with one PR, but partial refactors that cover some of below listed parts are also very welcome: - [ ] __No default role__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L776-L793 - [ ] __Default role with defaults__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L808-L828 - [ ] __Provider properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L839-L865 - [ ] __Function properties__: https://github.com/serverless/serverless/blob/aa42b3bb899b1b770aeea013b51a4a3f2916427e/lib/plugins/aws/package/lib/mergeIamTemplates.test.js#L890-L893
non_priority
tests refactor lib plugins aws package lib mergeiamtemplates test js most of the tests that currently cover core functionalities depend heavily on and sometimes test internal implementation characteristics while they should be testing whether given implementation produces desired outcome treating its implementation more as a black box current state of things is problematic for eventual internal improvements and refactors which occasionally we want to introduce as in most of such cases proposed improvements need to be accompanied with counterproductive numerous updates to tests which are covering otherwise not altered functionalities at some point we ve introduced a new black box based way of testing the internals it s through runserverless utlity which allows us create natural as in real world serverless instance and inspect the produced outcome for chosen command more details here this is issue is about refactoring to runserverless based variant it s needed so we cleanly move forward with to make refactor relatively easy spec for new tests is fully defined and is already committed in it s in scope of this describe block each test slot indicates which test it is about to replace it is indicated where and how many exactly runserverless runs are expected to happen against which fixture and command they should be based necessary hints are provided for every refactored test the old test should be removed refactor can be submitted with one pr but partial refactors that cover some of below listed parts are also very welcome no default role default role with defaults provider properties function properties
0
322,075
23,888,322,592
IssuesEvent
2022-09-08 09:26:51
k0sproject/k0s
https://api.github.com/repos/k0sproject/k0s
closed
Improve dual-stack documentation
documentation enhancement
### Is your feature request related to a problem? Please describe. * The template shown at https://docs.k0sproject.io/v1.24.3+k0s.0/dual-stack/ should specify `spec.network.provider = calico` for it to work. Otherwise the cluster will be using kube-proxy. * The template mentioned above says "Use the following k0s.yaml as a template", but it's only listing the `spec` object. Since k0s seems to accept partial configuration (see #2090), replace the example with an actually valid k0s.yaml file (with apiVersion, kind fields, etc.) * In the configuration docs, there is no mention of `spec.network.calico.mode` accepting the option `bird`. Only `vxlan` and `ipip` are shown to be acceptable. ### Describe the solution you would like _No response_ ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
Improve dual-stack documentation - ### Is your feature request related to a problem? Please describe. * The template shown at https://docs.k0sproject.io/v1.24.3+k0s.0/dual-stack/ should specify `spec.network.provider = calico` for it to work. Otherwise the cluster will be using kube-proxy. * The template mentioned above says "Use the following k0s.yaml as a template", but it's only listing the `spec` object. Since k0s seems to accept partial configuration (see #2090), replace the example with an actually valid k0s.yaml file (with apiVersion, kind fields, etc.) * In the configuration docs, there is no mention of `spec.network.calico.mode` accepting the option `bird`. Only `vxlan` and `ipip` are shown to be acceptable. ### Describe the solution you would like _No response_ ### Describe alternatives you've considered _No response_ ### Additional context _No response_
non_priority
improve dual stack documentation is your feature request related to a problem please describe the template shown at should specify spec network provider calico for it to work otherwise the cluster will be using kube proxy the template mentioned above says use the following yaml as a template but it s only listing the spec object since seems to accept partial configuration see replace the example with an actually valid yaml file with apiversion kind fields etc in the configuration docs there is no mention of spec network calico mode accepting the option bird only vxlan and ipip are shown to be acceptable describe the solution you would like no response describe alternatives you ve considered no response additional context no response
0
200,879
7,017,881,552
IssuesEvent
2017-12-21 11:22:06
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
Duplicated NIC when save a template
Category: CLI Priority: High Status: Accepted Type: Bug
<!--////////////////////////////////////////////--> <!-- COMPLETE ONLY ONE SECTION --> <!-- BUG REPORT *OR* ENHANCEMENT REQUEST --> <!--////////////////////////////////////////////--> <!--////////////////////////////////////////////--> <!-- COMPLETE THIS SECTION FOR BUG REPORTS --> <!--////////////////////////////////////////////--> # Bug Report ## Version of OpenNebula <!--Mark the relevant versions affected with [X] --> - [ ] 5.2.2 - [X] 5.4.0 - [X] 5.4.1 - [X] 5.4.2 - [X] 5.4.3 - [ ] Development build ## Component <!-- Mark the relevant versions affected with [X] --> - [ ] Authorization (LDAP, x509 certs...) - [X] Command Line Interface (CLI) - [ ] Contextualization - [ ] Documentation - [ ] Federation and HA - [ ] Host, Clusters and Monitorization - [ ] KVM - [ ] Networking - [ ] Orchestration (OpenNebula Flow) - [ ] Packages - [ ] Scheduler - [ ] Storage & Images - [ ] Sunstone - [ ] Upgrades - [ ] User, Groups, VDCs and ACL - [ ] vCenter ## Description <!-- Brief description of your problem --> When you save a virtual Machine with two or more nics associated, the template generated has double the nics of the original machine. ### Expected Behavior ### Actual Behavior ## How to reproduce <!-- Steps to reproduce the issue --> Save a Virtual Machine with two NICs associated. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> # Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Duplicated NIC when save a template - <!--////////////////////////////////////////////--> <!-- COMPLETE ONLY ONE SECTION --> <!-- BUG REPORT *OR* ENHANCEMENT REQUEST --> <!--////////////////////////////////////////////--> <!--////////////////////////////////////////////--> <!-- COMPLETE THIS SECTION FOR BUG REPORTS --> <!--////////////////////////////////////////////--> # Bug Report ## Version of OpenNebula <!--Mark the relevant versions affected with [X] --> - [ ] 5.2.2 - [X] 5.4.0 - [X] 5.4.1 - [X] 5.4.2 - [X] 5.4.3 - [ ] Development build ## Component <!-- Mark the relevant versions affected with [X] --> - [ ] Authorization (LDAP, x509 certs...) - [X] Command Line Interface (CLI) - [ ] Contextualization - [ ] Documentation - [ ] Federation and HA - [ ] Host, Clusters and Monitorization - [ ] KVM - [ ] Networking - [ ] Orchestration (OpenNebula Flow) - [ ] Packages - [ ] Scheduler - [ ] Storage & Images - [ ] Sunstone - [ ] Upgrades - [ ] User, Groups, VDCs and ACL - [ ] vCenter ## Description <!-- Brief description of your problem --> When you save a virtual Machine with two or more nics associated, the template generated has double the nics of the original machine. ### Expected Behavior ### Actual Behavior ## How to reproduce <!-- Steps to reproduce the issue --> Save a Virtual Machine with two NICs associated. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> # Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
duplicated nic when save a template bug report version of opennebula development build component authorization ldap certs command line interface cli contextualization documentation federation and ha host clusters and monitorization kvm networking orchestration opennebula flow packages scheduler storage images sunstone upgrades user groups vdcs and acl vcenter description when you save a virtual machine with two or more nics associated the template generated has double the nics of the original machine expected behavior actual behavior how to reproduce save a virtual machine with two nics associated progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
451,449
32,028,893,581
IssuesEvent
2023-09-22 10:49:18
airalab/robonomics-wiki
https://api.github.com/repos/airalab/robonomics-wiki
closed
[Deprecation]: Article "Robonomics hass integration" is deprecated due to a release "Robonomics Integration 1.6.0" in "homeassistant-robonomics-integration".
documentation deprecation
### Issue description Article ["Robonomics hass integration"](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) has been automatically marked as deprecated due to a recent release ["Robonomics Integration 1.6.0"](https://github.com/airalab/homeassistant-robonomics-integration/releases/tag/1.6.0) in ["homeassistant-robonomics-integration"](https://github.com/airalab/homeassistant-robonomics-integration). An action has been requested from the article contributors: - @LoSk-p - @nakata5321 - @Fingerling42 Deprecation reference id: e4ad529468ba0f2bb19b8aa6ee47922c ### Doc Page [Robonomics hass integration](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) ### Note This issue has been automatically created by the [Wiki deprecation bot](https://github.com/airalab/robonomics-wiki-deprecation-notifier). If this is a false alarm and the article is not actually deprecated - feel free to close this issue. It will not be created again unless a new release comes out in [homeassistant-robonomics-integration](https://github.com/airalab/homeassistant-robonomics-integration).
1.0
[Deprecation]: Article "Robonomics hass integration" is deprecated due to a release "Robonomics Integration 1.6.0" in "homeassistant-robonomics-integration". - ### Issue description Article ["Robonomics hass integration"](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) has been automatically marked as deprecated due to a recent release ["Robonomics Integration 1.6.0"](https://github.com/airalab/homeassistant-robonomics-integration/releases/tag/1.6.0) in ["homeassistant-robonomics-integration"](https://github.com/airalab/homeassistant-robonomics-integration). An action has been requested from the article contributors: - @LoSk-p - @nakata5321 - @Fingerling42 Deprecation reference id: e4ad529468ba0f2bb19b8aa6ee47922c ### Doc Page [Robonomics hass integration](https://raw.githubusercontent.com/airalab/robonomics-wiki/master/docs/robonomics-hass-integration.md) ### Note This issue has been automatically created by the [Wiki deprecation bot](https://github.com/airalab/robonomics-wiki-deprecation-notifier). If this is a false alarm and the article is not actually deprecated - feel free to close this issue. It will not be created again unless a new release comes out in [homeassistant-robonomics-integration](https://github.com/airalab/homeassistant-robonomics-integration).
non_priority
article robonomics hass integration is deprecated due to a release robonomics integration in homeassistant robonomics integration issue description article has been automatically marked as deprecated due to a recent release in an action has been requested from the article contributors losk p deprecation reference id doc page note this issue has been automatically created by the if this is a false alarm and the article is not actually deprecated feel free to close this issue it will not be created again unless a new release comes out in
0
216,372
7,307,313,093
IssuesEvent
2018-02-28 02:06:21
tgstation/tgstation-server
https://api.github.com/repos/tgstation/tgstation-server
closed
SSL issues after the PR attempts to merge (Test merge)
Bug Priority: High Review
https://i.imgur.com/fjRL2oQ.png This is after updating to the latest TGS3 and updating the JSON.
1.0
SSL issues after the PR attempts to merge (Test merge) - https://i.imgur.com/fjRL2oQ.png This is after updating to the latest TGS3 and updating the JSON.
priority
ssl issues after the pr attempts to merge test merge this is after updating to the latest and updating the json
1
667,341
22,466,486,712
IssuesEvent
2022-06-22 02:33:43
ml4ai/tomcat
https://api.github.com/repos/ml4ai/tomcat
closed
Connect audioStreamer to ASR agent
bug Priority: High
Test session: April 12, 2020 Before starting the trial, speechAgent connected to 2 client maps. During the trial, the speechAgent still only connected to 2 client maps, and did not connect to the 3rd client map (on Tiger). We need a consistent procedure for Clientmaps to connect to speechAgent.
1.0
Connect audioStreamer to ASR agent - Test session: April 12, 2020 Before starting the trial, speechAgent connected to 2 client maps. During the trial, the speechAgent still only connected to 2 client maps, and did not connect to the 3rd client map (on Tiger). We need a consistent procedure for Clientmaps to connect to speechAgent.
priority
connect audiostreamer to asr agent test session april before starting the trial speechagent connected to client maps during the trial the speechagent still only connected to client maps and did not connect to the client map on tiger we need a consistent procedure for clientmaps to connect to speechagent
1
59,589
3,114,532,498
IssuesEvent
2015-09-03 09:21:12
HubTurbo/HubTurbo
https://api.github.com/repos/HubTurbo/HubTurbo
closed
Ensure that globalConfigTest passes
priority.medium
Or at least most of the time, it seems to fail ~half the time now? ``` guitests.UseGlobalConfigsTest > globalConfigTest FAILED org.loadui.testfx.exceptions.NoNodesVisibleException: Matching nodes were found, but none of them were visible. Screenshot saved as /home/travis/build/HubTurbo/HubTurbo/screenshot1441084988176.png. at org.loadui.testfx.GuiTest.getVisibleNodes(GuiTest.java:238) at org.loadui.testfx.GuiTest.findAll(GuiTest.java:251) at org.loadui.testfx.GuiTest.findByCssSelector(GuiTest.java:432) at org.loadui.testfx.GuiTest.find(GuiTest.java:300) at org.loadui.testfx.GuiTest.pointFor(GuiTest.java:1159) at org.loadui.testfx.GuiTest.move(GuiTest.java:849) at org.loadui.testfx.GuiTest.click(GuiTest.java:627) at guitests.UseGlobalConfigsTest.globalConfigTest(UseGlobalConfigsTest.java:68) ```
1.0
Ensure that globalConfigTest passes - Or at least most of the time, it seems to fail ~half the time now? ``` guitests.UseGlobalConfigsTest > globalConfigTest FAILED org.loadui.testfx.exceptions.NoNodesVisibleException: Matching nodes were found, but none of them were visible. Screenshot saved as /home/travis/build/HubTurbo/HubTurbo/screenshot1441084988176.png. at org.loadui.testfx.GuiTest.getVisibleNodes(GuiTest.java:238) at org.loadui.testfx.GuiTest.findAll(GuiTest.java:251) at org.loadui.testfx.GuiTest.findByCssSelector(GuiTest.java:432) at org.loadui.testfx.GuiTest.find(GuiTest.java:300) at org.loadui.testfx.GuiTest.pointFor(GuiTest.java:1159) at org.loadui.testfx.GuiTest.move(GuiTest.java:849) at org.loadui.testfx.GuiTest.click(GuiTest.java:627) at guitests.UseGlobalConfigsTest.globalConfigTest(UseGlobalConfigsTest.java:68) ```
priority
ensure that globalconfigtest passes or at least most of the time it seems to fail half the time now guitests useglobalconfigstest globalconfigtest failed org loadui testfx exceptions nonodesvisibleexception matching nodes were found but none of them were visible screenshot saved as home travis build hubturbo hubturbo png at org loadui testfx guitest getvisiblenodes guitest java at org loadui testfx guitest findall guitest java at org loadui testfx guitest findbycssselector guitest java at org loadui testfx guitest find guitest java at org loadui testfx guitest pointfor guitest java at org loadui testfx guitest move guitest java at org loadui testfx guitest click guitest java at guitests useglobalconfigstest globalconfigtest useglobalconfigstest java
1
292,350
8,956,651,425
IssuesEvent
2019-01-26 19:20:25
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256
medium priority
## 🐛 Bug torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256 ## To Reproduce ``` import torch batch_size = 256*256 transform_parameters = torch.cuda.FloatTensor([[1,0,0],[0,1,0]]) transform_parameters = torch.stack([transform_parameters] * batch_size, 0).contiguous() resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2))) print(resampling_grids.size()) ``` crashes with the following error: ``` Traceback (most recent call last): File "bug.py", line 5, in <module> resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2))) File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/functional.py", line 2615, in affine_grid return vision.affine_grid_generator(theta, size) File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/_functions/vision.py", line 10, in affine_grid_generator ret = torch.cudnn_affine_grid_generator(theta, N, C, H, W) RuntimeError: cuDNN error: CUDNN_STATUS_EXECUTION_FAILED ``` If I set batch_size to 256*256 - 1 or do operation on a CPU everything works fine. Behavior is very similar to the torch.inverse bug #13276 , but with a different function and a different error message. ## Expected behavior The code should print `torch.Size([65536, 2, 2, 2])` ## Environment PyTorch version: 1.0.0 (checked with pytorch-nightly of 24.01.2019) Is debug build: No CUDA used to build PyTorch: 10.0.130 OS: Ubuntu 18.04.1 LTS GCC version: (Ubuntu 7.3.0-27ubuntu1~18.04) 7.3.0 CMake version: Could not collect Python version: 3.7 Is CUDA available: Yes CUDA runtime version: 10.0.130 GPU models and configuration: GPU 0: GeForce GTX 1080 Ti Nvidia driver version: 410.79 cuDNN version: 7401 (installed by pytorch itself) Versions of relevant libraries: [pip3] maskrcnn-benchmark (0.1, /media/aosokin/kingston2tb/software/pytorch/maskrcnn-benchmark) [pip3] numpy (1.14.0) [pip3] numpydoc (0.7.0) [pip3] torch (1.0.0a0+db5d313) [pip3] torchfile (0.1.0) [pip3] torchvision (0.2.1) [conda] blas 1.0 mkl [conda] cuda100 1.0 0 pytorch [conda] mkl 2019.1 144 [conda] mkl_fft 1.0.6 py37hd81dba3_0 [conda] mkl_random 1.0.2 py37hd81dba3_0 [conda] pytorch 1.0.0 py3.7_cuda10.0.130_cudnn7.4.1_1 [cuda100] pytorch [conda] torchvision 0.2.1 py_2 pytorch
1.0
torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256 - ## 🐛 Bug torch.nn.functional.affine_grid on GPU crashes when batch size >= 256*256 ## To Reproduce ``` import torch batch_size = 256*256 transform_parameters = torch.cuda.FloatTensor([[1,0,0],[0,1,0]]) transform_parameters = torch.stack([transform_parameters] * batch_size, 0).contiguous() resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2))) print(resampling_grids.size()) ``` crashes with the following error: ``` Traceback (most recent call last): File "bug.py", line 5, in <module> resampling_grids = torch.nn.functional.affine_grid(transform_parameters, torch.Size((batch_size, 1, 2, 2))) File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/functional.py", line 2615, in affine_grid return vision.affine_grid_generator(theta, size) File "/home/aosokin/local/software/anaconda3/envs/pytorch/lib/python3.7/site-packages/torch/nn/_functions/vision.py", line 10, in affine_grid_generator ret = torch.cudnn_affine_grid_generator(theta, N, C, H, W) RuntimeError: cuDNN error: CUDNN_STATUS_EXECUTION_FAILED ``` If I set batch_size to 256*256 - 1 or do operation on a CPU everything works fine. Behavior is very similar to the torch.inverse bug #13276 , but with a different function and a different error message. ## Expected behavior The code should print `torch.Size([65536, 2, 2, 2])` ## Environment PyTorch version: 1.0.0 (checked with pytorch-nightly of 24.01.2019) Is debug build: No CUDA used to build PyTorch: 10.0.130 OS: Ubuntu 18.04.1 LTS GCC version: (Ubuntu 7.3.0-27ubuntu1~18.04) 7.3.0 CMake version: Could not collect Python version: 3.7 Is CUDA available: Yes CUDA runtime version: 10.0.130 GPU models and configuration: GPU 0: GeForce GTX 1080 Ti Nvidia driver version: 410.79 cuDNN version: 7401 (installed by pytorch itself) Versions of relevant libraries: [pip3] maskrcnn-benchmark (0.1, /media/aosokin/kingston2tb/software/pytorch/maskrcnn-benchmark) [pip3] numpy (1.14.0) [pip3] numpydoc (0.7.0) [pip3] torch (1.0.0a0+db5d313) [pip3] torchfile (0.1.0) [pip3] torchvision (0.2.1) [conda] blas 1.0 mkl [conda] cuda100 1.0 0 pytorch [conda] mkl 2019.1 144 [conda] mkl_fft 1.0.6 py37hd81dba3_0 [conda] mkl_random 1.0.2 py37hd81dba3_0 [conda] pytorch 1.0.0 py3.7_cuda10.0.130_cudnn7.4.1_1 [cuda100] pytorch [conda] torchvision 0.2.1 py_2 pytorch
priority
torch nn functional affine grid on gpu crashes when batch size 🐛 bug torch nn functional affine grid on gpu crashes when batch size to reproduce import torch batch size transform parameters torch cuda floattensor transform parameters torch stack batch size contiguous resampling grids torch nn functional affine grid transform parameters torch size batch size print resampling grids size crashes with the following error traceback most recent call last file bug py line in resampling grids torch nn functional affine grid transform parameters torch size batch size file home aosokin local software envs pytorch lib site packages torch nn functional py line in affine grid return vision affine grid generator theta size file home aosokin local software envs pytorch lib site packages torch nn functions vision py line in affine grid generator ret torch cudnn affine grid generator theta n c h w runtimeerror cudnn error cudnn status execution failed if i set batch size to or do operation on a cpu everything works fine behavior is very similar to the torch inverse bug but with a different function and a different error message expected behavior the code should print torch size environment pytorch version checked with pytorch nightly of is debug build no cuda used to build pytorch os ubuntu lts gcc version ubuntu cmake version could not collect python version is cuda available yes cuda runtime version gpu models and configuration gpu geforce gtx ti nvidia driver version cudnn version installed by pytorch itself versions of relevant libraries maskrcnn benchmark media aosokin software pytorch maskrcnn benchmark numpy numpydoc torch torchfile torchvision blas mkl pytorch mkl mkl fft mkl random pytorch pytorch torchvision py pytorch
1
165,456
14,002,622,331
IssuesEvent
2020-10-28 14:59:12
tezos-reward-distributor-organization/tezos-reward-distributor
https://api.github.com/repos/tezos-reward-distributor-organization/tezos-reward-distributor
closed
Exception: Lock file present. Another process is running...
documentation
I keep getting when I run the script: ``` 2020-10-10 01:28:33,474 - MainThread - -------------------------------------------- Traceback (most recent call last): File "src/main.py", line 225, in <module> main(args) File "src/main.py", line 146, in main life_cycle.start(not dry_run) File "/Users/-/tezos-reward-distributor/src/util/process_life_cycle.py", line 20, in start self.lock_file.lock() File "/Users/-/tezos-reward-distributor/src/util/lock_file.py", line 10, in lock self.tryLock() File "/Users/-tezos-reward-distributor/src/util/lock_file.py", line 20, in tryLock raise Exception("Lock file present. Another process is running...") Exception: Lock file present. Another process is running... ``` Not sure how this happened, but I have to imagine the last script didn't end neatly. How do I go about reseting this? Can't seem to find a command. Thanks!
1.0
Exception: Lock file present. Another process is running... - I keep getting when I run the script: ``` 2020-10-10 01:28:33,474 - MainThread - -------------------------------------------- Traceback (most recent call last): File "src/main.py", line 225, in <module> main(args) File "src/main.py", line 146, in main life_cycle.start(not dry_run) File "/Users/-/tezos-reward-distributor/src/util/process_life_cycle.py", line 20, in start self.lock_file.lock() File "/Users/-/tezos-reward-distributor/src/util/lock_file.py", line 10, in lock self.tryLock() File "/Users/-tezos-reward-distributor/src/util/lock_file.py", line 20, in tryLock raise Exception("Lock file present. Another process is running...") Exception: Lock file present. Another process is running... ``` Not sure how this happened, but I have to imagine the last script didn't end neatly. How do I go about reseting this? Can't seem to find a command. Thanks!
non_priority
exception lock file present another process is running i keep getting when i run the script mainthread traceback most recent call last file src main py line in main args file src main py line in main life cycle start not dry run file users tezos reward distributor src util process life cycle py line in start self lock file lock file users tezos reward distributor src util lock file py line in lock self trylock file users tezos reward distributor src util lock file py line in trylock raise exception lock file present another process is running exception lock file present another process is running not sure how this happened but i have to imagine the last script didn t end neatly how do i go about reseting this can t seem to find a command thanks
0
85,248
15,736,660,755
IssuesEvent
2021-03-30 01:08:59
jgeraigery/k8-secrets-as-hyperledger-fabric-wallet
https://api.github.com/repos/jgeraigery/k8-secrets-as-hyperledger-fabric-wallet
opened
CVE-2020-13935 (High) detected in tomcat-embed-websocket-9.0.27.jar
security vulnerability
## CVE-2020-13935 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-websocket-9.0.27.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: k8-secrets-as-hyperledger-fabric-wallet/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-websocket/9.0.27/tomcat-embed-websocket-9.0.27.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.1.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.2.1.RELEASE.jar - :x: **tomcat-embed-websocket-9.0.27.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service. <p>Publish Date: 2020-07-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935>CVE-2020-13935</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2020-07-14</p> <p>Fix Resolution: org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-websocket","packageVersion":"9.0.27","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.2.1.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.2.1.RELEASE;org.apache.tomcat.embed:tomcat-embed-websocket:9.0.27","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13935","vulnerabilityDetails":"The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-13935 (High) detected in tomcat-embed-websocket-9.0.27.jar - ## CVE-2020-13935 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-websocket-9.0.27.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: k8-secrets-as-hyperledger-fabric-wallet/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-websocket/9.0.27/tomcat-embed-websocket-9.0.27.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.1.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.2.1.RELEASE.jar - :x: **tomcat-embed-websocket-9.0.27.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service. <p>Publish Date: 2020-07-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935>CVE-2020-13935</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/rd48c72bd3255bda87564d4da3791517c074d94f8a701f93b85752651%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2020-07-14</p> <p>Fix Resolution: org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-websocket","packageVersion":"9.0.27","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.2.1.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.2.1.RELEASE;org.apache.tomcat.embed:tomcat-embed-websocket:9.0.27","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat:tomcat-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7;org.apache.tomcat.embed:tomcat-embed-websocket:7.0.105,8.5.57,9.0.37,10.0.0-M7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13935","vulnerabilityDetails":"The payload length in a WebSocket frame was not correctly validated in Apache Tomcat 10.0.0-M1 to 10.0.0-M6, 9.0.0.M1 to 9.0.36, 8.5.0 to 8.5.56 and 7.0.27 to 7.0.104. Invalid payload lengths could trigger an infinite loop. Multiple requests with invalid payload lengths could lead to a denial of service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13935","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in tomcat embed websocket jar cve high severity vulnerability vulnerable library tomcat embed websocket jar core tomcat implementation path to dependency file secrets as hyperledger fabric wallet pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed websocket tomcat embed websocket jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed websocket jar vulnerable library found in base branch master vulnerability details the payload length in a websocket frame was not correctly validated in apache tomcat to to to and to invalid payload lengths could trigger an infinite loop multiple requests with invalid payload lengths could lead to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat websocket org apache tomcat embed tomcat embed websocket isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed websocket isminimumfixversionavailable true minimumfixversion org apache tomcat tomcat websocket org apache tomcat embed tomcat embed websocket basebranches vulnerabilityidentifier cve vulnerabilitydetails the payload length in a websocket frame was not correctly validated in apache tomcat to to to and to invalid payload lengths could trigger an infinite loop multiple requests with invalid payload lengths could lead to a denial of service vulnerabilityurl
0
75,444
3,462,478,282
IssuesEvent
2015-12-20 23:37:50
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Unable to echo unsigned ints at compile-time
Medium Priority Stdlib
```nim import macros macro foo(): stmt = var x = 8'u8 echo(x) foo() ``` ``` c:\users\dominik\appdata\local\temp\aporia\a14.nim(7, 3) Info: template/generic instantiation from here c:\users\dominik\git\nimrod\lib\system\repr.nim(25, 10) Error: cannot 'importc' variable at compile time ``` If anything, at least the error should be better.
1.0
Unable to echo unsigned ints at compile-time - ```nim import macros macro foo(): stmt = var x = 8'u8 echo(x) foo() ``` ``` c:\users\dominik\appdata\local\temp\aporia\a14.nim(7, 3) Info: template/generic instantiation from here c:\users\dominik\git\nimrod\lib\system\repr.nim(25, 10) Error: cannot 'importc' variable at compile time ``` If anything, at least the error should be better.
priority
unable to echo unsigned ints at compile time nim import macros macro foo stmt var x echo x foo c users dominik appdata local temp aporia nim info template generic instantiation from here c users dominik git nimrod lib system repr nim error cannot importc variable at compile time if anything at least the error should be better
1
49,565
7,522,142,282
IssuesEvent
2018-04-12 19:26:47
patternplate/patternplate
https://api.github.com/repos/patternplate/patternplate
closed
Terminal GIF out of sync with documentation
documentation help wanted
The GIF in the documentation here: https://patternplate.github.io/doc/docs/advanced/deploy?guides-enabled=true&advanced-enabled=true#build-a-static-interface Is out of sync with the written documentation. How was the GIF created? Happy to fix myself.
1.0
Terminal GIF out of sync with documentation - The GIF in the documentation here: https://patternplate.github.io/doc/docs/advanced/deploy?guides-enabled=true&advanced-enabled=true#build-a-static-interface Is out of sync with the written documentation. How was the GIF created? Happy to fix myself.
non_priority
terminal gif out of sync with documentation the gif in the documentation here is out of sync with the written documentation how was the gif created happy to fix myself
0
322,347
23,903,634,165
IssuesEvent
2022-09-08 21:22:01
neurobagel/concept_query
https://api.github.com/repos/neurobagel/concept_query
opened
Pushing Docker image to Docker hub
documentation deployment
Once the Dockerfile is in our repo and builds a local Docker image, let's - [] Push the image to Dockerhub - [] Write documentation for users of the repo to use `docker pull` the image instead of having to build it from the Dockerfile - [] In building this image, ensure that versions of base image, python, and required packages are locked in for consistent behavior
1.0
Pushing Docker image to Docker hub - Once the Dockerfile is in our repo and builds a local Docker image, let's - [] Push the image to Dockerhub - [] Write documentation for users of the repo to use `docker pull` the image instead of having to build it from the Dockerfile - [] In building this image, ensure that versions of base image, python, and required packages are locked in for consistent behavior
non_priority
pushing docker image to docker hub once the dockerfile is in our repo and builds a local docker image let s push the image to dockerhub write documentation for users of the repo to use docker pull the image instead of having to build it from the dockerfile in building this image ensure that versions of base image python and required packages are locked in for consistent behavior
0
384,426
11,392,342,154
IssuesEvent
2020-01-30 02:10:42
michaelsanford/bittray
https://api.github.com/repos/michaelsanford/bittray
closed
Fix rsrc syso inclusion
bug ci high-priority
It seems the existing way of doing that no longer works with go 1.13.6. Try with `rsrc` - https://github.com/akavel/rsrc - https://github.com/lxn/walk/issues/28 (good reference 👍 ) - https://www.programming-books.io/essential/go/a-first-window-app-e87631a7f8a94e8f92f9114e8e6baf0a
1.0
Fix rsrc syso inclusion - It seems the existing way of doing that no longer works with go 1.13.6. Try with `rsrc` - https://github.com/akavel/rsrc - https://github.com/lxn/walk/issues/28 (good reference 👍 ) - https://www.programming-books.io/essential/go/a-first-window-app-e87631a7f8a94e8f92f9114e8e6baf0a
priority
fix rsrc syso inclusion it seems the existing way of doing that no longer works with go try with rsrc good reference 👍
1
55,865
6,926,283,297
IssuesEvent
2017-11-30 18:37:54
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
Streak Bonus snack notification does not include Gold icon
POST-REDESIGN priority: medium section: Task Page status: issue: in progress
Bonus Streak snacks do not contain the Gold icon. The icon should be identical to the earned/lost snack sizing. ![image](https://user-images.githubusercontent.com/2686586/33393018-c4307b0c-d50b-11e7-810d-7d4d00e99e2d.png)
1.0
Streak Bonus snack notification does not include Gold icon - Bonus Streak snacks do not contain the Gold icon. The icon should be identical to the earned/lost snack sizing. ![image](https://user-images.githubusercontent.com/2686586/33393018-c4307b0c-d50b-11e7-810d-7d4d00e99e2d.png)
non_priority
streak bonus snack notification does not include gold icon bonus streak snacks do not contain the gold icon the icon should be identical to the earned lost snack sizing
0
361,949
10,721,605,323
IssuesEvent
2019-10-27 04:21:14
AY1920S1-CS2103T-W11-2/main
https://api.github.com/repos/AY1920S1-CS2103T-W11-2/main
closed
Amend List command to display either a list of activities or contacts
priority.High type.Task
Amend the existing List command to allow a user to select between displaying the list of contacts and list of activities.
1.0
Amend List command to display either a list of activities or contacts - Amend the existing List command to allow a user to select between displaying the list of contacts and list of activities.
priority
amend list command to display either a list of activities or contacts amend the existing list command to allow a user to select between displaying the list of contacts and list of activities
1
809,705
30,206,385,112
IssuesEvent
2023-07-05 09:42:52
informalsystems/quint
https://api.github.com/repos/informalsystems/quint
closed
Topological sort of definitions in phase 2
W8 Fresolver (phase 2) impact-medium effort-hard product-priority-general
The language manual lets the user to write definitions out of order. For instance: ```scala module OutOfOrder { val n1 = e1 val n2 = n3 + 1 val n3 = n1 * 2 } ``` Currently, the parser (phase 1) and the name resolver (phase 2) keep the order of definitions. Moreover, the name resolver does not check for cyclic definitions. As a result, when writing a later phase (e.g., the simulator) with `IRVisitor`, I have to think about the out-of-order definitions. It looks like we have enough information to make a topological sort of definitions in phase 2. If we do so, using `IRVisitor` in later phases would be simpler, as we can assume that all names are defined before they are used (in the traversal order of `IRVisitor`). Apalache is doing a topological sort of definitions already, perhaps not optimally, so we can get some inspiration from [DeclarationSorter.scala](https://github.com/informalsystems/apalache/blob/unstable/tlair/src/main/scala/at/forsyte/apalache/tla/lir/transformations/standard/DeclarationSorter.scala). Once we have topological sort, we can detect cyclic references in phase 2.
1.0
Topological sort of definitions in phase 2 - The language manual lets the user to write definitions out of order. For instance: ```scala module OutOfOrder { val n1 = e1 val n2 = n3 + 1 val n3 = n1 * 2 } ``` Currently, the parser (phase 1) and the name resolver (phase 2) keep the order of definitions. Moreover, the name resolver does not check for cyclic definitions. As a result, when writing a later phase (e.g., the simulator) with `IRVisitor`, I have to think about the out-of-order definitions. It looks like we have enough information to make a topological sort of definitions in phase 2. If we do so, using `IRVisitor` in later phases would be simpler, as we can assume that all names are defined before they are used (in the traversal order of `IRVisitor`). Apalache is doing a topological sort of definitions already, perhaps not optimally, so we can get some inspiration from [DeclarationSorter.scala](https://github.com/informalsystems/apalache/blob/unstable/tlair/src/main/scala/at/forsyte/apalache/tla/lir/transformations/standard/DeclarationSorter.scala). Once we have topological sort, we can detect cyclic references in phase 2.
priority
topological sort of definitions in phase the language manual lets the user to write definitions out of order for instance scala module outoforder val val val currently the parser phase and the name resolver phase keep the order of definitions moreover the name resolver does not check for cyclic definitions as a result when writing a later phase e g the simulator with irvisitor i have to think about the out of order definitions it looks like we have enough information to make a topological sort of definitions in phase if we do so using irvisitor in later phases would be simpler as we can assume that all names are defined before they are used in the traversal order of irvisitor apalache is doing a topological sort of definitions already perhaps not optimally so we can get some inspiration from once we have topological sort we can detect cyclic references in phase
1
103,703
12,963,534,326
IssuesEvent
2020-07-20 18:56:59
microsoft/WSL
https://api.github.com/repos/microsoft/WSL
closed
generateResolvConf = false did not work in wsl.conf
bydesign
<!-- 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 I ACKNOWLEDGE THE FOLLOWING BEFORE PROCEEDING: 1. If I delete this entire template and go my own path, the core team may close my issue without further explanation or engagement. 2. If I list multiple bugs/concerns in this one issue, the core team may close my issue without further explanation or engagement. 3. If I write an issue that has many duplicates, the core team may close my issue without further explanation or engagement (and without necessarily spending time to find the exact duplicate ID number). 4. If I leave the title incomplete when filing the issue, the core team may close my issue without further explanation or engagement. 5. If I file something completely blank in the body, the core team may close my issue without further explanation or engagement. All good? Then proceed! --> I have WSL2 with debian -all good. I created a /etc/wsl.conf file it contains ``` [network] generateResolvConf = false ``` as per https://docs.microsoft.com/en-us/windows/wsl/wsl-config#network if i delete /etc/resolv.conf (or edit and save) on start of the next debia WSL2 session it has been regenerated again. running WSL2 in windows 10 build 20170.rs_prerelease.200710-1419 <!-- This bug tracker is monitored by Windows Subsystem for Linux development team and other technical folks. Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues. Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. Ideally, please configure your machine to capture minidumps, repro the issue, and send the minidump from "C:\Windows\minidump\". You can find instructions to do that here: https://support.microsoft.com/en-us/help/315263/how-to-read-the-small-memory-dump-file-that-is-created-by-windows-if-a If this is a console issue (a problem with layout, rendering, colors, etc.), please post the issue to the Terminal tracker: https://github.com/microsoft/terminal/issues For documentation improvements, please post to the documentation tracker: https://github.com/MicrosoftDocs/WSL/issues For any other questions on contributing please see our contribution guidelines: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md Please fill out the items below. --> # Environment ```none Windows build number: Microsoft Windows [Version 10.0.20170.1000] Your Distribution version: PRETTY_NAME="Debian GNU/Linux 10 (buster)" NAME="Debian GNU/Linux" VERSION_ID="10" VERSION="10 (buster)" VERSION_CODENAME=buster ID=debian WSL 2 ``` # Steps to reproduce <!-- What you're doing and what's happening. Copy&paste the full set of specific command-line steps necessary to reproduce the behavior, and their output. Include screenshots if that helps demonstrate the problem. --> <!-- If you'd like to provide logs you can provide an `strace(1)` log of the failing command (if `some_command` is failing, then run `strace -o some_command.strace -f some_command some_args`, and link the contents of `some_command.strace` in a gist. More info on `strace` can be found here: https://www.man7.org/linux/man-pages/man1/strace.1.html You can use Github gists to share the output: https://gist.github.com/ Additionally, For WSL launch issues, please collect detailed logs, instructions here: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md#8-detailed-logs --> # Expected behavior <!-- A description of what you're expecting, possibly containing screenshots or reference material. --> # Actual behavior <!-- What's actually happening? -->
1.0
generateResolvConf = false did not work in wsl.conf - <!-- 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 I ACKNOWLEDGE THE FOLLOWING BEFORE PROCEEDING: 1. If I delete this entire template and go my own path, the core team may close my issue without further explanation or engagement. 2. If I list multiple bugs/concerns in this one issue, the core team may close my issue without further explanation or engagement. 3. If I write an issue that has many duplicates, the core team may close my issue without further explanation or engagement (and without necessarily spending time to find the exact duplicate ID number). 4. If I leave the title incomplete when filing the issue, the core team may close my issue without further explanation or engagement. 5. If I file something completely blank in the body, the core team may close my issue without further explanation or engagement. All good? Then proceed! --> I have WSL2 with debian -all good. I created a /etc/wsl.conf file it contains ``` [network] generateResolvConf = false ``` as per https://docs.microsoft.com/en-us/windows/wsl/wsl-config#network if i delete /etc/resolv.conf (or edit and save) on start of the next debia WSL2 session it has been regenerated again. running WSL2 in windows 10 build 20170.rs_prerelease.200710-1419 <!-- This bug tracker is monitored by Windows Subsystem for Linux development team and other technical folks. Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues. Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. Ideally, please configure your machine to capture minidumps, repro the issue, and send the minidump from "C:\Windows\minidump\". You can find instructions to do that here: https://support.microsoft.com/en-us/help/315263/how-to-read-the-small-memory-dump-file-that-is-created-by-windows-if-a If this is a console issue (a problem with layout, rendering, colors, etc.), please post the issue to the Terminal tracker: https://github.com/microsoft/terminal/issues For documentation improvements, please post to the documentation tracker: https://github.com/MicrosoftDocs/WSL/issues For any other questions on contributing please see our contribution guidelines: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md Please fill out the items below. --> # Environment ```none Windows build number: Microsoft Windows [Version 10.0.20170.1000] Your Distribution version: PRETTY_NAME="Debian GNU/Linux 10 (buster)" NAME="Debian GNU/Linux" VERSION_ID="10" VERSION="10 (buster)" VERSION_CODENAME=buster ID=debian WSL 2 ``` # Steps to reproduce <!-- What you're doing and what's happening. Copy&paste the full set of specific command-line steps necessary to reproduce the behavior, and their output. Include screenshots if that helps demonstrate the problem. --> <!-- If you'd like to provide logs you can provide an `strace(1)` log of the failing command (if `some_command` is failing, then run `strace -o some_command.strace -f some_command some_args`, and link the contents of `some_command.strace` in a gist. More info on `strace` can be found here: https://www.man7.org/linux/man-pages/man1/strace.1.html You can use Github gists to share the output: https://gist.github.com/ Additionally, For WSL launch issues, please collect detailed logs, instructions here: https://github.com/Microsoft/WSL/blob/master/CONTRIBUTING.md#8-detailed-logs --> # Expected behavior <!-- A description of what you're expecting, possibly containing screenshots or reference material. --> # Actual behavior <!-- What's actually happening? -->
non_priority
generateresolvconf false did not work in wsl conf 🚨🚨🚨🚨🚨🚨🚨🚨🚨🚨 i acknowledge the following before proceeding if i delete this entire template and go my own path the core team may close my issue without further explanation or engagement if i list multiple bugs concerns in this one issue the core team may close my issue without further explanation or engagement if i write an issue that has many duplicates the core team may close my issue without further explanation or engagement and without necessarily spending time to find the exact duplicate id number if i leave the title incomplete when filing the issue the core team may close my issue without further explanation or engagement if i file something completely blank in the body the core team may close my issue without further explanation or engagement all good then proceed i have with debian all good i created a etc wsl conf file it contains generateresolvconf false as per if i delete etc resolv conf or edit and save on start of the next debia session it has been regenerated again running in windows build rs prerelease this bug tracker is monitored by windows subsystem for linux development team and other technical folks important when reporting bsods or security issues do not attach memory dumps logs or traces to github issues instead send dumps traces to secure microsoft com referencing this github issue ideally please configure your machine to capture minidumps repro the issue and send the minidump from c windows minidump you can find instructions to do that here if this is a console issue a problem with layout rendering colors etc please post the issue to the terminal tracker for documentation improvements please post to the documentation tracker for any other questions on contributing please see our contribution guidelines please fill out the items below environment none windows build number microsoft windows your distribution version pretty name debian gnu linux buster name debian gnu linux version id version buster version codename buster id debian wsl steps to reproduce if you d like to provide logs you can provide an strace log of the failing command if some command is failing then run strace o some command strace f some command some args and link the contents of some command strace in a gist more info on strace can be found here you can use github gists to share the output additionally for wsl launch issues please collect detailed logs instructions here expected behavior actual behavior
0
790,635
27,830,950,291
IssuesEvent
2023-03-20 04:49:49
AY2223S2-CS2103T-W14-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-W14-1/tp
closed
Refactor UI package to be more extensible
priority.med type.refactor severity.Medium
UI package is a messy platter of java files with little to no organisation.
1.0
Refactor UI package to be more extensible - UI package is a messy platter of java files with little to no organisation.
priority
refactor ui package to be more extensible ui package is a messy platter of java files with little to no organisation
1
75,802
3,476,131,524
IssuesEvent
2015-12-26 14:23:50
Stephane-D/SGDK
https://api.github.com/repos/Stephane-D/SGDK
closed
recomp sprite collision, wrong label name
Priority-Medium
Labels name are generated multiple time: see https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1247 https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1259 using this .res file: https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-res
1.0
recomp sprite collision, wrong label name - Labels name are generated multiple time: see https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1247 https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-s-L1259 using this .res file: https://gist.github.com/Ced2911/78f848ebb4bb1dba6a47#file-gfx-res
priority
recomp sprite collision wrong label name labels name are generated multiple time see using this res file
1
351,054
25,011,721,862
IssuesEvent
2022-11-03 15:41:27
deforum-art/deforum-for-automatic1111-webui
https://api.github.com/repos/deforum-art/deforum-for-automatic1111-webui
closed
Add tooltips
documentation enhancement good first issue help wanted
Tooltips would definitely help the newcomers, but the list is pretty large, so if you have some time to spare, I'd welcome your contributions. The list of tooltips is a simple JSON file in a form `"label":"tooltip"`. The names of the widgets lie in https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum.py and https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum/args.py. Examples from the main webui ``` titles = { "Sampling steps": "How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results", "Sampling method": "Which algorithm to use to produce the image", "GFPGAN": "Restore low quality faces using GFPGAN neural network", "Euler a": "Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help", "DDIM": "Denoising Diffusion Implicit Models - best at inpainting", } ```
1.0
Add tooltips - Tooltips would definitely help the newcomers, but the list is pretty large, so if you have some time to spare, I'd welcome your contributions. The list of tooltips is a simple JSON file in a form `"label":"tooltip"`. The names of the widgets lie in https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum.py and https://github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/scripts/deforum/args.py. Examples from the main webui ``` titles = { "Sampling steps": "How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results", "Sampling method": "Which algorithm to use to produce the image", "GFPGAN": "Restore low quality faces using GFPGAN neural network", "Euler a": "Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help", "DDIM": "Denoising Diffusion Implicit Models - best at inpainting", } ```
non_priority
add tooltips tooltips would definitely help the newcomers but the list is pretty large so if you have some time to spare i d welcome your contributions the list of tooltips is a simple json file in a form label tooltip the names of the widgets lie in and examples from the main webui titles sampling steps how many times to improve the generated image iteratively higher values take longer very low values can produce bad results sampling method which algorithm to use to produce the image gfpgan restore low quality faces using gfpgan neural network euler a euler ancestral very creative each can get a completely different picture depending on step count setting steps to higher than does not help ddim denoising diffusion implicit models best at inpainting
0
355,778
10,584,634,504
IssuesEvent
2019-10-08 15:48:53
angular/angular-cli
https://api.github.com/repos/angular/angular-cli
closed
Unable to spyOn Math.floor
priority: 2 (required) type: bug/fix
### Versions ``` @angular/cli: 1.4.9 node: 6.9.3 os: win32 x64 @angular/animations: 4.4.6 @angular/cdk: 2.0.0-beta.12 @angular/common: 4.4.6 @angular/compiler: 4.4.6 @angular/core: 4.4.6 @angular/forms: 4.4.6 @angular/flex-layout: 2.0.0-beta.10-4905443 @angular/http: 4.4.6 @angular/material: 2.0.0-beta.12 @angular/platform-browser: 4.4.6 @angular/platform-browser-dynamic: 4.4.6 @angular/router: 4.4.6 @angular/cli: 1.4.9 @angular/compiler-cli: 4.4.6 @angular/language-service: 4.4.6 typescript: 2.5.3 ``` ### Repro steps Create a simple test and attempt to do `spyOn(Math, 'floor').and.returnValue(10); Run ng test ### Observed behavior ``` RangeError: Maximum call stack size exceeded at c (node_modules/source-map-support/browser-source-map-support.js:53:483) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) ``` ### Desired behavior I should be able to spyOn Math.floor similar to other Math functions ### Mention any other details that might be useful (optional) Spying on Math.ceil works as expected and running without source maps `ng test --sm=false` also works.
1.0
Unable to spyOn Math.floor - ### Versions ``` @angular/cli: 1.4.9 node: 6.9.3 os: win32 x64 @angular/animations: 4.4.6 @angular/cdk: 2.0.0-beta.12 @angular/common: 4.4.6 @angular/compiler: 4.4.6 @angular/core: 4.4.6 @angular/forms: 4.4.6 @angular/flex-layout: 2.0.0-beta.10-4905443 @angular/http: 4.4.6 @angular/material: 2.0.0-beta.12 @angular/platform-browser: 4.4.6 @angular/platform-browser-dynamic: 4.4.6 @angular/router: 4.4.6 @angular/cli: 1.4.9 @angular/compiler-cli: 4.4.6 @angular/language-service: 4.4.6 typescript: 2.5.3 ``` ### Repro steps Create a simple test and attempt to do `spyOn(Math, 'floor').and.returnValue(10); Run ng test ### Observed behavior ``` RangeError: Maximum call stack size exceeded at c (node_modules/source-map-support/browser-source-map-support.js:53:483) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) at c (node_modules/source-map-support/browser-source-map-support.js:54:101) ``` ### Desired behavior I should be able to spyOn Math.floor similar to other Math functions ### Mention any other details that might be useful (optional) Spying on Math.ceil works as expected and running without source maps `ng test --sm=false` also works.
priority
unable to spyon math floor versions angular cli node os angular animations angular cdk beta angular common angular compiler angular core angular forms angular flex layout beta angular http angular material beta angular platform browser angular platform browser dynamic angular router angular cli angular compiler cli angular language service typescript repro steps create a simple test and attempt to do spyon math floor and returnvalue run ng test observed behavior rangeerror maximum call stack size exceeded at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js at c node modules source map support browser source map support js desired behavior i should be able to spyon math floor similar to other math functions mention any other details that might be useful optional spying on math ceil works as expected and running without source maps ng test sm false also works
1
2,093
4,381,721,899
IssuesEvent
2016-08-06 12:14:02
support-project/knowledge
https://api.github.com/repos/support-project/knowledge
closed
ファイルサーバへのリンク
[Status] 3.merged [Type] 1.requirement
Markdown本文の中で、ファイルサーバーへのリンクを記載したい。 以下のようなネットワークディレクトリの表記を行うと、Markdownのパース処理か XSS対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという 報告があった。 \\server_host\hoge\fuga\sample.txt file://server_host/hoge/fuga/sample.txt
1.0
ファイルサーバへのリンク - Markdown本文の中で、ファイルサーバーへのリンクを記載したい。 以下のようなネットワークディレクトリの表記を行うと、Markdownのパース処理か XSS対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという 報告があった。 \\server_host\hoge\fuga\sample.txt file://server_host/hoge/fuga/sample.txt
non_priority
ファイルサーバへのリンク markdown本文の中で、ファイルサーバーへのリンクを記載したい。 以下のようなネットワークディレクトリの表記を行うと、markdownのパース処理か xss対策処理の中で、パスを壊してしまうようで、リンクが変更されてしまうという 報告があった。 server host hoge fuga sample txt file server host hoge fuga sample txt
0
777,049
27,266,788,975
IssuesEvent
2023-02-22 18:41:57
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
MSRuleCleaner potentially not cleaning unlocked input dataset rules
BUG High Priority MSRuleCleaner
**Impact of the bug** MSRuleCleaner **Describe the bug** A possible issue with MSRuleCleaner has been reported in the past week, where it leaves Rucio rules behind for some of the input datasets. Amanda initially provided these 3 RelVal input datasets (blocks): ``` /RelValZMM_14/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#621aab39-d57d-4308-b38c-fa8bfeb65fdb /RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun4_realistic_v8_2026D88noPU-v1/GEN-SIM-DIGI-RAW#fef08bee-3d76-485d-b4b4-68ed0b181b70 /RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#4c88cda3-eebb-4e7f-9860-ef201108d9d8 ``` which are still locked under the account `wmcore_transferor`, even though all the relevant workflows processing it are no longer active in the system (they have been archived). A non-relval input dataset example is: https://cmsweb.cern.ch/reqmgr2/data/request?inputdataset=/BTohh_hToMuNu_BsBdMixture_modHadLifetime_TuneCP5_13TeV-pythia8-evtgen/RunIISummer20UL17RECO-106X_mc2017_realistic_v6-v3/AODSIM which has been archived in December 2021, leaving these two rucio rules behind: ``` c55dcaa65f2b443b9824fff1983228e1 61fa02b8b4694f27b80f91adec4d1591 ``` **How to reproduce it** No clue **Expected behavior** From a quick debugging that I have done so far, MSRuleCleaner seem to be properly deleting rules for the input dataset. So there might be a specific scenario that triggers this ill behaviour. On what concerns this issue, we should: * improve logging, such that we can explicitly see whenever an input data is looked up, and under which classification (primary, parent or secondary); * ensure that rules allowed for deletion are actually removed, especially for input dataset **Additional context and error message** None
1.0
MSRuleCleaner potentially not cleaning unlocked input dataset rules - **Impact of the bug** MSRuleCleaner **Describe the bug** A possible issue with MSRuleCleaner has been reported in the past week, where it leaves Rucio rules behind for some of the input datasets. Amanda initially provided these 3 RelVal input datasets (blocks): ``` /RelValZMM_14/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#621aab39-d57d-4308-b38c-fa8bfeb65fdb /RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun4_realistic_v8_2026D88noPU-v1/GEN-SIM-DIGI-RAW#fef08bee-3d76-485d-b4b4-68ed0b181b70 /RelValTTbar_14TeV/CMSSW_12_5_0_pre3-124X_mcRun3_2022_realistic_v8-v2/GEN-SIM-DIGI-RAW#4c88cda3-eebb-4e7f-9860-ef201108d9d8 ``` which are still locked under the account `wmcore_transferor`, even though all the relevant workflows processing it are no longer active in the system (they have been archived). A non-relval input dataset example is: https://cmsweb.cern.ch/reqmgr2/data/request?inputdataset=/BTohh_hToMuNu_BsBdMixture_modHadLifetime_TuneCP5_13TeV-pythia8-evtgen/RunIISummer20UL17RECO-106X_mc2017_realistic_v6-v3/AODSIM which has been archived in December 2021, leaving these two rucio rules behind: ``` c55dcaa65f2b443b9824fff1983228e1 61fa02b8b4694f27b80f91adec4d1591 ``` **How to reproduce it** No clue **Expected behavior** From a quick debugging that I have done so far, MSRuleCleaner seem to be properly deleting rules for the input dataset. So there might be a specific scenario that triggers this ill behaviour. On what concerns this issue, we should: * improve logging, such that we can explicitly see whenever an input data is looked up, and under which classification (primary, parent or secondary); * ensure that rules allowed for deletion are actually removed, especially for input dataset **Additional context and error message** None
priority
msrulecleaner potentially not cleaning unlocked input dataset rules impact of the bug msrulecleaner describe the bug a possible issue with msrulecleaner has been reported in the past week where it leaves rucio rules behind for some of the input datasets amanda initially provided these relval input datasets blocks relvalzmm cmssw realistic gen sim digi raw relvalttbar cmssw realistic gen sim digi raw relvalttbar cmssw realistic gen sim digi raw eebb which are still locked under the account wmcore transferor even though all the relevant workflows processing it are no longer active in the system they have been archived a non relval input dataset example is which has been archived in december leaving these two rucio rules behind how to reproduce it no clue expected behavior from a quick debugging that i have done so far msrulecleaner seem to be properly deleting rules for the input dataset so there might be a specific scenario that triggers this ill behaviour on what concerns this issue we should improve logging such that we can explicitly see whenever an input data is looked up and under which classification primary parent or secondary ensure that rules allowed for deletion are actually removed especially for input dataset additional context and error message none
1
47,977
25,288,583,756
IssuesEvent
2022-11-16 21:36:37
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
optimize_move_to_prewhere does not work with Engine=merge -> distributed -> mergeTree
performance
```sql drop table if exists foo_merge; drop table if exists foo_merge_distributed; drop table if exists foo_distributed; drop table if exists foo_distributed_merge; drop table if exists foo; create table foo(A Int64, V Int64, S String) Engine=MergeTree order by A; insert into foo select number, sipHash64(number), toString(arrayMap(x->sipHash64(number*x), range(10))) from numbers(10000000); -- merge -> distributed -> mergeTree create table foo_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo); create table foo_distributed_merge as foo ENGINE=Merge(currentDatabase(), '^foo_distributed$'); -- No optimize_move_to_prewhere :( select * from foo_distributed_merge where V = 666; 0 rows in set. Elapsed: 0.530 sec. Processed 10.00 million rows, 2.31 GB (18.85 million rows/s., 4.35 GB/s.) -- manual prewhere select * from foo_distributed_merge prewhere V = 666; 0 rows in set. Elapsed: 0.016 sec. Processed 10.00 million rows, 80.00 MB (617.71 million rows/s., 4.94 GB/s.) -- distributed -> merge -> mergeTree create table foo_merge as foo ENGINE=Merge(currentDatabase(), '^foo$'); create table foo_merge_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo_merge); -- optimize_move_to_prewhere :) select * from foo_merge_distributed where V = 666; 0 rows in set. Elapsed: 0.024 sec. Processed 10.00 million rows, 80.00 MB (411.88 million rows/s., 3.30 GB/s.) ```
True
optimize_move_to_prewhere does not work with Engine=merge -> distributed -> mergeTree - ```sql drop table if exists foo_merge; drop table if exists foo_merge_distributed; drop table if exists foo_distributed; drop table if exists foo_distributed_merge; drop table if exists foo; create table foo(A Int64, V Int64, S String) Engine=MergeTree order by A; insert into foo select number, sipHash64(number), toString(arrayMap(x->sipHash64(number*x), range(10))) from numbers(10000000); -- merge -> distributed -> mergeTree create table foo_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo); create table foo_distributed_merge as foo ENGINE=Merge(currentDatabase(), '^foo_distributed$'); -- No optimize_move_to_prewhere :( select * from foo_distributed_merge where V = 666; 0 rows in set. Elapsed: 0.530 sec. Processed 10.00 million rows, 2.31 GB (18.85 million rows/s., 4.35 GB/s.) -- manual prewhere select * from foo_distributed_merge prewhere V = 666; 0 rows in set. Elapsed: 0.016 sec. Processed 10.00 million rows, 80.00 MB (617.71 million rows/s., 4.94 GB/s.) -- distributed -> merge -> mergeTree create table foo_merge as foo ENGINE=Merge(currentDatabase(), '^foo$'); create table foo_merge_distributed as foo Engine = Distributed('test_shard_localhost', currentDatabase(), foo_merge); -- optimize_move_to_prewhere :) select * from foo_merge_distributed where V = 666; 0 rows in set. Elapsed: 0.024 sec. Processed 10.00 million rows, 80.00 MB (411.88 million rows/s., 3.30 GB/s.) ```
non_priority
optimize move to prewhere does not work with engine merge distributed mergetree sql drop table if exists foo merge drop table if exists foo merge distributed drop table if exists foo distributed drop table if exists foo distributed merge drop table if exists foo create table foo a v s string engine mergetree order by a insert into foo select number number tostring arraymap x number x range from numbers merge distributed mergetree create table foo distributed as foo engine distributed test shard localhost currentdatabase foo create table foo distributed merge as foo engine merge currentdatabase foo distributed no optimize move to prewhere select from foo distributed merge where v rows in set elapsed sec processed million rows gb million rows s gb s manual prewhere select from foo distributed merge prewhere v rows in set elapsed sec processed million rows mb million rows s gb s distributed merge mergetree create table foo merge as foo engine merge currentdatabase foo create table foo merge distributed as foo engine distributed test shard localhost currentdatabase foo merge optimize move to prewhere select from foo merge distributed where v rows in set elapsed sec processed million rows mb million rows s gb s
0
257,649
8,140,033,770
IssuesEvent
2018-08-20 19:42:26
nprapps/elections18-graphics
https://api.github.com/repos/nprapps/elections18-graphics
closed
Run an informational member-station webinar
effort:medium priority:high
We'll be working with Janeen and Franklin to coordinate a webinar with member station staff. At this webinar, we'll present our existing design work, and show how it could be embedded on their site as a standalone page or widget. We'll also solicit feedback from the member stations, and incorporate these into the issue matrix (priority vs effort). cc @constancebmiller
1.0
Run an informational member-station webinar - We'll be working with Janeen and Franklin to coordinate a webinar with member station staff. At this webinar, we'll present our existing design work, and show how it could be embedded on their site as a standalone page or widget. We'll also solicit feedback from the member stations, and incorporate these into the issue matrix (priority vs effort). cc @constancebmiller
priority
run an informational member station webinar we ll be working with janeen and franklin to coordinate a webinar with member station staff at this webinar we ll present our existing design work and show how it could be embedded on their site as a standalone page or widget we ll also solicit feedback from the member stations and incorporate these into the issue matrix priority vs effort cc constancebmiller
1
502,887
14,569,265,032
IssuesEvent
2020-12-17 12:51:02
staxrip/staxrip
https://api.github.com/repos/staxrip/staxrip
closed
AOMENC Better defaults with explanations.
feature request fixed priority medium
Change: Basic > Mode: Change from One-pass to Two-pass  Reason(s):  1. first pass is usually always very fast  2. Aomenc does not use alt ref frames when performing a single pass, therefore it is generally recommended to always perform a 2-pass encode, no matter what Rate Mode is chosen.  ------------  Change: Basic > Rate Mode:  1. Rename "Q" to "Q (Constant Quality)" 2. Rename "CQ" to "CQ (Constrained Quality)"  Reason(s):  1. This will help eliminate confusion because users coming from x265 might choose "CQ", expecting it to behave similar to x265's CRF (CQ), but it will result in a terrible quality encode due to it producing a constrained quality video.  ------------ Change: Basic > Rate Mode: Change from VBR to Q  Reason(s):  1. Aomenc is less likely to be used for targeting precise filesizes, and more likely to be used for targeting quality at imprecise, yet small filesizes.  ------------ Change: Basic > Bit Depth: Change from 8 to 10  Reason(s):  1. 10bit is basically the new norm as, there are almost no new devices/software that do not support 10bit decoding  2. 10bit almost always produces a better quality encode with smoother gradients in skies.  3. It's safer to do 10bit final encode than 8bit, even if converting an 8bit source video into a 10bit result. There will almost always be fewer anomalies going with a 10bit final encode. ------------ Change: Basic > CPU Used: Change from 8 to 4 (Medium)  Reason(s): 1. 8 Will produce a very low quality encode with poor VMAF scores in a lot of cases.  2. 4 is a good balance in terms of speed and quality and produces good VMAF scores.  3. Charts show that cpu speeds 0-4 all have very close VMAF scores, but cpu speeds 5-8 begin producing much worse VMAF scores, therefore CPU 4 should be the default (in my opinion) ------------  Change: Basic > CQ Level: Change from 0 to 25  Reason(s):  1. 0 is lossless, and lossless should never be the default for a lossy encoder. This shouldn't require a huge explanation, as the reasons should be obvious.  2. 25 is a good default quality level since it produces a high quality encode at a small filesize. AV1 excels at higher CQ levels compared to x265, and AV1 is primarily used to create low filesize encodes at a target quality. AV1 isn't well optimized for low CQ levels due to quantizer related denoising anomalies (especially on grainy content). The goal of AV1 is to denoise / remove granular details without becoming a blocky, swimming mess, and it is better able to achieve that goal at higher CQ levels, until some kinks have been ironed out. Believe me, I would LOVE to use a lower CQ level as a default, but CQ 30 just, across the board, produces more consistent results. 3. CQ Level 25 provides good quality for both 8bit and native 10bit content. According to user BlueSwordM, native 10bit content benefits more from a CQ level of 25 compared with higher CQ levels. 25 is going to be a safe starting point for content of either bit depth.  ------------ Change: Analysis > Tile Columns: Change from 0 to 2 Analysis > Tile Rows: Change from 0 to 1  Reason(s):  1. Assists with multi-threaded encoding by better utilizing cores on your cpu, with a downside of a small hit to encoding compression. However, many AV1 users agree that the speed gained is worth the tradeoff.  3. Reddit user said this:   "By keeping tiles as "square" as possible, you have the lowest edge/area ratio. Edges cause a small loss in coding efficiency. So for a 16:9 video, 2x1, 2x2 or 4x2 is recommended depending on resolution. Most times 2x1 is good for 1080p, 2x2 for 1440p and 4x2 for 2160p and higher."   4. Since most content being encoded is still 1080p, then 2x1 is probably a safe default.  ------------ Change: Slice Decision > Min GOP Size: Rename to Keyframe Min Distance Slice Decision > Max GOP Size: Rename to Keyframe Max Distance Reason(s): 1. This will better reflect the parameter name since AV1 refers to this as keyframe distance, rather than "Group of Picture" size  2. Easier to understand without inexperienced users having to resort to google to figure out what GOP means. ------------ Change: Slice Decision > Max GOP Size (i.e. Keyframe Max Distance): Change from 0 to 240  Reason(s):  1. I heard AV1 users at one point discussing this number being recommended by the AV1 developers, and also I believe this is a recommended number even for VP9.  2. It represents approximately maximum of 10 seconds between keyframes (if your framerate is approximately 24fps), otherwise keyframes could be placed at "who knows". Better to specify than to leave it unspecified, since the AV1 parameter doesn't say what the default value is. ------------  Change: Slice Decision > Lag in Frames: Change from 0 to 25  Reason(s): 1. --lag-in-frames parameter defines an upper limit on the number of frames into the future that the encoder can look 2. I have seen people preferring 25 for AV1.  3. For VP9, a value of 16 was recommended. I have no idea what the recommended or the max value is in AV1, but it seems safe to default this to 25. ------------  Change: Performance > Add a checkbox called "Automatically determine thread count". Enabling this would gray out/disallow user input for "Threads", and the Thread count would be automatically determined based on the CPU's multi-threading capability.  Reason(s):  1. 0 doesn't actually cause aomenc to auto multi thread. 0 disables multi-threading.  2. Sloppy Workaround: You could try setting default thread count to 8 since that's a safe number in 2020, since most people doing video conversion in 2020 have at least a quad core with hyper threading, however it would be best if StaxRip could automatically determine the cpu's maximum thread capability and automatically set this number accordingly. ------------ Change: Misc 2 > Auto Alt Ref: Change from 0 to 1  Reason(s):  1. It is better to have "auto alt ref" Enabled because according to some AV1 users in discord, having it disabled will result in a "completely shit encode"  2. If you're doing a single pass encode, auto-alt-ref is actually disabled by default, but you can force it to be enabled by forcibly specifying it.  3. alt-ref frames are used when using a 2-pass encode by default if it is not specified. Based on my limited understanding, you are generally better off just forcibly enabling this setting for most case scenarios.
1.0
AOMENC Better defaults with explanations. - Change: Basic > Mode: Change from One-pass to Two-pass  Reason(s):  1. first pass is usually always very fast  2. Aomenc does not use alt ref frames when performing a single pass, therefore it is generally recommended to always perform a 2-pass encode, no matter what Rate Mode is chosen.  ------------  Change: Basic > Rate Mode:  1. Rename "Q" to "Q (Constant Quality)" 2. Rename "CQ" to "CQ (Constrained Quality)"  Reason(s):  1. This will help eliminate confusion because users coming from x265 might choose "CQ", expecting it to behave similar to x265's CRF (CQ), but it will result in a terrible quality encode due to it producing a constrained quality video.  ------------ Change: Basic > Rate Mode: Change from VBR to Q  Reason(s):  1. Aomenc is less likely to be used for targeting precise filesizes, and more likely to be used for targeting quality at imprecise, yet small filesizes.  ------------ Change: Basic > Bit Depth: Change from 8 to 10  Reason(s):  1. 10bit is basically the new norm as, there are almost no new devices/software that do not support 10bit decoding  2. 10bit almost always produces a better quality encode with smoother gradients in skies.  3. It's safer to do 10bit final encode than 8bit, even if converting an 8bit source video into a 10bit result. There will almost always be fewer anomalies going with a 10bit final encode. ------------ Change: Basic > CPU Used: Change from 8 to 4 (Medium)  Reason(s): 1. 8 Will produce a very low quality encode with poor VMAF scores in a lot of cases.  2. 4 is a good balance in terms of speed and quality and produces good VMAF scores.  3. Charts show that cpu speeds 0-4 all have very close VMAF scores, but cpu speeds 5-8 begin producing much worse VMAF scores, therefore CPU 4 should be the default (in my opinion) ------------  Change: Basic > CQ Level: Change from 0 to 25  Reason(s):  1. 0 is lossless, and lossless should never be the default for a lossy encoder. This shouldn't require a huge explanation, as the reasons should be obvious.  2. 25 is a good default quality level since it produces a high quality encode at a small filesize. AV1 excels at higher CQ levels compared to x265, and AV1 is primarily used to create low filesize encodes at a target quality. AV1 isn't well optimized for low CQ levels due to quantizer related denoising anomalies (especially on grainy content). The goal of AV1 is to denoise / remove granular details without becoming a blocky, swimming mess, and it is better able to achieve that goal at higher CQ levels, until some kinks have been ironed out. Believe me, I would LOVE to use a lower CQ level as a default, but CQ 30 just, across the board, produces more consistent results. 3. CQ Level 25 provides good quality for both 8bit and native 10bit content. According to user BlueSwordM, native 10bit content benefits more from a CQ level of 25 compared with higher CQ levels. 25 is going to be a safe starting point for content of either bit depth.  ------------ Change: Analysis > Tile Columns: Change from 0 to 2 Analysis > Tile Rows: Change from 0 to 1  Reason(s):  1. Assists with multi-threaded encoding by better utilizing cores on your cpu, with a downside of a small hit to encoding compression. However, many AV1 users agree that the speed gained is worth the tradeoff.  3. Reddit user said this:   "By keeping tiles as "square" as possible, you have the lowest edge/area ratio. Edges cause a small loss in coding efficiency. So for a 16:9 video, 2x1, 2x2 or 4x2 is recommended depending on resolution. Most times 2x1 is good for 1080p, 2x2 for 1440p and 4x2 for 2160p and higher."   4. Since most content being encoded is still 1080p, then 2x1 is probably a safe default.  ------------ Change: Slice Decision > Min GOP Size: Rename to Keyframe Min Distance Slice Decision > Max GOP Size: Rename to Keyframe Max Distance Reason(s): 1. This will better reflect the parameter name since AV1 refers to this as keyframe distance, rather than "Group of Picture" size  2. Easier to understand without inexperienced users having to resort to google to figure out what GOP means. ------------ Change: Slice Decision > Max GOP Size (i.e. Keyframe Max Distance): Change from 0 to 240  Reason(s):  1. I heard AV1 users at one point discussing this number being recommended by the AV1 developers, and also I believe this is a recommended number even for VP9.  2. It represents approximately maximum of 10 seconds between keyframes (if your framerate is approximately 24fps), otherwise keyframes could be placed at "who knows". Better to specify than to leave it unspecified, since the AV1 parameter doesn't say what the default value is. ------------  Change: Slice Decision > Lag in Frames: Change from 0 to 25  Reason(s): 1. --lag-in-frames parameter defines an upper limit on the number of frames into the future that the encoder can look 2. I have seen people preferring 25 for AV1.  3. For VP9, a value of 16 was recommended. I have no idea what the recommended or the max value is in AV1, but it seems safe to default this to 25. ------------  Change: Performance > Add a checkbox called "Automatically determine thread count". Enabling this would gray out/disallow user input for "Threads", and the Thread count would be automatically determined based on the CPU's multi-threading capability.  Reason(s):  1. 0 doesn't actually cause aomenc to auto multi thread. 0 disables multi-threading.  2. Sloppy Workaround: You could try setting default thread count to 8 since that's a safe number in 2020, since most people doing video conversion in 2020 have at least a quad core with hyper threading, however it would be best if StaxRip could automatically determine the cpu's maximum thread capability and automatically set this number accordingly. ------------ Change: Misc 2 > Auto Alt Ref: Change from 0 to 1  Reason(s):  1. It is better to have "auto alt ref" Enabled because according to some AV1 users in discord, having it disabled will result in a "completely shit encode"  2. If you're doing a single pass encode, auto-alt-ref is actually disabled by default, but you can force it to be enabled by forcibly specifying it.  3. alt-ref frames are used when using a 2-pass encode by default if it is not specified. Based on my limited understanding, you are generally better off just forcibly enabling this setting for most case scenarios.
priority
aomenc better defaults with explanations change basic mode change from one pass to two pass  reason s  first pass is usually always very fast  aomenc does not use alt ref frames when performing a single pass therefore it is generally recommended to always perform a pass encode no matter what rate mode is chosen       change basic rate mode  rename q to q constant quality  rename cq to cq constrained quality  reason s  this will help eliminate confusion because users coming from might choose cq expecting it to behave similar to s crf cq but it will result in a terrible quality encode due to it producing a constrained quality video      change basic rate mode change from vbr to q  reason s  aomenc is less likely to be used for targeting precise filesizes and more likely to be used for targeting quality at imprecise yet small filesizes       change basic bit depth change from to  reason s  is basically the new norm as there are almost no new devices software that do not support decoding  almost always produces a better quality encode with smoother gradients in skies  it s safer to do final encode than even if converting an source video into a result there will almost always be fewer anomalies going with a final encode     change basic cpu used change from to medium  reason s will produce a very low quality encode with poor vmaf scores in a lot of cases   is a good balance in terms of speed and quality and produces good vmaf scores   charts show that cpu speeds all have very close vmaf scores but cpu speeds begin producing much worse vmaf scores therefore cpu should be the default in my opinion      change basic cq level change from to  reason s  is lossless and lossless should never be the default for a lossy encoder this shouldn t require a huge explanation as the reasons should be obvious  is a good default quality level since it produces a high quality encode at a small filesize excels at higher cq levels compared to and is primarily used to create low filesize encodes at a target quality isn t well optimized for low cq levels due to quantizer related denoising anomalies especially on grainy content the goal of is to denoise remove granular details without becoming a blocky swimming mess and it is better able to achieve that goal at higher cq levels until some kinks have been ironed out believe me i would love to use a lower cq level as a default but cq just across the board produces more consistent results cq level provides good quality for both and native content according to user blueswordm native content benefits more from a cq level of compared with higher cq levels is going to be a safe starting point for content of either bit depth      change analysis tile columns change from to analysis tile rows change from to   reason s  assists with multi threaded encoding by better utilizing cores on your cpu with a downside of a small hit to encoding compression however many users agree that the speed gained is worth the tradeoff  reddit user said this   by keeping tiles as square as possible you have the lowest edge area ratio edges cause a small loss in coding efficiency so for a video or is recommended depending on resolution most times is good for for and for and higher   since most content being encoded is still then is probably a safe default      change slice decision min gop size rename to keyframe min distance slice decision max gop size rename to keyframe max distance reason s this will better reflect the parameter name since refers to this as keyframe distance rather than group of picture size  easier to understand without inexperienced users having to resort to google to figure out what gop means      change slice decision max gop size i e keyframe max distance change from to  reason s  i heard users at one point discussing this number being recommended by the developers and also i believe this is a recommended number even for  it represents approximately maximum of seconds between keyframes if your framerate is approximately otherwise keyframes could be placed at who knows better to specify than to leave it unspecified since the parameter doesn t say what the default value is      change slice decision lag in frames change from to  reason s  lag in frames parameter defines an upper limit on the number of frames into the future that the encoder can look i have seen people preferring for  for a value of was recommended i have no idea what the recommended or the max value is in but it seems safe to default this to      change performance add a checkbox called automatically determine thread count enabling this would gray out disallow user input for threads and the thread count would be automatically determined based on the cpu s multi threading capability  reason s  doesn t actually cause aomenc to auto multi thread disables multi threading  sloppy workaround you could try setting default thread count to since that s a safe number in since most people doing video conversion in have at least a quad core with hyper threading however it would be best if staxrip could automatically determine the cpu s maximum thread capability and automatically set this number accordingly     change misc auto alt ref change from to  reason s  it is better to have auto alt ref enabled because according to some users in discord having it disabled will result in a completely shit encode  if you re doing a single pass encode auto alt ref is actually disabled by default but you can force it to be enabled by forcibly specifying it  alt ref frames are used when using a pass encode by default if it is not specified based on my limited understanding you are generally better off just forcibly enabling this setting for most case scenarios
1
24,949
24,512,381,144
IssuesEvent
2022-10-10 23:25:25
tailscale/tailscale
https://api.github.com/repos/tailscale/tailscale
closed
Raspberry Pi Libreelec does not login after reboot
L1 Very few P1 Nuisance T5 Usability bug
### What is the issue? I have installed Tailscale on Libreelec manually and I don't know if it's working correctly. To install I downloaded the arm64 static binary, unpacked it and made this service to automatically start tailscale on startup: ``` [Unit] Description=Tailscale node agent Documentation=https://tailscale.com/kb/ Wants=network-pre.target After=network-pre.target StartLimitIntervalSec=0 StartLimitBurst=0 [Service] EnvironmentFile=/storage/tailscale/tailscaled.defaults ExecStartPre=/storage/tailscale/tailscaled --cleanup ExecStart=/storage/tailscale/tailscaled --state=/var/lib/tailscale/tailscaled.state --socket=/run/tailscale/tailscaled.sock --port $PORT $FLAGS ExecStopPost=/storage/tailscale/tailscaled --cleanup Restart=on-failure RuntimeDirectory=tailscale RuntimeDirectoryMode=0777 StateDirectory=tailscale StateDirectoryMode=0777 CacheDirectory=tailscale CacheDirectoryMode=0777 Type=notify [Install] WantedBy=multi-user.target ``` Now when I try to login for the first time with `tailscale up` and authorize in the browser it is successful and there seems to be something safed to `/var/lib/tailscale/` as a new folder "files", tailscaled.state and two empty log files are created. However, when the Pi is rebooted the service runs fine but it seems like it is unable to login as the Pi is not connected. Running `tailscale up` again needs reauthentication and the IP changes which I try to avoid because I want to use a DNS. I tried to authorize with a timeout and a reuseable key at start, this works but it is also always a new IP. What I hope is that it can behave like on a laptop where I can disconnect and connect again without reauthorization, keeping the same IP. This is the log after rebooting: [log.txt](https://github.com/tailscale/tailscale/files/9717796/log.txt) ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS Linux ### OS version Libreelec 10.2 ### Tailscale version 1.30.2 ### Bug report BUG-a42c55eaa2754c6172171faab53e660ad278bb51e7f4c1ccf001b2343be12370-20221005174348Z-499137b2ec467d14
True
Raspberry Pi Libreelec does not login after reboot - ### What is the issue? I have installed Tailscale on Libreelec manually and I don't know if it's working correctly. To install I downloaded the arm64 static binary, unpacked it and made this service to automatically start tailscale on startup: ``` [Unit] Description=Tailscale node agent Documentation=https://tailscale.com/kb/ Wants=network-pre.target After=network-pre.target StartLimitIntervalSec=0 StartLimitBurst=0 [Service] EnvironmentFile=/storage/tailscale/tailscaled.defaults ExecStartPre=/storage/tailscale/tailscaled --cleanup ExecStart=/storage/tailscale/tailscaled --state=/var/lib/tailscale/tailscaled.state --socket=/run/tailscale/tailscaled.sock --port $PORT $FLAGS ExecStopPost=/storage/tailscale/tailscaled --cleanup Restart=on-failure RuntimeDirectory=tailscale RuntimeDirectoryMode=0777 StateDirectory=tailscale StateDirectoryMode=0777 CacheDirectory=tailscale CacheDirectoryMode=0777 Type=notify [Install] WantedBy=multi-user.target ``` Now when I try to login for the first time with `tailscale up` and authorize in the browser it is successful and there seems to be something safed to `/var/lib/tailscale/` as a new folder "files", tailscaled.state and two empty log files are created. However, when the Pi is rebooted the service runs fine but it seems like it is unable to login as the Pi is not connected. Running `tailscale up` again needs reauthentication and the IP changes which I try to avoid because I want to use a DNS. I tried to authorize with a timeout and a reuseable key at start, this works but it is also always a new IP. What I hope is that it can behave like on a laptop where I can disconnect and connect again without reauthorization, keeping the same IP. This is the log after rebooting: [log.txt](https://github.com/tailscale/tailscale/files/9717796/log.txt) ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS Linux ### OS version Libreelec 10.2 ### Tailscale version 1.30.2 ### Bug report BUG-a42c55eaa2754c6172171faab53e660ad278bb51e7f4c1ccf001b2343be12370-20221005174348Z-499137b2ec467d14
non_priority
raspberry pi libreelec does not login after reboot what is the issue i have installed tailscale on libreelec manually and i don t know if it s working correctly to install i downloaded the static binary unpacked it and made this service to automatically start tailscale on startup description tailscale node agent documentation wants network pre target after network pre target startlimitintervalsec startlimitburst environmentfile storage tailscale tailscaled defaults execstartpre storage tailscale tailscaled cleanup execstart storage tailscale tailscaled state var lib tailscale tailscaled state socket run tailscale tailscaled sock port port flags execstoppost storage tailscale tailscaled cleanup restart on failure runtimedirectory tailscale runtimedirectorymode statedirectory tailscale statedirectorymode cachedirectory tailscale cachedirectorymode type notify wantedby multi user target now when i try to login for the first time with tailscale up and authorize in the browser it is successful and there seems to be something safed to var lib tailscale as a new folder files tailscaled state and two empty log files are created however when the pi is rebooted the service runs fine but it seems like it is unable to login as the pi is not connected running tailscale up again needs reauthentication and the ip changes which i try to avoid because i want to use a dns i tried to authorize with a timeout and a reuseable key at start this works but it is also always a new ip what i hope is that it can behave like on a laptop where i can disconnect and connect again without reauthorization keeping the same ip this is the log after rebooting steps to reproduce no response are there any recent changes that introduced the issue no response os linux os version libreelec tailscale version bug report bug
0
134,855
19,399,949,497
IssuesEvent
2021-12-19 01:25:12
hackforla/website
https://api.github.com/repos/hackforla/website
closed
Create template for Reviewing the alt text on all pages
Status: Updated Feature: Accessibility role: design Size: Small
### Overview We need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way. ### Action Items - [x] Create template draft in comment below - [x] Talk to Matthew to find out if dev is going to make this list - [x] revise after Matthew conversation - [x] review with Bonnie - [x] have Bonnie help you make it into a template on the site - [x] Make for all relevant issues - [ ] List the issues under resources with checkboxes ### Resources/Instructions - [Entire worksheet Google Sheet](https://docs.google.com/spreadsheets/d/1abMlPOX88zHTV29E4BFxAYwahCxgJN28Mn95cXr8jak/edit#gid=308160713) - Issue we are modeling this template after: https://github.com/hackforla/website/issues/2354
1.0
Create template for Reviewing the alt text on all pages - ### Overview We need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way. ### Action Items - [x] Create template draft in comment below - [x] Talk to Matthew to find out if dev is going to make this list - [x] revise after Matthew conversation - [x] review with Bonnie - [x] have Bonnie help you make it into a template on the site - [x] Make for all relevant issues - [ ] List the issues under resources with checkboxes ### Resources/Instructions - [Entire worksheet Google Sheet](https://docs.google.com/spreadsheets/d/1abMlPOX88zHTV29E4BFxAYwahCxgJN28Mn95cXr8jak/edit#gid=308160713) - Issue we are modeling this template after: https://github.com/hackforla/website/issues/2354
non_priority
create template for reviewing the alt text on all pages overview we need to create a template so that we can easily make issue for the design team to check each page on the site in a consistent way action items create template draft in comment below talk to matthew to find out if dev is going to make this list revise after matthew conversation review with bonnie have bonnie help you make it into a template on the site make for all relevant issues list the issues under resources with checkboxes resources instructions issue we are modeling this template after
0
74,352
25,080,647,581
IssuesEvent
2022-11-07 19:00:27
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
opened
Strictness of event dispatch result for ORM checkRules()
defect
### Description Error > Cake\ORM\Table::checkRules(): Return value must be of type bool, null returned As the signature of that method is: ```php public function checkRules(...): bool ``` The code inside checkRules() reads ```php $event = $this->dispatchEvent( 'Model.afterRules', compact('entity', 'options', 'result', 'operation') ); if ($event->isStopped()) { return $event->getResult(); } ``` Maybe, to make this less hard to fail, we should allow null as OK value? return $event->getResult() ?? true; What do you think? ### CakePHP Version 4.4.7 ### PHP Version 8.1
1.0
Strictness of event dispatch result for ORM checkRules() - ### Description Error > Cake\ORM\Table::checkRules(): Return value must be of type bool, null returned As the signature of that method is: ```php public function checkRules(...): bool ``` The code inside checkRules() reads ```php $event = $this->dispatchEvent( 'Model.afterRules', compact('entity', 'options', 'result', 'operation') ); if ($event->isStopped()) { return $event->getResult(); } ``` Maybe, to make this less hard to fail, we should allow null as OK value? return $event->getResult() ?? true; What do you think? ### CakePHP Version 4.4.7 ### PHP Version 8.1
non_priority
strictness of event dispatch result for orm checkrules description error cake orm table checkrules return value must be of type bool null returned as the signature of that method is php public function checkrules bool the code inside checkrules reads php event this dispatchevent model afterrules compact entity options result operation if event isstopped return event getresult maybe to make this less hard to fail we should allow null as ok value return event getresult true what do you think cakephp version php version
0
62,875
3,193,579,159
IssuesEvent
2015-09-30 06:42:50
fusioninventory/fusioninventory-for-glpi
https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi
closed
Package Files : rename uncompress field
Category: Deploy Component: For junior contributor Priority: Normal Status: Closed Tracker: Feature
--- Author Name: **alexandre delaunay** (@orthagh) Original Redmine Issue: 1254, http://forge.fusioninventory.org/issues/1254 Original Date: 2011-10-13 --- None
1.0
Package Files : rename uncompress field - --- Author Name: **alexandre delaunay** (@orthagh) Original Redmine Issue: 1254, http://forge.fusioninventory.org/issues/1254 Original Date: 2011-10-13 --- None
priority
package files rename uncompress field author name alexandre delaunay orthagh original redmine issue original date none
1
679,704
23,242,466,059
IssuesEvent
2022-08-03 16:46:55
spiffe/spire
https://api.github.com/repos/spiffe/spire
closed
Upgrade to google.golang.org/protobuf 1.28.1 breaks unit test assertions
priority/backlog
Upgrading from [google.golang.org/protobuf](http://google.golang.org/protobuf) 1.28.0 to 1.28.1 and results in some unit test failures. Something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields. Example: [these assertions](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache_test.go#L540-L541) in `pkg/agent/manager/storecache:TestUpdateEntries` fail, likely because the [storecache.Record](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache.go#L21) and [cache.StaleEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/cache/cache.go#L126) structs include a *[common.RegistrationEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/proto/spire/common/common.proto#L59-L91) protobuf message field that is not cleanly matching using `reflect.DeepEqual()`. Digging into the failure, it looks like some internal proto field called `atomicMessageInfo` is not matching - one of the objects contains a pointer value for that field, the other a nil value. It is probably worth evaluating whether protobuf messages should be embedded in Go structs in SPIRE code because of the challenges involved with comparing Go structs containing several fields, where some of the fields are plain Go structs and some are protobuf messages.
1.0
Upgrade to google.golang.org/protobuf 1.28.1 breaks unit test assertions - Upgrading from [google.golang.org/protobuf](http://google.golang.org/protobuf) 1.28.0 to 1.28.1 and results in some unit test failures. Something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields. Example: [these assertions](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache_test.go#L540-L541) in `pkg/agent/manager/storecache:TestUpdateEntries` fail, likely because the [storecache.Record](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/storecache/cache.go#L21) and [cache.StaleEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/pkg/agent/manager/cache/cache.go#L126) structs include a *[common.RegistrationEntry](https://github.com/spiffe/spire/blob/3e56c0120cff69f53e202e6fb858a60ef7cd514c/proto/spire/common/common.proto#L59-L91) protobuf message field that is not cleanly matching using `reflect.DeepEqual()`. Digging into the failure, it looks like some internal proto field called `atomicMessageInfo` is not matching - one of the objects contains a pointer value for that field, the other a nil value. It is probably worth evaluating whether protobuf messages should be embedded in Go structs in SPIRE code because of the challenges involved with comparing Go structs containing several fields, where some of the fields are plain Go structs and some are protobuf messages.
priority
upgrade to google golang org protobuf breaks unit test assertions upgrading from to and results in some unit test failures something in this latest version of protobuf breaks some of the assertions in our code for types that embed protobuf messages as fields example in pkg agent manager storecache testupdateentries fail likely because the and structs include a protobuf message field that is not cleanly matching using reflect deepequal digging into the failure it looks like some internal proto field called atomicmessageinfo is not matching one of the objects contains a pointer value for that field the other a nil value it is probably worth evaluating whether protobuf messages should be embedded in go structs in spire code because of the challenges involved with comparing go structs containing several fields where some of the fields are plain go structs and some are protobuf messages
1
72,411
24,108,144,953
IssuesEvent
2022-09-20 09:09:18
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Element desktop frequently crashes
T-Defect
### Steps to reproduce 1. Open the app, wait around 10 seconds (sometimes a bit more, but not longer than a minute i guess). 2. Experience the crash This happens very reliably. I cannot use the app for more than a few seconds. I've tried deleting `~/.config/Elemen` and reinstalling the app without success. ### Outcome #### What did you expect? Element donesn't crash #### What happened instead? A crash ### Operating system Arch Linux ### Application version Element version: 1.11.4 Olm version: 3.2.12 ### How did you install the app? From the official Arch repo (community) ### Homeserver Private homeserver, Synapse 1.67.0 (from nixpkgs) ### Will you send logs? Yes
1.0
Element desktop frequently crashes - ### Steps to reproduce 1. Open the app, wait around 10 seconds (sometimes a bit more, but not longer than a minute i guess). 2. Experience the crash This happens very reliably. I cannot use the app for more than a few seconds. I've tried deleting `~/.config/Elemen` and reinstalling the app without success. ### Outcome #### What did you expect? Element donesn't crash #### What happened instead? A crash ### Operating system Arch Linux ### Application version Element version: 1.11.4 Olm version: 3.2.12 ### How did you install the app? From the official Arch repo (community) ### Homeserver Private homeserver, Synapse 1.67.0 (from nixpkgs) ### Will you send logs? Yes
non_priority
element desktop frequently crashes steps to reproduce open the app wait around seconds sometimes a bit more but not longer than a minute i guess experience the crash this happens very reliably i cannot use the app for more than a few seconds i ve tried deleting config elemen and reinstalling the app without success outcome what did you expect element donesn t crash what happened instead a crash operating system arch linux application version element version olm version how did you install the app from the official arch repo community homeserver private homeserver synapse from nixpkgs will you send logs yes
0
51,192
6,150,536,753
IssuesEvent
2017-06-27 22:54:06
SavageCore/new-retro-arcade-neon-attract-screen-tool
https://api.github.com/repos/SavageCore/new-retro-arcade-neon-attract-screen-tool
closed
App fails to load for certain people
bug critical needs-testing
![Imgur](http://i.imgur.com/Ujru1xc.jpg) Disabling `updateChecker()` resolves the issue for [mrleelee123](http://steamcommunity.com/app/465780/discussions/0/343787283758035953/?tscn=1479142088#c217691032444349428)
1.0
App fails to load for certain people - ![Imgur](http://i.imgur.com/Ujru1xc.jpg) Disabling `updateChecker()` resolves the issue for [mrleelee123](http://steamcommunity.com/app/465780/discussions/0/343787283758035953/?tscn=1479142088#c217691032444349428)
non_priority
app fails to load for certain people disabling updatechecker resolves the issue for
0
708,994
24,364,483,305
IssuesEvent
2022-10-03 14:13:02
IAmTamal/Milan
https://api.github.com/repos/IAmTamal/Milan
opened
[DOCS] Add setup instruction in the readme file
📄 aspect: text ✨ goal: improvement 🟨 priority: medium 🛠 status : under development
### Description There is no setup instructions in the readme file. i think we can add both setup steps in the readme file, one with npm and one with docker. ### Screenshots _No response_ ### Additional information _No response_
1.0
[DOCS] Add setup instruction in the readme file - ### Description There is no setup instructions in the readme file. i think we can add both setup steps in the readme file, one with npm and one with docker. ### Screenshots _No response_ ### Additional information _No response_
priority
add setup instruction in the readme file description there is no setup instructions in the readme file i think we can add both setup steps in the readme file one with npm and one with docker screenshots no response additional information no response
1
85,063
15,731,185,231
IssuesEvent
2021-03-29 16:46:12
wrbejar/bag-of-holding
https://api.github.com/repos/wrbejar/bag-of-holding
opened
CVE-2020-11022 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-2.1.3.min.js</b>, <b>jquery-1.10.2.min.js</b></p></summary> <p> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.1.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p> <p>Path to vulnerable library: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.3.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.10.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html</p> <p>Path to vulnerable library: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html,bag-of-holding/bower_components/select2-bootstrap-css/_jekyll/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.10.2.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","packageFilePaths":["/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.3","packageFilePaths":["/bower_components/Chart.js/samples/line-customTooltips.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html","/bower_components/select2-bootstrap-css/_jekyll/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11022 (Medium) detected in multiple libraries - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-2.1.3.min.js</b>, <b>jquery-1.10.2.min.js</b></p></summary> <p> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: bag-of-holding/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.1.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.3/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p> <p>Path to vulnerable library: bag-of-holding/bower_components/Chart.js/samples/line-customTooltips.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.3.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.10.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.10.2/jquery.min.js</a></p> <p>Path to dependency file: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html</p> <p>Path to vulnerable library: bag-of-holding/bower_components/select2-bootstrap-css/docs/3.4.4.html,bag-of-holding/bower_components/select2-bootstrap-css/_jekyll/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.10.2.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","packageFilePaths":["/node_modules/bower/lib/node_modules/redeyed/examples/browser/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.3","packageFilePaths":["/bower_components/Chart.js/samples/line-customTooltips.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.10.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html","/bower_components/select2-bootstrap-css/_jekyll/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding node modules bower lib node modules redeyed examples browser index html path to vulnerable library bag of holding node modules bower lib node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding bower components chart js samples line customtooltips html path to vulnerable library bag of holding bower components chart js samples line customtooltips html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file bag of holding bower components bootstrap css docs html path to vulnerable library bag of holding bower components bootstrap css docs html bag of holding bower components bootstrap css jekyll index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery basebranches vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl
0
66,222
20,072,280,336
IssuesEvent
2022-02-04 08:40:33
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Message search initialisation failed (invalid passphrase)
T-Defect X-Needs-Info S-Major A-E2EE A-Search A-Indexing X-Needs-Investigation O-Uncommon
### Steps to reproduce 1. Where are you starting? What can you see? If I try to search messages in a non group chat, I get `Message search initialisation failed, check your settings for more information`. If I click the settings, I can see: `Message search initialisation failed`. If I click the drop down I see `Error opening the database: DatabaseUnlockError("Invalid passphrase")`. I have no idea what passphrase this is. When I launch element, I am prompted for a password, which I type in and everything launches ok. I dont understand what passphrase this is that it is missing. 2. What do you click? 3. More steps… ### Outcome #### What did you expect? Search to work #### What happened instead? Search didnt work ### Operating system Ubuntu 20.04 lts ### Application version Element Desktop Element version: 1.9.9 Olm version: 3.2.8 ### How did you install the app? debian package ### Homeserver Non-public, not sure ### Will you send logs? No
1.0
Message search initialisation failed (invalid passphrase) - ### Steps to reproduce 1. Where are you starting? What can you see? If I try to search messages in a non group chat, I get `Message search initialisation failed, check your settings for more information`. If I click the settings, I can see: `Message search initialisation failed`. If I click the drop down I see `Error opening the database: DatabaseUnlockError("Invalid passphrase")`. I have no idea what passphrase this is. When I launch element, I am prompted for a password, which I type in and everything launches ok. I dont understand what passphrase this is that it is missing. 2. What do you click? 3. More steps… ### Outcome #### What did you expect? Search to work #### What happened instead? Search didnt work ### Operating system Ubuntu 20.04 lts ### Application version Element Desktop Element version: 1.9.9 Olm version: 3.2.8 ### How did you install the app? debian package ### Homeserver Non-public, not sure ### Will you send logs? No
non_priority
message search initialisation failed invalid passphrase steps to reproduce where are you starting what can you see if i try to search messages in a non group chat i get message search initialisation failed check your settings for more information if i click the settings i can see message search initialisation failed if i click the drop down i see error opening the database databaseunlockerror invalid passphrase i have no idea what passphrase this is when i launch element i am prompted for a password which i type in and everything launches ok i dont understand what passphrase this is that it is missing what do you click more steps… outcome what did you expect search to work what happened instead search didnt work operating system ubuntu lts application version element desktop element version olm version how did you install the app debian package homeserver non public not sure will you send logs no
0
236,454
18,097,964,142
IssuesEvent
2021-09-22 11:13:12
alphagov/govuk-frontend
https://api.github.com/repos/alphagov/govuk-frontend
opened
Add info about automated tools to accessibility criteria document
documentation accessibility
## What Based on user feedback, add info about automated tools to [accessibility criteria doc](https://github.com/alphagov/govuk-frontend/blob/main/docs/contributing/test-components-using-accessibility-acceptance-criteria.md). ## Why After we got 2i approval to publish our accessibility criteria, we realised it lacked content about automated tools that help detect accessibility issues. ### Points to consider - according to [GOV.UK's blog on testing with a deliberately inaccessible webpage](https://accessibility.blog.gov.uk/2017/02/24/what-we-found-when-we-tested-tools-on-the-worlds-least-accessible-webpage/), 29% of the accessibility barriers went undetected by any of the 10 automated tools used - for more info, see the section titled 'Lots of the barriers weren’t found by any of the tools' - however, automated tools (for example, Axe, SiteImprove) are still useful for picking up basic issues - we'd like to hear from the Design System community about the automated tools they find useful ## Who needs to know about this Technical Writer, Community Manager, Developers ## Done when - [ ] Developer and Technical Writer draft update - [ ] Update receives review from teammate - [ ] Update passes 2i - [ ] We publish update
1.0
Add info about automated tools to accessibility criteria document - ## What Based on user feedback, add info about automated tools to [accessibility criteria doc](https://github.com/alphagov/govuk-frontend/blob/main/docs/contributing/test-components-using-accessibility-acceptance-criteria.md). ## Why After we got 2i approval to publish our accessibility criteria, we realised it lacked content about automated tools that help detect accessibility issues. ### Points to consider - according to [GOV.UK's blog on testing with a deliberately inaccessible webpage](https://accessibility.blog.gov.uk/2017/02/24/what-we-found-when-we-tested-tools-on-the-worlds-least-accessible-webpage/), 29% of the accessibility barriers went undetected by any of the 10 automated tools used - for more info, see the section titled 'Lots of the barriers weren’t found by any of the tools' - however, automated tools (for example, Axe, SiteImprove) are still useful for picking up basic issues - we'd like to hear from the Design System community about the automated tools they find useful ## Who needs to know about this Technical Writer, Community Manager, Developers ## Done when - [ ] Developer and Technical Writer draft update - [ ] Update receives review from teammate - [ ] Update passes 2i - [ ] We publish update
non_priority
add info about automated tools to accessibility criteria document what based on user feedback add info about automated tools to why after we got approval to publish our accessibility criteria we realised it lacked content about automated tools that help detect accessibility issues points to consider according to of the accessibility barriers went undetected by any of the automated tools used for more info see the section titled lots of the barriers weren’t found by any of the tools however automated tools for example axe siteimprove are still useful for picking up basic issues we d like to hear from the design system community about the automated tools they find useful who needs to know about this technical writer community manager developers done when developer and technical writer draft update update receives review from teammate update passes we publish update
0
713,018
24,515,057,791
IssuesEvent
2022-10-11 03:43:15
matrixorigin/matrixone
https://api.github.com/repos/matrixorigin/matrixone
closed
[Subtask]: remove update related
priority/p0 component/distributed tae kind/subtask
### Parent Issue #4628 ### Detail of Subtask Remove update-related code and test and refactor any code that depends on the code. ### Describe implementation you've considered _No response_ ### Additional information _No response_
1.0
[Subtask]: remove update related - ### Parent Issue #4628 ### Detail of Subtask Remove update-related code and test and refactor any code that depends on the code. ### Describe implementation you've considered _No response_ ### Additional information _No response_
priority
remove update related parent issue detail of subtask remove update related code and test and refactor any code that depends on the code describe implementation you ve considered no response additional information no response
1
357,242
10,604,023,074
IssuesEvent
2019-10-10 17:15:02
satcom886/rugl
https://api.github.com/repos/satcom886/rugl
closed
Make the labels a little more original... Maybe?
low-priority organisation
pycracker - What could be wrong? pylogin - What is this even for? pyparosci - How broken is it?
1.0
Make the labels a little more original... Maybe? - pycracker - What could be wrong? pylogin - What is this even for? pyparosci - How broken is it?
priority
make the labels a little more original maybe pycracker what could be wrong pylogin what is this even for pyparosci how broken is it
1
759,408
26,593,109,321
IssuesEvent
2023-01-23 10:19:09
status-im/status-desktop
https://api.github.com/repos/status-im/status-desktop
closed
[Performance] Improve Reply message author detail fetch
priority 2: medium E:PerformanceImprovements S:3 messenger
### Description Right now, when there is a reply message, we get the details for the author of the replied message using the pubkey only and calling the backend to get JSON. That's very inefficient, as we need to do a back and forth and then parse the data. Plus, I needed to do a iterator property just to tell the frontend to call the JSON again. One way I realized that will improve that a lot is to just populate it all in the backend (in the module most likely). We just need to use the new properties Pascal added here: https://github.com/status-im/status-desktop/pull/9111/files#diff-4f86d004be69cc9b51a70273a5ec7ab80e851a4231773dad079e8c01afa753b5R53-R54 When we detect that the author was updated (the model already has `modelContactUpdateIterator` taht is called in those situations), we just update the details directly instead of relying on the iterator.
1.0
[Performance] Improve Reply message author detail fetch - ### Description Right now, when there is a reply message, we get the details for the author of the replied message using the pubkey only and calling the backend to get JSON. That's very inefficient, as we need to do a back and forth and then parse the data. Plus, I needed to do a iterator property just to tell the frontend to call the JSON again. One way I realized that will improve that a lot is to just populate it all in the backend (in the module most likely). We just need to use the new properties Pascal added here: https://github.com/status-im/status-desktop/pull/9111/files#diff-4f86d004be69cc9b51a70273a5ec7ab80e851a4231773dad079e8c01afa753b5R53-R54 When we detect that the author was updated (the model already has `modelContactUpdateIterator` taht is called in those situations), we just update the details directly instead of relying on the iterator.
priority
improve reply message author detail fetch description right now when there is a reply message we get the details for the author of the replied message using the pubkey only and calling the backend to get json that s very inefficient as we need to do a back and forth and then parse the data plus i needed to do a iterator property just to tell the frontend to call the json again one way i realized that will improve that a lot is to just populate it all in the backend in the module most likely we just need to use the new properties pascal added here when we detect that the author was updated the model already has modelcontactupdateiterator taht is called in those situations we just update the details directly instead of relying on the iterator
1
317,538
9,666,447,637
IssuesEvent
2019-05-21 10:50:39
oceanprotocol/ocean
https://api.github.com/repos/oceanprotocol/ocean
opened
Integration of Aquarius with Elastic Search for Commons
L2 Metadata bug commons priority:medium
In order to support better search capabilities, it would be necessary to integrate Aquarius with Elastic Search in the `Duero` and `Nile` environments. The tasks included are: - [ ] Deployment of Elastic Search in `Duero` https://github.com/oceanprotocol/atlantic/issues/54 - [ ] Integration of Aquarius with Elastic Search in `Duero` https://github.com/oceanprotocol/aquarius/issues/184 - [ ] Deployment of Elastic Search in `Nile` - [ ] Integration of Aquarius with Elastic Search in `Nile`
1.0
Integration of Aquarius with Elastic Search for Commons - In order to support better search capabilities, it would be necessary to integrate Aquarius with Elastic Search in the `Duero` and `Nile` environments. The tasks included are: - [ ] Deployment of Elastic Search in `Duero` https://github.com/oceanprotocol/atlantic/issues/54 - [ ] Integration of Aquarius with Elastic Search in `Duero` https://github.com/oceanprotocol/aquarius/issues/184 - [ ] Deployment of Elastic Search in `Nile` - [ ] Integration of Aquarius with Elastic Search in `Nile`
priority
integration of aquarius with elastic search for commons in order to support better search capabilities it would be necessary to integrate aquarius with elastic search in the duero and nile environments the tasks included are deployment of elastic search in duero integration of aquarius with elastic search in duero deployment of elastic search in nile integration of aquarius with elastic search in nile
1
297,349
25,723,568,962
IssuesEvent
2022-12-07 15:13:57
vegaprotocol/frontend-monorepo
https://api.github.com/repos/vegaprotocol/frontend-monorepo
closed
Support starting the front-ends with a long-living token
Trading Testing 🧪 Wallet-Automation common
Blocked by: - https://github.com/vegaprotocol/frontend-monorepo/issues/2249 --- Long-living tokens are tokens that are pre-generated from the command line. ``` vega wallet api-token generate ``` These tokens allows headless software to query the service without human intervention. It primarily exists to support bots and automated scripts. It can also be of good use for UI testing. However, since this feature is designed with a slightly different request flow than the "interactive" one (manual approval and all), the front-ends needs to be able to account for that difference. Interactive workflow is: 1. `client.connect_wallet` -> get a connection token 2. `client.list_keys { "token": token }` 3. etc Headless workflow is: 1. `client.list_keys { "token": longLivingToken }` 2. etc. With a long-living token, there is no need to call `client.connect_wallet` since we already have the token. So the front-ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow.
1.0
Support starting the front-ends with a long-living token - Blocked by: - https://github.com/vegaprotocol/frontend-monorepo/issues/2249 --- Long-living tokens are tokens that are pre-generated from the command line. ``` vega wallet api-token generate ``` These tokens allows headless software to query the service without human intervention. It primarily exists to support bots and automated scripts. It can also be of good use for UI testing. However, since this feature is designed with a slightly different request flow than the "interactive" one (manual approval and all), the front-ends needs to be able to account for that difference. Interactive workflow is: 1. `client.connect_wallet` -> get a connection token 2. `client.list_keys { "token": token }` 3. etc Headless workflow is: 1. `client.list_keys { "token": longLivingToken }` 2. etc. With a long-living token, there is no need to call `client.connect_wallet` since we already have the token. So the front-ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow.
non_priority
support starting the front ends with a long living token blocked by long living tokens are tokens that are pre generated from the command line vega wallet api token generate these tokens allows headless software to query the service without human intervention it primarily exists to support bots and automated scripts it can also be of good use for ui testing however since this feature is designed with a slightly different request flow than the interactive one manual approval and all the front ends needs to be able to account for that difference interactive workflow is client connect wallet get a connection token client list keys token token etc headless workflow is client list keys token longlivingtoken etc with a long living token there is no need to call client connect wallet since we already have the token so the front ends must be able to start in a special testing mode to work with the headless workflow instead of the interactive workflow
0
72,507
3,386,588,651
IssuesEvent
2015-11-27 19:00:37
YetiForceCompany/YetiForceCRM
https://api.github.com/repos/YetiForceCompany/YetiForceCRM
closed
How to reset backup
Label::Module Priority::#3 High Type::Bug
I tried the backup option in 2.2 and it was hanging again. Even updated to 2.3 but still the same screen: Any idea how to reset this? ![backup](https://cloud.githubusercontent.com/assets/10379403/10480075/80a6f9e0-7268-11e5-9381-f4aa8a397f25.png)
1.0
How to reset backup - I tried the backup option in 2.2 and it was hanging again. Even updated to 2.3 but still the same screen: Any idea how to reset this? ![backup](https://cloud.githubusercontent.com/assets/10379403/10480075/80a6f9e0-7268-11e5-9381-f4aa8a397f25.png)
priority
how to reset backup i tried the backup option in and it was hanging again even updated to but still the same screen any idea how to reset this
1
523,219
15,175,711,590
IssuesEvent
2021-02-14 00:40:13
nischalshrestha/DataTutor
https://api.github.com/repos/nischalshrestha/DataTutor
closed
Stepper: make it work for R too.
priority
Upon detection of language engine, we should have the same functionality of parse/execute/store outputs, and stepping to work with R as well. It could also be easier to do this because R is easier to introspect and no worries about output rendering as much.
1.0
Stepper: make it work for R too. - Upon detection of language engine, we should have the same functionality of parse/execute/store outputs, and stepping to work with R as well. It could also be easier to do this because R is easier to introspect and no worries about output rendering as much.
priority
stepper make it work for r too upon detection of language engine we should have the same functionality of parse execute store outputs and stepping to work with r as well it could also be easier to do this because r is easier to introspect and no worries about output rendering as much
1
309,516
23,297,904,423
IssuesEvent
2022-08-06 21:55:31
MattTheLegoman/RealmsInExile
https://api.github.com/repos/MattTheLegoman/RealmsInExile
opened
Suggestion: More forms of Executions
documentation help wanted localization lore suggestion research
Some people in discussion mentioned recently and i've heard it before that a lot of our executions tends to be by burning at the stakes for most faiths. This is something we likely we can change and do more flavorful. I'll keep this as a suggestion if someone is interested. I guess this is most likely a localization heavy issue, but research into lore and so on is still needed.
1.0
Suggestion: More forms of Executions - Some people in discussion mentioned recently and i've heard it before that a lot of our executions tends to be by burning at the stakes for most faiths. This is something we likely we can change and do more flavorful. I'll keep this as a suggestion if someone is interested. I guess this is most likely a localization heavy issue, but research into lore and so on is still needed.
non_priority
suggestion more forms of executions some people in discussion mentioned recently and i ve heard it before that a lot of our executions tends to be by burning at the stakes for most faiths this is something we likely we can change and do more flavorful i ll keep this as a suggestion if someone is interested i guess this is most likely a localization heavy issue but research into lore and so on is still needed
0
349,026
24,931,147,096
IssuesEvent
2022-10-31 11:44:09
ToolJet/ToolJet
https://api.github.com/repos/ToolJet/ToolJet
closed
[docs]: add image to button group docs
documentation good first issue hacktoberfest up-for-grabs
### Summary There is no image of the button group widget. Add the following image in the button group doc below the description: ![Screenshot 2022-10-27 at 3 10 44 PM](https://user-images.githubusercontent.com/25361949/198250963-c29a2806-6c5e-465a-8942-772b7e0c526d.png) ### Issue Type Documentation bug ### The entire URL of the documentation with the issue https://docs.tooljet.com/docs/widgets/button-group ### Steps to reproduce the issue _No response_ ### Additional Information _No response_ ### If the issue is confirmed, would you be willing to submit a pull request? No ### Code of Conduct - [X] I agree to follow the ToolJet Code of Conduct
1.0
[docs]: add image to button group docs - ### Summary There is no image of the button group widget. Add the following image in the button group doc below the description: ![Screenshot 2022-10-27 at 3 10 44 PM](https://user-images.githubusercontent.com/25361949/198250963-c29a2806-6c5e-465a-8942-772b7e0c526d.png) ### Issue Type Documentation bug ### The entire URL of the documentation with the issue https://docs.tooljet.com/docs/widgets/button-group ### Steps to reproduce the issue _No response_ ### Additional Information _No response_ ### If the issue is confirmed, would you be willing to submit a pull request? No ### Code of Conduct - [X] I agree to follow the ToolJet Code of Conduct
non_priority
add image to button group docs summary there is no image of the button group widget add the following image in the button group doc below the description issue type documentation bug the entire url of the documentation with the issue steps to reproduce the issue no response additional information no response if the issue is confirmed would you be willing to submit a pull request no code of conduct i agree to follow the tooljet code of conduct
0
211,519
16,448,252,862
IssuesEvent
2021-05-20 23:02:23
insertcode/insertcode-website
https://api.github.com/repos/insertcode/insertcode-website
closed
Mejorar el README
documentation good first issue
Sería bueno incluir al menos: - Instrucciones para desarrollo local - Stack actual - Instrucciones para contribuir
1.0
Mejorar el README - Sería bueno incluir al menos: - Instrucciones para desarrollo local - Stack actual - Instrucciones para contribuir
non_priority
mejorar el readme sería bueno incluir al menos instrucciones para desarrollo local stack actual instrucciones para contribuir
0
51,554
3,013,031,880
IssuesEvent
2015-07-29 05:35:17
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Pagination missing from library packages
bug priority
When you click through to view the packages in the library you only see 30, and there is no pagination to see more. Confirmed on demo which has approx 1700 packages but I can only see 30 of each. Source: http://help.octopusdeploy.com/discussions/problems/36342
1.0
Pagination missing from library packages - When you click through to view the packages in the library you only see 30, and there is no pagination to see more. Confirmed on demo which has approx 1700 packages but I can only see 30 of each. Source: http://help.octopusdeploy.com/discussions/problems/36342
priority
pagination missing from library packages when you click through to view the packages in the library you only see and there is no pagination to see more confirmed on demo which has approx packages but i can only see of each source
1
539,294
15,786,562,035
IssuesEvent
2021-04-01 17:55:50
code4romania/de-urgenta-android
https://api.github.com/repos/code4romania/de-urgenta-android
opened
Implement proper error handling mechanism
high-priority :fire:
Implement an easy to use error handling mechanism. Should support displaying snackbars and dialogs. Check the [design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530) for hints on how to design them, but follow the android material design guidelines for them. These should be displayed easily in case of validation or api errors.
1.0
Implement proper error handling mechanism - Implement an easy to use error handling mechanism. Should support displaying snackbars and dialogs. Check the [design in Figma](https://www.figma.com/file/d6n3MPVRWbL5jP2NmvW96b/De_Urgenta?node-id=1466%3A530) for hints on how to design them, but follow the android material design guidelines for them. These should be displayed easily in case of validation or api errors.
priority
implement proper error handling mechanism implement an easy to use error handling mechanism should support displaying snackbars and dialogs check the for hints on how to design them but follow the android material design guidelines for them these should be displayed easily in case of validation or api errors
1
238,432
26,107,631,719
IssuesEvent
2022-12-27 15:10:51
elastic/cloudbeat
https://api.github.com/repos/elastic/cloudbeat
opened
Launcher wrapper layer and the validator abstraction
Team:Cloud Security Posture
**Motivation** Today our `cloudbeat` is wrapped by a `launcher` component that manages the reconfigurations and runs the beat itself. This is different from other beats that do not have this wrapper logic and is also not part of the `beat` development generation code. The goal of the `launcher` was to be agnostic to the underlying managed `beat` and to eventually be transformed into the `beat` development generation code. Currently, this wrapper is being used only by our beat and there are no future plans for it to change (not that I'm aware of as of right now). Part of the `launcher` includes the implementation of an abstraction layer of `validator` that has the purpose of validating configurations for the `beat` instance. Our current implementation of the `validator` just calls the `config` constructor and doesn't do anything besides that. We should discuss if we still want to have this whole abstraction layer or if it is redundant. While working on this PR: - https://github.com/elastic/cloudbeat/pull/597 I've had multiple commits regarding different ways how to tackle this configuration issue. We've decided to keep the `validator` and leave the checks inside the `config` for now. **Definition of done** What needs to be completed at the end of this task - [] Decide the fate of the abstraction and act accordingly - [] If we decide to leave it then create a plan/roadmap to utilize it and design everything accordingly
True
Launcher wrapper layer and the validator abstraction - **Motivation** Today our `cloudbeat` is wrapped by a `launcher` component that manages the reconfigurations and runs the beat itself. This is different from other beats that do not have this wrapper logic and is also not part of the `beat` development generation code. The goal of the `launcher` was to be agnostic to the underlying managed `beat` and to eventually be transformed into the `beat` development generation code. Currently, this wrapper is being used only by our beat and there are no future plans for it to change (not that I'm aware of as of right now). Part of the `launcher` includes the implementation of an abstraction layer of `validator` that has the purpose of validating configurations for the `beat` instance. Our current implementation of the `validator` just calls the `config` constructor and doesn't do anything besides that. We should discuss if we still want to have this whole abstraction layer or if it is redundant. While working on this PR: - https://github.com/elastic/cloudbeat/pull/597 I've had multiple commits regarding different ways how to tackle this configuration issue. We've decided to keep the `validator` and leave the checks inside the `config` for now. **Definition of done** What needs to be completed at the end of this task - [] Decide the fate of the abstraction and act accordingly - [] If we decide to leave it then create a plan/roadmap to utilize it and design everything accordingly
non_priority
launcher wrapper layer and the validator abstraction motivation today our cloudbeat is wrapped by a launcher component that manages the reconfigurations and runs the beat itself this is different from other beats that do not have this wrapper logic and is also not part of the beat development generation code the goal of the launcher was to be agnostic to the underlying managed beat and to eventually be transformed into the beat development generation code currently this wrapper is being used only by our beat and there are no future plans for it to change not that i m aware of as of right now part of the launcher includes the implementation of an abstraction layer of validator that has the purpose of validating configurations for the beat instance our current implementation of the validator just calls the config constructor and doesn t do anything besides that we should discuss if we still want to have this whole abstraction layer or if it is redundant while working on this pr i ve had multiple commits regarding different ways how to tackle this configuration issue we ve decided to keep the validator and leave the checks inside the config for now definition of done what needs to be completed at the end of this task decide the fate of the abstraction and act accordingly if we decide to leave it then create a plan roadmap to utilize it and design everything accordingly
0
9,891
12,890,275,957
IssuesEvent
2020-07-13 15:45:05
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Token for the current date seems missing
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
Please also include a token for the current day. We already have a token for yyyymmdd. I don't know if its not supported or missing in documentation but we needed a token for just the day (dd) component of the date. [Enter feedback here] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Token for the current date seems missing - Please also include a token for the current day. We already have a token for yyyymmdd. I don't know if its not supported or missing in documentation but we needed a token for just the day (dd) component of the date. [Enter feedback here] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
non_priority
token for the current date seems missing please also include a token for the current day we already have a token for yyyymmdd i don t know if its not supported or missing in documentation but we needed a token for just the day dd component of the date document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
0
240,038
19,988,375,057
IssuesEvent
2022-01-31 00:43:03
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Tech Lead Java @ Mollica IT
CLT PJ Java Remoto Especialista Exterior Testes automatizados Alocado Redis SQL Git RabbitMQ Stale
## Nossa empresa Mais do que uma consultoria, temos uma solução de Recrutamento & Seleção, sendo uma conexão que une profissionais adequados com diversos clientes no Brasil e exterior. Somos mais que uma Consultoria, somos seus Parceiros! A Mollica conta com um DNA inovador, somos especialistas no recrutamento de profissionais de tecnologias, especializados em soluções de TI. OBS: Estamos trabalhando essa posição para a Plus IT com alocação no Banco Digio. ## Descrição da vaga Quer trabalhar em um banco digital, uma bantech que conta com a solidez de um banco e agilidade de uma fintech? Então vem com a gente e se liga só nessa oportunidade: Somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um Tech Lead com foco em Java. Qual será a sua missão: Definir e entender requisitos técnicos e funcionais com a equipe; Planejar e desenhar detalhes de componentes e serviços; Codificar, testar e publicar serviços; Participar de code e design review; ## Local Remoto - aceitamos de candidatos todo o Brasil ## Requisitos **Obrigatórios:** - Ensino superior completo (ou em andamento) em tecnologia da informação; - Mínimo de 7 anos de experiência com Desenvolvimento Web; - Conhecimentos avançados na Stack JAVA e seus principais frameworks; - Conhecimentos avançados em SQL Server (estrutura de dados, views e procedures); - Conhecimentos avançados no controle de versão Git; - Conhecimento de boas práticas de desenvolvimento de software, incluindo padrões de desenvolvimento, documentação, code reviews, controle de versão e testes ; **Diferenciais:** - Inglês intermediário para comunicação técnica com equipes de outros países; - Práticas de testes automatizados; - Conhecimento em arquitetura micro-serviços e soluções de cache (Redis) e mensageria (RabbitMQ); - Habilidade em lidar com problemas ambíguos/indefinidos e pensar de forma abstrata; ## Benefícios Benefícios para CLT: ✔️ Vale Refeição ✔️ Assistência Médica ✔️ Assistência Odontológica ✔️Seguro de vida ✔️Auxilio Creche Benefícios para PJ: ✔️ Férias remuneradas ## Contratação PJ ou CLT (a escolha do candidato) ## Como se candidatar Por favor envie um e-mail para thais@mollicait.com com seu CV anexado ou candidatar-se através do link: https://jobs.quickin.io/mollicait/jobs/619e85f74af9e9001cde5a57 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 05 dias após cada processo. E-mail para contato em caso de não haver resposta: thais@mollicait.com ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Alocado - Remoto #### Regime - CLT - PJ #### Nível - Sênior - Especialista
1.0
[Remoto] Tech Lead Java @ Mollica IT - ## Nossa empresa Mais do que uma consultoria, temos uma solução de Recrutamento & Seleção, sendo uma conexão que une profissionais adequados com diversos clientes no Brasil e exterior. Somos mais que uma Consultoria, somos seus Parceiros! A Mollica conta com um DNA inovador, somos especialistas no recrutamento de profissionais de tecnologias, especializados em soluções de TI. OBS: Estamos trabalhando essa posição para a Plus IT com alocação no Banco Digio. ## Descrição da vaga Quer trabalhar em um banco digital, uma bantech que conta com a solidez de um banco e agilidade de uma fintech? Então vem com a gente e se liga só nessa oportunidade: Somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um Tech Lead com foco em Java. Qual será a sua missão: Definir e entender requisitos técnicos e funcionais com a equipe; Planejar e desenhar detalhes de componentes e serviços; Codificar, testar e publicar serviços; Participar de code e design review; ## Local Remoto - aceitamos de candidatos todo o Brasil ## Requisitos **Obrigatórios:** - Ensino superior completo (ou em andamento) em tecnologia da informação; - Mínimo de 7 anos de experiência com Desenvolvimento Web; - Conhecimentos avançados na Stack JAVA e seus principais frameworks; - Conhecimentos avançados em SQL Server (estrutura de dados, views e procedures); - Conhecimentos avançados no controle de versão Git; - Conhecimento de boas práticas de desenvolvimento de software, incluindo padrões de desenvolvimento, documentação, code reviews, controle de versão e testes ; **Diferenciais:** - Inglês intermediário para comunicação técnica com equipes de outros países; - Práticas de testes automatizados; - Conhecimento em arquitetura micro-serviços e soluções de cache (Redis) e mensageria (RabbitMQ); - Habilidade em lidar com problemas ambíguos/indefinidos e pensar de forma abstrata; ## Benefícios Benefícios para CLT: ✔️ Vale Refeição ✔️ Assistência Médica ✔️ Assistência Odontológica ✔️Seguro de vida ✔️Auxilio Creche Benefícios para PJ: ✔️ Férias remuneradas ## Contratação PJ ou CLT (a escolha do candidato) ## Como se candidatar Por favor envie um e-mail para thais@mollicait.com com seu CV anexado ou candidatar-se através do link: https://jobs.quickin.io/mollicait/jobs/619e85f74af9e9001cde5a57 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 05 dias após cada processo. E-mail para contato em caso de não haver resposta: thais@mollicait.com ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Alocado - Remoto #### Regime - CLT - PJ #### Nível - Sênior - Especialista
non_priority
tech lead java mollica it nossa empresa mais do que uma consultoria temos uma solução de recrutamento seleção sendo uma conexão que une profissionais adequados com diversos clientes no brasil e exterior somos mais que uma consultoria somos seus parceiros a mollica conta com um dna inovador somos especialistas no recrutamento de profissionais de tecnologias especializados em soluções de ti obs estamos trabalhando essa posição para a plus it com alocação no banco digio descrição da vaga quer trabalhar em um banco digital uma bantech que conta com a solidez de um banco e agilidade de uma fintech então vem com a gente e se liga só nessa oportunidade somos uma plataforma digital de serviços financeiros com múltiplos produtos e serviços para consumidores e estamos em busca de um tech lead com foco em java qual será a sua missão definir e entender requisitos técnicos e funcionais com a equipe planejar e desenhar detalhes de componentes e serviços codificar testar e publicar serviços participar de code e design review local remoto aceitamos de candidatos todo o brasil requisitos obrigatórios ensino superior completo ou em andamento em tecnologia da informação mínimo de anos de experiência com desenvolvimento web conhecimentos avançados na stack java e seus principais frameworks conhecimentos avançados em sql server estrutura de dados views e procedures conhecimentos avançados no controle de versão git conhecimento de boas práticas de desenvolvimento de software incluindo padrões de desenvolvimento documentação code reviews controle de versão e testes diferenciais inglês intermediário para comunicação técnica com equipes de outros países práticas de testes automatizados conhecimento em arquitetura micro serviços e soluções de cache redis e mensageria rabbitmq habilidade em lidar com problemas ambíguos indefinidos e pensar de forma abstrata benefícios benefícios para clt ✔️ vale refeição ✔️ assistência médica ✔️ assistência odontológica ✔️seguro de vida ✔️auxilio creche benefícios para pj ✔️ férias remuneradas contratação pj ou clt a escolha do candidato como se candidatar por favor envie um e mail para thais mollicait com com seu cv anexado ou candidatar se através do link tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta thais mollicait com labels alocação alocado remoto regime clt pj nível sênior especialista
0
101,613
11,253,391,442
IssuesEvent
2020-01-11 15:58:40
vuepressjs/vuepress-theme-blog
https://api.github.com/repos/vuepressjs/vuepress-theme-blog
closed
Improving Documentation
documentation
A lot of the options and features are not properly explained in the documentation. It would serve the project better if we had more explicit and detailed documentation. Happy to take thison this issue.
1.0
Improving Documentation - A lot of the options and features are not properly explained in the documentation. It would serve the project better if we had more explicit and detailed documentation. Happy to take thison this issue.
non_priority
improving documentation a lot of the options and features are not properly explained in the documentation it would serve the project better if we had more explicit and detailed documentation happy to take thison this issue
0
226,724
7,522,532,874
IssuesEvent
2018-04-12 20:45:12
runconduit/conduit
https://api.github.com/repos/runconduit/conduit
closed
Public-api should query Prometheus directly
area/controller area/telemetry priority/P0
Since the telemetry service is going away, the public-api service should be updated to query Prometheus directly. The metrics available are described in #425 and #428. The result should be that all data in the Conduit web dashboard and in the Conduit CLI stat command are provided by these new metrics.
1.0
Public-api should query Prometheus directly - Since the telemetry service is going away, the public-api service should be updated to query Prometheus directly. The metrics available are described in #425 and #428. The result should be that all data in the Conduit web dashboard and in the Conduit CLI stat command are provided by these new metrics.
priority
public api should query prometheus directly since the telemetry service is going away the public api service should be updated to query prometheus directly the metrics available are described in and the result should be that all data in the conduit web dashboard and in the conduit cli stat command are provided by these new metrics
1
29,588
2,716,608,483
IssuesEvent
2015-04-10 20:12:18
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
Error on JSON serialization (used on REST services or client DB transactions) when Class has an internal parameterized type
bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.3
_From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 20, 2014 16:28:25_ Error when tryong to create a JSONSerializer for types like: public class MyClass{ private MyOtherClass<String> a; } _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=475_
1.0
Error on JSON serialization (used on REST services or client DB transactions) when Class has an internal parameterized type - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 20, 2014 16:28:25_ Error when tryong to create a JSONSerializer for types like: public class MyClass{ private MyOtherClass<String> a; } _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=475_
priority
error on json serialization used on rest services or client db transactions when class has an internal parameterized type from on august error when tryong to create a jsonserializer for types like public class myclass private myotherclass a original issue
1
130,695
5,120,219,969
IssuesEvent
2017-01-09 01:32:05
geekgirlbeta/linkapp
https://api.github.com/repos/geekgirlbeta/linkapp
opened
Deleting Links
Priority: High Status: Needs Research Type: Feature
Currently, there is no way for the user to delete a Link. We really won't want them to do this - we want our URLs to persist indefinitely. We need to allow the user to delete a post, but in a way that preserves the URLs.
1.0
Deleting Links - Currently, there is no way for the user to delete a Link. We really won't want them to do this - we want our URLs to persist indefinitely. We need to allow the user to delete a post, but in a way that preserves the URLs.
priority
deleting links currently there is no way for the user to delete a link we really won t want them to do this we want our urls to persist indefinitely we need to allow the user to delete a post but in a way that preserves the urls
1
248,216
7,928,274,384
IssuesEvent
2018-07-06 11:00:59
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Debug hit highlighting should go away when the flow is finish
Component/Composer Imported Priority/High Type/Bug
<a href="https://github.com/yasassri"><img src="https://avatars1.githubusercontent.com/u/7681361?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [yasassri](https://github.com/yasassri)** _Sunday Nov 05, 2017 at 09:54 GMT_ _Originally opened as https://github.com/ballerina-lang/composer/issues/4464_ ---- This happens in split view. ![screenshot from 2017-11-05 15-21-20](https://user-images.githubusercontent.com/7681361/32413701-4b1daa22-c23d-11e7-8e05-2c9713145efe.png)
1.0
Debug hit highlighting should go away when the flow is finish - <a href="https://github.com/yasassri"><img src="https://avatars1.githubusercontent.com/u/7681361?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [yasassri](https://github.com/yasassri)** _Sunday Nov 05, 2017 at 09:54 GMT_ _Originally opened as https://github.com/ballerina-lang/composer/issues/4464_ ---- This happens in split view. ![screenshot from 2017-11-05 15-21-20](https://user-images.githubusercontent.com/7681361/32413701-4b1daa22-c23d-11e7-8e05-2c9713145efe.png)
priority
debug hit highlighting should go away when the flow is finish issue by sunday nov at gmt originally opened as this happens in split view
1
667,787
22,500,379,733
IssuesEvent
2022-06-23 11:16:19
ita-social-projects/horondi_client_fe
https://api.github.com/repos/ita-social-projects/horondi_client_fe
closed
[Product Details Page] Incorrect vertical alignment of currency icon on Product Details Page
bug UI FrontEnd part priority: medium severity: trivial
**Environment** Windows 10, Chrome Version 102.0.5005.63 **Reproducible:** Always **Preconditions:** Go to https://horondi-front-staging.azurewebsites.net/ **Steps to reproduce:** Open Catalog Click on any product to open Product Details Page Pay attention to currency icon **Actual result:** Currency icon is higher than price ![2022-06-06_15h46_51](https://user-images.githubusercontent.com/33963679/172163343-925bd4b7-85b7-4615-bfaa-8acd49d321b3.png) **Expected result:** Currency icon and price should be at the same level
1.0
[Product Details Page] Incorrect vertical alignment of currency icon on Product Details Page - **Environment** Windows 10, Chrome Version 102.0.5005.63 **Reproducible:** Always **Preconditions:** Go to https://horondi-front-staging.azurewebsites.net/ **Steps to reproduce:** Open Catalog Click on any product to open Product Details Page Pay attention to currency icon **Actual result:** Currency icon is higher than price ![2022-06-06_15h46_51](https://user-images.githubusercontent.com/33963679/172163343-925bd4b7-85b7-4615-bfaa-8acd49d321b3.png) **Expected result:** Currency icon and price should be at the same level
priority
incorrect vertical alignment of currency icon on product details page environment windows chrome version reproducible always preconditions go to steps to reproduce open catalog click on any product to open product details page pay attention to currency icon actual result currency icon is higher than price expected result currency icon and price should be at the same level
1
759,985
26,621,834,415
IssuesEvent
2023-01-24 11:50:41
ethereum-push-notification-service/push-dapp
https://api.github.com/repos/ethereum-push-notification-service/push-dapp
closed
Delegate should be in CAIP10 format to ensure people understand
enhancement Priority 1
Rationale: current flow doesn't point out the delegations blockchain and since it's triggered and guided by the individual smart contract on that chain, it needs to be made clear for the user. <img width="1810" alt="Screenshot 2022-09-04 at 8 47 11 PM" src="https://user-images.githubusercontent.com/22663032/188324401-029b85b8-8127-4564-95f8-7e19498bc5c0.png">
1.0
Delegate should be in CAIP10 format to ensure people understand - Rationale: current flow doesn't point out the delegations blockchain and since it's triggered and guided by the individual smart contract on that chain, it needs to be made clear for the user. <img width="1810" alt="Screenshot 2022-09-04 at 8 47 11 PM" src="https://user-images.githubusercontent.com/22663032/188324401-029b85b8-8127-4564-95f8-7e19498bc5c0.png">
priority
delegate should be in format to ensure people understand rationale current flow doesn t point out the delegations blockchain and since it s triggered and guided by the individual smart contract on that chain it needs to be made clear for the user img width alt screenshot at pm src
1
45,620
5,722,505,801
IssuesEvent
2017-04-20 09:39:29
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
ClientReconnectTest.testReconnectToNewInstanceAtSameAddress
Team: Client Type: Test-Failure
https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/8400/ ``` java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds, count left: 1 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:899) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:892) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:884) at com.hazelcast.client.ClientReconnectTest.testReconnectToNewInstanceAtSameAddress(ClientReconnectTest.java:95) ```
1.0
ClientReconnectTest.testReconnectToNewInstanceAtSameAddress - https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/8400/ ``` java.lang.AssertionError: CountDownLatch failed to complete within 120 seconds, count left: 1 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:899) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:892) at com.hazelcast.test.HazelcastTestSupport.assertOpenEventually(HazelcastTestSupport.java:884) at com.hazelcast.client.ClientReconnectTest.testReconnectToNewInstanceAtSameAddress(ClientReconnectTest.java:95) ```
non_priority
clientreconnecttest testreconnecttonewinstanceatsameaddress java lang assertionerror countdownlatch failed to complete within seconds count left at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport assertopeneventually hazelcasttestsupport java at com hazelcast client clientreconnecttest testreconnecttonewinstanceatsameaddress clientreconnecttest java
0
460,485
13,210,768,437
IssuesEvent
2020-08-15 18:44:49
radiorabe/klangbecken
https://api.github.com/repos/radiorabe/klangbecken
closed
Counter for skipped tracks is not reset when a track was ok
bug priority/1
This track should not have been played, because it there was an accepted track in the meantime. ``` [root@vm-0016 ~]# journalctl --unit liquidsoap@klangbecken.service --since "2020-08-14 11:30" | grep -E "track was recently played: |too many skipped tracks" Aug 14 11:47:01 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 11:47:01 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/db378175-5d46-488a-b1ea-20bc7f56ae32.mp3 (128685 seconds ago) Aug 14 12:02:26 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:02:26 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/ea7f0b81-8316-4f59-9fa8-d8e5945c6aed.mp3 (27725 seconds ago) Aug 14 12:09:04 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:09:04 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/9c0e36cc-7798-42c0-883c-6129158506f9.mp3 (25581 seconds ago) Aug 14 12:12:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:12:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/a6683e6d-6def-45cb-b3b3-8b00be4eea6d.mp3 (238572 seconds ago) Aug 14 12:26:38 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:38 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/28200f9a-4704-44ba-ba9e-11a76b045b24.mp3 (69360 seconds ago) Aug 14 12:26:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/4ba09c6a-e922-4b7f-a999-163306f3cd75.mp3 (63754 seconds ago) Aug 14 12:47:25 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:47:25 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/0c8782db-e9c2-4447-92f6-207510c44194.mp3 (163224 seconds ago) Aug 14 12:50:16 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:50:16 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/c38972ed-af1f-4928-a266-838b53ef519c.mp3 (172108 seconds ago) Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 (172703 seconds ago) Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] too many skipped tracks, playing /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 anyway ```
1.0
Counter for skipped tracks is not reset when a track was ok - This track should not have been played, because it there was an accepted track in the meantime. ``` [root@vm-0016 ~]# journalctl --unit liquidsoap@klangbecken.service --since "2020-08-14 11:30" | grep -E "track was recently played: |too many skipped tracks" Aug 14 11:47:01 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 11:47:01 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/db378175-5d46-488a-b1ea-20bc7f56ae32.mp3 (128685 seconds ago) Aug 14 12:02:26 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:02:26 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/ea7f0b81-8316-4f59-9fa8-d8e5945c6aed.mp3 (27725 seconds ago) Aug 14 12:09:04 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:09:04 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/9c0e36cc-7798-42c0-883c-6129158506f9.mp3 (25581 seconds ago) Aug 14 12:12:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:12:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/a6683e6d-6def-45cb-b3b3-8b00be4eea6d.mp3 (238572 seconds ago) Aug 14 12:26:38 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:38 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/28200f9a-4704-44ba-ba9e-11a76b045b24.mp3 (69360 seconds ago) Aug 14 12:26:39 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:26:39 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/4ba09c6a-e922-4b7f-a999-163306f3cd75.mp3 (63754 seconds ago) Aug 14 12:47:25 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:47:25 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/0c8782db-e9c2-4447-92f6-207510c44194.mp3 (163224 seconds ago) Aug 14 12:50:16 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 12:50:16 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/c38972ed-af1f-4928-a266-838b53ef519c.mp3 (172108 seconds ago) Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] track was recently played: /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 (172703 seconds ago) Aug 14 14:14:50 vm-0016.audio.int.rabe.ch liquidsoap[11066]: 2020/08/14 14:14:50 [check_next_func:3] too many skipped tracks, playing /var/lib/klangbecken/music/36bd58bd-c562-49a3-a034-e8a5c83f18d8.mp3 anyway ```
priority
counter for skipped tracks is not reset when a track was ok this track should not have been played because it there was an accepted track in the meantime journalctl unit liquidsoap klangbecken service since grep e track was recently played too many skipped tracks aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap track was recently played var lib klangbecken music seconds ago aug vm audio int rabe ch liquidsoap too many skipped tracks playing var lib klangbecken music anyway
1
68,553
9,200,329,705
IssuesEvent
2019-03-07 16:49:01
terraform-providers/terraform-provider-google
https://api.github.com/repos/terraform-providers/terraform-provider-google
closed
google_container_cluster: network_policy_config defaults to disabled
documentation
The docs at [1] imply that network_policy_config defaults to enabled and can be disabled. In fact, it's the opposite. [1] https://www.terraform.io/docs/providers/google/r/container_cluster.html#network_policy_config ### Terraform Version ``` $ terraform -v Terraform v0.11.7 + provider.google v1.15.0 ```
1.0
google_container_cluster: network_policy_config defaults to disabled - The docs at [1] imply that network_policy_config defaults to enabled and can be disabled. In fact, it's the opposite. [1] https://www.terraform.io/docs/providers/google/r/container_cluster.html#network_policy_config ### Terraform Version ``` $ terraform -v Terraform v0.11.7 + provider.google v1.15.0 ```
non_priority
google container cluster network policy config defaults to disabled the docs at imply that network policy config defaults to enabled and can be disabled in fact it s the opposite terraform version terraform v terraform provider google
0
188,515
6,777,155,107
IssuesEvent
2017-10-27 20:50:53
buttercup/buttercup-mobile
https://api.github.com/repos/buttercup/buttercup-mobile
closed
Add padding to home page right-icon
Effort: Low Priority: High Status: Completed Type: Enhancement
For better touch area, add padding to the left side of the right-icon.
1.0
Add padding to home page right-icon - For better touch area, add padding to the left side of the right-icon.
priority
add padding to home page right icon for better touch area add padding to the left side of the right icon
1
92,767
8,377,788,214
IssuesEvent
2018-10-06 05:58:13
snext1220/stext
https://api.github.com/repos/snext1220/stext
closed
「七惑星の欠片」をシナリオ側で自動加算させる機能
Testing enhancement
「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+4する事」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。 (『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした) 戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。 (減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています)
1.0
「七惑星の欠片」をシナリオ側で自動加算させる機能 - 「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+4する事」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。 (『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした) 戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。 (減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています)
non_priority
「七惑星の欠片」をシナリオ側で自動加算させる機能 「戦国ソーサリアン外伝・上杉謙信の章」にて「全ての星を+ 」と指示がありますが、繰り返してプレイする際に、手動で<七惑星の欠片>を加算するのがやや手間に感じられました。 (『少女は英雄を語りて』でも同様の指示があり、こちらも手間でした) 戦闘時にドロップする<七惑星の欠片>は、自動加算されなくても良いと思うのですが、シナリオ中で<七惑星の欠片>の加算が指示されている場合、自動的に加算される機能があると便利だと思います。 (減算については、欠片がないのに減算処理→エラー発生となるような気がするので、今まで通り手動が望ましいと思っています)
0
125,483
10,344,625,754
IssuesEvent
2019-09-04 11:42:07
ansgohar/urbanharmony
https://api.github.com/repos/ansgohar/urbanharmony
closed
FrontEnd -In الفائزون في المسابقات , "المرصد الحضاري" " and "صحافة"- The picture is displayed as black although it was uploaded
Needs Re-testing State: Pending Testing bug
ندوات و مسابقات1. Then الفائزون في المسابقات The picture is displayed as black although it was uploaded as below ![image](https://user-images.githubusercontent.com/49487834/55976414-691a0300-5c8c-11e9-839f-198d89d69456.png) 2. المرصد الحضاري ![image](https://user-images.githubusercontent.com/49487834/56093645-87217680-5ecb-11e9-9dac-de6eda1b4437.png) 3. اخر الاخبار Then صحافة ![image](https://user-images.githubusercontent.com/49487834/56131292-53a42200-5f87-11e9-967d-07a53c9aa7ea.png)
2.0
FrontEnd -In الفائزون في المسابقات , "المرصد الحضاري" " and "صحافة"- The picture is displayed as black although it was uploaded - ندوات و مسابقات1. Then الفائزون في المسابقات The picture is displayed as black although it was uploaded as below ![image](https://user-images.githubusercontent.com/49487834/55976414-691a0300-5c8c-11e9-839f-198d89d69456.png) 2. المرصد الحضاري ![image](https://user-images.githubusercontent.com/49487834/56093645-87217680-5ecb-11e9-9dac-de6eda1b4437.png) 3. اخر الاخبار Then صحافة ![image](https://user-images.githubusercontent.com/49487834/56131292-53a42200-5f87-11e9-967d-07a53c9aa7ea.png)
non_priority
frontend in الفائزون في المسابقات المرصد الحضاري and صحافة the picture is displayed as black although it was uploaded ندوات و then الفائزون في المسابقات the picture is displayed as black although it was uploaded as below المرصد الحضاري اخر الاخبار then صحافة
0
283,117
24,525,354,088
IssuesEvent
2022-10-11 12:45:37
ns1labs/orb
https://api.github.com/repos/ns1labs/orb
opened
Tests: create automated tests to validate UI filters
test
Agents: - name [ ] - tags [ ] - status [ ] -policies [ ] Groups: - Name [ ] - Tags [ ] Policies: - Name [ ] - Description [ ] - Tags [ ] - Version [ ] Sinks: - Name - Tags - Status - Backend
1.0
Tests: create automated tests to validate UI filters - Agents: - name [ ] - tags [ ] - status [ ] -policies [ ] Groups: - Name [ ] - Tags [ ] Policies: - Name [ ] - Description [ ] - Tags [ ] - Version [ ] Sinks: - Name - Tags - Status - Backend
non_priority
tests create automated tests to validate ui filters agents name tags status policies groups name tags policies name description tags version sinks name tags status backend
0
176,206
6,557,270,117
IssuesEvent
2017-09-06 16:47:44
openshift/origin
https://api.github.com/repos/openshift/origin
closed
OpenShift hitting external kubernetes for internal OpenShift resources
area/security component/auth priority/P2
Running against external kube version: ``` kubectl version Client Version: version.Info{Major:"1", Minor:"3+", GitVersion:"v1.3.0-alpha.5-dirty", GitCommit:"914163247d9a16b46921e83f7dbedb572859b3e4", GitTreeState:"dirty", BuildDate:"2016-06-14T18:43:00Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.5", GitCommit:"b0deb2eb8f4037421077f77cb163dbb4c0a2a9f5", GitTreeState:"clean", BuildDate:"2016-08-11T20:21:58Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"} ``` SecurityContextConstraint does not exist in Kubernetes API, I think. So it's trying to hit external kube for something that is openshift specific.
1.0
OpenShift hitting external kubernetes for internal OpenShift resources - Running against external kube version: ``` kubectl version Client Version: version.Info{Major:"1", Minor:"3+", GitVersion:"v1.3.0-alpha.5-dirty", GitCommit:"914163247d9a16b46921e83f7dbedb572859b3e4", GitTreeState:"dirty", BuildDate:"2016-06-14T18:43:00Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.5", GitCommit:"b0deb2eb8f4037421077f77cb163dbb4c0a2a9f5", GitTreeState:"clean", BuildDate:"2016-08-11T20:21:58Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"} ``` SecurityContextConstraint does not exist in Kubernetes API, I think. So it's trying to hit external kube for something that is openshift specific.
priority
openshift hitting external kubernetes for internal openshift resources running against external kube version kubectl version client version version info major minor gitversion alpha dirty gitcommit gittreestate dirty builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux securitycontextconstraint does not exist in kubernetes api i think so it s trying to hit external kube for something that is openshift specific
1
373,908
11,052,371,223
IssuesEvent
2019-12-10 09:16:31
incognitochain/incognito-chain
https://api.github.com/repos/incognitochain/incognito-chain
opened
[Testnet] Panic by 'fatal error: concurrent map read and map write' on synker.go
Priority: High Type: Bug
fatal error: concurrent map read and map write goroutine 227 [running]: runtime.throw(0x13f372d, 0x21) /usr/local/go/src/runtime/panic.go:617 +0x72 fp=0xc001930b48 sp=0xc001930b18 pc=0x4301a2 runtime.mapaccess2(0x1202220, 0xc01782e270, 0xc002382dd0, 0xc002382dd0, 0x1b0) /usr/local/go/src/runtime/map.go:472 +0x284 fp=0xc001930b90 sp=0xc001930b48 pc=0x410684 reflect.mapaccess(0x1202220, 0xc01782e270, 0xc002382dd0, 0x13e38b4) /usr/local/go/src/runtime/map.go:1321 +0x3f fp=0xc001930bc8 sp=0xc001930b90 pc=0x4128df reflect.Value.MapIndex(0x1202220, 0xc000448eb0, 0x195, 0x119c5a0, 0xc002382dd0, 0x98, 0x119c820, 0xc0077e8d50, 0x8b) /usr/local/go/src/reflect/value.go:1112 +0x10a fp=0xc001930c28 sp=0xc001930bc8 pc=0x49d28a encoding/json.mapEncoder.encode(0x1568cc0, 0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100) /usr/local/go/src/encoding/json/encode.go:702 +0x303 fp=0xc001930d70 sp=0xc001930c28 pc=0x5cfe73 encoding/json.mapEncoder.encode-fm(0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100) /usr/local/go/src/encoding/json/encode.go:678 +0x64 fp=0xc001930db0 sp=0xc001930d70 pc=0x5db8e4 encoding/json.structEncoder.encode(0xc000584600, 0x17, 0x21, 0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0x5c0100) /usr/local/go/src/encoding/json/encode.go:660 +0x304 fp=0xc001930e50 sp=0xc001930db0 pc=0x5cf944 encoding/json.structEncoder.encode-fm(0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0xc001df0100) /usr/local/go/src/encoding/json/encode.go:631 +0x76 fp=0xc001930ea0 sp=0xc001930e50 pc=0x5db866 encoding/json.ptrEncoder.encode(0xc00067afc0, 0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x1170100) /usr/local/go/src/encoding/json/encode.go:806 +0xb2 fp=0xc001930ee8 sp=0xc001930ea0 pc=0x5d0dc2 encoding/json.ptrEncoder.encode-fm(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0xc001930100) /usr/local/go/src/encoding/json/encode.go:801 +0x64 fp=0xc001930f28 sp=0xc001930ee8 pc=0x5dba64 encoding/json.(*encodeState).reflectValue(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x2400100) /usr/local/go/src/encoding/json/encode.go:334 +0x82 fp=0xc001930f60 sp=0xc001930f28 pc=0x5cd112 encoding/json.(*encodeState).marshal(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x400100, 0x0, 0x0) /usr/local/go/src/encoding/json/encode.go:306 +0xee fp=0xc001930fa8 sp=0xc001930f60 pc=0x5ccbbe encoding/json.Marshal(0x1170020, 0xc001dfa430, 0x1, 0xf7218f3940a6636e, 0xc001931088, 0x4102df, 0xc001931106) /usr/local/go/src/encoding/json/encode.go:160 +0x52 fp=0xc001931020 sp=0xc001930fa8 pc=0x5cc1e2 github.com/incognitochain/incognito-chain/blockchain.(*ShardBestState).MarshalJSON(0xc000448dc0, 0xc0002d6870, 0xc001931106, 0xc000427a20, 0xc0077e8d40, 0x0) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/shardbeststate.go:244 +0x60 fp=0xc001931098 sp=0xc001931020 pc=0xc07100 github.com/incognitochain/incognito-chain/blockchain.(*Synker).UpdateStatev2(0xc000136628) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:719 +0x8a1 fp=0xc001931cf0 sp=0xc001931098 pc=0xc3e971 github.com/incognitochain/incognito-chain/blockchain.(*Synker).Start(0xc000136628) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:163 +0xa29 fp=0xc001931fd8 sp=0xc001931cf0 pc=0xc3d2c9 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1337 +0x1 fp=0xc001931fe0 sp=0xc001931fd8 pc=0x45f3a1 created by main.Server.Start /root/go/src/github.com/incognitochain/incognito-chain/server.go:733 +0x3a9 goroutine 1 [chan receive, 15 minutes]: main.mainMaster(0x0, 0x0, 0x0) /root/go/src/github.com/incognitochain/incognito-chain/incognito.go:132 +0x8eb main.main() /root/go/src/github.com/incognitochain/incognito-chain/incognito.go:178 +0x246
1.0
[Testnet] Panic by 'fatal error: concurrent map read and map write' on synker.go - fatal error: concurrent map read and map write goroutine 227 [running]: runtime.throw(0x13f372d, 0x21) /usr/local/go/src/runtime/panic.go:617 +0x72 fp=0xc001930b48 sp=0xc001930b18 pc=0x4301a2 runtime.mapaccess2(0x1202220, 0xc01782e270, 0xc002382dd0, 0xc002382dd0, 0x1b0) /usr/local/go/src/runtime/map.go:472 +0x284 fp=0xc001930b90 sp=0xc001930b48 pc=0x410684 reflect.mapaccess(0x1202220, 0xc01782e270, 0xc002382dd0, 0x13e38b4) /usr/local/go/src/runtime/map.go:1321 +0x3f fp=0xc001930bc8 sp=0xc001930b90 pc=0x4128df reflect.Value.MapIndex(0x1202220, 0xc000448eb0, 0x195, 0x119c5a0, 0xc002382dd0, 0x98, 0x119c820, 0xc0077e8d50, 0x8b) /usr/local/go/src/reflect/value.go:1112 +0x10a fp=0xc001930c28 sp=0xc001930bc8 pc=0x49d28a encoding/json.mapEncoder.encode(0x1568cc0, 0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100) /usr/local/go/src/encoding/json/encode.go:702 +0x303 fp=0xc001930d70 sp=0xc001930c28 pc=0x5cfe73 encoding/json.mapEncoder.encode-fm(0xc0082a33b0, 0x1202220, 0xc000448eb0, 0x195, 0x100) /usr/local/go/src/encoding/json/encode.go:678 +0x64 fp=0xc001930db0 sp=0xc001930d70 pc=0x5db8e4 encoding/json.structEncoder.encode(0xc000584600, 0x17, 0x21, 0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0x5c0100) /usr/local/go/src/encoding/json/encode.go:660 +0x304 fp=0xc001930e50 sp=0xc001930db0 pc=0x5cf944 encoding/json.structEncoder.encode-fm(0xc0082a33b0, 0x123afc0, 0xc001dfa430, 0x199, 0xc001df0100) /usr/local/go/src/encoding/json/encode.go:631 +0x76 fp=0xc001930ea0 sp=0xc001930e50 pc=0x5db866 encoding/json.ptrEncoder.encode(0xc00067afc0, 0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x1170100) /usr/local/go/src/encoding/json/encode.go:806 +0xb2 fp=0xc001930ee8 sp=0xc001930ea0 pc=0x5d0dc2 encoding/json.ptrEncoder.encode-fm(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0xc001930100) /usr/local/go/src/encoding/json/encode.go:801 +0x64 fp=0xc001930f28 sp=0xc001930ee8 pc=0x5dba64 encoding/json.(*encodeState).reflectValue(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x16, 0x2400100) /usr/local/go/src/encoding/json/encode.go:334 +0x82 fp=0xc001930f60 sp=0xc001930f28 pc=0x5cd112 encoding/json.(*encodeState).marshal(0xc0082a33b0, 0x1170020, 0xc001dfa430, 0x400100, 0x0, 0x0) /usr/local/go/src/encoding/json/encode.go:306 +0xee fp=0xc001930fa8 sp=0xc001930f60 pc=0x5ccbbe encoding/json.Marshal(0x1170020, 0xc001dfa430, 0x1, 0xf7218f3940a6636e, 0xc001931088, 0x4102df, 0xc001931106) /usr/local/go/src/encoding/json/encode.go:160 +0x52 fp=0xc001931020 sp=0xc001930fa8 pc=0x5cc1e2 github.com/incognitochain/incognito-chain/blockchain.(*ShardBestState).MarshalJSON(0xc000448dc0, 0xc0002d6870, 0xc001931106, 0xc000427a20, 0xc0077e8d40, 0x0) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/shardbeststate.go:244 +0x60 fp=0xc001931098 sp=0xc001931020 pc=0xc07100 github.com/incognitochain/incognito-chain/blockchain.(*Synker).UpdateStatev2(0xc000136628) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:719 +0x8a1 fp=0xc001931cf0 sp=0xc001931098 pc=0xc3e971 github.com/incognitochain/incognito-chain/blockchain.(*Synker).Start(0xc000136628) /root/go/src/github.com/incognitochain/incognito-chain/blockchain/synker.go:163 +0xa29 fp=0xc001931fd8 sp=0xc001931cf0 pc=0xc3d2c9 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1337 +0x1 fp=0xc001931fe0 sp=0xc001931fd8 pc=0x45f3a1 created by main.Server.Start /root/go/src/github.com/incognitochain/incognito-chain/server.go:733 +0x3a9 goroutine 1 [chan receive, 15 minutes]: main.mainMaster(0x0, 0x0, 0x0) /root/go/src/github.com/incognitochain/incognito-chain/incognito.go:132 +0x8eb main.main() /root/go/src/github.com/incognitochain/incognito-chain/incognito.go:178 +0x246
priority
panic by fatal error concurrent map read and map write on synker go fatal error concurrent map read and map write goroutine runtime throw usr local go src runtime panic go fp sp pc runtime usr local go src runtime map go fp sp pc reflect mapaccess usr local go src runtime map go fp sp pc reflect value mapindex usr local go src reflect value go fp sp pc encoding json mapencoder encode usr local go src encoding json encode go fp sp pc encoding json mapencoder encode fm usr local go src encoding json encode go fp sp pc encoding json structencoder encode usr local go src encoding json encode go fp sp pc encoding json structencoder encode fm usr local go src encoding json encode go fp sp pc encoding json ptrencoder encode usr local go src encoding json encode go fp sp pc encoding json ptrencoder encode fm usr local go src encoding json encode go fp sp pc encoding json encodestate reflectvalue usr local go src encoding json encode go fp sp pc encoding json encodestate marshal usr local go src encoding json encode go fp sp pc encoding json marshal usr local go src encoding json encode go fp sp pc github com incognitochain incognito chain blockchain shardbeststate marshaljson root go src github com incognitochain incognito chain blockchain shardbeststate go fp sp pc github com incognitochain incognito chain blockchain synker root go src github com incognitochain incognito chain blockchain synker go fp sp pc github com incognitochain incognito chain blockchain synker start root go src github com incognitochain incognito chain blockchain synker go fp sp pc runtime goexit usr local go src runtime asm s fp sp pc created by main server start root go src github com incognitochain incognito chain server go goroutine main mainmaster root go src github com incognitochain incognito chain incognito go main main root go src github com incognitochain incognito chain incognito go
1
355,185
10,577,268,740
IssuesEvent
2019-10-07 19:44:06
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
must gather for OCP fails to run during deploy
Medium Priority
`oc adm must-gather` requires `KUBECONFIG` environment variable to properly run. It is set by `OCP.set_kubeconfig()`, but **only** if path to file exists. This file is created early in `openshift-install create cluster` execution, but earliest `OCP.set_kubeconfig()` call is usually near the end of `deploy` method (through `test_cluster()`). In other words, errors in `openshift-install` will trigger must gather, but without KUBECONFIG variable set. This causes must gather to fail. How to reproduce: 1. In config, set `gather_on_deploy_failure` to True and `installer_version` to `4.2.0-0.nightly` 2. Ensure your pull-secret **does not** have `registry.svc.ci.openshift.org` entry (this is sure way to make installer fail) 3. Run `run-ci -m deployment <options> --deploy` ``` $ run-ci -m deployment --ocsci-conf ocs_ci/framework/conf/config.yaml --cluster-conf conf/ocs_basic_install.yml --cluster-name mzalewsk-test --cluster-path /home/mzalewsk/ocs-test/ --deploy # snip 18:59:49 - MainThread - ocs_ci.utility.utils - INFO - Executing command: /home/mzalewsk/sources/ocs-ci/bin/openshift-install create cluster --dir /home/mzalewsk/ocs-test/ --log-level INFO 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather. Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file: 1. Via the command-line flag --config 2. Via the KUBECONFIG environment variable 3. In your home directory as ~/.kube/config To view or setup config directly use the 'config' command. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker 19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker. Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file: 1. Via the command-line flag --config 2. Via the KUBECONFIG environment variable 3. In your home directory as ~/.kube/config To view or setup config directly use the 'config' command. ```
1.0
must gather for OCP fails to run during deploy - `oc adm must-gather` requires `KUBECONFIG` environment variable to properly run. It is set by `OCP.set_kubeconfig()`, but **only** if path to file exists. This file is created early in `openshift-install create cluster` execution, but earliest `OCP.set_kubeconfig()` call is usually near the end of `deploy` method (through `test_cluster()`). In other words, errors in `openshift-install` will trigger must gather, but without KUBECONFIG variable set. This causes must gather to fail. How to reproduce: 1. In config, set `gather_on_deploy_failure` to True and `installer_version` to `4.2.0-0.nightly` 2. Ensure your pull-secret **does not** have `registry.svc.ci.openshift.org` entry (this is sure way to make installer fail) 3. Run `run-ci -m deployment <options> --deploy` ``` $ run-ci -m deployment --ocsci-conf ocs_ci/framework/conf/config.yaml --cluster-conf conf/ocs_basic_install.yml --cluster-name mzalewsk-test --cluster-path /home/mzalewsk/ocs-test/ --deploy # snip 18:59:49 - MainThread - ocs_ci.utility.utils - INFO - Executing command: /home/mzalewsk/sources/ocs-ci/bin/openshift-install create cluster --dir /home/mzalewsk/ocs-test/ --log-level INFO 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather. Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file: 1. Via the command-line flag --config 2. Via the KUBECONFIG environment variable 3. In your home directory as ~/.kube/config To view or setup config directly use the 'config' command. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - Must gather image: quay.io/openshift/origin-must-gather will be used. 19:05:55 - MainThread - ocs_ci.ocs.utils - INFO - OCS logs will be placed in location /tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather 19:05:55 - MainThread - ocs_ci.utility.utils - INFO - Executing command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker 19:05:55 - MainThread - ocs_ci.ocs.utils - ERROR - Failed during must gather logs! Error: Error during execution of command: oc adm must-gather --image=quay.io/openshift/origin-must-gather --dest-dir=/tmp/failed_testcase_ocs_logs_1570208363/deployment_ocs_logs/ocp_must_gather -- /usr/bin/gather_service_logs worker. Error is error: Missing or incomplete configuration info. Please login or point to an existing, complete config file: 1. Via the command-line flag --config 2. Via the KUBECONFIG environment variable 3. In your home directory as ~/.kube/config To view or setup config directly use the 'config' command. ```
priority
must gather for ocp fails to run during deploy oc adm must gather requires kubeconfig environment variable to properly run it is set by ocp set kubeconfig but only if path to file exists this file is created early in openshift install create cluster execution but earliest ocp set kubeconfig call is usually near the end of deploy method through test cluster in other words errors in openshift install will trigger must gather but without kubeconfig variable set this causes must gather to fail how to reproduce in config set gather on deploy failure to true and installer version to nightly ensure your pull secret does not have registry svc ci openshift org entry this is sure way to make installer fail run run ci m deployment deploy run ci m deployment ocsci conf ocs ci framework conf config yaml cluster conf conf ocs basic install yml cluster name mzalewsk test cluster path home mzalewsk ocs test deploy snip mainthread ocs ci utility utils info executing command home mzalewsk sources ocs ci bin openshift install create cluster dir home mzalewsk ocs test log level info mainthread ocs ci ocs utils info must gather image quay io openshift origin must gather will be used mainthread ocs ci ocs utils info ocs logs will be placed in location tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci utility utils info executing command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci ocs utils error failed during must gather logs error error during execution of command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather error is error missing or incomplete configuration info please login or point to an existing complete config file via the command line flag config via the kubeconfig environment variable in your home directory as kube config to view or setup config directly use the config command mainthread ocs ci ocs utils info must gather image quay io openshift origin must gather will be used mainthread ocs ci ocs utils info ocs logs will be placed in location tmp failed testcase ocs logs deployment ocs logs ocp must gather mainthread ocs ci utility utils info executing command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather usr bin gather service logs worker mainthread ocs ci ocs utils error failed during must gather logs error error during execution of command oc adm must gather image quay io openshift origin must gather dest dir tmp failed testcase ocs logs deployment ocs logs ocp must gather usr bin gather service logs worker error is error missing or incomplete configuration info please login or point to an existing complete config file via the command line flag config via the kubeconfig environment variable in your home directory as kube config to view or setup config directly use the config command
1
114,223
11,841,098,679
IssuesEvent
2020-03-23 20:07:52
gaugsouza/petshow-backend
https://api.github.com/repos/gaugsouza/petshow-backend
closed
Definir e escrever sobre Gestão de tempo (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.)
documentation
(exemplos: cronograma, diagrama de Gantt, formato de sprints etc.)
1.0
Definir e escrever sobre Gestão de tempo (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.) - (exemplos: cronograma, diagrama de Gantt, formato de sprints etc.)
non_priority
definir e escrever sobre gestão de tempo exemplos cronograma diagrama de gantt formato de sprints etc exemplos cronograma diagrama de gantt formato de sprints etc
0
712,545
24,498,660,617
IssuesEvent
2022-10-10 10:54:17
AY2223S1-CS2103T-W15-4/tp
https://api.github.com/repos/AY2223S1-CS2103T-W15-4/tp
closed
As a user, I am given descriptive and helpful error messages when I give an invalid or incorrect command
type.Story priority.High
...so that I can troubleshoot easily without consulting the User Guide every time
1.0
As a user, I am given descriptive and helpful error messages when I give an invalid or incorrect command - ...so that I can troubleshoot easily without consulting the User Guide every time
priority
as a user i am given descriptive and helpful error messages when i give an invalid or incorrect command so that i can troubleshoot easily without consulting the user guide every time
1