Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
1,763
2,518,902,252
IssuesEvent
2015-01-17 06:39:52
rbsl-gaurav/rbframework
https://api.github.com/repos/rbsl-gaurav/rbframework
opened
Properly Alignment formal parameter of Controller's and view's methods
Priority-Normal task
<a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)** _Monday Jun 23, 2014 at 10:57 GMT_ _Originally opened as https://github.com/readybytes/rbframework/issues/42_ ----
1.0
Properly Alignment formal parameter of Controller's and view's methods - <a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)** _Monday Jun 23, 2014 at 10:57 GMT_ _Originally opened as https://github.com/readybytes/rbframework/issues/42_ ----
priority
properly alignment formal parameter of controller s and view s methods issue by monday jun at gmt originally opened as
1
118,614
25,338,664,701
IssuesEvent
2022-11-18 19:14:33
raupargor/Friendsn-t-Games
https://api.github.com/repos/raupargor/Friendsn-t-Games
closed
4.1.Objetos y armas: Sprites
code Priority: medium
Desarrollar animaciones para los objetos y armas a través de sprites
1.0
4.1.Objetos y armas: Sprites - Desarrollar animaciones para los objetos y armas a través de sprites
non_priority
objetos y armas sprites desarrollar animaciones para los objetos y armas a través de sprites
0
370,976
10,959,523,608
IssuesEvent
2019-11-27 11:35:48
highcharts/highcharts
https://api.github.com/repos/highcharts/highcharts
closed
venn.src.js:723 Uncaught Error: f(a) and f(b) must have opposite signs.
Priority: High Product: Highcharts Status: Review Type: Regression
#### Expected behaviour Venn diagram is displayed #### Actual behaviour Venn diagtam is not displayed + exception in console: Uncaught Error: f(a) and f(b) must have opposite signs. at L (venn.src.js:723) at f (venn.src.js:906) at M (venn.src.js:923) at venn.src.js:961 at Array.reduce (<anonymous>) at X (venn.src.js:937) at d.translate (venn.src.js:1429) at highcharts.src.js:24833 at Array.forEach (<anonymous>) at c.Chart.renderSeries (highcharts.src.js:24832) #### Live demo with steps to reproduce https://jsfiddle.net/smLn35cz/ Chart options: {"series":[ {"type":"venn", "data":[{"sets":["A"],"value":269}, {"sets":["B"],"value":4350}, {"sets":["C"],"value":2330}, {"sets":["A","B"],"value":131}, {"sets":["A","C"],"value":9}, {"sets":["B","C"],"value":12}, {"sets":["A","B","C"],"value":8}] }] } #### Product version Highcharts 7.2.0 #### Affected browser(s) Chrome
1.0
venn.src.js:723 Uncaught Error: f(a) and f(b) must have opposite signs. - #### Expected behaviour Venn diagram is displayed #### Actual behaviour Venn diagtam is not displayed + exception in console: Uncaught Error: f(a) and f(b) must have opposite signs. at L (venn.src.js:723) at f (venn.src.js:906) at M (venn.src.js:923) at venn.src.js:961 at Array.reduce (<anonymous>) at X (venn.src.js:937) at d.translate (venn.src.js:1429) at highcharts.src.js:24833 at Array.forEach (<anonymous>) at c.Chart.renderSeries (highcharts.src.js:24832) #### Live demo with steps to reproduce https://jsfiddle.net/smLn35cz/ Chart options: {"series":[ {"type":"venn", "data":[{"sets":["A"],"value":269}, {"sets":["B"],"value":4350}, {"sets":["C"],"value":2330}, {"sets":["A","B"],"value":131}, {"sets":["A","C"],"value":9}, {"sets":["B","C"],"value":12}, {"sets":["A","B","C"],"value":8}] }] } #### Product version Highcharts 7.2.0 #### Affected browser(s) Chrome
priority
venn src js uncaught error f a and f b must have opposite signs expected behaviour venn diagram is displayed actual behaviour venn diagtam is not displayed exception in console uncaught error f a and f b must have opposite signs at l venn src js at f venn src js at m venn src js at venn src js at array reduce at x venn src js at d translate venn src js at highcharts src js at array foreach at c chart renderseries highcharts src js live demo with steps to reproduce chart options series type venn data value sets value sets value sets value sets value sets value sets value product version highcharts affected browser s chrome
1
666,714
22,364,939,419
IssuesEvent
2022-06-16 02:18:07
apache/incubator-kyuubi
https://api.github.com/repos/apache/incubator-kyuubi
opened
[Bug] kyuubi loads the Listener to start an error
kind:bug priority:major
### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) ### Search before asking - [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues. ### Describe the bug kyuubi loads the Listener to start an error. When I use yarn client mode, it is normal, but cluster mode fails to start. ### Affects Version(s) 1.5.1 ### Kyuubi Server Log Output ```logtalk 09:20:04.816 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server HZ3-BD-2204-V2809.lianlianpay-dc.com/10.90.37.115:2181, sessionid = 0x100faf3ff4b0001, negotiated timeout = 60000 09:20:04.817 INFO org.apache.zookeeper.server.ZooKeeperServer: Established session 0x100faf3ff4b0001 with negotiated timeout 60000 for client /10.90.37.115:37874 09:20:04.818 INFO org.apache.curator.framework.state.ConnectionStateManager: State change: CONNECTED 09:20:04.845 INFO org.apache.kyuubi.engine.EngineRef: Launching engine: /home/dubbo/spark-3.2.1-bin-hadoop2.7/bin/spark-submit \ --class org.apache.kyuubi.engine.spark.SparkSQLEngine \ --conf spark.kyuubi.session.engine.idle.timeout=PT30M \ --conf spark.kyuubi.session.engine.check.interval=PT5M \ --conf spark.sql.hive.convertMetastoreParquet=false \ --conf spark.hive.server2.thrift.resultset.default.fetch.size=1000 \ --conf spark.kyuubi.ha.zookeeper.quorum=HZ3-BD-2204-V280:2181 \ --conf spark.kyuubi.sparklens.jars.location=/home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/lens_2.12-1.0.2-SNAPSHOT.jar \ --conf spark.extraListeners=com.lly.lens.JobListener \ --conf spark.kyuubi.client.ip=10.90.37.115 \ --conf spark.yarn.queue=llyt \ --conf spark.kyuubi.engine.submit.time=1655342404826 \ --conf spark.app.name=dwd_evt_payout_deposit_di \ --conf spark.executorEnv.HADOOP_USER_NAME=hdfs \ --conf spark.driver.memory=4G \ --conf spark.executor.instances=20 \ --conf spark.kyuubi.ha.engine.ref.id=964041b0-08c9-4663-9d02-ed5e4001bf54 \ --conf spark.executorEnv.SPARK_USER=hdfs \ --conf spark.kyuubi.session.conf.advisor=org.apache.kyuubi.plugin.HSessionConfAdvisor \ --conf spark.driver.cores=2 \ --conf spark.kyuubi.ha.zookeeper.auth.type=NONE \ --conf spark.submit.deployMode=cluster \ --conf spark.master=yarn \ --conf spark.yarn.tags=KYUUBI \ --conf spark.kyuubi.engine.share.level=CONNECTION \ --conf spark.kyuubi.ha.zookeeper.namespace=/kyuubi_1.5.1-SNAPSHOT_CONNECTION_SPARK_SQL/hdfs/964041b0-08c9-4663-9d02-ed5e4001bf54 \ --conf spark.executor.memory=15G \ --conf spark.executor.cores=4 \ --conf spark.yarn.executor.memoryOverhead=6G \ --conf spark.memory.fraction=0.8 \ --conf spark.kyuubi.engine.type=SPARK_SQL \ --conf spark.executor.memory=12 \ --conf spark.sql.shuffle.partitions=300 \ --jars /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/sparklens_2.12-1.0.2-SNAPSHOT.jar \ --proxy-user hdfs /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/externals/engines/spark/kyuubi-spark-sql-engine_2.12-1.5.1-SNAPSHOT.jar 09:20:04.857 INFO org.apache.kyuubi.engine.ProcBuilder: Logging to /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/work/hdfs/kyuubi-spark-sql-engine.log.2 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.646 INFO org.apache.zookeeper.server.PrepRequestProcessor: Processed session termination for sessionid: 0x100faf3ff4b0001 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001 09:20:24.649 INFO org.apache.zookeeper.server.NIOServerCnxn: Closed socket connection for client /10.90.37.115:37874 which had sessionid 0x100faf3ff4b0001 09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds 09:20:24.677 INFO org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Received request of closing SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] 09:20:24.679 INFO org.apache.kyuubi.session.KyuubiSessionManager: SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] is closed, current opening sessions 0 09:20:24.683 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: ERROR_STATE -> CLOSE ``` ### Kyuubi Engine Log Output ```logtalk 22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0 22/06/16 09:20:15 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:16 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:17 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001 09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds 22/06/16 09:20:20 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:21 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:22 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:23 INFO Client: Application report for application_1654630917149_16762 (state: FAILED) 22/06/16 09:20:23 INFO Client: client token: N/A diagnostics: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13 Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch. Container id: container_e82_1654630917149_16762_02_000001 Exit code: 13 [2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc [2022-06-16 09:20:23.055]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc For more detailed output, check the application tracking page: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 Then click on links to logs of each attempt. . Failing the application. ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: ll start time: 1655342409537 final status: FAILED tracking URL: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 user: hdfs 22/06/16 09:20:23 INFO Client: Deleted staging directory hdfs://hz3bdcrossborder01/user/hdfs/.sparkStaging/application_1654630917149_16762 22/06/16 09:20:23 ERROR Client: Application diagnostics message: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13 Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch. Container id: container_e82_1654630917149_16762_02_000001 Exit code: 13 [2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc ``` ### Kyuubi Server Configurations _No response_ ### Kyuubi Engine Configurations _No response_ ### Additional context _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR!
1.0
[Bug] kyuubi loads the Listener to start an error - ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) ### Search before asking - [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues. ### Describe the bug kyuubi loads the Listener to start an error. When I use yarn client mode, it is normal, but cluster mode fails to start. ### Affects Version(s) 1.5.1 ### Kyuubi Server Log Output ```logtalk 09:20:04.816 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server HZ3-BD-2204-V2809.lianlianpay-dc.com/10.90.37.115:2181, sessionid = 0x100faf3ff4b0001, negotiated timeout = 60000 09:20:04.817 INFO org.apache.zookeeper.server.ZooKeeperServer: Established session 0x100faf3ff4b0001 with negotiated timeout 60000 for client /10.90.37.115:37874 09:20:04.818 INFO org.apache.curator.framework.state.ConnectionStateManager: State change: CONNECTED 09:20:04.845 INFO org.apache.kyuubi.engine.EngineRef: Launching engine: /home/dubbo/spark-3.2.1-bin-hadoop2.7/bin/spark-submit \ --class org.apache.kyuubi.engine.spark.SparkSQLEngine \ --conf spark.kyuubi.session.engine.idle.timeout=PT30M \ --conf spark.kyuubi.session.engine.check.interval=PT5M \ --conf spark.sql.hive.convertMetastoreParquet=false \ --conf spark.hive.server2.thrift.resultset.default.fetch.size=1000 \ --conf spark.kyuubi.ha.zookeeper.quorum=HZ3-BD-2204-V280:2181 \ --conf spark.kyuubi.sparklens.jars.location=/home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/lens_2.12-1.0.2-SNAPSHOT.jar \ --conf spark.extraListeners=com.lly.lens.JobListener \ --conf spark.kyuubi.client.ip=10.90.37.115 \ --conf spark.yarn.queue=llyt \ --conf spark.kyuubi.engine.submit.time=1655342404826 \ --conf spark.app.name=dwd_evt_payout_deposit_di \ --conf spark.executorEnv.HADOOP_USER_NAME=hdfs \ --conf spark.driver.memory=4G \ --conf spark.executor.instances=20 \ --conf spark.kyuubi.ha.engine.ref.id=964041b0-08c9-4663-9d02-ed5e4001bf54 \ --conf spark.executorEnv.SPARK_USER=hdfs \ --conf spark.kyuubi.session.conf.advisor=org.apache.kyuubi.plugin.HSessionConfAdvisor \ --conf spark.driver.cores=2 \ --conf spark.kyuubi.ha.zookeeper.auth.type=NONE \ --conf spark.submit.deployMode=cluster \ --conf spark.master=yarn \ --conf spark.yarn.tags=KYUUBI \ --conf spark.kyuubi.engine.share.level=CONNECTION \ --conf spark.kyuubi.ha.zookeeper.namespace=/kyuubi_1.5.1-SNAPSHOT_CONNECTION_SPARK_SQL/hdfs/964041b0-08c9-4663-9d02-ed5e4001bf54 \ --conf spark.executor.memory=15G \ --conf spark.executor.cores=4 \ --conf spark.yarn.executor.memoryOverhead=6G \ --conf spark.memory.fraction=0.8 \ --conf spark.kyuubi.engine.type=SPARK_SQL \ --conf spark.executor.memory=12 \ --conf spark.sql.shuffle.partitions=300 \ --jars /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/sparklens_2.12-1.0.2-SNAPSHOT.jar \ --proxy-user hdfs /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/externals/engines/spark/kyuubi-spark-sql-engine_2.12-1.5.1-SNAPSHOT.jar 09:20:04.857 INFO org.apache.kyuubi.engine.ProcBuilder: Logging to /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/work/hdfs/kyuubi-spark-sql-engine.log.2 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.646 INFO org.apache.zookeeper.server.PrepRequestProcessor: Processed session termination for sessionid: 0x100faf3ff4b0001 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001 09:20:24.649 INFO org.apache.zookeeper.server.NIOServerCnxn: Closed socket connection for client /10.90.37.115:37874 which had sessionid 0x100faf3ff4b0001 09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds 09:20:24.677 INFO org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Received request of closing SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] 09:20:24.679 INFO org.apache.kyuubi.session.KyuubiSessionManager: SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] is closed, current opening sessions 0 09:20:24.683 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: ERROR_STATE -> CLOSE ``` ### Kyuubi Engine Log Output ```logtalk 22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0 22/06/16 09:20:15 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:16 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:17 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting 09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed 09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001 09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds 22/06/16 09:20:20 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:21 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:22 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED) 22/06/16 09:20:23 INFO Client: Application report for application_1654630917149_16762 (state: FAILED) 22/06/16 09:20:23 INFO Client: client token: N/A diagnostics: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13 Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch. Container id: container_e82_1654630917149_16762_02_000001 Exit code: 13 [2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc [2022-06-16 09:20:23.055]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc For more detailed output, check the application tracking page: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 Then click on links to logs of each attempt. . Failing the application. ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: ll start time: 1655342409537 final status: FAILED tracking URL: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 user: hdfs 22/06/16 09:20:23 INFO Client: Deleted staging directory hdfs://hz3bdcrossborder01/user/hdfs/.sparkStaging/application_1654630917149_16762 22/06/16 09:20:23 ERROR Client: Application diagnostics message: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13 Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch. Container id: container_e82_1654630917149_16762_02_000001 Exit code: 13 [2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err. Last 4096 bytes of prelaunch.err : Last 4096 bytes of stderr : e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152) at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634) at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295) at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683) at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267) at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019) at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) at scala.util.Try$.apply(Try.scala:213) at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188) at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178) at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54) 22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac 22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc ``` ### Kyuubi Server Configurations _No response_ ### Kyuubi Engine Configurations _No response_ ### Additional context _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR!
priority
kyuubi loads the listener to start an error code of conduct i agree to follow this project s search before asking i have searched in the and found no similar issues describe the bug kyuubi loads the listener to start an error when i use yarn client mode it is normal but cluster mode fails to start affects version s kyuubi server log output logtalk info org apache zookeeper clientcnxn session establishment complete on server bd lianlianpay dc com sessionid negotiated timeout info org apache zookeeper server zookeeperserver established session with negotiated timeout for client info org apache curator framework state connectionstatemanager state change connected info org apache kyuubi engine engineref launching engine home dubbo spark bin bin spark submit class org apache kyuubi engine spark sparksqlengine conf spark kyuubi session engine idle timeout conf spark kyuubi session engine check interval conf spark sql hive convertmetastoreparquet false conf spark hive thrift resultset default fetch size conf spark kyuubi ha zookeeper quorum bd conf spark kyuubi sparklens jars location home dubbo apache kyuubi incubating bin external jars lens snapshot jar conf spark extralisteners com lly lens joblistener conf spark kyuubi client ip conf spark yarn queue llyt conf spark kyuubi engine submit time conf spark app name dwd evt payout deposit di conf spark executorenv hadoop user name hdfs conf spark driver memory conf spark executor instances conf spark kyuubi ha engine ref id conf spark executorenv spark user hdfs conf spark kyuubi session conf advisor org apache kyuubi plugin hsessionconfadvisor conf spark driver cores conf spark kyuubi ha zookeeper auth type none conf spark submit deploymode cluster conf spark master yarn conf spark yarn tags kyuubi conf spark kyuubi engine share level connection conf spark kyuubi ha zookeeper namespace kyuubi snapshot connection spark sql hdfs conf spark executor memory conf spark executor cores conf spark yarn executor memoryoverhead conf spark memory fraction conf spark kyuubi engine type spark sql conf spark executor memory conf spark sql shuffle partitions jars home dubbo apache kyuubi incubating bin external jars sparklens snapshot jar proxy user hdfs home dubbo apache kyuubi incubating bin externals engines spark kyuubi spark sql engine snapshot jar info org apache kyuubi engine procbuilder logging to home dubbo apache kyuubi incubating bin work hdfs kyuubi spark sql engine log info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper server preprequestprocessor processed session termination for sessionid info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info org apache zookeeper server nioservercnxn closed socket connection for client which had sessionid info org apache kyuubi operation launchengine processing hdfs s query running state error state statement launch engine time taken seconds info org apache kyuubi server kyuubitbinaryfrontendservice received request of closing sessionhandle info org apache kyuubi session kyuubisessionmanager sessionhandle is closed current opening sessions info org apache kyuubi operation launchengine processing hdfs s query error state close kyuubi engine log output logtalk info client application report for application state accepted info client application report for application state accepted info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info org apache kyuubi operation launchengine processing hdfs s query running state error state statement launch engine time taken seconds info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state failed info client client token n a diagnostics application application failed times due to am container for appattempt exited with exitcode failing this attempt diagnostics exception from container launch container id container exit code container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark for more detailed output check the application tracking page then click on links to logs of each attempt failing the application applicationmaster host n a applicationmaster rpc port queue ll start time final status failed tracking url user hdfs info client deleted staging directory hdfs user hdfs sparkstaging application error client application diagnostics message application application failed times due to am container for appattempt exited with exitcode failing this attempt diagnostics exception from container launch container id container exit code container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark kyuubi server configurations no response kyuubi engine configurations no response additional context no response are you willing to submit pr yes i am willing to submit a pr
1
685,492
23,458,304,277
IssuesEvent
2022-08-16 10:53:08
nanoframework/Home
https://api.github.com/repos/nanoframework/Home
closed
VS2022 extension broken on 17.3
Type: Bug Status: In progress Area: Visual Studio extension Priority: Critical
### Tool Visual Studio extension ### Description When opening a nfproj in VS2022 17.3 the extension fails. ### How to reproduce 1. Open a nfproj 2. VS output LimitedFunctionality System.ArgumentNullException: Value cannot be null. Parameter name: commandLine at Microsoft.VisualStudio.LanguageServices.Implementation.ProjectSystem.VisualStudioProjectOptionsProcessor.SetCommandLine(String commandLine) at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.VisualStudio.ProjectSystem.CommonProjectSystemTools.Rethrow(Exception ex) at Microsoft.VisualStudio.ProjectSystem.ExceptionFilter.<>cDisplayClass2_0.<Guard>gAction|0() at GuardMethodClass.GuardMethod(Func1 , Func2 , Func`2 ) ### Expected behaviour _No response_ ### Screenshots _No response_ ### Aditional context _No response_
1.0
VS2022 extension broken on 17.3 - ### Tool Visual Studio extension ### Description When opening a nfproj in VS2022 17.3 the extension fails. ### How to reproduce 1. Open a nfproj 2. VS output LimitedFunctionality System.ArgumentNullException: Value cannot be null. Parameter name: commandLine at Microsoft.VisualStudio.LanguageServices.Implementation.ProjectSystem.VisualStudioProjectOptionsProcessor.SetCommandLine(String commandLine) at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at Microsoft.VisualStudio.ProjectSystem.CommonProjectSystemTools.Rethrow(Exception ex) at Microsoft.VisualStudio.ProjectSystem.ExceptionFilter.<>cDisplayClass2_0.<Guard>gAction|0() at GuardMethodClass.GuardMethod(Func1 , Func2 , Func`2 ) ### Expected behaviour _No response_ ### Screenshots _No response_ ### Aditional context _No response_
priority
extension broken on tool visual studio extension description when opening a nfproj in the extension fails how to reproduce open a nfproj vs output limitedfunctionality system argumentnullexception value cannot be null parameter name commandline at microsoft visualstudio languageservices implementation projectsystem visualstudioprojectoptionsprocessor setcommandline string commandline at microsoft visualstudio projectsystem languageservices workspaceprojectcontextprovider movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft visualstudio projectsystem languageservices workspaceprojectcontextprovider movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft visualstudio projectsystem commonprojectsystemtools rethrow exception ex at microsoft visualstudio projectsystem exceptionfilter gaction at guardmethodclass guardmethod func expected behaviour no response screenshots no response aditional context no response
1
411,632
12,026,896,010
IssuesEvent
2020-04-12 16:03:22
ooni/probe
https://api.github.com/repos/ooni/probe
closed
Feature request: Run All Tests button
effort/L ooni/probe-mobile priority/medium ux
We have received community requests to include a "Run All Tests" button in the revamped app (on both Android and iOS), to run all OONI Probe tests in sequence, by tapping one button (instead of having to manually run each test individually).
1.0
Feature request: Run All Tests button - We have received community requests to include a "Run All Tests" button in the revamped app (on both Android and iOS), to run all OONI Probe tests in sequence, by tapping one button (instead of having to manually run each test individually).
priority
feature request run all tests button we have received community requests to include a run all tests button in the revamped app on both android and ios to run all ooni probe tests in sequence by tapping one button instead of having to manually run each test individually
1
182,612
30,873,966,286
IssuesEvent
2023-08-03 13:12:37
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
When provisioning an RKE1 custom cluster, provide a visual indication that a user has clicked the clipboard icon
[zube]: Done status/wontfix kind/design QA/XS kind/enhancement team/area2 ember
<!--------- For bugs and general issues ---------> **Setup** - Rancher version: `v2.6-head` https://github.com/rancher/rancher/commit/f4d09c73e1a609f5c5fe0848ebd6e4b059056264 Browser: Chrome - Browser type & version: Chrome **Describe the bug** <!--A clear and concise description of what the bug is.--> Not necessarily a bug, more of an enhancement. When provisioning a downstream RKE1 custom cluster, we have the following screen: ![Screen Shot 2022-06-08 at 12 02 28 PM](https://user-images.githubusercontent.com/52591819/172728360-71adc6c3-c7ce-413e-af59-0556819af959.png) When you click the clipboard icon to copy the registration command, there is no illustration that it has been copied. Throughout the UI, we indicate a green checkmark once a user clicks the clipboard icon. We should follow the same practice here to add overall cohesiveness. **To Reproduce** <!--Steps to reproduce the behavior--> 1. Setup Rancher and navigate to the UI in a browser. 2. Provision a downstream RKE1 custom cluster. 3. Click the clipboard icon to copy the registration command. Note that you will have no visual indication that it has been copied to your clipboard. **Result** No visual indication that I have successfully copied the registration command to my clipboard. **Expected Result** <!--A clear and concise description of what you expected to happen.--> When I click the clipboard, I should see a green checkmark or a message nothing that the registration command has been copied to my clipboard.
1.0
When provisioning an RKE1 custom cluster, provide a visual indication that a user has clicked the clipboard icon - <!--------- For bugs and general issues ---------> **Setup** - Rancher version: `v2.6-head` https://github.com/rancher/rancher/commit/f4d09c73e1a609f5c5fe0848ebd6e4b059056264 Browser: Chrome - Browser type & version: Chrome **Describe the bug** <!--A clear and concise description of what the bug is.--> Not necessarily a bug, more of an enhancement. When provisioning a downstream RKE1 custom cluster, we have the following screen: ![Screen Shot 2022-06-08 at 12 02 28 PM](https://user-images.githubusercontent.com/52591819/172728360-71adc6c3-c7ce-413e-af59-0556819af959.png) When you click the clipboard icon to copy the registration command, there is no illustration that it has been copied. Throughout the UI, we indicate a green checkmark once a user clicks the clipboard icon. We should follow the same practice here to add overall cohesiveness. **To Reproduce** <!--Steps to reproduce the behavior--> 1. Setup Rancher and navigate to the UI in a browser. 2. Provision a downstream RKE1 custom cluster. 3. Click the clipboard icon to copy the registration command. Note that you will have no visual indication that it has been copied to your clipboard. **Result** No visual indication that I have successfully copied the registration command to my clipboard. **Expected Result** <!--A clear and concise description of what you expected to happen.--> When I click the clipboard, I should see a green checkmark or a message nothing that the registration command has been copied to my clipboard.
non_priority
when provisioning an custom cluster provide a visual indication that a user has clicked the clipboard icon setup rancher version head browser chrome browser type version chrome describe the bug not necessarily a bug more of an enhancement when provisioning a downstream custom cluster we have the following screen when you click the clipboard icon to copy the registration command there is no illustration that it has been copied throughout the ui we indicate a green checkmark once a user clicks the clipboard icon we should follow the same practice here to add overall cohesiveness to reproduce setup rancher and navigate to the ui in a browser provision a downstream custom cluster click the clipboard icon to copy the registration command note that you will have no visual indication that it has been copied to your clipboard result no visual indication that i have successfully copied the registration command to my clipboard expected result when i click the clipboard i should see a green checkmark or a message nothing that the registration command has been copied to my clipboard
0
539,908
15,796,628,881
IssuesEvent
2021-04-02 15:15:04
refgenie/refgenie
https://api.github.com/repos/refgenie/refgenie
opened
auto documenting buildable assets
docs enhancement priority-low
It would be nice to generate some sort of [this page](http://refgenie.databio.org/en/latest/available_assets/) automatically, based on the set of available recipes. This will come in handy once the recipes are no longer a part of refgenie package, but are served with refgenieserver. It should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in [this issue](https://github.com/refgenie/refgenie/issues/198#issue-683071342) are defined. But we could also provide means to add more detailed documentation for each recipe. For example, `docs` key that is a freeform text, later interpreted as markdown, or something.
1.0
auto documenting buildable assets - It would be nice to generate some sort of [this page](http://refgenie.databio.org/en/latest/available_assets/) automatically, based on the set of available recipes. This will come in handy once the recipes are no longer a part of refgenie package, but are served with refgenieserver. It should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in [this issue](https://github.com/refgenie/refgenie/issues/198#issue-683071342) are defined. But we could also provide means to add more detailed documentation for each recipe. For example, `docs` key that is a freeform text, later interpreted as markdown, or something.
priority
auto documenting buildable assets it would be nice to generate some sort of automatically based on the set of available recipes this will come in handy once the recipes are no longer a part of refgenie package but are served with refgenieserver it should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in are defined but we could also provide means to add more detailed documentation for each recipe for example docs key that is a freeform text later interpreted as markdown or something
1
372,987
26,030,482,122
IssuesEvent
2022-12-21 20:39:57
solidjs/solid-start
https://api.github.com/repos/solidjs/solid-start
closed
Linking to pdf files inside `public` folder doesn't work on routing.
documentation
I have a repo that builds on top of solid-start here: https://github.com/nikitavoloboev/nikitavoloboev I have a pdf file in public called `cv.pdf`: https://github.com/nikitavoloboev/nikitavoloboev/blob/main/public/cv.pdf If I for example do this: https://github.com/nikitavoloboev/nikitavoloboev/blob/a0ab2a594e77e0cbcc450de56099a15aa4691c06/src/components/Nav.tsx#L37 ``` <a class="nav-link" href="./cv.pdf"> CV </a> ``` Inside the app if I go to this page, it will 404: https://www.loom.com/share/900b34ae19904c59aed372d2a7457103 Not sure what I am doing wrong. I need to refresh the page for the PDF to actually show.
1.0
Linking to pdf files inside `public` folder doesn't work on routing. - I have a repo that builds on top of solid-start here: https://github.com/nikitavoloboev/nikitavoloboev I have a pdf file in public called `cv.pdf`: https://github.com/nikitavoloboev/nikitavoloboev/blob/main/public/cv.pdf If I for example do this: https://github.com/nikitavoloboev/nikitavoloboev/blob/a0ab2a594e77e0cbcc450de56099a15aa4691c06/src/components/Nav.tsx#L37 ``` <a class="nav-link" href="./cv.pdf"> CV </a> ``` Inside the app if I go to this page, it will 404: https://www.loom.com/share/900b34ae19904c59aed372d2a7457103 Not sure what I am doing wrong. I need to refresh the page for the PDF to actually show.
non_priority
linking to pdf files inside public folder doesn t work on routing i have a repo that builds on top of solid start here i have a pdf file in public called cv pdf if i for example do this cv inside the app if i go to this page it will not sure what i am doing wrong i need to refresh the page for the pdf to actually show
0
277,160
8,621,305,231
IssuesEvent
2018-11-20 17:00:27
supergiant/supergiant
https://api.github.com/repos/supergiant/supergiant
closed
2.0: UI - Pagination needs to be removed from Apps store
Low Priority
![Screen Shot 2018-11-09 at 10:00 PM](http://marker.screenshots.prod.s3.amazonaws.com/a323c991a39244c2bbed37ab08432857-1541800855028.png) --- ### More Info **Reported by:** eldridge henley (eldridge@qbox.io) **Source URL**: [http://localhost:4200/apps/supergiant](http://localhost:4200/apps/supergiant) <table><tr><td><strong>Browser</strong></td><td>Chrome 70.0.3538.77</td></tr><tr><td><strong>Screen Size</strong></td><td>1920 x 1080</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.14.0</td></tr><tr><td><strong>Viewport Size</strong></td><td>1610 x 1027</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr></table>
1.0
2.0: UI - Pagination needs to be removed from Apps store - ![Screen Shot 2018-11-09 at 10:00 PM](http://marker.screenshots.prod.s3.amazonaws.com/a323c991a39244c2bbed37ab08432857-1541800855028.png) --- ### More Info **Reported by:** eldridge henley (eldridge@qbox.io) **Source URL**: [http://localhost:4200/apps/supergiant](http://localhost:4200/apps/supergiant) <table><tr><td><strong>Browser</strong></td><td>Chrome 70.0.3538.77</td></tr><tr><td><strong>Screen Size</strong></td><td>1920 x 1080</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.14.0</td></tr><tr><td><strong>Viewport Size</strong></td><td>1610 x 1027</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr></table>
priority
ui pagination needs to be removed from apps store more info reported by eldridge henley eldridge qbox io source url browser chrome screen size x os os x viewport size x zoom level pixel ratio
1
673,778
23,031,149,402
IssuesEvent
2022-07-22 14:02:19
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.apps.disneyplus.com - site is not usable
os-ios browser-firefox-ios priority-important device-tablet
<!-- @browser: Safari 13.1 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/107754 --> <!-- @extra_labels: browser-firefox-ios, device-tablet --> **URL**: https://www.apps.disneyplus.com/eg/onboarding **Browser / Version**: Safari 13.1 **Operating System**: Mac OS X 10.15.4 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: None of the links are working, not even on my MacBook <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.apps.disneyplus.com - site is not usable - <!-- @browser: Safari 13.1 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/107754 --> <!-- @extra_labels: browser-firefox-ios, device-tablet --> **URL**: https://www.apps.disneyplus.com/eg/onboarding **Browser / Version**: Safari 13.1 **Operating System**: Mac OS X 10.15.4 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: None of the links are working, not even on my MacBook <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version safari operating system mac os x tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce none of the links are working not even on my macbook browser configuration none from with ❤️
1
446,305
12,854,821,155
IssuesEvent
2020-07-09 03:10:49
web-cat/code-workout
https://api.github.com/repos/web-cat/code-workout
closed
Add current exercise score to exercise feedback display
Staged feature priority: normal
Add a radial progress bar showing the current score (same as in thumbnails) to the right of the "Feedback" title in the exercise feedback (generated in the sse/ajax_feedback view).
1.0
Add current exercise score to exercise feedback display - Add a radial progress bar showing the current score (same as in thumbnails) to the right of the "Feedback" title in the exercise feedback (generated in the sse/ajax_feedback view).
priority
add current exercise score to exercise feedback display add a radial progress bar showing the current score same as in thumbnails to the right of the feedback title in the exercise feedback generated in the sse ajax feedback view
1
24,997
24,559,224,225
IssuesEvent
2022-10-12 18:38:38
JuliaReach/LazySets.jl
https://api.github.com/repos/JuliaReach/LazySets.jl
opened
PolarDirections and SphericalDirections use different interpretations of Nφ
usability :computer_mouse: fix :face_with_head_bandage:
In `PolarDirections`, the range has length `Nφ+1`. In `SphericalDirections`, the range has length `Nφ`.
True
PolarDirections and SphericalDirections use different interpretations of Nφ - In `PolarDirections`, the range has length `Nφ+1`. In `SphericalDirections`, the range has length `Nφ`.
non_priority
polardirections and sphericaldirections use different interpretations of nφ in polardirections the range has length nφ in sphericaldirections the range has length nφ
0
31,453
14,970,563,595
IssuesEvent
2021-01-27 19:48:18
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
My app is slow or missing frames (metabug)
created via performance template macos-metal perf: speed severe: performance
This is a meta-issue to track reproducible reports of jank in Flutter apps. If you are experiencing jank in your app: 1. Try to reproduce the problem in a test app. Either run `flutter create janktest` and recreate the situation you are experiencing in that app, or clone your app and delete code until you have the jank reproducing with a single .dart file. 2. [File a bug](https://github.com/flutter/flutter/issues/new?assignees=&labels=created+via+performance+template&template=5_performance_speed.md&title=) and include your .dart file demonstrating the problem. If you need more than just a .dart file (for example, assets are needed to reproduce the issue, or plugins/packages are needed to reproduce the issue) then create a GitHub repository and upload the app there. Make sure to include the `flutter doctor -v` output and any logs from `flutter run` and `flutter analyze`. 3. Switch flutter to master channel and run this app on a physical device using profile mode with Skia tracing enabled, as follows: `flutter channel master` `flutter run --profile --trace-skia` Then press ‘P’ to enable the performance overlay. The bleeding edge master channel is encouraged here because Flutter is constantly fixing bugs and improving its performance. Your problem in an older Flutter version may have already been solved in the master channel. 4. Record a video of the performance issue using another phone so we can have an intuitive understanding of what happened. Don’t use "adb screenrecord", as that affects the performance of the profile run. Attach the video to your bug. 5. Open Observatory and save a timeline trace of the performance issue so we know which functions might be causing it. See "How to Collect and Read Timeline Traces" on this blog post: https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499 Attach the JSON file containing your trace to your bug. You may also wish to include a screenshot of the part of the trace showing the problem you are seeing, just so that people can see at a glance what kind of performance issue the bug is about. 6. Mention _this_ bug in your bug, so that GitHub includes a link to it here. Please avoid commenting on this bug. Keep each issue separate so that we can examine each specific problem individually. Having one issue that contains comments about multiple problems make the issue intractable.
True
My app is slow or missing frames (metabug) - This is a meta-issue to track reproducible reports of jank in Flutter apps. If you are experiencing jank in your app: 1. Try to reproduce the problem in a test app. Either run `flutter create janktest` and recreate the situation you are experiencing in that app, or clone your app and delete code until you have the jank reproducing with a single .dart file. 2. [File a bug](https://github.com/flutter/flutter/issues/new?assignees=&labels=created+via+performance+template&template=5_performance_speed.md&title=) and include your .dart file demonstrating the problem. If you need more than just a .dart file (for example, assets are needed to reproduce the issue, or plugins/packages are needed to reproduce the issue) then create a GitHub repository and upload the app there. Make sure to include the `flutter doctor -v` output and any logs from `flutter run` and `flutter analyze`. 3. Switch flutter to master channel and run this app on a physical device using profile mode with Skia tracing enabled, as follows: `flutter channel master` `flutter run --profile --trace-skia` Then press ‘P’ to enable the performance overlay. The bleeding edge master channel is encouraged here because Flutter is constantly fixing bugs and improving its performance. Your problem in an older Flutter version may have already been solved in the master channel. 4. Record a video of the performance issue using another phone so we can have an intuitive understanding of what happened. Don’t use "adb screenrecord", as that affects the performance of the profile run. Attach the video to your bug. 5. Open Observatory and save a timeline trace of the performance issue so we know which functions might be causing it. See "How to Collect and Read Timeline Traces" on this blog post: https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499 Attach the JSON file containing your trace to your bug. You may also wish to include a screenshot of the part of the trace showing the problem you are seeing, just so that people can see at a glance what kind of performance issue the bug is about. 6. Mention _this_ bug in your bug, so that GitHub includes a link to it here. Please avoid commenting on this bug. Keep each issue separate so that we can examine each specific problem individually. Having one issue that contains comments about multiple problems make the issue intractable.
non_priority
my app is slow or missing frames metabug this is a meta issue to track reproducible reports of jank in flutter apps if you are experiencing jank in your app try to reproduce the problem in a test app either run flutter create janktest and recreate the situation you are experiencing in that app or clone your app and delete code until you have the jank reproducing with a single dart file and include your dart file demonstrating the problem if you need more than just a dart file for example assets are needed to reproduce the issue or plugins packages are needed to reproduce the issue then create a github repository and upload the app there make sure to include the flutter doctor v output and any logs from flutter run and flutter analyze switch flutter to master channel and run this app on a physical device using profile mode with skia tracing enabled as follows flutter channel master flutter run profile trace skia then press ‘p’ to enable the performance overlay the bleeding edge master channel is encouraged here because flutter is constantly fixing bugs and improving its performance your problem in an older flutter version may have already been solved in the master channel record a video of the performance issue using another phone so we can have an intuitive understanding of what happened don’t use adb screenrecord as that affects the performance of the profile run attach the video to your bug open observatory and save a timeline trace of the performance issue so we know which functions might be causing it see how to collect and read timeline traces on this blog post attach the json file containing your trace to your bug you may also wish to include a screenshot of the part of the trace showing the problem you are seeing just so that people can see at a glance what kind of performance issue the bug is about mention this bug in your bug so that github includes a link to it here please avoid commenting on this bug keep each issue separate so that we can examine each specific problem individually having one issue that contains comments about multiple problems make the issue intractable
0
820,981
30,798,127,426
IssuesEvent
2023-07-31 21:45:24
PazerOP/tf2_bot_detector
https://api.github.com/repos/PazerOP/tf2_bot_detector
opened
[BUG] Bot Detector will not update
Type: Bug Priority: Medium
Any time I boot up the bot detector, it comes up with this Error. I don't think the bot detector even works while it's in this state. I saw others using the bot detector yesterday and, while I had mine open, it didn't do anything. Yes, I did boot TF2 from the bot detector. ``Update check failed: - class tf2_bot_detector::http_error - Failed to HTTP GET https://tf2bd-util.pazer.us:443/AppInstaller/LatestVersion.json?type=Public: <UNKNOWN>(HTTP 523)`` Log: [2023-07-31_22-35-39.log](https://github.com/PazerOP/tf2_bot_detector/files/12222482/2023-07-31_22-35-39.log) I am currently on the latest version, since someone suggested to reinstall it and I've deleted every instance I am aware of, and it hasn't fixed the issue.
1.0
[BUG] Bot Detector will not update - Any time I boot up the bot detector, it comes up with this Error. I don't think the bot detector even works while it's in this state. I saw others using the bot detector yesterday and, while I had mine open, it didn't do anything. Yes, I did boot TF2 from the bot detector. ``Update check failed: - class tf2_bot_detector::http_error - Failed to HTTP GET https://tf2bd-util.pazer.us:443/AppInstaller/LatestVersion.json?type=Public: <UNKNOWN>(HTTP 523)`` Log: [2023-07-31_22-35-39.log](https://github.com/PazerOP/tf2_bot_detector/files/12222482/2023-07-31_22-35-39.log) I am currently on the latest version, since someone suggested to reinstall it and I've deleted every instance I am aware of, and it hasn't fixed the issue.
priority
bot detector will not update any time i boot up the bot detector it comes up with this error i don t think the bot detector even works while it s in this state i saw others using the bot detector yesterday and while i had mine open it didn t do anything yes i did boot from the bot detector update check failed class bot detector http error failed to http get http log i am currently on the latest version since someone suggested to reinstall it and i ve deleted every instance i am aware of and it hasn t fixed the issue
1
512,311
14,893,807,569
IssuesEvent
2021-01-21 06:17:07
project-serum/anchor
https://api.github.com/repos/project-serum/anchor
opened
Provide separate Account and AccountBox types
priority:2
Currently we use `ProgramAccount` to refer to accounts owned by the current program. This type boxes it's inner account by default (needed to sometimes get around BPF stack size limits). Instead, we should provide a non-boxed version with `type AccountBox = Box<NonBoxedVersion>;` as the boxed variant.
1.0
Provide separate Account and AccountBox types - Currently we use `ProgramAccount` to refer to accounts owned by the current program. This type boxes it's inner account by default (needed to sometimes get around BPF stack size limits). Instead, we should provide a non-boxed version with `type AccountBox = Box<NonBoxedVersion>;` as the boxed variant.
priority
provide separate account and accountbox types currently we use programaccount to refer to accounts owned by the current program this type boxes it s inner account by default needed to sometimes get around bpf stack size limits instead we should provide a non boxed version with type accountbox box as the boxed variant
1
168,425
14,148,187,917
IssuesEvent
2020-11-10 22:05:33
bounswe/bounswe2020group2
https://api.github.com/repos/bounswe/bounswe2020group2
closed
Review of Project Plan
documentation effort: medium priority: low type: enhancement who: group
- [x] Resource names of [Project Plan](https://github.com/bounswe/bounswe2020group2/wiki/Project-Plan) should be changed according to subgroups for Android, back end and front end development. - [ ] Tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader. **_Due Date: 02.11.2020 Monday @23.59_**
1.0
Review of Project Plan - - [x] Resource names of [Project Plan](https://github.com/bounswe/bounswe2020group2/wiki/Project-Plan) should be changed according to subgroups for Android, back end and front end development. - [ ] Tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader. **_Due Date: 02.11.2020 Monday @23.59_**
non_priority
review of project plan resource names of should be changed according to subgroups for android back end and front end development tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader due date monday
0
54,098
13,391,229,413
IssuesEvent
2020-09-02 22:05:12
solex2006/SELIProject
https://api.github.com/repos/solex2006/SELIProject
opened
Audio is being requested for all scenes
:bug:bug P2 - High S2 - Minor Defects env:next release module:STORING TELLING type:functional
## Expected Behavior <!--- Tell us what should happen --> If a video with audio is already upload in scene, an audio file should not be required ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> I can't publish a story while not all scenes has an audio file. ## Environment <!--- Please tell us about your environment --> <!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView --> https://vm2161.kaj.pouta.csc.fi/ ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- Provide a detailed description of the change or addition you are proposing --> Include a `input type=checkbox` asking if the uploaded video is a video-only file (this will be important for accessibility). If user check the input, then audio must be required. ## Related issues <!--- Not obligatory, but reference related issues -->
1.0
Audio is being requested for all scenes - ## Expected Behavior <!--- Tell us what should happen --> If a video with audio is already upload in scene, an audio file should not be required ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> I can't publish a story while not all scenes has an audio file. ## Environment <!--- Please tell us about your environment --> <!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView --> https://vm2161.kaj.pouta.csc.fi/ ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- Provide a detailed description of the change or addition you are proposing --> Include a `input type=checkbox` asking if the uploaded video is a video-only file (this will be important for accessibility). If user check the input, then audio must be required. ## Related issues <!--- Not obligatory, but reference related issues -->
non_priority
audio is being requested for all scenes expected behavior if a video with audio is already upload in scene an audio file should not be required current behavior i can t publish a story while not all scenes has an audio file environment possible solution include a input type checkbox asking if the uploaded video is a video only file this will be important for accessibility if user check the input then audio must be required related issues
0
253,655
8,058,843,353
IssuesEvent
2018-08-02 19:51:24
18F/10x-qualitative-data
https://api.github.com/repos/18F/10x-qualitative-data
opened
Define requirements for Camille and OCE
Sprint 3 priority product
As an *analytics and content team member* I want *to have access to Qualtrics* so that *I can view reports, curate data and manage surveys*. ## Acceptance criteria - [ ] Request *Qualtrics* access for all members of Analytics and Content teams Qualtrics - [ ] Determine if we require any additional *Qualtrics* services that are not currently supported by our account.
1.0
Define requirements for Camille and OCE - As an *analytics and content team member* I want *to have access to Qualtrics* so that *I can view reports, curate data and manage surveys*. ## Acceptance criteria - [ ] Request *Qualtrics* access for all members of Analytics and Content teams Qualtrics - [ ] Determine if we require any additional *Qualtrics* services that are not currently supported by our account.
priority
define requirements for camille and oce as an analytics and content team member i want to have access to qualtrics so that i can view reports curate data and manage surveys acceptance criteria request qualtrics access for all members of analytics and content teams qualtrics determine if we require any additional qualtrics services that are not currently supported by our account
1
85,355
15,736,693,043
IssuesEvent
2021-03-30 01:13:22
benlazarine/datahog
https://api.github.com/repos/benlazarine/datahog
opened
CVE-2019-19118 (Medium) detected in Django-2.1.11.tar.gz
security vulnerability
## CVE-2019-19118 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.1.11.tar.gz</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz">https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz</a></p> <p>Path to dependency file: datahog/django/requirements.txt</p> <p>Path to vulnerable library: datahog/django/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-2.1.11.tar.gz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Django 2.1 before 2.1.15 and 2.2 before 2.2.8 allows unintended model editing. A Django model admin displaying inline related models, where the user has view-only permissions to a parent model but edit permissions to the inline model, would be presented with an editing UI, allowing POST requests, for updating the inline model. Directly editing the view-only parent model was not possible, but the parent model's save() method was called, triggering potential side effects, and causing pre and post-save signal handlers to be invoked. (To resolve this, the Django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable.) <p>Publish Date: 2019-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19118>CVE-2019-19118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118</a></p> <p>Release Date: 2019-12-02</p> <p>Fix Resolution: 2.1.15,2.2.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19118 (Medium) detected in Django-2.1.11.tar.gz - ## CVE-2019-19118 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.1.11.tar.gz</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz">https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz</a></p> <p>Path to dependency file: datahog/django/requirements.txt</p> <p>Path to vulnerable library: datahog/django/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-2.1.11.tar.gz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Django 2.1 before 2.1.15 and 2.2 before 2.2.8 allows unintended model editing. A Django model admin displaying inline related models, where the user has view-only permissions to a parent model but edit permissions to the inline model, would be presented with an editing UI, allowing POST requests, for updating the inline model. Directly editing the view-only parent model was not possible, but the parent model's save() method was called, triggering potential side effects, and causing pre and post-save signal handlers to be invoked. (To resolve this, the Django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable.) <p>Publish Date: 2019-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19118>CVE-2019-19118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118</a></p> <p>Release Date: 2019-12-02</p> <p>Fix Resolution: 2.1.15,2.2.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in django tar gz cve medium severity vulnerability vulnerable library django tar gz a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file datahog django requirements txt path to vulnerable library datahog django requirements txt dependency hierarchy x django tar gz vulnerable library vulnerability details django before and before allows unintended model editing a django model admin displaying inline related models where the user has view only permissions to a parent model but edit permissions to the inline model would be presented with an editing ui allowing post requests for updating the inline model directly editing the view only parent model was not possible but the parent model s save method was called triggering potential side effects and causing pre and post save signal handlers to be invoked to resolve this the django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
242,867
7,849,642,409
IssuesEvent
2018-06-20 05:04:01
borela/naomi
https://api.github.com/repos/borela/naomi
closed
Goto Definition from JSX to component
enhancement priority: medium
Hey, I would first like to say that I really like your package. My react code (with fragments, ligatures, flow, es6, ..) looks really good now thanks to you. The only thing I'm missing is to jump from JSX code to the used component. Below is a simple example of a component. It would be great if I could simple navigate to the Input or Label component with Goto Definition. Is this something that this package could do? Or should this be done by a different package? Or maybe this can't be done? ``` import * as React from "react"; import {uniqueId} from "lodash"; import {Label} from "./core/Label"; import {Input} from "./core/Input"; const InputWithLabel = ({text, value, onChange, children}) => { const id = uniqueId("id_") let label = children if (!label) label = text + ':' return <> <Label htmlFor={id}>{label}</Label> <Input id={id} type={type} value={value} onChange={onChange} placeholder={text}/> </> } ```
1.0
Goto Definition from JSX to component - Hey, I would first like to say that I really like your package. My react code (with fragments, ligatures, flow, es6, ..) looks really good now thanks to you. The only thing I'm missing is to jump from JSX code to the used component. Below is a simple example of a component. It would be great if I could simple navigate to the Input or Label component with Goto Definition. Is this something that this package could do? Or should this be done by a different package? Or maybe this can't be done? ``` import * as React from "react"; import {uniqueId} from "lodash"; import {Label} from "./core/Label"; import {Input} from "./core/Input"; const InputWithLabel = ({text, value, onChange, children}) => { const id = uniqueId("id_") let label = children if (!label) label = text + ':' return <> <Label htmlFor={id}>{label}</Label> <Input id={id} type={type} value={value} onChange={onChange} placeholder={text}/> </> } ```
priority
goto definition from jsx to component hey i would first like to say that i really like your package my react code with fragments ligatures flow looks really good now thanks to you the only thing i m missing is to jump from jsx code to the used component below is a simple example of a component it would be great if i could simple navigate to the input or label component with goto definition is this something that this package could do or should this be done by a different package or maybe this can t be done import as react from react import uniqueid from lodash import label from core label import input from core input const inputwithlabel text value onchange children const id uniqueid id let label children if label label text return label input id id type type value value onchange onchange placeholder text
1
649,346
21,280,069,093
IssuesEvent
2022-04-14 00:08:10
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
Group config check items by type in AppAssessment report
medium priority app-development
CPU and Memory Requests, CPU and Memory limits, Probes, etc. DoD - [ ] Group table items if applicable (resource usage, security, etc) (high)
1.0
Group config check items by type in AppAssessment report - CPU and Memory Requests, CPU and Memory limits, Probes, etc. DoD - [ ] Group table items if applicable (resource usage, security, etc) (high)
priority
group config check items by type in appassessment report cpu and memory requests cpu and memory limits probes etc dod group table items if applicable resource usage security etc high
1
198,065
6,969,402,027
IssuesEvent
2017-12-11 05:08:40
adventistmedia/adventisthub-support
https://api.github.com/repos/adventistmedia/adventisthub-support
closed
Colour Scheme
enhancement priority
Update colour scheme to closer resemble church brand and/or other church websites (e.g. search.adventistchurch.com)
1.0
Colour Scheme - Update colour scheme to closer resemble church brand and/or other church websites (e.g. search.adventistchurch.com)
priority
colour scheme update colour scheme to closer resemble church brand and or other church websites e g search adventistchurch com
1
477,666
13,766,032,431
IssuesEvent
2020-10-07 14:06:21
adaptlearning/adapt_framework
https://api.github.com/repos/adaptlearning/adapt_framework
closed
adapt-contrib-spoor: scorm tracking of course state not working if _shouldStoreResponses: false
bug high priority
### Subject of the issue/enhancement/features The tracking of which blocks are complete/incomplete doesn't appear to be working as it should ### Your environment * Framework master branch (but also occurs with v5.6.1)/spoor v3.5.0 (also checked master, still has issue) * Checked in IE11, FF, Chrome * Not checked devices but don't expect it to be any different ### Steps to reproduce Use `adapt create course` to set up a fresh install of Adapt. In config.json, set `_spoor._tracking._shouldSubmitScore` to `false`. run `grunt dev` then `grunt server-scorm`. Go into Presentation Components and complete a couple of interactive components. Reload the course. ### Expected behaviour The components completed prior to the reload should show as completed still ### Actual behaviour No components show as completed. ### Additional info The 'completion string' that is output to the console is correct BUT the value of `"c"` in the suspend_data never changes from its initial value of `"hAA"`, which doesn't seem anything close to being correct: ![image](https://user-images.githubusercontent.com/1249597/92953774-344e4480-f45a-11ea-8734-89f8ce05bfc8.png)
1.0
adapt-contrib-spoor: scorm tracking of course state not working if _shouldStoreResponses: false - ### Subject of the issue/enhancement/features The tracking of which blocks are complete/incomplete doesn't appear to be working as it should ### Your environment * Framework master branch (but also occurs with v5.6.1)/spoor v3.5.0 (also checked master, still has issue) * Checked in IE11, FF, Chrome * Not checked devices but don't expect it to be any different ### Steps to reproduce Use `adapt create course` to set up a fresh install of Adapt. In config.json, set `_spoor._tracking._shouldSubmitScore` to `false`. run `grunt dev` then `grunt server-scorm`. Go into Presentation Components and complete a couple of interactive components. Reload the course. ### Expected behaviour The components completed prior to the reload should show as completed still ### Actual behaviour No components show as completed. ### Additional info The 'completion string' that is output to the console is correct BUT the value of `"c"` in the suspend_data never changes from its initial value of `"hAA"`, which doesn't seem anything close to being correct: ![image](https://user-images.githubusercontent.com/1249597/92953774-344e4480-f45a-11ea-8734-89f8ce05bfc8.png)
priority
adapt contrib spoor scorm tracking of course state not working if shouldstoreresponses false subject of the issue enhancement features the tracking of which blocks are complete incomplete doesn t appear to be working as it should your environment framework master branch but also occurs with spoor also checked master still has issue checked in ff chrome not checked devices but don t expect it to be any different steps to reproduce use adapt create course to set up a fresh install of adapt in config json set spoor tracking shouldsubmitscore to false run grunt dev then grunt server scorm go into presentation components and complete a couple of interactive components reload the course expected behaviour the components completed prior to the reload should show as completed still actual behaviour no components show as completed additional info the completion string that is output to the console is correct but the value of c in the suspend data never changes from its initial value of haa which doesn t seem anything close to being correct
1
295,162
25,458,887,762
IssuesEvent
2022-11-24 16:30:44
DataDog/dd-trace-java
https://api.github.com/repos/DataDog/dd-trace-java
closed
TracerDebuggerIntegrationTest is flaky
flaky-test disabled
``` java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertFalse(Assert.java:64) at org.junit.Assert.assertFalse(Assert.java:74) at datadog.smoketest.TracerDebuggerIntegrationTest.testTracer(TracerDebuggerIntegrationTest.java:88) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) ```
1.0
TracerDebuggerIntegrationTest is flaky - ``` java.lang.AssertionError at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertFalse(Assert.java:64) at org.junit.Assert.assertFalse(Assert.java:74) at datadog.smoketest.TracerDebuggerIntegrationTest.testTracer(TracerDebuggerIntegrationTest.java:88) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84) at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115) at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104) at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) ```
non_priority
tracerdebuggerintegrationtest is flaky java lang assertionerror at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert assertfalse assert java at org junit assert assertfalse assert java at datadog smoketest tracerdebuggerintegrationtest testtracer tracerdebuggerintegrationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit platform commons util reflectionutils invokemethod reflectionutils java at org junit jupiter engine execution methodinvocation proceed methodinvocation java at org junit jupiter engine execution invocationinterceptorchain validatinginvocation proceed invocationinterceptorchain java at org junit jupiter engine extension timeoutextension intercept timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestablemethod timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestmethod timeoutextension java at org junit jupiter engine execution executableinvoker reflectiveinterceptorcall lambda ofvoidmethod executableinvoker java at org junit jupiter engine execution executableinvoker lambda invoke executableinvoker java at org junit jupiter engine execution invocationinterceptorchain interceptedinvocation proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain chainandinvoke invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain invoke invocationinterceptorchain java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine descriptor testmethodtestdescriptor lambda invoketestmethod testmethodtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor testmethodtestdescriptor invoketestmethod testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java
0
112,883
17,109,369,377
IssuesEvent
2021-07-10 01:45:43
vyas0189/CougarCS-Backend
https://api.github.com/repos/vyas0189/CougarCS-Backend
closed
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed
security vulnerability
## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: CougarCS-Backend/package.json</p> <p>Path to vulnerable library: CougarCS-Backend/node_modules/path-parse</p> <p> Dependency Hierarchy: - preset-env-7.14.7.tgz (Root Library) - babel-plugin-polyfill-corejs3-0.2.3.tgz - helper-define-polyfill-provider-0.2.3.tgz - resolve-1.20.0.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vyas0189/CougarCS-Backend/commit/41e2657038d5a56936fef604251da533f6670866">41e2657038d5a56936fef604251da533f6670866</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed - ## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: CougarCS-Backend/package.json</p> <p>Path to vulnerable library: CougarCS-Backend/node_modules/path-parse</p> <p> Dependency Hierarchy: - preset-env-7.14.7.tgz (Root Library) - babel-plugin-polyfill-corejs3-0.2.3.tgz - helper-define-polyfill-provider-0.2.3.tgz - resolve-1.20.0.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vyas0189/CougarCS-Backend/commit/41e2657038d5a56936fef604251da533f6670866">41e2657038d5a56936fef604251da533f6670866</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in path parse tgz autoclosed cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file cougarcs backend package json path to vulnerable library cougarcs backend node modules path parse dependency hierarchy preset env tgz root library babel plugin polyfill tgz helper define polyfill provider tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse step up your open source security game with whitesource
0
68,700
29,482,149,162
IssuesEvent
2023-06-02 06:51:44
IITC-CE/iitc-app-status
https://api.github.com/repos/IITC-CE/iitc-app-status
closed
🛑 Weblate service is down
status weblate-service
In [`6c04ec2`](https://github.com/IITC-CE/iitc-app-status/commit/6c04ec2feece866524484ba2e0fd37674631ef81 ), Weblate service (https://weblate.iitc.app) was **down**: - HTTP code: 500 - Response time: 1500 ms
1.0
🛑 Weblate service is down - In [`6c04ec2`](https://github.com/IITC-CE/iitc-app-status/commit/6c04ec2feece866524484ba2e0fd37674631ef81 ), Weblate service (https://weblate.iitc.app) was **down**: - HTTP code: 500 - Response time: 1500 ms
non_priority
🛑 weblate service is down in weblate service was down http code response time ms
0
239,032
7,785,947,066
IssuesEvent
2018-06-06 17:22:50
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
Websocket is not working with https
Priority:High Team:Frontend Type:Bug
The maps and graphics are not receiving data from websocket when used Https.
1.0
Websocket is not working with https - The maps and graphics are not receiving data from websocket when used Https.
priority
websocket is not working with https the maps and graphics are not receiving data from websocket when used https
1
223,650
17,614,306,251
IssuesEvent
2021-08-18 07:48:26
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
Fail to paste one secondary ADLS Gen2 blob/folder
🧪 testing :gear: adls gen2
**Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210818.3 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 20.04/MacOS Big Sur 11.4 **Architecture**: ia32\x64 **How Found**: From running test cases **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one ADLS Gen2 storage account -> Blob Containers. 2. Open one blob container -> Upload one blob to it. 3. Right click the blob container -> Click 'Open Secondary Location'. 4. Copy the blob -> Paste it to another blob container. 5. Check whether succeed to paste. ## Expected Experience ## Succeed to paste. ## Actual Experience ## Fail to paste with the below error. Error detail: ` "message": "\"failed to perform copy command due to error: cannot start job due to error: cannot list files due to reason -> github.com/Azure/azure-storage-blob-go/azblob.newStorageError, /home/vsts/go/pkg/mod/github.com/!azure/azure-storage-blob-go@v0.10.1-0.20210407023846` ![image](https://user-images.githubusercontent.com/41351993/129858215-d2df10ef-4616-4f76-8ac7-28a885149d94.png)
1.0
Fail to paste one secondary ADLS Gen2 blob/folder - **Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210818.3 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 20.04/MacOS Big Sur 11.4 **Architecture**: ia32\x64 **How Found**: From running test cases **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one ADLS Gen2 storage account -> Blob Containers. 2. Open one blob container -> Upload one blob to it. 3. Right click the blob container -> Click 'Open Secondary Location'. 4. Copy the blob -> Paste it to another blob container. 5. Check whether succeed to paste. ## Expected Experience ## Succeed to paste. ## Actual Experience ## Fail to paste with the below error. Error detail: ` "message": "\"failed to perform copy command due to error: cannot start job due to error: cannot list files due to reason -> github.com/Azure/azure-storage-blob-go/azblob.newStorageError, /home/vsts/go/pkg/mod/github.com/!azure/azure-storage-blob-go@v0.10.1-0.20210407023846` ![image](https://user-images.githubusercontent.com/41351993/129858215-d2df10ef-4616-4f76-8ac7-28a885149d94.png)
non_priority
fail to paste one secondary adls blob folder storage explorer version dev build number branch main platform os windows linux ubuntu macos big sur architecture how found from running test cases regression from not a regression steps to reproduce expand one adls storage account blob containers open one blob container upload one blob to it right click the blob container click open secondary location copy the blob paste it to another blob container check whether succeed to paste expected experience succeed to paste actual experience fail to paste with the below error error detail message failed to perform copy command due to error cannot start job due to error cannot list files due to reason github com azure azure storage blob go azblob newstorageerror home vsts go pkg mod github com azure azure storage blob go
0
241,114
7,808,907,450
IssuesEvent
2018-06-11 21:48:59
tomaka/winit
https://api.github.com/repos/tomaka/winit
closed
Segmentation fault in XIFreeDeviceInfo
difficulty: average platform: X11 priority: normal status: waiting type: bug
This was posted on jwilm/alacritty#895: > Version: master branch, revision f68558e9dcce97e0d4d842a2d5b62e31e4124210 > > I ran it with `rust-gdb ./target/debug/alacritty`, waited for the crash, and got this backtrace: ``` Thread 1 "alacritty" received signal SIGSEGV, Segmentation fault. 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6 (gdb) bt #0 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6 #1 0x0000555555d86967 in winit::platform::platform::x11::{{impl}}::drop (self=0x7fffffff5070) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:641 #2 0x0000555555d54395 in core::ptr::drop_in_place<winit::platform::platform::x11::DeviceInfo> () at /checkout/src/libcore/ptr.rs:60 #3 0x0000555555d86353 in winit::platform::platform::x11::EventsLoop::init_device (self=0x7fffffff8768, device=17) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:575 #4 0x0000555555745c9a in winit::platform::platform::x11::EventsLoop::process_event<&mut closure> (self=0x7fffffff8768, xev=0x7fffffff7818, callback=0x7fffffff78e0) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:553 #5 0x0000555555742313 in winit::platform::platform::x11::EventsLoop::run_forever<closure> (self=0x7fffffff8768, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:174 #6 0x000055555574201a in winit::platform::platform::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/mod.rs:387 #7 0x0000555555741e5d in winit::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/lib.rs:219 #8 0x000055555579d7ed in alacritty::window::Window::wait_events<closure> (self=0x7fffffff8760, func=closure = {...}) at /home/mrmetric/Downloads/source/alacritty/src/window.rs:274 #9 0x00005555556b0101 in alacritty::event::Processor<alacritty::event_loop::Notifier>::process_events<alacritty::event_loop::Notifier> (self=0x7fffffffc178, term=0x7fffef8d5e10, window=0x7fffffff8760) at /home/mrmetric/Downloads/source/alacritty/src/event.rs:375 #10 0x00005555557a0f52 in alacritty::run (config=Config = {...}, options=Options = {...}) at src/main.rs:170 #11 0x000055555579faac in alacritty::main () at src/main.rs:45 ``` I haven't had a chance to look at it myself yet.
1.0
Segmentation fault in XIFreeDeviceInfo - This was posted on jwilm/alacritty#895: > Version: master branch, revision f68558e9dcce97e0d4d842a2d5b62e31e4124210 > > I ran it with `rust-gdb ./target/debug/alacritty`, waited for the crash, and got this backtrace: ``` Thread 1 "alacritty" received signal SIGSEGV, Segmentation fault. 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6 (gdb) bt #0 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6 #1 0x0000555555d86967 in winit::platform::platform::x11::{{impl}}::drop (self=0x7fffffff5070) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:641 #2 0x0000555555d54395 in core::ptr::drop_in_place<winit::platform::platform::x11::DeviceInfo> () at /checkout/src/libcore/ptr.rs:60 #3 0x0000555555d86353 in winit::platform::platform::x11::EventsLoop::init_device (self=0x7fffffff8768, device=17) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:575 #4 0x0000555555745c9a in winit::platform::platform::x11::EventsLoop::process_event<&mut closure> (self=0x7fffffff8768, xev=0x7fffffff7818, callback=0x7fffffff78e0) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:553 #5 0x0000555555742313 in winit::platform::platform::x11::EventsLoop::run_forever<closure> (self=0x7fffffff8768, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:174 #6 0x000055555574201a in winit::platform::platform::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/mod.rs:387 #7 0x0000555555741e5d in winit::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...}) at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/lib.rs:219 #8 0x000055555579d7ed in alacritty::window::Window::wait_events<closure> (self=0x7fffffff8760, func=closure = {...}) at /home/mrmetric/Downloads/source/alacritty/src/window.rs:274 #9 0x00005555556b0101 in alacritty::event::Processor<alacritty::event_loop::Notifier>::process_events<alacritty::event_loop::Notifier> (self=0x7fffffffc178, term=0x7fffef8d5e10, window=0x7fffffff8760) at /home/mrmetric/Downloads/source/alacritty/src/event.rs:375 #10 0x00005555557a0f52 in alacritty::run (config=Config = {...}, options=Options = {...}) at src/main.rs:170 #11 0x000055555579faac in alacritty::main () at src/main.rs:45 ``` I haven't had a chance to look at it myself yet.
priority
segmentation fault in xifreedeviceinfo this was posted on jwilm alacritty version master branch revision i ran it with rust gdb target debug alacritty waited for the crash and got this backtrace thread alacritty received signal sigsegv segmentation fault in xifreedeviceinfo from usr lib libxi so gdb bt in xifreedeviceinfo from usr lib libxi so in winit platform platform impl drop self at home mrmetric cargo registry src github com winit src platform linux mod rs in core ptr drop in place at checkout src libcore ptr rs in winit platform platform eventsloop init device self device at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop process event self xev callback at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src platform linux mod rs in winit eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src lib rs in alacritty window window wait events self func closure at home mrmetric downloads source alacritty src window rs in alacritty event processor process events self term window at home mrmetric downloads source alacritty src event rs in alacritty run config config options options at src main rs in alacritty main at src main rs i haven t had a chance to look at it myself yet
1
281,458
8,695,531,110
IssuesEvent
2018-12-04 15:23:08
cosmos/voyager
https://api.github.com/repos/cosmos/voyager
closed
Proposal Creation feedbacks
bug 🐞 governance-1 :ballot_box: high priority
UI Version: 0.X.X Description: Governance/Create Proposal After entering 65 characters in the Title or 200 characters in the Description, the Create Proposal button is disabled but is not clear for which reason It would be nice to have feedbacks and/or highlighting the box that is blocking the creation. Voting starting date is always invalid. Subset of #1690
1.0
Proposal Creation feedbacks - UI Version: 0.X.X Description: Governance/Create Proposal After entering 65 characters in the Title or 200 characters in the Description, the Create Proposal button is disabled but is not clear for which reason It would be nice to have feedbacks and/or highlighting the box that is blocking the creation. Voting starting date is always invalid. Subset of #1690
priority
proposal creation feedbacks ui version x x description governance create proposal after entering characters in the title or characters in the description the create proposal button is disabled but is not clear for which reason it would be nice to have feedbacks and or highlighting the box that is blocking the creation voting starting date is always invalid subset of
1
47,392
24,985,021,003
IssuesEvent
2022-11-02 14:30:11
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
scale testing: Gigarepo indexing is not completing
team/search-core performance
Indexing of the gigarepo on the scale testing instance does not appear to be completing: https://scaletesting.sgdev.org/gigarepo/-/settings/index
True
scale testing: Gigarepo indexing is not completing - Indexing of the gigarepo on the scale testing instance does not appear to be completing: https://scaletesting.sgdev.org/gigarepo/-/settings/index
non_priority
scale testing gigarepo indexing is not completing indexing of the gigarepo on the scale testing instance does not appear to be completing
0
288,214
21,691,033,520
IssuesEvent
2022-05-09 15:22:53
GaloyMoney/galoy
https://api.github.com/repos/GaloyMoney/galoy
opened
possible race condition in payment
bug documentation need investigation lightning
I had an issue where: 1/ the recipient was using BBW (using the tipping page) 2/ two person were trying to pay the invoice at the same time. one with BBW. one with an external wallet. both payment initiated within less than 1s timeframe. the external wallet was probably initiated first. the payment with the external wallet went through (success), the payment from BBW showed an error. when looking at both wallets, both had a payment sent. it was expected for the external wallet because the payment succeed. for the BBW wallet, the payment showed as a failure, so this was not expected. it may be a race condition when an invoice is being paid at the same time both from outside, using lnd, and inside with an intraledger payment.
1.0
possible race condition in payment - I had an issue where: 1/ the recipient was using BBW (using the tipping page) 2/ two person were trying to pay the invoice at the same time. one with BBW. one with an external wallet. both payment initiated within less than 1s timeframe. the external wallet was probably initiated first. the payment with the external wallet went through (success), the payment from BBW showed an error. when looking at both wallets, both had a payment sent. it was expected for the external wallet because the payment succeed. for the BBW wallet, the payment showed as a failure, so this was not expected. it may be a race condition when an invoice is being paid at the same time both from outside, using lnd, and inside with an intraledger payment.
non_priority
possible race condition in payment i had an issue where the recipient was using bbw using the tipping page two person were trying to pay the invoice at the same time one with bbw one with an external wallet both payment initiated within less than timeframe the external wallet was probably initiated first the payment with the external wallet went through success the payment from bbw showed an error when looking at both wallets both had a payment sent it was expected for the external wallet because the payment succeed for the bbw wallet the payment showed as a failure so this was not expected it may be a race condition when an invoice is being paid at the same time both from outside using lnd and inside with an intraledger payment
0
311,889
9,540,011,972
IssuesEvent
2019-04-30 18:24:31
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
PortSwitch primitive system
priority: medium team: russ type: feature request
As discussed with a few of you, I've got the "PortSwitch" system mostly coded up now, and will PR it this weekend. It's a simple system that offers many ports on the input, but passes-through only one of them to the output (with one additional input for the port_selector). I imagine this being useful for combining many low-level controller primitives (e.g. DifferentialIK, PlanEval, hybrid control, etc) with all of them wired up, but only one of them getting called thanks to this switch. cc @kmuhlrad @pangtao22
1.0
PortSwitch primitive system - As discussed with a few of you, I've got the "PortSwitch" system mostly coded up now, and will PR it this weekend. It's a simple system that offers many ports on the input, but passes-through only one of them to the output (with one additional input for the port_selector). I imagine this being useful for combining many low-level controller primitives (e.g. DifferentialIK, PlanEval, hybrid control, etc) with all of them wired up, but only one of them getting called thanks to this switch. cc @kmuhlrad @pangtao22
priority
portswitch primitive system as discussed with a few of you i ve got the portswitch system mostly coded up now and will pr it this weekend it s a simple system that offers many ports on the input but passes through only one of them to the output with one additional input for the port selector i imagine this being useful for combining many low level controller primitives e g differentialik planeval hybrid control etc with all of them wired up but only one of them getting called thanks to this switch cc kmuhlrad
1
311,070
9,527,870,130
IssuesEvent
2019-04-29 06:29:32
Steven-Wagner/Movie-Suggestor
https://api.github.com/repos/Steven-Wagner/Movie-Suggestor
opened
Show a response each time user gets a duplicate error in Reviews
High Priority bug
When a user submits a review that has an error; the first time the user sees the change on the screen and is properly informed. However, if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time. Leading the user to think that the event was not registered by the client.
1.0
Show a response each time user gets a duplicate error in Reviews - When a user submits a review that has an error; the first time the user sees the change on the screen and is properly informed. However, if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time. Leading the user to think that the event was not registered by the client.
priority
show a response each time user gets a duplicate error in reviews when a user submits a review that has an error the first time the user sees the change on the screen and is properly informed however if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time leading the user to think that the event was not registered by the client
1
149,832
5,729,234,361
IssuesEvent
2017-04-21 05:09:34
dhowe/ReadersJS
https://api.github.com/repos/dhowe/ReadersJS
closed
Update website to point to readers-js-live
priority: High
LOCATIONS (others?): http://thereadersproject.org/elc3.html#ELC3 http://thereadersproject.org/index.html (add link to left nav?)
1.0
Update website to point to readers-js-live - LOCATIONS (others?): http://thereadersproject.org/elc3.html#ELC3 http://thereadersproject.org/index.html (add link to left nav?)
priority
update website to point to readers js live locations others add link to left nav
1
3,825
6,334,056,315
IssuesEvent
2017-07-26 15:53:05
opensensorhub/sensorml-editor
https://api.github.com/repos/opensensorhub/sensorml-editor
closed
Bring Ontology Search capabilities back into editor
Requirement
We previously had the ability to search an ontology for terms and them select. Previously this was done by downloading the ontology and doing a search directly on the "local" version. This search however, can be done by using the search capabilities on the Ontology (ORR) web services. For example, see: http://sensorml.com/orr/#st This search can also be done using sparQL: http://sensorml.com/ont/sparql.html
1.0
Bring Ontology Search capabilities back into editor - We previously had the ability to search an ontology for terms and them select. Previously this was done by downloading the ontology and doing a search directly on the "local" version. This search however, can be done by using the search capabilities on the Ontology (ORR) web services. For example, see: http://sensorml.com/orr/#st This search can also be done using sparQL: http://sensorml.com/ont/sparql.html
non_priority
bring ontology search capabilities back into editor we previously had the ability to search an ontology for terms and them select previously this was done by downloading the ontology and doing a search directly on the local version this search however can be done by using the search capabilities on the ontology orr web services for example see this search can also be done using sparql
0
629,309
20,029,055,061
IssuesEvent
2022-02-02 01:45:53
apcountryman/picolibrary
https://api.github.com/repos/apcountryman/picolibrary
closed
Add GPIO I/O pin concept
priority-normal status-awaiting_review type-feature
Add GPIO I/O pin concept (`::picolibrary::GPIO::IO_Pin_Concept`) and associated mock (`::picolibrary::Testing::Unit::GPIO::Mock_IO_Pin`). - [x] The `IO_Pin_Concept` class should be defined in the `include/picolibrary/gpio.h`/`source/picolibrary/gpio.cc` header/source file pair - [x] The `Mock_IO_Pin` class should be defined in the `include/picolibrary/testing/unit/gpio.h`/`source/picolibrary/testing/unit/gpio.cc` header/source file pair - [x] IO pins should be expected to support the following operations: - [x] `IO_Pin_Concept() noexcept = default;` - [x] `IO_Pin_Concept( IO_Pin_Concept && source ) noexcept = default;` - [x] `~IO_Pin_Concept() noexcept = default;` - [x] `auto operator=( IO_Pin_Concept && expression ) noexcept -> IO_Pin_Concept & = default;` - [x] `void initialize( Initial_Pin_State initial_pin_state = Initial_Pin_State::LOW ) noexcept;` Initialize the pin's hardware - [x] `auto is_low() const noexcept -> bool;`: Check if the pin is in the low state - [x] `auto is_high() const noexcept -> bool;`: Check if the pin is in the high state - [x] `void transition_to_low() noexcept;`: Transition the pin to the low state - [x] `void transition_to_high() noexcept;`: Transition the pin to the high state - [x] `void toggle() noexcept;`: Toggle the pin state - [x] The `Mock_IO_Pin` class should include a movable mock handle
1.0
Add GPIO I/O pin concept - Add GPIO I/O pin concept (`::picolibrary::GPIO::IO_Pin_Concept`) and associated mock (`::picolibrary::Testing::Unit::GPIO::Mock_IO_Pin`). - [x] The `IO_Pin_Concept` class should be defined in the `include/picolibrary/gpio.h`/`source/picolibrary/gpio.cc` header/source file pair - [x] The `Mock_IO_Pin` class should be defined in the `include/picolibrary/testing/unit/gpio.h`/`source/picolibrary/testing/unit/gpio.cc` header/source file pair - [x] IO pins should be expected to support the following operations: - [x] `IO_Pin_Concept() noexcept = default;` - [x] `IO_Pin_Concept( IO_Pin_Concept && source ) noexcept = default;` - [x] `~IO_Pin_Concept() noexcept = default;` - [x] `auto operator=( IO_Pin_Concept && expression ) noexcept -> IO_Pin_Concept & = default;` - [x] `void initialize( Initial_Pin_State initial_pin_state = Initial_Pin_State::LOW ) noexcept;` Initialize the pin's hardware - [x] `auto is_low() const noexcept -> bool;`: Check if the pin is in the low state - [x] `auto is_high() const noexcept -> bool;`: Check if the pin is in the high state - [x] `void transition_to_low() noexcept;`: Transition the pin to the low state - [x] `void transition_to_high() noexcept;`: Transition the pin to the high state - [x] `void toggle() noexcept;`: Toggle the pin state - [x] The `Mock_IO_Pin` class should include a movable mock handle
priority
add gpio i o pin concept add gpio i o pin concept picolibrary gpio io pin concept and associated mock picolibrary testing unit gpio mock io pin the io pin concept class should be defined in the include picolibrary gpio h source picolibrary gpio cc header source file pair the mock io pin class should be defined in the include picolibrary testing unit gpio h source picolibrary testing unit gpio cc header source file pair io pins should be expected to support the following operations io pin concept noexcept default io pin concept io pin concept source noexcept default io pin concept noexcept default auto operator io pin concept expression noexcept io pin concept default void initialize initial pin state initial pin state initial pin state low noexcept initialize the pin s hardware auto is low const noexcept bool check if the pin is in the low state auto is high const noexcept bool check if the pin is in the high state void transition to low noexcept transition the pin to the low state void transition to high noexcept transition the pin to the high state void toggle noexcept toggle the pin state the mock io pin class should include a movable mock handle
1
75,205
14,409,046,922
IssuesEvent
2020-12-04 01:15:33
UBC-Thunderbots/Software
https://api.github.com/repos/UBC-Thunderbots/Software
closed
Investigate Warnings from SimulatedTests
Difficulty - 13 G3 - Code Quality G6 - Gameplay and Navigation T - Maintenance
### Description of the task Several SimulatedTests cause warnings that fill the logs and are highly suspicious and may indicate a bug. One example is `free_kick_play_test`, which causes `Could not draw a line from the goalie to the ball through the defender crease` and `Error updating robot state, stopping` warnings. To start this investigation, run `//software/simulated_tests:free_kick_play_test` in debug mode in gdb with breakpoints where those warnings are logged. <!-- What does this work depend on? What interface will this work use or create? What are the main components of the task? Where does this work fit in the larger project? It is important to define this task sufficiently so that an untrained team member can take it on and know where to start. Feel free to link to resources or other team member which could guide the assignee to complete the task --> ### Acceptance criteria <!-- Checkbox list that outlines what needs to be done in order for this task to be considered "complete". Specify any implementation requirements such as data structures, functionalities, testing requirements, documentation, etc. --> - [ ] Explain why we get so many warnings from simulated tests - [ ] Figure out the best way to avoid these warnings, e.g. fixing an underlying bug or suppressing warnings that aren't helpful ### Blocked By <!-- List all other issues that need to be completed before this one, ex: - #123 - #374 -->
1.0
Investigate Warnings from SimulatedTests - ### Description of the task Several SimulatedTests cause warnings that fill the logs and are highly suspicious and may indicate a bug. One example is `free_kick_play_test`, which causes `Could not draw a line from the goalie to the ball through the defender crease` and `Error updating robot state, stopping` warnings. To start this investigation, run `//software/simulated_tests:free_kick_play_test` in debug mode in gdb with breakpoints where those warnings are logged. <!-- What does this work depend on? What interface will this work use or create? What are the main components of the task? Where does this work fit in the larger project? It is important to define this task sufficiently so that an untrained team member can take it on and know where to start. Feel free to link to resources or other team member which could guide the assignee to complete the task --> ### Acceptance criteria <!-- Checkbox list that outlines what needs to be done in order for this task to be considered "complete". Specify any implementation requirements such as data structures, functionalities, testing requirements, documentation, etc. --> - [ ] Explain why we get so many warnings from simulated tests - [ ] Figure out the best way to avoid these warnings, e.g. fixing an underlying bug or suppressing warnings that aren't helpful ### Blocked By <!-- List all other issues that need to be completed before this one, ex: - #123 - #374 -->
non_priority
investigate warnings from simulatedtests description of the task several simulatedtests cause warnings that fill the logs and are highly suspicious and may indicate a bug one example is free kick play test which causes could not draw a line from the goalie to the ball through the defender crease and error updating robot state stopping warnings to start this investigation run software simulated tests free kick play test in debug mode in gdb with breakpoints where those warnings are logged what does this work depend on what interface will this work use or create what are the main components of the task where does this work fit in the larger project it is important to define this task sufficiently so that an untrained team member can take it on and know where to start feel free to link to resources or other team member which could guide the assignee to complete the task acceptance criteria checkbox list that outlines what needs to be done in order for this task to be considered complete specify any implementation requirements such as data structures functionalities testing requirements documentation etc explain why we get so many warnings from simulated tests figure out the best way to avoid these warnings e g fixing an underlying bug or suppressing warnings that aren t helpful blocked by list all other issues that need to be completed before this one ex
0
135,889
11,028,500,045
IssuesEvent
2019-12-06 11:50:20
UglyToad/PdfPig
https://api.github.com/repos/UglyToad/PdfPig
closed
InvalidOperationException: 'No charstring sequence with the name /one in this font.'
bug testing
Opening the page 192 of [this](https://www.toshiba-storage.com/wp-content/uploads/2018/01/UM_Canvio_Basics.pdf) document throws `System.InvalidOperationException: 'No charstring sequence with the name /one in this font.'` in `Type2CharStrings.Generate(string name)`. __Full details:__ System.InvalidOperationException HResult=0x80131509 Message=No charstring sequence with the name /one in this font. Source=UglyToad.PdfPig StackTrace: at UglyToad.PdfPig.Fonts.CompactFontFormat.CharStrings.Type2CharStrings.Generate(String name) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CharStrings\Type2CharStrings.cs:line 63 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.<>c__DisplayClass16_0.<GetCharacterBoundingBox>b__1(Type2CharStrings x) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 41 at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 40 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFontProgram.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFontProgram.cs:line 48 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.<>c__DisplayClass20_0.<GetBoundingBoxInGlyphSpace>b__1(CompactFontFormatFontProgram x) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 185 at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBoxInGlyphSpace(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 169 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBox(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 124 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ShowText(IInputBytes bytes) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 198 at UglyToad.PdfPig.Graphics.Operations.TextShowing.ShowText.Run(IOperationContext operationContext) in \PdfPig\src\UglyToad.PdfPig\Graphics\Operations\TextShowing\ShowText.cs:line 68 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ProcessOperations(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 111 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.Process(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 102 at UglyToad.PdfPig.Parser.PageFactory.GetContent(IReadOnlyList'1 contentBytes, CropBox cropBox, UserSpaceUnit userSpaceUnit, PageRotationDegrees rotation, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 129 at UglyToad.PdfPig.Parser.PageFactory.Create(Int32 number, DictionaryToken dictionary, PageTreeMembers pageTreeMembers, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 112 at UglyToad.PdfPig.Content.Pages.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\Content\Pages.cs:line 66 at UglyToad.PdfPig.PdfDocument.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\PdfDocument.cs:line 150
1.0
InvalidOperationException: 'No charstring sequence with the name /one in this font.' - Opening the page 192 of [this](https://www.toshiba-storage.com/wp-content/uploads/2018/01/UM_Canvio_Basics.pdf) document throws `System.InvalidOperationException: 'No charstring sequence with the name /one in this font.'` in `Type2CharStrings.Generate(string name)`. __Full details:__ System.InvalidOperationException HResult=0x80131509 Message=No charstring sequence with the name /one in this font. Source=UglyToad.PdfPig StackTrace: at UglyToad.PdfPig.Fonts.CompactFontFormat.CharStrings.Type2CharStrings.Generate(String name) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CharStrings\Type2CharStrings.cs:line 63 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.<>c__DisplayClass16_0.<GetCharacterBoundingBox>b__1(Type2CharStrings x) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 41 at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 40 at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFontProgram.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFontProgram.cs:line 48 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.<>c__DisplayClass20_0.<GetBoundingBoxInGlyphSpace>b__1(CompactFontFormatFontProgram x) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 185 at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBoxInGlyphSpace(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 169 at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBox(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 124 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ShowText(IInputBytes bytes) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 198 at UglyToad.PdfPig.Graphics.Operations.TextShowing.ShowText.Run(IOperationContext operationContext) in \PdfPig\src\UglyToad.PdfPig\Graphics\Operations\TextShowing\ShowText.cs:line 68 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ProcessOperations(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 111 at UglyToad.PdfPig.Graphics.ContentStreamProcessor.Process(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 102 at UglyToad.PdfPig.Parser.PageFactory.GetContent(IReadOnlyList'1 contentBytes, CropBox cropBox, UserSpaceUnit userSpaceUnit, PageRotationDegrees rotation, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 129 at UglyToad.PdfPig.Parser.PageFactory.Create(Int32 number, DictionaryToken dictionary, PageTreeMembers pageTreeMembers, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 112 at UglyToad.PdfPig.Content.Pages.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\Content\Pages.cs:line 66 at UglyToad.PdfPig.PdfDocument.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\PdfDocument.cs:line 150
non_priority
invalidoperationexception no charstring sequence with the name one in this font opening the page of document throws system invalidoperationexception no charstring sequence with the name one in this font in generate string name full details system invalidoperationexception hresult message no charstring sequence with the name one in this font source uglytoad pdfpig stacktrace at uglytoad pdfpig fonts compactfontformat charstrings generate string name in pdfpig src uglytoad pdfpig fonts compactfontformat charstrings cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfont c b x in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfont cs line at uglytoad pdfpig util union match action first action second in pdfpig src uglytoad pdfpig util union cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfont getcharacterboundingbox string charactername in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfont cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfontprogram getcharacterboundingbox string charactername in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfontprogram cs line at uglytoad pdfpig fonts simple c b compactfontformatfontprogram x in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig util union match action first action second in pdfpig src uglytoad pdfpig util union cs line at uglytoad pdfpig fonts simple getboundingboxinglyphspace charactercode in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig fonts simple getboundingbox charactercode in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig graphics contentstreamprocessor showtext iinputbytes bytes in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig graphics operations textshowing showtext run ioperationcontext operationcontext in pdfpig src uglytoad pdfpig graphics operations textshowing showtext cs line at uglytoad pdfpig graphics contentstreamprocessor processoperations ireadonlylist operations in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig graphics contentstreamprocessor process ireadonlylist operations in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig parser pagefactory getcontent ireadonlylist contentbytes cropbox cropbox userspaceunit userspaceunit pagerotationdegrees rotation boolean islenientparsing in pdfpig src uglytoad pdfpig parser pagefactory cs line at uglytoad pdfpig parser pagefactory create number dictionarytoken dictionary pagetreemembers pagetreemembers boolean islenientparsing in pdfpig src uglytoad pdfpig parser pagefactory cs line at uglytoad pdfpig content pages getpage pagenumber in pdfpig src uglytoad pdfpig content pages cs line at uglytoad pdfpig pdfdocument getpage pagenumber in pdfpig src uglytoad pdfpig pdfdocument cs line
0
293,580
8,997,804,453
IssuesEvent
2019-02-02 15:59:44
Qiskit/qiskit-terra
https://api.github.com/repos/Qiskit/qiskit-terra
closed
change/extend how execution timeouts work
priority: low type: discussion
This issue is meant to start a discussion about how execution timeouts work. Currently, the timeout setting of a job starts the clock as soon as a job is submitted. If the job times out, the SDK stops waiting for the result but does not remove the job from the remote queue (this is not currently possible with the API). It may be better to add a timeout, or change the timeout, such that it starts when the circuits in the qobj start executing. update 12/2018: With asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work, however there is still a use case. A user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs. Some queuing systems such as LSF support setting a run time limit.
1.0
change/extend how execution timeouts work - This issue is meant to start a discussion about how execution timeouts work. Currently, the timeout setting of a job starts the clock as soon as a job is submitted. If the job times out, the SDK stops waiting for the result but does not remove the job from the remote queue (this is not currently possible with the API). It may be better to add a timeout, or change the timeout, such that it starts when the circuits in the qobj start executing. update 12/2018: With asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work, however there is still a use case. A user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs. Some queuing systems such as LSF support setting a run time limit.
priority
change extend how execution timeouts work this issue is meant to start a discussion about how execution timeouts work currently the timeout setting of a job starts the clock as soon as a job is submitted if the job times out the sdk stops waiting for the result but does not remove the job from the remote queue this is not currently possible with the api it may be better to add a timeout or change the timeout such that it starts when the circuits in the qobj start executing update with asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work however there is still a use case a user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs some queuing systems such as lsf support setting a run time limit
1
565,103
16,748,971,751
IssuesEvent
2021-06-11 19:38:01
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Make linked system better
Category: Gameplay Needs Task Plan Priority: Low Squad: Mountain Goat Status: Won't Fix Type: Feature
Inspired by this changes https://github.com/StrangeLoopGames/Eco/pull/6541 We have another items with this problem Crafting stations with fuel, some lights(braziers) and I think we will have a lot of another objects in future. So, currentrly we have 2 option for inventories in our techtree: None and Linked. I assume we can add another one - 'Excluded' that allow to see another **linked** inventories in its storage, but it will not shown in another storages. That you did for Pots, but add this customization to all items in techtree.
1.0
Make linked system better - Inspired by this changes https://github.com/StrangeLoopGames/Eco/pull/6541 We have another items with this problem Crafting stations with fuel, some lights(braziers) and I think we will have a lot of another objects in future. So, currentrly we have 2 option for inventories in our techtree: None and Linked. I assume we can add another one - 'Excluded' that allow to see another **linked** inventories in its storage, but it will not shown in another storages. That you did for Pots, but add this customization to all items in techtree.
priority
make linked system better inspired by this changes we have another items with this problem crafting stations with fuel some lights braziers and i think we will have a lot of another objects in future so currentrly we have option for inventories in our techtree none and linked i assume we can add another one excluded that allow to see another linked inventories in its storage but it will not shown in another storages that you did for pots but add this customization to all items in techtree
1
482,603
13,910,475,133
IssuesEvent
2020-10-20 16:07:09
stevencohn/OneMore
https://api.github.com/repos/stevencohn/OneMore
closed
Inline markdown conversions
enhancement low priority
Inline markdown conversions. This is a biggie and I'm not sure of its usefulness since the OneNote UI is so feature rich already. But the idea is to associate markdown directives with styles and/or custom styles, invoke an Apply Markdown command and replace all inline directives with styling. I think there are alternatives that convert markdown to HTML and then you can paste HTML into OneNote.
1.0
Inline markdown conversions - Inline markdown conversions. This is a biggie and I'm not sure of its usefulness since the OneNote UI is so feature rich already. But the idea is to associate markdown directives with styles and/or custom styles, invoke an Apply Markdown command and replace all inline directives with styling. I think there are alternatives that convert markdown to HTML and then you can paste HTML into OneNote.
priority
inline markdown conversions inline markdown conversions this is a biggie and i m not sure of its usefulness since the onenote ui is so feature rich already but the idea is to associate markdown directives with styles and or custom styles invoke an apply markdown command and replace all inline directives with styling i think there are alternatives that convert markdown to html and then you can paste html into onenote
1
566,479
16,822,517,730
IssuesEvent
2021-06-17 14:36:36
AugurProject/turbo
https://api.github.com/repos/AugurProject/turbo
opened
Update the add liquidity button on the market page - needs to stand out more for LPs
High Priority enhancement
https://www.figma.com/file/6y4nvjfeVZwzwKcoXB0neq/Augur-Simplified-UI?node-id=2919%3A5640 ![image](https://user-images.githubusercontent.com/39496225/122417766-a459a000-cf81-11eb-988d-58fe0ca53877.png)
1.0
Update the add liquidity button on the market page - needs to stand out more for LPs - https://www.figma.com/file/6y4nvjfeVZwzwKcoXB0neq/Augur-Simplified-UI?node-id=2919%3A5640 ![image](https://user-images.githubusercontent.com/39496225/122417766-a459a000-cf81-11eb-988d-58fe0ca53877.png)
priority
update the add liquidity button on the market page needs to stand out more for lps
1
301,465
26,050,447,828
IssuesEvent
2022-12-22 18:06:13
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: ledger/nodes=6/multi-az failed
C-test-failure O-robot O-roachtest branch-master T-kv
roachtest.ledger/nodes=6/multi-az [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/ledger/nodes=6/multi-az) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22): ``` test artifacts and logs in: /artifacts/ledger/nodes=6/multi-az/run_1 (test_impl.go:291).Fatal: output in run_163643.333865619_n7_workload_run_ledger: ./workload run ledger --init --histograms=perf/stats.json --concurrency=192 --duration=10m {pgurl:1-2} returned: COMMAND_PROBLEM: exit status 1 (test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*ledger/nodes=6/multi-az.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22677 Epic CRDB-18656
2.0
roachtest: ledger/nodes=6/multi-az failed - roachtest.ledger/nodes=6/multi-az [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/ledger/nodes=6/multi-az) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22): ``` test artifacts and logs in: /artifacts/ledger/nodes=6/multi-az/run_1 (test_impl.go:291).Fatal: output in run_163643.333865619_n7_workload_run_ledger: ./workload run ledger --init --histograms=perf/stats.json --concurrency=192 --duration=10m {pgurl:1-2} returned: COMMAND_PROBLEM: exit status 1 (test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*ledger/nodes=6/multi-az.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22677 Epic CRDB-18656
non_priority
roachtest ledger nodes multi az failed roachtest ledger nodes multi az with on master test artifacts and logs in artifacts ledger nodes multi az run test impl go fatal output in run workload run ledger workload run ledger init histograms perf stats json concurrency duration pgurl returned command problem exit status test impl go fatal monitor failure monitor task failed t fatal was called parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb kv triage jira issue crdb epic crdb
0
824,120
31,141,641,677
IssuesEvent
2023-08-16 00:51:09
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
Bestial Wrath
Class: Hunter Spell Priority: High
**Description:** if you try to use Bestial Wrath to break from paladin's repentance it says you can't use that while stunned,didn't try it with cyclone yet. Edit : doesn't work with Gouge too **How to reproduce:** cast repentance on BM hunter and use bestial wrath (The Beast Within sub-talent is needed to break the CC's) **How it should work:** Since Wotlk expansion patch changes : Bestial Wrath and Ability hunter beastwithin [[The Beast Within]](https://wowpedia.fandom.com/wiki/The_Beast_Within) changed to function like PvP Trinkets, removing any movement impairing effects and giving immunity to these effects for 10 seconds. The tooltip doesn't say this, but it also now breaks all forms of CC that a PvP trinket would break (including Cyclone, Sap, etc.). **Database links:** https://wowpedia.fandom.com/wiki/Bestial_Wrath Edit: found these videos so far https://www.youtube.com/watch?app=desktop&v=XmsUQZWqSQs&ab_channel=JohnDaniel time 0:22 slow from dk time 2:28 rogue cheap shot time 3:10 dk slow+warrior stun time 5:37 druid cyclone time 8:35 mage deep freeze https://www.youtube.com/watch?app=desktop&v=5qx2a0_xmFg&ab_channel=JohnDaniel time 7:47 feral slow (deep wounds) https://www.youtube.com/watch?app=desktop&v=X9jqRLTgcLU&ab_channel=RandomPeopleISH Kazzak PvP server---time 4:32 where he speaks about using Bestial Wrath as a trinket=breaks any stun/slow/snares etc also gathered the informations from one "blizzlike" server where they fixed Bestial Wrath according to these videos and few more comments.
1.0
Bestial Wrath - **Description:** if you try to use Bestial Wrath to break from paladin's repentance it says you can't use that while stunned,didn't try it with cyclone yet. Edit : doesn't work with Gouge too **How to reproduce:** cast repentance on BM hunter and use bestial wrath (The Beast Within sub-talent is needed to break the CC's) **How it should work:** Since Wotlk expansion patch changes : Bestial Wrath and Ability hunter beastwithin [[The Beast Within]](https://wowpedia.fandom.com/wiki/The_Beast_Within) changed to function like PvP Trinkets, removing any movement impairing effects and giving immunity to these effects for 10 seconds. The tooltip doesn't say this, but it also now breaks all forms of CC that a PvP trinket would break (including Cyclone, Sap, etc.). **Database links:** https://wowpedia.fandom.com/wiki/Bestial_Wrath Edit: found these videos so far https://www.youtube.com/watch?app=desktop&v=XmsUQZWqSQs&ab_channel=JohnDaniel time 0:22 slow from dk time 2:28 rogue cheap shot time 3:10 dk slow+warrior stun time 5:37 druid cyclone time 8:35 mage deep freeze https://www.youtube.com/watch?app=desktop&v=5qx2a0_xmFg&ab_channel=JohnDaniel time 7:47 feral slow (deep wounds) https://www.youtube.com/watch?app=desktop&v=X9jqRLTgcLU&ab_channel=RandomPeopleISH Kazzak PvP server---time 4:32 where he speaks about using Bestial Wrath as a trinket=breaks any stun/slow/snares etc also gathered the informations from one "blizzlike" server where they fixed Bestial Wrath according to these videos and few more comments.
priority
bestial wrath description if you try to use bestial wrath to break from paladin s repentance it says you can t use that while stunned didn t try it with cyclone yet edit doesn t work with gouge too how to reproduce cast repentance on bm hunter and use bestial wrath the beast within sub talent is needed to break the cc s how it should work since wotlk expansion patch changes bestial wrath and ability hunter beastwithin changed to function like pvp trinkets removing any movement impairing effects and giving immunity to these effects for seconds the tooltip doesn t say this but it also now breaks all forms of cc that a pvp trinket would break including cyclone sap etc database links edit found these videos so far time slow from dk time rogue cheap shot time dk slow warrior stun time druid cyclone time mage deep freeze time feral slow deep wounds kazzak pvp server time where he speaks about using bestial wrath as a trinket breaks any stun slow snares etc also gathered the informations from one blizzlike server where they fixed bestial wrath according to these videos and few more comments
1
748,716
26,134,474,976
IssuesEvent
2022-12-29 10:15:38
exyleio/exyleio
https://api.github.com/repos/exyleio/exyleio
closed
Set up SMTP server
priority 1 - high tag:development type:task area:scripts
- Requires: exyle.io domain As recommended by the [pocketbase documentation](https://pocketbase.io/docs/going-to-production/#use-smtp-mail-server), having a SMTP server prevents deliverability issues (email being sent to spam, etc.) when sending account verification email. Use [listmonk](https://github.com/knadh/listmonk) for it.
1.0
Set up SMTP server - - Requires: exyle.io domain As recommended by the [pocketbase documentation](https://pocketbase.io/docs/going-to-production/#use-smtp-mail-server), having a SMTP server prevents deliverability issues (email being sent to spam, etc.) when sending account verification email. Use [listmonk](https://github.com/knadh/listmonk) for it.
priority
set up smtp server requires exyle io domain as recommended by the having a smtp server prevents deliverability issues email being sent to spam etc when sending account verification email use for it
1
38,670
5,194,529,444
IssuesEvent
2017-01-23 04:22:17
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Skipping tests on Jenkins
area/tests priority/P3
We're seeing a bunch of skipped tests on Jenkins that I ~~don't~~ see skipped locally as well... ``` Of 3620 tests executed in 580.208s, 3613 succeeded, 0 failed, and 7 were skipped. In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestNewOpenStack" was skipped: === RUN TestNewOpenStack --- SKIP: TestNewOpenStack (0.00s) openstack_test.go:113: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestInstances" was skipped: === RUN TestInstances --- SKIP: TestInstances (0.00s) openstack_test.go:125: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestLoadBalancer" was skipped: === RUN TestLoadBalancer --- SKIP: TestLoadBalancer (0.00s) openstack_test.go:157: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestVolumes" was skipped: === RUN TestVolumes --- SKIP: TestVolumes (0.00s) openstack_test.go:205: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestNewRackspace" was skipped: === RUN TestNewRackspace --- SKIP: TestNewRackspace (0.00s) rackspace_test.go:113: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestInstances" was skipped: === RUN TestInstances --- SKIP: TestInstances (0.00s) rackspace_test.go:125: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/util/dbus", test case "TestRealDBus" was skipped: === RUN TestRealDBus --- SKIP: TestRealDBus (0.01s) dbus_test.go:73: D-Bus session bus is not available: exec: "dbus-launch": executable file not found in $PATH ``` /cc @smarterclayton @rhcarvalho iirc you knew something about this?
1.0
Skipping tests on Jenkins - We're seeing a bunch of skipped tests on Jenkins that I ~~don't~~ see skipped locally as well... ``` Of 3620 tests executed in 580.208s, 3613 succeeded, 0 failed, and 7 were skipped. In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestNewOpenStack" was skipped: === RUN TestNewOpenStack --- SKIP: TestNewOpenStack (0.00s) openstack_test.go:113: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestInstances" was skipped: === RUN TestInstances --- SKIP: TestInstances (0.00s) openstack_test.go:125: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestLoadBalancer" was skipped: === RUN TestLoadBalancer --- SKIP: TestLoadBalancer (0.00s) openstack_test.go:157: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestVolumes" was skipped: === RUN TestVolumes --- SKIP: TestVolumes (0.00s) openstack_test.go:205: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestNewRackspace" was skipped: === RUN TestNewRackspace --- SKIP: TestNewRackspace (0.00s) rackspace_test.go:113: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestInstances" was skipped: === RUN TestInstances --- SKIP: TestInstances (0.00s) rackspace_test.go:125: No config found in environment In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/util/dbus", test case "TestRealDBus" was skipped: === RUN TestRealDBus --- SKIP: TestRealDBus (0.01s) dbus_test.go:73: D-Bus session bus is not available: exec: "dbus-launch": executable file not found in $PATH ``` /cc @smarterclayton @rhcarvalho iirc you knew something about this?
non_priority
skipping tests on jenkins we re seeing a bunch of skipped tests on jenkins that i don t see skipped locally as well of tests executed in succeeded failed and were skipped in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testnewopenstack was skipped run testnewopenstack skip testnewopenstack openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testinstances was skipped run testinstances skip testinstances openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testloadbalancer was skipped run testloadbalancer skip testloadbalancer openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testvolumes was skipped run testvolumes skip testvolumes openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers rackspace test case testnewrackspace was skipped run testnewrackspace skip testnewrackspace rackspace test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers rackspace test case testinstances was skipped run testinstances skip testinstances rackspace test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg util dbus test case testrealdbus was skipped run testrealdbus skip testrealdbus dbus test go d bus session bus is not available exec dbus launch executable file not found in path cc smarterclayton rhcarvalho iirc you knew something about this
0
168,799
6,386,703,018
IssuesEvent
2017-08-03 11:53:00
ampproject/amp-by-example
https://api.github.com/repos/ampproject/amp-by-example
closed
Product Page - Add to cart is broken
bug P1: High Priority sample
Looking at https://ampbyexample.com/samples_templates/product_page/preview/ `Add to card` is supposed to redirect the user however it looks like the XHR to `https://ampbyexample.com/samples_templates/product_page/add_to_cart?__amp_source_origin=https%3A%2F%2Fampbyexample.com` does not return a JSON response and therefore we get `Failed to parse response JSON: SyntaxError: Unexpected token < in JSON at position 0​​​` error` and user is not redirected.
1.0
Product Page - Add to cart is broken - Looking at https://ampbyexample.com/samples_templates/product_page/preview/ `Add to card` is supposed to redirect the user however it looks like the XHR to `https://ampbyexample.com/samples_templates/product_page/add_to_cart?__amp_source_origin=https%3A%2F%2Fampbyexample.com` does not return a JSON response and therefore we get `Failed to parse response JSON: SyntaxError: Unexpected token < in JSON at position 0​​​` error` and user is not redirected.
priority
product page add to cart is broken looking at add to card is supposed to redirect the user however it looks like the xhr to does not return a json response and therefore we get failed to parse response json syntaxerror unexpected token in json at position ​​​ error and user is not redirected
1
191,200
22,215,673,027
IssuesEvent
2022-06-08 01:11:46
uniquelyparticular/zendesk-magento-m1-request
https://api.github.com/repos/uniquelyparticular/zendesk-magento-m1-request
closed
CVE-2019-13173 (High) detected in fstream-1.0.11.tgz - autoclosed
security vulnerability
## CVE-2019-13173 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-1.0.11.tgz</b></p></summary> <p>Advanced file system stream things</p> <p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz">https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz</a></p> <p>Path to dependency file: /zendesk-magento-m1-request/package.json</p> <p>Path to vulnerable library: /tmp/git/zendesk-magento-m1-request/node_modules/npm/node_modules/fstream/package.json</p> <p> Dependency Hierarchy: - semantic-release-15.13.16.tgz (Root Library) - npm-5.1.10.tgz - npm-6.9.0.tgz - node-gyp-3.8.0.tgz - :x: **fstream-1.0.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/zendesk-magento-m1-request/commit/281dde4039452ceffbf3f483ab9f672be345fcd3">281dde4039452ceffbf3f483ab9f672be345fcd3</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable. <p>Publish Date: 2019-07-02 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p> <p>Release Date: 2019-07-02</p> <p>Fix Resolution: 1.0.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-13173 (High) detected in fstream-1.0.11.tgz - autoclosed - ## CVE-2019-13173 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-1.0.11.tgz</b></p></summary> <p>Advanced file system stream things</p> <p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz">https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz</a></p> <p>Path to dependency file: /zendesk-magento-m1-request/package.json</p> <p>Path to vulnerable library: /tmp/git/zendesk-magento-m1-request/node_modules/npm/node_modules/fstream/package.json</p> <p> Dependency Hierarchy: - semantic-release-15.13.16.tgz (Root Library) - npm-5.1.10.tgz - npm-6.9.0.tgz - node-gyp-3.8.0.tgz - :x: **fstream-1.0.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/zendesk-magento-m1-request/commit/281dde4039452ceffbf3f483ab9f672be345fcd3">281dde4039452ceffbf3f483ab9f672be345fcd3</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable. <p>Publish Date: 2019-07-02 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p> <p>Release Date: 2019-07-02</p> <p>Fix Resolution: 1.0.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in fstream tgz autoclosed cve high severity vulnerability vulnerable library fstream tgz advanced file system stream things library home page a href path to dependency file zendesk magento request package json path to vulnerable library tmp git zendesk magento request node modules npm node modules fstream package json dependency hierarchy semantic release tgz root library npm tgz npm tgz node gyp tgz x fstream tgz vulnerable library found in head commit a href vulnerability details fstream before is vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file the fstream dirwriter function is vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
23,179
10,852,345,588
IssuesEvent
2019-11-13 12:37:21
ChenLuigi/TestingPOM
https://api.github.com/repos/ChenLuigi/TestingPOM
opened
CVE-2019-0232 (High) detected in tomcat-catalina-7.0.42.jar
security vulnerability
## CVE-2019-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/). <p>Publish Date: 2019-04-15 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232>CVE-2019-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p> <p>Release Date: 2019-04-15</p> <p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-0232 (High) detected in tomcat-catalina-7.0.42.jar - ## CVE-2019-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/). <p>Publish Date: 2019-04-15 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232>CVE-2019-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p> <p>Release Date: 2019-04-15</p> <p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in tomcat catalina jar cve high severity vulnerability vulnerable library tomcat catalina jar tomcat servlet engine core classes and standard implementations dependency hierarchy x tomcat catalina jar vulnerable library found in head commit a href vulnerability details when running on windows with enablecmdlinearguments enabled the cgi servlet in apache tomcat to to and to is vulnerable to remote code execution due to a bug in the way the jre passes command line arguments to windows the cgi servlet is disabled by default the cgi option enablecmdlinearguments is disable by default in tomcat x and will be disabled by default in all versions in response to this vulnerability for a detailed explanation of the jre behaviour see markus wulftange s blog and this archived msdn blog publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
376,084
11,138,211,998
IssuesEvent
2019-12-20 21:39:06
thibaultmeyer/sparrow
https://api.github.com/repos/thibaultmeyer/sparrow
closed
Implementing full-kotlin version of B-encode/B-decode
area/tracker kind/enhancement kind/refactor priority/low
Implementing full-kotlin version of B-encode/B-decode without any other dependencies.
1.0
Implementing full-kotlin version of B-encode/B-decode - Implementing full-kotlin version of B-encode/B-decode without any other dependencies.
priority
implementing full kotlin version of b encode b decode implementing full kotlin version of b encode b decode without any other dependencies
1
198,853
15,725,894,052
IssuesEvent
2021-03-29 10:33:38
Kong/charts
https://api.github.com/repos/Kong/charts
reopened
[kong] separate README.md into multiple guides
documentation stale
README.md is long and ever-growing. Many sections are somewhat specialized, and README.md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics. The [ingress controller documentation](https://github.com/Kong/kubernetes-ingress-controller/tree/master/docs) is a good model, where there are several types of documents (references for, well, reference, concepts for high-level explanation, guides for step-by-step instructions for specific tasks) linked from README.md.
1.0
[kong] separate README.md into multiple guides - README.md is long and ever-growing. Many sections are somewhat specialized, and README.md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics. The [ingress controller documentation](https://github.com/Kong/kubernetes-ingress-controller/tree/master/docs) is a good model, where there are several types of documents (references for, well, reference, concepts for high-level explanation, guides for step-by-step instructions for specific tasks) linked from README.md.
non_priority
separate readme md into multiple guides readme md is long and ever growing many sections are somewhat specialized and readme md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics the is a good model where there are several types of documents references for well reference concepts for high level explanation guides for step by step instructions for specific tasks linked from readme md
0
169,040
13,111,600,260
IssuesEvent
2020-08-04 23:30:27
Thy-Vipe/BeastsOfBermuda-issues
https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues
closed
[Major Bug] All-female egg glitch
Fixed! Gameplay Major Bug Potential fix public_testing
_Originally written by **The Frog Archer | 76561198201256457**_ Game Version: 1.1.985 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-9700 CPU @ 3.00GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1660 GPU Driver Info: Unknown Num CPU Cores: 8 ===================* Map: Ancestral_Plains *Expected Results:* 50/50 split between male and female babies *Actual Results:* Way to get 100% Female ory *Replication: Go into an ory egg and before hatching press B. You should always hatch as a female and get the female nesting prompt Location: X=-147341.609 Y=-8489.048 Z=5590.843 Could be a major balancing glitch with people wanting all female
1.0
[Major Bug] All-female egg glitch - _Originally written by **The Frog Archer | 76561198201256457**_ Game Version: 1.1.985 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-9700 CPU @ 3.00GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1660 GPU Driver Info: Unknown Num CPU Cores: 8 ===================* Map: Ancestral_Plains *Expected Results:* 50/50 split between male and female babies *Actual Results:* Way to get 100% Female ory *Replication: Go into an ory egg and before hatching press B. You should always hatch as a female and get the female nesting prompt Location: X=-147341.609 Y=-8489.048 Z=5590.843 Could be a major balancing glitch with people wanting all female
non_priority
all female egg glitch originally written by the frog archer game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand nvidia geforce gtx gpu driver info unknown num cpu cores map ancestral plains expected results split between male and female babies actual results way to get female ory replication go into an ory egg and before hatching press b you should always hatch as a female and get the female nesting prompt location x y z could be a major balancing glitch with people wanting all female
0
400,262
11,771,312,043
IssuesEvent
2020-03-15 23:20:27
AY1920S2-CS2103T-W17-2/main
https://api.github.com/repos/AY1920S2-CS2103T-W17-2/main
closed
Create StringCorrectionEngine class
priority.High status.Ongoing type.Enhancement
EDITED: Renamed `CommandSuggestionEngine` to `StringCorrectionEngine` As described in #97 , I'll be implementing the `StringCorrectionEngine` class. Proposed API (I'm only showing the important methods below): ```java class StringCorrectionEngine implements CorrectionEngine<String> { public StringCorrectionEngine(List<String> correctStrings) { } public CorrectionResult<String> correct(String uncorrected) { } } ``` #### How to use 1. I'd imagine that you instantiate this class during the creation of both `SuggestionEngine` and `Parser`. During this instantiation phase, you'll supply the list of correct commands as strings (notice the `correctStrings` constructor argument). 2. After that, whenever you receive user input, you'll call the `correct` method. You're guaranteed to receive a corrected result (or error if `uncorrected` is not understandable)
1.0
Create StringCorrectionEngine class - EDITED: Renamed `CommandSuggestionEngine` to `StringCorrectionEngine` As described in #97 , I'll be implementing the `StringCorrectionEngine` class. Proposed API (I'm only showing the important methods below): ```java class StringCorrectionEngine implements CorrectionEngine<String> { public StringCorrectionEngine(List<String> correctStrings) { } public CorrectionResult<String> correct(String uncorrected) { } } ``` #### How to use 1. I'd imagine that you instantiate this class during the creation of both `SuggestionEngine` and `Parser`. During this instantiation phase, you'll supply the list of correct commands as strings (notice the `correctStrings` constructor argument). 2. After that, whenever you receive user input, you'll call the `correct` method. You're guaranteed to receive a corrected result (or error if `uncorrected` is not understandable)
priority
create stringcorrectionengine class edited renamed commandsuggestionengine to stringcorrectionengine as described in i ll be implementing the stringcorrectionengine class proposed api i m only showing the important methods below java class stringcorrectionengine implements correctionengine public stringcorrectionengine list correctstrings public correctionresult correct string uncorrected how to use i d imagine that you instantiate this class during the creation of both suggestionengine and parser during this instantiation phase you ll supply the list of correct commands as strings notice the correctstrings constructor argument after that whenever you receive user input you ll call the correct method you re guaranteed to receive a corrected result or error if uncorrected is not understandable
1
26,592
11,351,177,493
IssuesEvent
2020-01-24 10:35:51
PeterNgTr/pizza-ui-tests
https://api.github.com/repos/PeterNgTr/pizza-ui-tests
closed
CVE-2018-11762 (Medium) detected in tika-core-1.18.jar
bug security vulnerability
## CVE-2018-11762 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.18.jar</b></p></summary> <p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also includes the core facades for the Tika API.</p> <p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p> <p>Path to vulnerable library: /pizza-ui-tests/node_modules/allure-commandline/dist/lib/tika-core-1.18.jar</p> <p> Dependency Hierarchy: - :x: **tika-core-1.18.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PeterNgTr/pizza-ui-tests/commit/b784718100dc8ca278a0aad1c22fba2cd661b638">b784718100dc8ca278a0aad1c22fba2cd661b638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tika 0.9 to 1.18, in a rare edge case where a user does not specify an extract directory on the commandline (--extract-dir=) and the input file has an embedded file with an absolute path, such as "C:/evil.bat", tika-app would overwrite that file. <p>Publish Date: 2018-09-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11762>CVE-2018-11762</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762</a></p> <p>Release Date: 2018-09-19</p> <p>Fix Resolution: 1.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11762 (Medium) detected in tika-core-1.18.jar - ## CVE-2018-11762 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.18.jar</b></p></summary> <p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also includes the core facades for the Tika API.</p> <p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p> <p>Path to vulnerable library: /pizza-ui-tests/node_modules/allure-commandline/dist/lib/tika-core-1.18.jar</p> <p> Dependency Hierarchy: - :x: **tika-core-1.18.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PeterNgTr/pizza-ui-tests/commit/b784718100dc8ca278a0aad1c22fba2cd661b638">b784718100dc8ca278a0aad1c22fba2cd661b638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tika 0.9 to 1.18, in a rare edge case where a user does not specify an extract directory on the commandline (--extract-dir=) and the input file has an embedded file with an absolute path, such as "C:/evil.bat", tika-app would overwrite that file. <p>Publish Date: 2018-09-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11762>CVE-2018-11762</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762</a></p> <p>Release Date: 2018-09-19</p> <p>Fix Resolution: 1.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in tika core jar cve medium severity vulnerability vulnerable library tika core jar this is the core apache tika™ toolkit library from which all other modules inherit functionality it also includes the core facades for the tika api library home page a href path to vulnerable library pizza ui tests node modules allure commandline dist lib tika core jar dependency hierarchy x tika core jar vulnerable library found in head commit a href vulnerability details in apache tika to in a rare edge case where a user does not specify an extract directory on the commandline extract dir and the input file has an embedded file with an absolute path such as c evil bat tika app would overwrite that file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
327,245
24,124,696,292
IssuesEvent
2022-09-20 22:26:34
cloudflare/cloudflare-docs
https://api.github.com/repos/cloudflare/cloudflare-docs
opened
[R2] Workers Runtime APIs layout & formatting
documentation content:edit
### Which Cloudflare product does this pertain to? R2 ### Existing documentation URL(s) - https://developers.cloudflare.com/r2/data-access/workers-api/workers-api-reference/ ### Section that requires update Layout and formatting. ### What needs to change? The current layout, whilst satisfactory for a type reference, will struggle to scale as the APIs gain more options and require more explanations/walkthroughs on how to properly utilise them. As an example, the addition of the new checksums supported in `R2PutOptions` means that either: 1) Every checksum is listed on it's own, with an aside that only one can be used at a time. 2) The type for the 'checksum' in `R2PutOptions` becomes `md5 | sha1 | sha256 | sha384 | sha512` Either way, it becomes very verbose and will easily confuse people. ### How should it change? Taking inspiration from well-regarded documentation layouts such as `docs.rs` (i.e https://doc.rust-lang.org/stable/std/collections/struct.HashMap.html), we should come up with a consistent structure on documenting Worker's Runtime APIs that we can then extend to refactoring the pages for KV, Durable Objects and the like. We want a good middle-ground between providing a reference, and providing a how-to, without going too in-depth that the material would be better split off into it's own how-to page. As an example of one potential layout that was discussed by myself and @Frederik-Baetens: <img width="643" alt="image" src="https://user-images.githubusercontent.com/94662631/191375548-9e5244fe-85be-4ca1-929a-4af793699941.png"> Using the [<details>](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/details) element, this allows us to: 1) Provide a lot of value without cluttering the rest of the page, as it's expandable. 2) Preserve the 'searchability' of the content, as CTRL+F on the majority of browsers will 'look inside' a `<details>` element. 3) Additionally include small examples, not in-depth or niche enough for their own page, under additional `<details>` elements. <img width="654" alt="image" src="https://user-images.githubusercontent.com/94662631/191375800-c884bf3c-dade-43a3-8aed-f3e48f1d6b21.png"> ### Additional information Original discussion: https://github.com/cloudflare/cloudflare-docs/pull/5897
1.0
[R2] Workers Runtime APIs layout & formatting - ### Which Cloudflare product does this pertain to? R2 ### Existing documentation URL(s) - https://developers.cloudflare.com/r2/data-access/workers-api/workers-api-reference/ ### Section that requires update Layout and formatting. ### What needs to change? The current layout, whilst satisfactory for a type reference, will struggle to scale as the APIs gain more options and require more explanations/walkthroughs on how to properly utilise them. As an example, the addition of the new checksums supported in `R2PutOptions` means that either: 1) Every checksum is listed on it's own, with an aside that only one can be used at a time. 2) The type for the 'checksum' in `R2PutOptions` becomes `md5 | sha1 | sha256 | sha384 | sha512` Either way, it becomes very verbose and will easily confuse people. ### How should it change? Taking inspiration from well-regarded documentation layouts such as `docs.rs` (i.e https://doc.rust-lang.org/stable/std/collections/struct.HashMap.html), we should come up with a consistent structure on documenting Worker's Runtime APIs that we can then extend to refactoring the pages for KV, Durable Objects and the like. We want a good middle-ground between providing a reference, and providing a how-to, without going too in-depth that the material would be better split off into it's own how-to page. As an example of one potential layout that was discussed by myself and @Frederik-Baetens: <img width="643" alt="image" src="https://user-images.githubusercontent.com/94662631/191375548-9e5244fe-85be-4ca1-929a-4af793699941.png"> Using the [<details>](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/details) element, this allows us to: 1) Provide a lot of value without cluttering the rest of the page, as it's expandable. 2) Preserve the 'searchability' of the content, as CTRL+F on the majority of browsers will 'look inside' a `<details>` element. 3) Additionally include small examples, not in-depth or niche enough for their own page, under additional `<details>` elements. <img width="654" alt="image" src="https://user-images.githubusercontent.com/94662631/191375800-c884bf3c-dade-43a3-8aed-f3e48f1d6b21.png"> ### Additional information Original discussion: https://github.com/cloudflare/cloudflare-docs/pull/5897
non_priority
workers runtime apis layout formatting which cloudflare product does this pertain to existing documentation url s section that requires update layout and formatting what needs to change the current layout whilst satisfactory for a type reference will struggle to scale as the apis gain more options and require more explanations walkthroughs on how to properly utilise them as an example the addition of the new checksums supported in means that either every checksum is listed on it s own with an aside that only one can be used at a time the type for the checksum in becomes either way it becomes very verbose and will easily confuse people how should it change taking inspiration from well regarded documentation layouts such as docs rs i e we should come up with a consistent structure on documenting worker s runtime apis that we can then extend to refactoring the pages for kv durable objects and the like we want a good middle ground between providing a reference and providing a how to without going too in depth that the material would be better split off into it s own how to page as an example of one potential layout that was discussed by myself and frederik baetens img width alt image src using the element this allows us to provide a lot of value without cluttering the rest of the page as it s expandable preserve the searchability of the content as ctrl f on the majority of browsers will look inside a element additionally include small examples not in depth or niche enough for their own page under additional elements img width alt image src additional information original discussion
0
649,873
21,328,831,080
IssuesEvent
2022-04-18 04:57:14
googleapis/java-spanner
https://api.github.com/repos/googleapis/java-spanner
closed
The build failed
type: bug priority: p1 api: spanner flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7465d23f-64d4-4e6e-b1f1-43af7e97fca6), [Sponge](http://sponge2/7465d23f-64d4-4e6e-b1f1-43af7e97fca6) status: failed <details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100) at com.google.cloud.spanner.it.ITCommitTimestampTest.setUp(ITCommitTimestampTest.java:69) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137) ... 35 more Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350) at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222) at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.addListener(AbstractFuture.java:761) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.addListener(FluentFuture.java:114) at com.google.common.util.concurrent.ForwardingListenableFuture.addListener(ForwardingListenableFuture.java:47) at com.google.api.core.ApiFutureToListenableFuture.addListener(ApiFutureToListenableFuture.java:51) at com.google.common.util.concurrent.AbstractCatchingFuture.create(AbstractCatchingFuture.java:46) at com.google.common.util.concurrent.Futures.catching(Futures.java:306) at com.google.api.core.ApiFutures.catching(ApiFutures.java:98) at com.google.api.gax.longrunning.OperationFutureImpl.<init>(OperationFutureImpl.java:95) at com.google.cloud.spanner.DatabaseAdminClientImpl.createDatabase(DatabaseAdminClientImpl.java:348) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:136) ... 35 more Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at io.grpc.Status.asRuntimeException(Status.java:535) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) </pre></details>
1.0
The build failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7465d23f-64d4-4e6e-b1f1-43af7e97fca6), [Sponge](http://sponge2/7465d23f-64d4-4e6e-b1f1-43af7e97fca6) status: failed <details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100) at com.google.cloud.spanner.it.ITCommitTimestampTest.setUp(ITCommitTimestampTest.java:69) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137) ... 35 more Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350) at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222) at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.addListener(AbstractFuture.java:761) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.addListener(FluentFuture.java:114) at com.google.common.util.concurrent.ForwardingListenableFuture.addListener(ForwardingListenableFuture.java:47) at com.google.api.core.ApiFutureToListenableFuture.addListener(ApiFutureToListenableFuture.java:51) at com.google.common.util.concurrent.AbstractCatchingFuture.create(AbstractCatchingFuture.java:46) at com.google.common.util.concurrent.Futures.catching(Futures.java:306) at com.google.api.core.ApiFutures.catching(ApiFutures.java:98) at com.google.api.gax.longrunning.OperationFutureImpl.<init>(OperationFutureImpl.java:95) at com.google.cloud.spanner.DatabaseAdminClientImpl.createDatabase(DatabaseAdminClientImpl.java:348) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:136) ... 35 more Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at io.grpc.Status.asRuntimeException(Status.java:535) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) </pre></details>
priority
the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output com google cloud spanner spannerexception resource exhausted com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner it itcommittimestamptest setup itcommittimestamptest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java util concurrent executionexception com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java more caused by com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory fromapiexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner databaseadminclientimpl lambda createdatabase databaseadminclientimpl java at com google api core apifutures apifunctiontoguavafunction apply apifutures java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture run abstractcatchingfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture addlistener abstractfuture java at com google common util concurrent fluentfuture trustedfuture addlistener fluentfuture java at com google common util concurrent forwardinglistenablefuture addlistener forwardinglistenablefuture java at com google api core apifuturetolistenablefuture addlistener apifuturetolistenablefuture java at com google common util concurrent abstractcatchingfuture create abstractcatchingfuture java at com google common util concurrent futures catching futures java at com google api core apifutures catching apifutures java at com google api gax longrunning operationfutureimpl operationfutureimpl java at com google cloud spanner databaseadminclientimpl createdatabase databaseadminclientimpl java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java more caused by io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at io grpc status asruntimeexception status java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google cloud spanner spi spannererrorinterceptor onclose spannererrorinterceptor java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java base java util concurrent executors runnableadapter call executors java at java base java util concurrent futuretask run futuretask java at java base java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java
1
400,604
11,777,613,859
IssuesEvent
2020-03-16 15:04:35
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
admin.microsoft.com - site is not usable
browser-firefox engine-gecko priority-critical
<!-- @browser: Firefox 74.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 --> <!-- @reported_with: addon-reporter-firefox --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50285 --> **URL**: https://admin.microsoft.com/AdminPortal/Home#/homepage **Browser / Version**: Firefox 74.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Site refuses to load **Steps to Reproduce**: I'm attempting to use the 365 administrator portal but recently it has stopped working in Firefox. I've disabled all blocking or privacy extensions and all I'm getting is a white screen. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/87407f95-677c-40ea-b9dd-5ab0c0d34505.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
admin.microsoft.com - site is not usable - <!-- @browser: Firefox 74.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 --> <!-- @reported_with: addon-reporter-firefox --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50285 --> **URL**: https://admin.microsoft.com/AdminPortal/Home#/homepage **Browser / Version**: Firefox 74.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Site refuses to load **Steps to Reproduce**: I'm attempting to use the 365 administrator portal but recently it has stopped working in Firefox. I've disabled all blocking or privacy extensions and all I'm getting is a white screen. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/87407f95-677c-40ea-b9dd-5ab0c0d34505.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
admin microsoft com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description site refuses to load steps to reproduce i m attempting to use the administrator portal but recently it has stopped working in firefox i ve disabled all blocking or privacy extensions and all i m getting is a white screen view the screenshot img alt screenshot src browser configuration none from with ❤️
1
256,342
8,127,347,823
IssuesEvent
2018-08-17 07:43:52
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Fix Manta stuff in Molecule plot's CMakeLists.txt
Bug Likelihood: 3 - Occasional Priority: Normal Severity: 2 - Minor Irritation
The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build. I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1418 Status: Resolved Project: VisIt Tracker: Bug Priority: Normal Subject: Fix Manta stuff in Molecule plot's CMakeLists.txt Assigned to: Kathleen Biagas Category: Target version: 2.7 Author: Brad Whitlock Start: 04/10/2013 Due date: % Done: 0 Estimated time: Created: 04/10/2013 05:23 pm Updated: 10/09/2013 05:08 pm Likelihood: 3 - Occasional Severity: 2 - Minor Irritation Found in version: 2.6.2 Impact: Expected Use: OS: All Support Group: Any Description: The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build. I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file. Comments: Added ability for plugins to use codefiles for conditional logic that is to be added to cmake files.For Molecule plot, the logic is:SVN Revisions 22043 (2.7RC) 22045 (trunk)
1.0
Fix Manta stuff in Molecule plot's CMakeLists.txt - The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build. I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1418 Status: Resolved Project: VisIt Tracker: Bug Priority: Normal Subject: Fix Manta stuff in Molecule plot's CMakeLists.txt Assigned to: Kathleen Biagas Category: Target version: 2.7 Author: Brad Whitlock Start: 04/10/2013 Due date: % Done: 0 Estimated time: Created: 04/10/2013 05:23 pm Updated: 10/09/2013 05:08 pm Likelihood: 3 - Occasional Severity: 2 - Minor Irritation Found in version: 2.6.2 Impact: Expected Use: OS: All Support Group: Any Description: The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build. I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file. Comments: Added ability for plugins to use codefiles for conditional logic that is to be added to cmake files.For Molecule plot, the logic is:SVN Revisions 22043 (2.7RC) 22045 (trunk)
priority
fix manta stuff in molecule plot s cmakelists txt the molecule plot s cmakelists txt file cannot be automatically regenerated because of some manta related coding if visit is being built with manta then some additional source files are added to the plugin s build i suggest that the manta sources always get added so the cmakelists txt file can be autogenerated then some conditional compilation in the manta source files can determine whether the file actually produces anything in its object file redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject fix manta stuff in molecule plot s cmakelists txt assigned to kathleen biagas category target version author brad whitlock start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description the molecule plot s cmakelists txt file cannot be automatically regenerated because of some manta related coding if visit is being built with manta then some additional source files are added to the plugin s build i suggest that the manta sources always get added so the cmakelists txt file can be autogenerated then some conditional compilation in the manta source files can determine whether the file actually produces anything in its object file comments added ability for plugins to use codefiles for conditional logic that is to be added to cmake files for molecule plot the logic is svn revisions trunk
1
346,304
30,883,363,599
IssuesEvent
2023-08-03 19:28:36
ray-project/ray
https://api.github.com/repos/ray-project/ray
opened
Release test rllib_stress_tests.aws failed
bug P0 rllib release-test
Release test **rllib_stress_tests.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/2019#0189bcd7-94cc-4087-bb27-d79bdde9aa23 for more details. Managed by OSS Test Policy
1.0
Release test rllib_stress_tests.aws failed - Release test **rllib_stress_tests.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/2019#0189bcd7-94cc-4087-bb27-d79bdde9aa23 for more details. Managed by OSS Test Policy
non_priority
release test rllib stress tests aws failed release test rllib stress tests aws failed see for more details managed by oss test policy
0
67,834
17,085,376,328
IssuesEvent
2021-07-08 11:06:55
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
opened
[Bug] Multiple bindings don't work inside the child of the list
Bug List Widget UI Building UI Building Pod Widgets
Multiple bindings inside the list widget don't work.
2.0
[Bug] Multiple bindings don't work inside the child of the list - Multiple bindings inside the list widget don't work.
non_priority
multiple bindings don t work inside the child of the list multiple bindings inside the list widget don t work
0
60,046
17,023,320,551
IssuesEvent
2021-07-03 01:25:01
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
mapnik ignores layer-tags of rivers
Component: mapnik Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 2.03pm, Wednesday, 12th November 2008]** Mapnik seems to ignore the "layer=*"-tags of rivers. [http://www.openstreetmap.org/?lat=48.051367&lon=8.208524&zoom=18&layers=B000FTF Here] is an example of this behavior. The river is tagged with layer=-1, the building is tagged with layer=1 (and i just added an area highway=pedestrian on level 0 and put a section of the river into a tunnel, but that was of course too late for the weekly dump, so you won't see it in mapnik now).
1.0
mapnik ignores layer-tags of rivers - **[Submitted to the original trac issue database at 2.03pm, Wednesday, 12th November 2008]** Mapnik seems to ignore the "layer=*"-tags of rivers. [http://www.openstreetmap.org/?lat=48.051367&lon=8.208524&zoom=18&layers=B000FTF Here] is an example of this behavior. The river is tagged with layer=-1, the building is tagged with layer=1 (and i just added an area highway=pedestrian on level 0 and put a section of the river into a tunnel, but that was of course too late for the weekly dump, so you won't see it in mapnik now).
non_priority
mapnik ignores layer tags of rivers mapnik seems to ignore the layer tags of rivers is an example of this behavior the river is tagged with layer the building is tagged with layer and i just added an area highway pedestrian on level and put a section of the river into a tunnel but that was of course too late for the weekly dump so you won t see it in mapnik now
0
65,608
8,823,269,928
IssuesEvent
2019-01-02 12:58:33
sinnerschrader/feature-hub
https://api.github.com/repos/sinnerschrader/feature-hub
closed
Website: Document integrator consumer ID usage
documentation
- [x] Will be closed by PR #208 How to deal with multiple integrators on one host? (e.g. LocalStorage)
1.0
Website: Document integrator consumer ID usage - - [x] Will be closed by PR #208 How to deal with multiple integrators on one host? (e.g. LocalStorage)
non_priority
website document integrator consumer id usage will be closed by pr how to deal with multiple integrators on one host e g localstorage
0
366,428
10,820,664,131
IssuesEvent
2019-11-08 16:51:23
AY1920S1-CS2113T-W17-4/main
https://api.github.com/repos/AY1920S1-CS2113T-W17-4/main
closed
As a Computing student, I can add my estimated time taken to complete a task
priority.Low type.Story
so that I know how much free time I would have.
1.0
As a Computing student, I can add my estimated time taken to complete a task - so that I know how much free time I would have.
priority
as a computing student i can add my estimated time taken to complete a task so that i know how much free time i would have
1
807,507
30,006,071,349
IssuesEvent
2023-06-26 12:32:37
GoogleCloudPlatform/cloud-sql-python-connector
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-python-connector
closed
system.test_pytds_connection: test_pooled_connection_with_pytds failed
type: bug priority: p2 flakybot: issue flakybot: flaky
Note: #672 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: fe64fb37d98efa76d8027077d3a7e52bf1e7a15d buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/5367360775 status: failed <details><summary>Test output</summary><br><pre>@pytest.fixture(name="pool") def setup() -> Generator: pool = init_connection_engine() > with pool.connect() as conn: tests/system/test_pytds_connection.py:65: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3264: in connect return self._connection_cls(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:145: in __init__ self._dbapi_connection = engine.raw_connection() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3288: in raw_connection return self.pool.connect() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:452: in connect return _ConnectionFairy._checkout(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:1267: in _checkout fairy = _ConnectionRecord.checkout(pool) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:716: in checkout rec = pool._do_get() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:169: in _do_get with util.safe_reraise(): .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__ raise exc_value.with_traceback(exc_tb) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:167: in _do_get return self._create_connection() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:393: in _create_connection return _ConnectionRecord(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:678: in __init__ self.__connect() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:902: in __connect with util.safe_reraise(): .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__ raise exc_value.with_traceback(exc_tb) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:898: in __connect self.dbapi_connection = connection = pool._invoke_creator(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:365: in <lambda> return lambda rec: creator_fn() tests/system/test_pytds_connection.py:40: in getconn conn = connector.connect( google/cloud/sql/connector/connector.py:159: in connect return connect_task.result() /Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:456: in result return self.__get_result() /Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:401: in __get_result raise self._exception google/cloud/sql/connector/connector.py:204: in connect_async instance = Instance( google/cloud/sql/connector/instance.py:253: in __init__ self._credentials = _auth_init(credentials) google/cloud/sql/connector/utils.py:124: in _auth_init credentials, _ = default(scopes=scopes) .nox/system-3-11/lib/python3.11/site-packages/google/auth/_default.py:675: in default project_id = credentials.get_project_id(request=request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:334: in get_project_id self.before_request(request, "GET", url, headers) .nox/system-3-11/lib/python3.11/site-packages/google/auth/credentials.py:151: in before_request self.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:355: in refresh self._impersonated_credentials.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:247: in refresh self._update_token(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:260: in _update_token self._source_credentials.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:368: in refresh subject_token=self.retrieve_subject_token(request), .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:158: in retrieve_subject_token self._get_token_data(request), .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:167: in _get_token_data return self._get_url_data( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.auth.identity_pool.Credentials object at 0x10d3ff3d0> request = <google.auth.transport.requests.Request object at 0x10d3fde50> url = 'https://pipelines.actions.githubusercontent.com/umAmnh0OhcfbtGEt7J16Yga6HsgM8dYIhPxbPiOYFLVwMnfbKz/00000000-0000-0000...cations%2Fglobal%2FworkloadIdentityPools%2Fgh-13a715-cloud-sql-pyt-dd1c5f%2Fproviders%2Fgh-13a715-cloud-sql-pyt-dd1c5f' headers = {'Authorization': 'Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6ImVCWl9jbjNzWFlBZDBjaDRUSEJLSElnT3dPRSJ9.eyJuYW1...rQ4yUX2hGS2UBW_MgcvJ0VQpGlt1nFMDC8DzhKGNCN6fDKQB8yJA2bq-ddCqnrOfrB_fFwkBEBhlh8BWKlZX8A8X2bNT8M9MOh0o3Xu4Okb895WQV9zRA'} def _get_url_data(self, request, url, headers): response = request(url=url, method="GET", headers=headers) # support both string and bytes type response.data response_body = ( response.data.decode("utf-8") if hasattr(response.data, "decode") else response.data ) if response.status != 200: > raise exceptions.RefreshError( "Unable to retrieve Identity Pool subject token", response_body ) E google.auth.exceptions.RefreshError: ('Unable to retrieve Identity Pool subject token', '{ "message": "GitHub Actions is temporarily unavailable. Please visit https://www.githubstatus.com/ for the status of our services.", "ref": "Ref A: F3B1C190713A4756A6359D9977F85EDD Ref B: CO1EDGE2015 Ref C: 2023-06-25T02:34:36Z" }') .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:189: RefreshError</pre></details>
1.0
system.test_pytds_connection: test_pooled_connection_with_pytds failed - Note: #672 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: fe64fb37d98efa76d8027077d3a7e52bf1e7a15d buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/5367360775 status: failed <details><summary>Test output</summary><br><pre>@pytest.fixture(name="pool") def setup() -> Generator: pool = init_connection_engine() > with pool.connect() as conn: tests/system/test_pytds_connection.py:65: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3264: in connect return self._connection_cls(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:145: in __init__ self._dbapi_connection = engine.raw_connection() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3288: in raw_connection return self.pool.connect() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:452: in connect return _ConnectionFairy._checkout(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:1267: in _checkout fairy = _ConnectionRecord.checkout(pool) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:716: in checkout rec = pool._do_get() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:169: in _do_get with util.safe_reraise(): .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__ raise exc_value.with_traceback(exc_tb) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:167: in _do_get return self._create_connection() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:393: in _create_connection return _ConnectionRecord(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:678: in __init__ self.__connect() .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:902: in __connect with util.safe_reraise(): .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__ raise exc_value.with_traceback(exc_tb) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:898: in __connect self.dbapi_connection = connection = pool._invoke_creator(self) .nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:365: in <lambda> return lambda rec: creator_fn() tests/system/test_pytds_connection.py:40: in getconn conn = connector.connect( google/cloud/sql/connector/connector.py:159: in connect return connect_task.result() /Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:456: in result return self.__get_result() /Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:401: in __get_result raise self._exception google/cloud/sql/connector/connector.py:204: in connect_async instance = Instance( google/cloud/sql/connector/instance.py:253: in __init__ self._credentials = _auth_init(credentials) google/cloud/sql/connector/utils.py:124: in _auth_init credentials, _ = default(scopes=scopes) .nox/system-3-11/lib/python3.11/site-packages/google/auth/_default.py:675: in default project_id = credentials.get_project_id(request=request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:334: in get_project_id self.before_request(request, "GET", url, headers) .nox/system-3-11/lib/python3.11/site-packages/google/auth/credentials.py:151: in before_request self.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:355: in refresh self._impersonated_credentials.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:247: in refresh self._update_token(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:260: in _update_token self._source_credentials.refresh(request) .nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:368: in refresh subject_token=self.retrieve_subject_token(request), .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:158: in retrieve_subject_token self._get_token_data(request), .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:167: in _get_token_data return self._get_url_data( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.auth.identity_pool.Credentials object at 0x10d3ff3d0> request = <google.auth.transport.requests.Request object at 0x10d3fde50> url = 'https://pipelines.actions.githubusercontent.com/umAmnh0OhcfbtGEt7J16Yga6HsgM8dYIhPxbPiOYFLVwMnfbKz/00000000-0000-0000...cations%2Fglobal%2FworkloadIdentityPools%2Fgh-13a715-cloud-sql-pyt-dd1c5f%2Fproviders%2Fgh-13a715-cloud-sql-pyt-dd1c5f' headers = {'Authorization': 'Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6ImVCWl9jbjNzWFlBZDBjaDRUSEJLSElnT3dPRSJ9.eyJuYW1...rQ4yUX2hGS2UBW_MgcvJ0VQpGlt1nFMDC8DzhKGNCN6fDKQB8yJA2bq-ddCqnrOfrB_fFwkBEBhlh8BWKlZX8A8X2bNT8M9MOh0o3Xu4Okb895WQV9zRA'} def _get_url_data(self, request, url, headers): response = request(url=url, method="GET", headers=headers) # support both string and bytes type response.data response_body = ( response.data.decode("utf-8") if hasattr(response.data, "decode") else response.data ) if response.status != 200: > raise exceptions.RefreshError( "Unable to retrieve Identity Pool subject token", response_body ) E google.auth.exceptions.RefreshError: ('Unable to retrieve Identity Pool subject token', '{ "message": "GitHub Actions is temporarily unavailable. Please visit https://www.githubstatus.com/ for the status of our services.", "ref": "Ref A: F3B1C190713A4756A6359D9977F85EDD Ref B: CO1EDGE2015 Ref C: 2023-06-25T02:34:36Z" }') .nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:189: RefreshError</pre></details>
priority
system test pytds connection test pooled connection with pytds failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output pytest fixture name pool def setup generator pool init connection engine with pool connect as conn tests system test pytds connection py nox system lib site packages sqlalchemy engine base py in connect return self connection cls self nox system lib site packages sqlalchemy engine base py in init self dbapi connection engine raw connection nox system lib site packages sqlalchemy engine base py in raw connection return self pool connect nox system lib site packages sqlalchemy pool base py in connect return connectionfairy checkout self nox system lib site packages sqlalchemy pool base py in checkout fairy connectionrecord checkout pool nox system lib site packages sqlalchemy pool base py in checkout rec pool do get nox system lib site packages sqlalchemy pool impl py in do get with util safe reraise nox system lib site packages sqlalchemy util langhelpers py in exit raise exc value with traceback exc tb nox system lib site packages sqlalchemy pool impl py in do get return self create connection nox system lib site packages sqlalchemy pool base py in create connection return connectionrecord self nox system lib site packages sqlalchemy pool base py in init self connect nox system lib site packages sqlalchemy pool base py in connect with util safe reraise nox system lib site packages sqlalchemy util langhelpers py in exit raise exc value with traceback exc tb nox system lib site packages sqlalchemy pool base py in connect self dbapi connection connection pool invoke creator self nox system lib site packages sqlalchemy pool base py in return lambda rec creator fn tests system test pytds connection py in getconn conn connector connect google cloud sql connector connector py in connect return connect task result library frameworks python framework versions lib concurrent futures base py in result return self get result library frameworks python framework versions lib concurrent futures base py in get result raise self exception google cloud sql connector connector py in connect async instance instance google cloud sql connector instance py in init self credentials auth init credentials google cloud sql connector utils py in auth init credentials default scopes scopes nox system lib site packages google auth default py in default project id credentials get project id request request nox system lib site packages google auth external account py in get project id self before request request get url headers nox system lib site packages google auth credentials py in before request self refresh request nox system lib site packages google auth external account py in refresh self impersonated credentials refresh request nox system lib site packages google auth impersonated credentials py in refresh self update token request nox system lib site packages google auth impersonated credentials py in update token self source credentials refresh request nox system lib site packages google auth external account py in refresh subject token self retrieve subject token request nox system lib site packages google auth identity pool py in retrieve subject token self get token data request nox system lib site packages google auth identity pool py in get token data return self get url data self request url headers authorization bearer ddcqnrofrb def get url data self request url headers response request url url method get headers headers support both string and bytes type response data response body response data decode utf if hasattr response data decode else response data if response status raise exceptions refresherror unable to retrieve identity pool subject token response body e google auth exceptions refresherror unable to retrieve identity pool subject token message github actions is temporarily unavailable please visit for the status of our services ref ref a ref b ref c nox system lib site packages google auth identity pool py refresherror
1
561,035
16,609,840,751
IssuesEvent
2021-06-02 10:05:36
svthalia/Reaxit
https://api.github.com/repos/svthalia/Reaxit
opened
Pretty splash screens
feature priority: high
We currently have a simple image as a splashcreen with no transition to the first frame. We should at least make the image (esp. on android) nicer, and it would also be nice to fade to the first frame.
1.0
Pretty splash screens - We currently have a simple image as a splashcreen with no transition to the first frame. We should at least make the image (esp. on android) nicer, and it would also be nice to fade to the first frame.
priority
pretty splash screens we currently have a simple image as a splashcreen with no transition to the first frame we should at least make the image esp on android nicer and it would also be nice to fade to the first frame
1
58,025
3,087,082,880
IssuesEvent
2015-08-25 09:15:34
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
Индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении
bug imported Priority-Medium
_From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on October 01, 2013 02:05:31_ 1. Подключиться к новому хабу на который вас не пускают по нику или тэгу, по размеру шары, либо по количеству слотов. 2. добавить хаб в избранные. 3. изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили, закрыть окно настроек. 4. дождаться автоматического переподключения к хабу. 5. хаб не пускает, т.к. ваши изменения не актуализировались. 6. нажать кнопку переподключения к хабу. 7. новые настройки вступают в силу и подключение происходит. FlylinkDC++ r502 -rc2-x64 build 15562 Compiled on: 2013-09-28 _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1322_
1.0
Индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении - _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on October 01, 2013 02:05:31_ 1. Подключиться к новому хабу на который вас не пускают по нику или тэгу, по размеру шары, либо по количеству слотов. 2. добавить хаб в избранные. 3. изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили, закрыть окно настроек. 4. дождаться автоматического переподключения к хабу. 5. хаб не пускает, т.к. ваши изменения не актуализировались. 6. нажать кнопку переподключения к хабу. 7. новые настройки вступают в силу и подключение происходит. FlylinkDC++ r502 -rc2-x64 build 15562 Compiled on: 2013-09-28 _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1322_
priority
индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении from on october подключиться к новому хабу на который вас не пускают по нику или тэгу по размеру шары либо по количеству слотов добавить хаб в избранные изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили закрыть окно настроек дождаться автоматического переподключения к хабу хаб не пускает т к ваши изменения не актуализировались нажать кнопку переподключения к хабу новые настройки вступают в силу и подключение происходит flylinkdc build compiled on original issue
1
245,822
18,795,361,614
IssuesEvent
2021-11-08 21:35:54
4kills/go-zlib
https://api.github.com/repos/4kills/go-zlib
closed
There is no `ReadBytes` method
documentation
README.md shows an alternative method for decompressing data using `ReadBytes()` method. Was this method removed?
1.0
There is no `ReadBytes` method - README.md shows an alternative method for decompressing data using `ReadBytes()` method. Was this method removed?
non_priority
there is no readbytes method readme md shows an alternative method for decompressing data using readbytes method was this method removed
0
552,105
16,195,203,499
IssuesEvent
2021-05-04 13:49:41
Edgeryders-Participio/multi-dreams
https://api.github.com/repos/Edgeryders-Participio/multi-dreams
closed
Require the user to approve the ToS and privacy policy on sign up
Priority: 1 (now - within 1 month)
We need to implement approving our TOS. https://www.iubenda.com/terms-and-conditions/58637640
1.0
Require the user to approve the ToS and privacy policy on sign up - We need to implement approving our TOS. https://www.iubenda.com/terms-and-conditions/58637640
priority
require the user to approve the tos and privacy policy on sign up we need to implement approving our tos
1
346,470
10,412,685,756
IssuesEvent
2019-09-13 16:33:49
Sp2000/colplus-repo
https://api.github.com/repos/Sp2000/colplus-repo
opened
Fix split-subgenera data bugs
bug high priority
I found another case where a species is placed on the classification in the wrong location by CoL+. This is similar to the [inc. sed. bug](https://github.com/Sp2000/colplus-backend/issues/477), but in this case it's caused by [split-subgenera](https://www.col.plus/dataset/1106/duplicates?category=uninomial&limit=50&minSize=2&mode=STRICT&offset=0&rank=subgenus&withDecision=false). Here's the output from the diff_key generator: diff_key | hierarchy | p8 | p7 | p6 | p5 | p4 | p3 | p2 | p1 | name | hash | url -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- 685deb96725d46c6ff6cd24d9c7158ab | Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | | Animalia | Echinodermata | Echinoidea | Spatangoida | Brissidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | 0a01d690cc0ad948cebb7267ef98854b | http://workbench.catalogueoflife.org/col_plus/details/species/id/0a01d690cc0ad948cebb7267ef98854b ecf1add23e3f3abae62addddd00674a1 | Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | Animalia | Echinodermata | Echinoidea | Spatangoida | Spatangoidea | Eupatagidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | d81d04e18a36e88d32021bfcc5025332 | http://www.catalogueoflife.org/annual-checklist/2019/details/species/id/d81d04e18a36e88d32021bfcc5025332 Here's the raw data from the ACEF 106.tar.gz file: ``` "755133" "Animalia" "Echinodermata" "Echinoidea" "Spatangoida" "Spatangoidea" "Eupatagidae" "Brissoides" "Rhabdobrissus" "aloysii" "Pijpers, 1933" "" "accepted name" "1" "1" "0" "marine" "" "Kroh, Andreas" "2014-03-05" "http://www.marinespecies.org/echinoidea/aphia.php?p=taxdetails&id=755133" "" "urn:lsid:marinespecies.org:taxname:755133" ``` [CoL+ places it incorrectly in Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae](https://www.col.plus/dataset/1106/taxon/755133) instead of Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae. @mdoering, this is a data quality issue that we'd have to report to the GSD and block one of the 2 from appearing in CoL+, but also might commonly happen so placing split subgenera in the classification correctly might still be important.
1.0
Fix split-subgenera data bugs - I found another case where a species is placed on the classification in the wrong location by CoL+. This is similar to the [inc. sed. bug](https://github.com/Sp2000/colplus-backend/issues/477), but in this case it's caused by [split-subgenera](https://www.col.plus/dataset/1106/duplicates?category=uninomial&limit=50&minSize=2&mode=STRICT&offset=0&rank=subgenus&withDecision=false). Here's the output from the diff_key generator: diff_key | hierarchy | p8 | p7 | p6 | p5 | p4 | p3 | p2 | p1 | name | hash | url -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- 685deb96725d46c6ff6cd24d9c7158ab | Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | | Animalia | Echinodermata | Echinoidea | Spatangoida | Brissidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | 0a01d690cc0ad948cebb7267ef98854b | http://workbench.catalogueoflife.org/col_plus/details/species/id/0a01d690cc0ad948cebb7267ef98854b ecf1add23e3f3abae62addddd00674a1 | Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | Animalia | Echinodermata | Echinoidea | Spatangoida | Spatangoidea | Eupatagidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | d81d04e18a36e88d32021bfcc5025332 | http://www.catalogueoflife.org/annual-checklist/2019/details/species/id/d81d04e18a36e88d32021bfcc5025332 Here's the raw data from the ACEF 106.tar.gz file: ``` "755133" "Animalia" "Echinodermata" "Echinoidea" "Spatangoida" "Spatangoidea" "Eupatagidae" "Brissoides" "Rhabdobrissus" "aloysii" "Pijpers, 1933" "" "accepted name" "1" "1" "0" "marine" "" "Kroh, Andreas" "2014-03-05" "http://www.marinespecies.org/echinoidea/aphia.php?p=taxdetails&id=755133" "" "urn:lsid:marinespecies.org:taxname:755133" ``` [CoL+ places it incorrectly in Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae](https://www.col.plus/dataset/1106/taxon/755133) instead of Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae. @mdoering, this is a data quality issue that we'd have to report to the GSD and block one of the 2 from appearing in CoL+, but also might commonly happen so placing split subgenera in the classification correctly might still be important.
priority
fix split subgenera data bugs i found another case where a species is placed on the classification in the wrong location by col this is similar to the but in this case it s caused by here s the output from the diff key generator diff key hierarchy name hash url animalia echinodermata echinoidea spatangoida brissidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida brissidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides brissoides rhabdobrissus aloysii here s the raw data from the acef tar gz file animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides rhabdobrissus aloysii pijpers accepted name marine kroh andreas urn lsid marinespecies org taxname instead of animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae mdoering this is a data quality issue that we d have to report to the gsd and block one of the from appearing in col but also might commonly happen so placing split subgenera in the classification correctly might still be important
1
103,369
11,355,091,898
IssuesEvent
2020-01-24 19:12:03
wisslab/judaicalink-labs
https://api.github.com/repos/wisslab/judaicalink-labs
opened
Documentation for backend app
documentation
The whole backend app with the messaging and task API, as well as the data management itself needs proper documentaion. Related to #6
1.0
Documentation for backend app - The whole backend app with the messaging and task API, as well as the data management itself needs proper documentaion. Related to #6
non_priority
documentation for backend app the whole backend app with the messaging and task api as well as the data management itself needs proper documentaion related to
0
666,076
22,341,644,715
IssuesEvent
2022-06-15 01:53:56
Couchers-org/web-frontend
https://api.github.com/repos/Couchers-org/web-frontend
opened
Event join button should join on first click, then show menu next clicks
good first issue improvement priority: normal
Instead of always showing the menu
1.0
Event join button should join on first click, then show menu next clicks - Instead of always showing the menu
priority
event join button should join on first click then show menu next clicks instead of always showing the menu
1
58,858
3,092,229,573
IssuesEvent
2015-08-26 16:45:29
brian-team/brian2
https://api.github.com/repos/brian-team/brian2
reopened
Some pre-synaptic updates give incorrect warnings
bug high priority
I'm using the following simple Synapse: ``` synapses = br2.Synapses(neurons, pre='v -= 0.2') ``` and I get a warning: ``` WARNING brian2.codegen.generators.base: Came across an abstract code block that is not well-defined: the outcome may depend on the order of execution. 2 lines of abstract code, first line is: "v -= w (in-place)" ``` However, if I make the following trivial change: ``` synapses = br2.Synapses(neurons, pre='v += -0.2') ``` The warning disappears.
1.0
Some pre-synaptic updates give incorrect warnings - I'm using the following simple Synapse: ``` synapses = br2.Synapses(neurons, pre='v -= 0.2') ``` and I get a warning: ``` WARNING brian2.codegen.generators.base: Came across an abstract code block that is not well-defined: the outcome may depend on the order of execution. 2 lines of abstract code, first line is: "v -= w (in-place)" ``` However, if I make the following trivial change: ``` synapses = br2.Synapses(neurons, pre='v += -0.2') ``` The warning disappears.
priority
some pre synaptic updates give incorrect warnings i m using the following simple synapse synapses synapses neurons pre v and i get a warning warning codegen generators base came across an abstract code block that is not well defined the outcome may depend on the order of execution lines of abstract code first line is v w in place however if i make the following trivial change synapses synapses neurons pre v the warning disappears
1
789,442
27,789,952,678
IssuesEvent
2023-03-17 08:05:27
AY2223S2-CS2103T-W12-2/tp
https://api.github.com/repos/AY2223S2-CS2103T-W12-2/tp
opened
Empty string fields in `json` save file
type.Bug priority.High
# The problem Currently, when a person-entry lacks some optional fields, their saved data in `/data/addressbook.json` looks like this: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "phone" : "", "email" : "", "address" : "", "remark" : "im a cuck", "tagged" : [ ] } ] } ``` where the missing fields are empty strings `""`. This means when they're loaded, those fields are initialised to empty string (eg. `new Address("")`) instead of `Optional.empty()`. And those empty-string fields are currently allowed because we changed the `isValid` methods of the fields to return true when it's an empty string. ```java /** * Returns true if a given string is a valid email. */ public static boolean isValidAddress(String test) { return test.matches(VALIDATION_REGEX) || test.isBlank(); } ``` <br> # Possible solution We could instead represent missing optional fields as `null` in the `.json` save file like: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "phone" : null, "email" : null, "address" : null, "remark" : "im a cuck", "tagged" : [ ] } ] } ``` Or perhaps lose those fields in the `json` file all together: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "remark" : "im a cuck", "tagged" : [ ] } ] } ```
1.0
Empty string fields in `json` save file - # The problem Currently, when a person-entry lacks some optional fields, their saved data in `/data/addressbook.json` looks like this: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "phone" : "", "email" : "", "address" : "", "remark" : "im a cuck", "tagged" : [ ] } ] } ``` where the missing fields are empty strings `""`. This means when they're loaded, those fields are initialised to empty string (eg. `new Address("")`) instead of `Optional.empty()`. And those empty-string fields are currently allowed because we changed the `isValid` methods of the fields to return true when it's an empty string. ```java /** * Returns true if a given string is a valid email. */ public static boolean isValidAddress(String test) { return test.matches(VALIDATION_REGEX) || test.isBlank(); } ``` <br> # Possible solution We could instead represent missing optional fields as `null` in the `.json` save file like: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "phone" : null, "email" : null, "address" : null, "remark" : "im a cuck", "tagged" : [ ] } ] } ``` Or perhaps lose those fields in the `json` file all together: ```json { "persons" : [ { "name" : "Netorare Enjoyer", "remark" : "im a cuck", "tagged" : [ ] } ] } ```
priority
empty string fields in json save file the problem currently when a person entry lacks some optional fields their saved data in data addressbook json looks like this json persons name netorare enjoyer phone email address remark im a cuck tagged where the missing fields are empty strings this means when they re loaded those fields are initialised to empty string eg new address instead of optional empty and those empty string fields are currently allowed because we changed the isvalid methods of the fields to return true when it s an empty string java returns true if a given string is a valid email public static boolean isvalidaddress string test return test matches validation regex test isblank possible solution we could instead represent missing optional fields as null in the json save file like json persons name netorare enjoyer phone null email null address null remark im a cuck tagged or perhaps lose those fields in the json file all together json persons name netorare enjoyer remark im a cuck tagged
1
145,639
11,701,362,557
IssuesEvent
2020-03-06 19:31:10
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Add helm version selection on catalog create
[zube]: To Test team/ui
**What kind of request is this (question/bug/enhancement/feature request):** Enhancement **Helm Version selection** - When adding a catalog, user needs to be able to set the helm version for that catalog - This should be set on create and user should be warned they cannot update it - This cannot be updated - When selecting an app to deploy, user needs to see what version of helm that template is using - When user is selecting a template version, user needs to see what version of helm that template version uses
1.0
Add helm version selection on catalog create - **What kind of request is this (question/bug/enhancement/feature request):** Enhancement **Helm Version selection** - When adding a catalog, user needs to be able to set the helm version for that catalog - This should be set on create and user should be warned they cannot update it - This cannot be updated - When selecting an app to deploy, user needs to see what version of helm that template is using - When user is selecting a template version, user needs to see what version of helm that template version uses
non_priority
add helm version selection on catalog create what kind of request is this question bug enhancement feature request enhancement helm version selection when adding a catalog user needs to be able to set the helm version for that catalog this should be set on create and user should be warned they cannot update it this cannot be updated when selecting an app to deploy user needs to see what version of helm that template is using when user is selecting a template version user needs to see what version of helm that template version uses
0
219,698
17,106,421,045
IssuesEvent
2021-07-09 18:31:54
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
opened
Exception when using "/plot visit <number>", where <number> is 2,147,483,648 or greater
Requires Testing
### Server Implementation Paper ### Server Version 1.16.5 ### Describe the bug Title says it all really. This is the exception: https://pastebin.com/bS0DL5uH ### To Reproduce 1. Type "/plot visit <number>", where <number> is 2,147,483,648 or greater ### Expected behaviour Some kind of catch for the exception, I guess. ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/05e1898542f34713a9553c6cb471e7c0 ### PlotSquared Version 6.0.5-Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? _No response_
1.0
Exception when using "/plot visit <number>", where <number> is 2,147,483,648 or greater - ### Server Implementation Paper ### Server Version 1.16.5 ### Describe the bug Title says it all really. This is the exception: https://pastebin.com/bS0DL5uH ### To Reproduce 1. Type "/plot visit <number>", where <number> is 2,147,483,648 or greater ### Expected behaviour Some kind of catch for the exception, I guess. ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/05e1898542f34713a9553c6cb471e7c0 ### PlotSquared Version 6.0.5-Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? _No response_
non_priority
exception when using plot visit where is or greater server implementation paper server version describe the bug title says it all really this is the exception to reproduce type plot visit where is or greater expected behaviour some kind of catch for the exception i guess screenshots videos no response error log if applicable no response plot debugpaste plotsquared version premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else no response
0
22,990
10,832,242,463
IssuesEvent
2019-11-11 10:10:37
alt-how/altinn-studio
https://api.github.com/repos/alt-how/altinn-studio
opened
CVE-2019-10219 (Medium) detected in hibernate-validator-6.0.17.Final.jar
security vulnerability
## CVE-2019-10219 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-6.0.17.Final.jar</b></p></summary> <p>Hibernate's Bean Validation (JSR-380) reference implementation.</p> <p>Library home page: <a href="http://hibernate.org/validator/">http://hibernate.org/validator/</a></p> <p>Path to dependency file: /tmp/ws-scm/altinn-studio/src/Altinn.Platform/Altinn.Platform.PDF/pom.xml</p> <p>Path to vulnerable library: epository/org/hibernate/validator/hibernate-validator/6.0.17.Final/hibernate-validator-6.0.17.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-validator-6.0.17.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/alt-how/altinn-studio/commit/01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3">01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Hibernate-Validator. The SafeHtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in HTML comments and instructions. This vulnerability can result in an XSS attack. <p>Publish Date: 2019-11-08 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219>CVE-2019-10219</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219</a></p> <p>Release Date: 2019-11-08</p> <p>Fix Resolution: 6.1.0.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10219 (Medium) detected in hibernate-validator-6.0.17.Final.jar - ## CVE-2019-10219 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-6.0.17.Final.jar</b></p></summary> <p>Hibernate's Bean Validation (JSR-380) reference implementation.</p> <p>Library home page: <a href="http://hibernate.org/validator/">http://hibernate.org/validator/</a></p> <p>Path to dependency file: /tmp/ws-scm/altinn-studio/src/Altinn.Platform/Altinn.Platform.PDF/pom.xml</p> <p>Path to vulnerable library: epository/org/hibernate/validator/hibernate-validator/6.0.17.Final/hibernate-validator-6.0.17.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-validator-6.0.17.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/alt-how/altinn-studio/commit/01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3">01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Hibernate-Validator. The SafeHtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in HTML comments and instructions. This vulnerability can result in an XSS attack. <p>Publish Date: 2019-11-08 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219>CVE-2019-10219</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219</a></p> <p>Release Date: 2019-11-08</p> <p>Fix Resolution: 6.1.0.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in hibernate validator final jar cve medium severity vulnerability vulnerable library hibernate validator final jar hibernate s bean validation jsr reference implementation library home page a href path to dependency file tmp ws scm altinn studio src altinn platform altinn platform pdf pom xml path to vulnerable library epository org hibernate validator hibernate validator final hibernate validator final jar dependency hierarchy x hibernate validator final jar vulnerable library found in head commit a href vulnerability details a vulnerability was found in hibernate validator the safehtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in html comments and instructions this vulnerability can result in an xss attack publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution final step up your open source security game with whitesource
0
66,001
3,249,485,993
IssuesEvent
2015-10-18 07:01:52
boarder2/Latest-Chatty-8
https://api.github.com/repos/boarder2/Latest-Chatty-8
opened
Update image assets
low priority
They're all low resolution and don't look good on high dpi devices. This will be tedious.
1.0
Update image assets - They're all low resolution and don't look good on high dpi devices. This will be tedious.
priority
update image assets they re all low resolution and don t look good on high dpi devices this will be tedious
1
61,947
14,643,852,457
IssuesEvent
2020-12-25 19:10:25
AlexRogalskiy/spark4you
https://api.github.com/repos/AlexRogalskiy/spark4you
opened
CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar
security vulnerability
## CVE-2009-2625 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary> <p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.</p> <p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p> <p> Dependency Hierarchy: - spark-core_2.12-3.0.1.jar (Root Library) - hadoop-client-2.7.4.jar - hadoop-hdfs-2.7.4.jar - :x: **xercesImpl-2.9.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework. <p>Publish Date: 2009-08-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux: * JDK and JRE 6 Update 15 or later * JDK and JRE 5.0 Update 20 or later Java SE releases are available at: JDK and JRE 6 Update 15: http://java.sun.com/javase/downloads/index.jsp JRE 6 Update 15: http://java.com/ through the Java Update tool for Microsoft Windows users. JDK 6 Update 15 for Solaris is available in the following patches: * Java SE 6 Update 15 (as delivered in patch 125136-16) * Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit)) * Java SE 6_x86 Update 15 (as delivered in patch 125138-16) * Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit)) JDK and JRE 5.0 Update 20: http://java.sun.com/javase/downloads/index_jdk5.jsp JDK 5.0 Update 20 for Solaris is available in the following patches: * J2SE 5.0 Update 18 (as delivered in patch 118666-21) * J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit)) * J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21) * J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit)) Java SE for Business releases are available at: http://www.sun.com/software/javaseforbusiness/getit_download.jsp Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see: http://www.java.com/en/download/help/5000010800.xml The vendor's advisory is available at: http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar - ## CVE-2009-2625 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary> <p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.</p> <p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p> <p> Dependency Hierarchy: - spark-core_2.12-3.0.1.jar (Root Library) - hadoop-client-2.7.4.jar - hadoop-hdfs-2.7.4.jar - :x: **xercesImpl-2.9.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework. <p>Publish Date: 2009-08-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux: * JDK and JRE 6 Update 15 or later * JDK and JRE 5.0 Update 20 or later Java SE releases are available at: JDK and JRE 6 Update 15: http://java.sun.com/javase/downloads/index.jsp JRE 6 Update 15: http://java.com/ through the Java Update tool for Microsoft Windows users. JDK 6 Update 15 for Solaris is available in the following patches: * Java SE 6 Update 15 (as delivered in patch 125136-16) * Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit)) * Java SE 6_x86 Update 15 (as delivered in patch 125138-16) * Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit)) JDK and JRE 5.0 Update 20: http://java.sun.com/javase/downloads/index_jdk5.jsp JDK 5.0 Update 20 for Solaris is available in the following patches: * J2SE 5.0 Update 18 (as delivered in patch 118666-21) * J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit)) * J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21) * J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit)) Java SE for Business releases are available at: http://www.sun.com/software/javaseforbusiness/getit_download.jsp Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see: http://www.java.com/en/download/help/5000010800.xml The vendor's advisory is available at: http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in xercesimpl jar cve medium severity vulnerability vulnerable library xercesimpl jar is the next generation of high performance fully compliant xml parsers in the apache xerces family this new version of xerces introduces the xerces native interface xni a complete framework for building parser components and configurations that is extremely modular and easy to program path to dependency file modules java pom xml path to vulnerable library home wss scanner repository xerces xercesimpl xercesimpl jar home wss scanner repository xerces xercesimpl xercesimpl jar dependency hierarchy spark core jar root library hadoop client jar hadoop hdfs jar x xercesimpl jar vulnerable library found in head commit a href found in base branch master vulnerability details xmlscanner java in apache java as used in sun java runtime environment jre in jdk and jre before update and jdk and jre before update and in other products allows remote attackers to cause a denial of service infinite loop and application hang via malformed xml input as demonstrated by the codenomicon xml fuzzing framework publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution the vendor has issued a fix for windows solaris and linux jdk and jre update or later jdk and jre update or later java se releases are available at jdk and jre update jre update through the java update tool for microsoft windows users jdk update for solaris is available in the following patches java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch jdk and jre update jdk update for solaris is available in the following patches update as delivered in patch update as delivered in patch update as delivered in patch update as delivered in patch java se for business releases are available at note when installing a new version of the product from a source other than a solaris patch it is recommended that the old affected versions be removed from your system to remove old affected versions on the windows platform please see the vendor s advisory is available at step up your open source security game with whitesource
0
403,475
11,841,558,030
IssuesEvent
2020-03-23 21:02:00
StudioTBA/CoronaIO
https://api.github.com/repos/StudioTBA/CoronaIO
opened
Make PoliceAgent implement Agent
Priority: Critical Refactor
**Is your feature request related to a problem? Please describe.** All agents should implement the abstract class Agent. **Describe the solution you would like** To make things simpler, besides making PoliceAgent implement Agent, create a `Police` folder in `Scripts/Agent/` that contains the contents of `Scripts/FSM`to centralize the scripts related to agents.
1.0
Make PoliceAgent implement Agent - **Is your feature request related to a problem? Please describe.** All agents should implement the abstract class Agent. **Describe the solution you would like** To make things simpler, besides making PoliceAgent implement Agent, create a `Police` folder in `Scripts/Agent/` that contains the contents of `Scripts/FSM`to centralize the scripts related to agents.
priority
make policeagent implement agent is your feature request related to a problem please describe all agents should implement the abstract class agent describe the solution you would like to make things simpler besides making policeagent implement agent create a police folder in scripts agent that contains the contents of scripts fsm to centralize the scripts related to agents
1
45,280
11,623,954,238
IssuesEvent
2020-02-27 09:53:59
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Cannot build Bisq rpm on Debian 10.3 VM
in:build on:Linux
<!-- SUPPORT REQUESTS: This is for reporting bugs in the Bisq app. If you have a support request, please join #support on Bisq's Keybase team at https://keybase.io/team/Bisq --> ### Description The desktop/package/linux/package.sh script fails while trying to create an rpm for version=1.2.7-SNAPSHOT error: line 3: Illegal char '-' (0x2d) in: Version: 1.2.7-SNAPSHOT #### Version 1.2.7-SNAPSHOT ### Steps to reproduce 1. Build src using OpenJDK 11 (on Debian 10.3) 2. Set JAVA_HOME to Oracle JDK 10 3. Run desktop/package/linux/package.sh ### Expected behaviour Bisq-1.2.7-SNAPSHOT.deb and Bisq-1.2.7-SNAPSHOT.rpm are created with no errors. ### Actual behaviour Bisq-1.2.7-SNAPSHOT.rpm could not be created because of the hyphen in the rpm version. #### Device or machine Debian 10.3 VM running inside Ubuntu 18.0.4 Intel i7, Nvidia GTX 960 #### Additional info package.sh debug output: [rpm-build-error.log](https://github.com/bisq-network/bisq/files/4246509/rpm-build-error.log) According to https://twiki.cern.ch/twiki/bin/view/Main/RPMAndDebVersioning RPM (4.7.1) version format version Mandatory Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot) Spaces not allowed release Optional Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot) Spaces not allowed According to https://www.debian.org/doc/debian-policy/ch-controlfields.html#s-f-version: upstream_version The upstream_version may contain only alphanumerics [6] and the characters . + - ~ (full stop, plus, hyphen, tilde) and should start with a digit. *** If there is no debian_revision then hyphens are not allowed. *** debian_revision This part of the version number specifies the version of the Debian package based on the upstream version. It may contain only alphanumerics and the characters + . ~ (plus, full stop, tilde) and is compared in the same way as the upstream_version is. Debian 10.3 can create Bisq-1.2.7-SNAPSHOT.deb, but not Bisq-1.2.7-SNAPSHOT.rpm Workaround: I created the rpm Bisq-1.2.7+SNAPSHOT.rpm on Debian 10.3 by creating a variable in linux/package.sh that replaces "-" with a valid "+" rpmVersion=${version//-/+} and substituting $rpmVersion for $version in the rpm generation related code in package.sh. I also created Bisq-1.2.7SNAPSHOT.rpm by replacing "-" with "". ------- The Bisq-1.2.7+SNAPSHOT.rpm file I created on the debian vm was installed using dnf on another rpm based fedora vm; it created the fedora menu item and bisq started with no problems.
1.0
Cannot build Bisq rpm on Debian 10.3 VM - <!-- SUPPORT REQUESTS: This is for reporting bugs in the Bisq app. If you have a support request, please join #support on Bisq's Keybase team at https://keybase.io/team/Bisq --> ### Description The desktop/package/linux/package.sh script fails while trying to create an rpm for version=1.2.7-SNAPSHOT error: line 3: Illegal char '-' (0x2d) in: Version: 1.2.7-SNAPSHOT #### Version 1.2.7-SNAPSHOT ### Steps to reproduce 1. Build src using OpenJDK 11 (on Debian 10.3) 2. Set JAVA_HOME to Oracle JDK 10 3. Run desktop/package/linux/package.sh ### Expected behaviour Bisq-1.2.7-SNAPSHOT.deb and Bisq-1.2.7-SNAPSHOT.rpm are created with no errors. ### Actual behaviour Bisq-1.2.7-SNAPSHOT.rpm could not be created because of the hyphen in the rpm version. #### Device or machine Debian 10.3 VM running inside Ubuntu 18.0.4 Intel i7, Nvidia GTX 960 #### Additional info package.sh debug output: [rpm-build-error.log](https://github.com/bisq-network/bisq/files/4246509/rpm-build-error.log) According to https://twiki.cern.ch/twiki/bin/view/Main/RPMAndDebVersioning RPM (4.7.1) version format version Mandatory Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot) Spaces not allowed release Optional Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot) Spaces not allowed According to https://www.debian.org/doc/debian-policy/ch-controlfields.html#s-f-version: upstream_version The upstream_version may contain only alphanumerics [6] and the characters . + - ~ (full stop, plus, hyphen, tilde) and should start with a digit. *** If there is no debian_revision then hyphens are not allowed. *** debian_revision This part of the version number specifies the version of the Debian package based on the upstream version. It may contain only alphanumerics and the characters + . ~ (plus, full stop, tilde) and is compared in the same way as the upstream_version is. Debian 10.3 can create Bisq-1.2.7-SNAPSHOT.deb, but not Bisq-1.2.7-SNAPSHOT.rpm Workaround: I created the rpm Bisq-1.2.7+SNAPSHOT.rpm on Debian 10.3 by creating a variable in linux/package.sh that replaces "-" with a valid "+" rpmVersion=${version//-/+} and substituting $rpmVersion for $version in the rpm generation related code in package.sh. I also created Bisq-1.2.7SNAPSHOT.rpm by replacing "-" with "". ------- The Bisq-1.2.7+SNAPSHOT.rpm file I created on the debian vm was installed using dnf on another rpm based fedora vm; it created the fedora menu item and bisq started with no problems.
non_priority
cannot build bisq rpm on debian vm support requests this is for reporting bugs in the bisq app if you have a support request please join support on bisq s keybase team at description the desktop package linux package sh script fails while trying to create an rpm for version snapshot error line illegal char in version snapshot version snapshot steps to reproduce build src using openjdk on debian set java home to oracle jdk run desktop package linux package sh expected behaviour bisq snapshot deb and bisq snapshot rpm are created with no errors actual behaviour bisq snapshot rpm could not be created because of the hyphen in the rpm version device or machine debian vm running inside ubuntu intel nvidia gtx additional info package sh debug output according to rpm version format version mandatory characters not allowed hyphen slash tilde double dot spaces not allowed release optional characters not allowed hyphen slash tilde double dot spaces not allowed according to upstream version the upstream version may contain only alphanumerics and the characters full stop plus hyphen tilde and should start with a digit if there is no debian revision then hyphens are not allowed debian revision this part of the version number specifies the version of the debian package based on the upstream version it may contain only alphanumerics and the characters plus full stop tilde and is compared in the same way as the upstream version is debian can create bisq snapshot deb but not bisq snapshot rpm workaround i created the rpm bisq snapshot rpm on debian by creating a variable in linux package sh that replaces with a valid rpmversion version and substituting rpmversion for version in the rpm generation related code in package sh i also created bisq rpm by replacing with the bisq snapshot rpm file i created on the debian vm was installed using dnf on another rpm based fedora vm it created the fedora menu item and bisq started with no problems
0
298,822
22,574,033,044
IssuesEvent
2022-06-28 05:04:47
OTRF/OSSEM-DD
https://api.github.com/repos/OTRF/OSSEM-DD
closed
Update Sysmon Dictionary (Windows) - Event 3
documentation
Add log example in XML format Can use the following reference: ``` <EventData> <Data Name="RuleName">-</Data> <Data Name="UtcTime">2021-11-03 05:01:26.226</Data> <Data Name="ProcessGuid">{3710b5c6-f3dd-6181-4000-000000000a00}</Data> <Data Name="ProcessId">2484</Data> <Data Name="Image">C:\Windows\System32\svchost.exe</Data> <Data Name="User">NT AUTHORITY\SYSTEM</Data> <Data Name="Protocol">tcp</Data> <Data Name="Initiated">true</Data> <Data Name="SourceIsIpv6">false</Data> <Data Name="SourceIp">192.168.10.8</Data> <Data Name="SourceHostname">-</Data> <Data Name="SourcePort">50026</Data> <Data Name="SourcePortName">-</Data> <Data Name="DestinationIsIpv6">false</Data> <Data Name="DestinationIp">52.167.249.196</Data> <Data Name="DestinationHostname">-</Data> <Data Name="DestinationPort">443</Data> <Data Name="DestinationPortName">-</Data> </EventData> ```
1.0
Update Sysmon Dictionary (Windows) - Event 3 - Add log example in XML format Can use the following reference: ``` <EventData> <Data Name="RuleName">-</Data> <Data Name="UtcTime">2021-11-03 05:01:26.226</Data> <Data Name="ProcessGuid">{3710b5c6-f3dd-6181-4000-000000000a00}</Data> <Data Name="ProcessId">2484</Data> <Data Name="Image">C:\Windows\System32\svchost.exe</Data> <Data Name="User">NT AUTHORITY\SYSTEM</Data> <Data Name="Protocol">tcp</Data> <Data Name="Initiated">true</Data> <Data Name="SourceIsIpv6">false</Data> <Data Name="SourceIp">192.168.10.8</Data> <Data Name="SourceHostname">-</Data> <Data Name="SourcePort">50026</Data> <Data Name="SourcePortName">-</Data> <Data Name="DestinationIsIpv6">false</Data> <Data Name="DestinationIp">52.167.249.196</Data> <Data Name="DestinationHostname">-</Data> <Data Name="DestinationPort">443</Data> <Data Name="DestinationPortName">-</Data> </EventData> ```
non_priority
update sysmon dictionary windows event add log example in xml format can use the following reference c windows svchost exe nt authority system tcp true false false
0
22,851
11,795,605,010
IssuesEvent
2020-03-18 09:17:07
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Issues after resetting existing service principal credentials
Pri2 container-service/svc cxp product-question triaged
By following this guide, I have resat the service principal credentials for 3 different Kubernetes clusters running in AKS. They had different sizes, VM types and were running different Kubernetes versions. After updating the clusters became unresponsive. I couldn't schedule new pods to run, update docker images for deployments, etc. That issue disappeared automatically after about an hour. After that I had an issue with running the Kubernetes Dashboard. When opening the proxy link, the dashboard would randomly fail loading with an error that the kubernetes-service did not exist. I also started to get error logs about services not being able to find each other. It seemed as if DNS failed between random Kubernetes services. After some debugging I found out that for 50 % of the nodes, the kube-proxy pod from the kube-system namespace was in Pending state. By manually deleting these Pending pods, they were recreated automatically and got into a working state, which solved the issue. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 0b842222-f6ac-5266-227f-7e400610e07b * Version Independent ID: 429cd3bf-2691-f633-2e0d-e70e5d8d0e03 * Content: [Reset the credentials for an Azure Kubernetes Service (AKS) cluster - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/update-credentials#feedback) * Content Source: [articles/aks/update-credentials.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/update-credentials.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
1.0
Issues after resetting existing service principal credentials - By following this guide, I have resat the service principal credentials for 3 different Kubernetes clusters running in AKS. They had different sizes, VM types and were running different Kubernetes versions. After updating the clusters became unresponsive. I couldn't schedule new pods to run, update docker images for deployments, etc. That issue disappeared automatically after about an hour. After that I had an issue with running the Kubernetes Dashboard. When opening the proxy link, the dashboard would randomly fail loading with an error that the kubernetes-service did not exist. I also started to get error logs about services not being able to find each other. It seemed as if DNS failed between random Kubernetes services. After some debugging I found out that for 50 % of the nodes, the kube-proxy pod from the kube-system namespace was in Pending state. By manually deleting these Pending pods, they were recreated automatically and got into a working state, which solved the issue. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 0b842222-f6ac-5266-227f-7e400610e07b * Version Independent ID: 429cd3bf-2691-f633-2e0d-e70e5d8d0e03 * Content: [Reset the credentials for an Azure Kubernetes Service (AKS) cluster - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/update-credentials#feedback) * Content Source: [articles/aks/update-credentials.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/update-credentials.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
non_priority
issues after resetting existing service principal credentials by following this guide i have resat the service principal credentials for different kubernetes clusters running in aks they had different sizes vm types and were running different kubernetes versions after updating the clusters became unresponsive i couldn t schedule new pods to run update docker images for deployments etc that issue disappeared automatically after about an hour after that i had an issue with running the kubernetes dashboard when opening the proxy link the dashboard would randomly fail loading with an error that the kubernetes service did not exist i also started to get error logs about services not being able to find each other it seemed as if dns failed between random kubernetes services after some debugging i found out that for of the nodes the kube proxy pod from the kube system namespace was in pending state by manually deleting these pending pods they were recreated automatically and got into a working state which solved the issue document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned
0
757,208
26,500,530,640
IssuesEvent
2023-01-18 09:56:36
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Bluetooth: host: central connection context leak
bug priority: low area: Bluetooth area: Bluetooth Host
**Describe the bug** I've been seeing this issue with the zephyr host. I am using NCS (2c8ef25e7) and the two samples for direction finding. The connection context for the central part is left in an incomplete state when the connection is timed out. **To Reproduce** `west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_peripheral -d peripheral -- -DCONFIG_BT_LL_SW_SPLIT=y` and `west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_central -d central -- -DCONFIG_BT_LL_SW_SPLIT=y` Flashing both samples, and resetting the peripheral side using the RESET button on the DK the central eventually runs into a problem where it is unable to create the connection. ``` [01:13:22.249,298] <wrn> bt_conn: Found valid connection in disconnected state [01:13:22.566,467] <wrn> b[DEVICE]: 17:5D:F5:1E:99:7D (random), AD evt type 3, AD data len 31, RSSI -82 [DEVICE]: D8:9A:D6:93:B8:0E (random), AD evt type 0, AD data len 8, RSSI -30 [AD]: 1 data_len 1 [AD]: 39 data_len 3 Create conn failed (err -22) ``` **Impact** Connection context is leaking, creating a new connection is no longer possible when this happens. **Environment (please complete the following information):** - OS: Linux - Toolchain: Zephyr SDK 0.13.1 - Commit SHA or Version used: NCS (2c8ef25e7)
1.0
Bluetooth: host: central connection context leak - **Describe the bug** I've been seeing this issue with the zephyr host. I am using NCS (2c8ef25e7) and the two samples for direction finding. The connection context for the central part is left in an incomplete state when the connection is timed out. **To Reproduce** `west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_peripheral -d peripheral -- -DCONFIG_BT_LL_SW_SPLIT=y` and `west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_central -d central -- -DCONFIG_BT_LL_SW_SPLIT=y` Flashing both samples, and resetting the peripheral side using the RESET button on the DK the central eventually runs into a problem where it is unable to create the connection. ``` [01:13:22.249,298] <wrn> bt_conn: Found valid connection in disconnected state [01:13:22.566,467] <wrn> b[DEVICE]: 17:5D:F5:1E:99:7D (random), AD evt type 3, AD data len 31, RSSI -82 [DEVICE]: D8:9A:D6:93:B8:0E (random), AD evt type 0, AD data len 8, RSSI -30 [AD]: 1 data_len 1 [AD]: 39 data_len 3 Create conn failed (err -22) ``` **Impact** Connection context is leaking, creating a new connection is no longer possible when this happens. **Environment (please complete the following information):** - OS: Linux - Toolchain: Zephyr SDK 0.13.1 - Commit SHA or Version used: NCS (2c8ef25e7)
priority
bluetooth host central connection context leak describe the bug i ve been seeing this issue with the zephyr host i am using ncs and the two samples for direction finding the connection context for the central part is left in an incomplete state when the connection is timed out to reproduce west build b zephyr samples bluetooth direction finding peripheral d peripheral dconfig bt ll sw split y and west build b zephyr samples bluetooth direction finding central d central dconfig bt ll sw split y flashing both samples and resetting the peripheral side using the reset button on the dk the central eventually runs into a problem where it is unable to create the connection bt conn found valid connection in disconnected state b random ad evt type ad data len rssi random ad evt type ad data len rssi data len data len create conn failed err impact connection context is leaking creating a new connection is no longer possible when this happens environment please complete the following information os linux toolchain zephyr sdk commit sha or version used ncs
1
275,394
8,575,623,432
IssuesEvent
2018-11-12 17:48:38
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
Vector transform code for AsDirection needs changes to account for scale.
Bug Likelihood: 3 - Occasional Priority: Normal Severity: 4 - Crash / Wrong Results
Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection) needs to be changed from V' = [T(P+epsilon*V) - T(P)] / epsilon to: V' = [T(P+epsilon*Vn) - T(P)] / epsilon where Vn = V/|V| -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 507 Status: Resolved Project: VisIt Tracker: Bug Priority: Normal Subject: Vector transform code for AsDirection needs changes to account for scale. Assigned to: Kathleen Biagas Category: Target version: 2.1.2 Author: Kathleen Biagas Start: 12/09/2010 Due date: % Done: 0 Estimated time: 2.0 Created: 12/09/2010 08:06 pm Updated: 12/10/2010 01:30 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.0.1 Impact: Expected Use: OS: All Support Group: Any Description: Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection) needs to be changed from V' = [T(P+epsilon*V) - T(P)] / epsilon to: V' = [T(P+epsilon*Vn) - T(P)] / epsilon where Vn = V/|V| Comments: Implemented the changes suggested, with the addition of re-scaling the results by the vector magnitude as follows:V' = ([T(P+epsilon*Vn) - T(P)] / epsilon) * |V|SVN Revisions 13236 (2.1RC) and 13238 (trunk). M /src/avt/Filters/avtCoordSystemConvert.C M /src/avt/Filters/avtTransform.C M /src/operators/Project/avtProjectFilter.C
1.0
Vector transform code for AsDirection needs changes to account for scale. - Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection) needs to be changed from V' = [T(P+epsilon*V) - T(P)] / epsilon to: V' = [T(P+epsilon*Vn) - T(P)] / epsilon where Vn = V/|V| -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 507 Status: Resolved Project: VisIt Tracker: Bug Priority: Normal Subject: Vector transform code for AsDirection needs changes to account for scale. Assigned to: Kathleen Biagas Category: Target version: 2.1.2 Author: Kathleen Biagas Start: 12/09/2010 Due date: % Done: 0 Estimated time: 2.0 Created: 12/09/2010 08:06 pm Updated: 12/10/2010 01:30 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.0.1 Impact: Expected Use: OS: All Support Group: Any Description: Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection) needs to be changed from V' = [T(P+epsilon*V) - T(P)] / epsilon to: V' = [T(P+epsilon*Vn) - T(P)] / epsilon where Vn = V/|V| Comments: Implemented the changes suggested, with the addition of re-scaling the results by the vector magnitude as follows:V' = ([T(P+epsilon*Vn) - T(P)] / epsilon) * |V|SVN Revisions 13236 (2.1RC) and 13238 (trunk). M /src/avt/Filters/avtCoordSystemConvert.C M /src/avt/Filters/avtTransform.C M /src/operators/Project/avtProjectFilter.C
priority
vector transform code for asdirection needs changes to account for scale per discussion between matt o brien and jeremyon visit users list vector transform case treat asdirection needs to be changed from v epsilon to v epsilon where vn v v redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject vector transform code for asdirection needs changes to account for scale assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description per discussion between matt o brien and jeremyon visit users list vector transform case treat asdirection needs to be changed from v epsilon to v epsilon where vn v v comments implemented the changes suggested with the addition of re scaling the results by the vector magnitude as follows v epsilon v svn revisions and trunk m src avt filters avtcoordsystemconvert c m src avt filters avttransform c m src operators project avtprojectfilter c
1
18,532
13,055,715,385
IssuesEvent
2020-07-30 02:31:12
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
No intellisense in morris UI
bug interface/infrastructure
When typing parameters into the morris UI, there is no intellisense popup.
1.0
No intellisense in morris UI - When typing parameters into the morris UI, there is no intellisense popup.
non_priority
no intellisense in morris ui when typing parameters into the morris ui there is no intellisense popup
0
199,128
15,024,954,891
IssuesEvent
2021-02-01 20:22:17
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts - Kibana Home click on collapse on observability in side nav to test a11y of collapse button
Feature:Home Project:Accessibility failed-test
A test failed on a tracked branch ``` Error: a11y report: VIOLATION [aria-required-parent]: Ensures elements with an ARIA role that require parent roles are contained by them Help: https://dequeuniversity.com/rules/axe/4.0/aria-required-parent?application=axeAPI Elements: - .euiTab-isSelected - a[href$="explorer"] > .euiTab[role="tab"] - a[href$="settings"] > .euiTab[role="tab"] at Accessibility.testAxeReport (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:91:15) at Accessibility.testAppSnapshot (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:60:18) at process._tickCallback (internal/process/next_tick.js:68:7) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/8857/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts","test.name":"Kibana Home click on collapse on observability in side nav to test a11y of collapse button","test.failCount":1}} -->
1.0
Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts - Kibana Home click on collapse on observability in side nav to test a11y of collapse button - A test failed on a tracked branch ``` Error: a11y report: VIOLATION [aria-required-parent]: Ensures elements with an ARIA role that require parent roles are contained by them Help: https://dequeuniversity.com/rules/axe/4.0/aria-required-parent?application=axeAPI Elements: - .euiTab-isSelected - a[href$="explorer"] > .euiTab[role="tab"] - a[href$="settings"] > .euiTab[role="tab"] at Accessibility.testAxeReport (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:91:15) at Accessibility.testAppSnapshot (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:60:18) at process._tickCallback (internal/process/next_tick.js:68:7) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/8857/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts","test.name":"Kibana Home click on collapse on observability in side nav to test a11y of collapse button","test.failCount":1}} -->
non_priority
failing test x pack accessibility tests x pack test accessibility apps home·ts kibana home click on collapse on observability in side nav to test of collapse button a test failed on a tracked branch error report violation ensures elements with an aria role that require parent roles are contained by them help elements euitab isselected a euitab a euitab at accessibility testaxereport dev shm workspace parallel kibana test accessibility services ts at accessibility testappsnapshot dev shm workspace parallel kibana test accessibility services ts at process tickcallback internal process next tick js first failure
0
453,962
13,092,596,751
IssuesEvent
2020-08-03 08:53:09
Candice-Louw/SORMAS.DE
https://api.github.com/repos/Candice-Louw/SORMAS.DE
closed
33| DE | Cases and Contacts | Option to document that quarantine talk took place with date and comment field
Super High Priority question
### Problem Description ### As a Gesundheitsamt reporter I would like to indicate that a discussion about quarantine has taken place (check-box) with a case/contact on a specific date (date selector field) and add further comments to the discussion if necessary (free-type text box). ### Steps to Reproduce ### Steps to reach section: Cases > Select case > Case Information: CASE tab > scroll down to "Quarantine" Contacts > Select contact > Contact Information: CONTACT tab > scroll down to "Quarantine" ### Suggested Addition ### Addition of 'quarantine discussion' drop-down field. Options in drop-down field: (DE)> " ", Ja, Nein, -nicht ermittelbar- (EN)> " ", Yes, No, Unknown IF "Ja/Yes" answered in drop-down field do the following { Make date field visible with caption: "Date of quarantine discussion". Make free-text comment field visible with caption: "Comments". } ### Screenshots ### <img width="1146" alt="QuarantineDiscussion" src="https://user-images.githubusercontent.com/65529128/88832399-0cb85b00-d1d1-11ea-9a1e-8b4b6b2dbaa4.png">
1.0
33| DE | Cases and Contacts | Option to document that quarantine talk took place with date and comment field - ### Problem Description ### As a Gesundheitsamt reporter I would like to indicate that a discussion about quarantine has taken place (check-box) with a case/contact on a specific date (date selector field) and add further comments to the discussion if necessary (free-type text box). ### Steps to Reproduce ### Steps to reach section: Cases > Select case > Case Information: CASE tab > scroll down to "Quarantine" Contacts > Select contact > Contact Information: CONTACT tab > scroll down to "Quarantine" ### Suggested Addition ### Addition of 'quarantine discussion' drop-down field. Options in drop-down field: (DE)> " ", Ja, Nein, -nicht ermittelbar- (EN)> " ", Yes, No, Unknown IF "Ja/Yes" answered in drop-down field do the following { Make date field visible with caption: "Date of quarantine discussion". Make free-text comment field visible with caption: "Comments". } ### Screenshots ### <img width="1146" alt="QuarantineDiscussion" src="https://user-images.githubusercontent.com/65529128/88832399-0cb85b00-d1d1-11ea-9a1e-8b4b6b2dbaa4.png">
priority
de cases and contacts option to document that quarantine talk took place with date and comment field problem description as a gesundheitsamt reporter i would like to indicate that a discussion about quarantine has taken place check box with a case contact on a specific date date selector field and add further comments to the discussion if necessary free type text box steps to reproduce steps to reach section cases select case case information case tab scroll down to quarantine contacts select contact contact information contact tab scroll down to quarantine suggested addition addition of quarantine discussion drop down field options in drop down field de ja nein nicht ermittelbar en yes no unknown if ja yes answered in drop down field do the following make date field visible with caption date of quarantine discussion make free text comment field visible with caption comments screenshots img width alt quarantinediscussion src
1
12,096
3,040,941,462
IssuesEvent
2015-08-07 18:16:27
aspnet/dnx
https://api.github.com/repos/aspnet/dnx
opened
Add error codes
needs design
It would be helpful if dnx errors/warnings each had their own unique error codes. This will be helpful in case Visual Studio needs to de-duplicate errors and it also enables a help link for each error message. cc @glennc @BillHiebert @davidfowl @DamianEdwards
1.0
Add error codes - It would be helpful if dnx errors/warnings each had their own unique error codes. This will be helpful in case Visual Studio needs to de-duplicate errors and it also enables a help link for each error message. cc @glennc @BillHiebert @davidfowl @DamianEdwards
non_priority
add error codes it would be helpful if dnx errors warnings each had their own unique error codes this will be helpful in case visual studio needs to de duplicate errors and it also enables a help link for each error message cc glennc billhiebert davidfowl damianedwards
0
56,824
13,929,986,730
IssuesEvent
2020-10-22 01:10:54
RTXteam/RTX
https://api.github.com/repos/RTXteam/RTX
reopened
Address warnings in `filter_kg_and_remap_predicates.log`
housekeeping kg2 verify this fix in next KG2 build
The following warnings appear in `filter_kg_and_remap_predicates.log` from KG2.3.5 build. ``` 118 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_stressor_via 119 relation curie is missing from the YAML config file: SO:derives_from 120 relation curie is missing from the YAML config file: SO:has_part 121 relation curie is missing from the YAML config file: SO:adjacent_to 122 relation curie is missing from the YAML config file: SO:guided_by 123 relation curie is missing from the YAML config file: SO:has_quality 124 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with 125 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_receptor_via 126 relation curie is missing from the YAML config file: SO:contains 127 relation curie is missing from the YAML config file: RO:0002491 128 relation curie is missing from the YAML config file: SO:non_functional_homolog_of 129 relation curie is missing from the YAML config file: SO:transcribed_to 130 relation curie is missing from the YAML config file: RO:0002309 131 relation curie is missing from the YAML config file: SO:transcribed_from 132 relation curie is missing from the YAML config file: SO:member_of 133 relation curie is missing from the YAML config file: SO:overlaps 134 relation curie is missing from the YAML config file: MONDO:has_onset ```
1.0
Address warnings in `filter_kg_and_remap_predicates.log` - The following warnings appear in `filter_kg_and_remap_predicates.log` from KG2.3.5 build. ``` 118 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_stressor_via 119 relation curie is missing from the YAML config file: SO:derives_from 120 relation curie is missing from the YAML config file: SO:has_part 121 relation curie is missing from the YAML config file: SO:adjacent_to 122 relation curie is missing from the YAML config file: SO:guided_by 123 relation curie is missing from the YAML config file: SO:has_quality 124 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with 125 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_receptor_via 126 relation curie is missing from the YAML config file: SO:contains 127 relation curie is missing from the YAML config file: RO:0002491 128 relation curie is missing from the YAML config file: SO:non_functional_homolog_of 129 relation curie is missing from the YAML config file: SO:transcribed_to 130 relation curie is missing from the YAML config file: RO:0002309 131 relation curie is missing from the YAML config file: SO:transcribed_from 132 relation curie is missing from the YAML config file: SO:member_of 133 relation curie is missing from the YAML config file: SO:overlaps 134 relation curie is missing from the YAML config file: MONDO:has_onset ```
non_priority
address warnings in filter kg and remap predicates log the following warnings appear in filter kg and remap predicates log from build relation curie is missing from the yaml config file obo exo obo interacts with an exposure stressor via relation curie is missing from the yaml config file so derives from relation curie is missing from the yaml config file so has part relation curie is missing from the yaml config file so adjacent to relation curie is missing from the yaml config file so guided by relation curie is missing from the yaml config file so has quality relation curie is missing from the yaml config file obo exo obo interacts with relation curie is missing from the yaml config file obo exo obo interacts with an exposure receptor via relation curie is missing from the yaml config file so contains relation curie is missing from the yaml config file ro relation curie is missing from the yaml config file so non functional homolog of relation curie is missing from the yaml config file so transcribed to relation curie is missing from the yaml config file ro relation curie is missing from the yaml config file so transcribed from relation curie is missing from the yaml config file so member of relation curie is missing from the yaml config file so overlaps relation curie is missing from the yaml config file mondo has onset
0
184,629
6,714,778,794
IssuesEvent
2017-10-13 18:18:41
Datawheel/codelife
https://api.github.com/repos/Datawheel/codelife
closed
Add "disable sharing" on user page for logged-in admins
priority: high type: feature
In response to repeated offenses, admins should be able to disable any user's ability to share projects and codeblocks via a switch on that user's profile page.
1.0
Add "disable sharing" on user page for logged-in admins - In response to repeated offenses, admins should be able to disable any user's ability to share projects and codeblocks via a switch on that user's profile page.
priority
add disable sharing on user page for logged in admins in response to repeated offenses admins should be able to disable any user s ability to share projects and codeblocks via a switch on that user s profile page
1
24,621
7,531,279,202
IssuesEvent
2018-04-15 03:25:21
travis-ci/travis-ci
https://api.github.com/repos/travis-ci/travis-ci
closed
Travis -proxy and -proxy-tunnel flags for connecting with Sauce Labs
feature-request stale travis-build
We are behind a proxy and are trying to get sauce labs to work. For this, we need to be able to set the `-proxy` and `-proxy-tunnel` flags (we have got it to work via command line tool running `sc -u username -k key -proxy proxy:port -proxy-tunnel`). However, setting these flags does not seem possible in Travis ([additional options](https://docs.travis-ci.com/user/sauce-connect/#Additional-options)). Will this be supported, or is there a workaround for this?
1.0
Travis -proxy and -proxy-tunnel flags for connecting with Sauce Labs - We are behind a proxy and are trying to get sauce labs to work. For this, we need to be able to set the `-proxy` and `-proxy-tunnel` flags (we have got it to work via command line tool running `sc -u username -k key -proxy proxy:port -proxy-tunnel`). However, setting these flags does not seem possible in Travis ([additional options](https://docs.travis-ci.com/user/sauce-connect/#Additional-options)). Will this be supported, or is there a workaround for this?
non_priority
travis proxy and proxy tunnel flags for connecting with sauce labs we are behind a proxy and are trying to get sauce labs to work for this we need to be able to set the proxy and proxy tunnel flags we have got it to work via command line tool running sc u username k key proxy proxy port proxy tunnel however setting these flags does not seem possible in travis will this be supported or is there a workaround for this
0
200,895
22,916,013,757
IssuesEvent
2022-07-17 01:08:53
nanopathi/system_bt_AOSP10_r33_CVE-2021-0522
https://api.github.com/repos/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522
reopened
CVE-2019-2227 (Medium) detected in https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33
security vulnerability
## CVE-2019-2227 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33</b></p></summary> <p> <p>AOSP Platform System bt</p> <p>Library home page: <a href=https://source.codeaurora.org/external/imx/aosp/platform/system/bt/>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522/commit/4f42a84fe2cb3c64302354c9278d446ed65c76f5">4f42a84fe2cb3c64302354c9278d446ed65c76f5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/bta/av/bta_av_aact.cc</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In DeepCopy of btif_av.cc, there is a possible out of bounds read due to improper casting. This could lead to remote information disclosure over Bluetooth with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10Android ID: A-140768453 <p>Publish Date: 2019-12-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2227>CVE-2019-2227</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227</a></p> <p>Release Date: 2019-12-06</p> <p>Fix Resolution: android-9.0.0_r51;android-10.0.0_r17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-2227 (Medium) detected in https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33 - ## CVE-2019-2227 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33</b></p></summary> <p> <p>AOSP Platform System bt</p> <p>Library home page: <a href=https://source.codeaurora.org/external/imx/aosp/platform/system/bt/>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522/commit/4f42a84fe2cb3c64302354c9278d446ed65c76f5">4f42a84fe2cb3c64302354c9278d446ed65c76f5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/bta/av/bta_av_aact.cc</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In DeepCopy of btif_av.cc, there is a possible out of bounds read due to improper casting. This could lead to remote information disclosure over Bluetooth with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10Android ID: A-140768453 <p>Publish Date: 2019-12-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2227>CVE-2019-2227</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227</a></p> <p>Release Date: 2019-12-06</p> <p>Fix Resolution: android-9.0.0_r51;android-10.0.0_r17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in cve medium severity vulnerability vulnerable library aosp platform system bt library home page a href found in head commit a href found in base branch master vulnerable source files bta av bta av aact cc vulnerability details in deepcopy of btif av cc there is a possible out of bounds read due to improper casting this could lead to remote information disclosure over bluetooth with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android step up your open source security game with whitesource
0
799,227
28,302,713,029
IssuesEvent
2023-04-10 07:54:47
AY2223S2-CS2113-W12-3/tp
https://api.github.com/repos/AY2223S2-CS2113-W12-3/tp
closed
Bug: Possible to add item with price/quantity = 0 but InventoryData file becomes corrupted
type.Bug priority.High
It's possible to add an item with price = 0 or quantity = 0 successfully through the program, but on startup the inventory data file's status will be corrupted. Probably because `checkFileValidSession` will return corrupt file status when `qty <= 0`/`price <= 0` . ![quantity_0_corrupt](https://user-images.githubusercontent.com/88529907/230792040-2219c89d-14d2-4550-85ed-5b6da9fb419f.PNG)
1.0
Bug: Possible to add item with price/quantity = 0 but InventoryData file becomes corrupted - It's possible to add an item with price = 0 or quantity = 0 successfully through the program, but on startup the inventory data file's status will be corrupted. Probably because `checkFileValidSession` will return corrupt file status when `qty <= 0`/`price <= 0` . ![quantity_0_corrupt](https://user-images.githubusercontent.com/88529907/230792040-2219c89d-14d2-4550-85ed-5b6da9fb419f.PNG)
priority
bug possible to add item with price quantity but inventorydata file becomes corrupted it s possible to add an item with price or quantity successfully through the program but on startup the inventory data file s status will be corrupted probably because checkfilevalidsession will return corrupt file status when qty price
1
570,467
17,023,118,477
IssuesEvent
2021-07-03 00:27:28
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
OpenLayers compatibility - javascript API for embedding OSM in webpages
Component: api Priority: major Resolution: fixed Type: enhancement
**[Submitted to the original trac issue database at 3.42pm, Monday, 29th May 2006]** Decoding what crschmidt said here: http://lists.openstreetmap.org/pipermail/talk/2006-May/004348.html so that we can use OpenLayers on the OSM website, and so that OSM's WMS server works with OpenLayers for everyone. Current tile implementation is broken for adding markers and info windows to the map. OpenLayers will fix that.
1.0
OpenLayers compatibility - javascript API for embedding OSM in webpages - **[Submitted to the original trac issue database at 3.42pm, Monday, 29th May 2006]** Decoding what crschmidt said here: http://lists.openstreetmap.org/pipermail/talk/2006-May/004348.html so that we can use OpenLayers on the OSM website, and so that OSM's WMS server works with OpenLayers for everyone. Current tile implementation is broken for adding markers and info windows to the map. OpenLayers will fix that.
priority
openlayers compatibility javascript api for embedding osm in webpages decoding what crschmidt said here so that we can use openlayers on the osm website and so that osm s wms server works with openlayers for everyone current tile implementation is broken for adding markers and info windows to the map openlayers will fix that
1
767,487
26,927,623,433
IssuesEvent
2023-02-07 14:48:17
daisy/ebraille
https://api.github.com/repos/daisy/ebraille
opened
Automatic, converted, and human prepared files will have different levels of quality and expectations
use case High Priority content spec metadata
I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format. *Detail* The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format. *Proposal* Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion. With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
1.0
Automatic, converted, and human prepared files will have different levels of quality and expectations - I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format. *Detail* The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format. *Proposal* Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion. With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
priority
automatic converted and human prepared files will have different levels of quality and expectations i am a braille user and my expectations for the quality of file i am receiving will vary based on whether the file was prepared by a transcriber an automatic process or converted from a braille file in an older format detail the best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber next will likely be those converted from an automatic process finally there will be files prepared by converting a braille file in an older format proposal metadata should indicate whether the file was prepared by a human an automatic process or converted from an older braille file if a human it would be good if it indicated the person s level of braille competence professional vs amateur though this aspect needs more thought if an automatic process it would be good if it indicated the file type of the source file as each will have strengths and weaknesses the program used for the automatic process would also be good to indicate finally files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion with this information schools libraries and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes programs and starting files work best for them and their needs
1
39,581
16,048,098,769
IssuesEvent
2021-04-22 15:44:46
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
support for secrets_id in aws_dms_endpoint
enhancement service/databasemigrationservice
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The aws_dms_endpoint allows to configure AWS Secrets manager secret with database details like host, user, password. Please add this to the resource. ### New or Affected Resource(s) * aws_dms_endpoint ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. For # security, you can also encrypt the files using our GPG public key. ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> * #0000
2.0
support for secrets_id in aws_dms_endpoint - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The aws_dms_endpoint allows to configure AWS Secrets manager secret with database details like host, user, password. Please add this to the resource. ### New or Affected Resource(s) * aws_dms_endpoint ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. For # security, you can also encrypt the files using our GPG public key. ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> * #0000
non_priority
support for secrets id in aws dms endpoint community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the aws dms endpoint allows to configure aws secrets manager secret with database details like host user password please add this to the resource new or affected resource s aws dms endpoint potential terraform configuration hcl copy paste your terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
85,175
16,611,209,225
IssuesEvent
2021-06-02 11:45:42
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[Perf][Mono] Regressions in System.Drawing.Tests.Perf_Color
area-Codegen-Interpreter-mono runtime-mono tenet-performance untriaged
It looks like this regression was caused by https://github.com/dotnet/runtime/commit/89498d8f6095511d1553fabe39fa8f0579b1cf2d @L2 Is there anything to do about this? ### Run Information Architecture | x64 -- | -- OS | ubuntu 18.04 Baseline | [a3e0894f38b1ab558653e7c1b7322a8c4db92af1](https://github.com/dotnet/runtime/commit/a3e0894f38b1ab558653e7c1b7322a8c4db92af1) Compare | [6bfc5f21dea7b550f1c807454d45408ef34764e1](https://github.com/dotnet/runtime/commit/6bfc5f21dea7b550f1c807454d45408ef34764e1) Diff | [Diff](https://github.com/dotnet/runtime/compare/a3e0894f38b1ab558653e7c1b7322a8c4db92af1...6bfc5f21dea7b550f1c807454d45408ef34764e1) ### Regressions in System.Drawing.Tests.Perf_Color Benchmark | Baseline | Test | Test/Base | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL -- | -- | -- | -- | -- | -- | -- | -- | -- [GetHue](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetHue.html>) | 1.60 μs | 1.93 μs | 1.20 | | | [GetBrightness](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetBrightness.html>) | 1.18 μs | 1.42 μs | 1.20 | | | ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_1.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_2.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_3.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_4.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_5.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_6.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_7.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_8.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_9.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_10.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_11.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_12.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_13.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_14.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_15.png>) [Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color.html>) ### Repro ```cmd git clone https://github.com/dotnet/performance.git python3 .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Drawing.Tests.Perf_Color*' ``` <details> ### Payloads [Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-4f289a83-011e-4af2-93d6-e54544451aaef8e9425d5ed45ccaa/6d6b6cd4-7320-4acb-8ebe-0f24ed4209e4.zip?sv=2019-07-07&se=2021-05-05T19%3A40%3A38Z&sr=c&sp=rl&sig=jpXdspM56nw22cwLixbAlaT6r3ta5%2BXlPj373BUFgwc%3D>) [Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-7428b546-ea89-43f0-9d23-b45da4cd3a0731d56e954b543338d/57a3901c-6af6-49e3-aa41-64712f588ca5.zip?sv=2019-07-07&se=2021-05-06T16%3A34%3A52Z&sr=c&sp=rl&sig=5LEDRnUTyNz%2FIubzfnjqjefoZlGJSeiY5c10wJtIwj0%3D>) ### Histogram #### System.Drawing.Tests.Perf_Color.GetHue ```log ``` #### System.Drawing.Tests.Perf_Color.GetBrightness ```log ``` ### Docs [Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md) [Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md) </details>
1.0
[Perf][Mono] Regressions in System.Drawing.Tests.Perf_Color - It looks like this regression was caused by https://github.com/dotnet/runtime/commit/89498d8f6095511d1553fabe39fa8f0579b1cf2d @L2 Is there anything to do about this? ### Run Information Architecture | x64 -- | -- OS | ubuntu 18.04 Baseline | [a3e0894f38b1ab558653e7c1b7322a8c4db92af1](https://github.com/dotnet/runtime/commit/a3e0894f38b1ab558653e7c1b7322a8c4db92af1) Compare | [6bfc5f21dea7b550f1c807454d45408ef34764e1](https://github.com/dotnet/runtime/commit/6bfc5f21dea7b550f1c807454d45408ef34764e1) Diff | [Diff](https://github.com/dotnet/runtime/compare/a3e0894f38b1ab558653e7c1b7322a8c4db92af1...6bfc5f21dea7b550f1c807454d45408ef34764e1) ### Regressions in System.Drawing.Tests.Perf_Color Benchmark | Baseline | Test | Test/Base | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL -- | -- | -- | -- | -- | -- | -- | -- | -- [GetHue](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetHue.html>) | 1.60 μs | 1.93 μs | 1.20 | | | [GetBrightness](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetBrightness.html>) | 1.18 μs | 1.42 μs | 1.20 | | | ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_1.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_2.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_3.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_4.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_5.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_6.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_7.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_8.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_9.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_10.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_11.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_12.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_13.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_14.png>) ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color_15.png>) [Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color.html>) ### Repro ```cmd git clone https://github.com/dotnet/performance.git python3 .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Drawing.Tests.Perf_Color*' ``` <details> ### Payloads [Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-4f289a83-011e-4af2-93d6-e54544451aaef8e9425d5ed45ccaa/6d6b6cd4-7320-4acb-8ebe-0f24ed4209e4.zip?sv=2019-07-07&se=2021-05-05T19%3A40%3A38Z&sr=c&sp=rl&sig=jpXdspM56nw22cwLixbAlaT6r3ta5%2BXlPj373BUFgwc%3D>) [Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-7428b546-ea89-43f0-9d23-b45da4cd3a0731d56e954b543338d/57a3901c-6af6-49e3-aa41-64712f588ca5.zip?sv=2019-07-07&se=2021-05-06T16%3A34%3A52Z&sr=c&sp=rl&sig=5LEDRnUTyNz%2FIubzfnjqjefoZlGJSeiY5c10wJtIwj0%3D>) ### Histogram #### System.Drawing.Tests.Perf_Color.GetHue ```log ``` #### System.Drawing.Tests.Perf_Color.GetBrightness ```log ``` ### Docs [Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md) [Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md) </details>
non_priority
regressions in system drawing tests perf color it looks like this regression was caused by is there anything to do about this run information architecture os ubuntu baseline compare diff regressions in system drawing tests perf color benchmark baseline test test base baseline ir compare ir ir ratio baseline etl compare etl μs μs μs μs repro cmd git clone performance scripts benchmarks ci py f filter system drawing tests perf color payloads histogram system drawing tests perf color gethue log system drawing tests perf color getbrightness log docs
0
57,112
7,033,879,080
IssuesEvent
2017-12-27 13:36:19
cpprefjp/kunai
https://api.github.com/repos/cpprefjp/kunai
closed
コードサンプル実行: Try Code ボタンが別のコードブロックに表示される
bug design
以下のページで起こっている問題です。Chrome 63.0.3239.84 / Firefox 57.0.2 (on Win 10) で確認しました。 * http://cpprefjp.github.io/lang/cpp11/char16_32.html * http://cpprefjp.github.io/lang/cpp11/utf8_string_literals.html リスト内に ```` ```cpp ~``` ```` があるとき、 1. Try Code ボタンが表示されるコードブロックが一斉にずれています。 2. また Try Code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。 (ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません)
1.0
コードサンプル実行: Try Code ボタンが別のコードブロックに表示される - 以下のページで起こっている問題です。Chrome 63.0.3239.84 / Firefox 57.0.2 (on Win 10) で確認しました。 * http://cpprefjp.github.io/lang/cpp11/char16_32.html * http://cpprefjp.github.io/lang/cpp11/utf8_string_literals.html リスト内に ```` ```cpp ~``` ```` があるとき、 1. Try Code ボタンが表示されるコードブロックが一斉にずれています。 2. また Try Code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。 (ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません)
non_priority
コードサンプル実行 try code ボタンが別のコードブロックに表示される 以下のページで起こっている問題です。chrome firefox on win で確認しました。 リスト内に cpp ~ があるとき、 try code ボタンが表示されるコードブロックが一斉にずれています。 また try code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。 ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません
0