Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,763 | 2,518,902,252 | IssuesEvent | 2015-01-17 06:39:52 | rbsl-gaurav/rbframework | https://api.github.com/repos/rbsl-gaurav/rbframework | opened | Properly Alignment formal parameter of Controller's and view's methods | Priority-Normal task | <a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)**
_Monday Jun 23, 2014 at 10:57 GMT_
_Originally opened as https://github.com/readybytes/rbframework/issues/42_
----
| 1.0 | Properly Alignment formal parameter of Controller's and view's methods - <a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)**
_Monday Jun 23, 2014 at 10:57 GMT_
_Originally opened as https://github.com/readybytes/rbframework/issues/42_
----
| priority | properly alignment formal parameter of controller s and view s methods issue by monday jun at gmt originally opened as | 1 |
118,614 | 25,338,664,701 | IssuesEvent | 2022-11-18 19:14:33 | raupargor/Friendsn-t-Games | https://api.github.com/repos/raupargor/Friendsn-t-Games | closed | 4.1.Objetos y armas: Sprites | code Priority: medium | Desarrollar animaciones para los objetos y armas a través de sprites | 1.0 | 4.1.Objetos y armas: Sprites - Desarrollar animaciones para los objetos y armas a través de sprites | non_priority | objetos y armas sprites desarrollar animaciones para los objetos y armas a través de sprites | 0 |
370,976 | 10,959,523,608 | IssuesEvent | 2019-11-27 11:35:48 | highcharts/highcharts | https://api.github.com/repos/highcharts/highcharts | closed | venn.src.js:723 Uncaught Error: f(a) and f(b) must have opposite signs. | Priority: High Product: Highcharts Status: Review Type: Regression | #### Expected behaviour
Venn diagram is displayed
#### Actual behaviour
Venn diagtam is not displayed + exception in console:
Uncaught Error: f(a) and f(b) must have opposite signs.
at L (venn.src.js:723)
at f (venn.src.js:906)
at M (venn.src.js:923)
at venn.src.js:961
at Array.reduce (<anonymous>)
at X (venn.src.js:937)
at d.translate (venn.src.js:1429)
at highcharts.src.js:24833
at Array.forEach (<anonymous>)
at c.Chart.renderSeries (highcharts.src.js:24832)
#### Live demo with steps to reproduce
https://jsfiddle.net/smLn35cz/
Chart options:
{"series":[
{"type":"venn",
"data":[{"sets":["A"],"value":269},
{"sets":["B"],"value":4350},
{"sets":["C"],"value":2330},
{"sets":["A","B"],"value":131},
{"sets":["A","C"],"value":9},
{"sets":["B","C"],"value":12},
{"sets":["A","B","C"],"value":8}]
}]
}
#### Product version
Highcharts 7.2.0
#### Affected browser(s)
Chrome
| 1.0 | venn.src.js:723 Uncaught Error: f(a) and f(b) must have opposite signs. - #### Expected behaviour
Venn diagram is displayed
#### Actual behaviour
Venn diagtam is not displayed + exception in console:
Uncaught Error: f(a) and f(b) must have opposite signs.
at L (venn.src.js:723)
at f (venn.src.js:906)
at M (venn.src.js:923)
at venn.src.js:961
at Array.reduce (<anonymous>)
at X (venn.src.js:937)
at d.translate (venn.src.js:1429)
at highcharts.src.js:24833
at Array.forEach (<anonymous>)
at c.Chart.renderSeries (highcharts.src.js:24832)
#### Live demo with steps to reproduce
https://jsfiddle.net/smLn35cz/
Chart options:
{"series":[
{"type":"venn",
"data":[{"sets":["A"],"value":269},
{"sets":["B"],"value":4350},
{"sets":["C"],"value":2330},
{"sets":["A","B"],"value":131},
{"sets":["A","C"],"value":9},
{"sets":["B","C"],"value":12},
{"sets":["A","B","C"],"value":8}]
}]
}
#### Product version
Highcharts 7.2.0
#### Affected browser(s)
Chrome
| priority | venn src js uncaught error f a and f b must have opposite signs expected behaviour venn diagram is displayed actual behaviour venn diagtam is not displayed exception in console uncaught error f a and f b must have opposite signs at l venn src js at f venn src js at m venn src js at venn src js at array reduce at x venn src js at d translate venn src js at highcharts src js at array foreach at c chart renderseries highcharts src js live demo with steps to reproduce chart options series type venn data value sets value sets value sets value sets value sets value sets value product version highcharts affected browser s chrome | 1 |
666,714 | 22,364,939,419 | IssuesEvent | 2022-06-16 02:18:07 | apache/incubator-kyuubi | https://api.github.com/repos/apache/incubator-kyuubi | opened | [Bug] kyuubi loads the Listener to start an error | kind:bug priority:major | ### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
### Search before asking
- [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues.
### Describe the bug
kyuubi loads the Listener to start an error. When I use yarn client mode, it is normal, but cluster mode fails to start.
### Affects Version(s)
1.5.1
### Kyuubi Server Log Output
```logtalk
09:20:04.816 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server HZ3-BD-2204-V2809.lianlianpay-dc.com/10.90.37.115:2181, sessionid = 0x100faf3ff4b0001, negotiated timeout = 60000
09:20:04.817 INFO org.apache.zookeeper.server.ZooKeeperServer: Established session 0x100faf3ff4b0001 with negotiated timeout 60000 for client /10.90.37.115:37874
09:20:04.818 INFO org.apache.curator.framework.state.ConnectionStateManager: State change: CONNECTED
09:20:04.845 INFO org.apache.kyuubi.engine.EngineRef: Launching engine:
/home/dubbo/spark-3.2.1-bin-hadoop2.7/bin/spark-submit \
--class org.apache.kyuubi.engine.spark.SparkSQLEngine \
--conf spark.kyuubi.session.engine.idle.timeout=PT30M \
--conf spark.kyuubi.session.engine.check.interval=PT5M \
--conf spark.sql.hive.convertMetastoreParquet=false \
--conf spark.hive.server2.thrift.resultset.default.fetch.size=1000 \
--conf spark.kyuubi.ha.zookeeper.quorum=HZ3-BD-2204-V280:2181 \
--conf spark.kyuubi.sparklens.jars.location=/home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/lens_2.12-1.0.2-SNAPSHOT.jar \
--conf spark.extraListeners=com.lly.lens.JobListener \
--conf spark.kyuubi.client.ip=10.90.37.115 \
--conf spark.yarn.queue=llyt \
--conf spark.kyuubi.engine.submit.time=1655342404826 \
--conf spark.app.name=dwd_evt_payout_deposit_di \
--conf spark.executorEnv.HADOOP_USER_NAME=hdfs \
--conf spark.driver.memory=4G \
--conf spark.executor.instances=20 \
--conf spark.kyuubi.ha.engine.ref.id=964041b0-08c9-4663-9d02-ed5e4001bf54 \
--conf spark.executorEnv.SPARK_USER=hdfs \
--conf spark.kyuubi.session.conf.advisor=org.apache.kyuubi.plugin.HSessionConfAdvisor \
--conf spark.driver.cores=2 \
--conf spark.kyuubi.ha.zookeeper.auth.type=NONE \
--conf spark.submit.deployMode=cluster \
--conf spark.master=yarn \
--conf spark.yarn.tags=KYUUBI \
--conf spark.kyuubi.engine.share.level=CONNECTION \
--conf spark.kyuubi.ha.zookeeper.namespace=/kyuubi_1.5.1-SNAPSHOT_CONNECTION_SPARK_SQL/hdfs/964041b0-08c9-4663-9d02-ed5e4001bf54 \
--conf spark.executor.memory=15G \
--conf spark.executor.cores=4 \
--conf spark.yarn.executor.memoryOverhead=6G \
--conf spark.memory.fraction=0.8 \
--conf spark.kyuubi.engine.type=SPARK_SQL \
--conf spark.executor.memory=12 \
--conf spark.sql.shuffle.partitions=300 \
--jars /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/sparklens_2.12-1.0.2-SNAPSHOT.jar \
--proxy-user hdfs /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/externals/engines/spark/kyuubi-spark-sql-engine_2.12-1.5.1-SNAPSHOT.jar
09:20:04.857 INFO org.apache.kyuubi.engine.ProcBuilder: Logging to /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/work/hdfs/kyuubi-spark-sql-engine.log.2
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.646 INFO org.apache.zookeeper.server.PrepRequestProcessor: Processed session termination for sessionid: 0x100faf3ff4b0001
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001
09:20:24.649 INFO org.apache.zookeeper.server.NIOServerCnxn: Closed socket connection for client /10.90.37.115:37874 which had sessionid 0x100faf3ff4b0001
09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds
09:20:24.677 INFO org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Received request of closing SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64]
09:20:24.679 INFO org.apache.kyuubi.session.KyuubiSessionManager: SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] is closed, current opening sessions 0
09:20:24.683 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: ERROR_STATE -> CLOSE
```
### Kyuubi Engine Log Output
```logtalk
22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0
22/06/16 09:20:15 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:16 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:17 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001
09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds
22/06/16 09:20:20 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:21 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:22 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:23 INFO Client: Application report for application_1654630917149_16762 (state: FAILED)
22/06/16 09:20:23 INFO Client:
client token: N/A
diagnostics: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch.
Container id: container_e82_1654630917149_16762_02_000001
Exit code: 13
[2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
[2022-06-16 09:20:23.055]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
For more detailed output, check the application tracking page: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 Then click on links to logs of each attempt.
. Failing the application.
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: ll
start time: 1655342409537
final status: FAILED
tracking URL: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762
user: hdfs
22/06/16 09:20:23 INFO Client: Deleted staging directory hdfs://hz3bdcrossborder01/user/hdfs/.sparkStaging/application_1654630917149_16762
22/06/16 09:20:23 ERROR Client: Application diagnostics message: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch.
Container id: container_e82_1654630917149_16762_02_000001
Exit code: 13
[2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
```
### Kyuubi Server Configurations
_No response_
### Kyuubi Engine Configurations
_No response_
### Additional context
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR! | 1.0 | [Bug] kyuubi loads the Listener to start an error - ### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
### Search before asking
- [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues.
### Describe the bug
kyuubi loads the Listener to start an error. When I use yarn client mode, it is normal, but cluster mode fails to start.
### Affects Version(s)
1.5.1
### Kyuubi Server Log Output
```logtalk
09:20:04.816 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server HZ3-BD-2204-V2809.lianlianpay-dc.com/10.90.37.115:2181, sessionid = 0x100faf3ff4b0001, negotiated timeout = 60000
09:20:04.817 INFO org.apache.zookeeper.server.ZooKeeperServer: Established session 0x100faf3ff4b0001 with negotiated timeout 60000 for client /10.90.37.115:37874
09:20:04.818 INFO org.apache.curator.framework.state.ConnectionStateManager: State change: CONNECTED
09:20:04.845 INFO org.apache.kyuubi.engine.EngineRef: Launching engine:
/home/dubbo/spark-3.2.1-bin-hadoop2.7/bin/spark-submit \
--class org.apache.kyuubi.engine.spark.SparkSQLEngine \
--conf spark.kyuubi.session.engine.idle.timeout=PT30M \
--conf spark.kyuubi.session.engine.check.interval=PT5M \
--conf spark.sql.hive.convertMetastoreParquet=false \
--conf spark.hive.server2.thrift.resultset.default.fetch.size=1000 \
--conf spark.kyuubi.ha.zookeeper.quorum=HZ3-BD-2204-V280:2181 \
--conf spark.kyuubi.sparklens.jars.location=/home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/lens_2.12-1.0.2-SNAPSHOT.jar \
--conf spark.extraListeners=com.lly.lens.JobListener \
--conf spark.kyuubi.client.ip=10.90.37.115 \
--conf spark.yarn.queue=llyt \
--conf spark.kyuubi.engine.submit.time=1655342404826 \
--conf spark.app.name=dwd_evt_payout_deposit_di \
--conf spark.executorEnv.HADOOP_USER_NAME=hdfs \
--conf spark.driver.memory=4G \
--conf spark.executor.instances=20 \
--conf spark.kyuubi.ha.engine.ref.id=964041b0-08c9-4663-9d02-ed5e4001bf54 \
--conf spark.executorEnv.SPARK_USER=hdfs \
--conf spark.kyuubi.session.conf.advisor=org.apache.kyuubi.plugin.HSessionConfAdvisor \
--conf spark.driver.cores=2 \
--conf spark.kyuubi.ha.zookeeper.auth.type=NONE \
--conf spark.submit.deployMode=cluster \
--conf spark.master=yarn \
--conf spark.yarn.tags=KYUUBI \
--conf spark.kyuubi.engine.share.level=CONNECTION \
--conf spark.kyuubi.ha.zookeeper.namespace=/kyuubi_1.5.1-SNAPSHOT_CONNECTION_SPARK_SQL/hdfs/964041b0-08c9-4663-9d02-ed5e4001bf54 \
--conf spark.executor.memory=15G \
--conf spark.executor.cores=4 \
--conf spark.yarn.executor.memoryOverhead=6G \
--conf spark.memory.fraction=0.8 \
--conf spark.kyuubi.engine.type=SPARK_SQL \
--conf spark.executor.memory=12 \
--conf spark.sql.shuffle.partitions=300 \
--jars /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/external_jars/sparklens_2.12-1.0.2-SNAPSHOT.jar \
--proxy-user hdfs /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/externals/engines/spark/kyuubi-spark-sql-engine_2.12-1.5.1-SNAPSHOT.jar
09:20:04.857 INFO org.apache.kyuubi.engine.ProcBuilder: Logging to /home/dubbo/apache-kyuubi-1.5.1-incubating-bin/work/hdfs/kyuubi-spark-sql-engine.log.2
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.646 INFO org.apache.zookeeper.server.PrepRequestProcessor: Processed session termination for sessionid: 0x100faf3ff4b0001
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001
09:20:24.649 INFO org.apache.zookeeper.server.NIOServerCnxn: Closed socket connection for client /10.90.37.115:37874 which had sessionid 0x100faf3ff4b0001
09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds
09:20:24.677 INFO org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Received request of closing SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64]
09:20:24.679 INFO org.apache.kyuubi.session.KyuubiSessionManager: SessionHandle [f07458c0-c2f0-4af8-89db-0c6d3e321e64] is closed, current opening sessions 0
09:20:24.683 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: ERROR_STATE -> CLOSE
```
### Kyuubi Engine Log Output
```logtalk
22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0
22/06/16 09:20:15 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:16 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:17 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:18 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:19 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
09:20:24.644 INFO org.apache.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
09:20:24.648 INFO org.apache.zookeeper.ZooKeeper: Session: 0x100faf3ff4b0001 closed
09:20:24.648 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down for session: 0x100faf3ff4b0001
09:20:24.654 INFO org.apache.kyuubi.operation.LaunchEngine: Processing hdfs's query[71bbed2b-6424-4db3-b771-a2eaf8012f8c]: RUNNING_STATE -> ERROR_STATE, statement: LAUNCH_ENGINE, time taken: 19.854 seconds
22/06/16 09:20:20 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:21 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:22 INFO Client: Application report for application_1654630917149_16762 (state: ACCEPTED)
22/06/16 09:20:23 INFO Client: Application report for application_1654630917149_16762 (state: FAILED)
22/06/16 09:20:23 INFO Client:
client token: N/A
diagnostics: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch.
Container id: container_e82_1654630917149_16762_02_000001
Exit code: 13
[2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
[2022-06-16 09:20:23.055]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
For more detailed output, check the application tracking page: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762 Then click on links to logs of each attempt.
. Failing the application.
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: ll
start time: 1655342409537
final status: FAILED
tracking URL: http://HZ3-BD-2004-P1392:8088/cluster/app/application_1654630917149_16762
user: hdfs
22/06/16 09:20:23 INFO Client: Deleted staging directory hdfs://hz3bdcrossborder01/user/hdfs/.sparkStaging/application_1654630917149_16762
22/06/16 09:20:23 ERROR Client: Application diagnostics message: Application application_1654630917149_16762 failed 2 times due to AM Container for appattempt_1654630917149_16762_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: [2022-06-16 09:20:23.053]Exception from container-launch.
Container id: container_e82_1654630917149_16762_02_000001
Exit code: 13
[2022-06-16 09:20:23.054]Container exited with a non-zero exit code 13. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
e.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
at org.apache.hadoop.tracing.SpanReceiverHost.get(SpanReceiverHost.java:79)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:634)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
at org.apache.spark.deploy.yarn.ApplicationMaster.cleanupStagingDir(ApplicationMaster.scala:683)
at org.apache.spark.deploy.yarn.ApplicationMaster.$anonfun$run$2(ApplicationMaster.scala:267)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)
at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
22/06/16 09:20:22 INFO ShutdownHookManager: Shutdown hook called
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data8/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-310f70f7-060b-4c37-9b40-29c68f8737aa
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data2/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a2725161-0f29-470d-95d4-ad7762057e4b
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data3/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-49c7dee2-2a77-49d0-b307-73a72cf4e5dc
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data1/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-0dd13a24-0c92-4ec6-a5da-e21de71588e3
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data5/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-51ee1773-aa91-4f56-b43a-697a6a1e02a0
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data6/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-28474654-2043-4f6e-8f4d-1e5481ae9b05
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data7/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-771f67ef-3ac4-41fe-9cef-0daffc55865a
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data11/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-55c32301-2f79-4138-bb0a-1ad037b26df5
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data10/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-b4b76b3a-1207-4f00-a12b-3a569c8e500d
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data12/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-aab310b3-23d5-4ee9-a60b-b95e39313d77
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data9/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-a50fa878-0fe5-48df-9e18-d710167a2cac
22/06/16 09:20:22 INFO ShutdownHookManager: Deleting directory /data4/hadoop/yarn/local/usercache/hdfs/appcache/application_1654630917149_16762/spark-505d373b-94f1-4292-8b01-dbbaf301eecc
```
### Kyuubi Server Configurations
_No response_
### Kyuubi Engine Configurations
_No response_
### Additional context
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR! | priority | kyuubi loads the listener to start an error code of conduct i agree to follow this project s search before asking i have searched in the and found no similar issues describe the bug kyuubi loads the listener to start an error when i use yarn client mode it is normal but cluster mode fails to start affects version s kyuubi server log output logtalk info org apache zookeeper clientcnxn session establishment complete on server bd lianlianpay dc com sessionid negotiated timeout info org apache zookeeper server zookeeperserver established session with negotiated timeout for client info org apache curator framework state connectionstatemanager state change connected info org apache kyuubi engine engineref launching engine home dubbo spark bin bin spark submit class org apache kyuubi engine spark sparksqlengine conf spark kyuubi session engine idle timeout conf spark kyuubi session engine check interval conf spark sql hive convertmetastoreparquet false conf spark hive thrift resultset default fetch size conf spark kyuubi ha zookeeper quorum bd conf spark kyuubi sparklens jars location home dubbo apache kyuubi incubating bin external jars lens snapshot jar conf spark extralisteners com lly lens joblistener conf spark kyuubi client ip conf spark yarn queue llyt conf spark kyuubi engine submit time conf spark app name dwd evt payout deposit di conf spark executorenv hadoop user name hdfs conf spark driver memory conf spark executor instances conf spark kyuubi ha engine ref id conf spark executorenv spark user hdfs conf spark kyuubi session conf advisor org apache kyuubi plugin hsessionconfadvisor conf spark driver cores conf spark kyuubi ha zookeeper auth type none conf spark submit deploymode cluster conf spark master yarn conf spark yarn tags kyuubi conf spark kyuubi engine share level connection conf spark kyuubi ha zookeeper namespace kyuubi snapshot connection spark sql hdfs conf spark executor memory conf spark executor cores conf spark yarn executor memoryoverhead conf spark memory fraction conf spark kyuubi engine type spark sql conf spark executor memory conf spark sql shuffle partitions jars home dubbo apache kyuubi incubating bin external jars sparklens snapshot jar proxy user hdfs home dubbo apache kyuubi incubating bin externals engines spark kyuubi spark sql engine snapshot jar info org apache kyuubi engine procbuilder logging to home dubbo apache kyuubi incubating bin work hdfs kyuubi spark sql engine log info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper server preprequestprocessor processed session termination for sessionid info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info org apache zookeeper server nioservercnxn closed socket connection for client which had sessionid info org apache kyuubi operation launchengine processing hdfs s query running state error state statement launch engine time taken seconds info org apache kyuubi server kyuubitbinaryfrontendservice received request of closing sessionhandle info org apache kyuubi session kyuubisessionmanager sessionhandle is closed current opening sessions info org apache kyuubi operation launchengine processing hdfs s query error state close kyuubi engine log output logtalk info client application report for application state accepted info client application report for application state accepted info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info org apache curator framework imps curatorframeworkimpl backgroundoperationsloop exiting info org apache zookeeper zookeeper session closed info org apache zookeeper clientcnxn eventthread shut down for session info org apache kyuubi operation launchengine processing hdfs s query running state error state statement launch engine time taken seconds info client application report for application state accepted info client application report for application state accepted info client application report for application state accepted info client application report for application state failed info client client token n a diagnostics application application failed times due to am container for appattempt exited with exitcode failing this attempt diagnostics exception from container launch container id container exit code container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark for more detailed output check the application tracking page then click on links to logs of each attempt failing the application applicationmaster host n a applicationmaster rpc port queue ll start time final status failed tracking url user hdfs info client deleted staging directory hdfs user hdfs sparkstaging application error client application diagnostics message application application failed times due to am container for appattempt exited with exitcode failing this attempt diagnostics exception from container launch container id container exit code container exited with a non zero exit code error file prelaunch err last bytes of prelaunch err last bytes of stderr e hadoop util shutdownhookmanager addshutdownhook shutdownhookmanager java at org apache hadoop tracing spanreceiverhost get spanreceiverhost java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs dfsclient dfsclient java at org apache hadoop hdfs distributedfilesystem initialize distributedfilesystem java at org apache hadoop fs filesystem createfilesystem filesystem java at org apache hadoop fs filesystem access filesystem java at org apache hadoop fs filesystem cache getinternal filesystem java at org apache hadoop fs filesystem cache get filesystem java at org apache hadoop fs filesystem get filesystem java at org apache hadoop fs path getfilesystem path java at org apache spark deploy yarn applicationmaster cleanupstagingdir applicationmaster scala at org apache spark deploy yarn applicationmaster anonfun run applicationmaster scala at org apache spark util sparkshutdownhook run shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at org apache spark util utils loguncaughtexceptions utils scala at org apache spark util sparkshutdownhookmanager anonfun runall shutdownhookmanager scala at scala runtime mcv sp apply mcv sp java at scala util try apply try scala at org apache spark util sparkshutdownhookmanager runall shutdownhookmanager scala at org apache spark util sparkshutdownhookmanager anon run shutdownhookmanager scala at org apache hadoop util shutdownhookmanager run shutdownhookmanager java info shutdownhookmanager shutdown hook called info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark info shutdownhookmanager deleting directory hadoop yarn local usercache hdfs appcache application spark kyuubi server configurations no response kyuubi engine configurations no response additional context no response are you willing to submit pr yes i am willing to submit a pr | 1 |
685,492 | 23,458,304,277 | IssuesEvent | 2022-08-16 10:53:08 | nanoframework/Home | https://api.github.com/repos/nanoframework/Home | closed | VS2022 extension broken on 17.3 | Type: Bug Status: In progress Area: Visual Studio extension Priority: Critical | ### Tool
Visual Studio extension
### Description
When opening a nfproj in VS2022 17.3 the extension fails.
### How to reproduce
1. Open a nfproj
2. VS output
LimitedFunctionality
System.ArgumentNullException: Value cannot be null.
Parameter name: commandLine
at Microsoft.VisualStudio.LanguageServices.Implementation.ProjectSystem.VisualStudioProjectOptionsProcessor.SetCommandLine(String commandLine)
at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.VisualStudio.ProjectSystem.CommonProjectSystemTools.Rethrow(Exception ex)
at Microsoft.VisualStudio.ProjectSystem.ExceptionFilter.<>cDisplayClass2_0.<Guard>gAction|0()
at GuardMethodClass.GuardMethod(Func1 , Func2 , Func`2 )
### Expected behaviour
_No response_
### Screenshots
_No response_
### Aditional context
_No response_ | 1.0 | VS2022 extension broken on 17.3 - ### Tool
Visual Studio extension
### Description
When opening a nfproj in VS2022 17.3 the extension fails.
### How to reproduce
1. Open a nfproj
2. VS output
LimitedFunctionality
System.ArgumentNullException: Value cannot be null.
Parameter name: commandLine
at Microsoft.VisualStudio.LanguageServices.Implementation.ProjectSystem.VisualStudioProjectOptionsProcessor.SetCommandLine(String commandLine)
at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.VisualStudio.ProjectSystem.LanguageServices.WorkspaceProjectContextProvider.<CreateProjectContextHandlingFaultAsync>d7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at Microsoft.VisualStudio.ProjectSystem.CommonProjectSystemTools.Rethrow(Exception ex)
at Microsoft.VisualStudio.ProjectSystem.ExceptionFilter.<>cDisplayClass2_0.<Guard>gAction|0()
at GuardMethodClass.GuardMethod(Func1 , Func2 , Func`2 )
### Expected behaviour
_No response_
### Screenshots
_No response_
### Aditional context
_No response_ | priority | extension broken on tool visual studio extension description when opening a nfproj in the extension fails how to reproduce open a nfproj vs output limitedfunctionality system argumentnullexception value cannot be null parameter name commandline at microsoft visualstudio languageservices implementation projectsystem visualstudioprojectoptionsprocessor setcommandline string commandline at microsoft visualstudio projectsystem languageservices workspaceprojectcontextprovider movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft visualstudio projectsystem languageservices workspaceprojectcontextprovider movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at microsoft visualstudio projectsystem commonprojectsystemtools rethrow exception ex at microsoft visualstudio projectsystem exceptionfilter gaction at guardmethodclass guardmethod func expected behaviour no response screenshots no response aditional context no response | 1 |
411,632 | 12,026,896,010 | IssuesEvent | 2020-04-12 16:03:22 | ooni/probe | https://api.github.com/repos/ooni/probe | closed | Feature request: Run All Tests button | effort/L ooni/probe-mobile priority/medium ux | We have received community requests to include a "Run All Tests" button in the revamped app (on both Android and iOS), to run all OONI Probe tests in sequence, by tapping one button (instead of having to manually run each test individually). | 1.0 | Feature request: Run All Tests button - We have received community requests to include a "Run All Tests" button in the revamped app (on both Android and iOS), to run all OONI Probe tests in sequence, by tapping one button (instead of having to manually run each test individually). | priority | feature request run all tests button we have received community requests to include a run all tests button in the revamped app on both android and ios to run all ooni probe tests in sequence by tapping one button instead of having to manually run each test individually | 1 |
182,612 | 30,873,966,286 | IssuesEvent | 2023-08-03 13:12:37 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | When provisioning an RKE1 custom cluster, provide a visual indication that a user has clicked the clipboard icon | [zube]: Done status/wontfix kind/design QA/XS kind/enhancement team/area2 ember | <!--------- For bugs and general issues --------->
**Setup**
- Rancher version: `v2.6-head` https://github.com/rancher/rancher/commit/f4d09c73e1a609f5c5fe0848ebd6e4b059056264
Browser: Chrome
- Browser type & version: Chrome
**Describe the bug**
<!--A clear and concise description of what the bug is.-->
Not necessarily a bug, more of an enhancement. When provisioning a downstream RKE1 custom cluster, we have the following screen:

When you click the clipboard icon to copy the registration command, there is no illustration that it has been copied. Throughout the UI, we indicate a green checkmark once a user clicks the clipboard icon. We should follow the same practice here to add overall cohesiveness.
**To Reproduce**
<!--Steps to reproduce the behavior-->
1. Setup Rancher and navigate to the UI in a browser.
2. Provision a downstream RKE1 custom cluster.
3. Click the clipboard icon to copy the registration command. Note that you will have no visual indication that it has been copied to your clipboard.
**Result**
No visual indication that I have successfully copied the registration command to my clipboard.
**Expected Result**
<!--A clear and concise description of what you expected to happen.-->
When I click the clipboard, I should see a green checkmark or a message nothing that the registration command has been copied to my clipboard. | 1.0 | When provisioning an RKE1 custom cluster, provide a visual indication that a user has clicked the clipboard icon - <!--------- For bugs and general issues --------->
**Setup**
- Rancher version: `v2.6-head` https://github.com/rancher/rancher/commit/f4d09c73e1a609f5c5fe0848ebd6e4b059056264
Browser: Chrome
- Browser type & version: Chrome
**Describe the bug**
<!--A clear and concise description of what the bug is.-->
Not necessarily a bug, more of an enhancement. When provisioning a downstream RKE1 custom cluster, we have the following screen:

When you click the clipboard icon to copy the registration command, there is no illustration that it has been copied. Throughout the UI, we indicate a green checkmark once a user clicks the clipboard icon. We should follow the same practice here to add overall cohesiveness.
**To Reproduce**
<!--Steps to reproduce the behavior-->
1. Setup Rancher and navigate to the UI in a browser.
2. Provision a downstream RKE1 custom cluster.
3. Click the clipboard icon to copy the registration command. Note that you will have no visual indication that it has been copied to your clipboard.
**Result**
No visual indication that I have successfully copied the registration command to my clipboard.
**Expected Result**
<!--A clear and concise description of what you expected to happen.-->
When I click the clipboard, I should see a green checkmark or a message nothing that the registration command has been copied to my clipboard. | non_priority | when provisioning an custom cluster provide a visual indication that a user has clicked the clipboard icon setup rancher version head browser chrome browser type version chrome describe the bug not necessarily a bug more of an enhancement when provisioning a downstream custom cluster we have the following screen when you click the clipboard icon to copy the registration command there is no illustration that it has been copied throughout the ui we indicate a green checkmark once a user clicks the clipboard icon we should follow the same practice here to add overall cohesiveness to reproduce setup rancher and navigate to the ui in a browser provision a downstream custom cluster click the clipboard icon to copy the registration command note that you will have no visual indication that it has been copied to your clipboard result no visual indication that i have successfully copied the registration command to my clipboard expected result when i click the clipboard i should see a green checkmark or a message nothing that the registration command has been copied to my clipboard | 0 |
539,908 | 15,796,628,881 | IssuesEvent | 2021-04-02 15:15:04 | refgenie/refgenie | https://api.github.com/repos/refgenie/refgenie | opened | auto documenting buildable assets | docs enhancement priority-low | It would be nice to generate some sort of [this page](http://refgenie.databio.org/en/latest/available_assets/) automatically, based on the set of available recipes. This will come in handy once the recipes are no longer a part of refgenie package, but are served with refgenieserver.
It should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in [this issue](https://github.com/refgenie/refgenie/issues/198#issue-683071342) are defined. But we could also provide means to add more detailed documentation for each recipe. For example, `docs` key that is a freeform text, later interpreted as markdown, or something. | 1.0 | auto documenting buildable assets - It would be nice to generate some sort of [this page](http://refgenie.databio.org/en/latest/available_assets/) automatically, based on the set of available recipes. This will come in handy once the recipes are no longer a part of refgenie package, but are served with refgenieserver.
It should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in [this issue](https://github.com/refgenie/refgenie/issues/198#issue-683071342) are defined. But we could also provide means to add more detailed documentation for each recipe. For example, `docs` key that is a freeform text, later interpreted as markdown, or something. | priority | auto documenting buildable assets it would be nice to generate some sort of automatically based on the set of available recipes this will come in handy once the recipes are no longer a part of refgenie package but are served with refgenieserver it should be relatively easy to produce at least a minimal markdown entry for each asset in case if only the keys suggested in are defined but we could also provide means to add more detailed documentation for each recipe for example docs key that is a freeform text later interpreted as markdown or something | 1 |
372,987 | 26,030,482,122 | IssuesEvent | 2022-12-21 20:39:57 | solidjs/solid-start | https://api.github.com/repos/solidjs/solid-start | closed | Linking to pdf files inside `public` folder doesn't work on routing. | documentation | I have a repo that builds on top of solid-start here: https://github.com/nikitavoloboev/nikitavoloboev
I have a pdf file in public called `cv.pdf`: https://github.com/nikitavoloboev/nikitavoloboev/blob/main/public/cv.pdf
If I for example do this:
https://github.com/nikitavoloboev/nikitavoloboev/blob/a0ab2a594e77e0cbcc450de56099a15aa4691c06/src/components/Nav.tsx#L37
```
<a class="nav-link" href="./cv.pdf">
CV
</a>
```
Inside the app if I go to this page, it will 404:
https://www.loom.com/share/900b34ae19904c59aed372d2a7457103
Not sure what I am doing wrong. I need to refresh the page for the PDF to actually show. | 1.0 | Linking to pdf files inside `public` folder doesn't work on routing. - I have a repo that builds on top of solid-start here: https://github.com/nikitavoloboev/nikitavoloboev
I have a pdf file in public called `cv.pdf`: https://github.com/nikitavoloboev/nikitavoloboev/blob/main/public/cv.pdf
If I for example do this:
https://github.com/nikitavoloboev/nikitavoloboev/blob/a0ab2a594e77e0cbcc450de56099a15aa4691c06/src/components/Nav.tsx#L37
```
<a class="nav-link" href="./cv.pdf">
CV
</a>
```
Inside the app if I go to this page, it will 404:
https://www.loom.com/share/900b34ae19904c59aed372d2a7457103
Not sure what I am doing wrong. I need to refresh the page for the PDF to actually show. | non_priority | linking to pdf files inside public folder doesn t work on routing i have a repo that builds on top of solid start here i have a pdf file in public called cv pdf if i for example do this cv inside the app if i go to this page it will not sure what i am doing wrong i need to refresh the page for the pdf to actually show | 0 |
277,160 | 8,621,305,231 | IssuesEvent | 2018-11-20 17:00:27 | supergiant/supergiant | https://api.github.com/repos/supergiant/supergiant | closed | 2.0: UI - Pagination needs to be removed from Apps store | Low Priority | 
---
### More Info
**Reported by:** eldridge henley (eldridge@qbox.io)
**Source URL**: [http://localhost:4200/apps/supergiant](http://localhost:4200/apps/supergiant)
<table><tr><td><strong>Browser</strong></td><td>Chrome 70.0.3538.77</td></tr><tr><td><strong>Screen Size</strong></td><td>1920 x 1080</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.14.0</td></tr><tr><td><strong>Viewport Size</strong></td><td>1610 x 1027</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr></table> | 1.0 | 2.0: UI - Pagination needs to be removed from Apps store - 
---
### More Info
**Reported by:** eldridge henley (eldridge@qbox.io)
**Source URL**: [http://localhost:4200/apps/supergiant](http://localhost:4200/apps/supergiant)
<table><tr><td><strong>Browser</strong></td><td>Chrome 70.0.3538.77</td></tr><tr><td><strong>Screen Size</strong></td><td>1920 x 1080</td></tr><tr><td><strong>OS</strong></td><td>OS X 10.14.0</td></tr><tr><td><strong>Viewport Size</strong></td><td>1610 x 1027</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr></table> | priority | ui pagination needs to be removed from apps store more info reported by eldridge henley eldridge qbox io source url browser chrome screen size x os os x viewport size x zoom level pixel ratio | 1 |
673,778 | 23,031,149,402 | IssuesEvent | 2022-07-22 14:02:19 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.apps.disneyplus.com - site is not usable | os-ios browser-firefox-ios priority-important device-tablet | <!-- @browser: Safari 13.1 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/107754 -->
<!-- @extra_labels: browser-firefox-ios, device-tablet -->
**URL**: https://www.apps.disneyplus.com/eg/onboarding
**Browser / Version**: Safari 13.1
**Operating System**: Mac OS X 10.15.4
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
None of the links are working, not even on my MacBook
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.apps.disneyplus.com - site is not usable - <!-- @browser: Safari 13.1 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/107754 -->
<!-- @extra_labels: browser-firefox-ios, device-tablet -->
**URL**: https://www.apps.disneyplus.com/eg/onboarding
**Browser / Version**: Safari 13.1
**Operating System**: Mac OS X 10.15.4
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
None of the links are working, not even on my MacBook
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version safari operating system mac os x tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce none of the links are working not even on my macbook browser configuration none from with ❤️ | 1 |
446,305 | 12,854,821,155 | IssuesEvent | 2020-07-09 03:10:49 | web-cat/code-workout | https://api.github.com/repos/web-cat/code-workout | closed | Add current exercise score to exercise feedback display | Staged feature priority: normal | Add a radial progress bar showing the current score (same as in thumbnails) to the right of the "Feedback" title in the exercise feedback (generated in the sse/ajax_feedback view).
| 1.0 | Add current exercise score to exercise feedback display - Add a radial progress bar showing the current score (same as in thumbnails) to the right of the "Feedback" title in the exercise feedback (generated in the sse/ajax_feedback view).
| priority | add current exercise score to exercise feedback display add a radial progress bar showing the current score same as in thumbnails to the right of the feedback title in the exercise feedback generated in the sse ajax feedback view | 1 |
24,997 | 24,559,224,225 | IssuesEvent | 2022-10-12 18:38:38 | JuliaReach/LazySets.jl | https://api.github.com/repos/JuliaReach/LazySets.jl | opened | PolarDirections and SphericalDirections use different interpretations of Nφ | usability :computer_mouse: fix :face_with_head_bandage: | In `PolarDirections`, the range has length `Nφ+1`.
In `SphericalDirections`, the range has length `Nφ`. | True | PolarDirections and SphericalDirections use different interpretations of Nφ - In `PolarDirections`, the range has length `Nφ+1`.
In `SphericalDirections`, the range has length `Nφ`. | non_priority | polardirections and sphericaldirections use different interpretations of nφ in polardirections the range has length nφ in sphericaldirections the range has length nφ | 0 |
31,453 | 14,970,563,595 | IssuesEvent | 2021-01-27 19:48:18 | flutter/flutter | https://api.github.com/repos/flutter/flutter | opened | My app is slow or missing frames (metabug) | created via performance template macos-metal perf: speed severe: performance | This is a meta-issue to track reproducible reports of jank in Flutter apps.
If you are experiencing jank in your app:
1. Try to reproduce the problem in a test app. Either run `flutter create janktest` and recreate the situation you are experiencing in that app, or clone your app and delete code until you have the jank reproducing with a single .dart file.
2. [File a bug](https://github.com/flutter/flutter/issues/new?assignees=&labels=created+via+performance+template&template=5_performance_speed.md&title=) and include your .dart file demonstrating the problem. If you need more than just a .dart file (for example, assets are needed to reproduce the issue, or plugins/packages are needed to reproduce the issue) then create a GitHub repository and upload the app there.
Make sure to include the `flutter doctor -v` output and any logs from `flutter run` and `flutter analyze`.
3. Switch flutter to master channel and run this app on a physical device using profile mode with Skia tracing enabled, as follows:
`flutter channel master`
`flutter run --profile --trace-skia`
Then press ‘P’ to enable the performance overlay.
The bleeding edge master channel is encouraged here because Flutter is constantly fixing bugs and improving its performance. Your problem in an older Flutter version may have already been solved in the master channel.
4. Record a video of the performance issue using another phone so we can have an intuitive understanding of what happened. Don’t use "adb screenrecord", as that affects the performance of the profile run. Attach the video to your bug.
5. Open Observatory and save a timeline trace of the performance issue so we know which functions might be causing it. See "How to Collect and Read Timeline Traces" on this blog post:
https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499
Attach the JSON file containing your trace to your bug. You may also wish to include a screenshot of the part of the trace showing the problem you are seeing, just so that people can see at a glance what kind of performance issue the bug is about.
6. Mention _this_ bug in your bug, so that GitHub includes a link to it here.
Please avoid commenting on this bug. Keep each issue separate so that we can examine each specific problem individually. Having one issue that contains comments about multiple problems make the issue intractable. | True | My app is slow or missing frames (metabug) - This is a meta-issue to track reproducible reports of jank in Flutter apps.
If you are experiencing jank in your app:
1. Try to reproduce the problem in a test app. Either run `flutter create janktest` and recreate the situation you are experiencing in that app, or clone your app and delete code until you have the jank reproducing with a single .dart file.
2. [File a bug](https://github.com/flutter/flutter/issues/new?assignees=&labels=created+via+performance+template&template=5_performance_speed.md&title=) and include your .dart file demonstrating the problem. If you need more than just a .dart file (for example, assets are needed to reproduce the issue, or plugins/packages are needed to reproduce the issue) then create a GitHub repository and upload the app there.
Make sure to include the `flutter doctor -v` output and any logs from `flutter run` and `flutter analyze`.
3. Switch flutter to master channel and run this app on a physical device using profile mode with Skia tracing enabled, as follows:
`flutter channel master`
`flutter run --profile --trace-skia`
Then press ‘P’ to enable the performance overlay.
The bleeding edge master channel is encouraged here because Flutter is constantly fixing bugs and improving its performance. Your problem in an older Flutter version may have already been solved in the master channel.
4. Record a video of the performance issue using another phone so we can have an intuitive understanding of what happened. Don’t use "adb screenrecord", as that affects the performance of the profile run. Attach the video to your bug.
5. Open Observatory and save a timeline trace of the performance issue so we know which functions might be causing it. See "How to Collect and Read Timeline Traces" on this blog post:
https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499
Attach the JSON file containing your trace to your bug. You may also wish to include a screenshot of the part of the trace showing the problem you are seeing, just so that people can see at a glance what kind of performance issue the bug is about.
6. Mention _this_ bug in your bug, so that GitHub includes a link to it here.
Please avoid commenting on this bug. Keep each issue separate so that we can examine each specific problem individually. Having one issue that contains comments about multiple problems make the issue intractable. | non_priority | my app is slow or missing frames metabug this is a meta issue to track reproducible reports of jank in flutter apps if you are experiencing jank in your app try to reproduce the problem in a test app either run flutter create janktest and recreate the situation you are experiencing in that app or clone your app and delete code until you have the jank reproducing with a single dart file and include your dart file demonstrating the problem if you need more than just a dart file for example assets are needed to reproduce the issue or plugins packages are needed to reproduce the issue then create a github repository and upload the app there make sure to include the flutter doctor v output and any logs from flutter run and flutter analyze switch flutter to master channel and run this app on a physical device using profile mode with skia tracing enabled as follows flutter channel master flutter run profile trace skia then press ‘p’ to enable the performance overlay the bleeding edge master channel is encouraged here because flutter is constantly fixing bugs and improving its performance your problem in an older flutter version may have already been solved in the master channel record a video of the performance issue using another phone so we can have an intuitive understanding of what happened don’t use adb screenrecord as that affects the performance of the profile run attach the video to your bug open observatory and save a timeline trace of the performance issue so we know which functions might be causing it see how to collect and read timeline traces on this blog post attach the json file containing your trace to your bug you may also wish to include a screenshot of the part of the trace showing the problem you are seeing just so that people can see at a glance what kind of performance issue the bug is about mention this bug in your bug so that github includes a link to it here please avoid commenting on this bug keep each issue separate so that we can examine each specific problem individually having one issue that contains comments about multiple problems make the issue intractable | 0 |
820,981 | 30,798,127,426 | IssuesEvent | 2023-07-31 21:45:24 | PazerOP/tf2_bot_detector | https://api.github.com/repos/PazerOP/tf2_bot_detector | opened | [BUG] Bot Detector will not update | Type: Bug Priority: Medium | Any time I boot up the bot detector, it comes up with this Error. I don't think the bot detector even works while it's in this state. I saw others using the bot detector yesterday and, while I had mine open, it didn't do anything. Yes, I did boot TF2 from the bot detector.
``Update check failed:
- class tf2_bot_detector::http_error
- Failed to HTTP GET https://tf2bd-util.pazer.us:443/AppInstaller/LatestVersion.json?type=Public:
<UNKNOWN>(HTTP 523)``
Log:
[2023-07-31_22-35-39.log](https://github.com/PazerOP/tf2_bot_detector/files/12222482/2023-07-31_22-35-39.log)
I am currently on the latest version, since someone suggested to reinstall it and I've deleted every instance I am aware of, and it hasn't fixed the issue. | 1.0 | [BUG] Bot Detector will not update - Any time I boot up the bot detector, it comes up with this Error. I don't think the bot detector even works while it's in this state. I saw others using the bot detector yesterday and, while I had mine open, it didn't do anything. Yes, I did boot TF2 from the bot detector.
``Update check failed:
- class tf2_bot_detector::http_error
- Failed to HTTP GET https://tf2bd-util.pazer.us:443/AppInstaller/LatestVersion.json?type=Public:
<UNKNOWN>(HTTP 523)``
Log:
[2023-07-31_22-35-39.log](https://github.com/PazerOP/tf2_bot_detector/files/12222482/2023-07-31_22-35-39.log)
I am currently on the latest version, since someone suggested to reinstall it and I've deleted every instance I am aware of, and it hasn't fixed the issue. | priority | bot detector will not update any time i boot up the bot detector it comes up with this error i don t think the bot detector even works while it s in this state i saw others using the bot detector yesterday and while i had mine open it didn t do anything yes i did boot from the bot detector update check failed class bot detector http error failed to http get http log i am currently on the latest version since someone suggested to reinstall it and i ve deleted every instance i am aware of and it hasn t fixed the issue | 1 |
512,311 | 14,893,807,569 | IssuesEvent | 2021-01-21 06:17:07 | project-serum/anchor | https://api.github.com/repos/project-serum/anchor | opened | Provide separate Account and AccountBox types | priority:2 | Currently we use `ProgramAccount` to refer to accounts owned by the current program. This type boxes it's inner account by default (needed to sometimes get around BPF stack size limits). Instead, we should provide a non-boxed version with `type AccountBox = Box<NonBoxedVersion>;` as the boxed variant. | 1.0 | Provide separate Account and AccountBox types - Currently we use `ProgramAccount` to refer to accounts owned by the current program. This type boxes it's inner account by default (needed to sometimes get around BPF stack size limits). Instead, we should provide a non-boxed version with `type AccountBox = Box<NonBoxedVersion>;` as the boxed variant. | priority | provide separate account and accountbox types currently we use programaccount to refer to accounts owned by the current program this type boxes it s inner account by default needed to sometimes get around bpf stack size limits instead we should provide a non boxed version with type accountbox box as the boxed variant | 1 |
168,425 | 14,148,187,917 | IssuesEvent | 2020-11-10 22:05:33 | bounswe/bounswe2020group2 | https://api.github.com/repos/bounswe/bounswe2020group2 | closed | Review of Project Plan | documentation effort: medium priority: low type: enhancement who: group | - [x] Resource names of [Project Plan](https://github.com/bounswe/bounswe2020group2/wiki/Project-Plan) should be changed according to subgroups for Android, back end and front end development.
- [ ] Tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader.
**_Due Date: 02.11.2020 Monday @23.59_** | 1.0 | Review of Project Plan - - [x] Resource names of [Project Plan](https://github.com/bounswe/bounswe2020group2/wiki/Project-Plan) should be changed according to subgroups for Android, back end and front end development.
- [ ] Tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader.
**_Due Date: 02.11.2020 Monday @23.59_** | non_priority | review of project plan resource names of should be changed according to subgroups for android back end and front end development tasks of future process should be updated after finding the plan of each subgroup out from its corresponding leader due date monday | 0 |
54,098 | 13,391,229,413 | IssuesEvent | 2020-09-02 22:05:12 | solex2006/SELIProject | https://api.github.com/repos/solex2006/SELIProject | opened | Audio is being requested for all scenes | :bug:bug P2 - High S2 - Minor Defects env:next release module:STORING TELLING type:functional | ## Expected Behavior
<!--- Tell us what should happen -->
If a video with audio is already upload in scene, an audio file should not be required
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
I can't publish a story while not all scenes has an audio file.
## Environment
<!--- Please tell us about your environment -->
<!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView -->
https://vm2161.kaj.pouta.csc.fi/
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- Provide a detailed description of the change or addition you are proposing -->
Include a `input type=checkbox` asking if the uploaded video is a video-only file (this will be important for accessibility).
If user check the input, then audio must be required.
## Related issues
<!--- Not obligatory, but reference related issues -->
| 1.0 | Audio is being requested for all scenes - ## Expected Behavior
<!--- Tell us what should happen -->
If a video with audio is already upload in scene, an audio file should not be required
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
I can't publish a story while not all scenes has an audio file.
## Environment
<!--- Please tell us about your environment -->
<!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView -->
https://vm2161.kaj.pouta.csc.fi/
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- Provide a detailed description of the change or addition you are proposing -->
Include a `input type=checkbox` asking if the uploaded video is a video-only file (this will be important for accessibility).
If user check the input, then audio must be required.
## Related issues
<!--- Not obligatory, but reference related issues -->
| non_priority | audio is being requested for all scenes expected behavior if a video with audio is already upload in scene an audio file should not be required current behavior i can t publish a story while not all scenes has an audio file environment possible solution include a input type checkbox asking if the uploaded video is a video only file this will be important for accessibility if user check the input then audio must be required related issues | 0 |
253,655 | 8,058,843,353 | IssuesEvent | 2018-08-02 19:51:24 | 18F/10x-qualitative-data | https://api.github.com/repos/18F/10x-qualitative-data | opened | Define requirements for Camille and OCE | Sprint 3 priority product | As an *analytics and content team member*
I want *to have access to Qualtrics*
so that *I can view reports, curate data and manage surveys*.
## Acceptance criteria
- [ ] Request *Qualtrics* access for all members of Analytics and Content teams Qualtrics
- [ ] Determine if we require any additional *Qualtrics* services that are not currently supported by our account.
| 1.0 | Define requirements for Camille and OCE - As an *analytics and content team member*
I want *to have access to Qualtrics*
so that *I can view reports, curate data and manage surveys*.
## Acceptance criteria
- [ ] Request *Qualtrics* access for all members of Analytics and Content teams Qualtrics
- [ ] Determine if we require any additional *Qualtrics* services that are not currently supported by our account.
| priority | define requirements for camille and oce as an analytics and content team member i want to have access to qualtrics so that i can view reports curate data and manage surveys acceptance criteria request qualtrics access for all members of analytics and content teams qualtrics determine if we require any additional qualtrics services that are not currently supported by our account | 1 |
85,355 | 15,736,693,043 | IssuesEvent | 2021-03-30 01:13:22 | benlazarine/datahog | https://api.github.com/repos/benlazarine/datahog | opened | CVE-2019-19118 (Medium) detected in Django-2.1.11.tar.gz | security vulnerability | ## CVE-2019-19118 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.1.11.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz">https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz</a></p>
<p>Path to dependency file: datahog/django/requirements.txt</p>
<p>Path to vulnerable library: datahog/django/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.1.11.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Django 2.1 before 2.1.15 and 2.2 before 2.2.8 allows unintended model editing. A Django model admin displaying inline related models, where the user has view-only permissions to a parent model but edit permissions to the inline model, would be presented with an editing UI, allowing POST requests, for updating the inline model. Directly editing the view-only parent model was not possible, but the parent model's save() method was called, triggering potential side effects, and causing pre and post-save signal handlers to be invoked. (To resolve this, the Django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable.)
<p>Publish Date: 2019-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19118>CVE-2019-19118</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118</a></p>
<p>Release Date: 2019-12-02</p>
<p>Fix Resolution: 2.1.15,2.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19118 (Medium) detected in Django-2.1.11.tar.gz - ## CVE-2019-19118 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.1.11.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz">https://files.pythonhosted.org/packages/e0/e9/7e6008abee3eb2a40704c95a5cfc8a9627012df1580289d3df0f34c99766/Django-2.1.11.tar.gz</a></p>
<p>Path to dependency file: datahog/django/requirements.txt</p>
<p>Path to vulnerable library: datahog/django/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.1.11.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Django 2.1 before 2.1.15 and 2.2 before 2.2.8 allows unintended model editing. A Django model admin displaying inline related models, where the user has view-only permissions to a parent model but edit permissions to the inline model, would be presented with an editing UI, allowing POST requests, for updating the inline model. Directly editing the view-only parent model was not possible, but the parent model's save() method was called, triggering potential side effects, and causing pre and post-save signal handlers to be invoked. (To resolve this, the Django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable.)
<p>Publish Date: 2019-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19118>CVE-2019-19118</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19118</a></p>
<p>Release Date: 2019-12-02</p>
<p>Fix Resolution: 2.1.15,2.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in django tar gz cve medium severity vulnerability vulnerable library django tar gz a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file datahog django requirements txt path to vulnerable library datahog django requirements txt dependency hierarchy x django tar gz vulnerable library vulnerability details django before and before allows unintended model editing a django model admin displaying inline related models where the user has view only permissions to a parent model but edit permissions to the inline model would be presented with an editing ui allowing post requests for updating the inline model directly editing the view only parent model was not possible but the parent model s save method was called triggering potential side effects and causing pre and post save signal handlers to be invoked to resolve this the django admin is adjusted to require edit permissions on the parent model in order for inline models to be editable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
242,867 | 7,849,642,409 | IssuesEvent | 2018-06-20 05:04:01 | borela/naomi | https://api.github.com/repos/borela/naomi | closed | Goto Definition from JSX to component | enhancement priority: medium | Hey,
I would first like to say that I really like your package.
My react code (with fragments, ligatures, flow, es6, ..) looks really good now thanks to you.
The only thing I'm missing is to jump from JSX code to the used component.
Below is a simple example of a component. It would be great if I could simple navigate to the Input or Label component with Goto Definition.
Is this something that this package could do? Or should this be done by a different package?
Or maybe this can't be done?
```
import * as React from "react";
import {uniqueId} from "lodash";
import {Label} from "./core/Label";
import {Input} from "./core/Input";
const InputWithLabel = ({text, value, onChange, children}) => {
const id = uniqueId("id_")
let label = children
if (!label) label = text + ':'
return <>
<Label htmlFor={id}>{label}</Label>
<Input id={id}
type={type}
value={value}
onChange={onChange}
placeholder={text}/>
</>
}
```
| 1.0 | Goto Definition from JSX to component - Hey,
I would first like to say that I really like your package.
My react code (with fragments, ligatures, flow, es6, ..) looks really good now thanks to you.
The only thing I'm missing is to jump from JSX code to the used component.
Below is a simple example of a component. It would be great if I could simple navigate to the Input or Label component with Goto Definition.
Is this something that this package could do? Or should this be done by a different package?
Or maybe this can't be done?
```
import * as React from "react";
import {uniqueId} from "lodash";
import {Label} from "./core/Label";
import {Input} from "./core/Input";
const InputWithLabel = ({text, value, onChange, children}) => {
const id = uniqueId("id_")
let label = children
if (!label) label = text + ':'
return <>
<Label htmlFor={id}>{label}</Label>
<Input id={id}
type={type}
value={value}
onChange={onChange}
placeholder={text}/>
</>
}
```
| priority | goto definition from jsx to component hey i would first like to say that i really like your package my react code with fragments ligatures flow looks really good now thanks to you the only thing i m missing is to jump from jsx code to the used component below is a simple example of a component it would be great if i could simple navigate to the input or label component with goto definition is this something that this package could do or should this be done by a different package or maybe this can t be done import as react from react import uniqueid from lodash import label from core label import input from core input const inputwithlabel text value onchange children const id uniqueid id let label children if label label text return label input id id type type value value onchange onchange placeholder text | 1 |
649,346 | 21,280,069,093 | IssuesEvent | 2022-04-14 00:08:10 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Group config check items by type in AppAssessment report | medium priority app-development | CPU and Memory Requests, CPU and Memory limits, Probes, etc.
DoD
- [ ] Group table items if applicable (resource usage, security, etc) (high) | 1.0 | Group config check items by type in AppAssessment report - CPU and Memory Requests, CPU and Memory limits, Probes, etc.
DoD
- [ ] Group table items if applicable (resource usage, security, etc) (high) | priority | group config check items by type in appassessment report cpu and memory requests cpu and memory limits probes etc dod group table items if applicable resource usage security etc high | 1 |
198,065 | 6,969,402,027 | IssuesEvent | 2017-12-11 05:08:40 | adventistmedia/adventisthub-support | https://api.github.com/repos/adventistmedia/adventisthub-support | closed | Colour Scheme | enhancement priority | Update colour scheme to closer resemble church brand and/or other church websites (e.g. search.adventistchurch.com) | 1.0 | Colour Scheme - Update colour scheme to closer resemble church brand and/or other church websites (e.g. search.adventistchurch.com) | priority | colour scheme update colour scheme to closer resemble church brand and or other church websites e g search adventistchurch com | 1 |
477,666 | 13,766,032,431 | IssuesEvent | 2020-10-07 14:06:21 | adaptlearning/adapt_framework | https://api.github.com/repos/adaptlearning/adapt_framework | closed | adapt-contrib-spoor: scorm tracking of course state not working if _shouldStoreResponses: false | bug high priority | ### Subject of the issue/enhancement/features
The tracking of which blocks are complete/incomplete doesn't appear to be working as it should
### Your environment
* Framework master branch (but also occurs with v5.6.1)/spoor v3.5.0 (also checked master, still has issue)
* Checked in IE11, FF, Chrome
* Not checked devices but don't expect it to be any different
### Steps to reproduce
Use `adapt create course` to set up a fresh install of Adapt. In config.json, set `_spoor._tracking._shouldSubmitScore` to `false`. run `grunt dev` then `grunt server-scorm`. Go into Presentation Components and complete a couple of interactive components. Reload the course.
### Expected behaviour
The components completed prior to the reload should show as completed still
### Actual behaviour
No components show as completed.
### Additional info
The 'completion string' that is output to the console is correct BUT the value of `"c"` in the suspend_data never changes from its initial value of `"hAA"`, which doesn't seem anything close to being correct:

| 1.0 | adapt-contrib-spoor: scorm tracking of course state not working if _shouldStoreResponses: false - ### Subject of the issue/enhancement/features
The tracking of which blocks are complete/incomplete doesn't appear to be working as it should
### Your environment
* Framework master branch (but also occurs with v5.6.1)/spoor v3.5.0 (also checked master, still has issue)
* Checked in IE11, FF, Chrome
* Not checked devices but don't expect it to be any different
### Steps to reproduce
Use `adapt create course` to set up a fresh install of Adapt. In config.json, set `_spoor._tracking._shouldSubmitScore` to `false`. run `grunt dev` then `grunt server-scorm`. Go into Presentation Components and complete a couple of interactive components. Reload the course.
### Expected behaviour
The components completed prior to the reload should show as completed still
### Actual behaviour
No components show as completed.
### Additional info
The 'completion string' that is output to the console is correct BUT the value of `"c"` in the suspend_data never changes from its initial value of `"hAA"`, which doesn't seem anything close to being correct:

| priority | adapt contrib spoor scorm tracking of course state not working if shouldstoreresponses false subject of the issue enhancement features the tracking of which blocks are complete incomplete doesn t appear to be working as it should your environment framework master branch but also occurs with spoor also checked master still has issue checked in ff chrome not checked devices but don t expect it to be any different steps to reproduce use adapt create course to set up a fresh install of adapt in config json set spoor tracking shouldsubmitscore to false run grunt dev then grunt server scorm go into presentation components and complete a couple of interactive components reload the course expected behaviour the components completed prior to the reload should show as completed still actual behaviour no components show as completed additional info the completion string that is output to the console is correct but the value of c in the suspend data never changes from its initial value of haa which doesn t seem anything close to being correct | 1 |
295,162 | 25,458,887,762 | IssuesEvent | 2022-11-24 16:30:44 | DataDog/dd-trace-java | https://api.github.com/repos/DataDog/dd-trace-java | closed | TracerDebuggerIntegrationTest is flaky | flaky-test disabled | ```
java.lang.AssertionError
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertFalse(Assert.java:64)
at org.junit.Assert.assertFalse(Assert.java:74)
at datadog.smoketest.TracerDebuggerIntegrationTest.testTracer(TracerDebuggerIntegrationTest.java:88)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
``` | 1.0 | TracerDebuggerIntegrationTest is flaky - ```
java.lang.AssertionError
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertFalse(Assert.java:64)
at org.junit.Assert.assertFalse(Assert.java:74)
at datadog.smoketest.TracerDebuggerIntegrationTest.testTracer(TracerDebuggerIntegrationTest.java:88)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:725)
at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131)
at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:149)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:140)
at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:84)
at org.junit.jupiter.engine.execution.ExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(ExecutableInvoker.java:115)
at org.junit.jupiter.engine.execution.ExecutableInvoker.lambda$invoke$0(ExecutableInvoker.java:105)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45)
at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:104)
at org.junit.jupiter.engine.execution.ExecutableInvoker.invoke(ExecutableInvoker.java:98)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:214)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:210)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:135)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
``` | non_priority | tracerdebuggerintegrationtest is flaky java lang assertionerror at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert assertfalse assert java at org junit assert assertfalse assert java at datadog smoketest tracerdebuggerintegrationtest testtracer tracerdebuggerintegrationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit platform commons util reflectionutils invokemethod reflectionutils java at org junit jupiter engine execution methodinvocation proceed methodinvocation java at org junit jupiter engine execution invocationinterceptorchain validatinginvocation proceed invocationinterceptorchain java at org junit jupiter engine extension timeoutextension intercept timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestablemethod timeoutextension java at org junit jupiter engine extension timeoutextension intercepttestmethod timeoutextension java at org junit jupiter engine execution executableinvoker reflectiveinterceptorcall lambda ofvoidmethod executableinvoker java at org junit jupiter engine execution executableinvoker lambda invoke executableinvoker java at org junit jupiter engine execution invocationinterceptorchain interceptedinvocation proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain proceed invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain chainandinvoke invocationinterceptorchain java at org junit jupiter engine execution invocationinterceptorchain invoke invocationinterceptorchain java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine execution executableinvoker invoke executableinvoker java at org junit jupiter engine descriptor testmethodtestdescriptor lambda invoketestmethod testmethodtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor testmethodtestdescriptor invoketestmethod testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor execute testmethodtestdescriptor java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java | 0 |
112,883 | 17,109,369,377 | IssuesEvent | 2021-07-10 01:45:43 | vyas0189/CougarCS-Backend | https://api.github.com/repos/vyas0189/CougarCS-Backend | closed | CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed | security vulnerability | ## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>Path to dependency file: CougarCS-Backend/package.json</p>
<p>Path to vulnerable library: CougarCS-Backend/node_modules/path-parse</p>
<p>
Dependency Hierarchy:
- preset-env-7.14.7.tgz (Root Library)
- babel-plugin-polyfill-corejs3-0.2.3.tgz
- helper-define-polyfill-provider-0.2.3.tgz
- resolve-1.20.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vyas0189/CougarCS-Backend/commit/41e2657038d5a56936fef604251da533f6670866">41e2657038d5a56936fef604251da533f6670866</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed - ## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>Path to dependency file: CougarCS-Backend/package.json</p>
<p>Path to vulnerable library: CougarCS-Backend/node_modules/path-parse</p>
<p>
Dependency Hierarchy:
- preset-env-7.14.7.tgz (Root Library)
- babel-plugin-polyfill-corejs3-0.2.3.tgz
- helper-define-polyfill-provider-0.2.3.tgz
- resolve-1.20.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vyas0189/CougarCS-Backend/commit/41e2657038d5a56936fef604251da533f6670866">41e2657038d5a56936fef604251da533f6670866</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in path parse tgz autoclosed cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file cougarcs backend package json path to vulnerable library cougarcs backend node modules path parse dependency hierarchy preset env tgz root library babel plugin polyfill tgz helper define polyfill provider tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse step up your open source security game with whitesource | 0 |
68,700 | 29,482,149,162 | IssuesEvent | 2023-06-02 06:51:44 | IITC-CE/iitc-app-status | https://api.github.com/repos/IITC-CE/iitc-app-status | closed | 🛑 Weblate service is down | status weblate-service | In [`6c04ec2`](https://github.com/IITC-CE/iitc-app-status/commit/6c04ec2feece866524484ba2e0fd37674631ef81
), Weblate service (https://weblate.iitc.app) was **down**:
- HTTP code: 500
- Response time: 1500 ms
| 1.0 | 🛑 Weblate service is down - In [`6c04ec2`](https://github.com/IITC-CE/iitc-app-status/commit/6c04ec2feece866524484ba2e0fd37674631ef81
), Weblate service (https://weblate.iitc.app) was **down**:
- HTTP code: 500
- Response time: 1500 ms
| non_priority | 🛑 weblate service is down in weblate service was down http code response time ms | 0 |
239,032 | 7,785,947,066 | IssuesEvent | 2018-06-06 17:22:50 | dojot/dojot | https://api.github.com/repos/dojot/dojot | closed | Websocket is not working with https | Priority:High Team:Frontend Type:Bug | The maps and graphics are not receiving data from websocket when used Https. | 1.0 | Websocket is not working with https - The maps and graphics are not receiving data from websocket when used Https. | priority | websocket is not working with https the maps and graphics are not receiving data from websocket when used https | 1 |
223,650 | 17,614,306,251 | IssuesEvent | 2021-08-18 07:48:26 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | Fail to paste one secondary ADLS Gen2 blob/folder | 🧪 testing :gear: adls gen2 | **Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210818.3
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 20.04/MacOS Big Sur 11.4
**Architecture**: ia32\x64
**How Found**: From running test cases
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one ADLS Gen2 storage account -> Blob Containers.
2. Open one blob container -> Upload one blob to it.
3. Right click the blob container -> Click 'Open Secondary Location'.
4. Copy the blob -> Paste it to another blob container.
5. Check whether succeed to paste.
## Expected Experience ##
Succeed to paste.
## Actual Experience ##
Fail to paste with the below error.
Error detail:
` "message": "\"failed to perform copy command due to error: cannot start job due to error: cannot list files due to reason -> github.com/Azure/azure-storage-blob-go/azblob.newStorageError, /home/vsts/go/pkg/mod/github.com/!azure/azure-storage-blob-go@v0.10.1-0.20210407023846`

| 1.0 | Fail to paste one secondary ADLS Gen2 blob/folder - **Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210818.3
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 20.04/MacOS Big Sur 11.4
**Architecture**: ia32\x64
**How Found**: From running test cases
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one ADLS Gen2 storage account -> Blob Containers.
2. Open one blob container -> Upload one blob to it.
3. Right click the blob container -> Click 'Open Secondary Location'.
4. Copy the blob -> Paste it to another blob container.
5. Check whether succeed to paste.
## Expected Experience ##
Succeed to paste.
## Actual Experience ##
Fail to paste with the below error.
Error detail:
` "message": "\"failed to perform copy command due to error: cannot start job due to error: cannot list files due to reason -> github.com/Azure/azure-storage-blob-go/azblob.newStorageError, /home/vsts/go/pkg/mod/github.com/!azure/azure-storage-blob-go@v0.10.1-0.20210407023846`

| non_priority | fail to paste one secondary adls blob folder storage explorer version dev build number branch main platform os windows linux ubuntu macos big sur architecture how found from running test cases regression from not a regression steps to reproduce expand one adls storage account blob containers open one blob container upload one blob to it right click the blob container click open secondary location copy the blob paste it to another blob container check whether succeed to paste expected experience succeed to paste actual experience fail to paste with the below error error detail message failed to perform copy command due to error cannot start job due to error cannot list files due to reason github com azure azure storage blob go azblob newstorageerror home vsts go pkg mod github com azure azure storage blob go | 0 |
241,114 | 7,808,907,450 | IssuesEvent | 2018-06-11 21:48:59 | tomaka/winit | https://api.github.com/repos/tomaka/winit | closed | Segmentation fault in XIFreeDeviceInfo | difficulty: average platform: X11 priority: normal status: waiting type: bug | This was posted on jwilm/alacritty#895:
> Version: master branch, revision f68558e9dcce97e0d4d842a2d5b62e31e4124210
>
> I ran it with `rust-gdb ./target/debug/alacritty`, waited for the crash, and got this backtrace:
```
Thread 1 "alacritty" received signal SIGSEGV, Segmentation fault.
0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6
(gdb) bt
#0 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6
#1 0x0000555555d86967 in winit::platform::platform::x11::{{impl}}::drop (self=0x7fffffff5070)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:641
#2 0x0000555555d54395 in core::ptr::drop_in_place<winit::platform::platform::x11::DeviceInfo> () at /checkout/src/libcore/ptr.rs:60
#3 0x0000555555d86353 in winit::platform::platform::x11::EventsLoop::init_device (self=0x7fffffff8768, device=17)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:575
#4 0x0000555555745c9a in winit::platform::platform::x11::EventsLoop::process_event<&mut closure> (self=0x7fffffff8768, xev=0x7fffffff7818, callback=0x7fffffff78e0)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:553
#5 0x0000555555742313 in winit::platform::platform::x11::EventsLoop::run_forever<closure> (self=0x7fffffff8768, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:174
#6 0x000055555574201a in winit::platform::platform::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/mod.rs:387
#7 0x0000555555741e5d in winit::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/lib.rs:219
#8 0x000055555579d7ed in alacritty::window::Window::wait_events<closure> (self=0x7fffffff8760, func=closure = {...}) at /home/mrmetric/Downloads/source/alacritty/src/window.rs:274
#9 0x00005555556b0101 in alacritty::event::Processor<alacritty::event_loop::Notifier>::process_events<alacritty::event_loop::Notifier> (self=0x7fffffffc178, term=0x7fffef8d5e10,
window=0x7fffffff8760) at /home/mrmetric/Downloads/source/alacritty/src/event.rs:375
#10 0x00005555557a0f52 in alacritty::run (config=Config = {...}, options=Options = {...}) at src/main.rs:170
#11 0x000055555579faac in alacritty::main () at src/main.rs:45
```
I haven't had a chance to look at it myself yet. | 1.0 | Segmentation fault in XIFreeDeviceInfo - This was posted on jwilm/alacritty#895:
> Version: master branch, revision f68558e9dcce97e0d4d842a2d5b62e31e4124210
>
> I ran it with `rust-gdb ./target/debug/alacritty`, waited for the crash, and got this backtrace:
```
Thread 1 "alacritty" received signal SIGSEGV, Segmentation fault.
0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6
(gdb) bt
#0 0x00007ffff363b9fc in XIFreeDeviceInfo () from /usr/lib/libXi.so.6
#1 0x0000555555d86967 in winit::platform::platform::x11::{{impl}}::drop (self=0x7fffffff5070)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:641
#2 0x0000555555d54395 in core::ptr::drop_in_place<winit::platform::platform::x11::DeviceInfo> () at /checkout/src/libcore/ptr.rs:60
#3 0x0000555555d86353 in winit::platform::platform::x11::EventsLoop::init_device (self=0x7fffffff8768, device=17)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:575
#4 0x0000555555745c9a in winit::platform::platform::x11::EventsLoop::process_event<&mut closure> (self=0x7fffffff8768, xev=0x7fffffff7818, callback=0x7fffffff78e0)
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:553
#5 0x0000555555742313 in winit::platform::platform::x11::EventsLoop::run_forever<closure> (self=0x7fffffff8768, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/x11/mod.rs:174
#6 0x000055555574201a in winit::platform::platform::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/platform/linux/mod.rs:387
#7 0x0000555555741e5d in winit::EventsLoop::run_forever<closure> (self=0x7fffffff8760, callback=closure = {...})
at /home/mrmetric/.cargo/registry/src/github.com-1ecc6299db9ec823/winit-0.7.6/src/lib.rs:219
#8 0x000055555579d7ed in alacritty::window::Window::wait_events<closure> (self=0x7fffffff8760, func=closure = {...}) at /home/mrmetric/Downloads/source/alacritty/src/window.rs:274
#9 0x00005555556b0101 in alacritty::event::Processor<alacritty::event_loop::Notifier>::process_events<alacritty::event_loop::Notifier> (self=0x7fffffffc178, term=0x7fffef8d5e10,
window=0x7fffffff8760) at /home/mrmetric/Downloads/source/alacritty/src/event.rs:375
#10 0x00005555557a0f52 in alacritty::run (config=Config = {...}, options=Options = {...}) at src/main.rs:170
#11 0x000055555579faac in alacritty::main () at src/main.rs:45
```
I haven't had a chance to look at it myself yet. | priority | segmentation fault in xifreedeviceinfo this was posted on jwilm alacritty version master branch revision i ran it with rust gdb target debug alacritty waited for the crash and got this backtrace thread alacritty received signal sigsegv segmentation fault in xifreedeviceinfo from usr lib libxi so gdb bt in xifreedeviceinfo from usr lib libxi so in winit platform platform impl drop self at home mrmetric cargo registry src github com winit src platform linux mod rs in core ptr drop in place at checkout src libcore ptr rs in winit platform platform eventsloop init device self device at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop process event self xev callback at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src platform linux mod rs in winit platform platform eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src platform linux mod rs in winit eventsloop run forever self callback closure at home mrmetric cargo registry src github com winit src lib rs in alacritty window window wait events self func closure at home mrmetric downloads source alacritty src window rs in alacritty event processor process events self term window at home mrmetric downloads source alacritty src event rs in alacritty run config config options options at src main rs in alacritty main at src main rs i haven t had a chance to look at it myself yet | 1 |
281,458 | 8,695,531,110 | IssuesEvent | 2018-12-04 15:23:08 | cosmos/voyager | https://api.github.com/repos/cosmos/voyager | closed | Proposal Creation feedbacks | bug 🐞 governance-1 :ballot_box: high priority | UI Version: 0.X.X
Description:
Governance/Create Proposal
After entering 65 characters in the Title or 200 characters in the Description, the Create Proposal button is disabled but is not clear for which reason
It would be nice to have feedbacks and/or highlighting the box that is blocking the creation.
Voting starting date is always invalid.
Subset of #1690 | 1.0 | Proposal Creation feedbacks - UI Version: 0.X.X
Description:
Governance/Create Proposal
After entering 65 characters in the Title or 200 characters in the Description, the Create Proposal button is disabled but is not clear for which reason
It would be nice to have feedbacks and/or highlighting the box that is blocking the creation.
Voting starting date is always invalid.
Subset of #1690 | priority | proposal creation feedbacks ui version x x description governance create proposal after entering characters in the title or characters in the description the create proposal button is disabled but is not clear for which reason it would be nice to have feedbacks and or highlighting the box that is blocking the creation voting starting date is always invalid subset of | 1 |
47,392 | 24,985,021,003 | IssuesEvent | 2022-11-02 14:30:11 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | scale testing: Gigarepo indexing is not completing | team/search-core performance | Indexing of the gigarepo on the scale testing instance does not appear to be completing:
https://scaletesting.sgdev.org/gigarepo/-/settings/index
| True | scale testing: Gigarepo indexing is not completing - Indexing of the gigarepo on the scale testing instance does not appear to be completing:
https://scaletesting.sgdev.org/gigarepo/-/settings/index
| non_priority | scale testing gigarepo indexing is not completing indexing of the gigarepo on the scale testing instance does not appear to be completing | 0 |
288,214 | 21,691,033,520 | IssuesEvent | 2022-05-09 15:22:53 | GaloyMoney/galoy | https://api.github.com/repos/GaloyMoney/galoy | opened | possible race condition in payment | bug documentation need investigation lightning | I had an issue where:
1/ the recipient was using BBW (using the tipping page)
2/ two person were trying to pay the invoice at the same time. one with BBW. one with an external wallet.
both payment initiated within less than 1s timeframe. the external wallet was probably initiated first.
the payment with the external wallet went through (success), the payment from BBW showed an error.
when looking at both wallets, both had a payment sent.
it was expected for the external wallet because the payment succeed.
for the BBW wallet, the payment showed as a failure, so this was not expected.
it may be a race condition when an invoice is being paid at the same time both from outside, using lnd, and inside with an intraledger payment. | 1.0 | possible race condition in payment - I had an issue where:
1/ the recipient was using BBW (using the tipping page)
2/ two person were trying to pay the invoice at the same time. one with BBW. one with an external wallet.
both payment initiated within less than 1s timeframe. the external wallet was probably initiated first.
the payment with the external wallet went through (success), the payment from BBW showed an error.
when looking at both wallets, both had a payment sent.
it was expected for the external wallet because the payment succeed.
for the BBW wallet, the payment showed as a failure, so this was not expected.
it may be a race condition when an invoice is being paid at the same time both from outside, using lnd, and inside with an intraledger payment. | non_priority | possible race condition in payment i had an issue where the recipient was using bbw using the tipping page two person were trying to pay the invoice at the same time one with bbw one with an external wallet both payment initiated within less than timeframe the external wallet was probably initiated first the payment with the external wallet went through success the payment from bbw showed an error when looking at both wallets both had a payment sent it was expected for the external wallet because the payment succeed for the bbw wallet the payment showed as a failure so this was not expected it may be a race condition when an invoice is being paid at the same time both from outside using lnd and inside with an intraledger payment | 0 |
311,889 | 9,540,011,972 | IssuesEvent | 2019-04-30 18:24:31 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | PortSwitch primitive system | priority: medium team: russ type: feature request | As discussed with a few of you, I've got the "PortSwitch" system mostly coded up now, and will PR it this weekend. It's a simple system that offers many ports on the input, but passes-through only one of them to the output (with one additional input for the port_selector). I imagine this being useful for combining many low-level controller primitives (e.g. DifferentialIK, PlanEval, hybrid control, etc) with all of them wired up, but only one of them getting called thanks to this switch.
cc @kmuhlrad @pangtao22 | 1.0 | PortSwitch primitive system - As discussed with a few of you, I've got the "PortSwitch" system mostly coded up now, and will PR it this weekend. It's a simple system that offers many ports on the input, but passes-through only one of them to the output (with one additional input for the port_selector). I imagine this being useful for combining many low-level controller primitives (e.g. DifferentialIK, PlanEval, hybrid control, etc) with all of them wired up, but only one of them getting called thanks to this switch.
cc @kmuhlrad @pangtao22 | priority | portswitch primitive system as discussed with a few of you i ve got the portswitch system mostly coded up now and will pr it this weekend it s a simple system that offers many ports on the input but passes through only one of them to the output with one additional input for the port selector i imagine this being useful for combining many low level controller primitives e g differentialik planeval hybrid control etc with all of them wired up but only one of them getting called thanks to this switch cc kmuhlrad | 1 |
311,070 | 9,527,870,130 | IssuesEvent | 2019-04-29 06:29:32 | Steven-Wagner/Movie-Suggestor | https://api.github.com/repos/Steven-Wagner/Movie-Suggestor | opened | Show a response each time user gets a duplicate error in Reviews | High Priority bug | When a user submits a review that has an error; the first time the user sees the change on the screen and is properly informed. However, if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time. Leading the user to think that the event was not registered by the client. | 1.0 | Show a response each time user gets a duplicate error in Reviews - When a user submits a review that has an error; the first time the user sees the change on the screen and is properly informed. However, if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time. Leading the user to think that the event was not registered by the client. | priority | show a response each time user gets a duplicate error in reviews when a user submits a review that has an error the first time the user sees the change on the screen and is properly informed however if they make the same error again the screen shows the proper error message but does not indicate that it is a different message from the one they received the first time leading the user to think that the event was not registered by the client | 1 |
149,832 | 5,729,234,361 | IssuesEvent | 2017-04-21 05:09:34 | dhowe/ReadersJS | https://api.github.com/repos/dhowe/ReadersJS | closed | Update website to point to readers-js-live | priority: High | LOCATIONS (others?):
http://thereadersproject.org/elc3.html#ELC3
http://thereadersproject.org/index.html (add link to left nav?) | 1.0 | Update website to point to readers-js-live - LOCATIONS (others?):
http://thereadersproject.org/elc3.html#ELC3
http://thereadersproject.org/index.html (add link to left nav?) | priority | update website to point to readers js live locations others add link to left nav | 1 |
3,825 | 6,334,056,315 | IssuesEvent | 2017-07-26 15:53:05 | opensensorhub/sensorml-editor | https://api.github.com/repos/opensensorhub/sensorml-editor | closed | Bring Ontology Search capabilities back into editor | Requirement | We previously had the ability to search an ontology for terms and them select. Previously this was done by downloading the ontology and doing a search directly on the "local" version. This search however, can be done by using the search capabilities on the Ontology (ORR) web services.
For example, see: http://sensorml.com/orr/#st
This search can also be done using sparQL: http://sensorml.com/ont/sparql.html | 1.0 | Bring Ontology Search capabilities back into editor - We previously had the ability to search an ontology for terms and them select. Previously this was done by downloading the ontology and doing a search directly on the "local" version. This search however, can be done by using the search capabilities on the Ontology (ORR) web services.
For example, see: http://sensorml.com/orr/#st
This search can also be done using sparQL: http://sensorml.com/ont/sparql.html | non_priority | bring ontology search capabilities back into editor we previously had the ability to search an ontology for terms and them select previously this was done by downloading the ontology and doing a search directly on the local version this search however can be done by using the search capabilities on the ontology orr web services for example see this search can also be done using sparql | 0 |
629,309 | 20,029,055,061 | IssuesEvent | 2022-02-02 01:45:53 | apcountryman/picolibrary | https://api.github.com/repos/apcountryman/picolibrary | closed | Add GPIO I/O pin concept | priority-normal status-awaiting_review type-feature | Add GPIO I/O pin concept (`::picolibrary::GPIO::IO_Pin_Concept`) and associated mock (`::picolibrary::Testing::Unit::GPIO::Mock_IO_Pin`).
- [x] The `IO_Pin_Concept` class should be defined in the `include/picolibrary/gpio.h`/`source/picolibrary/gpio.cc` header/source file pair
- [x] The `Mock_IO_Pin` class should be defined in the `include/picolibrary/testing/unit/gpio.h`/`source/picolibrary/testing/unit/gpio.cc` header/source file pair
- [x] IO pins should be expected to support the following operations:
- [x] `IO_Pin_Concept() noexcept = default;`
- [x] `IO_Pin_Concept( IO_Pin_Concept && source ) noexcept = default;`
- [x] `~IO_Pin_Concept() noexcept = default;`
- [x] `auto operator=( IO_Pin_Concept && expression ) noexcept -> IO_Pin_Concept & = default;`
- [x] `void initialize( Initial_Pin_State initial_pin_state = Initial_Pin_State::LOW ) noexcept;` Initialize the pin's hardware
- [x] `auto is_low() const noexcept -> bool;`: Check if the pin is in the low state
- [x] `auto is_high() const noexcept -> bool;`: Check if the pin is in the high state
- [x] `void transition_to_low() noexcept;`: Transition the pin to the low state
- [x] `void transition_to_high() noexcept;`: Transition the pin to the high state
- [x] `void toggle() noexcept;`: Toggle the pin state
- [x] The `Mock_IO_Pin` class should include a movable mock handle | 1.0 | Add GPIO I/O pin concept - Add GPIO I/O pin concept (`::picolibrary::GPIO::IO_Pin_Concept`) and associated mock (`::picolibrary::Testing::Unit::GPIO::Mock_IO_Pin`).
- [x] The `IO_Pin_Concept` class should be defined in the `include/picolibrary/gpio.h`/`source/picolibrary/gpio.cc` header/source file pair
- [x] The `Mock_IO_Pin` class should be defined in the `include/picolibrary/testing/unit/gpio.h`/`source/picolibrary/testing/unit/gpio.cc` header/source file pair
- [x] IO pins should be expected to support the following operations:
- [x] `IO_Pin_Concept() noexcept = default;`
- [x] `IO_Pin_Concept( IO_Pin_Concept && source ) noexcept = default;`
- [x] `~IO_Pin_Concept() noexcept = default;`
- [x] `auto operator=( IO_Pin_Concept && expression ) noexcept -> IO_Pin_Concept & = default;`
- [x] `void initialize( Initial_Pin_State initial_pin_state = Initial_Pin_State::LOW ) noexcept;` Initialize the pin's hardware
- [x] `auto is_low() const noexcept -> bool;`: Check if the pin is in the low state
- [x] `auto is_high() const noexcept -> bool;`: Check if the pin is in the high state
- [x] `void transition_to_low() noexcept;`: Transition the pin to the low state
- [x] `void transition_to_high() noexcept;`: Transition the pin to the high state
- [x] `void toggle() noexcept;`: Toggle the pin state
- [x] The `Mock_IO_Pin` class should include a movable mock handle | priority | add gpio i o pin concept add gpio i o pin concept picolibrary gpio io pin concept and associated mock picolibrary testing unit gpio mock io pin the io pin concept class should be defined in the include picolibrary gpio h source picolibrary gpio cc header source file pair the mock io pin class should be defined in the include picolibrary testing unit gpio h source picolibrary testing unit gpio cc header source file pair io pins should be expected to support the following operations io pin concept noexcept default io pin concept io pin concept source noexcept default io pin concept noexcept default auto operator io pin concept expression noexcept io pin concept default void initialize initial pin state initial pin state initial pin state low noexcept initialize the pin s hardware auto is low const noexcept bool check if the pin is in the low state auto is high const noexcept bool check if the pin is in the high state void transition to low noexcept transition the pin to the low state void transition to high noexcept transition the pin to the high state void toggle noexcept toggle the pin state the mock io pin class should include a movable mock handle | 1 |
75,205 | 14,409,046,922 | IssuesEvent | 2020-12-04 01:15:33 | UBC-Thunderbots/Software | https://api.github.com/repos/UBC-Thunderbots/Software | closed | Investigate Warnings from SimulatedTests | Difficulty - 13 G3 - Code Quality G6 - Gameplay and Navigation T - Maintenance | ### Description of the task
Several SimulatedTests cause warnings that fill the logs and are highly suspicious and may indicate a bug. One example is `free_kick_play_test`, which causes `Could not draw a line from the goalie to the ball through the defender crease` and `Error updating robot state, stopping` warnings. To start this investigation, run `//software/simulated_tests:free_kick_play_test` in debug mode in gdb with breakpoints where those warnings are logged.
<!--
What does this work depend on?
What interface will this work use or create?
What are the main components of the task?
Where does this work fit in the larger project?
It is important to define this task sufficiently so that an untrained
team member can take it on and know where to start. Feel free to
link to resources or other team member which could guide the assignee to
complete the task
-->
### Acceptance criteria
<!--
Checkbox list that outlines what needs to be done in order for this task
to be considered "complete".
Specify any implementation requirements such as data structures,
functionalities, testing requirements, documentation, etc.
-->
- [ ] Explain why we get so many warnings from simulated tests
- [ ] Figure out the best way to avoid these warnings, e.g. fixing an underlying bug or suppressing warnings that aren't helpful
### Blocked By
<!--
List all other issues that need to be completed before this one, ex:
- #123
- #374
-->
| 1.0 | Investigate Warnings from SimulatedTests - ### Description of the task
Several SimulatedTests cause warnings that fill the logs and are highly suspicious and may indicate a bug. One example is `free_kick_play_test`, which causes `Could not draw a line from the goalie to the ball through the defender crease` and `Error updating robot state, stopping` warnings. To start this investigation, run `//software/simulated_tests:free_kick_play_test` in debug mode in gdb with breakpoints where those warnings are logged.
<!--
What does this work depend on?
What interface will this work use or create?
What are the main components of the task?
Where does this work fit in the larger project?
It is important to define this task sufficiently so that an untrained
team member can take it on and know where to start. Feel free to
link to resources or other team member which could guide the assignee to
complete the task
-->
### Acceptance criteria
<!--
Checkbox list that outlines what needs to be done in order for this task
to be considered "complete".
Specify any implementation requirements such as data structures,
functionalities, testing requirements, documentation, etc.
-->
- [ ] Explain why we get so many warnings from simulated tests
- [ ] Figure out the best way to avoid these warnings, e.g. fixing an underlying bug or suppressing warnings that aren't helpful
### Blocked By
<!--
List all other issues that need to be completed before this one, ex:
- #123
- #374
-->
| non_priority | investigate warnings from simulatedtests description of the task several simulatedtests cause warnings that fill the logs and are highly suspicious and may indicate a bug one example is free kick play test which causes could not draw a line from the goalie to the ball through the defender crease and error updating robot state stopping warnings to start this investigation run software simulated tests free kick play test in debug mode in gdb with breakpoints where those warnings are logged what does this work depend on what interface will this work use or create what are the main components of the task where does this work fit in the larger project it is important to define this task sufficiently so that an untrained team member can take it on and know where to start feel free to link to resources or other team member which could guide the assignee to complete the task acceptance criteria checkbox list that outlines what needs to be done in order for this task to be considered complete specify any implementation requirements such as data structures functionalities testing requirements documentation etc explain why we get so many warnings from simulated tests figure out the best way to avoid these warnings e g fixing an underlying bug or suppressing warnings that aren t helpful blocked by list all other issues that need to be completed before this one ex | 0 |
135,889 | 11,028,500,045 | IssuesEvent | 2019-12-06 11:50:20 | UglyToad/PdfPig | https://api.github.com/repos/UglyToad/PdfPig | closed | InvalidOperationException: 'No charstring sequence with the name /one in this font.' | bug testing | Opening the page 192 of [this](https://www.toshiba-storage.com/wp-content/uploads/2018/01/UM_Canvio_Basics.pdf) document throws `System.InvalidOperationException: 'No charstring sequence with the name /one in this font.'` in `Type2CharStrings.Generate(string name)`.
__Full details:__
System.InvalidOperationException
HResult=0x80131509
Message=No charstring sequence with the name /one in this font.
Source=UglyToad.PdfPig
StackTrace:
at UglyToad.PdfPig.Fonts.CompactFontFormat.CharStrings.Type2CharStrings.Generate(String name) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CharStrings\Type2CharStrings.cs:line 63
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.<>c__DisplayClass16_0.<GetCharacterBoundingBox>b__1(Type2CharStrings x) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 41
at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 40
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFontProgram.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFontProgram.cs:line 48
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.<>c__DisplayClass20_0.<GetBoundingBoxInGlyphSpace>b__1(CompactFontFormatFontProgram x) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 185
at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBoxInGlyphSpace(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 169
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBox(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 124
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ShowText(IInputBytes bytes) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 198
at UglyToad.PdfPig.Graphics.Operations.TextShowing.ShowText.Run(IOperationContext operationContext) in \PdfPig\src\UglyToad.PdfPig\Graphics\Operations\TextShowing\ShowText.cs:line 68
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ProcessOperations(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 111
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.Process(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 102
at UglyToad.PdfPig.Parser.PageFactory.GetContent(IReadOnlyList'1 contentBytes, CropBox cropBox, UserSpaceUnit userSpaceUnit, PageRotationDegrees rotation, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 129
at UglyToad.PdfPig.Parser.PageFactory.Create(Int32 number, DictionaryToken dictionary, PageTreeMembers pageTreeMembers, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 112
at UglyToad.PdfPig.Content.Pages.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\Content\Pages.cs:line 66
at UglyToad.PdfPig.PdfDocument.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\PdfDocument.cs:line 150 | 1.0 | InvalidOperationException: 'No charstring sequence with the name /one in this font.' - Opening the page 192 of [this](https://www.toshiba-storage.com/wp-content/uploads/2018/01/UM_Canvio_Basics.pdf) document throws `System.InvalidOperationException: 'No charstring sequence with the name /one in this font.'` in `Type2CharStrings.Generate(string name)`.
__Full details:__
System.InvalidOperationException
HResult=0x80131509
Message=No charstring sequence with the name /one in this font.
Source=UglyToad.PdfPig
StackTrace:
at UglyToad.PdfPig.Fonts.CompactFontFormat.CharStrings.Type2CharStrings.Generate(String name) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CharStrings\Type2CharStrings.cs:line 63
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.<>c__DisplayClass16_0.<GetCharacterBoundingBox>b__1(Type2CharStrings x) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 41
at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFont.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFont.cs:line 40
at UglyToad.PdfPig.Fonts.CompactFontFormat.CompactFontFormatFontProgram.GetCharacterBoundingBox(String characterName) in \PdfPig\src\UglyToad.PdfPig\Fonts\CompactFontFormat\CompactFontFormatFontProgram.cs:line 48
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.<>c__DisplayClass20_0.<GetBoundingBoxInGlyphSpace>b__1(CompactFontFormatFontProgram x) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 185
at UglyToad.PdfPig.Util.Union'2.Case2.Match(Action'1 first, Action'1 second) in \PdfPig\src\UglyToad.PdfPig\Util\Union.cs:line 56
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBoxInGlyphSpace(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 169
at UglyToad.PdfPig.Fonts.Simple.Type1FontSimple.GetBoundingBox(Int32 characterCode) in \PdfPig\src\UglyToad.PdfPig\Fonts\Simple\Type1FontSimple.cs:line 124
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ShowText(IInputBytes bytes) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 198
at UglyToad.PdfPig.Graphics.Operations.TextShowing.ShowText.Run(IOperationContext operationContext) in \PdfPig\src\UglyToad.PdfPig\Graphics\Operations\TextShowing\ShowText.cs:line 68
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.ProcessOperations(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 111
at UglyToad.PdfPig.Graphics.ContentStreamProcessor.Process(IReadOnlyList'1 operations) in \PdfPig\src\UglyToad.PdfPig\Graphics\ContentStreamProcessor.cs:line 102
at UglyToad.PdfPig.Parser.PageFactory.GetContent(IReadOnlyList'1 contentBytes, CropBox cropBox, UserSpaceUnit userSpaceUnit, PageRotationDegrees rotation, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 129
at UglyToad.PdfPig.Parser.PageFactory.Create(Int32 number, DictionaryToken dictionary, PageTreeMembers pageTreeMembers, Boolean isLenientParsing) in \PdfPig\src\UglyToad.PdfPig\Parser\PageFactory.cs:line 112
at UglyToad.PdfPig.Content.Pages.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\Content\Pages.cs:line 66
at UglyToad.PdfPig.PdfDocument.GetPage(Int32 pageNumber) in \PdfPig\src\UglyToad.PdfPig\PdfDocument.cs:line 150 | non_priority | invalidoperationexception no charstring sequence with the name one in this font opening the page of document throws system invalidoperationexception no charstring sequence with the name one in this font in generate string name full details system invalidoperationexception hresult message no charstring sequence with the name one in this font source uglytoad pdfpig stacktrace at uglytoad pdfpig fonts compactfontformat charstrings generate string name in pdfpig src uglytoad pdfpig fonts compactfontformat charstrings cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfont c b x in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfont cs line at uglytoad pdfpig util union match action first action second in pdfpig src uglytoad pdfpig util union cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfont getcharacterboundingbox string charactername in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfont cs line at uglytoad pdfpig fonts compactfontformat compactfontformatfontprogram getcharacterboundingbox string charactername in pdfpig src uglytoad pdfpig fonts compactfontformat compactfontformatfontprogram cs line at uglytoad pdfpig fonts simple c b compactfontformatfontprogram x in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig util union match action first action second in pdfpig src uglytoad pdfpig util union cs line at uglytoad pdfpig fonts simple getboundingboxinglyphspace charactercode in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig fonts simple getboundingbox charactercode in pdfpig src uglytoad pdfpig fonts simple cs line at uglytoad pdfpig graphics contentstreamprocessor showtext iinputbytes bytes in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig graphics operations textshowing showtext run ioperationcontext operationcontext in pdfpig src uglytoad pdfpig graphics operations textshowing showtext cs line at uglytoad pdfpig graphics contentstreamprocessor processoperations ireadonlylist operations in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig graphics contentstreamprocessor process ireadonlylist operations in pdfpig src uglytoad pdfpig graphics contentstreamprocessor cs line at uglytoad pdfpig parser pagefactory getcontent ireadonlylist contentbytes cropbox cropbox userspaceunit userspaceunit pagerotationdegrees rotation boolean islenientparsing in pdfpig src uglytoad pdfpig parser pagefactory cs line at uglytoad pdfpig parser pagefactory create number dictionarytoken dictionary pagetreemembers pagetreemembers boolean islenientparsing in pdfpig src uglytoad pdfpig parser pagefactory cs line at uglytoad pdfpig content pages getpage pagenumber in pdfpig src uglytoad pdfpig content pages cs line at uglytoad pdfpig pdfdocument getpage pagenumber in pdfpig src uglytoad pdfpig pdfdocument cs line | 0 |
293,580 | 8,997,804,453 | IssuesEvent | 2019-02-02 15:59:44 | Qiskit/qiskit-terra | https://api.github.com/repos/Qiskit/qiskit-terra | closed | change/extend how execution timeouts work | priority: low type: discussion | This issue is meant to start a discussion about how execution timeouts work.
Currently, the timeout setting of a job starts the clock as soon as a job is submitted. If the job times out, the SDK stops waiting for the result but does not remove the job from the remote queue (this is not currently possible with the API). It may be better to add a timeout, or change the timeout, such that it starts when the circuits in the qobj start executing.
update 12/2018:
With asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work, however there is still a use case. A user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs. Some queuing systems such as LSF support setting a run time limit. | 1.0 | change/extend how execution timeouts work - This issue is meant to start a discussion about how execution timeouts work.
Currently, the timeout setting of a job starts the clock as soon as a job is submitted. If the job times out, the SDK stops waiting for the result but does not remove the job from the remote queue (this is not currently possible with the API). It may be better to add a timeout, or change the timeout, such that it starts when the circuits in the qobj start executing.
update 12/2018:
With asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work, however there is still a use case. A user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs. Some queuing systems such as LSF support setting a run time limit. | priority | change extend how execution timeouts work this issue is meant to start a discussion about how execution timeouts work currently the timeout setting of a job starts the clock as soon as a job is submitted if the job times out the sdk stops waiting for the result but does not remove the job from the remote queue this is not currently possible with the api it may be better to add a timeout or change the timeout such that it starts when the circuits in the qobj start executing update with asynchronous jobs and the ability to cancel jobs there is less of a need to change how timeouts work however there is still a use case a user could want to limit the time taken for a job to run on the backend which might be desirable if the backend charges the user for longer running jobs some queuing systems such as lsf support setting a run time limit | 1 |
565,103 | 16,748,971,751 | IssuesEvent | 2021-06-11 19:38:01 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Make linked system better | Category: Gameplay Needs Task Plan Priority: Low Squad: Mountain Goat Status: Won't Fix Type: Feature | Inspired by this changes https://github.com/StrangeLoopGames/Eco/pull/6541
We have another items with this problem Crafting stations with fuel, some lights(braziers) and I think we will have a lot of another objects in future. So, currentrly we have 2 option for inventories in our techtree: None and Linked. I assume we can add another one - 'Excluded' that allow to see another **linked** inventories in its storage, but it will not shown in another storages. That you did for Pots, but add this customization to all items in techtree. | 1.0 | Make linked system better - Inspired by this changes https://github.com/StrangeLoopGames/Eco/pull/6541
We have another items with this problem Crafting stations with fuel, some lights(braziers) and I think we will have a lot of another objects in future. So, currentrly we have 2 option for inventories in our techtree: None and Linked. I assume we can add another one - 'Excluded' that allow to see another **linked** inventories in its storage, but it will not shown in another storages. That you did for Pots, but add this customization to all items in techtree. | priority | make linked system better inspired by this changes we have another items with this problem crafting stations with fuel some lights braziers and i think we will have a lot of another objects in future so currentrly we have option for inventories in our techtree none and linked i assume we can add another one excluded that allow to see another linked inventories in its storage but it will not shown in another storages that you did for pots but add this customization to all items in techtree | 1 |
482,603 | 13,910,475,133 | IssuesEvent | 2020-10-20 16:07:09 | stevencohn/OneMore | https://api.github.com/repos/stevencohn/OneMore | closed | Inline markdown conversions | enhancement low priority | Inline markdown conversions. This is a biggie and I'm not sure of its usefulness since the OneNote UI is so feature rich already. But the idea is to associate markdown directives with styles and/or custom styles, invoke an Apply Markdown command and replace all inline directives with styling.
I think there are alternatives that convert markdown to HTML and then you can paste HTML into OneNote.
| 1.0 | Inline markdown conversions - Inline markdown conversions. This is a biggie and I'm not sure of its usefulness since the OneNote UI is so feature rich already. But the idea is to associate markdown directives with styles and/or custom styles, invoke an Apply Markdown command and replace all inline directives with styling.
I think there are alternatives that convert markdown to HTML and then you can paste HTML into OneNote.
| priority | inline markdown conversions inline markdown conversions this is a biggie and i m not sure of its usefulness since the onenote ui is so feature rich already but the idea is to associate markdown directives with styles and or custom styles invoke an apply markdown command and replace all inline directives with styling i think there are alternatives that convert markdown to html and then you can paste html into onenote | 1 |
566,479 | 16,822,517,730 | IssuesEvent | 2021-06-17 14:36:36 | AugurProject/turbo | https://api.github.com/repos/AugurProject/turbo | opened | Update the add liquidity button on the market page - needs to stand out more for LPs | High Priority enhancement | https://www.figma.com/file/6y4nvjfeVZwzwKcoXB0neq/Augur-Simplified-UI?node-id=2919%3A5640

| 1.0 | Update the add liquidity button on the market page - needs to stand out more for LPs - https://www.figma.com/file/6y4nvjfeVZwzwKcoXB0neq/Augur-Simplified-UI?node-id=2919%3A5640

| priority | update the add liquidity button on the market page needs to stand out more for lps | 1 |
301,465 | 26,050,447,828 | IssuesEvent | 2022-12-22 18:06:13 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: ledger/nodes=6/multi-az failed | C-test-failure O-robot O-roachtest branch-master T-kv | roachtest.ledger/nodes=6/multi-az [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/ledger/nodes=6/multi-az) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22):
```
test artifacts and logs in: /artifacts/ledger/nodes=6/multi-az/run_1
(test_impl.go:291).Fatal: output in run_163643.333865619_n7_workload_run_ledger: ./workload run ledger --init --histograms=perf/stats.json --concurrency=192 --duration=10m {pgurl:1-2} returned: COMMAND_PROBLEM: exit status 1
(test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=16</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*ledger/nodes=6/multi-az.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22677
Epic CRDB-18656 | 2.0 | roachtest: ledger/nodes=6/multi-az failed - roachtest.ledger/nodes=6/multi-az [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/ledger/nodes=6/multi-az) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22):
```
test artifacts and logs in: /artifacts/ledger/nodes=6/multi-az/run_1
(test_impl.go:291).Fatal: output in run_163643.333865619_n7_workload_run_ledger: ./workload run ledger --init --histograms=perf/stats.json --concurrency=192 --duration=10m {pgurl:1-2} returned: COMMAND_PROBLEM: exit status 1
(test_impl.go:291).Fatal: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=16</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*ledger/nodes=6/multi-az.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22677
Epic CRDB-18656 | non_priority | roachtest ledger nodes multi az failed roachtest ledger nodes multi az with on master test artifacts and logs in artifacts ledger nodes multi az run test impl go fatal output in run workload run ledger workload run ledger init histograms perf stats json concurrency duration pgurl returned command problem exit status test impl go fatal monitor failure monitor task failed t fatal was called parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb kv triage jira issue crdb epic crdb | 0 |
824,120 | 31,141,641,677 | IssuesEvent | 2023-08-16 00:51:09 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | Bestial Wrath | Class: Hunter Spell Priority: High | **Description:**
if you try to use Bestial Wrath to break from paladin's repentance it says you can't use that while stunned,didn't try it with cyclone yet.
Edit : doesn't work with Gouge too
**How to reproduce:**
cast repentance on BM hunter and use bestial wrath (The Beast Within sub-talent is needed to break the CC's)
**How it should work:**
Since Wotlk expansion patch changes : Bestial Wrath and Ability hunter beastwithin [[The Beast Within]](https://wowpedia.fandom.com/wiki/The_Beast_Within) changed to function like PvP Trinkets, removing any movement impairing effects and giving immunity to these effects for 10 seconds. The tooltip doesn't say this, but it also now breaks all forms of CC that a PvP trinket would break (including Cyclone, Sap, etc.).
**Database links:**
https://wowpedia.fandom.com/wiki/Bestial_Wrath
Edit: found these videos so far https://www.youtube.com/watch?app=desktop&v=XmsUQZWqSQs&ab_channel=JohnDaniel
time 0:22 slow from dk
time 2:28 rogue cheap shot
time 3:10 dk slow+warrior stun
time 5:37 druid cyclone
time 8:35 mage deep freeze
https://www.youtube.com/watch?app=desktop&v=5qx2a0_xmFg&ab_channel=JohnDaniel
time 7:47 feral slow (deep wounds)
https://www.youtube.com/watch?app=desktop&v=X9jqRLTgcLU&ab_channel=RandomPeopleISH
Kazzak PvP server---time 4:32 where he speaks about using Bestial Wrath as a trinket=breaks any stun/slow/snares etc
also gathered the informations from one "blizzlike" server where they fixed Bestial Wrath according to these videos and few more comments. | 1.0 | Bestial Wrath - **Description:**
if you try to use Bestial Wrath to break from paladin's repentance it says you can't use that while stunned,didn't try it with cyclone yet.
Edit : doesn't work with Gouge too
**How to reproduce:**
cast repentance on BM hunter and use bestial wrath (The Beast Within sub-talent is needed to break the CC's)
**How it should work:**
Since Wotlk expansion patch changes : Bestial Wrath and Ability hunter beastwithin [[The Beast Within]](https://wowpedia.fandom.com/wiki/The_Beast_Within) changed to function like PvP Trinkets, removing any movement impairing effects and giving immunity to these effects for 10 seconds. The tooltip doesn't say this, but it also now breaks all forms of CC that a PvP trinket would break (including Cyclone, Sap, etc.).
**Database links:**
https://wowpedia.fandom.com/wiki/Bestial_Wrath
Edit: found these videos so far https://www.youtube.com/watch?app=desktop&v=XmsUQZWqSQs&ab_channel=JohnDaniel
time 0:22 slow from dk
time 2:28 rogue cheap shot
time 3:10 dk slow+warrior stun
time 5:37 druid cyclone
time 8:35 mage deep freeze
https://www.youtube.com/watch?app=desktop&v=5qx2a0_xmFg&ab_channel=JohnDaniel
time 7:47 feral slow (deep wounds)
https://www.youtube.com/watch?app=desktop&v=X9jqRLTgcLU&ab_channel=RandomPeopleISH
Kazzak PvP server---time 4:32 where he speaks about using Bestial Wrath as a trinket=breaks any stun/slow/snares etc
also gathered the informations from one "blizzlike" server where they fixed Bestial Wrath according to these videos and few more comments. | priority | bestial wrath description if you try to use bestial wrath to break from paladin s repentance it says you can t use that while stunned didn t try it with cyclone yet edit doesn t work with gouge too how to reproduce cast repentance on bm hunter and use bestial wrath the beast within sub talent is needed to break the cc s how it should work since wotlk expansion patch changes bestial wrath and ability hunter beastwithin changed to function like pvp trinkets removing any movement impairing effects and giving immunity to these effects for seconds the tooltip doesn t say this but it also now breaks all forms of cc that a pvp trinket would break including cyclone sap etc database links edit found these videos so far time slow from dk time rogue cheap shot time dk slow warrior stun time druid cyclone time mage deep freeze time feral slow deep wounds kazzak pvp server time where he speaks about using bestial wrath as a trinket breaks any stun slow snares etc also gathered the informations from one blizzlike server where they fixed bestial wrath according to these videos and few more comments | 1 |
748,716 | 26,134,474,976 | IssuesEvent | 2022-12-29 10:15:38 | exyleio/exyleio | https://api.github.com/repos/exyleio/exyleio | closed | Set up SMTP server | priority 1 - high tag:development type:task area:scripts | - Requires: exyle.io domain
As recommended by the [pocketbase documentation](https://pocketbase.io/docs/going-to-production/#use-smtp-mail-server), having a SMTP server prevents deliverability issues (email being sent to spam, etc.) when sending account verification email. Use [listmonk](https://github.com/knadh/listmonk) for it. | 1.0 | Set up SMTP server - - Requires: exyle.io domain
As recommended by the [pocketbase documentation](https://pocketbase.io/docs/going-to-production/#use-smtp-mail-server), having a SMTP server prevents deliverability issues (email being sent to spam, etc.) when sending account verification email. Use [listmonk](https://github.com/knadh/listmonk) for it. | priority | set up smtp server requires exyle io domain as recommended by the having a smtp server prevents deliverability issues email being sent to spam etc when sending account verification email use for it | 1 |
38,670 | 5,194,529,444 | IssuesEvent | 2017-01-23 04:22:17 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Skipping tests on Jenkins | area/tests priority/P3 | We're seeing a bunch of skipped tests on Jenkins that I ~~don't~~ see skipped locally as well...
```
Of 3620 tests executed in 580.208s, 3613 succeeded, 0 failed, and 7 were skipped.
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestNewOpenStack" was skipped:
=== RUN TestNewOpenStack
--- SKIP: TestNewOpenStack (0.00s)
openstack_test.go:113: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestInstances" was skipped:
=== RUN TestInstances
--- SKIP: TestInstances (0.00s)
openstack_test.go:125: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestLoadBalancer" was skipped:
=== RUN TestLoadBalancer
--- SKIP: TestLoadBalancer (0.00s)
openstack_test.go:157: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestVolumes" was skipped:
=== RUN TestVolumes
--- SKIP: TestVolumes (0.00s)
openstack_test.go:205: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestNewRackspace" was skipped:
=== RUN TestNewRackspace
--- SKIP: TestNewRackspace (0.00s)
rackspace_test.go:113: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestInstances" was skipped:
=== RUN TestInstances
--- SKIP: TestInstances (0.00s)
rackspace_test.go:125: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/util/dbus", test case "TestRealDBus" was skipped:
=== RUN TestRealDBus
--- SKIP: TestRealDBus (0.01s)
dbus_test.go:73: D-Bus session bus is not available: exec: "dbus-launch": executable file not found in $PATH
```
/cc @smarterclayton
@rhcarvalho iirc you knew something about this?
| 1.0 | Skipping tests on Jenkins - We're seeing a bunch of skipped tests on Jenkins that I ~~don't~~ see skipped locally as well...
```
Of 3620 tests executed in 580.208s, 3613 succeeded, 0 failed, and 7 were skipped.
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestNewOpenStack" was skipped:
=== RUN TestNewOpenStack
--- SKIP: TestNewOpenStack (0.00s)
openstack_test.go:113: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestInstances" was skipped:
=== RUN TestInstances
--- SKIP: TestInstances (0.00s)
openstack_test.go:125: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestLoadBalancer" was skipped:
=== RUN TestLoadBalancer
--- SKIP: TestLoadBalancer (0.00s)
openstack_test.go:157: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/openstack", test case "TestVolumes" was skipped:
=== RUN TestVolumes
--- SKIP: TestVolumes (0.00s)
openstack_test.go:205: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestNewRackspace" was skipped:
=== RUN TestNewRackspace
--- SKIP: TestNewRackspace (0.00s)
rackspace_test.go:113: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/cloudprovider/providers/rackspace", test case "TestInstances" was skipped:
=== RUN TestInstances
--- SKIP: TestInstances (0.00s)
rackspace_test.go:125: No config found in environment
In suite "github.com/openshift/origin/Godeps/_workspace/src/k8s.io/kubernetes/pkg/util/dbus", test case "TestRealDBus" was skipped:
=== RUN TestRealDBus
--- SKIP: TestRealDBus (0.01s)
dbus_test.go:73: D-Bus session bus is not available: exec: "dbus-launch": executable file not found in $PATH
```
/cc @smarterclayton
@rhcarvalho iirc you knew something about this?
| non_priority | skipping tests on jenkins we re seeing a bunch of skipped tests on jenkins that i don t see skipped locally as well of tests executed in succeeded failed and were skipped in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testnewopenstack was skipped run testnewopenstack skip testnewopenstack openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testinstances was skipped run testinstances skip testinstances openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testloadbalancer was skipped run testloadbalancer skip testloadbalancer openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers openstack test case testvolumes was skipped run testvolumes skip testvolumes openstack test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers rackspace test case testnewrackspace was skipped run testnewrackspace skip testnewrackspace rackspace test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg cloudprovider providers rackspace test case testinstances was skipped run testinstances skip testinstances rackspace test go no config found in environment in suite github com openshift origin godeps workspace src io kubernetes pkg util dbus test case testrealdbus was skipped run testrealdbus skip testrealdbus dbus test go d bus session bus is not available exec dbus launch executable file not found in path cc smarterclayton rhcarvalho iirc you knew something about this | 0 |
168,799 | 6,386,703,018 | IssuesEvent | 2017-08-03 11:53:00 | ampproject/amp-by-example | https://api.github.com/repos/ampproject/amp-by-example | closed | Product Page - Add to cart is broken | bug P1: High Priority sample | Looking at https://ampbyexample.com/samples_templates/product_page/preview/
`Add to card` is supposed to redirect the user however it looks like the XHR to `https://ampbyexample.com/samples_templates/product_page/add_to_cart?__amp_source_origin=https%3A%2F%2Fampbyexample.com` does not return a JSON response and therefore we get `Failed to parse response JSON: SyntaxError: Unexpected token < in JSON at position 0` error` and user is not redirected. | 1.0 | Product Page - Add to cart is broken - Looking at https://ampbyexample.com/samples_templates/product_page/preview/
`Add to card` is supposed to redirect the user however it looks like the XHR to `https://ampbyexample.com/samples_templates/product_page/add_to_cart?__amp_source_origin=https%3A%2F%2Fampbyexample.com` does not return a JSON response and therefore we get `Failed to parse response JSON: SyntaxError: Unexpected token < in JSON at position 0` error` and user is not redirected. | priority | product page add to cart is broken looking at add to card is supposed to redirect the user however it looks like the xhr to does not return a json response and therefore we get failed to parse response json syntaxerror unexpected token in json at position error and user is not redirected | 1 |
191,200 | 22,215,673,027 | IssuesEvent | 2022-06-08 01:11:46 | uniquelyparticular/zendesk-magento-m1-request | https://api.github.com/repos/uniquelyparticular/zendesk-magento-m1-request | closed | CVE-2019-13173 (High) detected in fstream-1.0.11.tgz - autoclosed | security vulnerability | ## CVE-2019-13173 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-1.0.11.tgz</b></p></summary>
<p>Advanced file system stream things</p>
<p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz">https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz</a></p>
<p>Path to dependency file: /zendesk-magento-m1-request/package.json</p>
<p>Path to vulnerable library: /tmp/git/zendesk-magento-m1-request/node_modules/npm/node_modules/fstream/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.13.16.tgz (Root Library)
- npm-5.1.10.tgz
- npm-6.9.0.tgz
- node-gyp-3.8.0.tgz
- :x: **fstream-1.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/zendesk-magento-m1-request/commit/281dde4039452ceffbf3f483ab9f672be345fcd3">281dde4039452ceffbf3f483ab9f672be345fcd3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.
<p>Publish Date: 2019-07-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p>
<p>Release Date: 2019-07-02</p>
<p>Fix Resolution: 1.0.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-13173 (High) detected in fstream-1.0.11.tgz - autoclosed - ## CVE-2019-13173 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-1.0.11.tgz</b></p></summary>
<p>Advanced file system stream things</p>
<p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz">https://registry.npmjs.org/fstream/-/fstream-1.0.11.tgz</a></p>
<p>Path to dependency file: /zendesk-magento-m1-request/package.json</p>
<p>Path to vulnerable library: /tmp/git/zendesk-magento-m1-request/node_modules/npm/node_modules/fstream/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-15.13.16.tgz (Root Library)
- npm-5.1.10.tgz
- npm-6.9.0.tgz
- node-gyp-3.8.0.tgz
- :x: **fstream-1.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/zendesk-magento-m1-request/commit/281dde4039452ceffbf3f483ab9f672be345fcd3">281dde4039452ceffbf3f483ab9f672be345fcd3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.
<p>Publish Date: 2019-07-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p>
<p>Release Date: 2019-07-02</p>
<p>Fix Resolution: 1.0.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in fstream tgz autoclosed cve high severity vulnerability vulnerable library fstream tgz advanced file system stream things library home page a href path to dependency file zendesk magento request package json path to vulnerable library tmp git zendesk magento request node modules npm node modules fstream package json dependency hierarchy semantic release tgz root library npm tgz npm tgz node gyp tgz x fstream tgz vulnerable library found in head commit a href vulnerability details fstream before is vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file the fstream dirwriter function is vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
23,179 | 10,852,345,588 | IssuesEvent | 2019-11-13 12:37:21 | ChenLuigi/TestingPOM | https://api.github.com/repos/ChenLuigi/TestingPOM | opened | CVE-2019-0232 (High) detected in tomcat-catalina-7.0.42.jar | security vulnerability | ## CVE-2019-0232 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-catalina-7.0.42.jar</b></p></summary>
<p>Tomcat Servlet Engine Core Classes and Standard implementations</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/).
<p>Publish Date: 2019-04-15
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232>CVE-2019-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p>
<p>Release Date: 2019-04-15</p>
<p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-0232 (High) detected in tomcat-catalina-7.0.42.jar - ## CVE-2019-0232 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-catalina-7.0.42.jar</b></p></summary>
<p>Tomcat Servlet Engine Core Classes and Standard implementations</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/).
<p>Publish Date: 2019-04-15
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232>CVE-2019-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p>
<p>Release Date: 2019-04-15</p>
<p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tomcat catalina jar cve high severity vulnerability vulnerable library tomcat catalina jar tomcat servlet engine core classes and standard implementations dependency hierarchy x tomcat catalina jar vulnerable library found in head commit a href vulnerability details when running on windows with enablecmdlinearguments enabled the cgi servlet in apache tomcat to to and to is vulnerable to remote code execution due to a bug in the way the jre passes command line arguments to windows the cgi servlet is disabled by default the cgi option enablecmdlinearguments is disable by default in tomcat x and will be disabled by default in all versions in response to this vulnerability for a detailed explanation of the jre behaviour see markus wulftange s blog and this archived msdn blog publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
376,084 | 11,138,211,998 | IssuesEvent | 2019-12-20 21:39:06 | thibaultmeyer/sparrow | https://api.github.com/repos/thibaultmeyer/sparrow | closed | Implementing full-kotlin version of B-encode/B-decode | area/tracker kind/enhancement kind/refactor priority/low | Implementing full-kotlin version of B-encode/B-decode without any other dependencies. | 1.0 | Implementing full-kotlin version of B-encode/B-decode - Implementing full-kotlin version of B-encode/B-decode without any other dependencies. | priority | implementing full kotlin version of b encode b decode implementing full kotlin version of b encode b decode without any other dependencies | 1 |
198,853 | 15,725,894,052 | IssuesEvent | 2021-03-29 10:33:38 | Kong/charts | https://api.github.com/repos/Kong/charts | reopened | [kong] separate README.md into multiple guides | documentation stale | README.md is long and ever-growing. Many sections are somewhat specialized, and README.md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics.
The [ingress controller documentation](https://github.com/Kong/kubernetes-ingress-controller/tree/master/docs) is a good model, where there are several types of documents (references for, well, reference, concepts for high-level explanation, guides for step-by-step instructions for specific tasks) linked from README.md. | 1.0 | [kong] separate README.md into multiple guides - README.md is long and ever-growing. Many sections are somewhat specialized, and README.md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics.
The [ingress controller documentation](https://github.com/Kong/kubernetes-ingress-controller/tree/master/docs) is a good model, where there are several types of documents (references for, well, reference, concepts for high-level explanation, guides for step-by-step instructions for specific tasks) linked from README.md. | non_priority | separate readme md into multiple guides readme md is long and ever growing many sections are somewhat specialized and readme md would benefit from being reduced to a shorter discussion of basic usage with links to other references for details and advanced topics the is a good model where there are several types of documents references for well reference concepts for high level explanation guides for step by step instructions for specific tasks linked from readme md | 0 |
169,040 | 13,111,600,260 | IssuesEvent | 2020-08-04 23:30:27 | Thy-Vipe/BeastsOfBermuda-issues | https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues | closed | [Major Bug] All-female egg glitch | Fixed! Gameplay Major Bug Potential fix public_testing | _Originally written by **The Frog Archer | 76561198201256457**_
Game Version: 1.1.985
*===== System Specs =====
CPU Brand: Intel(R) Core(TM) i7-9700 CPU @ 3.00GHz
Vendor: GenuineIntel
GPU Brand: NVIDIA GeForce GTX 1660
GPU Driver Info: Unknown
Num CPU Cores: 8
===================*
Map: Ancestral_Plains
*Expected Results:* 50/50 split between male and female babies
*Actual Results:* Way to get 100% Female ory
*Replication: Go into an ory egg and before hatching press B. You should always hatch as a female and get the female nesting prompt
Location: X=-147341.609 Y=-8489.048 Z=5590.843
Could be a major balancing glitch with people wanting all female | 1.0 | [Major Bug] All-female egg glitch - _Originally written by **The Frog Archer | 76561198201256457**_
Game Version: 1.1.985
*===== System Specs =====
CPU Brand: Intel(R) Core(TM) i7-9700 CPU @ 3.00GHz
Vendor: GenuineIntel
GPU Brand: NVIDIA GeForce GTX 1660
GPU Driver Info: Unknown
Num CPU Cores: 8
===================*
Map: Ancestral_Plains
*Expected Results:* 50/50 split between male and female babies
*Actual Results:* Way to get 100% Female ory
*Replication: Go into an ory egg and before hatching press B. You should always hatch as a female and get the female nesting prompt
Location: X=-147341.609 Y=-8489.048 Z=5590.843
Could be a major balancing glitch with people wanting all female | non_priority | all female egg glitch originally written by the frog archer game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand nvidia geforce gtx gpu driver info unknown num cpu cores map ancestral plains expected results split between male and female babies actual results way to get female ory replication go into an ory egg and before hatching press b you should always hatch as a female and get the female nesting prompt location x y z could be a major balancing glitch with people wanting all female | 0 |
400,262 | 11,771,312,043 | IssuesEvent | 2020-03-15 23:20:27 | AY1920S2-CS2103T-W17-2/main | https://api.github.com/repos/AY1920S2-CS2103T-W17-2/main | closed | Create StringCorrectionEngine class | priority.High status.Ongoing type.Enhancement | EDITED: Renamed `CommandSuggestionEngine` to `StringCorrectionEngine`
As described in #97 , I'll be implementing the `StringCorrectionEngine` class.
Proposed API (I'm only showing the important methods below):
```java
class StringCorrectionEngine implements CorrectionEngine<String> {
public StringCorrectionEngine(List<String> correctStrings) { }
public CorrectionResult<String> correct(String uncorrected) { }
}
```
#### How to use
1. I'd imagine that you instantiate this class during the creation of both `SuggestionEngine` and `Parser`. During this instantiation phase, you'll supply the list of correct commands as strings (notice the `correctStrings` constructor argument).
2. After that, whenever you receive user input, you'll call the `correct` method. You're guaranteed to receive a corrected result (or error if `uncorrected` is not understandable) | 1.0 | Create StringCorrectionEngine class - EDITED: Renamed `CommandSuggestionEngine` to `StringCorrectionEngine`
As described in #97 , I'll be implementing the `StringCorrectionEngine` class.
Proposed API (I'm only showing the important methods below):
```java
class StringCorrectionEngine implements CorrectionEngine<String> {
public StringCorrectionEngine(List<String> correctStrings) { }
public CorrectionResult<String> correct(String uncorrected) { }
}
```
#### How to use
1. I'd imagine that you instantiate this class during the creation of both `SuggestionEngine` and `Parser`. During this instantiation phase, you'll supply the list of correct commands as strings (notice the `correctStrings` constructor argument).
2. After that, whenever you receive user input, you'll call the `correct` method. You're guaranteed to receive a corrected result (or error if `uncorrected` is not understandable) | priority | create stringcorrectionengine class edited renamed commandsuggestionengine to stringcorrectionengine as described in i ll be implementing the stringcorrectionengine class proposed api i m only showing the important methods below java class stringcorrectionengine implements correctionengine public stringcorrectionengine list correctstrings public correctionresult correct string uncorrected how to use i d imagine that you instantiate this class during the creation of both suggestionengine and parser during this instantiation phase you ll supply the list of correct commands as strings notice the correctstrings constructor argument after that whenever you receive user input you ll call the correct method you re guaranteed to receive a corrected result or error if uncorrected is not understandable | 1 |
26,592 | 11,351,177,493 | IssuesEvent | 2020-01-24 10:35:51 | PeterNgTr/pizza-ui-tests | https://api.github.com/repos/PeterNgTr/pizza-ui-tests | closed | CVE-2018-11762 (Medium) detected in tika-core-1.18.jar | bug security vulnerability | ## CVE-2018-11762 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.18.jar</b></p></summary>
<p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also
includes the core facades for the Tika API.</p>
<p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p>
<p>Path to vulnerable library: /pizza-ui-tests/node_modules/allure-commandline/dist/lib/tika-core-1.18.jar</p>
<p>
Dependency Hierarchy:
- :x: **tika-core-1.18.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PeterNgTr/pizza-ui-tests/commit/b784718100dc8ca278a0aad1c22fba2cd661b638">b784718100dc8ca278a0aad1c22fba2cd661b638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Tika 0.9 to 1.18, in a rare edge case where a user does not specify an extract directory on the commandline (--extract-dir=) and the input file has an embedded file with an absolute path, such as "C:/evil.bat", tika-app would overwrite that file.
<p>Publish Date: 2018-09-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11762>CVE-2018-11762</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762</a></p>
<p>Release Date: 2018-09-19</p>
<p>Fix Resolution: 1.19</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-11762 (Medium) detected in tika-core-1.18.jar - ## CVE-2018-11762 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tika-core-1.18.jar</b></p></summary>
<p>This is the core Apache Tika™ toolkit library from which all other modules inherit functionality. It also
includes the core facades for the Tika API.</p>
<p>Library home page: <a href="http://tika.apache.org/">http://tika.apache.org/</a></p>
<p>Path to vulnerable library: /pizza-ui-tests/node_modules/allure-commandline/dist/lib/tika-core-1.18.jar</p>
<p>
Dependency Hierarchy:
- :x: **tika-core-1.18.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PeterNgTr/pizza-ui-tests/commit/b784718100dc8ca278a0aad1c22fba2cd661b638">b784718100dc8ca278a0aad1c22fba2cd661b638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Tika 0.9 to 1.18, in a rare edge case where a user does not specify an extract directory on the commandline (--extract-dir=) and the input file has an embedded file with an absolute path, such as "C:/evil.bat", tika-app would overwrite that file.
<p>Publish Date: 2018-09-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11762>CVE-2018-11762</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11762</a></p>
<p>Release Date: 2018-09-19</p>
<p>Fix Resolution: 1.19</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in tika core jar cve medium severity vulnerability vulnerable library tika core jar this is the core apache tika™ toolkit library from which all other modules inherit functionality it also includes the core facades for the tika api library home page a href path to vulnerable library pizza ui tests node modules allure commandline dist lib tika core jar dependency hierarchy x tika core jar vulnerable library found in head commit a href vulnerability details in apache tika to in a rare edge case where a user does not specify an extract directory on the commandline extract dir and the input file has an embedded file with an absolute path such as c evil bat tika app would overwrite that file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
327,245 | 24,124,696,292 | IssuesEvent | 2022-09-20 22:26:34 | cloudflare/cloudflare-docs | https://api.github.com/repos/cloudflare/cloudflare-docs | opened | [R2] Workers Runtime APIs layout & formatting | documentation content:edit | ### Which Cloudflare product does this pertain to?
R2
### Existing documentation URL(s)
- https://developers.cloudflare.com/r2/data-access/workers-api/workers-api-reference/
### Section that requires update
Layout and formatting.
### What needs to change?
The current layout, whilst satisfactory for a type reference, will struggle to scale as the APIs gain more options and require more explanations/walkthroughs on how to properly utilise them.
As an example, the addition of the new checksums supported in `R2PutOptions` means that either:
1) Every checksum is listed on it's own, with an aside that only one can be used at a time.
2) The type for the 'checksum' in `R2PutOptions` becomes `md5 | sha1 | sha256 | sha384 | sha512`
Either way, it becomes very verbose and will easily confuse people.
### How should it change?
Taking inspiration from well-regarded documentation layouts such as `docs.rs` (i.e https://doc.rust-lang.org/stable/std/collections/struct.HashMap.html), we should come up with a consistent structure on documenting Worker's Runtime APIs that we can then extend to refactoring the pages for KV, Durable Objects and the like.
We want a good middle-ground between providing a reference, and providing a how-to, without going too in-depth that the material would be better split off into it's own how-to page.
As an example of one potential layout that was discussed by myself and @Frederik-Baetens:
<img width="643" alt="image" src="https://user-images.githubusercontent.com/94662631/191375548-9e5244fe-85be-4ca1-929a-4af793699941.png">
Using the [<details>](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/details) element, this allows us to:
1) Provide a lot of value without cluttering the rest of the page, as it's expandable.
2) Preserve the 'searchability' of the content, as CTRL+F on the majority of browsers will 'look inside' a `<details>` element.
3) Additionally include small examples, not in-depth or niche enough for their own page, under additional `<details>` elements.
<img width="654" alt="image" src="https://user-images.githubusercontent.com/94662631/191375800-c884bf3c-dade-43a3-8aed-f3e48f1d6b21.png">
### Additional information
Original discussion: https://github.com/cloudflare/cloudflare-docs/pull/5897 | 1.0 | [R2] Workers Runtime APIs layout & formatting - ### Which Cloudflare product does this pertain to?
R2
### Existing documentation URL(s)
- https://developers.cloudflare.com/r2/data-access/workers-api/workers-api-reference/
### Section that requires update
Layout and formatting.
### What needs to change?
The current layout, whilst satisfactory for a type reference, will struggle to scale as the APIs gain more options and require more explanations/walkthroughs on how to properly utilise them.
As an example, the addition of the new checksums supported in `R2PutOptions` means that either:
1) Every checksum is listed on it's own, with an aside that only one can be used at a time.
2) The type for the 'checksum' in `R2PutOptions` becomes `md5 | sha1 | sha256 | sha384 | sha512`
Either way, it becomes very verbose and will easily confuse people.
### How should it change?
Taking inspiration from well-regarded documentation layouts such as `docs.rs` (i.e https://doc.rust-lang.org/stable/std/collections/struct.HashMap.html), we should come up with a consistent structure on documenting Worker's Runtime APIs that we can then extend to refactoring the pages for KV, Durable Objects and the like.
We want a good middle-ground between providing a reference, and providing a how-to, without going too in-depth that the material would be better split off into it's own how-to page.
As an example of one potential layout that was discussed by myself and @Frederik-Baetens:
<img width="643" alt="image" src="https://user-images.githubusercontent.com/94662631/191375548-9e5244fe-85be-4ca1-929a-4af793699941.png">
Using the [<details>](https://developer.mozilla.org/en-US/docs/Web/HTML/Element/details) element, this allows us to:
1) Provide a lot of value without cluttering the rest of the page, as it's expandable.
2) Preserve the 'searchability' of the content, as CTRL+F on the majority of browsers will 'look inside' a `<details>` element.
3) Additionally include small examples, not in-depth or niche enough for their own page, under additional `<details>` elements.
<img width="654" alt="image" src="https://user-images.githubusercontent.com/94662631/191375800-c884bf3c-dade-43a3-8aed-f3e48f1d6b21.png">
### Additional information
Original discussion: https://github.com/cloudflare/cloudflare-docs/pull/5897 | non_priority | workers runtime apis layout formatting which cloudflare product does this pertain to existing documentation url s section that requires update layout and formatting what needs to change the current layout whilst satisfactory for a type reference will struggle to scale as the apis gain more options and require more explanations walkthroughs on how to properly utilise them as an example the addition of the new checksums supported in means that either every checksum is listed on it s own with an aside that only one can be used at a time the type for the checksum in becomes either way it becomes very verbose and will easily confuse people how should it change taking inspiration from well regarded documentation layouts such as docs rs i e we should come up with a consistent structure on documenting worker s runtime apis that we can then extend to refactoring the pages for kv durable objects and the like we want a good middle ground between providing a reference and providing a how to without going too in depth that the material would be better split off into it s own how to page as an example of one potential layout that was discussed by myself and frederik baetens img width alt image src using the element this allows us to provide a lot of value without cluttering the rest of the page as it s expandable preserve the searchability of the content as ctrl f on the majority of browsers will look inside a element additionally include small examples not in depth or niche enough for their own page under additional elements img width alt image src additional information original discussion | 0 |
649,873 | 21,328,831,080 | IssuesEvent | 2022-04-18 04:57:14 | googleapis/java-spanner | https://api.github.com/repos/googleapis/java-spanner | closed | The build failed | type: bug priority: p1 api: spanner flakybot: issue | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7465d23f-64d4-4e6e-b1f1-43af7e97fca6), [Sponge](http://sponge2/7465d23f-64d4-4e6e-b1f1-43af7e97fca6)
status: failed
<details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100)
at com.google.cloud.spanner.it.ITCommitTimestampTest.setUp(ITCommitTimestampTest.java:69)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91)
at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66)
at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137)
... 35 more
Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110)
at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350)
at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239)
at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234)
at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222)
at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.addListener(AbstractFuture.java:761)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.addListener(FluentFuture.java:114)
at com.google.common.util.concurrent.ForwardingListenableFuture.addListener(ForwardingListenableFuture.java:47)
at com.google.api.core.ApiFutureToListenableFuture.addListener(ApiFutureToListenableFuture.java:51)
at com.google.common.util.concurrent.AbstractCatchingFuture.create(AbstractCatchingFuture.java:46)
at com.google.common.util.concurrent.Futures.catching(Futures.java:306)
at com.google.api.core.ApiFutures.catching(ApiFutures.java:98)
at com.google.api.gax.longrunning.OperationFutureImpl.<init>(OperationFutureImpl.java:95)
at com.google.cloud.spanner.DatabaseAdminClientImpl.createDatabase(DatabaseAdminClientImpl.java:348)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:136)
... 35 more
Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at io.grpc.Status.asRuntimeException(Status.java:535)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100)
at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463)
at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427)
at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
</pre></details> | 1.0 | The build failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7465d23f-64d4-4e6e-b1f1-43af7e97fca6), [Sponge](http://sponge2/7465d23f-64d4-4e6e-b1f1-43af7e97fca6)
status: failed
<details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100)
at com.google.cloud.spanner.it.ITCommitTimestampTest.setUp(ITCommitTimestampTest.java:69)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:547)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91)
at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66)
at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137)
... 35 more
Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110)
at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350)
at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239)
at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234)
at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222)
at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.addListener(AbstractFuture.java:761)
at com.google.common.util.concurrent.FluentFuture$TrustedFuture.addListener(FluentFuture.java:114)
at com.google.common.util.concurrent.ForwardingListenableFuture.addListener(ForwardingListenableFuture.java:47)
at com.google.api.core.ApiFutureToListenableFuture.addListener(ApiFutureToListenableFuture.java:51)
at com.google.common.util.concurrent.AbstractCatchingFuture.create(AbstractCatchingFuture.java:46)
at com.google.common.util.concurrent.Futures.catching(Futures.java:306)
at com.google.api.core.ApiFutures.catching(ApiFutures.java:98)
at com.google.api.gax.longrunning.OperationFutureImpl.<init>(OperationFutureImpl.java:95)
at com.google.cloud.spanner.DatabaseAdminClientImpl.createDatabase(DatabaseAdminClientImpl.java:348)
at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:136)
... 35 more
Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_841467678_0015' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance.
at io.grpc.Status.asRuntimeException(Status.java:535)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100)
at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463)
at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427)
at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
</pre></details> | priority | the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output com google cloud spanner spannerexception resource exhausted com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner it itcommittimestamptest setup itcommittimestamptest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java util concurrent executionexception com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java more caused by com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory fromapiexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner databaseadminclientimpl lambda createdatabase databaseadminclientimpl java at com google api core apifutures apifunctiontoguavafunction apply apifutures java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture run abstractcatchingfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture addlistener abstractfuture java at com google common util concurrent fluentfuture trustedfuture addlistener fluentfuture java at com google common util concurrent forwardinglistenablefuture addlistener forwardinglistenablefuture java at com google api core apifuturetolistenablefuture addlistener apifuturetolistenablefuture java at com google common util concurrent abstractcatchingfuture create abstractcatchingfuture java at com google common util concurrent futures catching futures java at com google api core apifutures catching apifutures java at com google api gax longrunning operationfutureimpl operationfutureimpl java at com google cloud spanner databaseadminclientimpl createdatabase databaseadminclientimpl java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java more caused by io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at io grpc status asruntimeexception status java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google cloud spanner spi spannererrorinterceptor onclose spannererrorinterceptor java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java base java util concurrent executors runnableadapter call executors java at java base java util concurrent futuretask run futuretask java at java base java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java | 1 |
400,604 | 11,777,613,859 | IssuesEvent | 2020-03-16 15:04:35 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | admin.microsoft.com - site is not usable | browser-firefox engine-gecko priority-critical | <!-- @browser: Firefox 74.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 -->
<!-- @reported_with: addon-reporter-firefox -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50285 -->
**URL**: https://admin.microsoft.com/AdminPortal/Home#/homepage
**Browser / Version**: Firefox 74.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Site refuses to load
**Steps to Reproduce**:
I'm attempting to use the 365 administrator portal but recently it has stopped working in Firefox. I've disabled all blocking or privacy extensions and all I'm getting is a white screen.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/87407f95-677c-40ea-b9dd-5ab0c0d34505.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | admin.microsoft.com - site is not usable - <!-- @browser: Firefox 74.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 -->
<!-- @reported_with: addon-reporter-firefox -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50285 -->
**URL**: https://admin.microsoft.com/AdminPortal/Home#/homepage
**Browser / Version**: Firefox 74.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Site refuses to load
**Steps to Reproduce**:
I'm attempting to use the 365 administrator portal but recently it has stopped working in Firefox. I've disabled all blocking or privacy extensions and all I'm getting is a white screen.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/87407f95-677c-40ea-b9dd-5ab0c0d34505.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | admin microsoft com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description site refuses to load steps to reproduce i m attempting to use the administrator portal but recently it has stopped working in firefox i ve disabled all blocking or privacy extensions and all i m getting is a white screen view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
256,342 | 8,127,347,823 | IssuesEvent | 2018-08-17 07:43:52 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Fix Manta stuff in Molecule plot's CMakeLists.txt | Bug Likelihood: 3 - Occasional Priority: Normal Severity: 2 - Minor Irritation | The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build.
I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1418
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Fix Manta stuff in Molecule plot's CMakeLists.txt
Assigned to: Kathleen Biagas
Category:
Target version: 2.7
Author: Brad Whitlock
Start: 04/10/2013
Due date:
% Done: 0
Estimated time:
Created: 04/10/2013 05:23 pm
Updated: 10/09/2013 05:08 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build.
I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file.
Comments:
Added ability for plugins to use codefiles for conditional logic that is to be added to cmake files.For Molecule plot, the logic is:SVN Revisions 22043 (2.7RC) 22045 (trunk)
| 1.0 | Fix Manta stuff in Molecule plot's CMakeLists.txt - The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build.
I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1418
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Fix Manta stuff in Molecule plot's CMakeLists.txt
Assigned to: Kathleen Biagas
Category:
Target version: 2.7
Author: Brad Whitlock
Start: 04/10/2013
Due date:
% Done: 0
Estimated time:
Created: 04/10/2013 05:23 pm
Updated: 10/09/2013 05:08 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The Molecule plot's CMakeLists.txt file cannot be automatically regenerated because of some Manta-related coding. If VisIt is being built with Manta then some additional source files are added to the plugin's build.
I suggest that the Manta sources always get added so the CMakeLists.txt file can be autogenerated. Then some conditional compilation in the Manta source files can determine whether the file actually produces anything in its object file.
Comments:
Added ability for plugins to use codefiles for conditional logic that is to be added to cmake files.For Molecule plot, the logic is:SVN Revisions 22043 (2.7RC) 22045 (trunk)
| priority | fix manta stuff in molecule plot s cmakelists txt the molecule plot s cmakelists txt file cannot be automatically regenerated because of some manta related coding if visit is being built with manta then some additional source files are added to the plugin s build i suggest that the manta sources always get added so the cmakelists txt file can be autogenerated then some conditional compilation in the manta source files can determine whether the file actually produces anything in its object file redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject fix manta stuff in molecule plot s cmakelists txt assigned to kathleen biagas category target version author brad whitlock start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description the molecule plot s cmakelists txt file cannot be automatically regenerated because of some manta related coding if visit is being built with manta then some additional source files are added to the plugin s build i suggest that the manta sources always get added so the cmakelists txt file can be autogenerated then some conditional compilation in the manta source files can determine whether the file actually produces anything in its object file comments added ability for plugins to use codefiles for conditional logic that is to be added to cmake files for molecule plot the logic is svn revisions trunk | 1 |
346,304 | 30,883,363,599 | IssuesEvent | 2023-08-03 19:28:36 | ray-project/ray | https://api.github.com/repos/ray-project/ray | opened | Release test rllib_stress_tests.aws failed | bug P0 rllib release-test | Release test **rllib_stress_tests.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/2019#0189bcd7-94cc-4087-bb27-d79bdde9aa23 for more details.
Managed by OSS Test Policy | 1.0 | Release test rllib_stress_tests.aws failed - Release test **rllib_stress_tests.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/2019#0189bcd7-94cc-4087-bb27-d79bdde9aa23 for more details.
Managed by OSS Test Policy | non_priority | release test rllib stress tests aws failed release test rllib stress tests aws failed see for more details managed by oss test policy | 0 |
67,834 | 17,085,376,328 | IssuesEvent | 2021-07-08 11:06:55 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | opened | [Bug] Multiple bindings don't work inside the child of the list | Bug List Widget UI Building UI Building Pod Widgets | Multiple bindings inside the list widget don't work.
| 2.0 | [Bug] Multiple bindings don't work inside the child of the list - Multiple bindings inside the list widget don't work.
| non_priority | multiple bindings don t work inside the child of the list multiple bindings inside the list widget don t work | 0 |
60,046 | 17,023,320,551 | IssuesEvent | 2021-07-03 01:25:01 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | mapnik ignores layer-tags of rivers | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 2.03pm, Wednesday, 12th November 2008]**
Mapnik seems to ignore the "layer=*"-tags of rivers. [http://www.openstreetmap.org/?lat=48.051367&lon=8.208524&zoom=18&layers=B000FTF Here] is an example of this behavior. The river is tagged with layer=-1, the building is tagged with layer=1 (and i just added an area highway=pedestrian on level 0 and put a section of the river into a tunnel, but that was of course too late for the weekly dump, so you won't see it in mapnik now). | 1.0 | mapnik ignores layer-tags of rivers - **[Submitted to the original trac issue database at 2.03pm, Wednesday, 12th November 2008]**
Mapnik seems to ignore the "layer=*"-tags of rivers. [http://www.openstreetmap.org/?lat=48.051367&lon=8.208524&zoom=18&layers=B000FTF Here] is an example of this behavior. The river is tagged with layer=-1, the building is tagged with layer=1 (and i just added an area highway=pedestrian on level 0 and put a section of the river into a tunnel, but that was of course too late for the weekly dump, so you won't see it in mapnik now). | non_priority | mapnik ignores layer tags of rivers mapnik seems to ignore the layer tags of rivers is an example of this behavior the river is tagged with layer the building is tagged with layer and i just added an area highway pedestrian on level and put a section of the river into a tunnel but that was of course too late for the weekly dump so you won t see it in mapnik now | 0 |
65,608 | 8,823,269,928 | IssuesEvent | 2019-01-02 12:58:33 | sinnerschrader/feature-hub | https://api.github.com/repos/sinnerschrader/feature-hub | closed | Website: Document integrator consumer ID usage | documentation | - [x] Will be closed by PR #208
How to deal with multiple integrators on one host? (e.g. LocalStorage) | 1.0 | Website: Document integrator consumer ID usage - - [x] Will be closed by PR #208
How to deal with multiple integrators on one host? (e.g. LocalStorage) | non_priority | website document integrator consumer id usage will be closed by pr how to deal with multiple integrators on one host e g localstorage | 0 |
366,428 | 10,820,664,131 | IssuesEvent | 2019-11-08 16:51:23 | AY1920S1-CS2113T-W17-4/main | https://api.github.com/repos/AY1920S1-CS2113T-W17-4/main | closed | As a Computing student, I can add my estimated time taken to complete a task | priority.Low type.Story | so that I know how much free time I would have.
| 1.0 | As a Computing student, I can add my estimated time taken to complete a task - so that I know how much free time I would have.
| priority | as a computing student i can add my estimated time taken to complete a task so that i know how much free time i would have | 1 |
807,507 | 30,006,071,349 | IssuesEvent | 2023-06-26 12:32:37 | GoogleCloudPlatform/cloud-sql-python-connector | https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-python-connector | closed | system.test_pytds_connection: test_pooled_connection_with_pytds failed | type: bug priority: p2 flakybot: issue flakybot: flaky | Note: #672 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: fe64fb37d98efa76d8027077d3a7e52bf1e7a15d
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/5367360775
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="pool")
def setup() -> Generator:
pool = init_connection_engine()
> with pool.connect() as conn:
tests/system/test_pytds_connection.py:65:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3264: in connect
return self._connection_cls(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:145: in __init__
self._dbapi_connection = engine.raw_connection()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3288: in raw_connection
return self.pool.connect()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:452: in connect
return _ConnectionFairy._checkout(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:1267: in _checkout
fairy = _ConnectionRecord.checkout(pool)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:716: in checkout
rec = pool._do_get()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:169: in _do_get
with util.safe_reraise():
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__
raise exc_value.with_traceback(exc_tb)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:167: in _do_get
return self._create_connection()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:393: in _create_connection
return _ConnectionRecord(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:678: in __init__
self.__connect()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:902: in __connect
with util.safe_reraise():
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__
raise exc_value.with_traceback(exc_tb)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:898: in __connect
self.dbapi_connection = connection = pool._invoke_creator(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:365: in <lambda>
return lambda rec: creator_fn()
tests/system/test_pytds_connection.py:40: in getconn
conn = connector.connect(
google/cloud/sql/connector/connector.py:159: in connect
return connect_task.result()
/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:456: in result
return self.__get_result()
/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:401: in __get_result
raise self._exception
google/cloud/sql/connector/connector.py:204: in connect_async
instance = Instance(
google/cloud/sql/connector/instance.py:253: in __init__
self._credentials = _auth_init(credentials)
google/cloud/sql/connector/utils.py:124: in _auth_init
credentials, _ = default(scopes=scopes)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/_default.py:675: in default
project_id = credentials.get_project_id(request=request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:334: in get_project_id
self.before_request(request, "GET", url, headers)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/credentials.py:151: in before_request
self.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:355: in refresh
self._impersonated_credentials.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:247: in refresh
self._update_token(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:260: in _update_token
self._source_credentials.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:368: in refresh
subject_token=self.retrieve_subject_token(request),
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:158: in retrieve_subject_token
self._get_token_data(request),
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:167: in _get_token_data
return self._get_url_data(
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.auth.identity_pool.Credentials object at 0x10d3ff3d0>
request = <google.auth.transport.requests.Request object at 0x10d3fde50>
url = 'https://pipelines.actions.githubusercontent.com/umAmnh0OhcfbtGEt7J16Yga6HsgM8dYIhPxbPiOYFLVwMnfbKz/00000000-0000-0000...cations%2Fglobal%2FworkloadIdentityPools%2Fgh-13a715-cloud-sql-pyt-dd1c5f%2Fproviders%2Fgh-13a715-cloud-sql-pyt-dd1c5f'
headers = {'Authorization': 'Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6ImVCWl9jbjNzWFlBZDBjaDRUSEJLSElnT3dPRSJ9.eyJuYW1...rQ4yUX2hGS2UBW_MgcvJ0VQpGlt1nFMDC8DzhKGNCN6fDKQB8yJA2bq-ddCqnrOfrB_fFwkBEBhlh8BWKlZX8A8X2bNT8M9MOh0o3Xu4Okb895WQV9zRA'}
def _get_url_data(self, request, url, headers):
response = request(url=url, method="GET", headers=headers)
# support both string and bytes type response.data
response_body = (
response.data.decode("utf-8")
if hasattr(response.data, "decode")
else response.data
)
if response.status != 200:
> raise exceptions.RefreshError(
"Unable to retrieve Identity Pool subject token", response_body
)
E google.auth.exceptions.RefreshError: ('Unable to retrieve Identity Pool subject token', '{ "message": "GitHub Actions is temporarily unavailable. Please visit https://www.githubstatus.com/ for the status of our services.", "ref": "Ref A: F3B1C190713A4756A6359D9977F85EDD Ref B: CO1EDGE2015 Ref C: 2023-06-25T02:34:36Z" }')
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:189: RefreshError</pre></details> | 1.0 | system.test_pytds_connection: test_pooled_connection_with_pytds failed - Note: #672 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: fe64fb37d98efa76d8027077d3a7e52bf1e7a15d
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/5367360775
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="pool")
def setup() -> Generator:
pool = init_connection_engine()
> with pool.connect() as conn:
tests/system/test_pytds_connection.py:65:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3264: in connect
return self._connection_cls(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:145: in __init__
self._dbapi_connection = engine.raw_connection()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/engine/base.py:3288: in raw_connection
return self.pool.connect()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:452: in connect
return _ConnectionFairy._checkout(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:1267: in _checkout
fairy = _ConnectionRecord.checkout(pool)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:716: in checkout
rec = pool._do_get()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:169: in _do_get
with util.safe_reraise():
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__
raise exc_value.with_traceback(exc_tb)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/impl.py:167: in _do_get
return self._create_connection()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:393: in _create_connection
return _ConnectionRecord(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:678: in __init__
self.__connect()
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:902: in __connect
with util.safe_reraise():
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/util/langhelpers.py:147: in __exit__
raise exc_value.with_traceback(exc_tb)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:898: in __connect
self.dbapi_connection = connection = pool._invoke_creator(self)
.nox/system-3-11/lib/python3.11/site-packages/sqlalchemy/pool/base.py:365: in <lambda>
return lambda rec: creator_fn()
tests/system/test_pytds_connection.py:40: in getconn
conn = connector.connect(
google/cloud/sql/connector/connector.py:159: in connect
return connect_task.result()
/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:456: in result
return self.__get_result()
/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/concurrent/futures/_base.py:401: in __get_result
raise self._exception
google/cloud/sql/connector/connector.py:204: in connect_async
instance = Instance(
google/cloud/sql/connector/instance.py:253: in __init__
self._credentials = _auth_init(credentials)
google/cloud/sql/connector/utils.py:124: in _auth_init
credentials, _ = default(scopes=scopes)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/_default.py:675: in default
project_id = credentials.get_project_id(request=request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:334: in get_project_id
self.before_request(request, "GET", url, headers)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/credentials.py:151: in before_request
self.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:355: in refresh
self._impersonated_credentials.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:247: in refresh
self._update_token(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/impersonated_credentials.py:260: in _update_token
self._source_credentials.refresh(request)
.nox/system-3-11/lib/python3.11/site-packages/google/auth/external_account.py:368: in refresh
subject_token=self.retrieve_subject_token(request),
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:158: in retrieve_subject_token
self._get_token_data(request),
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:167: in _get_token_data
return self._get_url_data(
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.auth.identity_pool.Credentials object at 0x10d3ff3d0>
request = <google.auth.transport.requests.Request object at 0x10d3fde50>
url = 'https://pipelines.actions.githubusercontent.com/umAmnh0OhcfbtGEt7J16Yga6HsgM8dYIhPxbPiOYFLVwMnfbKz/00000000-0000-0000...cations%2Fglobal%2FworkloadIdentityPools%2Fgh-13a715-cloud-sql-pyt-dd1c5f%2Fproviders%2Fgh-13a715-cloud-sql-pyt-dd1c5f'
headers = {'Authorization': 'Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6ImVCWl9jbjNzWFlBZDBjaDRUSEJLSElnT3dPRSJ9.eyJuYW1...rQ4yUX2hGS2UBW_MgcvJ0VQpGlt1nFMDC8DzhKGNCN6fDKQB8yJA2bq-ddCqnrOfrB_fFwkBEBhlh8BWKlZX8A8X2bNT8M9MOh0o3Xu4Okb895WQV9zRA'}
def _get_url_data(self, request, url, headers):
response = request(url=url, method="GET", headers=headers)
# support both string and bytes type response.data
response_body = (
response.data.decode("utf-8")
if hasattr(response.data, "decode")
else response.data
)
if response.status != 200:
> raise exceptions.RefreshError(
"Unable to retrieve Identity Pool subject token", response_body
)
E google.auth.exceptions.RefreshError: ('Unable to retrieve Identity Pool subject token', '{ "message": "GitHub Actions is temporarily unavailable. Please visit https://www.githubstatus.com/ for the status of our services.", "ref": "Ref A: F3B1C190713A4756A6359D9977F85EDD Ref B: CO1EDGE2015 Ref C: 2023-06-25T02:34:36Z" }')
.nox/system-3-11/lib/python3.11/site-packages/google/auth/identity_pool.py:189: RefreshError</pre></details> | priority | system test pytds connection test pooled connection with pytds failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output pytest fixture name pool def setup generator pool init connection engine with pool connect as conn tests system test pytds connection py nox system lib site packages sqlalchemy engine base py in connect return self connection cls self nox system lib site packages sqlalchemy engine base py in init self dbapi connection engine raw connection nox system lib site packages sqlalchemy engine base py in raw connection return self pool connect nox system lib site packages sqlalchemy pool base py in connect return connectionfairy checkout self nox system lib site packages sqlalchemy pool base py in checkout fairy connectionrecord checkout pool nox system lib site packages sqlalchemy pool base py in checkout rec pool do get nox system lib site packages sqlalchemy pool impl py in do get with util safe reraise nox system lib site packages sqlalchemy util langhelpers py in exit raise exc value with traceback exc tb nox system lib site packages sqlalchemy pool impl py in do get return self create connection nox system lib site packages sqlalchemy pool base py in create connection return connectionrecord self nox system lib site packages sqlalchemy pool base py in init self connect nox system lib site packages sqlalchemy pool base py in connect with util safe reraise nox system lib site packages sqlalchemy util langhelpers py in exit raise exc value with traceback exc tb nox system lib site packages sqlalchemy pool base py in connect self dbapi connection connection pool invoke creator self nox system lib site packages sqlalchemy pool base py in return lambda rec creator fn tests system test pytds connection py in getconn conn connector connect google cloud sql connector connector py in connect return connect task result library frameworks python framework versions lib concurrent futures base py in result return self get result library frameworks python framework versions lib concurrent futures base py in get result raise self exception google cloud sql connector connector py in connect async instance instance google cloud sql connector instance py in init self credentials auth init credentials google cloud sql connector utils py in auth init credentials default scopes scopes nox system lib site packages google auth default py in default project id credentials get project id request request nox system lib site packages google auth external account py in get project id self before request request get url headers nox system lib site packages google auth credentials py in before request self refresh request nox system lib site packages google auth external account py in refresh self impersonated credentials refresh request nox system lib site packages google auth impersonated credentials py in refresh self update token request nox system lib site packages google auth impersonated credentials py in update token self source credentials refresh request nox system lib site packages google auth external account py in refresh subject token self retrieve subject token request nox system lib site packages google auth identity pool py in retrieve subject token self get token data request nox system lib site packages google auth identity pool py in get token data return self get url data self request url headers authorization bearer ddcqnrofrb def get url data self request url headers response request url url method get headers headers support both string and bytes type response data response body response data decode utf if hasattr response data decode else response data if response status raise exceptions refresherror unable to retrieve identity pool subject token response body e google auth exceptions refresherror unable to retrieve identity pool subject token message github actions is temporarily unavailable please visit for the status of our services ref ref a ref b ref c nox system lib site packages google auth identity pool py refresherror | 1 |
561,035 | 16,609,840,751 | IssuesEvent | 2021-06-02 10:05:36 | svthalia/Reaxit | https://api.github.com/repos/svthalia/Reaxit | opened | Pretty splash screens | feature priority: high | We currently have a simple image as a splashcreen with no transition to the first frame. We should at least make the image (esp. on android) nicer, and it would also be nice to fade to the first frame.
| 1.0 | Pretty splash screens - We currently have a simple image as a splashcreen with no transition to the first frame. We should at least make the image (esp. on android) nicer, and it would also be nice to fade to the first frame.
| priority | pretty splash screens we currently have a simple image as a splashcreen with no transition to the first frame we should at least make the image esp on android nicer and it would also be nice to fade to the first frame | 1 |
58,025 | 3,087,082,880 | IssuesEvent | 2015-08-25 09:15:34 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | Индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении | bug imported Priority-Medium | _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on October 01, 2013 02:05:31_
1. Подключиться к новому хабу на который вас не пускают по нику или тэгу, по размеру шары, либо по количеству слотов.
2. добавить хаб в избранные.
3. изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили, закрыть окно настроек.
4. дождаться автоматического переподключения к хабу.
5. хаб не пускает, т.к. ваши изменения не актуализировались.
6. нажать кнопку переподключения к хабу.
7. новые настройки вступают в силу и подключение происходит.
FlylinkDC++ r502 -rc2-x64 build 15562 Compiled on: 2013-09-28
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1322_ | 1.0 | Индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении - _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on October 01, 2013 02:05:31_
1. Подключиться к новому хабу на который вас не пускают по нику или тэгу, по размеру шары, либо по количеству слотов.
2. добавить хаб в избранные.
3. изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили, закрыть окно настроек.
4. дождаться автоматического переподключения к хабу.
5. хаб не пускает, т.к. ваши изменения не актуализировались.
6. нажать кнопку переподключения к хабу.
7. новые настройки вступают в силу и подключение происходит.
FlylinkDC++ r502 -rc2-x64 build 15562 Compiled on: 2013-09-28
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1322_ | priority | индивидуальные настройки хабов не актуализируются при автоматическом пепеподключении from on october подключиться к новому хабу на который вас не пускают по нику или тэгу по размеру шары либо по количеству слотов добавить хаб в избранные изменить индивидуальные настройки хаба по правой кнопке на вкладке так чтобы на хаб пустили закрыть окно настроек дождаться автоматического переподключения к хабу хаб не пускает т к ваши изменения не актуализировались нажать кнопку переподключения к хабу новые настройки вступают в силу и подключение происходит flylinkdc build compiled on original issue | 1 |
245,822 | 18,795,361,614 | IssuesEvent | 2021-11-08 21:35:54 | 4kills/go-zlib | https://api.github.com/repos/4kills/go-zlib | closed | There is no `ReadBytes` method | documentation | README.md shows an alternative method for decompressing data using `ReadBytes()` method. Was this method removed? | 1.0 | There is no `ReadBytes` method - README.md shows an alternative method for decompressing data using `ReadBytes()` method. Was this method removed? | non_priority | there is no readbytes method readme md shows an alternative method for decompressing data using readbytes method was this method removed | 0 |
552,105 | 16,195,203,499 | IssuesEvent | 2021-05-04 13:49:41 | Edgeryders-Participio/multi-dreams | https://api.github.com/repos/Edgeryders-Participio/multi-dreams | closed | Require the user to approve the ToS and privacy policy on sign up | Priority: 1 (now - within 1 month) | We need to implement approving our TOS.
https://www.iubenda.com/terms-and-conditions/58637640 | 1.0 | Require the user to approve the ToS and privacy policy on sign up - We need to implement approving our TOS.
https://www.iubenda.com/terms-and-conditions/58637640 | priority | require the user to approve the tos and privacy policy on sign up we need to implement approving our tos | 1 |
346,470 | 10,412,685,756 | IssuesEvent | 2019-09-13 16:33:49 | Sp2000/colplus-repo | https://api.github.com/repos/Sp2000/colplus-repo | opened | Fix split-subgenera data bugs | bug high priority | I found another case where a species is placed on the classification in the wrong location by CoL+. This is similar to the [inc. sed. bug](https://github.com/Sp2000/colplus-backend/issues/477), but in this case it's caused by [split-subgenera](https://www.col.plus/dataset/1106/duplicates?category=uninomial&limit=50&minSize=2&mode=STRICT&offset=0&rank=subgenus&withDecision=false).
Here's the output from the diff_key generator:
diff_key | hierarchy | p8 | p7 | p6 | p5 | p4 | p3 | p2 | p1 | name | hash | url
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
685deb96725d46c6ff6cd24d9c7158ab | Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | | Animalia | Echinodermata | Echinoidea | Spatangoida | Brissidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | 0a01d690cc0ad948cebb7267ef98854b | http://workbench.catalogueoflife.org/col_plus/details/species/id/0a01d690cc0ad948cebb7267ef98854b
ecf1add23e3f3abae62addddd00674a1 | Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | Animalia | Echinodermata | Echinoidea | Spatangoida | Spatangoidea | Eupatagidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | d81d04e18a36e88d32021bfcc5025332 | http://www.catalogueoflife.org/annual-checklist/2019/details/species/id/d81d04e18a36e88d32021bfcc5025332
Here's the raw data from the ACEF 106.tar.gz file:
```
"755133" "Animalia" "Echinodermata" "Echinoidea" "Spatangoida" "Spatangoidea" "Eupatagidae" "Brissoides" "Rhabdobrissus" "aloysii" "Pijpers, 1933" "" "accepted name" "1" "1" "0" "marine" "" "Kroh, Andreas" "2014-03-05" "http://www.marinespecies.org/echinoidea/aphia.php?p=taxdetails&id=755133" "" "urn:lsid:marinespecies.org:taxname:755133"
```
[CoL+ places it incorrectly in Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae](https://www.col.plus/dataset/1106/taxon/755133) instead of Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae.
@mdoering, this is a data quality issue that we'd have to report to the GSD and block one of the 2 from appearing in CoL+, but also might commonly happen so placing split subgenera in the classification correctly might still be important. | 1.0 | Fix split-subgenera data bugs - I found another case where a species is placed on the classification in the wrong location by CoL+. This is similar to the [inc. sed. bug](https://github.com/Sp2000/colplus-backend/issues/477), but in this case it's caused by [split-subgenera](https://www.col.plus/dataset/1106/duplicates?category=uninomial&limit=50&minSize=2&mode=STRICT&offset=0&rank=subgenus&withDecision=false).
Here's the output from the diff_key generator:
diff_key | hierarchy | p8 | p7 | p6 | p5 | p4 | p3 | p2 | p1 | name | hash | url
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
685deb96725d46c6ff6cd24d9c7158ab | Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | | Animalia | Echinodermata | Echinoidea | Spatangoida | Brissidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | 0a01d690cc0ad948cebb7267ef98854b | http://workbench.catalogueoflife.org/col_plus/details/species/id/0a01d690cc0ad948cebb7267ef98854b
ecf1add23e3f3abae62addddd00674a1 | Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae_Brissoides_Brissoides (Rhabdobrissus) aloysii | | Animalia | Echinodermata | Echinoidea | Spatangoida | Spatangoidea | Eupatagidae | Brissoides | Brissoides (Rhabdobrissus) aloysii | d81d04e18a36e88d32021bfcc5025332 | http://www.catalogueoflife.org/annual-checklist/2019/details/species/id/d81d04e18a36e88d32021bfcc5025332
Here's the raw data from the ACEF 106.tar.gz file:
```
"755133" "Animalia" "Echinodermata" "Echinoidea" "Spatangoida" "Spatangoidea" "Eupatagidae" "Brissoides" "Rhabdobrissus" "aloysii" "Pijpers, 1933" "" "accepted name" "1" "1" "0" "marine" "" "Kroh, Andreas" "2014-03-05" "http://www.marinespecies.org/echinoidea/aphia.php?p=taxdetails&id=755133" "" "urn:lsid:marinespecies.org:taxname:755133"
```
[CoL+ places it incorrectly in Animalia_Echinodermata_Echinoidea_Spatangoida_Brissidae](https://www.col.plus/dataset/1106/taxon/755133) instead of Animalia_Echinodermata_Echinoidea_Spatangoida_Spatangoidea_Eupatagidae.
@mdoering, this is a data quality issue that we'd have to report to the GSD and block one of the 2 from appearing in CoL+, but also might commonly happen so placing split subgenera in the classification correctly might still be important. | priority | fix split subgenera data bugs i found another case where a species is placed on the classification in the wrong location by col this is similar to the but in this case it s caused by here s the output from the diff key generator diff key hierarchy name hash url animalia echinodermata echinoidea spatangoida brissidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida brissidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides brissoides rhabdobrissus aloysii animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides brissoides rhabdobrissus aloysii here s the raw data from the acef tar gz file animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae brissoides rhabdobrissus aloysii pijpers accepted name marine kroh andreas urn lsid marinespecies org taxname instead of animalia echinodermata echinoidea spatangoida spatangoidea eupatagidae mdoering this is a data quality issue that we d have to report to the gsd and block one of the from appearing in col but also might commonly happen so placing split subgenera in the classification correctly might still be important | 1 |
103,369 | 11,355,091,898 | IssuesEvent | 2020-01-24 19:12:03 | wisslab/judaicalink-labs | https://api.github.com/repos/wisslab/judaicalink-labs | opened | Documentation for backend app | documentation | The whole backend app with the messaging and task API, as well as the data management itself needs proper documentaion.
Related to #6 | 1.0 | Documentation for backend app - The whole backend app with the messaging and task API, as well as the data management itself needs proper documentaion.
Related to #6 | non_priority | documentation for backend app the whole backend app with the messaging and task api as well as the data management itself needs proper documentaion related to | 0 |
666,076 | 22,341,644,715 | IssuesEvent | 2022-06-15 01:53:56 | Couchers-org/web-frontend | https://api.github.com/repos/Couchers-org/web-frontend | opened | Event join button should join on first click, then show menu next clicks | good first issue improvement priority: normal | Instead of always showing the menu | 1.0 | Event join button should join on first click, then show menu next clicks - Instead of always showing the menu | priority | event join button should join on first click then show menu next clicks instead of always showing the menu | 1 |
58,858 | 3,092,229,573 | IssuesEvent | 2015-08-26 16:45:29 | brian-team/brian2 | https://api.github.com/repos/brian-team/brian2 | reopened | Some pre-synaptic updates give incorrect warnings | bug high priority | I'm using the following simple Synapse:
```
synapses = br2.Synapses(neurons, pre='v -= 0.2')
```
and I get a warning:
```
WARNING brian2.codegen.generators.base: Came across an abstract code block that is not well-defined: the outcome may depend on the order of execution. 2 lines of abstract code, first line is: "v -= w (in-place)"
```
However, if I make the following trivial change:
```
synapses = br2.Synapses(neurons, pre='v += -0.2')
```
The warning disappears. | 1.0 | Some pre-synaptic updates give incorrect warnings - I'm using the following simple Synapse:
```
synapses = br2.Synapses(neurons, pre='v -= 0.2')
```
and I get a warning:
```
WARNING brian2.codegen.generators.base: Came across an abstract code block that is not well-defined: the outcome may depend on the order of execution. 2 lines of abstract code, first line is: "v -= w (in-place)"
```
However, if I make the following trivial change:
```
synapses = br2.Synapses(neurons, pre='v += -0.2')
```
The warning disappears. | priority | some pre synaptic updates give incorrect warnings i m using the following simple synapse synapses synapses neurons pre v and i get a warning warning codegen generators base came across an abstract code block that is not well defined the outcome may depend on the order of execution lines of abstract code first line is v w in place however if i make the following trivial change synapses synapses neurons pre v the warning disappears | 1 |
789,442 | 27,789,952,678 | IssuesEvent | 2023-03-17 08:05:27 | AY2223S2-CS2103T-W12-2/tp | https://api.github.com/repos/AY2223S2-CS2103T-W12-2/tp | opened | Empty string fields in `json` save file | type.Bug priority.High | # The problem
Currently, when a person-entry lacks some optional fields, their saved data in `/data/addressbook.json` looks like this:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"phone" : "",
"email" : "",
"address" : "",
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
```
where the missing fields are empty strings `""`. This means when they're loaded, those fields are initialised to empty string (eg. `new Address("")`) instead of `Optional.empty()`.
And those empty-string fields are currently allowed because we changed the `isValid` methods of the fields to return true when it's an empty string.
```java
/**
* Returns true if a given string is a valid email.
*/
public static boolean isValidAddress(String test) {
return test.matches(VALIDATION_REGEX) || test.isBlank();
}
```
<br>
# Possible solution
We could instead represent missing optional fields as `null` in the `.json` save file like:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"phone" : null,
"email" : null,
"address" : null,
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
```
Or perhaps lose those fields in the `json` file all together:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
``` | 1.0 | Empty string fields in `json` save file - # The problem
Currently, when a person-entry lacks some optional fields, their saved data in `/data/addressbook.json` looks like this:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"phone" : "",
"email" : "",
"address" : "",
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
```
where the missing fields are empty strings `""`. This means when they're loaded, those fields are initialised to empty string (eg. `new Address("")`) instead of `Optional.empty()`.
And those empty-string fields are currently allowed because we changed the `isValid` methods of the fields to return true when it's an empty string.
```java
/**
* Returns true if a given string is a valid email.
*/
public static boolean isValidAddress(String test) {
return test.matches(VALIDATION_REGEX) || test.isBlank();
}
```
<br>
# Possible solution
We could instead represent missing optional fields as `null` in the `.json` save file like:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"phone" : null,
"email" : null,
"address" : null,
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
```
Or perhaps lose those fields in the `json` file all together:
```json
{
"persons" : [ {
"name" : "Netorare Enjoyer",
"remark" : "im a cuck",
"tagged" : [ ]
} ]
}
``` | priority | empty string fields in json save file the problem currently when a person entry lacks some optional fields their saved data in data addressbook json looks like this json persons name netorare enjoyer phone email address remark im a cuck tagged where the missing fields are empty strings this means when they re loaded those fields are initialised to empty string eg new address instead of optional empty and those empty string fields are currently allowed because we changed the isvalid methods of the fields to return true when it s an empty string java returns true if a given string is a valid email public static boolean isvalidaddress string test return test matches validation regex test isblank possible solution we could instead represent missing optional fields as null in the json save file like json persons name netorare enjoyer phone null email null address null remark im a cuck tagged or perhaps lose those fields in the json file all together json persons name netorare enjoyer remark im a cuck tagged | 1 |
145,639 | 11,701,362,557 | IssuesEvent | 2020-03-06 19:31:10 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Add helm version selection on catalog create | [zube]: To Test team/ui | **What kind of request is this (question/bug/enhancement/feature request):**
Enhancement
**Helm Version selection**
- When adding a catalog, user needs to be able to set the helm version for that catalog
- This should be set on create and user should be warned they cannot update it
- This cannot be updated
- When selecting an app to deploy, user needs to see what version of helm that template is using
- When user is selecting a template version, user needs to see what version of helm that template version uses
| 1.0 | Add helm version selection on catalog create - **What kind of request is this (question/bug/enhancement/feature request):**
Enhancement
**Helm Version selection**
- When adding a catalog, user needs to be able to set the helm version for that catalog
- This should be set on create and user should be warned they cannot update it
- This cannot be updated
- When selecting an app to deploy, user needs to see what version of helm that template is using
- When user is selecting a template version, user needs to see what version of helm that template version uses
| non_priority | add helm version selection on catalog create what kind of request is this question bug enhancement feature request enhancement helm version selection when adding a catalog user needs to be able to set the helm version for that catalog this should be set on create and user should be warned they cannot update it this cannot be updated when selecting an app to deploy user needs to see what version of helm that template is using when user is selecting a template version user needs to see what version of helm that template version uses | 0 |
219,698 | 17,106,421,045 | IssuesEvent | 2021-07-09 18:31:54 | IntellectualSites/PlotSquared | https://api.github.com/repos/IntellectualSites/PlotSquared | opened | Exception when using "/plot visit <number>", where <number> is 2,147,483,648 or greater | Requires Testing | ### Server Implementation
Paper
### Server Version
1.16.5
### Describe the bug
Title says it all really. This is the exception: https://pastebin.com/bS0DL5uH
### To Reproduce
1. Type "/plot visit <number>", where <number> is 2,147,483,648 or greater
### Expected behaviour
Some kind of catch for the exception, I guess.
### Screenshots / Videos
_No response_
### Error log (if applicable)
_No response_
### Plot Debugpaste
https://athion.net/ISPaster/paste/view/05e1898542f34713a9553c6cb471e7c0
### PlotSquared Version
6.0.5-Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_ | 1.0 | Exception when using "/plot visit <number>", where <number> is 2,147,483,648 or greater - ### Server Implementation
Paper
### Server Version
1.16.5
### Describe the bug
Title says it all really. This is the exception: https://pastebin.com/bS0DL5uH
### To Reproduce
1. Type "/plot visit <number>", where <number> is 2,147,483,648 or greater
### Expected behaviour
Some kind of catch for the exception, I guess.
### Screenshots / Videos
_No response_
### Error log (if applicable)
_No response_
### Plot Debugpaste
https://athion.net/ISPaster/paste/view/05e1898542f34713a9553c6cb471e7c0
### PlotSquared Version
6.0.5-Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_ | non_priority | exception when using plot visit where is or greater server implementation paper server version describe the bug title says it all really this is the exception to reproduce type plot visit where is or greater expected behaviour some kind of catch for the exception i guess screenshots videos no response error log if applicable no response plot debugpaste plotsquared version premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else no response | 0 |
22,990 | 10,832,242,463 | IssuesEvent | 2019-11-11 10:10:37 | alt-how/altinn-studio | https://api.github.com/repos/alt-how/altinn-studio | opened | CVE-2019-10219 (Medium) detected in hibernate-validator-6.0.17.Final.jar | security vulnerability | ## CVE-2019-10219 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-6.0.17.Final.jar</b></p></summary>
<p>Hibernate's Bean Validation (JSR-380) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator/">http://hibernate.org/validator/</a></p>
<p>Path to dependency file: /tmp/ws-scm/altinn-studio/src/Altinn.Platform/Altinn.Platform.PDF/pom.xml</p>
<p>Path to vulnerable library: epository/org/hibernate/validator/hibernate-validator/6.0.17.Final/hibernate-validator-6.0.17.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-6.0.17.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alt-how/altinn-studio/commit/01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3">01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in Hibernate-Validator. The SafeHtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in HTML comments and instructions. This vulnerability can result in an XSS attack.
<p>Publish Date: 2019-11-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219>CVE-2019-10219</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219</a></p>
<p>Release Date: 2019-11-08</p>
<p>Fix Resolution: 6.1.0.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-10219 (Medium) detected in hibernate-validator-6.0.17.Final.jar - ## CVE-2019-10219 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-6.0.17.Final.jar</b></p></summary>
<p>Hibernate's Bean Validation (JSR-380) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator/">http://hibernate.org/validator/</a></p>
<p>Path to dependency file: /tmp/ws-scm/altinn-studio/src/Altinn.Platform/Altinn.Platform.PDF/pom.xml</p>
<p>Path to vulnerable library: epository/org/hibernate/validator/hibernate-validator/6.0.17.Final/hibernate-validator-6.0.17.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-6.0.17.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alt-how/altinn-studio/commit/01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3">01718ddb8d46049a2cf1af8ca7919c0d7a58f4b3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in Hibernate-Validator. The SafeHtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in HTML comments and instructions. This vulnerability can result in an XSS attack.
<p>Publish Date: 2019-11-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219>CVE-2019-10219</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10219</a></p>
<p>Release Date: 2019-11-08</p>
<p>Fix Resolution: 6.1.0.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in hibernate validator final jar cve medium severity vulnerability vulnerable library hibernate validator final jar hibernate s bean validation jsr reference implementation library home page a href path to dependency file tmp ws scm altinn studio src altinn platform altinn platform pdf pom xml path to vulnerable library epository org hibernate validator hibernate validator final hibernate validator final jar dependency hierarchy x hibernate validator final jar vulnerable library found in head commit a href vulnerability details a vulnerability was found in hibernate validator the safehtml validator annotation fails to properly sanitize payloads consisting of potentially malicious code in html comments and instructions this vulnerability can result in an xss attack publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution final step up your open source security game with whitesource | 0 |
66,001 | 3,249,485,993 | IssuesEvent | 2015-10-18 07:01:52 | boarder2/Latest-Chatty-8 | https://api.github.com/repos/boarder2/Latest-Chatty-8 | opened | Update image assets | low priority | They're all low resolution and don't look good on high dpi devices. This will be tedious. | 1.0 | Update image assets - They're all low resolution and don't look good on high dpi devices. This will be tedious. | priority | update image assets they re all low resolution and don t look good on high dpi devices this will be tedious | 1 |
61,947 | 14,643,852,457 | IssuesEvent | 2020-12-25 19:10:25 | AlexRogalskiy/spark4you | https://api.github.com/repos/AlexRogalskiy/spark4you | opened | CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar | security vulnerability | ## CVE-2009-2625 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the
Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI),
a complete framework for building parser components and configurations that is extremely
modular and easy to program.</p>
<p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- spark-core_2.12-3.0.1.jar (Root Library)
- hadoop-client-2.7.4.jar
- hadoop-hdfs-2.7.4.jar
- :x: **xercesImpl-2.9.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework.
<p>Publish Date: 2009-08-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux:
* JDK and JRE 6 Update 15 or later
* JDK and JRE 5.0 Update 20 or later
Java SE releases are available at:
JDK and JRE 6 Update 15:
http://java.sun.com/javase/downloads/index.jsp
JRE 6 Update 15:
http://java.com/
through the Java Update tool for Microsoft Windows users.
JDK 6 Update 15 for Solaris is available in the following patches:
* Java SE 6 Update 15 (as delivered in patch 125136-16)
* Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit))
* Java SE 6_x86 Update 15 (as delivered in patch 125138-16)
* Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit))
JDK and JRE 5.0 Update 20:
http://java.sun.com/javase/downloads/index_jdk5.jsp
JDK 5.0 Update 20 for Solaris is available in the following patches:
* J2SE 5.0 Update 18 (as delivered in patch 118666-21)
* J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit))
* J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21)
* J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit))
Java SE for Business releases are available at:
http://www.sun.com/software/javaseforbusiness/getit_download.jsp
Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see:
http://www.java.com/en/download/help/5000010800.xml
The vendor's advisory is available at:
http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar - ## CVE-2009-2625 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the
Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI),
a complete framework for building parser components and configurations that is extremely
modular and easy to program.</p>
<p>Path to dependency file: spark4you/modules/spark4you-java/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- spark-core_2.12-3.0.1.jar (Root Library)
- hadoop-client-2.7.4.jar
- hadoop-hdfs-2.7.4.jar
- :x: **xercesImpl-2.9.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/spark4you/commit/cf3d65692017a95100e23ebd5ace43d46c143ca5">cf3d65692017a95100e23ebd5ace43d46c143ca5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework.
<p>Publish Date: 2009-08-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux:
* JDK and JRE 6 Update 15 or later
* JDK and JRE 5.0 Update 20 or later
Java SE releases are available at:
JDK and JRE 6 Update 15:
http://java.sun.com/javase/downloads/index.jsp
JRE 6 Update 15:
http://java.com/
through the Java Update tool for Microsoft Windows users.
JDK 6 Update 15 for Solaris is available in the following patches:
* Java SE 6 Update 15 (as delivered in patch 125136-16)
* Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit))
* Java SE 6_x86 Update 15 (as delivered in patch 125138-16)
* Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit))
JDK and JRE 5.0 Update 20:
http://java.sun.com/javase/downloads/index_jdk5.jsp
JDK 5.0 Update 20 for Solaris is available in the following patches:
* J2SE 5.0 Update 18 (as delivered in patch 118666-21)
* J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit))
* J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21)
* J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit))
Java SE for Business releases are available at:
http://www.sun.com/software/javaseforbusiness/getit_download.jsp
Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see:
http://www.java.com/en/download/help/5000010800.xml
The vendor's advisory is available at:
http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in xercesimpl jar cve medium severity vulnerability vulnerable library xercesimpl jar is the next generation of high performance fully compliant xml parsers in the apache xerces family this new version of xerces introduces the xerces native interface xni a complete framework for building parser components and configurations that is extremely modular and easy to program path to dependency file modules java pom xml path to vulnerable library home wss scanner repository xerces xercesimpl xercesimpl jar home wss scanner repository xerces xercesimpl xercesimpl jar dependency hierarchy spark core jar root library hadoop client jar hadoop hdfs jar x xercesimpl jar vulnerable library found in head commit a href found in base branch master vulnerability details xmlscanner java in apache java as used in sun java runtime environment jre in jdk and jre before update and jdk and jre before update and in other products allows remote attackers to cause a denial of service infinite loop and application hang via malformed xml input as demonstrated by the codenomicon xml fuzzing framework publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution the vendor has issued a fix for windows solaris and linux jdk and jre update or later jdk and jre update or later java se releases are available at jdk and jre update jre update through the java update tool for microsoft windows users jdk update for solaris is available in the following patches java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch jdk and jre update jdk update for solaris is available in the following patches update as delivered in patch update as delivered in patch update as delivered in patch update as delivered in patch java se for business releases are available at note when installing a new version of the product from a source other than a solaris patch it is recommended that the old affected versions be removed from your system to remove old affected versions on the windows platform please see the vendor s advisory is available at step up your open source security game with whitesource | 0 |
403,475 | 11,841,558,030 | IssuesEvent | 2020-03-23 21:02:00 | StudioTBA/CoronaIO | https://api.github.com/repos/StudioTBA/CoronaIO | opened | Make PoliceAgent implement Agent | Priority: Critical Refactor | **Is your feature request related to a problem? Please describe.**
All agents should implement the abstract class Agent.
**Describe the solution you would like**
To make things simpler, besides making PoliceAgent implement Agent, create a `Police` folder in `Scripts/Agent/` that contains the contents of `Scripts/FSM`to centralize the scripts related to agents.
| 1.0 | Make PoliceAgent implement Agent - **Is your feature request related to a problem? Please describe.**
All agents should implement the abstract class Agent.
**Describe the solution you would like**
To make things simpler, besides making PoliceAgent implement Agent, create a `Police` folder in `Scripts/Agent/` that contains the contents of `Scripts/FSM`to centralize the scripts related to agents.
| priority | make policeagent implement agent is your feature request related to a problem please describe all agents should implement the abstract class agent describe the solution you would like to make things simpler besides making policeagent implement agent create a police folder in scripts agent that contains the contents of scripts fsm to centralize the scripts related to agents | 1 |
45,280 | 11,623,954,238 | IssuesEvent | 2020-02-27 09:53:59 | bisq-network/bisq | https://api.github.com/repos/bisq-network/bisq | closed | Cannot build Bisq rpm on Debian 10.3 VM | in:build on:Linux | <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team at https://keybase.io/team/Bisq
-->
### Description
The desktop/package/linux/package.sh script fails while trying to create an rpm for version=1.2.7-SNAPSHOT
error: line 3: Illegal char '-' (0x2d) in: Version: 1.2.7-SNAPSHOT
#### Version
1.2.7-SNAPSHOT
### Steps to reproduce
1. Build src using OpenJDK 11 (on Debian 10.3)
2. Set JAVA_HOME to Oracle JDK 10
3. Run desktop/package/linux/package.sh
### Expected behaviour
Bisq-1.2.7-SNAPSHOT.deb and Bisq-1.2.7-SNAPSHOT.rpm are created with no errors.
### Actual behaviour
Bisq-1.2.7-SNAPSHOT.rpm could not be created because of the hyphen in the rpm version.
#### Device or machine
Debian 10.3 VM running inside Ubuntu 18.0.4
Intel i7, Nvidia GTX 960
#### Additional info
package.sh debug output:
[rpm-build-error.log](https://github.com/bisq-network/bisq/files/4246509/rpm-build-error.log)
According to https://twiki.cern.ch/twiki/bin/view/Main/RPMAndDebVersioning
RPM (4.7.1) version format
version
Mandatory
Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot)
Spaces not allowed
release
Optional
Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot)
Spaces not allowed
According to https://www.debian.org/doc/debian-policy/ch-controlfields.html#s-f-version:
upstream_version
The upstream_version may contain only alphanumerics [6] and the
characters . + - ~ (full stop, plus, hyphen, tilde) and should
start with a digit.
*** If there is no debian_revision then hyphens are not allowed. ***
debian_revision
This part of the version number specifies the version of the Debian
package based on the upstream version. It may contain only alphanumerics
and the characters + . ~ (plus, full stop, tilde) and is compared in the
same way as the upstream_version is.
Debian 10.3 can create Bisq-1.2.7-SNAPSHOT.deb, but not Bisq-1.2.7-SNAPSHOT.rpm
Workaround:
I created the rpm Bisq-1.2.7+SNAPSHOT.rpm on Debian 10.3 by creating a variable in linux/package.sh that replaces "-" with a valid "+"
rpmVersion=${version//-/+}
and substituting $rpmVersion for $version in the rpm generation related code in package.sh.
I also created Bisq-1.2.7SNAPSHOT.rpm by replacing "-" with "".
-------
The Bisq-1.2.7+SNAPSHOT.rpm file I created on the debian vm was installed using dnf on another rpm based fedora vm; it created the fedora menu item and bisq started with no problems. | 1.0 | Cannot build Bisq rpm on Debian 10.3 VM - <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team at https://keybase.io/team/Bisq
-->
### Description
The desktop/package/linux/package.sh script fails while trying to create an rpm for version=1.2.7-SNAPSHOT
error: line 3: Illegal char '-' (0x2d) in: Version: 1.2.7-SNAPSHOT
#### Version
1.2.7-SNAPSHOT
### Steps to reproduce
1. Build src using OpenJDK 11 (on Debian 10.3)
2. Set JAVA_HOME to Oracle JDK 10
3. Run desktop/package/linux/package.sh
### Expected behaviour
Bisq-1.2.7-SNAPSHOT.deb and Bisq-1.2.7-SNAPSHOT.rpm are created with no errors.
### Actual behaviour
Bisq-1.2.7-SNAPSHOT.rpm could not be created because of the hyphen in the rpm version.
#### Device or machine
Debian 10.3 VM running inside Ubuntu 18.0.4
Intel i7, Nvidia GTX 960
#### Additional info
package.sh debug output:
[rpm-build-error.log](https://github.com/bisq-network/bisq/files/4246509/rpm-build-error.log)
According to https://twiki.cern.ch/twiki/bin/view/Main/RPMAndDebVersioning
RPM (4.7.1) version format
version
Mandatory
Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot)
Spaces not allowed
release
Optional
Characters NOT allowed: - / ~ .. (hyphen, slash, tilde, double dot)
Spaces not allowed
According to https://www.debian.org/doc/debian-policy/ch-controlfields.html#s-f-version:
upstream_version
The upstream_version may contain only alphanumerics [6] and the
characters . + - ~ (full stop, plus, hyphen, tilde) and should
start with a digit.
*** If there is no debian_revision then hyphens are not allowed. ***
debian_revision
This part of the version number specifies the version of the Debian
package based on the upstream version. It may contain only alphanumerics
and the characters + . ~ (plus, full stop, tilde) and is compared in the
same way as the upstream_version is.
Debian 10.3 can create Bisq-1.2.7-SNAPSHOT.deb, but not Bisq-1.2.7-SNAPSHOT.rpm
Workaround:
I created the rpm Bisq-1.2.7+SNAPSHOT.rpm on Debian 10.3 by creating a variable in linux/package.sh that replaces "-" with a valid "+"
rpmVersion=${version//-/+}
and substituting $rpmVersion for $version in the rpm generation related code in package.sh.
I also created Bisq-1.2.7SNAPSHOT.rpm by replacing "-" with "".
-------
The Bisq-1.2.7+SNAPSHOT.rpm file I created on the debian vm was installed using dnf on another rpm based fedora vm; it created the fedora menu item and bisq started with no problems. | non_priority | cannot build bisq rpm on debian vm support requests this is for reporting bugs in the bisq app if you have a support request please join support on bisq s keybase team at description the desktop package linux package sh script fails while trying to create an rpm for version snapshot error line illegal char in version snapshot version snapshot steps to reproduce build src using openjdk on debian set java home to oracle jdk run desktop package linux package sh expected behaviour bisq snapshot deb and bisq snapshot rpm are created with no errors actual behaviour bisq snapshot rpm could not be created because of the hyphen in the rpm version device or machine debian vm running inside ubuntu intel nvidia gtx additional info package sh debug output according to rpm version format version mandatory characters not allowed hyphen slash tilde double dot spaces not allowed release optional characters not allowed hyphen slash tilde double dot spaces not allowed according to upstream version the upstream version may contain only alphanumerics and the characters full stop plus hyphen tilde and should start with a digit if there is no debian revision then hyphens are not allowed debian revision this part of the version number specifies the version of the debian package based on the upstream version it may contain only alphanumerics and the characters plus full stop tilde and is compared in the same way as the upstream version is debian can create bisq snapshot deb but not bisq snapshot rpm workaround i created the rpm bisq snapshot rpm on debian by creating a variable in linux package sh that replaces with a valid rpmversion version and substituting rpmversion for version in the rpm generation related code in package sh i also created bisq rpm by replacing with the bisq snapshot rpm file i created on the debian vm was installed using dnf on another rpm based fedora vm it created the fedora menu item and bisq started with no problems | 0 |
298,822 | 22,574,033,044 | IssuesEvent | 2022-06-28 05:04:47 | OTRF/OSSEM-DD | https://api.github.com/repos/OTRF/OSSEM-DD | closed | Update Sysmon Dictionary (Windows) - Event 3 | documentation | Add log example in XML format
Can use the following reference:
```
<EventData>
<Data Name="RuleName">-</Data>
<Data Name="UtcTime">2021-11-03 05:01:26.226</Data>
<Data Name="ProcessGuid">{3710b5c6-f3dd-6181-4000-000000000a00}</Data>
<Data Name="ProcessId">2484</Data>
<Data Name="Image">C:\Windows\System32\svchost.exe</Data>
<Data Name="User">NT AUTHORITY\SYSTEM</Data>
<Data Name="Protocol">tcp</Data>
<Data Name="Initiated">true</Data>
<Data Name="SourceIsIpv6">false</Data>
<Data Name="SourceIp">192.168.10.8</Data>
<Data Name="SourceHostname">-</Data>
<Data Name="SourcePort">50026</Data>
<Data Name="SourcePortName">-</Data>
<Data Name="DestinationIsIpv6">false</Data>
<Data Name="DestinationIp">52.167.249.196</Data>
<Data Name="DestinationHostname">-</Data>
<Data Name="DestinationPort">443</Data>
<Data Name="DestinationPortName">-</Data>
</EventData>
``` | 1.0 | Update Sysmon Dictionary (Windows) - Event 3 - Add log example in XML format
Can use the following reference:
```
<EventData>
<Data Name="RuleName">-</Data>
<Data Name="UtcTime">2021-11-03 05:01:26.226</Data>
<Data Name="ProcessGuid">{3710b5c6-f3dd-6181-4000-000000000a00}</Data>
<Data Name="ProcessId">2484</Data>
<Data Name="Image">C:\Windows\System32\svchost.exe</Data>
<Data Name="User">NT AUTHORITY\SYSTEM</Data>
<Data Name="Protocol">tcp</Data>
<Data Name="Initiated">true</Data>
<Data Name="SourceIsIpv6">false</Data>
<Data Name="SourceIp">192.168.10.8</Data>
<Data Name="SourceHostname">-</Data>
<Data Name="SourcePort">50026</Data>
<Data Name="SourcePortName">-</Data>
<Data Name="DestinationIsIpv6">false</Data>
<Data Name="DestinationIp">52.167.249.196</Data>
<Data Name="DestinationHostname">-</Data>
<Data Name="DestinationPort">443</Data>
<Data Name="DestinationPortName">-</Data>
</EventData>
``` | non_priority | update sysmon dictionary windows event add log example in xml format can use the following reference c windows svchost exe nt authority system tcp true false false | 0 |
22,851 | 11,795,605,010 | IssuesEvent | 2020-03-18 09:17:07 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Issues after resetting existing service principal credentials | Pri2 container-service/svc cxp product-question triaged | By following this guide, I have resat the service principal credentials for 3 different Kubernetes clusters running in AKS. They had different sizes, VM types and were running different Kubernetes versions. After updating the clusters became unresponsive. I couldn't schedule new pods to run, update docker images for deployments, etc. That issue disappeared automatically after about an hour. After that I had an issue with running the Kubernetes Dashboard. When opening the proxy link, the dashboard would randomly fail loading with an error that the kubernetes-service did not exist.
I also started to get error logs about services not being able to find each other. It seemed as if DNS failed between random Kubernetes services.
After some debugging I found out that for 50 % of the nodes, the kube-proxy pod from the kube-system namespace was in Pending state. By manually deleting these Pending pods, they were recreated automatically and got into a working state, which solved the issue.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 0b842222-f6ac-5266-227f-7e400610e07b
* Version Independent ID: 429cd3bf-2691-f633-2e0d-e70e5d8d0e03
* Content: [Reset the credentials for an Azure Kubernetes Service (AKS) cluster - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/update-credentials#feedback)
* Content Source: [articles/aks/update-credentials.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/update-credentials.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | 1.0 | Issues after resetting existing service principal credentials - By following this guide, I have resat the service principal credentials for 3 different Kubernetes clusters running in AKS. They had different sizes, VM types and were running different Kubernetes versions. After updating the clusters became unresponsive. I couldn't schedule new pods to run, update docker images for deployments, etc. That issue disappeared automatically after about an hour. After that I had an issue with running the Kubernetes Dashboard. When opening the proxy link, the dashboard would randomly fail loading with an error that the kubernetes-service did not exist.
I also started to get error logs about services not being able to find each other. It seemed as if DNS failed between random Kubernetes services.
After some debugging I found out that for 50 % of the nodes, the kube-proxy pod from the kube-system namespace was in Pending state. By manually deleting these Pending pods, they were recreated automatically and got into a working state, which solved the issue.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 0b842222-f6ac-5266-227f-7e400610e07b
* Version Independent ID: 429cd3bf-2691-f633-2e0d-e70e5d8d0e03
* Content: [Reset the credentials for an Azure Kubernetes Service (AKS) cluster - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/update-credentials#feedback)
* Content Source: [articles/aks/update-credentials.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/update-credentials.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | non_priority | issues after resetting existing service principal credentials by following this guide i have resat the service principal credentials for different kubernetes clusters running in aks they had different sizes vm types and were running different kubernetes versions after updating the clusters became unresponsive i couldn t schedule new pods to run update docker images for deployments etc that issue disappeared automatically after about an hour after that i had an issue with running the kubernetes dashboard when opening the proxy link the dashboard would randomly fail loading with an error that the kubernetes service did not exist i also started to get error logs about services not being able to find each other it seemed as if dns failed between random kubernetes services after some debugging i found out that for of the nodes the kube proxy pod from the kube system namespace was in pending state by manually deleting these pending pods they were recreated automatically and got into a working state which solved the issue document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned | 0 |
757,208 | 26,500,530,640 | IssuesEvent | 2023-01-18 09:56:36 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Bluetooth: host: central connection context leak | bug priority: low area: Bluetooth area: Bluetooth Host | **Describe the bug**
I've been seeing this issue with the zephyr host. I am using NCS (2c8ef25e7) and the two samples for direction finding.
The connection context for the central part is left in an incomplete state when the connection is timed out.
**To Reproduce**
`west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_peripheral -d peripheral -- -DCONFIG_BT_LL_SW_SPLIT=y`
and
`west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_central -d central -- -DCONFIG_BT_LL_SW_SPLIT=y`
Flashing both samples, and resetting the peripheral side using the RESET button on the DK the central eventually runs into a problem where it is unable to create the connection.
```
[01:13:22.249,298] <wrn> bt_conn: Found valid connection in disconnected state
[01:13:22.566,467] <wrn> b[DEVICE]: 17:5D:F5:1E:99:7D (random), AD evt type 3, AD data len 31, RSSI -82
[DEVICE]: D8:9A:D6:93:B8:0E (random), AD evt type 0, AD data len 8, RSSI -30
[AD]: 1 data_len 1
[AD]: 39 data_len 3
Create conn failed (err -22)
```
**Impact**
Connection context is leaking, creating a new connection is no longer possible when this happens.
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.13.1
- Commit SHA or Version used: NCS (2c8ef25e7)
| 1.0 | Bluetooth: host: central connection context leak - **Describe the bug**
I've been seeing this issue with the zephyr host. I am using NCS (2c8ef25e7) and the two samples for direction finding.
The connection context for the central part is left in an incomplete state when the connection is timed out.
**To Reproduce**
`west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_peripheral -d peripheral -- -DCONFIG_BT_LL_SW_SPLIT=y`
and
`west build -b nrf52833dk_nrf52833 ../zephyr/samples/bluetooth/direction_finding_central -d central -- -DCONFIG_BT_LL_SW_SPLIT=y`
Flashing both samples, and resetting the peripheral side using the RESET button on the DK the central eventually runs into a problem where it is unable to create the connection.
```
[01:13:22.249,298] <wrn> bt_conn: Found valid connection in disconnected state
[01:13:22.566,467] <wrn> b[DEVICE]: 17:5D:F5:1E:99:7D (random), AD evt type 3, AD data len 31, RSSI -82
[DEVICE]: D8:9A:D6:93:B8:0E (random), AD evt type 0, AD data len 8, RSSI -30
[AD]: 1 data_len 1
[AD]: 39 data_len 3
Create conn failed (err -22)
```
**Impact**
Connection context is leaking, creating a new connection is no longer possible when this happens.
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.13.1
- Commit SHA or Version used: NCS (2c8ef25e7)
| priority | bluetooth host central connection context leak describe the bug i ve been seeing this issue with the zephyr host i am using ncs and the two samples for direction finding the connection context for the central part is left in an incomplete state when the connection is timed out to reproduce west build b zephyr samples bluetooth direction finding peripheral d peripheral dconfig bt ll sw split y and west build b zephyr samples bluetooth direction finding central d central dconfig bt ll sw split y flashing both samples and resetting the peripheral side using the reset button on the dk the central eventually runs into a problem where it is unable to create the connection bt conn found valid connection in disconnected state b random ad evt type ad data len rssi random ad evt type ad data len rssi data len data len create conn failed err impact connection context is leaking creating a new connection is no longer possible when this happens environment please complete the following information os linux toolchain zephyr sdk commit sha or version used ncs | 1 |
275,394 | 8,575,623,432 | IssuesEvent | 2018-11-12 17:48:38 | aowen87/TicketTester | https://api.github.com/repos/aowen87/TicketTester | closed | Vector transform code for AsDirection needs changes to account for scale. | Bug Likelihood: 3 - Occasional Priority: Normal Severity: 4 - Crash / Wrong Results | Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection)
needs to be changed from
V' = [T(P+epsilon*V) - T(P)] / epsilon
to:
V' = [T(P+epsilon*Vn) - T(P)] / epsilon
where Vn = V/|V|
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 507
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Vector transform code for AsDirection needs changes to account for scale.
Assigned to: Kathleen Biagas
Category:
Target version: 2.1.2
Author: Kathleen Biagas
Start: 12/09/2010
Due date:
% Done: 0
Estimated time: 2.0
Created: 12/09/2010 08:06 pm
Updated: 12/10/2010 01:30 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection)
needs to be changed from
V' = [T(P+epsilon*V) - T(P)] / epsilon
to:
V' = [T(P+epsilon*Vn) - T(P)] / epsilon
where Vn = V/|V|
Comments:
Implemented the changes suggested, with the addition of re-scaling the results by the vector magnitude as follows:V' = ([T(P+epsilon*Vn) - T(P)] / epsilon) * |V|SVN Revisions 13236 (2.1RC) and 13238 (trunk). M /src/avt/Filters/avtCoordSystemConvert.C M /src/avt/Filters/avtTransform.C M /src/operators/Project/avtProjectFilter.C
| 1.0 | Vector transform code for AsDirection needs changes to account for scale. - Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection)
needs to be changed from
V' = [T(P+epsilon*V) - T(P)] / epsilon
to:
V' = [T(P+epsilon*Vn) - T(P)] / epsilon
where Vn = V/|V|
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 507
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: Vector transform code for AsDirection needs changes to account for scale.
Assigned to: Kathleen Biagas
Category:
Target version: 2.1.2
Author: Kathleen Biagas
Start: 12/09/2010
Due date:
% Done: 0
Estimated time: 2.0
Created: 12/09/2010 08:06 pm
Updated: 12/10/2010 01:30 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.0.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Per discussion between Matt O'Brien and Jeremyon visit-users list, Vector transform case 4 (treat AsDirection)
needs to be changed from
V' = [T(P+epsilon*V) - T(P)] / epsilon
to:
V' = [T(P+epsilon*Vn) - T(P)] / epsilon
where Vn = V/|V|
Comments:
Implemented the changes suggested, with the addition of re-scaling the results by the vector magnitude as follows:V' = ([T(P+epsilon*Vn) - T(P)] / epsilon) * |V|SVN Revisions 13236 (2.1RC) and 13238 (trunk). M /src/avt/Filters/avtCoordSystemConvert.C M /src/avt/Filters/avtTransform.C M /src/operators/Project/avtProjectFilter.C
| priority | vector transform code for asdirection needs changes to account for scale per discussion between matt o brien and jeremyon visit users list vector transform case treat asdirection needs to be changed from v epsilon to v epsilon where vn v v redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority normal subject vector transform code for asdirection needs changes to account for scale assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description per discussion between matt o brien and jeremyon visit users list vector transform case treat asdirection needs to be changed from v epsilon to v epsilon where vn v v comments implemented the changes suggested with the addition of re scaling the results by the vector magnitude as follows v epsilon v svn revisions and trunk m src avt filters avtcoordsystemconvert c m src avt filters avttransform c m src operators project avtprojectfilter c | 1 |
18,532 | 13,055,715,385 | IssuesEvent | 2020-07-30 02:31:12 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | No intellisense in morris UI | bug interface/infrastructure | When typing parameters into the morris UI, there is no intellisense popup. | 1.0 | No intellisense in morris UI - When typing parameters into the morris UI, there is no intellisense popup. | non_priority | no intellisense in morris ui when typing parameters into the morris ui there is no intellisense popup | 0 |
199,128 | 15,024,954,891 | IssuesEvent | 2021-02-01 20:22:17 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts - Kibana Home click on collapse on observability in side nav to test a11y of collapse button | Feature:Home Project:Accessibility failed-test | A test failed on a tracked branch
```
Error: a11y report:
VIOLATION
[aria-required-parent]: Ensures elements with an ARIA role that require parent roles are contained by them
Help: https://dequeuniversity.com/rules/axe/4.0/aria-required-parent?application=axeAPI
Elements:
- .euiTab-isSelected
- a[href$="explorer"] > .euiTab[role="tab"]
- a[href$="settings"] > .euiTab[role="tab"]
at Accessibility.testAxeReport (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:91:15)
at Accessibility.testAppSnapshot (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:60:18)
at process._tickCallback (internal/process/next_tick.js:68:7)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/8857/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts","test.name":"Kibana Home click on collapse on observability in side nav to test a11y of collapse button","test.failCount":1}} --> | 1.0 | Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts - Kibana Home click on collapse on observability in side nav to test a11y of collapse button - A test failed on a tracked branch
```
Error: a11y report:
VIOLATION
[aria-required-parent]: Ensures elements with an ARIA role that require parent roles are contained by them
Help: https://dequeuniversity.com/rules/axe/4.0/aria-required-parent?application=axeAPI
Elements:
- .euiTab-isSelected
- a[href$="explorer"] > .euiTab[role="tab"]
- a[href$="settings"] > .euiTab[role="tab"]
at Accessibility.testAxeReport (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:91:15)
at Accessibility.testAppSnapshot (/dev/shm/workspace/parallel/2/kibana/test/accessibility/services/a11y/a11y.ts:60:18)
at process._tickCallback (internal/process/next_tick.js:68:7)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/8857/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/home·ts","test.name":"Kibana Home click on collapse on observability in side nav to test a11y of collapse button","test.failCount":1}} --> | non_priority | failing test x pack accessibility tests x pack test accessibility apps home·ts kibana home click on collapse on observability in side nav to test of collapse button a test failed on a tracked branch error report violation ensures elements with an aria role that require parent roles are contained by them help elements euitab isselected a euitab a euitab at accessibility testaxereport dev shm workspace parallel kibana test accessibility services ts at accessibility testappsnapshot dev shm workspace parallel kibana test accessibility services ts at process tickcallback internal process next tick js first failure | 0 |
453,962 | 13,092,596,751 | IssuesEvent | 2020-08-03 08:53:09 | Candice-Louw/SORMAS.DE | https://api.github.com/repos/Candice-Louw/SORMAS.DE | closed | 33| DE | Cases and Contacts | Option to document that quarantine talk took place with date and comment field | Super High Priority question | ### Problem Description ###
As a Gesundheitsamt reporter I would like to indicate that a discussion about quarantine has taken place (check-box)
with a case/contact on a specific date (date selector field) and add further comments to the discussion if necessary (free-type text box).
### Steps to Reproduce ###
Steps to reach section:
Cases > Select case > Case Information: CASE tab > scroll down to "Quarantine"
Contacts > Select contact > Contact Information: CONTACT tab > scroll down to "Quarantine"
### Suggested Addition ###
Addition of 'quarantine discussion' drop-down field.
Options in drop-down field:
(DE)> " ", Ja, Nein, -nicht ermittelbar-
(EN)> " ", Yes, No, Unknown
IF "Ja/Yes" answered in drop-down field do the following
{
Make date field visible with caption: "Date of quarantine discussion".
Make free-text comment field visible with caption: "Comments".
}
### Screenshots ###
<img width="1146" alt="QuarantineDiscussion" src="https://user-images.githubusercontent.com/65529128/88832399-0cb85b00-d1d1-11ea-9a1e-8b4b6b2dbaa4.png"> | 1.0 | 33| DE | Cases and Contacts | Option to document that quarantine talk took place with date and comment field - ### Problem Description ###
As a Gesundheitsamt reporter I would like to indicate that a discussion about quarantine has taken place (check-box)
with a case/contact on a specific date (date selector field) and add further comments to the discussion if necessary (free-type text box).
### Steps to Reproduce ###
Steps to reach section:
Cases > Select case > Case Information: CASE tab > scroll down to "Quarantine"
Contacts > Select contact > Contact Information: CONTACT tab > scroll down to "Quarantine"
### Suggested Addition ###
Addition of 'quarantine discussion' drop-down field.
Options in drop-down field:
(DE)> " ", Ja, Nein, -nicht ermittelbar-
(EN)> " ", Yes, No, Unknown
IF "Ja/Yes" answered in drop-down field do the following
{
Make date field visible with caption: "Date of quarantine discussion".
Make free-text comment field visible with caption: "Comments".
}
### Screenshots ###
<img width="1146" alt="QuarantineDiscussion" src="https://user-images.githubusercontent.com/65529128/88832399-0cb85b00-d1d1-11ea-9a1e-8b4b6b2dbaa4.png"> | priority | de cases and contacts option to document that quarantine talk took place with date and comment field problem description as a gesundheitsamt reporter i would like to indicate that a discussion about quarantine has taken place check box with a case contact on a specific date date selector field and add further comments to the discussion if necessary free type text box steps to reproduce steps to reach section cases select case case information case tab scroll down to quarantine contacts select contact contact information contact tab scroll down to quarantine suggested addition addition of quarantine discussion drop down field options in drop down field de ja nein nicht ermittelbar en yes no unknown if ja yes answered in drop down field do the following make date field visible with caption date of quarantine discussion make free text comment field visible with caption comments screenshots img width alt quarantinediscussion src | 1 |
12,096 | 3,040,941,462 | IssuesEvent | 2015-08-07 18:16:27 | aspnet/dnx | https://api.github.com/repos/aspnet/dnx | opened | Add error codes | needs design | It would be helpful if dnx errors/warnings each had their own unique error codes. This will be helpful in case Visual Studio needs to de-duplicate errors and it also enables a help link for each error message.
cc @glennc @BillHiebert @davidfowl @DamianEdwards | 1.0 | Add error codes - It would be helpful if dnx errors/warnings each had their own unique error codes. This will be helpful in case Visual Studio needs to de-duplicate errors and it also enables a help link for each error message.
cc @glennc @BillHiebert @davidfowl @DamianEdwards | non_priority | add error codes it would be helpful if dnx errors warnings each had their own unique error codes this will be helpful in case visual studio needs to de duplicate errors and it also enables a help link for each error message cc glennc billhiebert davidfowl damianedwards | 0 |
56,824 | 13,929,986,730 | IssuesEvent | 2020-10-22 01:10:54 | RTXteam/RTX | https://api.github.com/repos/RTXteam/RTX | reopened | Address warnings in `filter_kg_and_remap_predicates.log` | housekeeping kg2 verify this fix in next KG2 build | The following warnings appear in `filter_kg_and_remap_predicates.log` from KG2.3.5 build.
```
118 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_stressor_via
119 relation curie is missing from the YAML config file: SO:derives_from
120 relation curie is missing from the YAML config file: SO:has_part
121 relation curie is missing from the YAML config file: SO:adjacent_to
122 relation curie is missing from the YAML config file: SO:guided_by
123 relation curie is missing from the YAML config file: SO:has_quality
124 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with
125 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_receptor_via
126 relation curie is missing from the YAML config file: SO:contains
127 relation curie is missing from the YAML config file: RO:0002491
128 relation curie is missing from the YAML config file: SO:non_functional_homolog_of
129 relation curie is missing from the YAML config file: SO:transcribed_to
130 relation curie is missing from the YAML config file: RO:0002309
131 relation curie is missing from the YAML config file: SO:transcribed_from
132 relation curie is missing from the YAML config file: SO:member_of
133 relation curie is missing from the YAML config file: SO:overlaps
134 relation curie is missing from the YAML config file: MONDO:has_onset
``` | 1.0 | Address warnings in `filter_kg_and_remap_predicates.log` - The following warnings appear in `filter_kg_and_remap_predicates.log` from KG2.3.5 build.
```
118 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_stressor_via
119 relation curie is missing from the YAML config file: SO:derives_from
120 relation curie is missing from the YAML config file: SO:has_part
121 relation curie is missing from the YAML config file: SO:adjacent_to
122 relation curie is missing from the YAML config file: SO:guided_by
123 relation curie is missing from the YAML config file: SO:has_quality
124 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with
125 relation curie is missing from the YAML config file: OBO:exo.obo#interacts_with_an_exposure_receptor_via
126 relation curie is missing from the YAML config file: SO:contains
127 relation curie is missing from the YAML config file: RO:0002491
128 relation curie is missing from the YAML config file: SO:non_functional_homolog_of
129 relation curie is missing from the YAML config file: SO:transcribed_to
130 relation curie is missing from the YAML config file: RO:0002309
131 relation curie is missing from the YAML config file: SO:transcribed_from
132 relation curie is missing from the YAML config file: SO:member_of
133 relation curie is missing from the YAML config file: SO:overlaps
134 relation curie is missing from the YAML config file: MONDO:has_onset
``` | non_priority | address warnings in filter kg and remap predicates log the following warnings appear in filter kg and remap predicates log from build relation curie is missing from the yaml config file obo exo obo interacts with an exposure stressor via relation curie is missing from the yaml config file so derives from relation curie is missing from the yaml config file so has part relation curie is missing from the yaml config file so adjacent to relation curie is missing from the yaml config file so guided by relation curie is missing from the yaml config file so has quality relation curie is missing from the yaml config file obo exo obo interacts with relation curie is missing from the yaml config file obo exo obo interacts with an exposure receptor via relation curie is missing from the yaml config file so contains relation curie is missing from the yaml config file ro relation curie is missing from the yaml config file so non functional homolog of relation curie is missing from the yaml config file so transcribed to relation curie is missing from the yaml config file ro relation curie is missing from the yaml config file so transcribed from relation curie is missing from the yaml config file so member of relation curie is missing from the yaml config file so overlaps relation curie is missing from the yaml config file mondo has onset | 0 |
184,629 | 6,714,778,794 | IssuesEvent | 2017-10-13 18:18:41 | Datawheel/codelife | https://api.github.com/repos/Datawheel/codelife | closed | Add "disable sharing" on user page for logged-in admins | priority: high type: feature | In response to repeated offenses, admins should be able to disable any user's ability to share projects and codeblocks via a switch on that user's profile page. | 1.0 | Add "disable sharing" on user page for logged-in admins - In response to repeated offenses, admins should be able to disable any user's ability to share projects and codeblocks via a switch on that user's profile page. | priority | add disable sharing on user page for logged in admins in response to repeated offenses admins should be able to disable any user s ability to share projects and codeblocks via a switch on that user s profile page | 1 |
24,621 | 7,531,279,202 | IssuesEvent | 2018-04-15 03:25:21 | travis-ci/travis-ci | https://api.github.com/repos/travis-ci/travis-ci | closed | Travis -proxy and -proxy-tunnel flags for connecting with Sauce Labs | feature-request stale travis-build | We are behind a proxy and are trying to get sauce labs to work. For this, we need to be able to set the `-proxy` and `-proxy-tunnel` flags (we have got it to work via command line tool running `sc -u username -k key -proxy proxy:port -proxy-tunnel`). However, setting these flags does not seem possible in Travis ([additional options](https://docs.travis-ci.com/user/sauce-connect/#Additional-options)). Will this be supported, or is there a workaround for this?
| 1.0 | Travis -proxy and -proxy-tunnel flags for connecting with Sauce Labs - We are behind a proxy and are trying to get sauce labs to work. For this, we need to be able to set the `-proxy` and `-proxy-tunnel` flags (we have got it to work via command line tool running `sc -u username -k key -proxy proxy:port -proxy-tunnel`). However, setting these flags does not seem possible in Travis ([additional options](https://docs.travis-ci.com/user/sauce-connect/#Additional-options)). Will this be supported, or is there a workaround for this?
| non_priority | travis proxy and proxy tunnel flags for connecting with sauce labs we are behind a proxy and are trying to get sauce labs to work for this we need to be able to set the proxy and proxy tunnel flags we have got it to work via command line tool running sc u username k key proxy proxy port proxy tunnel however setting these flags does not seem possible in travis will this be supported or is there a workaround for this | 0 |
200,895 | 22,916,013,757 | IssuesEvent | 2022-07-17 01:08:53 | nanopathi/system_bt_AOSP10_r33_CVE-2021-0522 | https://api.github.com/repos/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522 | reopened | CVE-2019-2227 (Medium) detected in https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33 | security vulnerability | ## CVE-2019-2227 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33</b></p></summary>
<p>
<p>AOSP Platform System bt</p>
<p>Library home page: <a href=https://source.codeaurora.org/external/imx/aosp/platform/system/bt/>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522/commit/4f42a84fe2cb3c64302354c9278d446ed65c76f5">4f42a84fe2cb3c64302354c9278d446ed65c76f5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/bta/av/bta_av_aact.cc</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In DeepCopy of btif_av.cc, there is a possible out of bounds read due to improper casting. This could lead to remote information disclosure over Bluetooth with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10Android ID: A-140768453
<p>Publish Date: 2019-12-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2227>CVE-2019-2227</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227</a></p>
<p>Release Date: 2019-12-06</p>
<p>Fix Resolution: android-9.0.0_r51;android-10.0.0_r17</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-2227 (Medium) detected in https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33 - ## CVE-2019-2227 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/android-10.0.0_r33</b></p></summary>
<p>
<p>AOSP Platform System bt</p>
<p>Library home page: <a href=https://source.codeaurora.org/external/imx/aosp/platform/system/bt/>https://source.codeaurora.org/external/imx/aosp/platform/system/bt/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/system_bt_AOSP10_r33_CVE-2021-0522/commit/4f42a84fe2cb3c64302354c9278d446ed65c76f5">4f42a84fe2cb3c64302354c9278d446ed65c76f5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/bta/av/bta_av_aact.cc</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In DeepCopy of btif_av.cc, there is a possible out of bounds read due to improper casting. This could lead to remote information disclosure over Bluetooth with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10Android ID: A-140768453
<p>Publish Date: 2019-12-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2227>CVE-2019-2227</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2227</a></p>
<p>Release Date: 2019-12-06</p>
<p>Fix Resolution: android-9.0.0_r51;android-10.0.0_r17</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in cve medium severity vulnerability vulnerable library aosp platform system bt library home page a href found in head commit a href found in base branch master vulnerable source files bta av bta av aact cc vulnerability details in deepcopy of btif av cc there is a possible out of bounds read due to improper casting this could lead to remote information disclosure over bluetooth with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android step up your open source security game with whitesource | 0 |
799,227 | 28,302,713,029 | IssuesEvent | 2023-04-10 07:54:47 | AY2223S2-CS2113-W12-3/tp | https://api.github.com/repos/AY2223S2-CS2113-W12-3/tp | closed | Bug: Possible to add item with price/quantity = 0 but InventoryData file becomes corrupted | type.Bug priority.High | It's possible to add an item with price = 0 or quantity = 0 successfully through the program, but on startup the inventory data file's status will be corrupted. Probably because `checkFileValidSession` will return corrupt file status when `qty <= 0`/`price <= 0` .

| 1.0 | Bug: Possible to add item with price/quantity = 0 but InventoryData file becomes corrupted - It's possible to add an item with price = 0 or quantity = 0 successfully through the program, but on startup the inventory data file's status will be corrupted. Probably because `checkFileValidSession` will return corrupt file status when `qty <= 0`/`price <= 0` .

| priority | bug possible to add item with price quantity but inventorydata file becomes corrupted it s possible to add an item with price or quantity successfully through the program but on startup the inventory data file s status will be corrupted probably because checkfilevalidsession will return corrupt file status when qty price | 1 |
570,467 | 17,023,118,477 | IssuesEvent | 2021-07-03 00:27:28 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | OpenLayers compatibility - javascript API for embedding OSM in webpages | Component: api Priority: major Resolution: fixed Type: enhancement | **[Submitted to the original trac issue database at 3.42pm, Monday, 29th May 2006]**
Decoding what crschmidt said here:
http://lists.openstreetmap.org/pipermail/talk/2006-May/004348.html
so that we can use OpenLayers on the OSM website, and so that OSM's WMS server works with OpenLayers for everyone.
Current tile implementation is broken for adding markers and info windows to the map. OpenLayers will fix that. | 1.0 | OpenLayers compatibility - javascript API for embedding OSM in webpages - **[Submitted to the original trac issue database at 3.42pm, Monday, 29th May 2006]**
Decoding what crschmidt said here:
http://lists.openstreetmap.org/pipermail/talk/2006-May/004348.html
so that we can use OpenLayers on the OSM website, and so that OSM's WMS server works with OpenLayers for everyone.
Current tile implementation is broken for adding markers and info windows to the map. OpenLayers will fix that. | priority | openlayers compatibility javascript api for embedding osm in webpages decoding what crschmidt said here so that we can use openlayers on the osm website and so that osm s wms server works with openlayers for everyone current tile implementation is broken for adding markers and info windows to the map openlayers will fix that | 1 |
767,487 | 26,927,623,433 | IssuesEvent | 2023-02-07 14:48:17 | daisy/ebraille | https://api.github.com/repos/daisy/ebraille | opened | Automatic, converted, and human prepared files will have different levels of quality and expectations | use case High Priority content spec metadata | I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format.
*Detail*
The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format.
*Proposal*
Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion.
With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
| 1.0 | Automatic, converted, and human prepared files will have different levels of quality and expectations - I am a braille user and my expectations for the quality of file I am receiving will vary based on whether the file was prepared by a transcriber, an automatic process, or converted from a braille file in an older format.
*Detail*
The best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber. Next will likely be those converted from an automatic process. Finally, there will be files prepared by converting a braille file in an older format.
*Proposal*
Metadata should indicate whether the file was prepared by a human, an automatic process, or converted from an older braille file. If a human, it would be good if it indicated the person's level of braille competence (professional vs amateur- though this aspect needs more thought). If an automatic process, it would be good if it indicated the file type of the source file as each will have strengths and weaknesses. The program used for the automatic process would also be good to indicate. Finally, files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion.
With this information, schools, libraries, and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes, programs, and starting files work best for them and their needs.
| priority | automatic converted and human prepared files will have different levels of quality and expectations i am a braille user and my expectations for the quality of file i am receiving will vary based on whether the file was prepared by a transcriber an automatic process or converted from a braille file in an older format detail the best braille files will be those prepared by a human and within that subset they will be those prepared by a certified braille transcriber next will likely be those converted from an automatic process finally there will be files prepared by converting a braille file in an older format proposal metadata should indicate whether the file was prepared by a human an automatic process or converted from an older braille file if a human it would be good if it indicated the person s level of braille competence professional vs amateur though this aspect needs more thought if an automatic process it would be good if it indicated the file type of the source file as each will have strengths and weaknesses the program used for the automatic process would also be good to indicate finally files converted from a braille file in an older format should indicate the file type of the source file and what program did the conversion with this information schools libraries and braille users could better understand the quality of the files that they receive and could make informed decisions about what processes programs and starting files work best for them and their needs | 1 |
39,581 | 16,048,098,769 | IssuesEvent | 2021-04-22 15:44:46 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | support for secrets_id in aws_dms_endpoint | enhancement service/databasemigrationservice | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The aws_dms_endpoint allows to configure AWS Secrets manager secret with database details like host, user, password. Please add this to the resource.
### New or Affected Resource(s)
* aws_dms_endpoint
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| 2.0 | support for secrets_id in aws_dms_endpoint - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The aws_dms_endpoint allows to configure AWS Secrets manager secret with database details like host, user, password. Please add this to the resource.
### New or Affected Resource(s)
* aws_dms_endpoint
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
* #0000
| non_priority | support for secrets id in aws dms endpoint community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the aws dms endpoint allows to configure aws secrets manager secret with database details like host user password please add this to the resource new or affected resource s aws dms endpoint potential terraform configuration hcl copy paste your terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example | 0 |
85,175 | 16,611,209,225 | IssuesEvent | 2021-06-02 11:45:42 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [Perf][Mono] Regressions in System.Drawing.Tests.Perf_Color | area-Codegen-Interpreter-mono runtime-mono tenet-performance untriaged | It looks like this regression was caused by https://github.com/dotnet/runtime/commit/89498d8f6095511d1553fabe39fa8f0579b1cf2d
@L2 Is there anything to do about this?
### Run Information
Architecture | x64
-- | --
OS | ubuntu 18.04
Baseline | [a3e0894f38b1ab558653e7c1b7322a8c4db92af1](https://github.com/dotnet/runtime/commit/a3e0894f38b1ab558653e7c1b7322a8c4db92af1)
Compare | [6bfc5f21dea7b550f1c807454d45408ef34764e1](https://github.com/dotnet/runtime/commit/6bfc5f21dea7b550f1c807454d45408ef34764e1)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a3e0894f38b1ab558653e7c1b7322a8c4db92af1...6bfc5f21dea7b550f1c807454d45408ef34764e1)
### Regressions in System.Drawing.Tests.Perf_Color
Benchmark | Baseline | Test | Test/Base | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | --
[GetHue](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetHue.html>) | 1.60 μs | 1.93 μs | 1.20 | | |
[GetBrightness](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetBrightness.html>) | 1.18 μs | 1.42 μs | 1.20 | | |















[Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
python3 .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Drawing.Tests.Perf_Color*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-4f289a83-011e-4af2-93d6-e54544451aaef8e9425d5ed45ccaa/6d6b6cd4-7320-4acb-8ebe-0f24ed4209e4.zip?sv=2019-07-07&se=2021-05-05T19%3A40%3A38Z&sr=c&sp=rl&sig=jpXdspM56nw22cwLixbAlaT6r3ta5%2BXlPj373BUFgwc%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-7428b546-ea89-43f0-9d23-b45da4cd3a0731d56e954b543338d/57a3901c-6af6-49e3-aa41-64712f588ca5.zip?sv=2019-07-07&se=2021-05-06T16%3A34%3A52Z&sr=c&sp=rl&sig=5LEDRnUTyNz%2FIubzfnjqjefoZlGJSeiY5c10wJtIwj0%3D>)
### Histogram
#### System.Drawing.Tests.Perf_Color.GetHue
```log
```
#### System.Drawing.Tests.Perf_Color.GetBrightness
```log
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| 1.0 | [Perf][Mono] Regressions in System.Drawing.Tests.Perf_Color - It looks like this regression was caused by https://github.com/dotnet/runtime/commit/89498d8f6095511d1553fabe39fa8f0579b1cf2d
@L2 Is there anything to do about this?
### Run Information
Architecture | x64
-- | --
OS | ubuntu 18.04
Baseline | [a3e0894f38b1ab558653e7c1b7322a8c4db92af1](https://github.com/dotnet/runtime/commit/a3e0894f38b1ab558653e7c1b7322a8c4db92af1)
Compare | [6bfc5f21dea7b550f1c807454d45408ef34764e1](https://github.com/dotnet/runtime/commit/6bfc5f21dea7b550f1c807454d45408ef34764e1)
Diff | [Diff](https://github.com/dotnet/runtime/compare/a3e0894f38b1ab558653e7c1b7322a8c4db92af1...6bfc5f21dea7b550f1c807454d45408ef34764e1)
### Regressions in System.Drawing.Tests.Perf_Color
Benchmark | Baseline | Test | Test/Base | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | --
[GetHue](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetHue.html>) | 1.60 μs | 1.93 μs | 1.20 | | |
[GetBrightness](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_x64_ubuntu 18.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono/System.Drawing.Tests.Perf_Color.GetBrightness.html>) | 1.18 μs | 1.42 μs | 1.20 | | |















[Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/04_12_2021/refs/heads/main_x64_ubuntu%2018.04_LLVM=false_MonoAOT=false_MonoInterpreter=false_RunKind=micro_mono_Regression/System.Drawing.Tests.Perf_Color.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
python3 .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Drawing.Tests.Perf_Color*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-4f289a83-011e-4af2-93d6-e54544451aaef8e9425d5ed45ccaa/6d6b6cd4-7320-4acb-8ebe-0f24ed4209e4.zip?sv=2019-07-07&se=2021-05-05T19%3A40%3A38Z&sr=c&sp=rl&sig=jpXdspM56nw22cwLixbAlaT6r3ta5%2BXlPj373BUFgwc%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-7428b546-ea89-43f0-9d23-b45da4cd3a0731d56e954b543338d/57a3901c-6af6-49e3-aa41-64712f588ca5.zip?sv=2019-07-07&se=2021-05-06T16%3A34%3A52Z&sr=c&sp=rl&sig=5LEDRnUTyNz%2FIubzfnjqjefoZlGJSeiY5c10wJtIwj0%3D>)
### Histogram
#### System.Drawing.Tests.Perf_Color.GetHue
```log
```
#### System.Drawing.Tests.Perf_Color.GetBrightness
```log
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| non_priority | regressions in system drawing tests perf color it looks like this regression was caused by is there anything to do about this run information architecture os ubuntu baseline compare diff regressions in system drawing tests perf color benchmark baseline test test base baseline ir compare ir ir ratio baseline etl compare etl μs μs μs μs repro cmd git clone performance scripts benchmarks ci py f filter system drawing tests perf color payloads histogram system drawing tests perf color gethue log system drawing tests perf color getbrightness log docs | 0 |
57,112 | 7,033,879,080 | IssuesEvent | 2017-12-27 13:36:19 | cpprefjp/kunai | https://api.github.com/repos/cpprefjp/kunai | closed | コードサンプル実行: Try Code ボタンが別のコードブロックに表示される | bug design | 以下のページで起こっている問題です。Chrome 63.0.3239.84 / Firefox 57.0.2 (on Win 10) で確認しました。
* http://cpprefjp.github.io/lang/cpp11/char16_32.html
* http://cpprefjp.github.io/lang/cpp11/utf8_string_literals.html
リスト内に ```` ```cpp ~``` ```` があるとき、
1. Try Code ボタンが表示されるコードブロックが一斉にずれています。
2. また Try Code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。
(ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません)
| 1.0 | コードサンプル実行: Try Code ボタンが別のコードブロックに表示される - 以下のページで起こっている問題です。Chrome 63.0.3239.84 / Firefox 57.0.2 (on Win 10) で確認しました。
* http://cpprefjp.github.io/lang/cpp11/char16_32.html
* http://cpprefjp.github.io/lang/cpp11/utf8_string_literals.html
リスト内に ```` ```cpp ~``` ```` があるとき、
1. Try Code ボタンが表示されるコードブロックが一斉にずれています。
2. また Try Code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。
(ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません)
| non_priority | コードサンプル実行 try code ボタンが別のコードブロックに表示される 以下のページで起こっている問題です。chrome firefox on win で確認しました。 リスト内に cpp ~ があるとき、 try code ボタンが表示されるコードブロックが一斉にずれています。 また try code ボタンを再度押して実行モードを閉じたとき、出力枠が表示されたまま残ってしまいます。 ちゃんと調べていないのですが、恐らく kunai の問題だろうと思ってここに書きます。違っていたらすみません | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.