Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
68,846 | 14,958,805,907 | IssuesEvent | 2021-01-27 01:38:09 | mihorsky/phpbb | https://api.github.com/repos/mihorsky/phpbb | opened | CVE-2020-7753 (High) detected in trim-0.0.1.tgz | security vulnerability | ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: phpbb/package.json</p>
<p>Path to vulnerable library: phpbb/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-13.2.1.tgz (Root Library)
- postcss-markdown-0.36.0.tgz
- remark-10.0.1.tgz
- remark-parse-6.0.3.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mihorsky/phpbb/commit/df86840378ba9a67de78b42213b0c85da2b3eadc">df86840378ba9a67de78b42213b0c85da2b3eadc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7753 (High) detected in trim-0.0.1.tgz - ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: phpbb/package.json</p>
<p>Path to vulnerable library: phpbb/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-13.2.1.tgz (Root Library)
- postcss-markdown-0.36.0.tgz
- remark-10.0.1.tgz
- remark-parse-6.0.3.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mihorsky/phpbb/commit/df86840378ba9a67de78b42213b0c85da2b3eadc">df86840378ba9a67de78b42213b0c85da2b3eadc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in trim tgz cve high severity vulnerability vulnerable library trim tgz trim string whitespace library home page a href path to dependency file phpbb package json path to vulnerable library phpbb node modules trim package json dependency hierarchy stylelint tgz root library postcss markdown tgz remark tgz remark parse tgz x trim tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package trim are vulnerable to regular expression denial of service redos via trim publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
73,149 | 7,328,206,464 | IssuesEvent | 2018-03-04 18:29:04 | openSUSE/trollolo | https://api.github.com/repos/openSUSE/trollolo | closed | Missing check for true in unit test. | good first issue :1st_place_medal: help wanted :two_hearts: junior :baby: test suite :syringe: | I think there should be `.to be true` in lines 20 & 21 expect statements because now the test would pass in any scenario. :smile:
https://github.com/openSUSE/trollolo/blob/d928c9417943ba2e8645cf08b2b30c7b1ebdd34e/spec/unit/cli_spec.rb#L14-L22 | 1.0 | Missing check for true in unit test. - I think there should be `.to be true` in lines 20 & 21 expect statements because now the test would pass in any scenario. :smile:
https://github.com/openSUSE/trollolo/blob/d928c9417943ba2e8645cf08b2b30c7b1ebdd34e/spec/unit/cli_spec.rb#L14-L22 | non_infrastructure | missing check for true in unit test i think there should be to be true in lines expect statements because now the test would pass in any scenario smile | 0 |
18,508 | 10,132,356,715 | IssuesEvent | 2019-08-01 22:09:20 | cBioPortal/cbioportal | https://api.github.com/repos/cBioPortal/cbioportal | opened | co-expression is slow? | performance results.coexpression | The co-expression seems slower than before. I am wondering if we can find a way to test? | True | co-expression is slow? - The co-expression seems slower than before. I am wondering if we can find a way to test? | non_infrastructure | co expression is slow the co expression seems slower than before i am wondering if we can find a way to test | 0 |
23,360 | 11,942,312,326 | IssuesEvent | 2020-04-02 20:03:46 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | Send amp-analytics metrics to viewer | Type: Feature Request WG: performance | ## Describe the new feature or change to an existing feature you'd like to see
Send performance metrics captured from amp-analytics via viewer messaging channel for debugging purposes.
We have an existing performance class which sends performance metrics to the viewer (if any). This issue proposes that we also send any metrics from amp-analytics extension via the same channel.
[List of performance metrics to send](https://github.com/ampproject/amphtml/blob/master/spec/amp-var-substitutions.md#performance)
## Describe alternatives you've considered
n/a
## Additional context
This will ensure the AMP team is looking at the same information as publishers and give us a wider range of insights into the performance of documents from the cache. It potentially allows us to set up additional alerting if necessary to deliver the best user experience for AMP documents.
Possibly related: https://github.com/ampproject/amphtml/issues/9018 | True | Send amp-analytics metrics to viewer - ## Describe the new feature or change to an existing feature you'd like to see
Send performance metrics captured from amp-analytics via viewer messaging channel for debugging purposes.
We have an existing performance class which sends performance metrics to the viewer (if any). This issue proposes that we also send any metrics from amp-analytics extension via the same channel.
[List of performance metrics to send](https://github.com/ampproject/amphtml/blob/master/spec/amp-var-substitutions.md#performance)
## Describe alternatives you've considered
n/a
## Additional context
This will ensure the AMP team is looking at the same information as publishers and give us a wider range of insights into the performance of documents from the cache. It potentially allows us to set up additional alerting if necessary to deliver the best user experience for AMP documents.
Possibly related: https://github.com/ampproject/amphtml/issues/9018 | non_infrastructure | send amp analytics metrics to viewer describe the new feature or change to an existing feature you d like to see send performance metrics captured from amp analytics via viewer messaging channel for debugging purposes we have an existing performance class which sends performance metrics to the viewer if any this issue proposes that we also send any metrics from amp analytics extension via the same channel describe alternatives you ve considered n a additional context this will ensure the amp team is looking at the same information as publishers and give us a wider range of insights into the performance of documents from the cache it potentially allows us to set up additional alerting if necessary to deliver the best user experience for amp documents possibly related | 0 |
73,301 | 24,554,429,969 | IssuesEvent | 2022-10-12 14:51:16 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation | I-defect needs-triaging | ### What happened?
I am trying to save a screenshot from my RemoteWebDriver session (GRID 4 distributed mode) and it looks like there is an issue with the websocket connection or the cdp version is not compatiable with selenium-java version: 4.5.0
### How can we reproduce the issue?
```shell
Java code:
driver = WebDriverManager.chromedriver()
.remoteAddress("http://seleniumgridtest.check24tech.de:4444")
.capabilities(DesiredCapabilitiesFactory.createChromeOptions())
.create();
WebDriver augmentedDriver = new Augmenter().augment(driver);
srcFile = ((TakesScreenshot) augmentedDriver).getScreenshotAs(OutputType.FILE);
```
```
### Relevant log output
```shell
Connected to the target VM, address: '127.0.0.1:64835', transport: 'socket'
Oct 12, 2022 3:58:12 PM org.openqa.selenium.remote.tracing.opentelemetry.OpenTelemetryTracer createTracer
INFO: Using OpenTelemetry for tracing
Oct 12, 2022 3:58:25 PM org.openqa.selenium.remote.ProtocolHandshake createSession
INFO: Detected upstream dialect: W3C
Oct 12, 2022 4:02:05 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
WARNING: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation
Oct 12, 2022 4:02:05 PM org.openqa.selenium.devtools.CdpVer
Oct 12, 2022 4:20:19 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
WARNING: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation
Oct 12, 2022 4:20:19 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
INFO: Unable to find CDP implementation matching 106.
Oct 12, 2022 4:20:29 PM org.openqa.selenium.remote.http.netty.NettyWebSocket lambda$new$0
WARNING: connection timed out: /172.20.0.5:4444
java.net.ConnectException: connection timed out: /172.20.0.5:4444
at org.asynchttpclient.netty.channel.NettyConnectListener.onFailure(NettyConnectListener.java:179)
at org.asynchttpclient.netty.channel.NettyChannelConnector$1.onFailure(NettyChannelConnector.java:108)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:28)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:20)
at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:578)
at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:571)
at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:550)
at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:491)
at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:616)
at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:609)
at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:262)
at io.netty.util.concurrent.PromiseTask.runTask(PromiseTask.java:98)
at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:153)
at io.netty.util.concurrent.AbstractEventExecutor.runTask$$$capture(AbstractEventExecutor.java:174)
at io.netty.util.concurrent.AbstractEventExecutor.runTask(AbstractEventExecutor.java)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute$$$capture(AbstractEventExecutor.java:167)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java)
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:470)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:569)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:832)
Caused by: io.netty.channel.ConnectTimeoutException: connection timed out: /172.20.0.5:4444
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:261)
... 12 more
Oct 12, 2022 4:20:29 PM org.openqa.selenium.remote.http.WebSocket$Listener onError
WARNING: connection timed out: /172.20.0.5:4444
java.net.ConnectException: connection timed out: /172.20.0.5:4444
at org.asynchttpclient.netty.channel.NettyConnectListener.onFailure(NettyConnectListener.java:179)
at org.asynchttpclient.netty.channel.NettyChannelConnector$1.onFailure(NettyChannelConnector.java:108)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:28)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:20)
at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:578)
at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:571)
at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:550)
at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:491)
at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:616)
at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:609)
at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:262)
at io.netty.util.concurrent.PromiseTask.runTask(PromiseTask.java:98)
at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:153)
at io.netty.util.concurrent.AbstractEventExecutor.runTask$$$capture(AbstractEventExecutor.java:174)
at io.netty.util.concurrent.AbstractEventExecutor.runTask(AbstractEventExecutor.java)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute$$$capture(AbstractEventExecutor.java:167)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java)
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:470)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:569)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:832)
Caused by: io.netty.channel.ConnectTimeoutException: connection timed out: /172.20.0.5:4444
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:261)
... 12 more
org.openqa.selenium.remote.http.ConnectionFailedException: Unable to establish websocket connection to http://172.20.0.5:4444/session/b47bfd6fa8d3d37c813e127f67f8cb0d/se/cdp
Build info: version: '4.5.0', revision: 'fe167b119a'
System info: os.name: 'Mac OS X', os.arch: 'aarch64', os.version: '12.5', java.version: '15.0.8'
Driver info: driver.version: unknown
at org.openqa.selenium.remote.http.netty.NettyWebSocket.<init>(NettyWebSocket.java:102)
at org.openqa.selenium.remote.http.netty.NettyWebSocket.lambda$create$3(NettyWebSocket.java:128)
at org.openqa.selenium.remote.http.netty.NettyClient.openSocket(NettyClient.java:107)
at org.openqa.selenium.devtools.Connection.<init>(Connection.java:77)
at org.openqa.selenium.devtools.SeleniumCdpConnection.<init>(SeleniumCdpConnection.java:34)
at org.openqa.selenium.devtools.SeleniumCdpConnection.lambda$create$0(SeleniumCdpConnection.java:56)
at java.base/java.util.Optional.map(Optional.java:258)
at org.openqa.selenium.devtools.SeleniumCdpConnection.create(SeleniumCdpConnection.java:54)
at org.openqa.selenium.devtools.SeleniumCdpConnection.create(SeleniumCdpConnection.java:47)
at org.openqa.selenium.devtools.DevToolsProvider.getImplementation(DevToolsProvider.java:50)
at org.openqa.selenium.devtools.DevToolsProvider.getImplementation(DevToolsProvider.java:31)
at org.openqa.selenium.remote.Augmenter.augment(Augmenter.java:186)
at de.check24.testumgebung.TestListener.savePng(TestListener.java:64)
at de.check24.testumgebung.TestListener.createScreenshot(TestListener.java:42)
at de.check24.testumgebung.TestListener.onTestFailure(TestListener.java:160)
at org.testng.internal.TestListenerHelper.runTestListeners(TestListenerHelper.java:96)
at org.testng.internal.TestInvoker.runTestResultListener(TestInvoker.java:219)
at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:834)
at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:146)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1511)
at org.testng.TestRunner.privateRun(TestRunner.java:794)
at org.testng.TestRunner.run(TestRunner.java:596)
at org.testng.SuiteRunner.runTest(SuiteRunner.java:377)
at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:371)
at org.testng.SuiteRunner.privateRun(SuiteRunner.java:332)
at org.testng.SuiteRunner.run(SuiteRunner.java:276)
at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53)
at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96)
at org.testng.TestNG.runSuitesSequentially(TestNG.java:1212)
at org.testng.TestNG.runSuitesLocally(TestNG.java:1134)
at org.testng.TestNG.runSuites(TestNG.java:1063)
at org.testng.TestNG.run(TestNG.java:1031)
at com.intellij.rt.testng.IDEARemoteTestNG.run(IDEARemoteTestNG.java:66)
at com.intellij.rt.testng.RemoteTestNGStarter.main(RemoteTestNGStarter.java:109)
Disconnected from the target VM, address: '127.0.0.1:64835', transport: 'socket'
Process finished with exit code 255
```
### Operating System
macOS Monterey Version 12.5
### Selenium version
Selenium 4.5.0 | openjdk 15.0.8 2022-07-19 | Chrome-driver 4.5.0
### What are the browser(s) and version(s) where you see this issue?
Chrome v.106.0
### What are the browser driver(s) and version(s) where you see this issue?
browserVersion= v.106.0.5249.91
### Are you using Selenium Grid?
Selenium Grid 4.5.0 (revision fe167b119a) | 1.0 | [🐛 Bug]: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation - ### What happened?
I am trying to save a screenshot from my RemoteWebDriver session (GRID 4 distributed mode) and it looks like there is an issue with the websocket connection or the cdp version is not compatiable with selenium-java version: 4.5.0
### How can we reproduce the issue?
```shell
Java code:
driver = WebDriverManager.chromedriver()
.remoteAddress("http://seleniumgridtest.check24tech.de:4444")
.capabilities(DesiredCapabilitiesFactory.createChromeOptions())
.create();
WebDriver augmentedDriver = new Augmenter().augment(driver);
srcFile = ((TakesScreenshot) augmentedDriver).getScreenshotAs(OutputType.FILE);
```
```
### Relevant log output
```shell
Connected to the target VM, address: '127.0.0.1:64835', transport: 'socket'
Oct 12, 2022 3:58:12 PM org.openqa.selenium.remote.tracing.opentelemetry.OpenTelemetryTracer createTracer
INFO: Using OpenTelemetry for tracing
Oct 12, 2022 3:58:25 PM org.openqa.selenium.remote.ProtocolHandshake createSession
INFO: Detected upstream dialect: W3C
Oct 12, 2022 4:02:05 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
WARNING: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation
Oct 12, 2022 4:02:05 PM org.openqa.selenium.devtools.CdpVer
Oct 12, 2022 4:20:19 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
WARNING: Unable to find an exact match for CDP version 106, so returning the closest version found: a no-op implementation
Oct 12, 2022 4:20:19 PM org.openqa.selenium.devtools.CdpVersionFinder findNearestMatch
INFO: Unable to find CDP implementation matching 106.
Oct 12, 2022 4:20:29 PM org.openqa.selenium.remote.http.netty.NettyWebSocket lambda$new$0
WARNING: connection timed out: /172.20.0.5:4444
java.net.ConnectException: connection timed out: /172.20.0.5:4444
at org.asynchttpclient.netty.channel.NettyConnectListener.onFailure(NettyConnectListener.java:179)
at org.asynchttpclient.netty.channel.NettyChannelConnector$1.onFailure(NettyChannelConnector.java:108)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:28)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:20)
at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:578)
at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:571)
at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:550)
at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:491)
at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:616)
at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:609)
at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:262)
at io.netty.util.concurrent.PromiseTask.runTask(PromiseTask.java:98)
at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:153)
at io.netty.util.concurrent.AbstractEventExecutor.runTask$$$capture(AbstractEventExecutor.java:174)
at io.netty.util.concurrent.AbstractEventExecutor.runTask(AbstractEventExecutor.java)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute$$$capture(AbstractEventExecutor.java:167)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java)
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:470)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:569)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:832)
Caused by: io.netty.channel.ConnectTimeoutException: connection timed out: /172.20.0.5:4444
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:261)
... 12 more
Oct 12, 2022 4:20:29 PM org.openqa.selenium.remote.http.WebSocket$Listener onError
WARNING: connection timed out: /172.20.0.5:4444
java.net.ConnectException: connection timed out: /172.20.0.5:4444
at org.asynchttpclient.netty.channel.NettyConnectListener.onFailure(NettyConnectListener.java:179)
at org.asynchttpclient.netty.channel.NettyChannelConnector$1.onFailure(NettyChannelConnector.java:108)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:28)
at org.asynchttpclient.netty.SimpleChannelFutureListener.operationComplete(SimpleChannelFutureListener.java:20)
at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:578)
at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:571)
at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:550)
at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:491)
at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:616)
at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:609)
at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:262)
at io.netty.util.concurrent.PromiseTask.runTask(PromiseTask.java:98)
at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:153)
at io.netty.util.concurrent.AbstractEventExecutor.runTask$$$capture(AbstractEventExecutor.java:174)
at io.netty.util.concurrent.AbstractEventExecutor.runTask(AbstractEventExecutor.java)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute$$$capture(AbstractEventExecutor.java:167)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java)
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:470)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:569)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:832)
Caused by: io.netty.channel.ConnectTimeoutException: connection timed out: /172.20.0.5:4444
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe$1.run(AbstractNioChannel.java:261)
... 12 more
org.openqa.selenium.remote.http.ConnectionFailedException: Unable to establish websocket connection to http://172.20.0.5:4444/session/b47bfd6fa8d3d37c813e127f67f8cb0d/se/cdp
Build info: version: '4.5.0', revision: 'fe167b119a'
System info: os.name: 'Mac OS X', os.arch: 'aarch64', os.version: '12.5', java.version: '15.0.8'
Driver info: driver.version: unknown
at org.openqa.selenium.remote.http.netty.NettyWebSocket.<init>(NettyWebSocket.java:102)
at org.openqa.selenium.remote.http.netty.NettyWebSocket.lambda$create$3(NettyWebSocket.java:128)
at org.openqa.selenium.remote.http.netty.NettyClient.openSocket(NettyClient.java:107)
at org.openqa.selenium.devtools.Connection.<init>(Connection.java:77)
at org.openqa.selenium.devtools.SeleniumCdpConnection.<init>(SeleniumCdpConnection.java:34)
at org.openqa.selenium.devtools.SeleniumCdpConnection.lambda$create$0(SeleniumCdpConnection.java:56)
at java.base/java.util.Optional.map(Optional.java:258)
at org.openqa.selenium.devtools.SeleniumCdpConnection.create(SeleniumCdpConnection.java:54)
at org.openqa.selenium.devtools.SeleniumCdpConnection.create(SeleniumCdpConnection.java:47)
at org.openqa.selenium.devtools.DevToolsProvider.getImplementation(DevToolsProvider.java:50)
at org.openqa.selenium.devtools.DevToolsProvider.getImplementation(DevToolsProvider.java:31)
at org.openqa.selenium.remote.Augmenter.augment(Augmenter.java:186)
at de.check24.testumgebung.TestListener.savePng(TestListener.java:64)
at de.check24.testumgebung.TestListener.createScreenshot(TestListener.java:42)
at de.check24.testumgebung.TestListener.onTestFailure(TestListener.java:160)
at org.testng.internal.TestListenerHelper.runTestListeners(TestListenerHelper.java:96)
at org.testng.internal.TestInvoker.runTestResultListener(TestInvoker.java:219)
at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:834)
at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:146)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1511)
at org.testng.TestRunner.privateRun(TestRunner.java:794)
at org.testng.TestRunner.run(TestRunner.java:596)
at org.testng.SuiteRunner.runTest(SuiteRunner.java:377)
at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:371)
at org.testng.SuiteRunner.privateRun(SuiteRunner.java:332)
at org.testng.SuiteRunner.run(SuiteRunner.java:276)
at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53)
at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96)
at org.testng.TestNG.runSuitesSequentially(TestNG.java:1212)
at org.testng.TestNG.runSuitesLocally(TestNG.java:1134)
at org.testng.TestNG.runSuites(TestNG.java:1063)
at org.testng.TestNG.run(TestNG.java:1031)
at com.intellij.rt.testng.IDEARemoteTestNG.run(IDEARemoteTestNG.java:66)
at com.intellij.rt.testng.RemoteTestNGStarter.main(RemoteTestNGStarter.java:109)
Disconnected from the target VM, address: '127.0.0.1:64835', transport: 'socket'
Process finished with exit code 255
```
### Operating System
macOS Monterey Version 12.5
### Selenium version
Selenium 4.5.0 | openjdk 15.0.8 2022-07-19 | Chrome-driver 4.5.0
### What are the browser(s) and version(s) where you see this issue?
Chrome v.106.0
### What are the browser driver(s) and version(s) where you see this issue?
browserVersion= v.106.0.5249.91
### Are you using Selenium Grid?
Selenium Grid 4.5.0 (revision fe167b119a) | non_infrastructure | unable to find an exact match for cdp version so returning the closest version found a no op implementation what happened i am trying to save a screenshot from my remotewebdriver session grid distributed mode and it looks like there is an issue with the websocket connection or the cdp version is not compatiable with selenium java version how can we reproduce the issue shell java code driver webdrivermanager chromedriver remoteaddress capabilities desiredcapabilitiesfactory createchromeoptions create webdriver augmenteddriver new augmenter augment driver srcfile takesscreenshot augmenteddriver getscreenshotas outputtype file relevant log output shell connected to the target vm address transport socket oct pm org openqa selenium remote tracing opentelemetry opentelemetrytracer createtracer info using opentelemetry for tracing oct pm org openqa selenium remote protocolhandshake createsession info detected upstream dialect oct pm org openqa selenium devtools cdpversionfinder findnearestmatch warning unable to find an exact match for cdp version so returning the closest version found a no op implementation oct pm org openqa selenium devtools cdpver oct pm org openqa selenium devtools cdpversionfinder findnearestmatch warning unable to find an exact match for cdp version so returning the closest version found a no op implementation oct pm org openqa selenium devtools cdpversionfinder findnearestmatch info unable to find cdp implementation matching oct pm org openqa selenium remote http netty nettywebsocket lambda new warning connection timed out java net connectexception connection timed out at org asynchttpclient netty channel nettyconnectlistener onfailure nettyconnectlistener java at org asynchttpclient netty channel nettychannelconnector onfailure nettychannelconnector java at org asynchttpclient netty simplechannelfuturelistener operationcomplete simplechannelfuturelistener java at org asynchttpclient netty simplechannelfuturelistener operationcomplete simplechannelfuturelistener java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise notifylistenersnow defaultpromise java at io netty util concurrent defaultpromise notifylisteners defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise tryfailure defaultpromise java at io netty channel nio abstractniochannel abstractniounsafe run abstractniochannel java at io netty util concurrent promisetask runtask promisetask java at io netty util concurrent scheduledfuturetask run scheduledfuturetask java at io netty util concurrent abstracteventexecutor runtask capture abstracteventexecutor java at io netty util concurrent abstracteventexecutor runtask abstracteventexecutor java at io netty util concurrent abstracteventexecutor safeexecute capture abstracteventexecutor java at io netty util concurrent abstracteventexecutor safeexecute abstracteventexecutor java at io netty util concurrent singlethreadeventexecutor runalltasks singlethreadeventexecutor java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by io netty channel connecttimeoutexception connection timed out at io netty channel nio abstractniochannel abstractniounsafe run abstractniochannel java more oct pm org openqa selenium remote http websocket listener onerror warning connection timed out java net connectexception connection timed out at org asynchttpclient netty channel nettyconnectlistener onfailure nettyconnectlistener java at org asynchttpclient netty channel nettychannelconnector onfailure nettychannelconnector java at org asynchttpclient netty simplechannelfuturelistener operationcomplete simplechannelfuturelistener java at org asynchttpclient netty simplechannelfuturelistener operationcomplete simplechannelfuturelistener java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise notifylistenersnow defaultpromise java at io netty util concurrent defaultpromise notifylisteners defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise tryfailure defaultpromise java at io netty channel nio abstractniochannel abstractniounsafe run abstractniochannel java at io netty util concurrent promisetask runtask promisetask java at io netty util concurrent scheduledfuturetask run scheduledfuturetask java at io netty util concurrent abstracteventexecutor runtask capture abstracteventexecutor java at io netty util concurrent abstracteventexecutor runtask abstracteventexecutor java at io netty util concurrent abstracteventexecutor safeexecute capture abstracteventexecutor java at io netty util concurrent abstracteventexecutor safeexecute abstracteventexecutor java at io netty util concurrent singlethreadeventexecutor runalltasks singlethreadeventexecutor java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by io netty channel connecttimeoutexception connection timed out at io netty channel nio abstractniochannel abstractniounsafe run abstractniochannel java more org openqa selenium remote http connectionfailedexception unable to establish websocket connection to build info version revision system info os name mac os x os arch os version java version driver info driver version unknown at org openqa selenium remote http netty nettywebsocket nettywebsocket java at org openqa selenium remote http netty nettywebsocket lambda create nettywebsocket java at org openqa selenium remote http netty nettyclient opensocket nettyclient java at org openqa selenium devtools connection connection java at org openqa selenium devtools seleniumcdpconnection seleniumcdpconnection java at org openqa selenium devtools seleniumcdpconnection lambda create seleniumcdpconnection java at java base java util optional map optional java at org openqa selenium devtools seleniumcdpconnection create seleniumcdpconnection java at org openqa selenium devtools seleniumcdpconnection create seleniumcdpconnection java at org openqa selenium devtools devtoolsprovider getimplementation devtoolsprovider java at org openqa selenium devtools devtoolsprovider getimplementation devtoolsprovider java at org openqa selenium remote augmenter augment augmenter java at de testumgebung testlistener savepng testlistener java at de testumgebung testlistener createscreenshot testlistener java at de testumgebung testlistener ontestfailure testlistener java at org testng internal testlistenerhelper runtestlisteners testlistenerhelper java at org testng internal testinvoker runtestresultlistener testinvoker java at org testng internal testinvoker methodinvocationagent invoke testinvoker java at org testng internal testinvoker invoketestmethods testinvoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at java base java util arraylist foreach arraylist java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at com intellij rt testng idearemotetestng run idearemotetestng java at com intellij rt testng remotetestngstarter main remotetestngstarter java disconnected from the target vm address transport socket process finished with exit code operating system macos monterey version selenium version selenium openjdk chrome driver what are the browser s and version s where you see this issue chrome v what are the browser driver s and version s where you see this issue browserversion v are you using selenium grid selenium grid revision | 0 |
4,274 | 4,955,701,228 | IssuesEvent | 2016-12-01 21:11:48 | mRemoteNG/mRemoteNG | https://api.github.com/repos/mRemoteNG/mRemoteNG | closed | BUG: HTTP links. Upgrade http links to https. | Project Infrastructure | GitHub automatically redirects to HTTPS and they even have HSTS deployed, just no preloading.
Can you check http links and upgrade them to https?
Places I noticed:
* Repo description: https://github.com/mRemoteNG/mRemoteNG
* Repo description: https://github.com/mRemoteNG
* README: http://waffle.io/mRemoteNG/mRemoteNG
* http://forum.mremoteng.org/ (https needs to be fully setup first before switching) | 1.0 | BUG: HTTP links. Upgrade http links to https. - GitHub automatically redirects to HTTPS and they even have HSTS deployed, just no preloading.
Can you check http links and upgrade them to https?
Places I noticed:
* Repo description: https://github.com/mRemoteNG/mRemoteNG
* Repo description: https://github.com/mRemoteNG
* README: http://waffle.io/mRemoteNG/mRemoteNG
* http://forum.mremoteng.org/ (https needs to be fully setup first before switching) | infrastructure | bug http links upgrade http links to https github automatically redirects to https and they even have hsts deployed just no preloading can you check http links and upgrade them to https places i noticed repo description repo description readme https needs to be fully setup first before switching | 1 |
433,861 | 12,511,664,615 | IssuesEvent | 2020-06-02 21:00:48 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Disable NANOAOD/NANOAODSIM output data placement in Unified | High Priority New Feature ReqMgr2MS Unified Porting | **Impact of the new feature**
Unified
**Is your feature request related to a problem? Please describe.**
It's not really a WMCore issue, but given that Unified has closer to 0 dev manpower now. I thought we could make this change for them, we have anyways investigated that code already.
**Describe the solution you'd like**
We should implement a configurable parameter for the Unified setup, such that we can skip any output data placements for a set of datatiers. Logging the datasets being skipped from the data placement.
Reason for that is, that it will be taken care of by the Rucio/DM team.
**Describe alternatives you've considered**
An alternative would be to hard-code those datatiers in the code... However, this functionality will likely be needed for other use cases, so having it configurable is likely a better idea.
**Additional context**
none | 1.0 | Disable NANOAOD/NANOAODSIM output data placement in Unified - **Impact of the new feature**
Unified
**Is your feature request related to a problem? Please describe.**
It's not really a WMCore issue, but given that Unified has closer to 0 dev manpower now. I thought we could make this change for them, we have anyways investigated that code already.
**Describe the solution you'd like**
We should implement a configurable parameter for the Unified setup, such that we can skip any output data placements for a set of datatiers. Logging the datasets being skipped from the data placement.
Reason for that is, that it will be taken care of by the Rucio/DM team.
**Describe alternatives you've considered**
An alternative would be to hard-code those datatiers in the code... However, this functionality will likely be needed for other use cases, so having it configurable is likely a better idea.
**Additional context**
none | non_infrastructure | disable nanoaod nanoaodsim output data placement in unified impact of the new feature unified is your feature request related to a problem please describe it s not really a wmcore issue but given that unified has closer to dev manpower now i thought we could make this change for them we have anyways investigated that code already describe the solution you d like we should implement a configurable parameter for the unified setup such that we can skip any output data placements for a set of datatiers logging the datasets being skipped from the data placement reason for that is that it will be taken care of by the rucio dm team describe alternatives you ve considered an alternative would be to hard code those datatiers in the code however this functionality will likely be needed for other use cases so having it configurable is likely a better idea additional context none | 0 |
22,356 | 19,144,942,898 | IssuesEvent | 2021-12-02 06:14:48 | friendsofagape/autographa | https://api.github.com/repos/friendsofagape/autographa | closed | Resources Tab- The Close button image not loading | usability cosmetic ag2.0.0-beta.1 | The close button icon does not loaded properly in the Resources Tab,
Impacts the following options
Translation Words
Notes
Bible
Questions
Map
Commentary

| True | Resources Tab- The Close button image not loading - The close button icon does not loaded properly in the Resources Tab,
Impacts the following options
Translation Words
Notes
Bible
Questions
Map
Commentary

| non_infrastructure | resources tab the close button image not loading the close button icon does not loaded properly in the resources tab impacts the following options translation words notes bible questions map commentary | 0 |
33,672 | 27,706,797,877 | IssuesEvent | 2023-03-14 11:46:53 | gap-system/gap | https://api.github.com/repos/gap-system/gap | closed | CI failure in `Wrap releases / Create Windows x86_64 installer` | os: windows topic: infrastructure topic: ci | These CI jobs started to fail a few days ago, including scheduled jobs on master. See e.g.
- https://github.com/gap-system/gap/actions/runs/4348534499/jobs/7597530064
- https://github.com/gap-system/gap/actions/runs/4350407611
Excerpt of the logs just before the failure:
```
config.status: creating doc/make_doc
/bin/sh: SAGE_ROOT: command not found
config.status: creating doc/versiondata
config.status: creating build/config.h
config.status: executing build/stamp-h commands
config.status: executing libtool commands
Makefile:249: update target 'envs/build-4.13dev-340-gd600d6b-x86_64/opt/gap-4.13dev-340-gd600d6b/Makefile' due to: target does not exist
tools/subcyg "envs/build-4.13dev-340-gd600d6b-x86_64" "cd $(SAGE_ROOT) && make -j2"
make: *** No targets specified and no makefile found. Stop.
make: *** [Makefile:249: envs/build-4.13dev-340-gd600d6b-x86_64/opt/gap-4.13dev-340-gd600d6b/Makefile] Error 2
sha256sum: Output/gap-4.13dev-340-gd600d6b-x86_64.exe: No such file or directory
D:\a\_temp\66c9931e-7e7d-4118-ae3b-369d073d7d19: line 5: Output/gap-4.13dev-340-gd600d6b-x86_64.exe.sha256: No such file or directory
```
This is hard to debug as the code we use is derived from SageMath code for this and uses a complex `Makefile`, which is not that easy to understand, see https://github.com/ChrisJefferson/sage-windows (related: #5288). It is complex to deal with the size of SageMath; but since GAP is much much smaller, it might be possible to rewrite it as, say, a simpler shell script, which then perhaps is easier to debug. @ChrisJefferson is in principle willing to work on this, but of course has other obligations and so it can take some time.
Until then, we'll just have to ignore this part of the tests. | 1.0 | CI failure in `Wrap releases / Create Windows x86_64 installer` - These CI jobs started to fail a few days ago, including scheduled jobs on master. See e.g.
- https://github.com/gap-system/gap/actions/runs/4348534499/jobs/7597530064
- https://github.com/gap-system/gap/actions/runs/4350407611
Excerpt of the logs just before the failure:
```
config.status: creating doc/make_doc
/bin/sh: SAGE_ROOT: command not found
config.status: creating doc/versiondata
config.status: creating build/config.h
config.status: executing build/stamp-h commands
config.status: executing libtool commands
Makefile:249: update target 'envs/build-4.13dev-340-gd600d6b-x86_64/opt/gap-4.13dev-340-gd600d6b/Makefile' due to: target does not exist
tools/subcyg "envs/build-4.13dev-340-gd600d6b-x86_64" "cd $(SAGE_ROOT) && make -j2"
make: *** No targets specified and no makefile found. Stop.
make: *** [Makefile:249: envs/build-4.13dev-340-gd600d6b-x86_64/opt/gap-4.13dev-340-gd600d6b/Makefile] Error 2
sha256sum: Output/gap-4.13dev-340-gd600d6b-x86_64.exe: No such file or directory
D:\a\_temp\66c9931e-7e7d-4118-ae3b-369d073d7d19: line 5: Output/gap-4.13dev-340-gd600d6b-x86_64.exe.sha256: No such file or directory
```
This is hard to debug as the code we use is derived from SageMath code for this and uses a complex `Makefile`, which is not that easy to understand, see https://github.com/ChrisJefferson/sage-windows (related: #5288). It is complex to deal with the size of SageMath; but since GAP is much much smaller, it might be possible to rewrite it as, say, a simpler shell script, which then perhaps is easier to debug. @ChrisJefferson is in principle willing to work on this, but of course has other obligations and so it can take some time.
Until then, we'll just have to ignore this part of the tests. | infrastructure | ci failure in wrap releases create windows installer these ci jobs started to fail a few days ago including scheduled jobs on master see e g excerpt of the logs just before the failure config status creating doc make doc bin sh sage root command not found config status creating doc versiondata config status creating build config h config status executing build stamp h commands config status executing libtool commands makefile update target envs build opt gap makefile due to target does not exist tools subcyg envs build cd sage root make make no targets specified and no makefile found stop make error output gap exe no such file or directory d a temp line output gap exe no such file or directory this is hard to debug as the code we use is derived from sagemath code for this and uses a complex makefile which is not that easy to understand see related it is complex to deal with the size of sagemath but since gap is much much smaller it might be possible to rewrite it as say a simpler shell script which then perhaps is easier to debug chrisjefferson is in principle willing to work on this but of course has other obligations and so it can take some time until then we ll just have to ignore this part of the tests | 1 |
10,196 | 8,410,299,587 | IssuesEvent | 2018-10-12 10:07:10 | ressec/hemajoo-gaming | https://api.github.com/repos/ressec/hemajoo-gaming | opened | GitHub not receiving anymore Travis notification | Bug Development High Infrastructure | ## Purpose
Repositories using **GitHub** Apps now receive Build status updates via the Checks API only.
If you're using protected branches or custom status automation, you may need to adjust them. | 1.0 | GitHub not receiving anymore Travis notification - ## Purpose
Repositories using **GitHub** Apps now receive Build status updates via the Checks API only.
If you're using protected branches or custom status automation, you may need to adjust them. | infrastructure | github not receiving anymore travis notification purpose repositories using github apps now receive build status updates via the checks api only if you re using protected branches or custom status automation you may need to adjust them | 1 |
5,461 | 2,775,072,167 | IssuesEvent | 2015-05-04 14:04:37 | timvideos/gst-switch | https://api.github.com/repos/timvideos/gst-switch | closed | Cleanup Python-API unit test output | cleanup python-api tests | The current unit test run output has a lot of "junk" in it, see below;
Issues seem to be;
- [ ] print on composite mode change `Set composite mode to 0`
- [ ] printing of some tuples; `(False,)`
- [ ] printing of `[<unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be90>, ...`
- [ ] printing of source ending - `End source with pattern 1`
- [ ] printing of "wave"
- [ ] printing of format strings `video/x-raw, format=(string)I420, width=300, height=200`
- [ ] an error message about gdouble issue; `/home/travis/build/mithro/gst-switch/python-api/gstswitch/testsource.py:203: Warning: value "65535.000000" of type 'gdouble' is invalid or out of range for property 'freq' of type 'gdouble'
element.set_property('freq', int(freq)) `
Full log;
```
============================= test session starts ==============================
platform linux2 -- Python 2.7.6 -- py-1.4.26 -- pytest-2.6.4 -- /usr/bin/python
cachedir: /home/travis/build/mithro/gst-switch/python-api/.cache
plugins: cov, cache, pep8
collecting 71 items
(gst-plugin-scanner:10242): Clutter-CRITICAL **: Unable to initialize Clutter: Unable to open display. You have to set the DISPLAY environment variable, or use the --display command line argument
libdc1394 error: Failed to initialize libdc1394
collected 183 items
tests/unittests/__init__.py::PEP8-check PASSED
tests/unittests/test_connection_unit.py::PEP8-check PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_null PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_colon PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_normal PASSED
tests/unittests/test_connection_unit.py::TestBusName::test_normal PASSED
tests/unittests/test_connection_unit.py::TestBusName::test_normal_none PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_blank PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_slash PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_normal PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_none PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_dot PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_normal PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address2 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address3 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_mock1 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_mock2 PASSED
tests/unittests/test_connection_unit.py::test_get_compose_port PASSED
tests/unittests/test_connection_unit.py::test_get_encode_port PASSED
tests/unittests/test_connection_unit.py::test_get_audio_port PASSED
tests/unittests/test_connection_unit.py::test_get_preview_ports PASSED
tests/unittests/test_connection_unit.py::test_set_composite_mode PASSED
tests/unittests/test_connection_unit.py::test_set_encode_mode PASSED
tests/unittests/test_connection_unit.py::test_new_record PASSED
tests/unittests/test_connection_unit.py::test_adjust_pip PASSED
tests/unittests/test_connection_unit.py::test_switch PASSED
tests/unittests/test_connection_unit.py::test_click_video PASSED
tests/unittests/test_connection_unit.py::test_mark_face PASSED
tests/unittests/test_connection_unit.py::test_mark_tracking PASSED
tests/unittests/test_controller_unit.py::PEP8-check PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_null PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_colon PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_normal PASSED
tests/unittests/test_controller_unit.py::TestBusName::test_normal PASSED
tests/unittests/test_controller_unit.py::TestBusName::test_normal_none PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_blank PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_slash PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_normal PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_none PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_dot PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_normal PASSED
tests/unittests/test_controller_unit.py::TestEstablishConnection::test_normal PASSED
tests/unittests/test_controller_unit.py::TestGetComposePort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetComposePort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetEncodePort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetEncodePort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetAudioPort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetAudioPort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetPreviewPorts::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetPreviewPorts::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestSetCompositeMode::test_unpack (False,)
PASSED
tests/unittests/test_controller_unit.py::TestSetCompositeMode::test_normal_unpack (true,)
Set composite mode to 0
PASSED
tests/unittests/test_controller_unit.py::TestSetEncodeMode::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestSetEncodeMode::test_normal_unpack Set encode mode to 1
PASSED
tests/unittests/test_controller_unit.py::TestNewRecord::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestNewRecord::test_normal_unpack New record
PASSED
tests/unittests/test_controller_unit.py::TestAdjustPIP::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestAdjustPIP::test_normal_unpack adjust pip xpos:1 ypos:2 w:3 h:4
PASSED
tests/unittests/test_controller_unit.py::TestSwitch::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestSwitch::test_normal_unpack Switch channel:65 port:2
PASSED
tests/unittests/test_controller_unit.py::TestClickVideo::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestClickVideo::test_normal_unpack Click video: xpos:1 ypos:2 width:3 height:4
PASSED
tests/unittests/test_controller_unit.py::TestMarkFaces::test_normal PASSED
tests/unittests/test_controller_unit.py::TestMarkTracking::test_normal PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_value_error PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_syntax_error PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_normal PASSED
tests/unittests/test_helpers_unit.py::PEP8-check PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_new_test_video PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_get_test_video 0 pattern: 1
1 pattern: 2
2 pattern: 3
3 pattern: 19
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_error_video PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_normal_video End source with pattern 1
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate1_video TESTS: [<unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be90>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be50>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4bd90>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be10>]
End source with pattern 1
End source with pattern 2
End source with pattern 3
End source with pattern 19
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate2_video TESTS: []
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_new_test_audio PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_get_test_audio 0 wave: 1
1 wave: 2
2 wave: 3
3 wave: 10
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_error_audio PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_normal_audio End source with wave 1
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate1_audio TESTS: [<unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b273d0>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b27550>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b27450>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b271d0>]
End source with wave 1
End source with wave 2
End source with wave 3
End source with wave 10
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate2_audio TESTS: []
PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_blank PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_run start preview
PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_terminate_fail PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_terminate_normal end preview
PASSED
tests/unittests/test_server_unit.py::PEP8-check PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_provided_slash PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_provided_no_slash PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_empty PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_null PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_type PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_range PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_null PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_type PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_range PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_null PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_type PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_range PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_false PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_true PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid_date PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid_space PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_invalid PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_slashes PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_terminate_fail Killing server
PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_kill_fail PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_kill PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_terminate Killing server
PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_gov_flush PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_gcov_flush_fail GCOV FLUSH
PASSED
tests/unittests/test_server_unit.py::TestRun::test_run Starting server
PASSED
tests/unittests/test_server_unit.py::TestRun::test_run_process PASSED
tests/unittests/test_server_unit.py::TestRun::test_start_process_error Creating process cmd
PASSED
tests/unittests/test_server_unit.py::TestRun::test_start_process_normal Creating process cmd
cmd
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_terminate Killing server
Server Killed
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_kill PASSED
tests/unittests/test_server_unit.py::TestNormal::test_terminate Killing server
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_terminate_cov Killing server
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_kill PASSED
tests/unittests/test_server_unit.py::TestNormal::test_kill_cov PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_gcov_flush GCOV FLUSH
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_make_coverage /home/travis/build/mithro/gst-switch/tools/
0
PASSED
tests/unittests/test_testsource_unit.py::PEP8-check PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_non_positive PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_normal video/x-raw, format=(string)I420, width=1000000.0, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=200, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_non_positive PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_normal video/x-raw, format=(string)I420, width=300, height=1000000.0
video/x-raw, format=(string)I420, width=300, height=300
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_range PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcTimeOverlay::test_fail PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcTimeOverlay::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcClockOverlay::test_fail PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcClockOverlay::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_run video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_pause video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_end video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_run PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_end PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_play PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_disable PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_1 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_2 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_3 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_4 video/x-raw, format=(string)I420, width=300, height=200
(python:10239): GStreamer-WARNING **: Name 'timeoverlay' is not unique in bin 'gstswitch+testsource+videopipeline23', not adding
(python:10239): GStreamer-WARNING **: Name 'clockoverlay' is not unique in bin 'gstswitch+testsource+videopipeline23', not adding
PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_normal /home/travis/build/mithro/gst-switch/python-api/gstswitch/testsource.py:203: Warning: value "65535.000000" of type 'gdouble' is invalid or out of range for property 'freq' of type 'gdouble'
element.set_property('freq', int(freq))
PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_run PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_end PASSED
--------------- coverage: platform linux2, python 2.7.6-final-0 ----------------
Name Stmts Miss Cover
------------------------------------------
gstswitch/__init__ 0 0 100%
gstswitch/connection 183 0 100%
gstswitch/controller 171 0 100%
gstswitch/exception 15 0 100%
gstswitch/helpers 129 0 100%
gstswitch/server 161 3 98%
gstswitch/testsource 328 0 100%
------------------------------------------
TOTAL 987 3 99%
==================== 183 passed, 1 warnings in 2.73 seconds ====================
The command "/trusty/run.py $PWD/.travis-run.sh" exited with 0.
``` | 1.0 | Cleanup Python-API unit test output - The current unit test run output has a lot of "junk" in it, see below;
Issues seem to be;
- [ ] print on composite mode change `Set composite mode to 0`
- [ ] printing of some tuples; `(False,)`
- [ ] printing of `[<unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be90>, ...`
- [ ] printing of source ending - `End source with pattern 1`
- [ ] printing of "wave"
- [ ] printing of format strings `video/x-raw, format=(string)I420, width=300, height=200`
- [ ] an error message about gdouble issue; `/home/travis/build/mithro/gst-switch/python-api/gstswitch/testsource.py:203: Warning: value "65535.000000" of type 'gdouble' is invalid or out of range for property 'freq' of type 'gdouble'
element.set_property('freq', int(freq)) `
Full log;
```
============================= test session starts ==============================
platform linux2 -- Python 2.7.6 -- py-1.4.26 -- pytest-2.6.4 -- /usr/bin/python
cachedir: /home/travis/build/mithro/gst-switch/python-api/.cache
plugins: cov, cache, pep8
collecting 71 items
(gst-plugin-scanner:10242): Clutter-CRITICAL **: Unable to initialize Clutter: Unable to open display. You have to set the DISPLAY environment variable, or use the --display command line argument
libdc1394 error: Failed to initialize libdc1394
collected 183 items
tests/unittests/__init__.py::PEP8-check PASSED
tests/unittests/test_connection_unit.py::PEP8-check PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_null PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_colon PASSED
tests/unittests/test_connection_unit.py::TestAddress::test_address_normal PASSED
tests/unittests/test_connection_unit.py::TestBusName::test_normal PASSED
tests/unittests/test_connection_unit.py::TestBusName::test_normal_none PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_blank PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_slash PASSED
tests/unittests/test_connection_unit.py::TestObjectPath::test_object_path_normal PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_none PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_dot PASSED
tests/unittests/test_connection_unit.py::TestInterface::test_interface_normal PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address2 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_bad_address3 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_mock1 PASSED
tests/unittests/test_connection_unit.py::TestConnectDBus::test_mock2 PASSED
tests/unittests/test_connection_unit.py::test_get_compose_port PASSED
tests/unittests/test_connection_unit.py::test_get_encode_port PASSED
tests/unittests/test_connection_unit.py::test_get_audio_port PASSED
tests/unittests/test_connection_unit.py::test_get_preview_ports PASSED
tests/unittests/test_connection_unit.py::test_set_composite_mode PASSED
tests/unittests/test_connection_unit.py::test_set_encode_mode PASSED
tests/unittests/test_connection_unit.py::test_new_record PASSED
tests/unittests/test_connection_unit.py::test_adjust_pip PASSED
tests/unittests/test_connection_unit.py::test_switch PASSED
tests/unittests/test_connection_unit.py::test_click_video PASSED
tests/unittests/test_connection_unit.py::test_mark_face PASSED
tests/unittests/test_connection_unit.py::test_mark_tracking PASSED
tests/unittests/test_controller_unit.py::PEP8-check PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_null PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_colon PASSED
tests/unittests/test_controller_unit.py::TestAddress::test_address_normal PASSED
tests/unittests/test_controller_unit.py::TestBusName::test_normal PASSED
tests/unittests/test_controller_unit.py::TestBusName::test_normal_none PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_blank PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_slash PASSED
tests/unittests/test_controller_unit.py::TestObjectPath::test_object_path_normal PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_none PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_dot PASSED
tests/unittests/test_controller_unit.py::TestInterface::test_interface_normal PASSED
tests/unittests/test_controller_unit.py::TestEstablishConnection::test_normal PASSED
tests/unittests/test_controller_unit.py::TestGetComposePort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetComposePort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetEncodePort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetEncodePort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetAudioPort::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetAudioPort::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetPreviewPorts::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestGetPreviewPorts::test_normal_unpack PASSED
tests/unittests/test_controller_unit.py::TestSetCompositeMode::test_unpack (False,)
PASSED
tests/unittests/test_controller_unit.py::TestSetCompositeMode::test_normal_unpack (true,)
Set composite mode to 0
PASSED
tests/unittests/test_controller_unit.py::TestSetEncodeMode::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestSetEncodeMode::test_normal_unpack Set encode mode to 1
PASSED
tests/unittests/test_controller_unit.py::TestNewRecord::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestNewRecord::test_normal_unpack New record
PASSED
tests/unittests/test_controller_unit.py::TestAdjustPIP::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestAdjustPIP::test_normal_unpack adjust pip xpos:1 ypos:2 w:3 h:4
PASSED
tests/unittests/test_controller_unit.py::TestSwitch::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestSwitch::test_normal_unpack Switch channel:65 port:2
PASSED
tests/unittests/test_controller_unit.py::TestClickVideo::test_unpack PASSED
tests/unittests/test_controller_unit.py::TestClickVideo::test_normal_unpack Click video: xpos:1 ypos:2 width:3 height:4
PASSED
tests/unittests/test_controller_unit.py::TestMarkFaces::test_normal PASSED
tests/unittests/test_controller_unit.py::TestMarkTracking::test_normal PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_value_error PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_syntax_error PASSED
tests/unittests/test_controller_unit.py::TestParsePreviewPorts::test_normal PASSED
tests/unittests/test_helpers_unit.py::PEP8-check PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesVideoPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestTestSourcesAudioPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_new_test_video PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_get_test_video 0 pattern: 1
1 pattern: 2
2 pattern: 3
3 pattern: 19
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_error_video PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_normal_video End source with pattern 1
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate1_video TESTS: [<unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be90>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be50>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4bd90>, <unittests.test_helpers_unit.MockTest object at 0x2b0b04b4be10>]
End source with pattern 1
End source with pattern 2
End source with pattern 3
End source with pattern 19
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate2_video TESTS: []
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_new_test_audio PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_get_test_audio 0 wave: 1
1 wave: 2
2 wave: 3
3 wave: 10
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_error_audio PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate_index_normal_audio End source with wave 1
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate1_audio TESTS: [<unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b273d0>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b27550>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b27450>, <unittests.test_helpers_unit.MockTest2 object at 0x2b0b04b271d0>]
End source with wave 1
End source with wave 2
End source with wave 3
End source with wave 10
PASSED
tests/unittests/test_helpers_unit.py::TestTestSources::test_terminate2_audio TESTS: []
PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_blank PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_range PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_invalid PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinksPreviewPort::test_normal PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_run start preview
PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_terminate_fail PASSED
tests/unittests/test_helpers_unit.py::TestPreviewSinks::test_terminate_normal end preview
PASSED
tests/unittests/test_server_unit.py::PEP8-check PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_provided_slash PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_provided_no_slash PASSED
tests/unittests/test_server_unit.py::TestPath::test_path_empty PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_null PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_type PASSED
tests/unittests/test_server_unit.py::TestVideoPort::test_invalid_video_port_range PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_null PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_type PASSED
tests/unittests/test_server_unit.py::TestAudioPort::test_invalid_audio_port_range PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_null PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_type PASSED
tests/unittests/test_server_unit.py::TestControlPort::test_invalid_control_port_range PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_false PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_true PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid_date PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_valid_space PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_invalid PASSED
tests/unittests/test_server_unit.py::TestRecordFile::test_record_file_slashes PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_terminate_fail Killing server
PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_kill_fail PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_kill PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_terminate Killing server
PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_no_process_gov_flush PASSED
tests/unittests/test_server_unit.py::TestKillTerminate::test_gcov_flush_fail GCOV FLUSH
PASSED
tests/unittests/test_server_unit.py::TestRun::test_run Starting server
PASSED
tests/unittests/test_server_unit.py::TestRun::test_run_process PASSED
tests/unittests/test_server_unit.py::TestRun::test_start_process_error Creating process cmd
PASSED
tests/unittests/test_server_unit.py::TestRun::test_start_process_normal Creating process cmd
cmd
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_terminate Killing server
Server Killed
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_kill PASSED
tests/unittests/test_server_unit.py::TestNormal::test_terminate Killing server
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_terminate_cov Killing server
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_kill PASSED
tests/unittests/test_server_unit.py::TestNormal::test_kill_cov PASSED
tests/unittests/test_server_unit.py::TestNormal::test_normal_gcov_flush GCOV FLUSH
PASSED
tests/unittests/test_server_unit.py::TestNormal::test_make_coverage /home/travis/build/mithro/gst-switch/tools/
0
PASSED
tests/unittests/test_testsource_unit.py::PEP8-check PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPort::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_non_positive PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcWidth::test_normal video/x-raw, format=(string)I420, width=1000000.0, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=200, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_non_positive PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcHeight::test_normal video/x-raw, format=(string)I420, width=300, height=1000000.0
video/x-raw, format=(string)I420, width=300, height=300
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_range PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPattern::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcTimeOverlay::test_fail PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcTimeOverlay::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcClockOverlay::test_fail PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcClockOverlay::test_normal video/x-raw, format=(string)I420, width=300, height=200
video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_run video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_pause video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoSrcPlay::test_end video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPort::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_run PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestPreviewPlay::test_end PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_play PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestBasePipeline::test_disable PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_1 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_2 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_3 video/x-raw, format=(string)I420, width=300, height=200
PASSED
tests/unittests/test_testsource_unit.py::TestVideoPipeline::test_permuate_time_clock_4 video/x-raw, format=(string)I420, width=300, height=200
(python:10239): GStreamer-WARNING **: Name 'timeoverlay' is not unique in bin 'gstswitch+testsource+videopipeline23', not adding
(python:10239): GStreamer-WARNING **: Name 'clockoverlay' is not unique in bin 'gstswitch+testsource+videopipeline23', not adding
PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPort::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_blank PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcFreq::test_normal /home/travis/build/mithro/gst-switch/python-api/gstswitch/testsource.py:203: Warning: value "65535.000000" of type 'gdouble' is invalid or out of range for property 'freq' of type 'gdouble'
element.set_property('freq', int(freq))
PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_range PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_invalid PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcWave::test_normal PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_run PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_pause PASSED
tests/unittests/test_testsource_unit.py::TestAudioSrcPlay::test_end PASSED
--------------- coverage: platform linux2, python 2.7.6-final-0 ----------------
Name Stmts Miss Cover
------------------------------------------
gstswitch/__init__ 0 0 100%
gstswitch/connection 183 0 100%
gstswitch/controller 171 0 100%
gstswitch/exception 15 0 100%
gstswitch/helpers 129 0 100%
gstswitch/server 161 3 98%
gstswitch/testsource 328 0 100%
------------------------------------------
TOTAL 987 3 99%
==================== 183 passed, 1 warnings in 2.73 seconds ====================
The command "/trusty/run.py $PWD/.travis-run.sh" exited with 0.
``` | non_infrastructure | cleanup python api unit test output the current unit test run output has a lot of junk in it see below issues seem to be print on composite mode change set composite mode to printing of some tuples false printing of printing of source ending end source with pattern printing of wave printing of format strings video x raw format string width height an error message about gdouble issue home travis build mithro gst switch python api gstswitch testsource py warning value of type gdouble is invalid or out of range for property freq of type gdouble element set property freq int freq full log test session starts platform python py pytest usr bin python cachedir home travis build mithro gst switch python api cache plugins cov cache collecting items gst plugin scanner clutter critical unable to initialize clutter unable to open display you have to set the display environment variable or use the display command line argument error failed to initialize collected items tests unittests init py check passed tests unittests test connection unit py check passed tests unittests test connection unit py testaddress test address null passed tests unittests test connection unit py testaddress test address colon passed tests unittests test connection unit py testaddress test address normal passed tests unittests test connection unit py testbusname test normal passed tests unittests test connection unit py testbusname test normal none passed tests unittests test connection unit py testobjectpath test object path blank passed tests unittests test connection unit py testobjectpath test object path slash passed tests unittests test connection unit py testobjectpath test object path normal passed tests unittests test connection unit py testinterface test interface none passed tests unittests test connection unit py testinterface test interface dot passed tests unittests test connection unit py testinterface test interface normal passed tests unittests test connection unit py testconnectdbus test bad address passed tests unittests test connection unit py testconnectdbus test bad passed tests unittests test connection unit py testconnectdbus test bad passed tests unittests test connection unit py testconnectdbus test passed tests unittests test connection unit py testconnectdbus test passed tests unittests test connection unit py test get compose port passed tests unittests test connection unit py test get encode port passed tests unittests test connection unit py test get audio port passed tests unittests test connection unit py test get preview ports passed tests unittests test connection unit py test set composite mode passed tests unittests test connection unit py test set encode mode passed tests unittests test connection unit py test new record passed tests unittests test connection unit py test adjust pip passed tests unittests test connection unit py test switch passed tests unittests test connection unit py test click video passed tests unittests test connection unit py test mark face passed tests unittests test connection unit py test mark tracking passed tests unittests test controller unit py check passed tests unittests test controller unit py testaddress test address null passed tests unittests test controller unit py testaddress test address colon passed tests unittests test controller unit py testaddress test address normal passed tests unittests test controller unit py testbusname test normal passed tests unittests test controller unit py testbusname test normal none passed tests unittests test controller unit py testobjectpath test object path blank passed tests unittests test controller unit py testobjectpath test object path slash passed tests unittests test controller unit py testobjectpath test object path normal passed tests unittests test controller unit py testinterface test interface none passed tests unittests test controller unit py testinterface test interface dot passed tests unittests test controller unit py testinterface test interface normal passed tests unittests test controller unit py testestablishconnection test normal passed tests unittests test controller unit py testgetcomposeport test unpack passed tests unittests test controller unit py testgetcomposeport test normal unpack passed tests unittests test controller unit py testgetencodeport test unpack passed tests unittests test controller unit py testgetencodeport test normal unpack passed tests unittests test controller unit py testgetaudioport test unpack passed tests unittests test controller unit py testgetaudioport test normal unpack passed tests unittests test controller unit py testgetpreviewports test unpack passed tests unittests test controller unit py testgetpreviewports test normal unpack passed tests unittests test controller unit py testsetcompositemode test unpack false passed tests unittests test controller unit py testsetcompositemode test normal unpack true set composite mode to passed tests unittests test controller unit py testsetencodemode test unpack passed tests unittests test controller unit py testsetencodemode test normal unpack set encode mode to passed tests unittests test controller unit py testnewrecord test unpack passed tests unittests test controller unit py testnewrecord test normal unpack new record passed tests unittests test controller unit py testadjustpip test unpack passed tests unittests test controller unit py testadjustpip test normal unpack adjust pip xpos ypos w h passed tests unittests test controller unit py testswitch test unpack passed tests unittests test controller unit py testswitch test normal unpack switch channel port passed tests unittests test controller unit py testclickvideo test unpack passed tests unittests test controller unit py testclickvideo test normal unpack click video xpos ypos width height passed tests unittests test controller unit py testmarkfaces test normal passed tests unittests test controller unit py testmarktracking test normal passed tests unittests test controller unit py testparsepreviewports test value error passed tests unittests test controller unit py testparsepreviewports test syntax error passed tests unittests test controller unit py testparsepreviewports test normal passed tests unittests test helpers unit py check passed tests unittests test helpers unit py testtestsourcesvideoport test range passed tests unittests test helpers unit py testtestsourcesvideoport test invalid passed tests unittests test helpers unit py testtestsourcesvideoport test normal passed tests unittests test helpers unit py testtestsourcesaudioport test range passed tests unittests test helpers unit py testtestsourcesaudioport test invalid passed tests unittests test helpers unit py testtestsourcesaudioport test normal passed tests unittests test helpers unit py testtestsources test new test video passed tests unittests test helpers unit py testtestsources test get test video pattern pattern pattern pattern passed tests unittests test helpers unit py testtestsources test terminate index error video passed tests unittests test helpers unit py testtestsources test terminate index normal video end source with pattern passed tests unittests test helpers unit py testtestsources test video tests end source with pattern end source with pattern end source with pattern end source with pattern passed tests unittests test helpers unit py testtestsources test video tests passed tests unittests test helpers unit py testtestsources test new test audio passed tests unittests test helpers unit py testtestsources test get test audio wave wave wave wave passed tests unittests test helpers unit py testtestsources test terminate index error audio passed tests unittests test helpers unit py testtestsources test terminate index normal audio end source with wave passed tests unittests test helpers unit py testtestsources test audio tests end source with wave end source with wave end source with wave end source with wave passed tests unittests test helpers unit py testtestsources test audio tests passed tests unittests test helpers unit py testpreviewsinkspreviewport test blank passed tests unittests test helpers unit py testpreviewsinkspreviewport test range passed tests unittests test helpers unit py testpreviewsinkspreviewport test invalid passed tests unittests test helpers unit py testpreviewsinkspreviewport test normal passed tests unittests test helpers unit py testpreviewsinks test run start preview passed tests unittests test helpers unit py testpreviewsinks test terminate fail passed tests unittests test helpers unit py testpreviewsinks test terminate normal end preview passed tests unittests test server unit py check passed tests unittests test server unit py testpath test path provided slash passed tests unittests test server unit py testpath test path provided no slash passed tests unittests test server unit py testpath test path empty passed tests unittests test server unit py testvideoport test invalid video port null passed tests unittests test server unit py testvideoport test invalid video port type passed tests unittests test server unit py testvideoport test invalid video port range passed tests unittests test server unit py testaudioport test invalid audio port null passed tests unittests test server unit py testaudioport test invalid audio port type passed tests unittests test server unit py testaudioport test invalid audio port range passed tests unittests test server unit py testcontrolport test invalid control port null passed tests unittests test server unit py testcontrolport test invalid control port type passed tests unittests test server unit py testcontrolport test invalid control port range passed tests unittests test server unit py testrecordfile test record file false passed tests unittests test server unit py testrecordfile test record file true passed tests unittests test server unit py testrecordfile test record file valid passed tests unittests test server unit py testrecordfile test record file valid date passed tests unittests test server unit py testrecordfile test record file valid space passed tests unittests test server unit py testrecordfile test record file invalid passed tests unittests test server unit py testrecordfile test record file slashes passed tests unittests test server unit py testkillterminate test terminate fail killing server passed tests unittests test server unit py testkillterminate test kill fail passed tests unittests test server unit py testkillterminate test no process kill passed tests unittests test server unit py testkillterminate test no process terminate killing server passed tests unittests test server unit py testkillterminate test no process gov flush passed tests unittests test server unit py testkillterminate test gcov flush fail gcov flush passed tests unittests test server unit py testrun test run starting server passed tests unittests test server unit py testrun test run process passed tests unittests test server unit py testrun test start process error creating process cmd passed tests unittests test server unit py testrun test start process normal creating process cmd cmd passed tests unittests test server unit py testnormal test normal terminate killing server server killed passed tests unittests test server unit py testnormal test normal kill passed tests unittests test server unit py testnormal test terminate killing server passed tests unittests test server unit py testnormal test terminate cov killing server passed tests unittests test server unit py testnormal test kill passed tests unittests test server unit py testnormal test kill cov passed tests unittests test server unit py testnormal test normal gcov flush gcov flush passed tests unittests test server unit py testnormal test make coverage home travis build mithro gst switch tools passed tests unittests test testsource unit py check passed tests unittests test testsource unit py testvideosrcport test blank passed tests unittests test testsource unit py testvideosrcport test range passed tests unittests test testsource unit py testvideosrcport test invalid passed tests unittests test testsource unit py testvideosrcport test normal video x raw format string width height video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrcwidth test blank passed tests unittests test testsource unit py testvideosrcwidth test non positive passed tests unittests test testsource unit py testvideosrcwidth test invalid passed tests unittests test testsource unit py testvideosrcwidth test normal video x raw format string width height video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrcheight test blank passed tests unittests test testsource unit py testvideosrcheight test non positive passed tests unittests test testsource unit py testvideosrcheight test invalid passed tests unittests test testsource unit py testvideosrcheight test normal video x raw format string width height video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrcpattern test range passed tests unittests test testsource unit py testvideosrcpattern test invalid passed tests unittests test testsource unit py testvideosrcpattern test normal video x raw format string width height video x raw format string width height video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrctimeoverlay test fail passed tests unittests test testsource unit py testvideosrctimeoverlay test normal video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrcclockoverlay test fail passed tests unittests test testsource unit py testvideosrcclockoverlay test normal video x raw format string width height video x raw format string width height passed tests unittests test testsource unit py testvideosrcplay test run video x raw format string width height passed tests unittests test testsource unit py testvideosrcplay test pause video x raw format string width height passed tests unittests test testsource unit py testvideosrcplay test end video x raw format string width height passed tests unittests test testsource unit py testpreviewport test blank passed tests unittests test testsource unit py testpreviewport test range passed tests unittests test testsource unit py testpreviewport test invalid passed tests unittests test testsource unit py testpreviewport test normal passed tests unittests test testsource unit py testpreviewplay test run passed tests unittests test testsource unit py testpreviewplay test pause passed tests unittests test testsource unit py testpreviewplay test end passed tests unittests test testsource unit py testbasepipeline test play passed tests unittests test testsource unit py testbasepipeline test pause passed tests unittests test testsource unit py testbasepipeline test disable passed tests unittests test testsource unit py testvideopipeline test permuate time clock video x raw format string width height passed tests unittests test testsource unit py testvideopipeline test permuate time clock video x raw format string width height passed tests unittests test testsource unit py testvideopipeline test permuate time clock video x raw format string width height passed tests unittests test testsource unit py testvideopipeline test permuate time clock video x raw format string width height python gstreamer warning name timeoverlay is not unique in bin gstswitch testsource not adding python gstreamer warning name clockoverlay is not unique in bin gstswitch testsource not adding passed tests unittests test testsource unit py testaudiosrcport test blank passed tests unittests test testsource unit py testaudiosrcport test range passed tests unittests test testsource unit py testaudiosrcport test invalid passed tests unittests test testsource unit py testaudiosrcport test normal passed tests unittests test testsource unit py testaudiosrcfreq test blank passed tests unittests test testsource unit py testaudiosrcfreq test range passed tests unittests test testsource unit py testaudiosrcfreq test invalid passed tests unittests test testsource unit py testaudiosrcfreq test normal home travis build mithro gst switch python api gstswitch testsource py warning value of type gdouble is invalid or out of range for property freq of type gdouble element set property freq int freq passed tests unittests test testsource unit py testaudiosrcwave test range passed tests unittests test testsource unit py testaudiosrcwave test invalid passed tests unittests test testsource unit py testaudiosrcwave test normal passed tests unittests test testsource unit py testaudiosrcplay test run passed tests unittests test testsource unit py testaudiosrcplay test pause passed tests unittests test testsource unit py testaudiosrcplay test end passed coverage platform python final name stmts miss cover gstswitch init gstswitch connection gstswitch controller gstswitch exception gstswitch helpers gstswitch server gstswitch testsource total passed warnings in seconds the command trusty run py pwd travis run sh exited with | 0 |
899 | 2,997,348,764 | IssuesEvent | 2015-07-23 06:50:10 | excelsior-oss/restler | https://api.github.com/repos/excelsior-oss/restler | closed | Release automation | Infrastructure starter | We need maven task to release new versions of libraray, which should do at least:
1) Increment version from current snapshot to current release according to semantic versioning
2) Build artifact
Additionally this task may:
1) Deploy artifact to github
2) Deploy artifact mavencentral and/or bintray
3) Provide tasks to increment, major, minor and patch versions according to semantic versioning | 1.0 | Release automation - We need maven task to release new versions of libraray, which should do at least:
1) Increment version from current snapshot to current release according to semantic versioning
2) Build artifact
Additionally this task may:
1) Deploy artifact to github
2) Deploy artifact mavencentral and/or bintray
3) Provide tasks to increment, major, minor and patch versions according to semantic versioning | infrastructure | release automation we need maven task to release new versions of libraray which should do at least increment version from current snapshot to current release according to semantic versioning build artifact additionally this task may deploy artifact to github deploy artifact mavencentral and or bintray provide tasks to increment major minor and patch versions according to semantic versioning | 1 |
16,632 | 12,069,997,999 | IssuesEvent | 2020-04-16 16:54:59 | bootstrapworld/curriculum | https://api.github.com/repos/bootstrapworld/curriculum | closed | Programmatically generated links to exercises/solutions | Infrastructure | - [x] Exercises table for language X includes all lesson links, even if they're behind a conditional for language Y
- [x] Building the exercise table for data-science fails with a racket contract violation
- [x] Sort exercises in some reasonable way (alphabetically? order in which they appear in the lesson?)
- [x] Sort lessons in the table based on position in `lesson-order.txt`
- [x] Make lesson titles links back to the original lessons
- [x] Fallback for exercise titles should use the link text
- [x] Group exercises under lesson headings
- [x] BUG: rows seem to be repeated, once for each lesson in the pathway
- [x] Can the `@exercise-link` directive handle external links, to exercises like a desmos or quizziz activity - answer from Dorai (NO)
- [x] The collated exercise table should have _rows_ for every exercise - not one row with _cells_ for each exercise.
- [x] I'm getting tons of `@workbook-link used for non-workbook page` warnings, even though I'm not using that directive at all. Testing suggests that the `@exercise-link` directive might be sharing some of the code-path of `workbook-link`
- [x] Add directive to generate this HTML, and use it in each pathway's teacher resources page
- [x] Do we need an `@exercise-link` after all to keep track of these? (after discussing with Dorai, the answer is yes)
- [x] All exercises should be collected across lessons, and listed (grouped under lesson heading) in the teacher resources page. | 1.0 | Programmatically generated links to exercises/solutions - - [x] Exercises table for language X includes all lesson links, even if they're behind a conditional for language Y
- [x] Building the exercise table for data-science fails with a racket contract violation
- [x] Sort exercises in some reasonable way (alphabetically? order in which they appear in the lesson?)
- [x] Sort lessons in the table based on position in `lesson-order.txt`
- [x] Make lesson titles links back to the original lessons
- [x] Fallback for exercise titles should use the link text
- [x] Group exercises under lesson headings
- [x] BUG: rows seem to be repeated, once for each lesson in the pathway
- [x] Can the `@exercise-link` directive handle external links, to exercises like a desmos or quizziz activity - answer from Dorai (NO)
- [x] The collated exercise table should have _rows_ for every exercise - not one row with _cells_ for each exercise.
- [x] I'm getting tons of `@workbook-link used for non-workbook page` warnings, even though I'm not using that directive at all. Testing suggests that the `@exercise-link` directive might be sharing some of the code-path of `workbook-link`
- [x] Add directive to generate this HTML, and use it in each pathway's teacher resources page
- [x] Do we need an `@exercise-link` after all to keep track of these? (after discussing with Dorai, the answer is yes)
- [x] All exercises should be collected across lessons, and listed (grouped under lesson heading) in the teacher resources page. | infrastructure | programmatically generated links to exercises solutions exercises table for language x includes all lesson links even if they re behind a conditional for language y building the exercise table for data science fails with a racket contract violation sort exercises in some reasonable way alphabetically order in which they appear in the lesson sort lessons in the table based on position in lesson order txt make lesson titles links back to the original lessons fallback for exercise titles should use the link text group exercises under lesson headings bug rows seem to be repeated once for each lesson in the pathway can the exercise link directive handle external links to exercises like a desmos or quizziz activity answer from dorai no the collated exercise table should have rows for every exercise not one row with cells for each exercise i m getting tons of workbook link used for non workbook page warnings even though i m not using that directive at all testing suggests that the exercise link directive might be sharing some of the code path of workbook link add directive to generate this html and use it in each pathway s teacher resources page do we need an exercise link after all to keep track of these after discussing with dorai the answer is yes all exercises should be collected across lessons and listed grouped under lesson heading in the teacher resources page | 1 |
417,448 | 28,110,501,177 | IssuesEvent | 2023-03-31 06:44:14 | JamesLiuZX/ped | https://api.github.com/repos/JamesLiuZX/ped | opened | User guide ambiguity (deleting remarks) | type.DocumentationBug severity.Medium | Instruction given to delete the remarks: (last line)

Executing the instruction according to the UG:

<!--session: 1680243365580-cc59f93e-3a86-4a5b-8196-5f6b0c24adcc-->
<!--Version: Web v3.4.7--> | 1.0 | User guide ambiguity (deleting remarks) - Instruction given to delete the remarks: (last line)

Executing the instruction according to the UG:

<!--session: 1680243365580-cc59f93e-3a86-4a5b-8196-5f6b0c24adcc-->
<!--Version: Web v3.4.7--> | non_infrastructure | user guide ambiguity deleting remarks instruction given to delete the remarks last line executing the instruction according to the ug | 0 |
144,129 | 19,273,828,752 | IssuesEvent | 2021-12-10 09:29:51 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 93: xorg-server-1.20.8: 4 advisories [7.8] | 1.severity: security | [search](https://search.nix.gsc.io/?q=xorg-server&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=xorg-server+in%3Apath&type=Code)
* [ ] [CVE-2020-14346](https://nvd.nist.gov/vuln/detail/CVE-2020-14346) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14361](https://nvd.nist.gov/vuln/detail/CVE-2020-14361) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14362](https://nvd.nist.gov/vuln/detail/CVE-2020-14362) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14347](https://nvd.nist.gov/vuln/detail/CVE-2020-14347) CVSSv3=5.5 (nixos-20.03, nixos-20.09, nixos-unstable)
Scanned versions: nixos-20.03: 0d0660fde3b; nixos-20.09: ad3a5d5092e; nixos-unstable: 84d74ae9c9c.
| True | Vulnerability roundup 93: xorg-server-1.20.8: 4 advisories [7.8] - [search](https://search.nix.gsc.io/?q=xorg-server&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=xorg-server+in%3Apath&type=Code)
* [ ] [CVE-2020-14346](https://nvd.nist.gov/vuln/detail/CVE-2020-14346) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14361](https://nvd.nist.gov/vuln/detail/CVE-2020-14361) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14362](https://nvd.nist.gov/vuln/detail/CVE-2020-14362) CVSSv3=7.8 (nixos-20.03, nixos-20.09, nixos-unstable)
* [ ] [CVE-2020-14347](https://nvd.nist.gov/vuln/detail/CVE-2020-14347) CVSSv3=5.5 (nixos-20.03, nixos-20.09, nixos-unstable)
Scanned versions: nixos-20.03: 0d0660fde3b; nixos-20.09: ad3a5d5092e; nixos-unstable: 84d74ae9c9c.
| non_infrastructure | vulnerability roundup xorg server advisories nixos nixos nixos unstable nixos nixos nixos unstable nixos nixos nixos unstable nixos nixos nixos unstable scanned versions nixos nixos nixos unstable | 0 |
4,845 | 5,293,262,009 | IssuesEvent | 2017-02-09 06:33:04 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | gene_ontology_write includes incorrect SubPropertyOf axioms causing incoherency | editors-discussion Infrastructure | * immediate-transformation-of SubPropertyOf immediately_preceded_by
* transformation-of SubPropertyOf preceded_by
This introduces an incoherency, since the domain/range of the former is continuant and occurrent for the latter.
This ends up breaking everyone's GAF inferences, as mentioned here:
https://github.com/geneontology/go-annotation/issues/1487
The source of this issue is an old version of RO, @balhoff pointed this out here: https://github.com/oborel/obo-relations/issues/126 -- fixed in RO, but there is a zombie relation kicking around.
We need to
1. [x] Remove the incorrect axioms
2. [ ] Avoid zombie relations
3. [x] Better detect incoherencies upstream
I will do 1, trivial.
For 2, I suspect this came from an old version of RO that was itself MIREOTed into an ontology that GO imports. Due to our hack to support OE, we end up injecting these into the editors file until they are manually cleared. Long term fix is to switch to Protege.
For 3, the problem here is that technically the ontology is not incoherent until these relations are actually *used*. This means we end up not catching them as part of the normal ontology development cycle, and they show up in everyone's GAF inferences, which is confusing for everyone.
What we need is a check of "unsatisfiable Object Properties". This is noted in this robot ticket: https://github.com/ontodev/robot/issues/104 | 1.0 | gene_ontology_write includes incorrect SubPropertyOf axioms causing incoherency - * immediate-transformation-of SubPropertyOf immediately_preceded_by
* transformation-of SubPropertyOf preceded_by
This introduces an incoherency, since the domain/range of the former is continuant and occurrent for the latter.
This ends up breaking everyone's GAF inferences, as mentioned here:
https://github.com/geneontology/go-annotation/issues/1487
The source of this issue is an old version of RO, @balhoff pointed this out here: https://github.com/oborel/obo-relations/issues/126 -- fixed in RO, but there is a zombie relation kicking around.
We need to
1. [x] Remove the incorrect axioms
2. [ ] Avoid zombie relations
3. [x] Better detect incoherencies upstream
I will do 1, trivial.
For 2, I suspect this came from an old version of RO that was itself MIREOTed into an ontology that GO imports. Due to our hack to support OE, we end up injecting these into the editors file until they are manually cleared. Long term fix is to switch to Protege.
For 3, the problem here is that technically the ontology is not incoherent until these relations are actually *used*. This means we end up not catching them as part of the normal ontology development cycle, and they show up in everyone's GAF inferences, which is confusing for everyone.
What we need is a check of "unsatisfiable Object Properties". This is noted in this robot ticket: https://github.com/ontodev/robot/issues/104 | infrastructure | gene ontology write includes incorrect subpropertyof axioms causing incoherency immediate transformation of subpropertyof immediately preceded by transformation of subpropertyof preceded by this introduces an incoherency since the domain range of the former is continuant and occurrent for the latter this ends up breaking everyone s gaf inferences as mentioned here the source of this issue is an old version of ro balhoff pointed this out here fixed in ro but there is a zombie relation kicking around we need to remove the incorrect axioms avoid zombie relations better detect incoherencies upstream i will do trivial for i suspect this came from an old version of ro that was itself mireoted into an ontology that go imports due to our hack to support oe we end up injecting these into the editors file until they are manually cleared long term fix is to switch to protege for the problem here is that technically the ontology is not incoherent until these relations are actually used this means we end up not catching them as part of the normal ontology development cycle and they show up in everyone s gaf inferences which is confusing for everyone what we need is a check of unsatisfiable object properties this is noted in this robot ticket | 1 |
718,698 | 24,728,940,714 | IssuesEvent | 2022-10-20 15:57:12 | kubernetes-sigs/cluster-api-provider-aws | https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-aws | closed | Upgrade to controller-tools v0.9.2 | priority/important-soon lifecycle/stale triage/accepted | Upgrade to controller-tools version that includes @chrischdi's PR that allows setting listMapKey, listType at both type and field level. Needed for using Server Side Apply with AWSCluster SubnetSpec.
* https://github.com/kubernetes-sigs/controller-tools/pull/692
Note: feel free to retitle with a correct controller-tools version when it is released.
| 1.0 | Upgrade to controller-tools v0.9.2 - Upgrade to controller-tools version that includes @chrischdi's PR that allows setting listMapKey, listType at both type and field level. Needed for using Server Side Apply with AWSCluster SubnetSpec.
* https://github.com/kubernetes-sigs/controller-tools/pull/692
Note: feel free to retitle with a correct controller-tools version when it is released.
| non_infrastructure | upgrade to controller tools upgrade to controller tools version that includes chrischdi s pr that allows setting listmapkey listtype at both type and field level needed for using server side apply with awscluster subnetspec note feel free to retitle with a correct controller tools version when it is released | 0 |
35,199 | 30,832,741,200 | IssuesEvent | 2023-08-02 04:02:43 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Rename Microsoft.AspNetCore.Testing | area-infrastructure | As per https://github.com/dotnet/extensions/issues/4057#issuecomment-1660927215, https://github.com/dotnet/aspnetcore/blob/main/src/Testing/src/Microsoft.AspNetCore.Testing.csproj is used for internal purposes, and being published to the BAR it causes a clash with the project coming out of dotnet/extensions (e.g., https://github.com/dotnet/dnceng/issues/174).
The project should be renamed or should not be published to the BAR.
/cc: @joperezr @Tratcher @wtgodbe | 1.0 | Rename Microsoft.AspNetCore.Testing - As per https://github.com/dotnet/extensions/issues/4057#issuecomment-1660927215, https://github.com/dotnet/aspnetcore/blob/main/src/Testing/src/Microsoft.AspNetCore.Testing.csproj is used for internal purposes, and being published to the BAR it causes a clash with the project coming out of dotnet/extensions (e.g., https://github.com/dotnet/dnceng/issues/174).
The project should be renamed or should not be published to the BAR.
/cc: @joperezr @Tratcher @wtgodbe | infrastructure | rename microsoft aspnetcore testing as per is used for internal purposes and being published to the bar it causes a clash with the project coming out of dotnet extensions e g the project should be renamed or should not be published to the bar cc joperezr tratcher wtgodbe | 1 |
7,146 | 6,794,401,281 | IssuesEvent | 2017-11-01 12:00:11 | Daniel-Mietchen/ideas | https://api.github.com/repos/Daniel-Mietchen/ideas | opened | Get MeSH into Mix'n Match | infrastructure literature Wikidata | as per
https://www.nlm.nih.gov/mesh/
and
https://tools.wmflabs.org/mix-n-match/
Some are already in:
- https://tools.wmflabs.org/mix-n-match/#/catalog/474
- https://tools.wmflabs.org/mix-n-match/#/catalog/519
- https://tools.wmflabs.org/mix-n-match/#/catalog/633
Some still need to be imported
- https://tools.wmflabs.org/mix-n-match/import.php
| 1.0 | Get MeSH into Mix'n Match - as per
https://www.nlm.nih.gov/mesh/
and
https://tools.wmflabs.org/mix-n-match/
Some are already in:
- https://tools.wmflabs.org/mix-n-match/#/catalog/474
- https://tools.wmflabs.org/mix-n-match/#/catalog/519
- https://tools.wmflabs.org/mix-n-match/#/catalog/633
Some still need to be imported
- https://tools.wmflabs.org/mix-n-match/import.php
| infrastructure | get mesh into mix n match as per and some are already in some still need to be imported | 1 |
26,639 | 20,373,079,829 | IssuesEvent | 2022-02-21 13:10:32 | lampepfl/dotty | https://api.github.com/repos/lampepfl/dotty | closed | sbt scala3-compiler-bootstrapped/run always recompiles ~100 scala files | stat:cannot reproduce area:infrastructure itype:performance | When running `sbt run` twice in the row from a fresh clone, the second run is much faster (working as intended!). But something seems to be off with the `scala3-compiler-bootstrapped` project, because it systematically recompiles ~100 scala files:
```scala
$ git clean -fdx
[...]
$ sbt scala3-compiler-bootstrapped/run
[...]
[info] compiling 1 Scala source to /dotty/project/project/target/scala-2.12/sbt-1.0/classes ...
[info] compiling 10 Scala sources and 1 Java source to /dotty/project/target/scala-2.12/sbt-1.0/classes ...
[info] compiling 7 Java sources to /dotty/interfaces/target/classes ...
[info] compiling 74 Scala sources and 25 Java sources to /dotty/library/target/scala-3.1.1-RC2/classes ...
[info] compiling 7 Scala sources to /dotty/tasty/target/scala-3.1.1-RC2/classes ...
[info] compiling 504 Scala sources and 6 Java sources to /dotty/compiler/target/scala-3.1.1-RC2/classes ...
[info] compiling 14 Java sources to /dotty/sbt-bridge/src/target/classes ...
[info] compiling 74 Scala sources and 25 Java sources to /dotty/out/bootstrap/scala3-library-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 7 Scala sources to /dotty/out/bootstrap/tasty-core-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 504 Scala sources and 6 Java sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 3 Scala sources to /dotty/out/bootstrap/scala3-tasty-inspector/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 5 Scala sources to /dotty/out/bootstrap/scala3-staging/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] running (fork) dotty.tools.dotc.Main -classpath /home/olivier/.cache/coursier/v1/https/repo1.maven.org/maven2/org/scala-lang/scala-library/2.13.6/scala-library-2.13.6.jar:/dotty/library/../out/bootstrap/scala3-library-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/scala3-library_3-3.1.2-RC1-bin-SNAPSHOT.jar
Usage: scalac <options> <source files>
$ sbt scala3-compiler-bootstrapped/run
[...]
[info] compiling 10 Scala sources to /dotty/compiler/target/scala-3.1.1-RC2/classes ...
[info] compiling 12 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bi
n-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 67 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bi
n-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 6 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bin
-SNAPSHOT-nonbootstrapped/classes ...
[info] running (fork) dotty.tools.dotc.Main -classpath /home/olivier/.cache/coursier/v1/https/repo1.maven.org/maven2/org/scala-lang/sca
la-library/2.13.6/scala-library-2.13.6.jar:/dotty/library/../out/bootstrap/scala3-library-bootstrapped/sc
ala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/scala3-library_3-3.1.2-RC1-bin-SNAPSHOT.jar
Usage: scalac <options> <source files>
```
| 1.0 | sbt scala3-compiler-bootstrapped/run always recompiles ~100 scala files - When running `sbt run` twice in the row from a fresh clone, the second run is much faster (working as intended!). But something seems to be off with the `scala3-compiler-bootstrapped` project, because it systematically recompiles ~100 scala files:
```scala
$ git clean -fdx
[...]
$ sbt scala3-compiler-bootstrapped/run
[...]
[info] compiling 1 Scala source to /dotty/project/project/target/scala-2.12/sbt-1.0/classes ...
[info] compiling 10 Scala sources and 1 Java source to /dotty/project/target/scala-2.12/sbt-1.0/classes ...
[info] compiling 7 Java sources to /dotty/interfaces/target/classes ...
[info] compiling 74 Scala sources and 25 Java sources to /dotty/library/target/scala-3.1.1-RC2/classes ...
[info] compiling 7 Scala sources to /dotty/tasty/target/scala-3.1.1-RC2/classes ...
[info] compiling 504 Scala sources and 6 Java sources to /dotty/compiler/target/scala-3.1.1-RC2/classes ...
[info] compiling 14 Java sources to /dotty/sbt-bridge/src/target/classes ...
[info] compiling 74 Scala sources and 25 Java sources to /dotty/out/bootstrap/scala3-library-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 7 Scala sources to /dotty/out/bootstrap/tasty-core-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 504 Scala sources and 6 Java sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 3 Scala sources to /dotty/out/bootstrap/scala3-tasty-inspector/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 5 Scala sources to /dotty/out/bootstrap/scala3-staging/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/classes ...
[info] running (fork) dotty.tools.dotc.Main -classpath /home/olivier/.cache/coursier/v1/https/repo1.maven.org/maven2/org/scala-lang/scala-library/2.13.6/scala-library-2.13.6.jar:/dotty/library/../out/bootstrap/scala3-library-bootstrapped/scala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/scala3-library_3-3.1.2-RC1-bin-SNAPSHOT.jar
Usage: scalac <options> <source files>
$ sbt scala3-compiler-bootstrapped/run
[...]
[info] compiling 10 Scala sources to /dotty/compiler/target/scala-3.1.1-RC2/classes ...
[info] compiling 12 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bi
n-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 67 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bi
n-SNAPSHOT-nonbootstrapped/classes ...
[info] compiling 6 Scala sources to /dotty/out/bootstrap/scala3-compiler-bootstrapped/scala-3.1.2-RC1-bin
-SNAPSHOT-nonbootstrapped/classes ...
[info] running (fork) dotty.tools.dotc.Main -classpath /home/olivier/.cache/coursier/v1/https/repo1.maven.org/maven2/org/scala-lang/sca
la-library/2.13.6/scala-library-2.13.6.jar:/dotty/library/../out/bootstrap/scala3-library-bootstrapped/sc
ala-3.1.2-RC1-bin-SNAPSHOT-nonbootstrapped/scala3-library_3-3.1.2-RC1-bin-SNAPSHOT.jar
Usage: scalac <options> <source files>
```
| infrastructure | sbt compiler bootstrapped run always recompiles scala files when running sbt run twice in the row from a fresh clone the second run is much faster working as intended but something seems to be off with the compiler bootstrapped project because it systematically recompiles scala files scala git clean fdx sbt compiler bootstrapped run compiling scala source to dotty project project target scala sbt classes compiling scala sources and java source to dotty project target scala sbt classes compiling java sources to dotty interfaces target classes compiling scala sources and java sources to dotty library target scala classes compiling scala sources to dotty tasty target scala classes compiling scala sources and java sources to dotty compiler target scala classes compiling java sources to dotty sbt bridge src target classes compiling scala sources and java sources to dotty out bootstrap library bootstrapped scala bin snapshot nonbootstrapped classes compiling scala sources to dotty out bootstrap tasty core bootstrapped scala bin snapshot nonbootstrapped classes compiling scala sources and java sources to dotty out bootstrap compiler bootstrapped scala bin snapshot nonbootstrapped classes compiling scala sources to dotty out bootstrap tasty inspector scala bin snapshot nonbootstrapped classes compiling scala sources to dotty out bootstrap staging scala bin snapshot nonbootstrapped classes running fork dotty tools dotc main classpath home olivier cache coursier https maven org org scala lang scala library scala library jar dotty library out bootstrap library bootstrapped scala bin snapshot nonbootstrapped library bin snapshot jar usage scalac sbt compiler bootstrapped run compiling scala sources to dotty compiler target scala classes compiling scala sources to dotty out bootstrap compiler bootstrapped scala bi n snapshot nonbootstrapped classes compiling scala sources to dotty out bootstrap compiler bootstrapped scala bi n snapshot nonbootstrapped classes compiling scala sources to dotty out bootstrap compiler bootstrapped scala bin snapshot nonbootstrapped classes running fork dotty tools dotc main classpath home olivier cache coursier https maven org org scala lang sca la library scala library jar dotty library out bootstrap library bootstrapped sc ala bin snapshot nonbootstrapped library bin snapshot jar usage scalac | 1 |
440,966 | 30,764,268,485 | IssuesEvent | 2023-07-30 04:50:34 | ZanattaMichael/SRDSC | https://api.github.com/repos/ZanattaMichael/SRDSC | closed | Update Documentation | documentation | Add `Install-Module PowerShell-YAML, xPSDesiredStateConfiguration` to the quick start guide. | 1.0 | Update Documentation - Add `Install-Module PowerShell-YAML, xPSDesiredStateConfiguration` to the quick start guide. | non_infrastructure | update documentation add install module powershell yaml xpsdesiredstateconfiguration to the quick start guide | 0 |
601,927 | 18,442,780,955 | IssuesEvent | 2021-10-14 20:18:21 | aegershman/yugabyte-boshrelease | https://api.github.com/repos/aegershman/yugabyte-boshrelease | closed | sample apps should be able to override hosts property | priority/just-a-thought status/probably-vaporware job/yb-sample-apps | in order to re-use the sample apps as smoke tests or as standalone jobs, they shouldn't be required to consume yb-tserver links directly and should be able to accept manually configured host strings | 1.0 | sample apps should be able to override hosts property - in order to re-use the sample apps as smoke tests or as standalone jobs, they shouldn't be required to consume yb-tserver links directly and should be able to accept manually configured host strings | non_infrastructure | sample apps should be able to override hosts property in order to re use the sample apps as smoke tests or as standalone jobs they shouldn t be required to consume yb tserver links directly and should be able to accept manually configured host strings | 0 |
11,952 | 9,533,620,074 | IssuesEvent | 2019-04-29 21:51:21 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Compile error in src/Microsoft.VisualBasic.Core/tests/VB with `dotnet test /p:targetgroup=netfx` | area-Infrastructure | Compile error in src/Microsoft.VisualBasic.Core/tests/VB with `dotnet test /p:targetgroup=netfx`:
```
>dotnet test /p:targetgroup=netfx
FileIOTestBase.vb(12,18): error BC30002: Type 'IO.FileCleanupTestBase' is not defined.
```
See https://github.com/dotnet/corefx/pull/37010#issuecomment-484955991. | 1.0 | Compile error in src/Microsoft.VisualBasic.Core/tests/VB with `dotnet test /p:targetgroup=netfx` - Compile error in src/Microsoft.VisualBasic.Core/tests/VB with `dotnet test /p:targetgroup=netfx`:
```
>dotnet test /p:targetgroup=netfx
FileIOTestBase.vb(12,18): error BC30002: Type 'IO.FileCleanupTestBase' is not defined.
```
See https://github.com/dotnet/corefx/pull/37010#issuecomment-484955991. | infrastructure | compile error in src microsoft visualbasic core tests vb with dotnet test p targetgroup netfx compile error in src microsoft visualbasic core tests vb with dotnet test p targetgroup netfx dotnet test p targetgroup netfx fileiotestbase vb error type io filecleanuptestbase is not defined see | 1 |
328,849 | 10,000,744,079 | IssuesEvent | 2019-07-12 14:05:35 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.ebay.com.au - design is broken | browser-firefox-tablet engine-gecko priority-important | <!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.ebay.com.au/b/Solar-Outdoor-Floodlights/183393/bn_10594105
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Content wider than viewable area
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/7/79dfb88d-9b79-4afe-a487-07d70a527dcf.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190626134844</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.ebay.com.au - design is broken - <!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.ebay.com.au/b/Solar-Outdoor-Floodlights/183393/bn_10594105
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Content wider than viewable area
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/7/79dfb88d-9b79-4afe-a487-07d70a527dcf.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190626134844</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | design is broken url browser version firefox mobile tablet operating system android tested another browser yes problem type design is broken description content wider than viewable area steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️ | 0 |
32,785 | 26,993,883,550 | IssuesEvent | 2023-02-09 22:27:53 | crossplane/docs | https://api.github.com/repos/crossplane/docs | closed | Can't distinguish commands from output | enhancement P3 infrastructure | ### What's Wrong?
<!-- What problem did you find? -->
In some (possibly all?) command snippets It's difficult to quickly visually distinguish which lines are commands and which are output. For example at https://docs.crossplane.io/v1.11/getting-started/provider-aws/ I see:
<img width="919" alt="Screenshot 2023-02-01 at 12 29 44 PM" src="https://user-images.githubusercontent.com/1049349/216156581-08e7626e-3083-460b-a596-a2af78e8d681.png">
I do note that when I click the "copy" button it only copies the command input, which is useful but was not the behaviour I expected. My intuition was that the copy button would copy the entire content of the code block.
I believe it's fairly common to see the command prefaced with a `$` to give a visual distinction. Is this something we should do? | 1.0 | Can't distinguish commands from output - ### What's Wrong?
<!-- What problem did you find? -->
In some (possibly all?) command snippets It's difficult to quickly visually distinguish which lines are commands and which are output. For example at https://docs.crossplane.io/v1.11/getting-started/provider-aws/ I see:
<img width="919" alt="Screenshot 2023-02-01 at 12 29 44 PM" src="https://user-images.githubusercontent.com/1049349/216156581-08e7626e-3083-460b-a596-a2af78e8d681.png">
I do note that when I click the "copy" button it only copies the command input, which is useful but was not the behaviour I expected. My intuition was that the copy button would copy the entire content of the code block.
I believe it's fairly common to see the command prefaced with a `$` to give a visual distinction. Is this something we should do? | infrastructure | can t distinguish commands from output what s wrong in some possibly all command snippets it s difficult to quickly visually distinguish which lines are commands and which are output for example at i see img width alt screenshot at pm src i do note that when i click the copy button it only copies the command input which is useful but was not the behaviour i expected my intuition was that the copy button would copy the entire content of the code block i believe it s fairly common to see the command prefaced with a to give a visual distinction is this something we should do | 1 |
8,704 | 7,572,784,833 | IssuesEvent | 2018-04-23 15:49:49 | cityofaustin/techstack | https://api.github.com/repos/cityofaustin/techstack | closed | Alpha Production and Staging Environments | Resource: Technical Infrastructure Size: L Team: Dev | Currently we automatically deploy to alpha.austin.gov when code is merged into cityofaustin/janis master branch. We should be more deliberate before updating alpha.
- [x] deploy joplin (the production heroku app)
- [x] add new janis project (`austindotgov` in heroku), hook up to GitHub, deploy
- add/change DNS entries
- [x] alpha.austin.gov => alpha.austin.gov.herokudns.com
- [x] staging.austin.gov => jstaging.austin.gov.herokudns.com
There are a few weird things going on here that would be nice to address:
1. **Heroku Build**
1. **Don't rebuild at each stage** Normally a Docker workflow would have us build once and then move the built image to additional stages. This ensures that what we're seeing in staging is what we'll get in production. Heroku doesn't support this and requires a rebuild for each stage.
2. **Build-time variables per stage** It would be nice to set different CMS URLs for review apps, for times when we want to test changing the CMS and the frontend at the same time. This would allow us to use a CMS review app URL to build the frontend without having to change code. Heroku doesn't support this.
2. **DNS** Currently we have to ping another team to change DNS entries. They have to update a zone file and redeploy. That group often has a lot going on, so would be nice to relieve them of this burden. Can we own the DNS records?
3. **Images** Eventually we'll want to put static assets on a CDN. In the meantime, we're serving images from joplin. This stinks because we have to keep joplin up and running, even though we're serving static content from janis. If we distribute the images with janis, we can remove the need to keep joplin running & scaled. | 1.0 | Alpha Production and Staging Environments - Currently we automatically deploy to alpha.austin.gov when code is merged into cityofaustin/janis master branch. We should be more deliberate before updating alpha.
- [x] deploy joplin (the production heroku app)
- [x] add new janis project (`austindotgov` in heroku), hook up to GitHub, deploy
- add/change DNS entries
- [x] alpha.austin.gov => alpha.austin.gov.herokudns.com
- [x] staging.austin.gov => jstaging.austin.gov.herokudns.com
There are a few weird things going on here that would be nice to address:
1. **Heroku Build**
1. **Don't rebuild at each stage** Normally a Docker workflow would have us build once and then move the built image to additional stages. This ensures that what we're seeing in staging is what we'll get in production. Heroku doesn't support this and requires a rebuild for each stage.
2. **Build-time variables per stage** It would be nice to set different CMS URLs for review apps, for times when we want to test changing the CMS and the frontend at the same time. This would allow us to use a CMS review app URL to build the frontend without having to change code. Heroku doesn't support this.
2. **DNS** Currently we have to ping another team to change DNS entries. They have to update a zone file and redeploy. That group often has a lot going on, so would be nice to relieve them of this burden. Can we own the DNS records?
3. **Images** Eventually we'll want to put static assets on a CDN. In the meantime, we're serving images from joplin. This stinks because we have to keep joplin up and running, even though we're serving static content from janis. If we distribute the images with janis, we can remove the need to keep joplin running & scaled. | infrastructure | alpha production and staging environments currently we automatically deploy to alpha austin gov when code is merged into cityofaustin janis master branch we should be more deliberate before updating alpha deploy joplin the production heroku app add new janis project austindotgov in heroku hook up to github deploy add change dns entries alpha austin gov alpha austin gov herokudns com staging austin gov jstaging austin gov herokudns com there are a few weird things going on here that would be nice to address heroku build don t rebuild at each stage normally a docker workflow would have us build once and then move the built image to additional stages this ensures that what we re seeing in staging is what we ll get in production heroku doesn t support this and requires a rebuild for each stage build time variables per stage it would be nice to set different cms urls for review apps for times when we want to test changing the cms and the frontend at the same time this would allow us to use a cms review app url to build the frontend without having to change code heroku doesn t support this dns currently we have to ping another team to change dns entries they have to update a zone file and redeploy that group often has a lot going on so would be nice to relieve them of this burden can we own the dns records images eventually we ll want to put static assets on a cdn in the meantime we re serving images from joplin this stinks because we have to keep joplin up and running even though we re serving static content from janis if we distribute the images with janis we can remove the need to keep joplin running scaled | 1 |
1,725 | 3,354,063,908 | IssuesEvent | 2015-11-18 10:12:13 | nilearn/nilearn | https://api.github.com/repos/nilearn/nilearn | closed | Plotting displays need refactoring | Enhancement Infrastructure New feature Quality | Some methods of `BaseSlicer`, (`add_contours` for example) works perfectly in `GlassBrainAxes` because it relies on `tranform_to_2d`. Somebody should take a deeper look into it and find a new inheritance tree for these classes. | 1.0 | Plotting displays need refactoring - Some methods of `BaseSlicer`, (`add_contours` for example) works perfectly in `GlassBrainAxes` because it relies on `tranform_to_2d`. Somebody should take a deeper look into it and find a new inheritance tree for these classes. | infrastructure | plotting displays need refactoring some methods of baseslicer add contours for example works perfectly in glassbrainaxes because it relies on tranform to somebody should take a deeper look into it and find a new inheritance tree for these classes | 1 |
278,420 | 24,152,626,790 | IssuesEvent | 2022-09-22 03:28:09 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | pkg/ccl/testccl/workload/schemachange/schemachange_test: TestWorkload failed | C-test-failure O-robot branch-master skipped-test GA-blocker branch-release-22.2 | pkg/ccl/testccl/workload/schemachange/schemachange_test.TestWorkload [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4672755&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4672755&tab=artifacts#/) on master @ [cb10bfe399528718776711068b2560e5f6d53b55](https://github.com/cockroachdb/cockroach/commits/cb10bfe399528718776711068b2560e5f6d53b55):
```
1, 29, lease, 11
1, 29, locations, 21
1, 29, migrations, 40
1, 29, namespace, 30
1, 29, protected_ts_meta, 31
1, 29, protected_ts_records, 32
1, 29, rangelog, 13
1, 29, replication_constraint_stats, 25
1, 29, replication_critical_localities, 26
1, 29, replication_stats, 27
1, 29, reports_meta, 28
1, 29, role_members, 23
1, 29, role_options, 33
1, 29, scheduled_jobs, 37
1, 29, settings, 6
1, 29, span_configurations, 47
1, 29, sql_instances, 46
1, 29, sqlliveness, 39
1, 29, statement_bundle_chunks, 34
1, 29, statement_diagnostics, 36
1, 29, statement_diagnostics_requests, 35
1, 29, statement_statistics, 42
1, 29, table_statistics, 20
1, 29, tenant_settings, 50
1, 29, tenant_usage, 45
1, 29, tenants, 8
1, 29, transaction_statistics, 43
1, 29, ui, 14
1, 29, users, 4
1, 29, web_sessions, 19
1, 29, zones, 5
100, 0, public, 101
102, 0, public, 103
104, 0, public, 105
104, 0, schema271, 168
104, 105, _crdb_internal_region, 111
104, 105, _enum189, 150
104, 105, _enum52, 115
104, 105, _enum59, 119
104, 105, crdb_internal_region, 110
104, 105, enum189, 149
104, 105, enum52, 114
104, 105, enum59, 118
104, 105, table191, 152
104, 105, table224, 157
104, 105, table238, 161
104, 105, table243, 162
104, 105, view277, 172
schema_change_external_test.go:76: backup in /artifacts/tmp/_tmp/e432d12c42e9f286daed215cc6401428/TestWorkload105635140/001
--- FAIL: TestWorkload (11.91s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #78400 pkg/ccl/testccl/workload/schemachange/schemachange_test: TestWorkload failed [C-test-failure O-robot branch-release-22.1]
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestWorkload.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-14146 | 2.0 | pkg/ccl/testccl/workload/schemachange/schemachange_test: TestWorkload failed - pkg/ccl/testccl/workload/schemachange/schemachange_test.TestWorkload [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4672755&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4672755&tab=artifacts#/) on master @ [cb10bfe399528718776711068b2560e5f6d53b55](https://github.com/cockroachdb/cockroach/commits/cb10bfe399528718776711068b2560e5f6d53b55):
```
1, 29, lease, 11
1, 29, locations, 21
1, 29, migrations, 40
1, 29, namespace, 30
1, 29, protected_ts_meta, 31
1, 29, protected_ts_records, 32
1, 29, rangelog, 13
1, 29, replication_constraint_stats, 25
1, 29, replication_critical_localities, 26
1, 29, replication_stats, 27
1, 29, reports_meta, 28
1, 29, role_members, 23
1, 29, role_options, 33
1, 29, scheduled_jobs, 37
1, 29, settings, 6
1, 29, span_configurations, 47
1, 29, sql_instances, 46
1, 29, sqlliveness, 39
1, 29, statement_bundle_chunks, 34
1, 29, statement_diagnostics, 36
1, 29, statement_diagnostics_requests, 35
1, 29, statement_statistics, 42
1, 29, table_statistics, 20
1, 29, tenant_settings, 50
1, 29, tenant_usage, 45
1, 29, tenants, 8
1, 29, transaction_statistics, 43
1, 29, ui, 14
1, 29, users, 4
1, 29, web_sessions, 19
1, 29, zones, 5
100, 0, public, 101
102, 0, public, 103
104, 0, public, 105
104, 0, schema271, 168
104, 105, _crdb_internal_region, 111
104, 105, _enum189, 150
104, 105, _enum52, 115
104, 105, _enum59, 119
104, 105, crdb_internal_region, 110
104, 105, enum189, 149
104, 105, enum52, 114
104, 105, enum59, 118
104, 105, table191, 152
104, 105, table224, 157
104, 105, table238, 161
104, 105, table243, 162
104, 105, view277, 172
schema_change_external_test.go:76: backup in /artifacts/tmp/_tmp/e432d12c42e9f286daed215cc6401428/TestWorkload105635140/001
--- FAIL: TestWorkload (11.91s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #78400 pkg/ccl/testccl/workload/schemachange/schemachange_test: TestWorkload failed [C-test-failure O-robot branch-release-22.1]
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestWorkload.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-14146 | non_infrastructure | pkg ccl testccl workload schemachange schemachange test testworkload failed pkg ccl testccl workload schemachange schemachange test testworkload with on master lease locations migrations namespace protected ts meta protected ts records rangelog replication constraint stats replication critical localities replication stats reports meta role members role options scheduled jobs settings span configurations sql instances sqlliveness statement bundle chunks statement diagnostics statement diagnostics requests statement statistics table statistics tenant settings tenant usage tenants transaction statistics ui users web sessions zones public public public crdb internal region crdb internal region schema change external test go backup in artifacts tmp tmp fail testworkload help see also parameters in this failure tags bazel gss same failure on other branches pkg ccl testccl workload schemachange schemachange test testworkload failed jira issue crdb | 0 |
28,709 | 23,456,992,089 | IssuesEvent | 2022-08-16 09:44:03 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | ninja: build stopped: subcommand failed. | question area-Infrastructure-coreclr | Command failed with following error :
build.cmd clr+libs -c debug -a x64
Error :
waiting for C:/Users/mayadav1/runtime/artifacts/obj/win-x64.Debug/corehost/test/typelibs\dlldata.c ...
ninja: build stopped: subcommand failed.
Failed to generate native component build project!
dlldata.c is present above path. | 1.0 | ninja: build stopped: subcommand failed. - Command failed with following error :
build.cmd clr+libs -c debug -a x64
Error :
waiting for C:/Users/mayadav1/runtime/artifacts/obj/win-x64.Debug/corehost/test/typelibs\dlldata.c ...
ninja: build stopped: subcommand failed.
Failed to generate native component build project!
dlldata.c is present above path. | infrastructure | ninja build stopped subcommand failed command failed with following error build cmd clr libs c debug a error waiting for c users runtime artifacts obj win debug corehost test typelibs dlldata c ninja build stopped subcommand failed failed to generate native component build project dlldata c is present above path | 1 |
617,173 | 19,344,395,998 | IssuesEvent | 2021-12-15 09:18:25 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.otodom.pl - site is not usable | browser-firefox-mobile priority-important engine-gecko | <!-- @browser: Firefox Mobile 95.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:95.0) Gecko/95.0 Firefox/95.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/96707 -->
**URL**: https://www.otodom.pl/pl/oferta/nowoczesne-domy-w-cichej-i-spokojnej-okolicy-ID4dUlR
**Browser / Version**: Firefox Mobile 95.0
**Operating System**: Android 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Gdpr/cookies popup cannot be closed as accept button is not working
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/1078586d-dc8b-4a93-a23e-6aae28566035.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211129150630</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/12/8e724365-e955-4d45-b0b3-e52b63839c92)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.otodom.pl - site is not usable - <!-- @browser: Firefox Mobile 95.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:95.0) Gecko/95.0 Firefox/95.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/96707 -->
**URL**: https://www.otodom.pl/pl/oferta/nowoczesne-domy-w-cichej-i-spokojnej-okolicy-ID4dUlR
**Browser / Version**: Firefox Mobile 95.0
**Operating System**: Android 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Gdpr/cookies popup cannot be closed as accept button is not working
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/12/1078586d-dc8b-4a93-a23e-6aae28566035.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211129150630</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/12/8e724365-e955-4d45-b0b3-e52b63839c92)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description buttons or links not working steps to reproduce gdpr cookies popup cannot be closed as accept button is not working view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
26,659 | 20,403,504,251 | IssuesEvent | 2022-02-23 00:41:15 | ZcashFoundation/zebra | https://api.github.com/repos/ZcashFoundation/zebra | closed | Use rust-only link-time optimisation in the zebrad build | E-help-wanted A-infrastructure C-enhancement P-Low :snowflake: | We can improve the performance of our release builds, by activating rust-only link-time optimisation (LTO):
https://doc.rust-lang.org/cargo/reference/profiles.html#lto
This setting activates LTO across all crates:
```
[profile.release]
lto = "fat"
```
If that's too costly, we can just perform LTO within a single crate:
```
[profile.release]
lto = "thin"
```
Here's how we can make that decision:
Compare release builds with these LTO settings:
- no LTO
- thin LTO
- fat LTO
And look at:
- compilation time
- performance
- binary size
Rust-only LTO is different from Rust/C LTO (#805), which is a lot more tricky, because it needs a compatible clang version. | 1.0 | Use rust-only link-time optimisation in the zebrad build - We can improve the performance of our release builds, by activating rust-only link-time optimisation (LTO):
https://doc.rust-lang.org/cargo/reference/profiles.html#lto
This setting activates LTO across all crates:
```
[profile.release]
lto = "fat"
```
If that's too costly, we can just perform LTO within a single crate:
```
[profile.release]
lto = "thin"
```
Here's how we can make that decision:
Compare release builds with these LTO settings:
- no LTO
- thin LTO
- fat LTO
And look at:
- compilation time
- performance
- binary size
Rust-only LTO is different from Rust/C LTO (#805), which is a lot more tricky, because it needs a compatible clang version. | infrastructure | use rust only link time optimisation in the zebrad build we can improve the performance of our release builds by activating rust only link time optimisation lto this setting activates lto across all crates lto fat if that s too costly we can just perform lto within a single crate lto thin here s how we can make that decision compare release builds with these lto settings no lto thin lto fat lto and look at compilation time performance binary size rust only lto is different from rust c lto which is a lot more tricky because it needs a compatible clang version | 1 |
19,321 | 13,214,189,049 | IssuesEvent | 2020-08-16 16:30:48 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Fastly Ansible Module | affects_2.8 bot_closed collection collection:community.general feature module needs_collection_redirect support:community web_infrastructure | ##### SUMMARY
Fastly Ansible Module
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the new feature, module, plugin or task below -->
fastly
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
I would like to do CDN operations on Ansible.
Fastly has code of VCL among CDN vendors so I think that there is a special benefit of coding.
Since there is an API, it can be implemented.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
- fastly:
name: "{{ name }}"
api_key: "{{ api_key }}"
resource: fastly_service_v1
domain:
name: "{{ domain_name }}"
comment: "{{ domain_coment }}"
vcl: "{{ lookup('file', VCL_FILE_PATH) }}"
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
see API
https://docs.fastly.com/api/ | 1.0 | Fastly Ansible Module - ##### SUMMARY
Fastly Ansible Module
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the new feature, module, plugin or task below -->
fastly
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
I would like to do CDN operations on Ansible.
Fastly has code of VCL among CDN vendors so I think that there is a special benefit of coding.
Since there is an API, it can be implemented.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
- fastly:
name: "{{ name }}"
api_key: "{{ api_key }}"
resource: fastly_service_v1
domain:
name: "{{ domain_name }}"
comment: "{{ domain_coment }}"
vcl: "{{ lookup('file', VCL_FILE_PATH) }}"
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
see API
https://docs.fastly.com/api/ | infrastructure | fastly ansible module summary fastly ansible module issue type feature idea component name fastly additional information i would like to do cdn operations on ansible fastly has code of vcl among cdn vendors so i think that there is a special benefit of coding since there is an api it can be implemented yaml fastly name name api key api key resource fastly service domain name domain name comment domain coment vcl lookup file vcl file path see api | 1 |
12,811 | 9,962,198,726 | IssuesEvent | 2019-07-07 12:34:16 | webpack/webpack.js.org | https://api.github.com/repos/webpack/webpack.js.org | closed | 👨🔧[node-version]: Newer node version not supported in next branch | Infrastructure | **Description**
I am using node version `10.12.0` while working on the `next` branch, I am not able to install node_modules.

**How I worked out with this**

**Expected Behaviour**
Should work with latest node LTS versions (10.x) | 1.0 | 👨🔧[node-version]: Newer node version not supported in next branch - **Description**
I am using node version `10.12.0` while working on the `next` branch, I am not able to install node_modules.

**How I worked out with this**

**Expected Behaviour**
Should work with latest node LTS versions (10.x) | infrastructure | 👨🔧 newer node version not supported in next branch description i am using node version while working on the next branch i am not able to install node modules how i worked out with this expected behaviour should work with latest node lts versions x | 1 |
31,472 | 25,782,375,161 | IssuesEvent | 2022-12-09 17:02:50 | Tonomy-Foundation/Tonomy-Contracts | https://api.github.com/repos/Tonomy-Foundation/Tonomy-Contracts | closed | Antelope staging server in Digital ocean | infrastructure | Definition of done
- [x] Current staging server are deleted
- [x] new droplet with 2Gb RAM, 2x CPU and minimum 100 Gb storage is created
- [x] eosio software installed and blockchain started with the Tonomy Contracts initialized
- [x] bash script file is created with exact commands for how the server was setup
- [x] README.md file points to the script and explains what it is for | 1.0 | Antelope staging server in Digital ocean - Definition of done
- [x] Current staging server are deleted
- [x] new droplet with 2Gb RAM, 2x CPU and minimum 100 Gb storage is created
- [x] eosio software installed and blockchain started with the Tonomy Contracts initialized
- [x] bash script file is created with exact commands for how the server was setup
- [x] README.md file points to the script and explains what it is for | infrastructure | antelope staging server in digital ocean definition of done current staging server are deleted new droplet with ram cpu and minimum gb storage is created eosio software installed and blockchain started with the tonomy contracts initialized bash script file is created with exact commands for how the server was setup readme md file points to the script and explains what it is for | 1 |
22,110 | 14,988,805,439 | IssuesEvent | 2021-01-29 02:10:13 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Unable to edit manager script | bug interface/infrastructure | After deleting the target model of a link in a manager script, the script tab of the manager UI is blank. | 1.0 | Unable to edit manager script - After deleting the target model of a link in a manager script, the script tab of the manager UI is blank. | infrastructure | unable to edit manager script after deleting the target model of a link in a manager script the script tab of the manager ui is blank | 1 |
26,494 | 20,158,877,699 | IssuesEvent | 2022-02-09 19:13:05 | xDarkyne/canwecoop | https://api.github.com/repos/xDarkyne/canwecoop | closed | Images get loaded before they are available resulting in a crash (vercel/next.js#33860) | 🐞 type/bug 🏠 area/infrastructure | **Describe the bug**
At random intervals the app crashes in production mode because an image is loaded before it is cached. This is known issue in nextjs 12.0.10.
Issue: vercel/next.js#33860
**To Reproduce**
Steps to reproduce the behavior:
1. run `yarn build`
2. run `yarn start`
3. wait
**Expected behavior**
App should not crash
**Screenshots**
/
**Desktop (please complete the following information):**
- OS: All
- Browser All
- Version All
**Smartphone (please complete the following information):**
- Device: All
- OS: All
- Browser All
- Version All
**Additional context**
/
| 1.0 | Images get loaded before they are available resulting in a crash (vercel/next.js#33860) - **Describe the bug**
At random intervals the app crashes in production mode because an image is loaded before it is cached. This is known issue in nextjs 12.0.10.
Issue: vercel/next.js#33860
**To Reproduce**
Steps to reproduce the behavior:
1. run `yarn build`
2. run `yarn start`
3. wait
**Expected behavior**
App should not crash
**Screenshots**
/
**Desktop (please complete the following information):**
- OS: All
- Browser All
- Version All
**Smartphone (please complete the following information):**
- Device: All
- OS: All
- Browser All
- Version All
**Additional context**
/
| infrastructure | images get loaded before they are available resulting in a crash vercel next js describe the bug at random intervals the app crashes in production mode because an image is loaded before it is cached this is known issue in nextjs issue vercel next js to reproduce steps to reproduce the behavior run yarn build run yarn start wait expected behavior app should not crash screenshots desktop please complete the following information os all browser all version all smartphone please complete the following information device all os all browser all version all additional context | 1 |
58,997 | 8,318,561,112 | IssuesEvent | 2018-09-25 14:57:18 | sky-uk/toolkit | https://api.github.com/repos/sky-uk/toolkit | opened | Remove legacy references within documentation | chore documentation | **Describe the bug**
There are a couple of references to deprecated non-publicly accessible libraries within the README and documentation. These should be updated or removed where applicable.
**To Reproduce**
n/a
**Expected behavior**
n/a
**Attempted Fixes**
n/a
**Screenshots**
n/a
**Desktop (please complete the following information):**
n/a
**Smartphone (please complete the following information):**
n/a
**Additional context**
n/a
| 1.0 | Remove legacy references within documentation - **Describe the bug**
There are a couple of references to deprecated non-publicly accessible libraries within the README and documentation. These should be updated or removed where applicable.
**To Reproduce**
n/a
**Expected behavior**
n/a
**Attempted Fixes**
n/a
**Screenshots**
n/a
**Desktop (please complete the following information):**
n/a
**Smartphone (please complete the following information):**
n/a
**Additional context**
n/a
| non_infrastructure | remove legacy references within documentation describe the bug there are a couple of references to deprecated non publicly accessible libraries within the readme and documentation these should be updated or removed where applicable to reproduce n a expected behavior n a attempted fixes n a screenshots n a desktop please complete the following information n a smartphone please complete the following information n a additional context n a | 0 |
23,324 | 4,006,959,132 | IssuesEvent | 2016-05-12 16:28:38 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | Unscripted Testing Vates | Quality: Manual Tests | 1. Read http://www.mantidproject.org/Unscripted_Manual_Testing
1. Comment against this ticket the OS environment you are testing against.
1. Don't spend more than a few hours on the testing as fatigue will kick in.
1. If you find errors in the possibly scant documentation, then correct them.
1. Close the this ticket once you are done.
If you find bugs:
* Search to see if an issue already exists
* create an issue if it is new
* Urgent bugs or crashes should by against the current release, and assigned to a developer, then go and talk to the developer if possible.
* Less urgent bugs should be against a subsequent release, and assigned if the correct developer is known.
### Specific Notes:
Including data loading in Vates via ParaView. For help see [[Testing_Notes_2.5.0|here]] | 1.0 | Unscripted Testing Vates - 1. Read http://www.mantidproject.org/Unscripted_Manual_Testing
1. Comment against this ticket the OS environment you are testing against.
1. Don't spend more than a few hours on the testing as fatigue will kick in.
1. If you find errors in the possibly scant documentation, then correct them.
1. Close the this ticket once you are done.
If you find bugs:
* Search to see if an issue already exists
* create an issue if it is new
* Urgent bugs or crashes should by against the current release, and assigned to a developer, then go and talk to the developer if possible.
* Less urgent bugs should be against a subsequent release, and assigned if the correct developer is known.
### Specific Notes:
Including data loading in Vates via ParaView. For help see [[Testing_Notes_2.5.0|here]] | non_infrastructure | unscripted testing vates read comment against this ticket the os environment you are testing against don t spend more than a few hours on the testing as fatigue will kick in if you find errors in the possibly scant documentation then correct them close the this ticket once you are done if you find bugs search to see if an issue already exists create an issue if it is new urgent bugs or crashes should by against the current release and assigned to a developer then go and talk to the developer if possible less urgent bugs should be against a subsequent release and assigned if the correct developer is known specific notes including data loading in vates via paraview for help see | 0 |
31,273 | 25,506,836,037 | IssuesEvent | 2022-11-28 10:05:15 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | ApiCompat does not compare implementations to netstandard when a more specific ref is present | area-Infrastructure-libraries | Issue dotnet/runtime#26484 was caused by the netfx build of System.Security.Cryptography.Pkcs running ApiCompat against the netfx ref for System.Security.Cryptography.Pkcs. Both of those libraries have System.Security.Cryptography.Pkcs.CmsSigner as a type-forward to System.Security.dll. There doesn't seem to be any portion of a build which compared the netfx build to the netstandard ref. It seems similarly possible that when a netcoreapp and a netstandard ref are both present for a library that the netcoreapp implementation will only compare against the netcoreapp ref. If the netstandard ref happened to be contradictory nothing would seem to notice.
It's probably sufficient for the ApiCompat task to compare "most specific" and "netstandard"; though perhaps a more generalized "all applicable RIDs" is warranted.
cc @weshaggard | 1.0 | ApiCompat does not compare implementations to netstandard when a more specific ref is present - Issue dotnet/runtime#26484 was caused by the netfx build of System.Security.Cryptography.Pkcs running ApiCompat against the netfx ref for System.Security.Cryptography.Pkcs. Both of those libraries have System.Security.Cryptography.Pkcs.CmsSigner as a type-forward to System.Security.dll. There doesn't seem to be any portion of a build which compared the netfx build to the netstandard ref. It seems similarly possible that when a netcoreapp and a netstandard ref are both present for a library that the netcoreapp implementation will only compare against the netcoreapp ref. If the netstandard ref happened to be contradictory nothing would seem to notice.
It's probably sufficient for the ApiCompat task to compare "most specific" and "netstandard"; though perhaps a more generalized "all applicable RIDs" is warranted.
cc @weshaggard | infrastructure | apicompat does not compare implementations to netstandard when a more specific ref is present issue dotnet runtime was caused by the netfx build of system security cryptography pkcs running apicompat against the netfx ref for system security cryptography pkcs both of those libraries have system security cryptography pkcs cmssigner as a type forward to system security dll there doesn t seem to be any portion of a build which compared the netfx build to the netstandard ref it seems similarly possible that when a netcoreapp and a netstandard ref are both present for a library that the netcoreapp implementation will only compare against the netcoreapp ref if the netstandard ref happened to be contradictory nothing would seem to notice it s probably sufficient for the apicompat task to compare most specific and netstandard though perhaps a more generalized all applicable rids is warranted cc weshaggard | 1 |
4,694 | 5,231,876,742 | IssuesEvent | 2017-01-30 06:14:40 | oppia/oppia | https://api.github.com/repos/oppia/oppia | closed | Create milestones from the requirements in the design doc for answer classification project. | loc: full-stack owner: @anmolshkl TODO: other type: infrastructure | We need to create specific milestones which are self contained and can be delivered as a whole for the [answer classification project](https://docs.google.com/document/d/1Pfc71Z8oD0kmHZK_nf21r03jBUvjQuCsZ0dZo8RnvRE/edit#). | 1.0 | Create milestones from the requirements in the design doc for answer classification project. - We need to create specific milestones which are self contained and can be delivered as a whole for the [answer classification project](https://docs.google.com/document/d/1Pfc71Z8oD0kmHZK_nf21r03jBUvjQuCsZ0dZo8RnvRE/edit#). | infrastructure | create milestones from the requirements in the design doc for answer classification project we need to create specific milestones which are self contained and can be delivered as a whole for the | 1 |
168,825 | 26,704,298,166 | IssuesEvent | 2023-01-27 16:47:35 | department-of-veterans-affairs/vets-design-system-documentation | https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation | closed | Write up a revised plan for migration to USWDS v3 | vsp-design-system-team | ## Description
With the many team changes that happened in Q4 2022, we need a revised plan for migrating to USWDS v3.
## Details
Here are confluence pages with information from previous discovery work as well as plans Katy had written up.
https://vfs.atlassian.net/l/cp/V9XAKj73
https://vfs.atlassian.net/l/cp/h4S56AHp
https://vfs.atlassian.net/l/cp/Vh0E2guw
https://vfs.atlassian.net/l/cp/CPoSuj1D
Katy also left some parting thoughts which I wasn't aware of. This should have some helpful information as well.
https://vfs.atlassian.net/l/cp/fDs2yxHv
## Tasks
- [x] Review the plans that Katy wrote up
- [x] Review the discovery work that was done in Q3 and Q4 2022
- [x] Review work that Brooks had done
- [ ] Draft a revised plan for migration to USWDS v3 based on findings and current state of the upgrade
- [ ] Review with engineers for feedback
- [ ] Present to PO and PM
## Acceptance Criteria
- [x] Presented to the DST | 1.0 | Write up a revised plan for migration to USWDS v3 - ## Description
With the many team changes that happened in Q4 2022, we need a revised plan for migrating to USWDS v3.
## Details
Here are confluence pages with information from previous discovery work as well as plans Katy had written up.
https://vfs.atlassian.net/l/cp/V9XAKj73
https://vfs.atlassian.net/l/cp/h4S56AHp
https://vfs.atlassian.net/l/cp/Vh0E2guw
https://vfs.atlassian.net/l/cp/CPoSuj1D
Katy also left some parting thoughts which I wasn't aware of. This should have some helpful information as well.
https://vfs.atlassian.net/l/cp/fDs2yxHv
## Tasks
- [x] Review the plans that Katy wrote up
- [x] Review the discovery work that was done in Q3 and Q4 2022
- [x] Review work that Brooks had done
- [ ] Draft a revised plan for migration to USWDS v3 based on findings and current state of the upgrade
- [ ] Review with engineers for feedback
- [ ] Present to PO and PM
## Acceptance Criteria
- [x] Presented to the DST | non_infrastructure | write up a revised plan for migration to uswds description with the many team changes that happened in we need a revised plan for migrating to uswds details here are confluence pages with information from previous discovery work as well as plans katy had written up katy also left some parting thoughts which i wasn t aware of this should have some helpful information as well tasks review the plans that katy wrote up review the discovery work that was done in and review work that brooks had done draft a revised plan for migration to uswds based on findings and current state of the upgrade review with engineers for feedback present to po and pm acceptance criteria presented to the dst | 0 |
35,674 | 32,019,047,168 | IssuesEvent | 2023-09-22 01:48:48 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | opened | Add ESLint rule to flag improper dispatching of actions | P2 Type: Infrastructure | ## Feature Description
There are essentially 2 ways to dispatch an action properly such that it behaves as expected:
- Using `registry.dispatch`
- Using `yield`
The latter is of course only possible in a generator function, but calling the registry's `dispatch` is also possible in the same context but will not be processed in the same way and should be avoided.
This is fairly easy to identify statically, so we should introduce an ESLint rule to flag it
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* A new ESLint rule `no-yield-dispatch` should be added to our bundled `eslint-plugin-sitekit`
* The rule should error if `yield registry.dispatch` or `yield dispatch` is detected in a generator function with the following message:
> Only plain objects should be yielded from action generator functions. `dispatch` always returns a Promise.
## Implementation Brief
* <!-- One or more bullet points for how to technically implement the feature. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the feature works. -->
## QA Brief
* <!-- One or more bullet points for how to test that the feature works as expected. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| 1.0 | Add ESLint rule to flag improper dispatching of actions - ## Feature Description
There are essentially 2 ways to dispatch an action properly such that it behaves as expected:
- Using `registry.dispatch`
- Using `yield`
The latter is of course only possible in a generator function, but calling the registry's `dispatch` is also possible in the same context but will not be processed in the same way and should be avoided.
This is fairly easy to identify statically, so we should introduce an ESLint rule to flag it
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* A new ESLint rule `no-yield-dispatch` should be added to our bundled `eslint-plugin-sitekit`
* The rule should error if `yield registry.dispatch` or `yield dispatch` is detected in a generator function with the following message:
> Only plain objects should be yielded from action generator functions. `dispatch` always returns a Promise.
## Implementation Brief
* <!-- One or more bullet points for how to technically implement the feature. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the feature works. -->
## QA Brief
* <!-- One or more bullet points for how to test that the feature works as expected. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| infrastructure | add eslint rule to flag improper dispatching of actions feature description there are essentially ways to dispatch an action properly such that it behaves as expected using registry dispatch using yield the latter is of course only possible in a generator function but calling the registry s dispatch is also possible in the same context but will not be processed in the same way and should be avoided this is fairly easy to identify statically so we should introduce an eslint rule to flag it do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria a new eslint rule no yield dispatch should be added to our bundled eslint plugin sitekit the rule should error if yield registry dispatch or yield dispatch is detected in a generator function with the following message only plain objects should be yielded from action generator functions dispatch always returns a promise implementation brief test coverage qa brief changelog entry | 1 |
35,379 | 31,150,321,439 | IssuesEvent | 2023-08-16 09:30:07 | Chaste/Chaste | https://api.github.com/repos/Chaste/Chaste | closed | Support VTK 9.x | component: infrastructure priority: high | This issue continues the discussion for legacy trac ticket 3088:
___
[fcooper8472](https://github.com/orgs/Chaste/people/fcooper8472) created the following ticket on 2022-02-02 at 13:26:24, it is owned by [fcooper8472](https://github.com/orgs/Chaste/people/fcooper8472)
Changes can be seen here: https://www.kitware.com/vtk-9-0-0-available-for-download/
This currently prevents Chaste working (easily) on macOS as Homebrew gives v9.
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-08-01 at 11:05:05
VTK9 was added to the rotation in February:
- [2d47416](https://github.com/Chaste/Chaste/commit/2d47416)
- [234186e](https://chaste.cs.ox.ac.uk/trac/changeset/234186e/infrastructure-scripts)
but it has yet to compile successfully: https://chaste.cs.ox.ac.uk/buildbot/builders/Portability%206_Sat?numbuilds=30
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-08-04 at 13:33:48
Core components that need VTK are building fine (`global, mesh,` etc.).
The `cell_based` component fails when including `VtkMeshReader.hpp`. This is probably because it doesn't have the correct include path, but I can't work out why that may be.
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-10-24 at 15:17:59
Replying to jmpf@…:
> Core components that need VTK are building fine (`global, mesh,` etc.).
>
> The `cell_based` component fails when including `VtkMeshReader.hpp`. This is probably because it doesn't have the correct include path, but I can't work out why that may be.
Note that this is a CMake configuration problem and is to do with the way CMake now detects libraries and include paths. | 1.0 | Support VTK 9.x - This issue continues the discussion for legacy trac ticket 3088:
___
[fcooper8472](https://github.com/orgs/Chaste/people/fcooper8472) created the following ticket on 2022-02-02 at 13:26:24, it is owned by [fcooper8472](https://github.com/orgs/Chaste/people/fcooper8472)
Changes can be seen here: https://www.kitware.com/vtk-9-0-0-available-for-download/
This currently prevents Chaste working (easily) on macOS as Homebrew gives v9.
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-08-01 at 11:05:05
VTK9 was added to the rotation in February:
- [2d47416](https://github.com/Chaste/Chaste/commit/2d47416)
- [234186e](https://chaste.cs.ox.ac.uk/trac/changeset/234186e/infrastructure-scripts)
but it has yet to compile successfully: https://chaste.cs.ox.ac.uk/buildbot/builders/Portability%206_Sat?numbuilds=30
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-08-04 at 13:33:48
Core components that need VTK are building fine (`global, mesh,` etc.).
The `cell_based` component fails when including `VtkMeshReader.hpp`. This is probably because it doesn't have the correct include path, but I can't work out why that may be.
___
Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2022-10-24 at 15:17:59
Replying to jmpf@…:
> Core components that need VTK are building fine (`global, mesh,` etc.).
>
> The `cell_based` component fails when including `VtkMeshReader.hpp`. This is probably because it doesn't have the correct include path, but I can't work out why that may be.
Note that this is a CMake configuration problem and is to do with the way CMake now detects libraries and include paths. | infrastructure | support vtk x this issue continues the discussion for legacy trac ticket created the following ticket on at it is owned by changes can be seen here this currently prevents chaste working easily on macos as homebrew gives comment by on at was added to the rotation in february but it has yet to compile successfully comment by on at core components that need vtk are building fine global mesh etc the cell based component fails when including vtkmeshreader hpp this is probably because it doesn t have the correct include path but i can t work out why that may be comment by on at replying to jmpf … core components that need vtk are building fine global mesh etc the cell based component fails when including vtkmeshreader hpp this is probably because it doesn t have the correct include path but i can t work out why that may be note that this is a cmake configuration problem and is to do with the way cmake now detects libraries and include paths | 1 |
3,511 | 4,365,362,339 | IssuesEvent | 2016-08-03 10:31:53 | OpenSCAP/scap-security-guide | https://api.github.com/repos/OpenSCAP/scap-security-guide | closed | [Blocker] [BugFix] Running 'make content' for RHEL/6 fails after introduction of shared/transforms/shared_shorthand2xccdf.xslt | BLOCKER bugfix Infrastructure | Currently (2016-06-16) it's not possible ```make content``` from RHEL/6 folder to succeed (behavior is identical on both openscap-1.0.x and openscap-1.2.x):
```
[root@localhost 6]# make content
xsltproc -o output/guide.xml ../../shared/transforms/includelogo.xslt input/guide.xml
xsltproc --param withtest "1" --stringparam SHARED_RP "/root/scap-security-guide/shared" -o output/shorthand.xml input/guide.xslt output/guide.xml
xmllint --format --output output/shorthand.xml output/shorthand.xml
# Create output/contributors.xml file:
# * holding <text> like root element and
# * names of the individual SSG contributor(s) within <contributor> element(s)
echo "<text>" > output/contributors.xml
sed -n -e 's/\* \(.*\)/<contributor>\1<\/contributor>/p' ../../shared/../Contributors.md >> output/contributors.xml
echo "</text>" >> output/contributors.xml
xsltproc --stringparam ssg_version "0.1.29" -o output/xccdf-unlinked-unresolved.xml transforms/shorthand2xccdf.xslt output/shorthand.xml
warning: failed to load external entity "../../shared/output/contributors.xml"
runtime error: file ../../shared/transforms/shared_shorthand2xccdf.xslt line 196 element value-of
Variable 'cisuri' has not been declared.
xmlXPathCompiledEval: evaluation failed
runtime error: file ../../shared/transforms/shared_shorthand2xccdf.xslt line 196 element value-of
XPath evaluation returned no result.
make: *** [output/xccdf-unlinked-unresolved.xml] Error 10
```
This is because after move of ```shorthand2xccdf.xslt``` to ```shared/transforms``` folder the path to $(OUT)/contributors.xml changed.
Strangely enough ```RHEL/7```'s ```make content``` still works properly:
```
[root@localhost 7]# make clean
rm -f output/*.xml output/*.html output/*.xhtml output/*.pdf output/*.spec output/*.tar output/*.gz output/*.ini output/*.csv
rm -rf dist/content dist/guide
rm -rf build
[root@localhost 7]# make content
xsltproc -o output/guide.xml ../../shared/transforms/includelogo.xslt input/guide.xml
xsltproc --param withtest "1" --stringparam SHARED_RP "/root/scap-security-guide/shared" -o output/shorthand.xml input/guide.xslt output/guide.xml
xmllint --format --output output/shorthand.xml output/shorthand.xml
# Create output/contributors.xml file:
# * holding <text> like root element and
# * names of the individual SSG contributor(s) within <contributor> element(s)
echo "<text>" > output/contributors.xml
sed -n -e 's/\* \(.*\)/<contributor>\1<\/contributor>/p' ../../shared/../Contributors.md >> output/contributors.xml
echo "</text>" >> output/contributors.xml
xsltproc --stringparam ssg_version "0.1.29" -o output/xccdf-unlinked-unresolved.xml transforms/shorthand2xccdf.xslt output/shorthand.xml
warning: failed to load external entity "../../shared/output/contributors.xml"
oscap xccdf resolve -o output/xccdf-unlinked-empty-groups.xml output/xccdf-unlinked-unresolved.xml
../../shared/utils/unselect-empty-xccdf-groups.py --input output/xccdf-unlinked-empty-groups.xml --output output/xccdf-unlinked-empty-groups-unselected.xml
Unselected empty groups in 'standard'.
Unselected empty groups in 'pci-dss'.
Unselected empty groups in 'C2S'.
Unselected empty groups in 'rht-ccp'.
Unselected empty groups in 'common'.
Unselected empty groups in 'stig-rhel7-workstation-upstream'.
Unselected empty groups in 'stig-rhel7-server-gui-upstream'.
Unselected empty groups in 'stig-rhel7-server-upstream'.
Unselected empty groups in 'ospp-rhel7-server'.
Unselected empty groups in 'nist-cl-il-al'.
Unselected empty groups in 'cjis-rhel7-server'.
oscap xccdf resolve -o output/xccdf-unlinked-resolved.xml output/xccdf-unlinked-empty-groups-unselected.xml
xsltproc -o output/ocil-unlinked.xml ../../shared/transforms/xccdf-create-ocil.xslt output/xccdf-unlinked-resolved.xml
xmllint --format --output output/unlinked-rhel7-ocil.xml output/ocil-unlinked.xml
xsltproc -stringparam product "rhel7" -o output/xccdf-unlinked-ocilrefs.xml ../../shared/transforms/xccdf-ocilcheck2ref.xslt output/xccdf-unlinked-resolved.xml
# Make intermediate build/remediations directory to hold final list of remediation scripts for rhel7
mkdir -p build/remediations
# Search ../../shared/remediations/bash and input/remediations directories to find all product specific remediation scripts,
# which are regular files (not symlinks). Merge the final list into build/remediations directory
find ../../shared/remediations/bash input/remediations/bash -maxdepth 1 -type f -name '*.sh' -exec cp {} build/remediations ';'
../../shared/transforms/combineremediations.py rhel7 build/remediations output/bash-remediations.xml
Notification: Removed the 'accounts_passwords_pam_unix_remember.sh' remediation script from merging as the platform identifier in the script is missing!
Notification: Merged 293 remediation scripts into XML document.
xsltproc -stringparam remediations "/root/scap-security-guide/RHEL/7/output/bash-remediations.xml" -o output/xccdf-unlinked-withremediations.xml ../../shared/transforms/xccdf-addremediations.xslt output/xccdf-unlinked-ocilrefs.xml
xmllint --format --output output/xccdf-unlinked-withremediations.xml output/xccdf-unlinked-withremediations.xml
# Make intermediate build/rhel7_oval directory to hold final list of OVAL checks for rhel7
mkdir -p build/rhel7_oval
# Search ../../shared/oval and input/oval directories to find all product specific OVAL checks,
# which are regular files (not symlinks). Merge the final list into build/rhel7_oval directory
find ../../shared/oval input/oval -maxdepth 1 -type f -name '*.xml' -exec cp {} build/rhel7_oval ';'
# If openscap on the system supports OVAL-5.11 language version, include also OVAL-5.11 checks
# into final list of OVAL checks
xmlwf build/rhel7_oval/*.xml
../../shared/transforms/combineovals.py ../../shared/../config rhel7 build/rhel7_oval > output/unlinked-rhel7-oval.xml
Notification: Merged 380 OVAL checks into OVAL document.
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_tcp_syncookies_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_secure_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: login_banner_text
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: test_auditd_conf_log_group_root
Notification: this ID is used more than once and should represent equivalent elements: object_auditd_conf_log_group_root
Notification: this ID is used more than once and should represent equivalent elements: test_bootloader_recovery_disabled
Notification: this ID is used more than once and should represent equivalent elements: object_bootloader_disable_recovery_argument
Notification: this ID is used more than once and should represent equivalent elements: state_bootloader_disable_recovery_argument
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_forwarding_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: object_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: test_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: object_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: state_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_alternative_names
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: variable_not_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: var_removable_partition
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_log_martians_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_ra_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_secure_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: object_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: test_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: object_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: state_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_alternative_names
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: variable_not_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: var_removable_partition
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_icmp_ignore_bogus_error_responses_value
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: obj_package_gpg-pubkey
Notification: this ID is used more than once and should represent equivalent elements: obj_package_gpg-pubkey
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_rp_filter_value
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_log_martians_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_ra_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_rp_filter_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_icmp_echo_ignore_broadcasts_value
xmllint --format --output output/unlinked-rhel7-oval.xml output/unlinked-rhel7-oval.xml
cp output/xccdf-unlinked-final.xml output/unlinked-rhel7-xccdf.xml
xsltproc -o output/unlinked-rhel7-xccdf-guide.xml transforms/xccdf-removeaux.xslt output/unlinked-rhel7-xccdf.xml
../../shared/transforms/cpe_generate.py output/unlinked-rhel7-oval.xml input/oval/platform/rhel7-cpe-dictionary.xml ssg
../../shared/transforms/relabelids.py unlinked-rhel7-xccdf.xml ssg
Removing <check-content> OVAL element for service_autofs_disabled.
Removing <check-content> OVAL element for service_debug-shell_disabled.
Removing <check-content> OVAL element for disable_ctrlaltdel_reboot.
Removing <check-content> OVAL element for smartcard_auth.
Removing <check-content> OVAL element for service_bluetooth_disabled.
Removing <check-content> OVAL element for service_firewalld_enabled.
Removing <check-content> OVAL element for service_rsyslog_enabled.
Removing <check-content> OVAL element for rsyslog_files_ownership.
Removing <check-content> OVAL element for rsyslog_files_groupownership.
Removing <check-content> OVAL element for rsyslog_files_permissions.
Removing <check-content> OVAL element for service_auditd_enabled.
Removing <check-content> OVAL element for service_xinetd_disabled.
Removing <check-content> OVAL element for package_tcp_wrappers_installed.
Removing <check-content> OVAL element for service_telnet_disabled.
Removing <check-content> OVAL element for service_rexec_disabled.
Removing <check-content> OVAL element for service_rsh_disabled.
Removing <check-content> OVAL element for service_rlogin_disabled.
Removing <check-content> OVAL element for service_ypbind_disabled.
Removing <check-content> OVAL element for service_tftp_disabled.
Removing <check-content> OVAL element for service_abrtd_disabled.
Removing <check-content> OVAL element for service_acpid_disabled.
Removing <check-content> OVAL element for service_certmonger_disabled.
Removing <check-content> OVAL element for service_cgconfig_disabled.
Removing <check-content> OVAL element for service_cgred_disabled.
Removing <check-content> OVAL element for service_cpupower_disabled.
Removing <check-content> OVAL element for service_irqbalance_enabled.
Removing <check-content> OVAL element for service_kdump_disabled.
Removing <check-content> OVAL element for service_mdmonitor_disabled.
Removing <check-content> OVAL element for service_messagebus_disabled.
Removing <check-content> OVAL element for service_netconsole_disabled.
Removing <check-content> OVAL element for service_ntpdate_disabled.
Removing <check-content> OVAL element for service_oddjobd_disabled.
Removing <check-content> OVAL element for service_portreserve_disabled.
Removing <check-content> OVAL element for service_psacct_enabled.
Removing <check-content> OVAL element for service_qpidd_disabled.
Removing <check-content> OVAL element for service_quota_nld_disabled.
Removing <check-content> OVAL element for service_rdisc_disabled.
Removing <check-content> OVAL element for service_rhnsd_disabled.
Removing <check-content> OVAL element for service_rhsmcertd_disabled.
Removing <check-content> OVAL element for service_saslauthd_disabled.
Removing <check-content> OVAL element for service_smartd_disabled.
Removing <check-content> OVAL element for service_sysstat_disabled.
Removing <check-content> OVAL element for service_crond_enabled.
Removing <check-content> OVAL element for service_atd_disabled.
Removing <check-content> OVAL element for package_openssh-server_installed.
Removing <check-content> OVAL element for service_sshd_enabled.
Removing <check-content> OVAL element for package_sssd_installed.
Removing <check-content> OVAL element for service_sssd_enabled.
Removing <check-content> OVAL element for xwindows_runlevel_setting.
Removing <check-content> OVAL element for service_avahi-daemon_disabled.
Removing <check-content> OVAL element for service_cups_disabled.
Removing <check-content> OVAL element for service_dhcpd_disabled.
Removing <check-content> OVAL element for service_chronyd_or_ntpd_enabled.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_remote_server.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_multiple_servers.
Removing <check-content> OVAL element for service_postfix_enabled.
Removing <check-content> OVAL element for postfix_network_listening_disabled.
Removing <check-content> OVAL element for service_nfslock_disabled.
Removing <check-content> OVAL element for service_rpcgssd_disabled.
Removing <check-content> OVAL element for service_rpcbind_disabled.
Removing <check-content> OVAL element for service_rpcidmapd_disabled.
Removing <check-content> OVAL element for service_nfs_disabled.
Removing <check-content> OVAL element for service_rpcsvcgssd_disabled.
Removing <check-content> OVAL element for service_named_disabled.
Removing <check-content> OVAL element for service_vsftpd_disabled.
Removing <check-content> OVAL element for service_httpd_disabled.
Removing <check-content> OVAL element for service_dovecot_disabled.
Removing <check-content> OVAL element for dovecot_enable_ssl.
Removing <check-content> OVAL element for dovecot_disable_plaintext_auth.
Removing <check-content> OVAL element for service_zebra_disabled.
Removing <check-content> OVAL element for package_quagga_removed.
Removing <check-content> OVAL element for service_smb_disabled.
Removing <check-content> OVAL element for package_samba_removed.
Removing <check-content> OVAL element for service_squid_disabled.
Removing <check-content> OVAL element for service_snmpd_disabled.
../../shared/transforms/relabelids.py xccdf-unlinked-ocilrefs.xml ssg
Removing <check-content> OVAL element for service_autofs_disabled.
Removing <check-content> OVAL element for service_debug-shell_disabled.
Removing <check-content> OVAL element for disable_ctrlaltdel_reboot.
Removing <check-content> OVAL element for smartcard_auth.
Removing <check-content> OVAL element for service_bluetooth_disabled.
Removing <check-content> OVAL element for service_firewalld_enabled.
Removing <check-content> OVAL element for service_rsyslog_enabled.
Removing <check-content> OVAL element for rsyslog_files_ownership.
Removing <check-content> OVAL element for rsyslog_files_groupownership.
Removing <check-content> OVAL element for rsyslog_files_permissions.
Removing <check-content> OVAL element for service_auditd_enabled.
Removing <check-content> OVAL element for service_xinetd_disabled.
Removing <check-content> OVAL element for package_tcp_wrappers_installed.
Removing <check-content> OVAL element for service_telnet_disabled.
Removing <check-content> OVAL element for service_rexec_disabled.
Removing <check-content> OVAL element for service_rsh_disabled.
Removing <check-content> OVAL element for service_rlogin_disabled.
Removing <check-content> OVAL element for service_ypbind_disabled.
Removing <check-content> OVAL element for service_tftp_disabled.
Removing <check-content> OVAL element for service_abrtd_disabled.
Removing <check-content> OVAL element for service_acpid_disabled.
Removing <check-content> OVAL element for service_certmonger_disabled.
Removing <check-content> OVAL element for service_cgconfig_disabled.
Removing <check-content> OVAL element for service_cgred_disabled.
Removing <check-content> OVAL element for service_cpupower_disabled.
Removing <check-content> OVAL element for service_irqbalance_enabled.
Removing <check-content> OVAL element for service_kdump_disabled.
Removing <check-content> OVAL element for service_mdmonitor_disabled.
Removing <check-content> OVAL element for service_messagebus_disabled.
Removing <check-content> OVAL element for service_netconsole_disabled.
Removing <check-content> OVAL element for service_ntpdate_disabled.
Removing <check-content> OVAL element for service_oddjobd_disabled.
Removing <check-content> OVAL element for service_portreserve_disabled.
Removing <check-content> OVAL element for service_psacct_enabled.
Removing <check-content> OVAL element for service_qpidd_disabled.
Removing <check-content> OVAL element for service_quota_nld_disabled.
Removing <check-content> OVAL element for service_rdisc_disabled.
Removing <check-content> OVAL element for service_rhnsd_disabled.
Removing <check-content> OVAL element for service_rhsmcertd_disabled.
Removing <check-content> OVAL element for service_saslauthd_disabled.
Removing <check-content> OVAL element for service_smartd_disabled.
Removing <check-content> OVAL element for service_sysstat_disabled.
Removing <check-content> OVAL element for service_crond_enabled.
Removing <check-content> OVAL element for service_atd_disabled.
Removing <check-content> OVAL element for package_openssh-server_installed.
Removing <check-content> OVAL element for service_sshd_enabled.
Removing <check-content> OVAL element for package_sssd_installed.
Removing <check-content> OVAL element for service_sssd_enabled.
Removing <check-content> OVAL element for xwindows_runlevel_setting.
Removing <check-content> OVAL element for service_avahi-daemon_disabled.
Removing <check-content> OVAL element for service_cups_disabled.
Removing <check-content> OVAL element for service_dhcpd_disabled.
Removing <check-content> OVAL element for service_chronyd_or_ntpd_enabled.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_remote_server.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_multiple_servers.
Removing <check-content> OVAL element for service_postfix_enabled.
Removing <check-content> OVAL element for postfix_network_listening_disabled.
Removing <check-content> OVAL element for service_nfslock_disabled.
Removing <check-content> OVAL element for service_rpcgssd_disabled.
Removing <check-content> OVAL element for service_rpcbind_disabled.
Removing <check-content> OVAL element for service_rpcidmapd_disabled.
Removing <check-content> OVAL element for service_nfs_disabled.
Removing <check-content> OVAL element for service_rpcsvcgssd_disabled.
Removing <check-content> OVAL element for service_named_disabled.
Removing <check-content> OVAL element for service_vsftpd_disabled.
Removing <check-content> OVAL element for service_httpd_disabled.
Removing <check-content> OVAL element for service_dovecot_disabled.
Removing <check-content> OVAL element for dovecot_enable_ssl.
Removing <check-content> OVAL element for dovecot_disable_plaintext_auth.
Removing <check-content> OVAL element for service_zebra_disabled.
Removing <check-content> OVAL element for package_quagga_removed.
Removing <check-content> OVAL element for service_smb_disabled.
Removing <check-content> OVAL element for package_samba_removed.
Removing <check-content> OVAL element for service_squid_disabled.
Removing <check-content> OVAL element for service_snmpd_disabled.
xsltproc -o output/ssg-rhel7-oval.xml transforms/oval-fix-test-attestation-urls.xslt output/ssg-rhel7-oval.xml
xsltproc /usr/share/openscap/xsl/xccdf_1.1_remove_dangling_sub.xsl output/ssg-rhel7-xccdf.xml \
> output/ssg-rhel7-xccdf-nodangles.xml
xsltproc --stringparam reverse_DNS org.ssgproject.content /usr/share/openscap/xsl/xccdf_1.1_to_1.2.xsl \
output/ssg-rhel7-xccdf-nodangles.xml > output/ssg-rhel7-xccdf-1.2.xml
sed -i '/idref="dangling reference to /d' output/ssg-rhel7-xccdf-1.2.xml
sed -i 's/style="SCAP_1.1"/style="SCAP_1.2"/' output/ssg-rhel7-xccdf-1.2.xml
oscap ds sds-compose output/ssg-rhel7-xccdf-1.2.xml output/ssg-rhel7-ds.xml
OpenSCAP Error: Unknown document type: 'ssg-rhel7-ocil.xml' [oscapxml.c:630]
sed -i 's/schematron-version="[0-9].[0-9]"/schematron-version="1.2"/' output/ssg-rhel7-ds.xml
oscap ds sds-add output/ssg-rhel7-cpe-dictionary.xml output/ssg-rhel7-ds.xml
oscap ds sds-add output/ssg-rhel7-oval.xml output/ssg-rhel7-ds.xml
# Fixes https://github.com/OpenSCAP/scap-security-guide/issues/1100
# Fixes https://github.com/OpenSCAP/scap-security-guide/issues/1101
../../shared/transforms/datastream_move_ocil_to_ds_checks.py output/ssg-rhel7-ds.xml output/ssg-rhel7-ds.xml
../../shared/utils/enable-derivatives.py --enable-centos -i output/ssg-rhel7-xccdf.xml -o output/ssg-centos7-xccdf.xml
../../shared/utils/enable-derivatives.py --enable-centos -i output/ssg-rhel7-ds.xml -o output/ssg-centos7-ds.xml
../../shared/utils/enable-derivatives.py --enable-sl -i output/ssg-rhel7-xccdf.xml -o output/ssg-sl7-xccdf.xml
../../shared/utils/enable-derivatives.py --enable-sl -i output/ssg-rhel7-ds.xml -o output/ssg-sl7-ds.xml
[root@localhost 7]# echo $?
0
[root@localhost 7]#
```
Will investigate further. | 1.0 | [Blocker] [BugFix] Running 'make content' for RHEL/6 fails after introduction of shared/transforms/shared_shorthand2xccdf.xslt - Currently (2016-06-16) it's not possible ```make content``` from RHEL/6 folder to succeed (behavior is identical on both openscap-1.0.x and openscap-1.2.x):
```
[root@localhost 6]# make content
xsltproc -o output/guide.xml ../../shared/transforms/includelogo.xslt input/guide.xml
xsltproc --param withtest "1" --stringparam SHARED_RP "/root/scap-security-guide/shared" -o output/shorthand.xml input/guide.xslt output/guide.xml
xmllint --format --output output/shorthand.xml output/shorthand.xml
# Create output/contributors.xml file:
# * holding <text> like root element and
# * names of the individual SSG contributor(s) within <contributor> element(s)
echo "<text>" > output/contributors.xml
sed -n -e 's/\* \(.*\)/<contributor>\1<\/contributor>/p' ../../shared/../Contributors.md >> output/contributors.xml
echo "</text>" >> output/contributors.xml
xsltproc --stringparam ssg_version "0.1.29" -o output/xccdf-unlinked-unresolved.xml transforms/shorthand2xccdf.xslt output/shorthand.xml
warning: failed to load external entity "../../shared/output/contributors.xml"
runtime error: file ../../shared/transforms/shared_shorthand2xccdf.xslt line 196 element value-of
Variable 'cisuri' has not been declared.
xmlXPathCompiledEval: evaluation failed
runtime error: file ../../shared/transforms/shared_shorthand2xccdf.xslt line 196 element value-of
XPath evaluation returned no result.
make: *** [output/xccdf-unlinked-unresolved.xml] Error 10
```
This is because after move of ```shorthand2xccdf.xslt``` to ```shared/transforms``` folder the path to $(OUT)/contributors.xml changed.
Strangely enough ```RHEL/7```'s ```make content``` still works properly:
```
[root@localhost 7]# make clean
rm -f output/*.xml output/*.html output/*.xhtml output/*.pdf output/*.spec output/*.tar output/*.gz output/*.ini output/*.csv
rm -rf dist/content dist/guide
rm -rf build
[root@localhost 7]# make content
xsltproc -o output/guide.xml ../../shared/transforms/includelogo.xslt input/guide.xml
xsltproc --param withtest "1" --stringparam SHARED_RP "/root/scap-security-guide/shared" -o output/shorthand.xml input/guide.xslt output/guide.xml
xmllint --format --output output/shorthand.xml output/shorthand.xml
# Create output/contributors.xml file:
# * holding <text> like root element and
# * names of the individual SSG contributor(s) within <contributor> element(s)
echo "<text>" > output/contributors.xml
sed -n -e 's/\* \(.*\)/<contributor>\1<\/contributor>/p' ../../shared/../Contributors.md >> output/contributors.xml
echo "</text>" >> output/contributors.xml
xsltproc --stringparam ssg_version "0.1.29" -o output/xccdf-unlinked-unresolved.xml transforms/shorthand2xccdf.xslt output/shorthand.xml
warning: failed to load external entity "../../shared/output/contributors.xml"
oscap xccdf resolve -o output/xccdf-unlinked-empty-groups.xml output/xccdf-unlinked-unresolved.xml
../../shared/utils/unselect-empty-xccdf-groups.py --input output/xccdf-unlinked-empty-groups.xml --output output/xccdf-unlinked-empty-groups-unselected.xml
Unselected empty groups in 'standard'.
Unselected empty groups in 'pci-dss'.
Unselected empty groups in 'C2S'.
Unselected empty groups in 'rht-ccp'.
Unselected empty groups in 'common'.
Unselected empty groups in 'stig-rhel7-workstation-upstream'.
Unselected empty groups in 'stig-rhel7-server-gui-upstream'.
Unselected empty groups in 'stig-rhel7-server-upstream'.
Unselected empty groups in 'ospp-rhel7-server'.
Unselected empty groups in 'nist-cl-il-al'.
Unselected empty groups in 'cjis-rhel7-server'.
oscap xccdf resolve -o output/xccdf-unlinked-resolved.xml output/xccdf-unlinked-empty-groups-unselected.xml
xsltproc -o output/ocil-unlinked.xml ../../shared/transforms/xccdf-create-ocil.xslt output/xccdf-unlinked-resolved.xml
xmllint --format --output output/unlinked-rhel7-ocil.xml output/ocil-unlinked.xml
xsltproc -stringparam product "rhel7" -o output/xccdf-unlinked-ocilrefs.xml ../../shared/transforms/xccdf-ocilcheck2ref.xslt output/xccdf-unlinked-resolved.xml
# Make intermediate build/remediations directory to hold final list of remediation scripts for rhel7
mkdir -p build/remediations
# Search ../../shared/remediations/bash and input/remediations directories to find all product specific remediation scripts,
# which are regular files (not symlinks). Merge the final list into build/remediations directory
find ../../shared/remediations/bash input/remediations/bash -maxdepth 1 -type f -name '*.sh' -exec cp {} build/remediations ';'
../../shared/transforms/combineremediations.py rhel7 build/remediations output/bash-remediations.xml
Notification: Removed the 'accounts_passwords_pam_unix_remember.sh' remediation script from merging as the platform identifier in the script is missing!
Notification: Merged 293 remediation scripts into XML document.
xsltproc -stringparam remediations "/root/scap-security-guide/RHEL/7/output/bash-remediations.xml" -o output/xccdf-unlinked-withremediations.xml ../../shared/transforms/xccdf-addremediations.xslt output/xccdf-unlinked-ocilrefs.xml
xmllint --format --output output/xccdf-unlinked-withremediations.xml output/xccdf-unlinked-withremediations.xml
# Make intermediate build/rhel7_oval directory to hold final list of OVAL checks for rhel7
mkdir -p build/rhel7_oval
# Search ../../shared/oval and input/oval directories to find all product specific OVAL checks,
# which are regular files (not symlinks). Merge the final list into build/rhel7_oval directory
find ../../shared/oval input/oval -maxdepth 1 -type f -name '*.xml' -exec cp {} build/rhel7_oval ';'
# If openscap on the system supports OVAL-5.11 language version, include also OVAL-5.11 checks
# into final list of OVAL checks
xmlwf build/rhel7_oval/*.xml
../../shared/transforms/combineovals.py ../../shared/../config rhel7 build/rhel7_oval > output/unlinked-rhel7-oval.xml
Notification: Merged 380 OVAL checks into OVAL document.
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_tcp_syncookies_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_secure_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: login_banner_text
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: test_auditd_conf_log_group_root
Notification: this ID is used more than once and should represent equivalent elements: object_auditd_conf_log_group_root
Notification: this ID is used more than once and should represent equivalent elements: test_bootloader_recovery_disabled
Notification: this ID is used more than once and should represent equivalent elements: object_bootloader_disable_recovery_argument
Notification: this ID is used more than once and should represent equivalent elements: state_bootloader_disable_recovery_argument
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_forwarding_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: object_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: test_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: object_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: state_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_alternative_names
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: variable_not_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: var_removable_partition
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_log_martians_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_ra_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_secure_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: test_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: object_removable_partition_doesnt_exist
Notification: this ID is used more than once and should represent equivalent elements: test_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: object_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: state_var_removable_partition_is_cd_dvd_drive
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_alternative_names
Notification: this ID is used more than once and should represent equivalent elements: variable_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: variable_not_cd_dvd_drive_regex_pattern
Notification: this ID is used more than once and should represent equivalent elements: var_removable_partition
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_icmp_ignore_bogus_error_responses_value
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: obj_package_gpg-pubkey
Notification: this ID is used more than once and should represent equivalent elements: obj_package_gpg-pubkey
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_rp_filter_value
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_first_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_second_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_third_digit_of_umask_from_var_accounts_user_umask
Notification: this ID is used more than once and should represent equivalent elements: var_accounts_user_umask_umask_as_number
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_log_martians_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_default_accept_redirects_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv6_conf_all_accept_ra_value
Notification: this ID is used more than once and should represent equivalent elements: test_unix_family
Notification: this ID is used more than once and should represent equivalent elements: state_unix_family
Notification: this ID is used more than once and should represent equivalent elements: obj_unix_family
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_all_rp_filter_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_conf_default_accept_source_route_value
Notification: this ID is used more than once and should represent equivalent elements: sysctl_net_ipv4_icmp_echo_ignore_broadcasts_value
xmllint --format --output output/unlinked-rhel7-oval.xml output/unlinked-rhel7-oval.xml
cp output/xccdf-unlinked-final.xml output/unlinked-rhel7-xccdf.xml
xsltproc -o output/unlinked-rhel7-xccdf-guide.xml transforms/xccdf-removeaux.xslt output/unlinked-rhel7-xccdf.xml
../../shared/transforms/cpe_generate.py output/unlinked-rhel7-oval.xml input/oval/platform/rhel7-cpe-dictionary.xml ssg
../../shared/transforms/relabelids.py unlinked-rhel7-xccdf.xml ssg
Removing <check-content> OVAL element for service_autofs_disabled.
Removing <check-content> OVAL element for service_debug-shell_disabled.
Removing <check-content> OVAL element for disable_ctrlaltdel_reboot.
Removing <check-content> OVAL element for smartcard_auth.
Removing <check-content> OVAL element for service_bluetooth_disabled.
Removing <check-content> OVAL element for service_firewalld_enabled.
Removing <check-content> OVAL element for service_rsyslog_enabled.
Removing <check-content> OVAL element for rsyslog_files_ownership.
Removing <check-content> OVAL element for rsyslog_files_groupownership.
Removing <check-content> OVAL element for rsyslog_files_permissions.
Removing <check-content> OVAL element for service_auditd_enabled.
Removing <check-content> OVAL element for service_xinetd_disabled.
Removing <check-content> OVAL element for package_tcp_wrappers_installed.
Removing <check-content> OVAL element for service_telnet_disabled.
Removing <check-content> OVAL element for service_rexec_disabled.
Removing <check-content> OVAL element for service_rsh_disabled.
Removing <check-content> OVAL element for service_rlogin_disabled.
Removing <check-content> OVAL element for service_ypbind_disabled.
Removing <check-content> OVAL element for service_tftp_disabled.
Removing <check-content> OVAL element for service_abrtd_disabled.
Removing <check-content> OVAL element for service_acpid_disabled.
Removing <check-content> OVAL element for service_certmonger_disabled.
Removing <check-content> OVAL element for service_cgconfig_disabled.
Removing <check-content> OVAL element for service_cgred_disabled.
Removing <check-content> OVAL element for service_cpupower_disabled.
Removing <check-content> OVAL element for service_irqbalance_enabled.
Removing <check-content> OVAL element for service_kdump_disabled.
Removing <check-content> OVAL element for service_mdmonitor_disabled.
Removing <check-content> OVAL element for service_messagebus_disabled.
Removing <check-content> OVAL element for service_netconsole_disabled.
Removing <check-content> OVAL element for service_ntpdate_disabled.
Removing <check-content> OVAL element for service_oddjobd_disabled.
Removing <check-content> OVAL element for service_portreserve_disabled.
Removing <check-content> OVAL element for service_psacct_enabled.
Removing <check-content> OVAL element for service_qpidd_disabled.
Removing <check-content> OVAL element for service_quota_nld_disabled.
Removing <check-content> OVAL element for service_rdisc_disabled.
Removing <check-content> OVAL element for service_rhnsd_disabled.
Removing <check-content> OVAL element for service_rhsmcertd_disabled.
Removing <check-content> OVAL element for service_saslauthd_disabled.
Removing <check-content> OVAL element for service_smartd_disabled.
Removing <check-content> OVAL element for service_sysstat_disabled.
Removing <check-content> OVAL element for service_crond_enabled.
Removing <check-content> OVAL element for service_atd_disabled.
Removing <check-content> OVAL element for package_openssh-server_installed.
Removing <check-content> OVAL element for service_sshd_enabled.
Removing <check-content> OVAL element for package_sssd_installed.
Removing <check-content> OVAL element for service_sssd_enabled.
Removing <check-content> OVAL element for xwindows_runlevel_setting.
Removing <check-content> OVAL element for service_avahi-daemon_disabled.
Removing <check-content> OVAL element for service_cups_disabled.
Removing <check-content> OVAL element for service_dhcpd_disabled.
Removing <check-content> OVAL element for service_chronyd_or_ntpd_enabled.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_remote_server.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_multiple_servers.
Removing <check-content> OVAL element for service_postfix_enabled.
Removing <check-content> OVAL element for postfix_network_listening_disabled.
Removing <check-content> OVAL element for service_nfslock_disabled.
Removing <check-content> OVAL element for service_rpcgssd_disabled.
Removing <check-content> OVAL element for service_rpcbind_disabled.
Removing <check-content> OVAL element for service_rpcidmapd_disabled.
Removing <check-content> OVAL element for service_nfs_disabled.
Removing <check-content> OVAL element for service_rpcsvcgssd_disabled.
Removing <check-content> OVAL element for service_named_disabled.
Removing <check-content> OVAL element for service_vsftpd_disabled.
Removing <check-content> OVAL element for service_httpd_disabled.
Removing <check-content> OVAL element for service_dovecot_disabled.
Removing <check-content> OVAL element for dovecot_enable_ssl.
Removing <check-content> OVAL element for dovecot_disable_plaintext_auth.
Removing <check-content> OVAL element for service_zebra_disabled.
Removing <check-content> OVAL element for package_quagga_removed.
Removing <check-content> OVAL element for service_smb_disabled.
Removing <check-content> OVAL element for package_samba_removed.
Removing <check-content> OVAL element for service_squid_disabled.
Removing <check-content> OVAL element for service_snmpd_disabled.
../../shared/transforms/relabelids.py xccdf-unlinked-ocilrefs.xml ssg
Removing <check-content> OVAL element for service_autofs_disabled.
Removing <check-content> OVAL element for service_debug-shell_disabled.
Removing <check-content> OVAL element for disable_ctrlaltdel_reboot.
Removing <check-content> OVAL element for smartcard_auth.
Removing <check-content> OVAL element for service_bluetooth_disabled.
Removing <check-content> OVAL element for service_firewalld_enabled.
Removing <check-content> OVAL element for service_rsyslog_enabled.
Removing <check-content> OVAL element for rsyslog_files_ownership.
Removing <check-content> OVAL element for rsyslog_files_groupownership.
Removing <check-content> OVAL element for rsyslog_files_permissions.
Removing <check-content> OVAL element for service_auditd_enabled.
Removing <check-content> OVAL element for service_xinetd_disabled.
Removing <check-content> OVAL element for package_tcp_wrappers_installed.
Removing <check-content> OVAL element for service_telnet_disabled.
Removing <check-content> OVAL element for service_rexec_disabled.
Removing <check-content> OVAL element for service_rsh_disabled.
Removing <check-content> OVAL element for service_rlogin_disabled.
Removing <check-content> OVAL element for service_ypbind_disabled.
Removing <check-content> OVAL element for service_tftp_disabled.
Removing <check-content> OVAL element for service_abrtd_disabled.
Removing <check-content> OVAL element for service_acpid_disabled.
Removing <check-content> OVAL element for service_certmonger_disabled.
Removing <check-content> OVAL element for service_cgconfig_disabled.
Removing <check-content> OVAL element for service_cgred_disabled.
Removing <check-content> OVAL element for service_cpupower_disabled.
Removing <check-content> OVAL element for service_irqbalance_enabled.
Removing <check-content> OVAL element for service_kdump_disabled.
Removing <check-content> OVAL element for service_mdmonitor_disabled.
Removing <check-content> OVAL element for service_messagebus_disabled.
Removing <check-content> OVAL element for service_netconsole_disabled.
Removing <check-content> OVAL element for service_ntpdate_disabled.
Removing <check-content> OVAL element for service_oddjobd_disabled.
Removing <check-content> OVAL element for service_portreserve_disabled.
Removing <check-content> OVAL element for service_psacct_enabled.
Removing <check-content> OVAL element for service_qpidd_disabled.
Removing <check-content> OVAL element for service_quota_nld_disabled.
Removing <check-content> OVAL element for service_rdisc_disabled.
Removing <check-content> OVAL element for service_rhnsd_disabled.
Removing <check-content> OVAL element for service_rhsmcertd_disabled.
Removing <check-content> OVAL element for service_saslauthd_disabled.
Removing <check-content> OVAL element for service_smartd_disabled.
Removing <check-content> OVAL element for service_sysstat_disabled.
Removing <check-content> OVAL element for service_crond_enabled.
Removing <check-content> OVAL element for service_atd_disabled.
Removing <check-content> OVAL element for package_openssh-server_installed.
Removing <check-content> OVAL element for service_sshd_enabled.
Removing <check-content> OVAL element for package_sssd_installed.
Removing <check-content> OVAL element for service_sssd_enabled.
Removing <check-content> OVAL element for xwindows_runlevel_setting.
Removing <check-content> OVAL element for service_avahi-daemon_disabled.
Removing <check-content> OVAL element for service_cups_disabled.
Removing <check-content> OVAL element for service_dhcpd_disabled.
Removing <check-content> OVAL element for service_chronyd_or_ntpd_enabled.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_remote_server.
Removing <check-content> OVAL element for chronyd_or_ntpd_specify_multiple_servers.
Removing <check-content> OVAL element for service_postfix_enabled.
Removing <check-content> OVAL element for postfix_network_listening_disabled.
Removing <check-content> OVAL element for service_nfslock_disabled.
Removing <check-content> OVAL element for service_rpcgssd_disabled.
Removing <check-content> OVAL element for service_rpcbind_disabled.
Removing <check-content> OVAL element for service_rpcidmapd_disabled.
Removing <check-content> OVAL element for service_nfs_disabled.
Removing <check-content> OVAL element for service_rpcsvcgssd_disabled.
Removing <check-content> OVAL element for service_named_disabled.
Removing <check-content> OVAL element for service_vsftpd_disabled.
Removing <check-content> OVAL element for service_httpd_disabled.
Removing <check-content> OVAL element for service_dovecot_disabled.
Removing <check-content> OVAL element for dovecot_enable_ssl.
Removing <check-content> OVAL element for dovecot_disable_plaintext_auth.
Removing <check-content> OVAL element for service_zebra_disabled.
Removing <check-content> OVAL element for package_quagga_removed.
Removing <check-content> OVAL element for service_smb_disabled.
Removing <check-content> OVAL element for package_samba_removed.
Removing <check-content> OVAL element for service_squid_disabled.
Removing <check-content> OVAL element for service_snmpd_disabled.
xsltproc -o output/ssg-rhel7-oval.xml transforms/oval-fix-test-attestation-urls.xslt output/ssg-rhel7-oval.xml
xsltproc /usr/share/openscap/xsl/xccdf_1.1_remove_dangling_sub.xsl output/ssg-rhel7-xccdf.xml \
> output/ssg-rhel7-xccdf-nodangles.xml
xsltproc --stringparam reverse_DNS org.ssgproject.content /usr/share/openscap/xsl/xccdf_1.1_to_1.2.xsl \
output/ssg-rhel7-xccdf-nodangles.xml > output/ssg-rhel7-xccdf-1.2.xml
sed -i '/idref="dangling reference to /d' output/ssg-rhel7-xccdf-1.2.xml
sed -i 's/style="SCAP_1.1"/style="SCAP_1.2"/' output/ssg-rhel7-xccdf-1.2.xml
oscap ds sds-compose output/ssg-rhel7-xccdf-1.2.xml output/ssg-rhel7-ds.xml
OpenSCAP Error: Unknown document type: 'ssg-rhel7-ocil.xml' [oscapxml.c:630]
sed -i 's/schematron-version="[0-9].[0-9]"/schematron-version="1.2"/' output/ssg-rhel7-ds.xml
oscap ds sds-add output/ssg-rhel7-cpe-dictionary.xml output/ssg-rhel7-ds.xml
oscap ds sds-add output/ssg-rhel7-oval.xml output/ssg-rhel7-ds.xml
# Fixes https://github.com/OpenSCAP/scap-security-guide/issues/1100
# Fixes https://github.com/OpenSCAP/scap-security-guide/issues/1101
../../shared/transforms/datastream_move_ocil_to_ds_checks.py output/ssg-rhel7-ds.xml output/ssg-rhel7-ds.xml
../../shared/utils/enable-derivatives.py --enable-centos -i output/ssg-rhel7-xccdf.xml -o output/ssg-centos7-xccdf.xml
../../shared/utils/enable-derivatives.py --enable-centos -i output/ssg-rhel7-ds.xml -o output/ssg-centos7-ds.xml
../../shared/utils/enable-derivatives.py --enable-sl -i output/ssg-rhel7-xccdf.xml -o output/ssg-sl7-xccdf.xml
../../shared/utils/enable-derivatives.py --enable-sl -i output/ssg-rhel7-ds.xml -o output/ssg-sl7-ds.xml
[root@localhost 7]# echo $?
0
[root@localhost 7]#
```
Will investigate further. | infrastructure | running make content for rhel fails after introduction of shared transforms shared xslt currently it s not possible make content from rhel folder to succeed behavior is identical on both openscap x and openscap x make content xsltproc o output guide xml shared transforms includelogo xslt input guide xml xsltproc param withtest stringparam shared rp root scap security guide shared o output shorthand xml input guide xslt output guide xml xmllint format output output shorthand xml output shorthand xml create output contributors xml file holding like root element and names of the individual ssg contributor s within element s echo output contributors xml sed n e s p shared contributors md output contributors xml echo output contributors xml xsltproc stringparam ssg version o output xccdf unlinked unresolved xml transforms xslt output shorthand xml warning failed to load external entity shared output contributors xml runtime error file shared transforms shared xslt line element value of variable cisuri has not been declared xmlxpathcompiledeval evaluation failed runtime error file shared transforms shared xslt line element value of xpath evaluation returned no result make error this is because after move of xslt to shared transforms folder the path to out contributors xml changed strangely enough rhel s make content still works properly make clean rm f output xml output html output xhtml output pdf output spec output tar output gz output ini output csv rm rf dist content dist guide rm rf build make content xsltproc o output guide xml shared transforms includelogo xslt input guide xml xsltproc param withtest stringparam shared rp root scap security guide shared o output shorthand xml input guide xslt output guide xml xmllint format output output shorthand xml output shorthand xml create output contributors xml file holding like root element and names of the individual ssg contributor s within element s echo output contributors xml sed n e s p shared contributors md output contributors xml echo output contributors xml xsltproc stringparam ssg version o output xccdf unlinked unresolved xml transforms xslt output shorthand xml warning failed to load external entity shared output contributors xml oscap xccdf resolve o output xccdf unlinked empty groups xml output xccdf unlinked unresolved xml shared utils unselect empty xccdf groups py input output xccdf unlinked empty groups xml output output xccdf unlinked empty groups unselected xml unselected empty groups in standard unselected empty groups in pci dss unselected empty groups in unselected empty groups in rht ccp unselected empty groups in common unselected empty groups in stig workstation upstream unselected empty groups in stig server gui upstream unselected empty groups in stig server upstream unselected empty groups in ospp server unselected empty groups in nist cl il al unselected empty groups in cjis server oscap xccdf resolve o output xccdf unlinked resolved xml output xccdf unlinked empty groups unselected xml xsltproc o output ocil unlinked xml shared transforms xccdf create ocil xslt output xccdf unlinked resolved xml xmllint format output output unlinked ocil xml output ocil unlinked xml xsltproc stringparam product o output xccdf unlinked ocilrefs xml shared transforms xccdf xslt output xccdf unlinked resolved xml make intermediate build remediations directory to hold final list of remediation scripts for mkdir p build remediations search shared remediations bash and input remediations directories to find all product specific remediation scripts which are regular files not symlinks merge the final list into build remediations directory find shared remediations bash input remediations bash maxdepth type f name sh exec cp build remediations shared transforms combineremediations py build remediations output bash remediations xml notification removed the accounts passwords pam unix remember sh remediation script from merging as the platform identifier in the script is missing notification merged remediation scripts into xml document xsltproc stringparam remediations root scap security guide rhel output bash remediations xml o output xccdf unlinked withremediations xml shared transforms xccdf addremediations xslt output xccdf unlinked ocilrefs xml xmllint format output output xccdf unlinked withremediations xml output xccdf unlinked withremediations xml make intermediate build oval directory to hold final list of oval checks for mkdir p build oval search shared oval and input oval directories to find all product specific oval checks which are regular files not symlinks merge the final list into build oval directory find shared oval input oval maxdepth type f name xml exec cp build oval if openscap on the system supports oval language version include also oval checks into final list of oval checks xmlwf build oval xml shared transforms combineovals py shared config build oval output unlinked oval xml notification merged oval checks into oval document notification this id is used more than once and should represent equivalent elements sysctl net tcp syncookies value notification this id is used more than once and should represent equivalent elements sysctl net conf default secure redirects value notification this id is used more than once and should represent equivalent elements login banner text notification this id is used more than once and should represent equivalent elements test unix family notification this id is used more than once and should represent equivalent elements state unix family notification this id is used more than once and should represent equivalent elements obj unix family notification this id is used more than once and should represent equivalent elements test auditd conf log group root notification this id is used more than once and should represent equivalent elements object auditd conf log group root notification this id is used more than once and should represent equivalent elements test bootloader recovery disabled notification this id is used more than once and should represent equivalent elements object bootloader disable recovery argument notification this id is used more than once and should represent equivalent elements state bootloader disable recovery argument notification this id is used more than once and should represent equivalent elements sysctl net conf all forwarding value notification this id is used more than once and should represent equivalent elements test unix family notification this id is used more than once and should represent equivalent elements state unix family notification this id is used more than once and should represent equivalent elements obj unix family notification this id is used more than once and should represent equivalent elements sysctl net conf all accept source route value notification this id is used more than once and should represent equivalent elements test removable partition doesnt exist notification this id is used more than once and should represent equivalent elements object removable partition doesnt exist notification this id is used more than once and should represent equivalent elements test var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements object var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements state var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements variable cd dvd drive alternative names notification this id is used more than once and should represent equivalent elements variable cd dvd drive regex pattern notification this id is used more than once and should represent equivalent elements variable not cd dvd drive regex pattern notification this id is used more than once and should represent equivalent elements var removable partition notification this id is used more than once and should represent equivalent elements sysctl net conf all accept redirects value notification this id is used more than once and should represent equivalent elements sysctl net conf default log martians value notification this id is used more than once and should represent equivalent elements sysctl net conf default accept ra value notification this id is used more than once and should represent equivalent elements sysctl net conf default accept redirects value notification this id is used more than once and should represent equivalent elements sysctl net conf all accept source route value notification this id is used more than once and should represent equivalent elements test unix family notification this id is used more than once and should represent equivalent elements state unix family notification this id is used more than once and should represent equivalent elements obj unix family notification this id is used more than once and should represent equivalent elements sysctl net conf all secure redirects value notification this id is used more than once and should represent equivalent elements sysctl net conf default accept source route value notification this id is used more than once and should represent equivalent elements test removable partition doesnt exist notification this id is used more than once and should represent equivalent elements object removable partition doesnt exist notification this id is used more than once and should represent equivalent elements test var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements object var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements state var removable partition is cd dvd drive notification this id is used more than once and should represent equivalent elements variable cd dvd drive alternative names notification this id is used more than once and should represent equivalent elements variable cd dvd drive regex pattern notification this id is used more than once and should represent equivalent elements variable not cd dvd drive regex pattern notification this id is used more than once and should represent equivalent elements var removable partition notification this id is used more than once and should represent equivalent elements sysctl net icmp ignore bogus error responses value notification this id is used more than once and should represent equivalent elements var accounts user umask notification this id is used more than once and should represent equivalent elements var first digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var second digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var third digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var accounts user umask umask as number notification this id is used more than once and should represent equivalent elements sysctl net conf all accept redirects value notification this id is used more than once and should represent equivalent elements obj package gpg pubkey notification this id is used more than once and should represent equivalent elements obj package gpg pubkey notification this id is used more than once and should represent equivalent elements sysctl net conf default rp filter value notification this id is used more than once and should represent equivalent elements var accounts user umask notification this id is used more than once and should represent equivalent elements var first digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var second digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var third digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var accounts user umask umask as number notification this id is used more than once and should represent equivalent elements var accounts user umask notification this id is used more than once and should represent equivalent elements var first digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var second digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var third digit of umask from var accounts user umask notification this id is used more than once and should represent equivalent elements var accounts user umask umask as number notification this id is used more than once and should represent equivalent elements sysctl net conf all log martians value notification this id is used more than once and should represent equivalent elements test unix family notification this id is used more than once and should represent equivalent elements obj unix family notification this id is used more than once and should represent equivalent elements state unix family notification this id is used more than once and should represent equivalent elements sysctl net conf default accept redirects value notification this id is used more than once and should represent equivalent elements sysctl net conf all accept ra value notification this id is used more than once and should represent equivalent elements test unix family notification this id is used more than once and should represent equivalent elements state unix family notification this id is used more than once and should represent equivalent elements obj unix family notification this id is used more than once and should represent equivalent elements sysctl net conf all rp filter value notification this id is used more than once and should represent equivalent elements sysctl net conf default accept source route value notification this id is used more than once and should represent equivalent elements sysctl net icmp echo ignore broadcasts value xmllint format output output unlinked oval xml output unlinked oval xml cp output xccdf unlinked final xml output unlinked xccdf xml xsltproc o output unlinked xccdf guide xml transforms xccdf removeaux xslt output unlinked xccdf xml shared transforms cpe generate py output unlinked oval xml input oval platform cpe dictionary xml ssg shared transforms relabelids py unlinked xccdf xml ssg removing oval element for service autofs disabled removing oval element for service debug shell disabled removing oval element for disable ctrlaltdel reboot removing oval element for smartcard auth removing oval element for service bluetooth disabled removing oval element for service firewalld enabled removing oval element for service rsyslog enabled removing oval element for rsyslog files ownership removing oval element for rsyslog files groupownership removing oval element for rsyslog files permissions removing oval element for service auditd enabled removing oval element for service xinetd disabled removing oval element for package tcp wrappers installed removing oval element for service telnet disabled removing oval element for service rexec disabled removing oval element for service rsh disabled removing oval element for service rlogin disabled removing oval element for service ypbind disabled removing oval element for service tftp disabled removing oval element for service abrtd disabled removing oval element for service acpid disabled removing oval element for service certmonger disabled removing oval element for service cgconfig disabled removing oval element for service cgred disabled removing oval element for service cpupower disabled removing oval element for service irqbalance enabled removing oval element for service kdump disabled removing oval element for service mdmonitor disabled removing oval element for service messagebus disabled removing oval element for service netconsole disabled removing oval element for service ntpdate disabled removing oval element for service oddjobd disabled removing oval element for service portreserve disabled removing oval element for service psacct enabled removing oval element for service qpidd disabled removing oval element for service quota nld disabled removing oval element for service rdisc disabled removing oval element for service rhnsd disabled removing oval element for service rhsmcertd disabled removing oval element for service saslauthd disabled removing oval element for service smartd disabled removing oval element for service sysstat disabled removing oval element for service crond enabled removing oval element for service atd disabled removing oval element for package openssh server installed removing oval element for service sshd enabled removing oval element for package sssd installed removing oval element for service sssd enabled removing oval element for xwindows runlevel setting removing oval element for service avahi daemon disabled removing oval element for service cups disabled removing oval element for service dhcpd disabled removing oval element for service chronyd or ntpd enabled removing oval element for chronyd or ntpd specify remote server removing oval element for chronyd or ntpd specify multiple servers removing oval element for service postfix enabled removing oval element for postfix network listening disabled removing oval element for service nfslock disabled removing oval element for service rpcgssd disabled removing oval element for service rpcbind disabled removing oval element for service rpcidmapd disabled removing oval element for service nfs disabled removing oval element for service rpcsvcgssd disabled removing oval element for service named disabled removing oval element for service vsftpd disabled removing oval element for service httpd disabled removing oval element for service dovecot disabled removing oval element for dovecot enable ssl removing oval element for dovecot disable plaintext auth removing oval element for service zebra disabled removing oval element for package quagga removed removing oval element for service smb disabled removing oval element for package samba removed removing oval element for service squid disabled removing oval element for service snmpd disabled shared transforms relabelids py xccdf unlinked ocilrefs xml ssg removing oval element for service autofs disabled removing oval element for service debug shell disabled removing oval element for disable ctrlaltdel reboot removing oval element for smartcard auth removing oval element for service bluetooth disabled removing oval element for service firewalld enabled removing oval element for service rsyslog enabled removing oval element for rsyslog files ownership removing oval element for rsyslog files groupownership removing oval element for rsyslog files permissions removing oval element for service auditd enabled removing oval element for service xinetd disabled removing oval element for package tcp wrappers installed removing oval element for service telnet disabled removing oval element for service rexec disabled removing oval element for service rsh disabled removing oval element for service rlogin disabled removing oval element for service ypbind disabled removing oval element for service tftp disabled removing oval element for service abrtd disabled removing oval element for service acpid disabled removing oval element for service certmonger disabled removing oval element for service cgconfig disabled removing oval element for service cgred disabled removing oval element for service cpupower disabled removing oval element for service irqbalance enabled removing oval element for service kdump disabled removing oval element for service mdmonitor disabled removing oval element for service messagebus disabled removing oval element for service netconsole disabled removing oval element for service ntpdate disabled removing oval element for service oddjobd disabled removing oval element for service portreserve disabled removing oval element for service psacct enabled removing oval element for service qpidd disabled removing oval element for service quota nld disabled removing oval element for service rdisc disabled removing oval element for service rhnsd disabled removing oval element for service rhsmcertd disabled removing oval element for service saslauthd disabled removing oval element for service smartd disabled removing oval element for service sysstat disabled removing oval element for service crond enabled removing oval element for service atd disabled removing oval element for package openssh server installed removing oval element for service sshd enabled removing oval element for package sssd installed removing oval element for service sssd enabled removing oval element for xwindows runlevel setting removing oval element for service avahi daemon disabled removing oval element for service cups disabled removing oval element for service dhcpd disabled removing oval element for service chronyd or ntpd enabled removing oval element for chronyd or ntpd specify remote server removing oval element for chronyd or ntpd specify multiple servers removing oval element for service postfix enabled removing oval element for postfix network listening disabled removing oval element for service nfslock disabled removing oval element for service rpcgssd disabled removing oval element for service rpcbind disabled removing oval element for service rpcidmapd disabled removing oval element for service nfs disabled removing oval element for service rpcsvcgssd disabled removing oval element for service named disabled removing oval element for service vsftpd disabled removing oval element for service httpd disabled removing oval element for service dovecot disabled removing oval element for dovecot enable ssl removing oval element for dovecot disable plaintext auth removing oval element for service zebra disabled removing oval element for package quagga removed removing oval element for service smb disabled removing oval element for package samba removed removing oval element for service squid disabled removing oval element for service snmpd disabled xsltproc o output ssg oval xml transforms oval fix test attestation urls xslt output ssg oval xml xsltproc usr share openscap xsl xccdf remove dangling sub xsl output ssg xccdf xml output ssg xccdf nodangles xml xsltproc stringparam reverse dns org ssgproject content usr share openscap xsl xccdf to xsl output ssg xccdf nodangles xml output ssg xccdf xml sed i idref dangling reference to d output ssg xccdf xml sed i s style scap style scap output ssg xccdf xml oscap ds sds compose output ssg xccdf xml output ssg ds xml openscap error unknown document type ssg ocil xml sed i s schematron version schematron version output ssg ds xml oscap ds sds add output ssg cpe dictionary xml output ssg ds xml oscap ds sds add output ssg oval xml output ssg ds xml fixes fixes shared transforms datastream move ocil to ds checks py output ssg ds xml output ssg ds xml shared utils enable derivatives py enable centos i output ssg xccdf xml o output ssg xccdf xml shared utils enable derivatives py enable centos i output ssg ds xml o output ssg ds xml shared utils enable derivatives py enable sl i output ssg xccdf xml o output ssg xccdf xml shared utils enable derivatives py enable sl i output ssg ds xml o output ssg ds xml echo will investigate further | 1 |
23,034 | 3,988,624,432 | IssuesEvent | 2016-05-09 10:38:34 | radare/radare2 | https://api.github.com/repos/radare/radare2 | closed | zignatures: loaded but not applied | bug test-required zignatures | test.c:
```
#include <stdio.h>
int g(int a) {
return a + 3;
}
int h(char *p) {
return atoi(p);
}
int f(int a, char *p) {
return g(a) + h(p);
}
int main(int argc, char **argv) {
f (argc, argv[1]);
return 0;
}
```
Compile it once with symbols (test) and once without (test_strip)
```
$ r2 -A test
> zg testzg /tmp/test.r2
> q
$ r2 test_strip
> . /tmp/test.r2
> z
Loaded 6 signatures
6 byte signatures
0 head signatures
0 func signatures
Found 0 matches
```
It seems like nothing was found. The content of /tmp/test.r2 is:
```
# Signatures
zn test
zb sym.func.100000f40 554889e54883ec20c745fc00000000897df8488975f08b7df8488b75f0488b7608e8aaffffff31ff8945ec89f84883c4205dc3
zb sym.func.100000f10 554889e54883ec20897dfc488975f08b7dfce8a9ffffff488b7df08945ece8bdffffff8b4dec01c189c84883c4205dc3
zb sym.func.100000ed0 554889e5897dfc8b7dfc81c70300000089f85dc3
zb sym.func.100000ef0 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
zb sym.imp.atoi ff2596000000
zb loc.100000f7c 4c8d1d850000004153ff2575000000
zn-
```
But if I search for any of this pattern, I can find it:
```
$ r2 ./test_strip
> /x 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
Searching 29 bytes in [0x100000ed0-0x100001018]
hits: 1
0x100000ef0 hit0_0 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
``` | 1.0 | zignatures: loaded but not applied - test.c:
```
#include <stdio.h>
int g(int a) {
return a + 3;
}
int h(char *p) {
return atoi(p);
}
int f(int a, char *p) {
return g(a) + h(p);
}
int main(int argc, char **argv) {
f (argc, argv[1]);
return 0;
}
```
Compile it once with symbols (test) and once without (test_strip)
```
$ r2 -A test
> zg testzg /tmp/test.r2
> q
$ r2 test_strip
> . /tmp/test.r2
> z
Loaded 6 signatures
6 byte signatures
0 head signatures
0 func signatures
Found 0 matches
```
It seems like nothing was found. The content of /tmp/test.r2 is:
```
# Signatures
zn test
zb sym.func.100000f40 554889e54883ec20c745fc00000000897df8488975f08b7df8488b75f0488b7608e8aaffffff31ff8945ec89f84883c4205dc3
zb sym.func.100000f10 554889e54883ec20897dfc488975f08b7dfce8a9ffffff488b7df08945ece8bdffffff8b4dec01c189c84883c4205dc3
zb sym.func.100000ed0 554889e5897dfc8b7dfc81c70300000089f85dc3
zb sym.func.100000ef0 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
zb sym.imp.atoi ff2596000000
zb loc.100000f7c 4c8d1d850000004153ff2575000000
zn-
```
But if I search for any of this pattern, I can find it:
```
$ r2 ./test_strip
> /x 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
Searching 29 bytes in [0x100000ed0-0x100001018]
hits: 1
0x100000ef0 hit0_0 554889e54883ec1048897df8488b7df8b000e86d0000004883c4105dc3
``` | non_infrastructure | zignatures loaded but not applied test c include int g int a return a int h char p return atoi p int f int a char p return g a h p int main int argc char argv f argc argv return compile it once with symbols test and once without test strip a test zg testzg tmp test q test strip tmp test z loaded signatures byte signatures head signatures func signatures found matches it seems like nothing was found the content of tmp test is signatures zn test zb sym func zb sym func zb sym func zb sym func zb sym imp atoi zb loc zn but if i search for any of this pattern i can find it test strip x searching bytes in hits | 0 |
169,468 | 6,402,623,233 | IssuesEvent | 2017-08-06 11:27:43 | OperationCode/operationcode_frontend | https://api.github.com/repos/OperationCode/operationcode_frontend | opened | Move MOOC entries to the backend | Priority: Medium Status: Available Type: Feature | <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
MOOCs under our CodeSchools component are currently hard coded which is confusing due to all other data coming from our API.
## What should your feature do?
Make sure we're able to serve MOOCs from the backend.
Remove hard coded values from https://github.com/OperationCode/operationcode_frontend/blob/master/src/scenes/home/codeSchools/onlineSchools/onlineSchools.js and replace with values from our API. You'll need to create a ticket in our backend repo to add the correct information. | 1.0 | Move MOOC entries to the backend - <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
MOOCs under our CodeSchools component are currently hard coded which is confusing due to all other data coming from our API.
## What should your feature do?
Make sure we're able to serve MOOCs from the backend.
Remove hard coded values from https://github.com/OperationCode/operationcode_frontend/blob/master/src/scenes/home/codeSchools/onlineSchools/onlineSchools.js and replace with values from our API. You'll need to create a ticket in our backend repo to add the correct information. | non_infrastructure | move mooc entries to the backend feature why is this feature being added moocs under our codeschools component are currently hard coded which is confusing due to all other data coming from our api what should your feature do make sure we re able to serve moocs from the backend remove hard coded values from and replace with values from our api you ll need to create a ticket in our backend repo to add the correct information | 0 |
277,211 | 21,032,718,001 | IssuesEvent | 2022-03-31 03:22:31 | jump-dev/MathOptInterface.jl | https://api.github.com/repos/jump-dev/MathOptInterface.jl | closed | Discover fixed excluded tests | Type: Documentation Submodule: Tests | It would be useful to have a keyword argument to `runtests` that is unset by default but when set it would run excluded tests as well and throw warning when excluded tests now work. | 1.0 | Discover fixed excluded tests - It would be useful to have a keyword argument to `runtests` that is unset by default but when set it would run excluded tests as well and throw warning when excluded tests now work. | non_infrastructure | discover fixed excluded tests it would be useful to have a keyword argument to runtests that is unset by default but when set it would run excluded tests as well and throw warning when excluded tests now work | 0 |
26,956 | 2,689,185,180 | IssuesEvent | 2015-03-31 08:27:09 | 52North/SOS | https://api.github.com/repos/52North/SOS | opened | Support for CompositePhenomenon | enhancement medium priority | Currently, the 52°North SOS 4.x only supports the insertion of observations with a simple/single observedProperty in a SweDataArray obsrevation with SplitDataArray extension.
But in the 52°North SOS 3.x and O&M 1.0.0 the CompositePhenomenon was supported which allows to insert multiple observedProperties in a single InsertObservation request.
There are two options to support this via the 52°North SOS 4.x:
- Implement SOS 1.0.0 InsertObservation support
- Enhance the SOS 2.0 InsertObservation implementation for SweDataArray to support multiple observedProperties which can be split into single observations. | 1.0 | Support for CompositePhenomenon - Currently, the 52°North SOS 4.x only supports the insertion of observations with a simple/single observedProperty in a SweDataArray obsrevation with SplitDataArray extension.
But in the 52°North SOS 3.x and O&M 1.0.0 the CompositePhenomenon was supported which allows to insert multiple observedProperties in a single InsertObservation request.
There are two options to support this via the 52°North SOS 4.x:
- Implement SOS 1.0.0 InsertObservation support
- Enhance the SOS 2.0 InsertObservation implementation for SweDataArray to support multiple observedProperties which can be split into single observations. | non_infrastructure | support for compositephenomenon currently the °north sos x only supports the insertion of observations with a simple single observedproperty in a swedataarray obsrevation with splitdataarray extension but in the °north sos x and o m the compositephenomenon was supported which allows to insert multiple observedproperties in a single insertobservation request there are two options to support this via the °north sos x implement sos insertobservation support enhance the sos insertobservation implementation for swedataarray to support multiple observedproperties which can be split into single observations | 0 |
762,023 | 26,706,467,490 | IssuesEvent | 2023-01-27 18:38:54 | Alluxio/alluxio | https://api.github.com/repos/Alluxio/alluxio | closed | Fuse IO stressbench hardening | priority-medium type-feature stale | **Is your feature request related to a problem? Please describe.**
[Fuse IO stressbench](https://docs.alluxio.io/os/user/stable/en/operation/StressBench.html#fuse-io-stress-bench) according to users including @maobaolong has some limitations
it requires
- can only test data written by Fuse IO stressbench, cannot test existing files
- Unstable, no retry logic for read/write/ls, if one error thrown, the whole test will failed.
- Require the cluster setup to be fixed. In cluster testing, each worker node needs to have one worker, one Fuse and one job worker. If one process down, the test will not be passed.
**Describe the solution you'd like**
Hardening the Fuse IO stressbench to make it more user-friendly
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Urgency**
Explain why the feature is important
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Fuse IO stressbench hardening - **Is your feature request related to a problem? Please describe.**
[Fuse IO stressbench](https://docs.alluxio.io/os/user/stable/en/operation/StressBench.html#fuse-io-stress-bench) according to users including @maobaolong has some limitations
it requires
- can only test data written by Fuse IO stressbench, cannot test existing files
- Unstable, no retry logic for read/write/ls, if one error thrown, the whole test will failed.
- Require the cluster setup to be fixed. In cluster testing, each worker node needs to have one worker, one Fuse and one job worker. If one process down, the test will not be passed.
**Describe the solution you'd like**
Hardening the Fuse IO stressbench to make it more user-friendly
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Urgency**
Explain why the feature is important
**Additional context**
Add any other context or screenshots about the feature request here.
| non_infrastructure | fuse io stressbench hardening is your feature request related to a problem please describe according to users including maobaolong has some limitations it requires can only test data written by fuse io stressbench cannot test existing files unstable no retry logic for read write ls if one error thrown the whole test will failed require the cluster setup to be fixed in cluster testing each worker node needs to have one worker one fuse and one job worker if one process down the test will not be passed describe the solution you d like hardening the fuse io stressbench to make it more user friendly describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered urgency explain why the feature is important additional context add any other context or screenshots about the feature request here | 0 |
27,041 | 21,051,901,025 | IssuesEvent | 2022-03-31 21:20:12 | dotnet/sdk | https://api.github.com/repos/dotnet/sdk | closed | Build failed: Validate-DotNet/main real signed, .NET 6, validation-sdk-.NET 6, sdk #sdk-102308-.NET6 | Area-Infrastructure untriaged | Build [#sdk-102308-.NET6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1299990) failed
## :x: : internal / Validate-DotNet failed
### Summary
**Finished** - Tue, 17 Aug 2021 23:48:43 GMT
**Duration** - 42 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Signing Ring
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1299990/logs/101) - PowerShell exited with code '1'.
#### Source Code Validation
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1299990/logs/86) - (NETCORE_ENGINEERING_TELEMETRY=Sdl) Last command failed with exit code 1.
### Changes
- [210a9396](https://dev.azure.com/dnceng/internal/_git/dotnet-release/commit/210a93961e47c09ad104bbaad364d01b8f7ff769) - Michelle McDaniel - Merged PR 17444: Push staged release to blob storage
| 1.0 | Build failed: Validate-DotNet/main real signed, .NET 6, validation-sdk-.NET 6, sdk #sdk-102308-.NET6 - Build [#sdk-102308-.NET6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1299990) failed
## :x: : internal / Validate-DotNet failed
### Summary
**Finished** - Tue, 17 Aug 2021 23:48:43 GMT
**Duration** - 42 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Signing Ring
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1299990/logs/101) - PowerShell exited with code '1'.
#### Source Code Validation
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1299990/logs/86) - (NETCORE_ENGINEERING_TELEMETRY=Sdl) Last command failed with exit code 1.
### Changes
- [210a9396](https://dev.azure.com/dnceng/internal/_git/dotnet-release/commit/210a93961e47c09ad104bbaad364d01b8f7ff769) - Michelle McDaniel - Merged PR 17444: Push staged release to blob storage
| infrastructure | build failed validate dotnet main real signed net validation sdk net sdk sdk build failed x internal validate dotnet failed summary finished tue aug gmt duration minutes requested for dotnet bot reason manual details signing ring x powershell exited with code source code validation x netcore engineering telemetry sdl last command failed with exit code changes michelle mcdaniel merged pr push staged release to blob storage | 1 |
159,128 | 20,036,649,411 | IssuesEvent | 2022-02-02 12:38:32 | kapseliboi/webuild | https://api.github.com/repos/kapseliboi/webuild | opened | WS-2019-0032 (High) detected in multiple libraries | security vulnerability | ## WS-2019-0032 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>js-yaml-3.12.0.tgz</b>, <b>js-yaml-3.5.5.tgz</b>, <b>js-yaml-3.4.6.tgz</b></p></summary>
<p>
<details><summary><b>js-yaml-3.12.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.12.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.12.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- coveralls-3.0.2.tgz (Root Library)
- :x: **js-yaml-3.12.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.5.5.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/grunt-jsbeautifier/node_modules/js-yaml/package.json,/node_modules/grunt/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-1.0.3.tgz (Root Library)
- :x: **js-yaml-3.5.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.4.6.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.4.6.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.4.6.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jscs/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jscs-3.0.1.tgz (Root Library)
- jscs-3.0.7.tgz
- :x: **js-yaml-3.4.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution (js-yaml): 3.13.0</p>
<p>Direct dependency fix Resolution (coveralls): 3.0.3</p><p>Fix Resolution (js-yaml): 3.13.0</p>
<p>Direct dependency fix Resolution (grunt): 1.0.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0032 (High) detected in multiple libraries - ## WS-2019-0032 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>js-yaml-3.12.0.tgz</b>, <b>js-yaml-3.5.5.tgz</b>, <b>js-yaml-3.4.6.tgz</b></p></summary>
<p>
<details><summary><b>js-yaml-3.12.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.12.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.12.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- coveralls-3.0.2.tgz (Root Library)
- :x: **js-yaml-3.12.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.5.5.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/grunt-jsbeautifier/node_modules/js-yaml/package.json,/node_modules/grunt/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-1.0.3.tgz (Root Library)
- :x: **js-yaml-3.5.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.4.6.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.4.6.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.4.6.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jscs/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jscs-3.0.1.tgz (Root Library)
- jscs-3.0.7.tgz
- :x: **js-yaml-3.4.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution (js-yaml): 3.13.0</p>
<p>Direct dependency fix Resolution (coveralls): 3.0.3</p><p>Fix Resolution (js-yaml): 3.13.0</p>
<p>Direct dependency fix Resolution (grunt): 1.0.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | ws high detected in multiple libraries ws high severity vulnerability vulnerable libraries js yaml tgz js yaml tgz js yaml tgz js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules js yaml package json dependency hierarchy coveralls tgz root library x js yaml tgz vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules grunt jsbeautifier node modules js yaml package json node modules grunt node modules js yaml package json dependency hierarchy grunt tgz root library x js yaml tgz vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules jscs node modules js yaml package json dependency hierarchy grunt jscs tgz root library jscs tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution js yaml direct dependency fix resolution coveralls fix resolution js yaml direct dependency fix resolution grunt step up your open source security game with whitesource | 0 |
33,521 | 27,544,439,655 | IssuesEvent | 2023-03-07 10:45:13 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | opened | Publish archive with ideType=IU on each commit into main branch | ctg-enhancement comp-infrastructure | **Description**
Publish ultimate version of the plugin on each commit into main branch.
**Expected behavior**
Two versions of plugin will be published for each commit into main branch:
- for IC
- for IU
**Environment**
Run workflow from main
**Context**
Related task:
- #1836 | 1.0 | Publish archive with ideType=IU on each commit into main branch - **Description**
Publish ultimate version of the plugin on each commit into main branch.
**Expected behavior**
Two versions of plugin will be published for each commit into main branch:
- for IC
- for IU
**Environment**
Run workflow from main
**Context**
Related task:
- #1836 | infrastructure | publish archive with idetype iu on each commit into main branch description publish ultimate version of the plugin on each commit into main branch expected behavior two versions of plugin will be published for each commit into main branch for ic for iu environment run workflow from main context related task | 1 |
12,457 | 9,792,466,538 | IssuesEvent | 2019-06-10 17:28:58 | forseti-security/forseti-security | https://api.github.com/repos/forseti-security/forseti-security | reopened | Signed releases | issue-review: future-milestone module: infrastructure priority: p1 triaged: yes | The currently published releases are just GitHub providing bundles for the tags. However, this means no hash is published allowing for more easily validating the bundle and means the releases aren't signed.
It would be great if you could leverage GitHub's support for signed releases so anyone has the opportunity to validate the release before executing any of the associated code. As far as I'm aware Travis can be leveraged to do this, by encrypting (`travis encrypt`) a GPG signing key into the repository (or some other location Travis can fetch from) and using that to sign the release. This would require [packaging an actual release](https://docs.travis-ci.com/user/deployment/releases/) though, which Travis supports, and not purely relying on the tag to bundle trick GitHub applies. Letting Travis decrypt and use that key could be a concern though but then it seems this would have to be a manual step. | 1.0 | Signed releases - The currently published releases are just GitHub providing bundles for the tags. However, this means no hash is published allowing for more easily validating the bundle and means the releases aren't signed.
It would be great if you could leverage GitHub's support for signed releases so anyone has the opportunity to validate the release before executing any of the associated code. As far as I'm aware Travis can be leveraged to do this, by encrypting (`travis encrypt`) a GPG signing key into the repository (or some other location Travis can fetch from) and using that to sign the release. This would require [packaging an actual release](https://docs.travis-ci.com/user/deployment/releases/) though, which Travis supports, and not purely relying on the tag to bundle trick GitHub applies. Letting Travis decrypt and use that key could be a concern though but then it seems this would have to be a manual step. | infrastructure | signed releases the currently published releases are just github providing bundles for the tags however this means no hash is published allowing for more easily validating the bundle and means the releases aren t signed it would be great if you could leverage github s support for signed releases so anyone has the opportunity to validate the release before executing any of the associated code as far as i m aware travis can be leveraged to do this by encrypting travis encrypt a gpg signing key into the repository or some other location travis can fetch from and using that to sign the release this would require though which travis supports and not purely relying on the tag to bundle trick github applies letting travis decrypt and use that key could be a concern though but then it seems this would have to be a manual step | 1 |
19,927 | 13,541,720,236 | IssuesEvent | 2020-09-16 16:15:30 | coq/coq | https://api.github.com/repos/coq/coq | reopened | Could we make coq-ltac2 a dummy package in 8.11 instead of requiring Coq < 8.11? | kind: infrastructure part: ltac2 resolved: moved | #### Description of the problem
I'm trying to support 8.9 through 8.11 in the same project. I use Ltac2, so I have a dependency on the coq-ltac2 package. But currently that package requires Coq < 8.11.
Could the coq-ltac2 be marked as compatible with 8.11 and made to install a dummy package there? Or is there a simpler solution that I overlooked?
cc @ppedrot | 1.0 | Could we make coq-ltac2 a dummy package in 8.11 instead of requiring Coq < 8.11? - #### Description of the problem
I'm trying to support 8.9 through 8.11 in the same project. I use Ltac2, so I have a dependency on the coq-ltac2 package. But currently that package requires Coq < 8.11.
Could the coq-ltac2 be marked as compatible with 8.11 and made to install a dummy package there? Or is there a simpler solution that I overlooked?
cc @ppedrot | infrastructure | could we make coq a dummy package in instead of requiring coq description of the problem i m trying to support through in the same project i use so i have a dependency on the coq package but currently that package requires coq could the coq be marked as compatible with and made to install a dummy package there or is there a simpler solution that i overlooked cc ppedrot | 1 |
329,998 | 24,242,402,985 | IssuesEvent | 2022-09-27 07:53:21 | Ashleykkkb/cuddly-adventure | https://api.github.com/repos/Ashleykkkb/cuddly-adventure | closed | License | bug documentation duplicate enhancement help wanted good first issue invalid | Copyright (c) GitHub, Inc.
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE. | 1.0 | License - Copyright (c) GitHub, Inc.
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE. | non_infrastructure | license copyright c github inc permission is hereby granted free of charge to any person obtaining a copy of this software and associated documentation files the software to deal in the software without restriction including without limitation the rights to use copy modify merge publish distribute sublicense and or sell copies of the software and to permit persons to whom the software is furnished to do so subject to the following conditions the above copyright notice and this permission notice shall be included in all copies or substantial portions of the software the software is provided as is without warranty of any kind express or implied including but not limited to the warranties of merchantability fitness for a particular purpose and noninfringement in no event shall the authors or copyright holders be liable for any claim damages or other liability whether in an action of contract tort or otherwise arising from out of or in connection with the software or the use or other dealings in the software | 0 |
774,129 | 27,184,326,205 | IssuesEvent | 2023-02-19 02:06:25 | Together-Java/TJ-Bot | https://api.github.com/repos/Together-Java/TJ-Bot | closed | ScamBlocker DM failure shouldnt result in an error | bug good first issue priority: normal help wanted | The `ScamBlocker` sends DMs to the authors of a message to inform them about the action. Thats totally fine and expected, but failure to DM a user (for example if they blocked DMs), it should not yield to an exception being thrown in the logs:
> ```java
> net.dv8tion.jda.api.requests.RestAction
> ERROR
> RestAction queue returned failure: [ErrorResponseException] 50007: Cannot send messages to this user
> net.dv8tion.jda.api.exceptions.ContextException
> at net.dv8tion.jda.api.exceptions.ContextException.here(ContextException.java:54)
> at net.dv8tion.jda.internal.requests.restaction.operator.RestActionOperator.contextWrap(RestActionOperator.java:123)
> at net.dv8tion.jda.internal.requests.restaction.operator.FlatMapRestAction.queue(FlatMapRestAction.java:50)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:572)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:538)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:256)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:252)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeActionAutoDeleteAndQuarantine(ScamBlocker.java:189)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeAction(ScamBl...
> 01/03/2023 5:52 PM
> ```
The actual DM sending should probably use `mapToResults()` and then maybe send a DEBUG level log message:
 | 1.0 | ScamBlocker DM failure shouldnt result in an error - The `ScamBlocker` sends DMs to the authors of a message to inform them about the action. Thats totally fine and expected, but failure to DM a user (for example if they blocked DMs), it should not yield to an exception being thrown in the logs:
> ```java
> net.dv8tion.jda.api.requests.RestAction
> ERROR
> RestAction queue returned failure: [ErrorResponseException] 50007: Cannot send messages to this user
> net.dv8tion.jda.api.exceptions.ContextException
> at net.dv8tion.jda.api.exceptions.ContextException.here(ContextException.java:54)
> at net.dv8tion.jda.internal.requests.restaction.operator.RestActionOperator.contextWrap(RestActionOperator.java:123)
> at net.dv8tion.jda.internal.requests.restaction.operator.FlatMapRestAction.queue(FlatMapRestAction.java:50)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:572)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:538)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:256)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:252)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeActionAutoDeleteAndQuarantine(ScamBlocker.java:189)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeAction(ScamBl...
> 01/03/2023 5:52 PM
> ```
The actual DM sending should probably use `mapToResults()` and then maybe send a DEBUG level log message:
 | non_infrastructure | scamblocker dm failure shouldnt result in an error the scamblocker sends dms to the authors of a message to inform them about the action thats totally fine and expected but failure to dm a user for example if they blocked dms it should not yield to an exception being thrown in the logs java net jda api requests restaction error restaction queue returned failure cannot send messages to this user net jda api exceptions contextexception at net jda api exceptions contextexception here contextexception java at net jda internal requests restaction operator restactionoperator contextwrap restactionoperator java at net jda internal requests restaction operator flatmaprestaction queue flatmaprestaction java at net jda api requests restaction queue restaction java at net jda api requests restaction queue restaction java at org togetherjava tjbot commands moderation scam scamblocker dmuser scamblocker java at org togetherjava tjbot commands moderation scam scamblocker dmuser scamblocker java at org togetherjava tjbot commands moderation scam scamblocker takeactionautodeleteandquarantine scamblocker java at org togetherjava tjbot commands moderation scam scamblocker takeaction scambl pm the actual dm sending should probably use maptoresults and then maybe send a debug level log message | 0 |
110,466 | 4,427,302,320 | IssuesEvent | 2016-08-16 20:59:32 | timwie/spoilerwatch | https://api.github.com/repos/timwie/spoilerwatch | closed | 'Not watched' selector availability | enhancement high priority | Since we have access to the watch history, we should prevent users from selecting episodes they haven't watched yet. | 1.0 | 'Not watched' selector availability - Since we have access to the watch history, we should prevent users from selecting episodes they haven't watched yet. | non_infrastructure | not watched selector availability since we have access to the watch history we should prevent users from selecting episodes they haven t watched yet | 0 |
154,474 | 24,302,521,929 | IssuesEvent | 2022-09-29 14:50:17 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | Surface related patterns in the template inspector | Needs Design Feedback [Feature] Patterns | When editing a template or template part, it would be nice to find associated patterns in the Inspector.
For template parts that might look something like:

It could work the same for templates too, but we'd need a way for patterns to be associated with specific templates. For example it wouldn't be helpful to be suggested 404 patterns while editing the Archive template. | 1.0 | Surface related patterns in the template inspector - When editing a template or template part, it would be nice to find associated patterns in the Inspector.
For template parts that might look something like:

It could work the same for templates too, but we'd need a way for patterns to be associated with specific templates. For example it wouldn't be helpful to be suggested 404 patterns while editing the Archive template. | non_infrastructure | surface related patterns in the template inspector when editing a template or template part it would be nice to find associated patterns in the inspector for template parts that might look something like it could work the same for templates too but we d need a way for patterns to be associated with specific templates for example it wouldn t be helpful to be suggested patterns while editing the archive template | 0 |
2,248 | 3,592,535,519 | IssuesEvent | 2016-02-01 16:22:29 | trackpete/exiletools-indexer | https://api.github.com/repos/trackpete/exiletools-indexer | closed | UIR: Load testing | enhancement infrastructure / exiletools.com setup reporting | When the report is complete, perform some basic load testing to ensure the system can handle initial expected load. Expect up to 100+ simultaneous requests. | 1.0 | UIR: Load testing - When the report is complete, perform some basic load testing to ensure the system can handle initial expected load. Expect up to 100+ simultaneous requests. | infrastructure | uir load testing when the report is complete perform some basic load testing to ensure the system can handle initial expected load expect up to simultaneous requests | 1 |
34,332 | 6,311,984,421 | IssuesEvent | 2017-07-24 00:29:45 | mangapress/mangapress | https://api.github.com/repos/mangapress/mangapress | opened | Lightbox integration | documentation enhancement question | Add tutorial on integrating a lightbox with Manga+Press, to allow a comic image to be full-screen.
Could this become a feature or separate plugin? | 1.0 | Lightbox integration - Add tutorial on integrating a lightbox with Manga+Press, to allow a comic image to be full-screen.
Could this become a feature or separate plugin? | non_infrastructure | lightbox integration add tutorial on integrating a lightbox with manga press to allow a comic image to be full screen could this become a feature or separate plugin | 0 |
8,206 | 7,290,693,435 | IssuesEvent | 2018-02-24 04:54:05 | CoolProp/CoolProp | https://api.github.com/repos/CoolProp/CoolProp | closed | Buildbot update | infrastructure | I recently updated my OSX builder and the old config seems to be incompatible with the new workers :-( . We might want to update the master and the master configuration accordingly. | 1.0 | Buildbot update - I recently updated my OSX builder and the old config seems to be incompatible with the new workers :-( . We might want to update the master and the master configuration accordingly. | infrastructure | buildbot update i recently updated my osx builder and the old config seems to be incompatible with the new workers we might want to update the master and the master configuration accordingly | 1 |
412,247 | 12,037,164,510 | IssuesEvent | 2020-04-13 21:14:53 | cybersemics/em | https://api.github.com/repos/cybersemics/em | opened | importText: Plaintext with <p> not imported correctly. | bug low-priority | ```
- A
- A paragraph tag is <p>
```
This text gets imported as a single thought since `<p>` makes it think it is HTML.
How to avoid false positive while preserving `isHTML` autodetect? | 1.0 | importText: Plaintext with <p> not imported correctly. - ```
- A
- A paragraph tag is <p>
```
This text gets imported as a single thought since `<p>` makes it think it is HTML.
How to avoid false positive while preserving `isHTML` autodetect? | non_infrastructure | importtext plaintext with not imported correctly a a paragraph tag is this text gets imported as a single thought since makes it think it is html how to avoid false positive while preserving ishtml autodetect | 0 |
18,139 | 12,804,161,531 | IssuesEvent | 2020-07-03 03:29:18 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | Models: Optimize searching enum-like and boolean fields where possible | category:models-and-data-infrastructure category:optimize-database status:stale | @mnelson4 reported in codebase 9144:
> For example, if we have the following query: `EEM_Registration::instance()->get_all( array( array( 'REG_deleted' => array( 'IN', array( true, false ) ) ) );` it may actually be quite a bit faster to automatically, behind the scenes, changes it to `EEM_Registration::instance()->get_all();`. This way MYSQL won't have to evaluate all the rows' `REG_deleted` cells to see if they're 0 or 1, which we intuitively know they are (but the machine might not).
> This might significantly help on big DBs especially. | 1.0 | Models: Optimize searching enum-like and boolean fields where possible - @mnelson4 reported in codebase 9144:
> For example, if we have the following query: `EEM_Registration::instance()->get_all( array( array( 'REG_deleted' => array( 'IN', array( true, false ) ) ) );` it may actually be quite a bit faster to automatically, behind the scenes, changes it to `EEM_Registration::instance()->get_all();`. This way MYSQL won't have to evaluate all the rows' `REG_deleted` cells to see if they're 0 or 1, which we intuitively know they are (but the machine might not).
> This might significantly help on big DBs especially. | infrastructure | models optimize searching enum like and boolean fields where possible reported in codebase for example if we have the following query eem registration instance get all array array reg deleted array in array true false it may actually be quite a bit faster to automatically behind the scenes changes it to eem registration instance get all this way mysql won t have to evaluate all the rows reg deleted cells to see if they re or which we intuitively know they are but the machine might not this might significantly help on big dbs especially | 1 |
22,218 | 15,052,239,397 | IssuesEvent | 2021-02-03 14:59:22 | airyhq/airy | https://api.github.com/repos/airyhq/airy | closed | The smoke test doesn't work anymore | bug infrastructure | After the latest release the `/infrastructure/scripts/smoke-test.sh` script fails when authenticating against the chatplugin.
Also we need to check if we still need to keep the `/infrasructure/scripts/chatplugin-generator.sh` script. | 1.0 | The smoke test doesn't work anymore - After the latest release the `/infrastructure/scripts/smoke-test.sh` script fails when authenticating against the chatplugin.
Also we need to check if we still need to keep the `/infrasructure/scripts/chatplugin-generator.sh` script. | infrastructure | the smoke test doesn t work anymore after the latest release the infrastructure scripts smoke test sh script fails when authenticating against the chatplugin also we need to check if we still need to keep the infrasructure scripts chatplugin generator sh script | 1 |
27,152 | 21,215,551,167 | IssuesEvent | 2022-04-11 06:54:44 | chatwoot/chatwoot | https://api.github.com/repos/chatwoot/chatwoot | closed | Compile Error | infrastructure | **Describe the bug**
When install and running I get a application error at this step
rake assets:precompile RAILS_ENV=production
```
db LICENSE SECURITY.md
chatwoot@chatwoot:~/chatwoot$ rake assets:precompile RAILS_ENV=production
yarn install v1.22.17
[1/5] Validating package.json...
[2/5] Resolving packages...
success Already up-to-date.
Done in 2.19s.
I, [2021-12-24T05:13:28.383261 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js
I, [2021-12-24T05:13:28.383627 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js.gz
I, [2021-12-24T05:13:28.394473 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js.gz
Compiling...
Compilation failed:
warn - The `conservative` purge mode will be removed in Tailwind 2.0.
warn - Please switch to the new `layers` mode instead.
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/assets/scss/app.scss 1:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/assets/scss/app.scss 1:9 @import
app/javascript/dashboard/App.vue 84:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/widgets/conversation/ConversationBox.vue 88:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/ChatList.vue 382:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/widgets/conversation/bubble/Actions.vue 133:9 root stylesheet
```
**To Reproduce**
Steps to reproduce the behavior:
4. See error
Share the server logs to debug the issue further
**Environment**
Self Hosted Ubuntu 20.4 / Linux VM
**Additional context**
Add any other context about the problem here.
| 1.0 | Compile Error - **Describe the bug**
When install and running I get a application error at this step
rake assets:precompile RAILS_ENV=production
```
db LICENSE SECURITY.md
chatwoot@chatwoot:~/chatwoot$ rake assets:precompile RAILS_ENV=production
yarn install v1.22.17
[1/5] Validating package.json...
[2/5] Resolving packages...
success Already up-to-date.
Done in 2.19s.
I, [2021-12-24T05:13:28.383261 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js
I, [2021-12-24T05:13:28.383627 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js.gz
I, [2021-12-24T05:13:28.394473 #251618] INFO -- : Writing /home/chatwoot/chatwoot/public/assets/administrate/application-784dfb416b2f7fc251341c48450416eeebe649439794dc5222ff57e0791d03da.js.gz
Compiling...
Compilation failed:
warn - The `conservative` purge mode will be removed in Tailwind 2.0.
warn - Please switch to the new `layers` mode instead.
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/assets/scss/app.scss 1:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/assets/scss/app.scss 1:9 @import
app/javascript/dashboard/App.vue 84:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/widgets/conversation/ConversationBox.vue 88:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/ChatList.vue 382:9 root stylesheet
WARNING: flex-grid-row-align(): flex-start is not a valid value for horizontal alignment. Use left, right, center, justify, or spaced.
node_modules/foundation-sites/scss/util/_flex.scss 43:7 flex-align()
app/javascript/dashboard/assets/scss/widgets/_modal.scss 88:5 @import
app/javascript/dashboard/assets/scss/_woot.scss 34:9 @import
app/javascript/dashboard/components/widgets/conversation/bubble/Actions.vue 133:9 root stylesheet
```
**To Reproduce**
Steps to reproduce the behavior:
4. See error
Share the server logs to debug the issue further
**Environment**
Self Hosted Ubuntu 20.4 / Linux VM
**Additional context**
Add any other context about the problem here.
| infrastructure | compile error describe the bug when install and running i get a application error at this step rake assets precompile rails env production db license security md chatwoot chatwoot chatwoot rake assets precompile rails env production yarn install validating package json resolving packages success already up to date done in i info writing home chatwoot chatwoot public assets administrate application js i info writing home chatwoot chatwoot public assets administrate application js gz i info writing home chatwoot chatwoot public assets administrate application js gz compiling compilation failed warn the conservative purge mode will be removed in tailwind warn please switch to the new layers mode instead warning flex grid row align flex start is not a valid value for horizontal alignment use left right center justify or spaced node modules foundation sites scss util flex scss flex align app javascript dashboard assets scss widgets modal scss import app javascript dashboard assets scss woot scss import app javascript dashboard assets scss app scss root stylesheet warning flex grid row align flex start is not a valid value for horizontal alignment use left right center justify or spaced node modules foundation sites scss util flex scss flex align app javascript dashboard assets scss widgets modal scss import app javascript dashboard assets scss woot scss import app javascript dashboard assets scss app scss import app javascript dashboard app vue root stylesheet warning flex grid row align flex start is not a valid value for horizontal alignment use left right center justify or spaced node modules foundation sites scss util flex scss flex align app javascript dashboard assets scss widgets modal scss import app javascript dashboard assets scss woot scss import app javascript dashboard components widgets conversation conversationbox vue root stylesheet warning flex grid row align flex start is not a valid value for horizontal alignment use left right center justify or spaced node modules foundation sites scss util flex scss flex align app javascript dashboard assets scss widgets modal scss import app javascript dashboard assets scss woot scss import app javascript dashboard components chatlist vue root stylesheet warning flex grid row align flex start is not a valid value for horizontal alignment use left right center justify or spaced node modules foundation sites scss util flex scss flex align app javascript dashboard assets scss widgets modal scss import app javascript dashboard assets scss woot scss import app javascript dashboard components widgets conversation bubble actions vue root stylesheet to reproduce steps to reproduce the behavior see error share the server logs to debug the issue further environment self hosted ubuntu linux vm additional context add any other context about the problem here | 1 |
63,448 | 26,399,374,168 | IssuesEvent | 2023-01-12 22:57:51 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | Support for "Standard test" as part of azurerm_application_insights_web_test | enhancement service/application-insights | ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Description
Azure added new `kind` of the availability tests called "Standard tests". It has some more options compared to the `ping` test (like sending payload & http headers, meaning it can be used for pinging restricted resources, which is exactly why we need it).
### New or Affected Resource(s)
* azurerm_application_insights_web_test
### References
* https://docs.microsoft.com/en-us/azure/azure-monitor/app/availability-standard-tests
| 1.0 | Support for "Standard test" as part of azurerm_application_insights_web_test - ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Description
Azure added new `kind` of the availability tests called "Standard tests". It has some more options compared to the `ping` test (like sending payload & http headers, meaning it can be used for pinging restricted resources, which is exactly why we need it).
### New or Affected Resource(s)
* azurerm_application_insights_web_test
### References
* https://docs.microsoft.com/en-us/azure/azure-monitor/app/availability-standard-tests
| non_infrastructure | support for standard test as part of azurerm application insights web test community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description azure added new kind of the availability tests called standard tests it has some more options compared to the ping test like sending payload http headers meaning it can be used for pinging restricted resources which is exactly why we need it new or affected resource s azurerm application insights web test references | 0 |
134,692 | 10,927,151,824 | IssuesEvent | 2019-11-22 16:06:00 | LiskHQ/lisk-sdk | https://api.github.com/repos/LiskHQ/lisk-sdk | closed | Add more test coverage for logic that uses delegateMinHeightActive value | type: test | ### Description
As a follow up to the #4473 we need to add missing unit and integration tests.
### Acceptance Criteria
- Add unit test coverage to DPoS module
- Add integration test coverage to DPoS module.
- Add unit test coverage to BFT module.
### Additional Information
| 1.0 | Add more test coverage for logic that uses delegateMinHeightActive value - ### Description
As a follow up to the #4473 we need to add missing unit and integration tests.
### Acceptance Criteria
- Add unit test coverage to DPoS module
- Add integration test coverage to DPoS module.
- Add unit test coverage to BFT module.
### Additional Information
| non_infrastructure | add more test coverage for logic that uses delegateminheightactive value description as a follow up to the we need to add missing unit and integration tests acceptance criteria add unit test coverage to dpos module add integration test coverage to dpos module add unit test coverage to bft module additional information | 0 |
400,210 | 27,272,850,503 | IssuesEvent | 2023-02-23 00:35:19 | AlaskaAirlines/auro-combobox | https://api.github.com/repos/AlaskaAirlines/auro-combobox | closed | Missing demo for combobox with menu | not-reviewed Type: Documentation auro-combobox | # General Support Request
Looking around the Auro docsite we are missing a demo using combobox with auro-menu options to select from.
## Support request
There is an example of typing into the combobox and getting a smaller list, but there are no examples for how this would be used in the scenario for displaying all menu options when clicking on the trigger.
## Possible Solution
Add a demo to illustrate this use case. | 1.0 | Missing demo for combobox with menu - # General Support Request
Looking around the Auro docsite we are missing a demo using combobox with auro-menu options to select from.
## Support request
There is an example of typing into the combobox and getting a smaller list, but there are no examples for how this would be used in the scenario for displaying all menu options when clicking on the trigger.
## Possible Solution
Add a demo to illustrate this use case. | non_infrastructure | missing demo for combobox with menu general support request looking around the auro docsite we are missing a demo using combobox with auro menu options to select from support request there is an example of typing into the combobox and getting a smaller list but there are no examples for how this would be used in the scenario for displaying all menu options when clicking on the trigger possible solution add a demo to illustrate this use case | 0 |
16,536 | 11,035,750,900 | IssuesEvent | 2019-12-07 15:54:00 | spyder-ide/spyder | https://api.github.com/repos/spyder-ide/spyder | closed | Hover tips interfere with editor text selection | component:Editor tag:Ux-usability | <!--- **PLEASE READ:** When submitting here, please ensure you've completed the following checklist and checked the boxes to confirm. Issue reports without it may be closed. Thanks! --->
### Issue Report Checklist
* [X] Searched the [issues page](https://github.com/spyder-ide/spyder/issues?q=is%3Aissue) for similar reports
* [X] Read the relevant sections of the [Spyder Troubleshooting Guide](https://github.com/spyder-ide/spyder/wiki/Troubleshooting-Guide-and-FAQ) and followed its advice
* [X] Reproduced the issue after updating with ``conda update spyder`` (or ``pip``, if not using Anaconda)
* [ ] Could not reproduce inside ``jupyter qtconsole`` (if console-related)
* [ ] Tried basic troubleshooting (if a bug/error)
* [ ] Restarted Spyder
* [ ] Reset preferences with ``spyder --reset``
* [ ] Reinstalled the latest version of [Anaconda](https://www.anaconda.com/download/)
* [ ] Tried the other applicable steps from the Troubleshooting Guide
* [X] Completed the **Problem Description**, **Steps to Reproduce** and **Version** sections below
## Problem Description
The hover hints interfere with highlighting text.


### What steps reproduce the problem?
1. Activate a hover
2. Click and drag to highlight
3. Can't see the highlighting
### What is the expected output? What do you see instead?
I think that if the hover windows disapapere if the user clicks in the editor that would be one way to solve it
## Versions
<!--- You can get this information from Help > About Spyder...
or (if Spyder won't launch) the "conda list" command
from the Anaconda Prompt/Terminal/command line. --->
* Spyder version: master
* Python version: 3.7.3 64-bit
* Qt version: Qt 5.9.6
* PyQt version: PyQt5 5.9.2
* Operating System name/version: Linux
### Dependencies
<!--- Please go to the menu entry Help > Dependencies,
press the Copy to clipboard button and paste below --->
```
PASTE DEPENDENCIES HERE
```
cloudpickle >=0.5.0 : 1.2.1 (OK)
pygments >=2.0 : 2.4.2 (OK)
qtconsole >=4.6.0 : 4.6.0 (OK)
nbconvert >=4.0 : 5.5.0 (OK)
sphinx >=0.6.6 : 2.1.2 (OK)
pylint >=0.25 : 2.3.1 (OK)
psutil >=0.3 : 5.6.3 (OK)
qtawesome >=0.5.7 : 0.5.7 (OK)
qtpy >=1.5.0 : 1.9.0 (OK)
pickleshare >=0.4 : 0.7.5 (OK)
zmq >=17 : 18.1.0 (OK)
chardet >=2.0.0 : 3.0.4 (OK)
numpydoc >=0.6.0 : 0.9.1 (OK)
spyder_kernels >=1.8.0;<2.0.0: 1.9.0.dev0 (OK)
qdarkstyle >=2.7 : 2.7 (OK)
atomicwrites >=1.2.0 : 1.3.0 (OK)
diff_match_patch >=20181111 : 20181111 (OK)
intervaltree : None (OK)
watchdog : None (OK)
keyring : None (OK)
pexpect >=4.4.0 : 4.7.0 (OK)
pympler : None (OK)
sympy >=0.7.3 : 1.4 (OK)
cython >=0.21 : 0.29.13 (OK)
IPython >=4.0 : 7.9.0 (OK)
matplotlib >=2.0.0 : 3.1.0 (OK)
pandas >=0.13.1 : 0.25.0 (OK)
numpy >=1.7 : 1.17.3 (OK)
scipy >=0.17.0 : 1.3.1 (OK)
pyls >=0.31.0;<0.32.0 : 0.31.1 (OK)
xdg >=0.26 : 0.26 (OK)
rtree >=0.8.3 : 0.8.3 (OK) | True | Hover tips interfere with editor text selection - <!--- **PLEASE READ:** When submitting here, please ensure you've completed the following checklist and checked the boxes to confirm. Issue reports without it may be closed. Thanks! --->
### Issue Report Checklist
* [X] Searched the [issues page](https://github.com/spyder-ide/spyder/issues?q=is%3Aissue) for similar reports
* [X] Read the relevant sections of the [Spyder Troubleshooting Guide](https://github.com/spyder-ide/spyder/wiki/Troubleshooting-Guide-and-FAQ) and followed its advice
* [X] Reproduced the issue after updating with ``conda update spyder`` (or ``pip``, if not using Anaconda)
* [ ] Could not reproduce inside ``jupyter qtconsole`` (if console-related)
* [ ] Tried basic troubleshooting (if a bug/error)
* [ ] Restarted Spyder
* [ ] Reset preferences with ``spyder --reset``
* [ ] Reinstalled the latest version of [Anaconda](https://www.anaconda.com/download/)
* [ ] Tried the other applicable steps from the Troubleshooting Guide
* [X] Completed the **Problem Description**, **Steps to Reproduce** and **Version** sections below
## Problem Description
The hover hints interfere with highlighting text.


### What steps reproduce the problem?
1. Activate a hover
2. Click and drag to highlight
3. Can't see the highlighting
### What is the expected output? What do you see instead?
I think that if the hover windows disapapere if the user clicks in the editor that would be one way to solve it
## Versions
<!--- You can get this information from Help > About Spyder...
or (if Spyder won't launch) the "conda list" command
from the Anaconda Prompt/Terminal/command line. --->
* Spyder version: master
* Python version: 3.7.3 64-bit
* Qt version: Qt 5.9.6
* PyQt version: PyQt5 5.9.2
* Operating System name/version: Linux
### Dependencies
<!--- Please go to the menu entry Help > Dependencies,
press the Copy to clipboard button and paste below --->
```
PASTE DEPENDENCIES HERE
```
cloudpickle >=0.5.0 : 1.2.1 (OK)
pygments >=2.0 : 2.4.2 (OK)
qtconsole >=4.6.0 : 4.6.0 (OK)
nbconvert >=4.0 : 5.5.0 (OK)
sphinx >=0.6.6 : 2.1.2 (OK)
pylint >=0.25 : 2.3.1 (OK)
psutil >=0.3 : 5.6.3 (OK)
qtawesome >=0.5.7 : 0.5.7 (OK)
qtpy >=1.5.0 : 1.9.0 (OK)
pickleshare >=0.4 : 0.7.5 (OK)
zmq >=17 : 18.1.0 (OK)
chardet >=2.0.0 : 3.0.4 (OK)
numpydoc >=0.6.0 : 0.9.1 (OK)
spyder_kernels >=1.8.0;<2.0.0: 1.9.0.dev0 (OK)
qdarkstyle >=2.7 : 2.7 (OK)
atomicwrites >=1.2.0 : 1.3.0 (OK)
diff_match_patch >=20181111 : 20181111 (OK)
intervaltree : None (OK)
watchdog : None (OK)
keyring : None (OK)
pexpect >=4.4.0 : 4.7.0 (OK)
pympler : None (OK)
sympy >=0.7.3 : 1.4 (OK)
cython >=0.21 : 0.29.13 (OK)
IPython >=4.0 : 7.9.0 (OK)
matplotlib >=2.0.0 : 3.1.0 (OK)
pandas >=0.13.1 : 0.25.0 (OK)
numpy >=1.7 : 1.17.3 (OK)
scipy >=0.17.0 : 1.3.1 (OK)
pyls >=0.31.0;<0.32.0 : 0.31.1 (OK)
xdg >=0.26 : 0.26 (OK)
rtree >=0.8.3 : 0.8.3 (OK) | non_infrastructure | hover tips interfere with editor text selection issue report checklist searched the for similar reports read the relevant sections of the and followed its advice reproduced the issue after updating with conda update spyder or pip if not using anaconda could not reproduce inside jupyter qtconsole if console related tried basic troubleshooting if a bug error restarted spyder reset preferences with spyder reset reinstalled the latest version of tried the other applicable steps from the troubleshooting guide completed the problem description steps to reproduce and version sections below problem description the hover hints interfere with highlighting text what steps reproduce the problem activate a hover click and drag to highlight can t see the highlighting what is the expected output what do you see instead i think that if the hover windows disapapere if the user clicks in the editor that would be one way to solve it versions about spyder or if spyder won t launch the conda list command from the anaconda prompt terminal command line spyder version master python version bit qt version qt pyqt version operating system name version linux dependencies dependencies press the copy to clipboard button and paste below paste dependencies here cloudpickle ok pygments ok qtconsole ok nbconvert ok sphinx ok pylint ok psutil ok qtawesome ok qtpy ok pickleshare ok zmq ok chardet ok numpydoc ok spyder kernels ok qdarkstyle ok atomicwrites ok diff match patch ok intervaltree none ok watchdog none ok keyring none ok pexpect ok pympler none ok sympy ok cython ok ipython ok matplotlib ok pandas ok numpy ok scipy ok pyls ok xdg ok rtree ok | 0 |
364,417 | 10,764,016,115 | IssuesEvent | 2019-11-01 06:44:31 | erxes/erxes | https://api.github.com/repos/erxes/erxes | closed | When the user switches between the board, pipeline, stages, and selected only board. The pipeline and stage are automatically selected and saved. | committed priority: High scope: Deal scope: Task scope: Ticket type: bug | 
| 1.0 | When the user switches between the board, pipeline, stages, and selected only board. The pipeline and stage are automatically selected and saved. - 
| non_infrastructure | when the user switches between the board pipeline stages and selected only board the pipeline and stage are automatically selected and saved | 0 |
265,948 | 20,116,521,216 | IssuesEvent | 2022-02-07 20:08:58 | cigie-shared-services/start-here | https://api.github.com/repos/cigie-shared-services/start-here | closed | Add Project Initiation Flowchart | documentation enhancement | <!-- Welcome! Thank you for contributing. These HTML comments will not render in the issue, but you can delete them once you've read them if you prefer!
Before creating a new issue:
* Search for relevant issues
* Need tips on issue writing? See the following Wiredcraft article for general guidelines:
https://wiredcraft.com/blog/how-we-write-our-github-issues/
-->
## Problem
<!-- Is your feature request related to a problem? Please describe a clear and concise description of what the problem is.
* I'm always frustrated when [...].
* I would like it if [...] happened when I [...] because [...].
-->
A clear illustration that shows how to submit a project proposal for review and approval does not exist.
## Proposed Solution
<!-- Describe the solution you'd like. A clear and concise description of what you want to happen.
* Add an option so that when [...] [...] will happen.
-->
Develop a flowchart diagram on Lucidchart to illustrate the project initiation process.
## Additional Context
<!-- Add any other context or screenshots about the feature request here.
* Another project [...] solved this by [...].
-->
[Lucidchart](https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&cad=rja&uact=8&ved=2ahUKEwjE2MGIsu71AhUUP30KHS4FBeoQFnoECAQQAQ&url=https%3A%2F%2Fwww.lucidchart.com%2Fpages%2F&usg=AOvVaw0AMip6vhaE-AZ-qlWAOqLm)
| 1.0 | Add Project Initiation Flowchart - <!-- Welcome! Thank you for contributing. These HTML comments will not render in the issue, but you can delete them once you've read them if you prefer!
Before creating a new issue:
* Search for relevant issues
* Need tips on issue writing? See the following Wiredcraft article for general guidelines:
https://wiredcraft.com/blog/how-we-write-our-github-issues/
-->
## Problem
<!-- Is your feature request related to a problem? Please describe a clear and concise description of what the problem is.
* I'm always frustrated when [...].
* I would like it if [...] happened when I [...] because [...].
-->
A clear illustration that shows how to submit a project proposal for review and approval does not exist.
## Proposed Solution
<!-- Describe the solution you'd like. A clear and concise description of what you want to happen.
* Add an option so that when [...] [...] will happen.
-->
Develop a flowchart diagram on Lucidchart to illustrate the project initiation process.
## Additional Context
<!-- Add any other context or screenshots about the feature request here.
* Another project [...] solved this by [...].
-->
[Lucidchart](https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&cad=rja&uact=8&ved=2ahUKEwjE2MGIsu71AhUUP30KHS4FBeoQFnoECAQQAQ&url=https%3A%2F%2Fwww.lucidchart.com%2Fpages%2F&usg=AOvVaw0AMip6vhaE-AZ-qlWAOqLm)
| non_infrastructure | add project initiation flowchart welcome thank you for contributing these html comments will not render in the issue but you can delete them once you ve read them if you prefer before creating a new issue search for relevant issues need tips on issue writing see the following wiredcraft article for general guidelines problem is your feature request related to a problem please describe a clear and concise description of what the problem is i m always frustrated when i would like it if happened when i because a clear illustration that shows how to submit a project proposal for review and approval does not exist proposed solution describe the solution you d like a clear and concise description of what you want to happen add an option so that when will happen develop a flowchart diagram on lucidchart to illustrate the project initiation process additional context add any other context or screenshots about the feature request here another project solved this by | 0 |
123,107 | 16,444,411,145 | IssuesEvent | 2021-05-20 17:46:30 | KeyWorksRW/wxUiEditor | https://api.github.com/repos/KeyWorksRW/wxUiEditor | closed | Generators need to use \t not 4 spaces for special indentation | design change | ### Description:
<!-- Provide a description of of what you want to happen here -->
A lot of the code generation is using spaces for indentation. That is what we want in the final file, but carefully prefixing strings with the exact number of spaces is a mistake waiting to happen. Currently, there's code like the following:
```c++
code << "\n " << GenerateQuotedString(msg);
```
In looking at the above code, you can't immediately tell that there are exactly 8 spaces. If there aren't, the spacing will be wrong in the output file. Even when clang-format support gets added, not all users will have it installed so we can't rely on it to fix the problem.
What we need is support for using '\t' and converting it to 4 spaces in the output file. The best location for this processing would probably be WriteCode::WriteCodeLine() and WriteCode::write(). Both these methods require ttlib::cview for the code since it can be a literal. In both cases, before `doWrite(code)` is called, a search for '\t' should be done and each one converted into 4 spaces if found.
Note that this assumes there aren't any real non-escaped tabs in the output string. If there are, we'll need to figure out how to special-case them. | 1.0 | Generators need to use \t not 4 spaces for special indentation - ### Description:
<!-- Provide a description of of what you want to happen here -->
A lot of the code generation is using spaces for indentation. That is what we want in the final file, but carefully prefixing strings with the exact number of spaces is a mistake waiting to happen. Currently, there's code like the following:
```c++
code << "\n " << GenerateQuotedString(msg);
```
In looking at the above code, you can't immediately tell that there are exactly 8 spaces. If there aren't, the spacing will be wrong in the output file. Even when clang-format support gets added, not all users will have it installed so we can't rely on it to fix the problem.
What we need is support for using '\t' and converting it to 4 spaces in the output file. The best location for this processing would probably be WriteCode::WriteCodeLine() and WriteCode::write(). Both these methods require ttlib::cview for the code since it can be a literal. In both cases, before `doWrite(code)` is called, a search for '\t' should be done and each one converted into 4 spaces if found.
Note that this assumes there aren't any real non-escaped tabs in the output string. If there are, we'll need to figure out how to special-case them. | non_infrastructure | generators need to use t not spaces for special indentation description a lot of the code generation is using spaces for indentation that is what we want in the final file but carefully prefixing strings with the exact number of spaces is a mistake waiting to happen currently there s code like the following c code n generatequotedstring msg in looking at the above code you can t immediately tell that there are exactly spaces if there aren t the spacing will be wrong in the output file even when clang format support gets added not all users will have it installed so we can t rely on it to fix the problem what we need is support for using t and converting it to spaces in the output file the best location for this processing would probably be writecode writecodeline and writecode write both these methods require ttlib cview for the code since it can be a literal in both cases before dowrite code is called a search for t should be done and each one converted into spaces if found note that this assumes there aren t any real non escaped tabs in the output string if there are we ll need to figure out how to special case them | 0 |
15,805 | 27,986,491,122 | IssuesEvent | 2023-03-26 18:56:34 | MonarchDevelopment/SquireCore | https://api.github.com/repos/MonarchDevelopment/SquireCore | opened | [Dev]: Round Editor | todo requirement SquireWeb | ## Unmet Need:
There needs to be a way to edit rounds. This should include dropping a player from a round, adding game wins for players, adding game draws, and confirming the results on behalf of each player, and, for convenience, a "confirm round" button that submits confirmation for all players.
## Challenges/Considerations:
There should probably be a "Confirm all rounds" button too, perhaps somewhere near the filter input.
| 1.0 | [Dev]: Round Editor - ## Unmet Need:
There needs to be a way to edit rounds. This should include dropping a player from a round, adding game wins for players, adding game draws, and confirming the results on behalf of each player, and, for convenience, a "confirm round" button that submits confirmation for all players.
## Challenges/Considerations:
There should probably be a "Confirm all rounds" button too, perhaps somewhere near the filter input.
| non_infrastructure | round editor unmet need there needs to be a way to edit rounds this should include dropping a player from a round adding game wins for players adding game draws and confirming the results on behalf of each player and for convenience a confirm round button that submits confirmation for all players challenges considerations there should probably be a confirm all rounds button too perhaps somewhere near the filter input | 0 |
3,486 | 6,555,794,276 | IssuesEvent | 2017-09-06 11:48:47 | zero-os/0-stor | https://api.github.com/repos/zero-os/0-stor | closed | Merge ObjectCreate with Write & ObjectGet with Read | process_wontfix type_bug | ObjectCreate & ObjectGet currently create & get data without processing it through pipes.
While Write&Get process it through pipes.
Client should merge this those methods, depend on the pipes existence | 1.0 | Merge ObjectCreate with Write & ObjectGet with Read - ObjectCreate & ObjectGet currently create & get data without processing it through pipes.
While Write&Get process it through pipes.
Client should merge this those methods, depend on the pipes existence | non_infrastructure | merge objectcreate with write objectget with read objectcreate objectget currently create get data without processing it through pipes while write get process it through pipes client should merge this those methods depend on the pipes existence | 0 |
28,314 | 23,146,767,445 | IssuesEvent | 2022-07-29 02:19:10 | deckhouse/deckhouse | https://api.github.com/repos/deckhouse/deckhouse | reopened | Choice "source of truth" of deckhouse registry and add d8-system/deckhouse-registry secret validation webhook | area/cluster-and-infrastructure area/core | ### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Version
v1.29.0
### Expected Behavior
Deckhouse registry for modules images and bashible should get from one source.
If set incorrect credentials for Secret `d8-system/deckhouse registry`, cluster should not break.
### Actual Behavior
For Deckhouse modules we get Deckhouse registry (registry host and registry path) from Deckhouse deployment, but for bashible steps we get registry from `d8-system/deckhouse-registry` secret. It is confusing. We need one source of true for it.
If we put incorrect registry credentials to `d8-system/deckhouse-registry` secret, we can break cluster. Why?
When `d8-system/deckhouse-registry` secret was changed, Deckhouse deployment will be restarted.
Dechouse controller will be restarted successfully, because we write deckhouse credentials into containerd config and image will be pulled, although `d8-system/deckhouse-registry` is imagePullSecret and Deckouse deployment has imagePullPolicy - `Always`.
After restart, Deckhouse controller will change `control-plane-manager` module with incorrect credentials and `control-plane-manager` change all static pods manifests (etcd, kube-apiserver...) and cluster became not working, because kubelet will not pull images.
It is not good for single-master clusters.
### Steps To Reproduce
_No response_
### Additional Information
For second case (with deckhouse-registry secret) we can add validation web hook.
### Logs
_No response_ | 1.0 | Choice "source of truth" of deckhouse registry and add d8-system/deckhouse-registry secret validation webhook - ### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Version
v1.29.0
### Expected Behavior
Deckhouse registry for modules images and bashible should get from one source.
If set incorrect credentials for Secret `d8-system/deckhouse registry`, cluster should not break.
### Actual Behavior
For Deckhouse modules we get Deckhouse registry (registry host and registry path) from Deckhouse deployment, but for bashible steps we get registry from `d8-system/deckhouse-registry` secret. It is confusing. We need one source of true for it.
If we put incorrect registry credentials to `d8-system/deckhouse-registry` secret, we can break cluster. Why?
When `d8-system/deckhouse-registry` secret was changed, Deckhouse deployment will be restarted.
Dechouse controller will be restarted successfully, because we write deckhouse credentials into containerd config and image will be pulled, although `d8-system/deckhouse-registry` is imagePullSecret and Deckouse deployment has imagePullPolicy - `Always`.
After restart, Deckhouse controller will change `control-plane-manager` module with incorrect credentials and `control-plane-manager` change all static pods manifests (etcd, kube-apiserver...) and cluster became not working, because kubelet will not pull images.
It is not good for single-master clusters.
### Steps To Reproduce
_No response_
### Additional Information
For second case (with deckhouse-registry secret) we can add validation web hook.
### Logs
_No response_ | infrastructure | choice source of truth of deckhouse registry and add system deckhouse registry secret validation webhook preflight checklist i agree to follow the that this project adheres to i have searched the for an issue that matches the one i want to file without success version expected behavior deckhouse registry for modules images and bashible should get from one source if set incorrect credentials for secret system deckhouse registry cluster should not break actual behavior for deckhouse modules we get deckhouse registry registry host and registry path from deckhouse deployment but for bashible steps we get registry from system deckhouse registry secret it is confusing we need one source of true for it if we put incorrect registry credentials to system deckhouse registry secret we can break cluster why when system deckhouse registry secret was changed deckhouse deployment will be restarted dechouse controller will be restarted successfully because we write deckhouse credentials into containerd config and image will be pulled although system deckhouse registry is imagepullsecret and deckouse deployment has imagepullpolicy always after restart deckhouse controller will change control plane manager module with incorrect credentials and control plane manager change all static pods manifests etcd kube apiserver and cluster became not working because kubelet will not pull images it is not good for single master clusters steps to reproduce no response additional information for second case with deckhouse registry secret we can add validation web hook logs no response | 1 |
5,226 | 8,029,426,362 | IssuesEvent | 2018-07-27 15:57:33 | GoogleCloudPlatform/google-cloud-python | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python | closed | Bigtable system tests fail creating tables with 503 | api: bigtable flaky testing type: process | CI failures for changes unrelated to Bigtable:
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6267
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6268 | 1.0 | Bigtable system tests fail creating tables with 503 - CI failures for changes unrelated to Bigtable:
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6267
- https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/6268 | non_infrastructure | bigtable system tests fail creating tables with ci failures for changes unrelated to bigtable | 0 |
27,108 | 21,180,088,125 | IssuesEvent | 2022-04-08 07:01:51 | 3PillarGlobal-Czechia/interview-app-api | https://api.github.com/repos/3PillarGlobal-Czechia/interview-app-api | closed | DevOps: Final steps for deployment automatization | enhancement infrastructure | To finish deployment scripts following steps need to be implemented for the deployment pipeline.
* stop app before deployment
* get the connection string from KeyVault
* execute migrations on server | 1.0 | DevOps: Final steps for deployment automatization - To finish deployment scripts following steps need to be implemented for the deployment pipeline.
* stop app before deployment
* get the connection string from KeyVault
* execute migrations on server | infrastructure | devops final steps for deployment automatization to finish deployment scripts following steps need to be implemented for the deployment pipeline stop app before deployment get the connection string from keyvault execute migrations on server | 1 |
26,123 | 19,680,956,792 | IssuesEvent | 2022-01-11 16:42:38 | bcgov/foi-flow | https://api.github.com/repos/bcgov/foi-flow | closed | Research Data Visualization tools for reporting | Task Data Infrastructure incomplete AC | Title of ticket:
#### Description
Determine which tool would be suitable for dashboard use within our core application
#### Dependencies
Are there any dependencies?
#### DOD
- [ ] List the items that need to be complete for this ticket to be considered done
- [ ]
- [ ]
- [ ]
- [ ]
| 1.0 | Research Data Visualization tools for reporting - Title of ticket:
#### Description
Determine which tool would be suitable for dashboard use within our core application
#### Dependencies
Are there any dependencies?
#### DOD
- [ ] List the items that need to be complete for this ticket to be considered done
- [ ]
- [ ]
- [ ]
- [ ]
| infrastructure | research data visualization tools for reporting title of ticket description determine which tool would be suitable for dashboard use within our core application dependencies are there any dependencies dod list the items that need to be complete for this ticket to be considered done | 1 |
260 | 2,598,997,988 | IssuesEvent | 2015-02-23 01:12:23 | RobDixonIII/Bloom | https://api.github.com/repos/RobDixonIII/Bloom | closed | Create Home Shared Module | infrastructure | Create the module to facilitate the home tab as described in issue #94 . | 1.0 | Create Home Shared Module - Create the module to facilitate the home tab as described in issue #94 . | infrastructure | create home shared module create the module to facilitate the home tab as described in issue | 1 |
845 | 4,434,960,338 | IssuesEvent | 2016-08-18 06:20:44 | Guite/MostGenerator | https://api.github.com/repos/Guite/MostGenerator | closed | Use dedicated config controller | architecture zk 1.4.x | Disallow entity name `config`.
Generate configuration handling into dedicated config controller.
Preparation for #715 | 1.0 | Use dedicated config controller - Disallow entity name `config`.
Generate configuration handling into dedicated config controller.
Preparation for #715 | non_infrastructure | use dedicated config controller disallow entity name config generate configuration handling into dedicated config controller preparation for | 0 |
180,213 | 6,647,321,577 | IssuesEvent | 2017-09-28 03:08:20 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.google.com.mx - site is not usable | browser-firefox priority-critical status-needstriage type-stylo | <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://www.google.com.mx/provisioning/index.php
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: This page does not load in this browser
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/9/0f45b43a-efc1-435e-bd55-8894649b6717.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.google.com.mx - site is not usable - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://www.google.com.mx/provisioning/index.php
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: This page does not load in this browser
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/9/0f45b43a-efc1-435e-bd55-8894649b6717.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description this page does not load in this browser steps to reproduce layout css servo enabled true from with ❤️ | 0 |
30,697 | 25,001,281,184 | IssuesEvent | 2022-11-03 08:09:54 | Altinn/altinn-platform | https://api.github.com/repos/Altinn/altinn-platform | closed | Add access for developer image to service owner storage account | solution/apps ops/infrastructure | ## What
Developers need to access storage account from developer image. Whitelist subnet for developer images on service owner storage account.
## How
Added developer image subnet to storage account
```
modules/aks/main.tf
```
Run pipelines:
```
altinn-apps-orchestrator-pipeline
run-apply-on-rings
```
Ref: https://pedia.altinn.cloud/altinn-3/ops/patching/terraform/apps/ | 1.0 | Add access for developer image to service owner storage account - ## What
Developers need to access storage account from developer image. Whitelist subnet for developer images on service owner storage account.
## How
Added developer image subnet to storage account
```
modules/aks/main.tf
```
Run pipelines:
```
altinn-apps-orchestrator-pipeline
run-apply-on-rings
```
Ref: https://pedia.altinn.cloud/altinn-3/ops/patching/terraform/apps/ | infrastructure | add access for developer image to service owner storage account what developers need to access storage account from developer image whitelist subnet for developer images on service owner storage account how added developer image subnet to storage account modules aks main tf run pipelines altinn apps orchestrator pipeline run apply on rings ref | 1 |
107,962 | 4,322,648,865 | IssuesEvent | 2016-07-25 14:42:41 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | opened | Impact Pathway: Pre-fill 2022 year when adding new Outcomes | Priority - High Type - Enhancement | **Feedback from Amanda Wyatt**:
> Outcomes are all 2022. Maybe target year could be pre-filled? | 1.0 | Impact Pathway: Pre-fill 2022 year when adding new Outcomes - **Feedback from Amanda Wyatt**:
> Outcomes are all 2022. Maybe target year could be pre-filled? | non_infrastructure | impact pathway pre fill year when adding new outcomes feedback from amanda wyatt outcomes are all maybe target year could be pre filled | 0 |
8,880 | 7,716,938,671 | IssuesEvent | 2018-05-23 12:21:19 | seqan/seqan3 | https://api.github.com/repos/seqan/seqan3 | opened | Linking libdivsufsort temporarily (for SDSL) | infrastructure | Anyone familiar with the cmake in SeqAn3? @marehr :grin: | 1.0 | Linking libdivsufsort temporarily (for SDSL) - Anyone familiar with the cmake in SeqAn3? @marehr :grin: | infrastructure | linking libdivsufsort temporarily for sdsl anyone familiar with the cmake in marehr grin | 1 |
528,656 | 15,371,562,213 | IssuesEvent | 2021-03-02 10:09:48 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | InstrumentView crash debug mode OSX | High Priority MantidPlot Stale | This issue was originally [TRAC 10963](http://trac.mantidproject.org/mantid/ticket/10963)
- Open MantidPlot
- Open a workspace in the AutoTest directory, I used one of the inter runs
- Open the instrument view -> Crash
```
Thread 0 Crashed:: Dispatch queue: com.apple.main-thread
0 QtCore 0x000000011519c650 QMetaObjectPrivate::indexOfSignalRelative(QMetaObject const**, char const*, bool) + 64
1 QtCore 0x00000001151ac343 QObject::connect(QObject const*, char const*, QObject const*, char const*, Qt::ConnectionType) + 483
2 org.mantidproject.MantidPlot 0x000000010f52f22f ProjectionSurface::ProjectionSurface(InstrumentActor const*) + 463 (ProjectionSurface.cpp:44)
3 org.mantidproject.MantidPlot 0x000000010f533987 ProjectionSurface::ProjectionSurface(InstrumentActor const*) + 35
4 org.mantidproject.MantidPlot 0x000000010f53576e Projection3D::Projection3D(InstrumentActor const*, int, int) + 66 (Projection3D.cpp:40)
5 org.mantidproject.MantidPlot 0x000000010f4ca09e InstrumentWindow::setSurfaceType(int) + 2840 (InstrumentWindow.cpp:374)
6 org.mantidproject.MantidPlot 0x000000010f4cac97 InstrumentWindow::setSurfaceType(QString const&) + 497 (InstrumentWindow.cpp:479)
7 org.mantidproject.MantidPlot 0x000000010f4c9107 InstrumentWindow::init(bool, bool, double, double, bool) + 1661 (InstrumentWindow.cpp:218)
8 org.mantidproject.MantidPlot 0x000000010f45cd01 MantidUI::getInstrumentView(QString const&, int) + 1821 (MantidUI.cpp:2023)
9 org.mantidproject.MantidPlot 0x000000010f45d0d4 MantidUI::showMantidInstrument(QString const&) + 42 (MantidUI.cpp:2054)
10 org.mantidproject.MantidPlot 0x000000010f45d449 MantidUI::showMantidInstrumentSelected() + 113 (MantidUI.cpp:2094)
11 org.mantidproject.MantidPlot 0x000000010f5e4e88 MantidUI::qt_static_metacall(QObject*, QMetaObject::Call, int, void**) + 12940 (moc_MantidUI.cxx:510)
```
This doesn't affect the last stable release, so either the problem has been introduced since then, or the compilation without the optimisations is responsible in some way.
| 1.0 | InstrumentView crash debug mode OSX - This issue was originally [TRAC 10963](http://trac.mantidproject.org/mantid/ticket/10963)
- Open MantidPlot
- Open a workspace in the AutoTest directory, I used one of the inter runs
- Open the instrument view -> Crash
```
Thread 0 Crashed:: Dispatch queue: com.apple.main-thread
0 QtCore 0x000000011519c650 QMetaObjectPrivate::indexOfSignalRelative(QMetaObject const**, char const*, bool) + 64
1 QtCore 0x00000001151ac343 QObject::connect(QObject const*, char const*, QObject const*, char const*, Qt::ConnectionType) + 483
2 org.mantidproject.MantidPlot 0x000000010f52f22f ProjectionSurface::ProjectionSurface(InstrumentActor const*) + 463 (ProjectionSurface.cpp:44)
3 org.mantidproject.MantidPlot 0x000000010f533987 ProjectionSurface::ProjectionSurface(InstrumentActor const*) + 35
4 org.mantidproject.MantidPlot 0x000000010f53576e Projection3D::Projection3D(InstrumentActor const*, int, int) + 66 (Projection3D.cpp:40)
5 org.mantidproject.MantidPlot 0x000000010f4ca09e InstrumentWindow::setSurfaceType(int) + 2840 (InstrumentWindow.cpp:374)
6 org.mantidproject.MantidPlot 0x000000010f4cac97 InstrumentWindow::setSurfaceType(QString const&) + 497 (InstrumentWindow.cpp:479)
7 org.mantidproject.MantidPlot 0x000000010f4c9107 InstrumentWindow::init(bool, bool, double, double, bool) + 1661 (InstrumentWindow.cpp:218)
8 org.mantidproject.MantidPlot 0x000000010f45cd01 MantidUI::getInstrumentView(QString const&, int) + 1821 (MantidUI.cpp:2023)
9 org.mantidproject.MantidPlot 0x000000010f45d0d4 MantidUI::showMantidInstrument(QString const&) + 42 (MantidUI.cpp:2054)
10 org.mantidproject.MantidPlot 0x000000010f45d449 MantidUI::showMantidInstrumentSelected() + 113 (MantidUI.cpp:2094)
11 org.mantidproject.MantidPlot 0x000000010f5e4e88 MantidUI::qt_static_metacall(QObject*, QMetaObject::Call, int, void**) + 12940 (moc_MantidUI.cxx:510)
```
This doesn't affect the last stable release, so either the problem has been introduced since then, or the compilation without the optimisations is responsible in some way.
| non_infrastructure | instrumentview crash debug mode osx this issue was originally open mantidplot open a workspace in the autotest directory i used one of the inter runs open the instrument view crash thread crashed dispatch queue com apple main thread qtcore qmetaobjectprivate indexofsignalrelative qmetaobject const char const bool qtcore qobject connect qobject const char const qobject const char const qt connectiontype org mantidproject mantidplot projectionsurface projectionsurface instrumentactor const projectionsurface cpp org mantidproject mantidplot projectionsurface projectionsurface instrumentactor const org mantidproject mantidplot instrumentactor const int int cpp org mantidproject mantidplot instrumentwindow setsurfacetype int instrumentwindow cpp org mantidproject mantidplot instrumentwindow setsurfacetype qstring const instrumentwindow cpp org mantidproject mantidplot instrumentwindow init bool bool double double bool instrumentwindow cpp org mantidproject mantidplot mantidui getinstrumentview qstring const int mantidui cpp org mantidproject mantidplot mantidui showmantidinstrument qstring const mantidui cpp org mantidproject mantidplot mantidui showmantidinstrumentselected mantidui cpp org mantidproject mantidplot mantidui qt static metacall qobject qmetaobject call int void moc mantidui cxx this doesn t affect the last stable release so either the problem has been introduced since then or the compilation without the optimisations is responsible in some way | 0 |
161,243 | 6,111,431,454 | IssuesEvent | 2017-06-21 17:01:20 | TerraFusion/basicFusion | https://api.github.com/repos/TerraFusion/basicFusion | opened | Change COMPUTE_TERRA variables | enhancement Medium Priority | The variables need to be changed so that they are initially empty, then users will set the variables to what they need. | 1.0 | Change COMPUTE_TERRA variables - The variables need to be changed so that they are initially empty, then users will set the variables to what they need. | non_infrastructure | change compute terra variables the variables need to be changed so that they are initially empty then users will set the variables to what they need | 0 |
21,372 | 14,542,209,998 | IssuesEvent | 2020-12-15 15:27:11 | robotology/QA | https://api.github.com/repos/robotology/QA | closed | yarp.it 403 Forbidden | infrastructure software | "Forbidden
You don't have permission to access / on this server."
It's been like that for a couple of days... Bad, because yarp.it is cool and it is also linked from https://en.wikipedia.org/wiki/YARP
| 1.0 | yarp.it 403 Forbidden - "Forbidden
You don't have permission to access / on this server."
It's been like that for a couple of days... Bad, because yarp.it is cool and it is also linked from https://en.wikipedia.org/wiki/YARP
| infrastructure | yarp it forbidden forbidden you don t have permission to access on this server it s been like that for a couple of days bad because yarp it is cool and it is also linked from | 1 |
108,777 | 16,822,670,882 | IssuesEvent | 2021-06-17 14:45:33 | idonthaveafifaaddiction/flink | https://api.github.com/repos/idonthaveafifaaddiction/flink | opened | CVE-2014-0114 (High) detected in commons-beanutils-1.8.3.jar, commons-beanutils-1.9.3.jar | security vulnerability | ## CVE-2014-0114 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-beanutils-1.8.3.jar</b>, <b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>
<details><summary><b>commons-beanutils-1.8.3.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: flink/flink-filesystems/flink-oss-fs-hadoop/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-client-2.8.1.jar (Root Library)
- hadoop-common-2.8.1.jar
- commons-configuration-1.7.jar
- :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library)
</details>
<details><summary><b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: flink/flink-filesystems/flink-s3-fs-base/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-common-3.1.0.jar (Root Library)
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.3","packageFilePaths":["/flink-filesystems/flink-oss-fs-hadoop/pom.xml","/flink-filesystems/flink-swift-fs-hadoop/pom.xml","/flink-filesystems/flink-azure-fs-hadoop/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.8.1;org.apache.hadoop:hadoop-common:2.8.1;commons-configuration:commons-configuration:1.7;commons-beanutils:commons-beanutils:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"},{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.9.3","packageFilePaths":["/flink-filesystems/flink-s3-fs-base/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-common:3.1.0;commons-beanutils:commons-beanutils:1.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> --> | True | CVE-2014-0114 (High) detected in commons-beanutils-1.8.3.jar, commons-beanutils-1.9.3.jar - ## CVE-2014-0114 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-beanutils-1.8.3.jar</b>, <b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>
<details><summary><b>commons-beanutils-1.8.3.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: flink/flink-filesystems/flink-oss-fs-hadoop/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar,/home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-client-2.8.1.jar (Root Library)
- hadoop-common-2.8.1.jar
- commons-configuration-1.7.jar
- :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library)
</details>
<details><summary><b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: flink/flink-filesystems/flink-s3-fs-base/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-common-3.1.0.jar (Root Library)
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.3","packageFilePaths":["/flink-filesystems/flink-oss-fs-hadoop/pom.xml","/flink-filesystems/flink-swift-fs-hadoop/pom.xml","/flink-filesystems/flink-azure-fs-hadoop/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.8.1;org.apache.hadoop:hadoop-common:2.8.1;commons-configuration:commons-configuration:1.7;commons-beanutils:commons-beanutils:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"},{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.9.3","packageFilePaths":["/flink-filesystems/flink-s3-fs-base/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-common:3.1.0;commons-beanutils:commons-beanutils:1.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in commons beanutils jar commons beanutils jar cve high severity vulnerability vulnerable libraries commons beanutils jar commons beanutils jar commons beanutils jar beanutils provides an easy to use but flexible wrapper around reflection and introspection path to dependency file flink flink filesystems flink oss fs hadoop pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar home wss scanner repository commons beanutils commons beanutils commons beanutils jar home wss scanner repository commons beanutils commons beanutils commons beanutils jar home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy hadoop client jar root library hadoop common jar commons configuration jar x commons beanutils jar vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file flink flink filesystems flink fs base pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy hadoop common jar root library x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils org apache struts core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache hadoop hadoop client org apache hadoop hadoop common commons configuration commons configuration commons beanutils commons beanutils isminimumfixversionavailable true minimumfixversion commons beanutils commons beanutils org apache struts core packagetype java groupid commons beanutils packagename commons beanutils packageversion packagefilepaths istransitivedependency true dependencytree org apache hadoop hadoop common commons beanutils commons beanutils isminimumfixversionavailable true minimumfixversion commons beanutils commons beanutils org apache struts core basebranches vulnerabilityidentifier cve vulnerabilitydetails apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts vulnerabilityurl | 0 |
32,802 | 27,004,545,589 | IssuesEvent | 2023-02-10 10:33:42 | zer0Kerbal/Dreamer | https://api.github.com/repos/zer0Kerbal/Dreamer | closed | Repo | semantic: adoption type: infrastructure | - [x] Create Local
- [x] folder structure
- [x] .github folder contents
- [x] .gitattributes
- [x] .gitignore
- [x] _release.json
- [x] readme.md
- [x] .version file
- [x] _releasenotes
- [x] Archive original release
- [x] _build
- [x] _buildJSON
- [x] _buildRelease
- [x] _buildDeploy
- [x] copy labels
- [x] create repo's Adoption and Initial Release project
- [x] create User Project - Adoption | 1.0 | Repo - - [x] Create Local
- [x] folder structure
- [x] .github folder contents
- [x] .gitattributes
- [x] .gitignore
- [x] _release.json
- [x] readme.md
- [x] .version file
- [x] _releasenotes
- [x] Archive original release
- [x] _build
- [x] _buildJSON
- [x] _buildRelease
- [x] _buildDeploy
- [x] copy labels
- [x] create repo's Adoption and Initial Release project
- [x] create User Project - Adoption | infrastructure | repo create local folder structure github folder contents gitattributes gitignore release json readme md version file releasenotes archive original release build buildjson buildrelease builddeploy copy labels create repo s adoption and initial release project create user project adoption | 1 |
16,924 | 12,152,149,494 | IssuesEvent | 2020-04-24 21:31:20 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Enable Pulp Repositories as a cluster upgrade source | Infrastructure closed feature | https://trello.com/c/F1FE0a1s/56-enable-pulp-repositories-as-a-cluster-upgrade-source
Enable a pulp source for cluster upgrades for versioning our upgrades.
| 1.0 | Enable Pulp Repositories as a cluster upgrade source - https://trello.com/c/F1FE0a1s/56-enable-pulp-repositories-as-a-cluster-upgrade-source
Enable a pulp source for cluster upgrades for versioning our upgrades.
| infrastructure | enable pulp repositories as a cluster upgrade source enable a pulp source for cluster upgrades for versioning our upgrades | 1 |
10,344 | 8,507,078,390 | IssuesEvent | 2018-10-30 18:07:20 | coddingtonbear/inthe.am | https://api.github.com/repos/coddingtonbear/inthe.am | closed | 500 Server Error when logging in with Google for first time | infrastructure maybe fixed? production incident | Never used taskwarrior or inthe.am until today. Tried to login with Google (I have never logged in to inthe.am before), and I am greeted with a white page that just says "Server Error (500)" | 1.0 | 500 Server Error when logging in with Google for first time - Never used taskwarrior or inthe.am until today. Tried to login with Google (I have never logged in to inthe.am before), and I am greeted with a white page that just says "Server Error (500)" | infrastructure | server error when logging in with google for first time never used taskwarrior or inthe am until today tried to login with google i have never logged in to inthe am before and i am greeted with a white page that just says server error | 1 |
886 | 8,597,479,777 | IssuesEvent | 2018-11-15 18:48:38 | mozilla-mobile/reference-browser | https://api.github.com/repos/mozilla-mobile/reference-browser | closed | Sign builds of reference browser | 🤖 automation | Until we get the real pipeline with upload to Google Play (~ December) we can setup signing on taskcluster with a dummy key from the secrets service. This will allow us to store signed APKs and make them available from a taskcluster index URL. | 1.0 | Sign builds of reference browser - Until we get the real pipeline with upload to Google Play (~ December) we can setup signing on taskcluster with a dummy key from the secrets service. This will allow us to store signed APKs and make them available from a taskcluster index URL. | non_infrastructure | sign builds of reference browser until we get the real pipeline with upload to google play december we can setup signing on taskcluster with a dummy key from the secrets service this will allow us to store signed apks and make them available from a taskcluster index url | 0 |
142,834 | 5,477,748,041 | IssuesEvent | 2017-03-12 11:46:46 | Rsl1122/Plan-PlayerAnalytics | https://api.github.com/repos/Rsl1122/Plan-PlayerAnalytics | closed | SQL Injection Possible with Nicknames | Priority: LOW Security Vulnerability | **Plan Version:** Affected versions: 2.0.0 onwards
**Server Version:** -
**Database Type:** both
**Description:**
SQL Injection is possible by changing nickname, although limited to 30 chars before 2.7.0 or 75 after.
| 1.0 | SQL Injection Possible with Nicknames - **Plan Version:** Affected versions: 2.0.0 onwards
**Server Version:** -
**Database Type:** both
**Description:**
SQL Injection is possible by changing nickname, although limited to 30 chars before 2.7.0 or 75 after.
| non_infrastructure | sql injection possible with nicknames plan version affected versions onwards server version database type both description sql injection is possible by changing nickname although limited to chars before or after | 0 |
6,548 | 6,507,302,288 | IssuesEvent | 2017-08-24 12:46:42 | AdguardTeam/AdguardFilters | https://api.github.com/repos/AdguardTeam/AdguardFilters | closed | Add BarbBlock filter list | Infrastructure | [BarbBlock](https://ssl.bblck.me) is a content blocking list that blacklists sites that have used spurious and invalid DMCA takedowns to force removal from other content blocking lists. Originally a Chrome extension, it has [recently been made available as a block list](https://github.com/paulgb/BarbBlock/issues/5).
This list is available in [both ABP and uBlock Origin formats](https://ssl.bblck.me/install). Would you please add this list so it is available in the AdGuard configuration UI and can be enabled by users via a checkbox in the usual filter selection settings screen?
I am not certain if this is the appropriate place to request filter list additions, but I could not locate a more appropriate repository. If there is a better place to make this request, please let me know. Thank you! | 1.0 | Add BarbBlock filter list - [BarbBlock](https://ssl.bblck.me) is a content blocking list that blacklists sites that have used spurious and invalid DMCA takedowns to force removal from other content blocking lists. Originally a Chrome extension, it has [recently been made available as a block list](https://github.com/paulgb/BarbBlock/issues/5).
This list is available in [both ABP and uBlock Origin formats](https://ssl.bblck.me/install). Would you please add this list so it is available in the AdGuard configuration UI and can be enabled by users via a checkbox in the usual filter selection settings screen?
I am not certain if this is the appropriate place to request filter list additions, but I could not locate a more appropriate repository. If there is a better place to make this request, please let me know. Thank you! | infrastructure | add barbblock filter list is a content blocking list that blacklists sites that have used spurious and invalid dmca takedowns to force removal from other content blocking lists originally a chrome extension it has this list is available in would you please add this list so it is available in the adguard configuration ui and can be enabled by users via a checkbox in the usual filter selection settings screen i am not certain if this is the appropriate place to request filter list additions but i could not locate a more appropriate repository if there is a better place to make this request please let me know thank you | 1 |
52,881 | 13,079,614,087 | IssuesEvent | 2020-08-01 04:04:30 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | opened | tensorflow-io 0.14.0 requires tensorflow<2.3.0,>=2.2.0, but you'll have tensorflow 2.3.0 which is incompatible | type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04
- TensorFlow installed from (source or binary): source
- TensorFlow version: 2.3.0
- Python version:3.8.2
- Installed using virtualenv? pip? conda?: pip
- CUDA/cuDNN version: 10.1
- GPU model and memory: 1650 4GB
**Describe the problem**
Cant use tensorflow-io , if I try to install tensorflow-io with pip tensorflow downgrades please fix this dependency issue.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
pip install tensorflow
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
pip install tensorflow
Defaulting to user installation because normal site-packages is not writeable
Collecting tensorflow
Using cached tensorflow-2.3.0-cp38-cp38-manylinux2010_x86_64.whl (320.5 MB)
Requirement already satisfied: wrapt>=1.11.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.12.1)
Requirement already satisfied: gast==0.3.3 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.3.3)
Requirement already satisfied: numpy<1.19.0,>=1.16.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.18.4)
Requirement already satisfied: tensorflow-estimator<2.4.0,>=2.3.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.3.0)
Requirement already satisfied: termcolor>=1.1.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.1.0)
Requirement already satisfied: astunparse==1.6.3 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.6.3)
Requirement already satisfied: google-pasta>=0.1.8 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.2.0)
Requirement already satisfied: keras-preprocessing<1.2,>=1.1.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.1.2)
Requirement already satisfied: protobuf>=3.9.2 in ./.local/lib/python3.8/site-packages (from tensorflow) (3.12.2)
Requirement already satisfied: h5py<2.11.0,>=2.10.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.10.0)
Requirement already satisfied: tensorboard<3,>=2.3.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.3.0)
Requirement already satisfied: scipy==1.4.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.4.1)
Requirement already satisfied: grpcio>=1.8.6 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.29.0)
Requirement already satisfied: six>=1.12.0 in /usr/lib/python3/dist-packages (from tensorflow) (1.14.0)
Requirement already satisfied: wheel>=0.26 in /usr/lib/python3/dist-packages (from tensorflow) (0.34.2)
Requirement already satisfied: opt-einsum>=2.3.2 in ./.local/lib/python3.8/site-packages (from tensorflow) (3.2.1)
Requirement already satisfied: absl-py>=0.7.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.9.0)
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from protobuf>=3.9.2->tensorflow) (45.2.0)
Requirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (0.4.1)
Requirement already satisfied: google-auth<2,>=1.6.3 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.16.0)
Requirement already satisfied: markdown>=2.6.8 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (3.2.2)
Requirement already satisfied: requests<3,>=2.21.0 in /usr/lib/python3/dist-packages (from tensorboard<3,>=2.3.0->tensorflow) (2.22.0)
Requirement already satisfied: tensorboard-plugin-wit>=1.6.0 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.6.0.post3)
Requirement already satisfied: werkzeug>=0.11.15 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.0.1)
Requirement already satisfied: requests-oauthlib>=0.7.0 in ./.local/lib/python3.8/site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard<3,>=2.3.0->tensorflow) (1.3.0)
Requirement already satisfied: rsa<4.1,>=3.1.4 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (4.0)
Requirement already satisfied: pyasn1-modules>=0.2.1 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (0.2.8)
Requirement already satisfied: cachetools<5.0,>=2.0.0 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (4.1.0)
Requirement already satisfied: oauthlib>=3.0.0 in /usr/lib/python3/dist-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard<3,>=2.3.0->tensorflow) (3.1.0)
Requirement already satisfied: pyasn1>=0.1.3 in ./.local/lib/python3.8/site-packages (from rsa<4.1,>=3.1.4->google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (0.4.8)
Installing collected packages: tensorflow
ERROR: After October 2020 you may experience errors when installing or updating packages. This is because pip will change the way that it resolves dependency conflicts.
We recommend you use --use-feature=2020-resolver to test your packages with the new resolver before it becomes the default.
tensorflow-io 0.14.0 requires tensorflow<2.3.0,>=2.2.0, but you'll have tensorflow 2.3.0 which is incompatible | 1.0 | tensorflow-io 0.14.0 requires tensorflow<2.3.0,>=2.2.0, but you'll have tensorflow 2.3.0 which is incompatible - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 20.04
- TensorFlow installed from (source or binary): source
- TensorFlow version: 2.3.0
- Python version:3.8.2
- Installed using virtualenv? pip? conda?: pip
- CUDA/cuDNN version: 10.1
- GPU model and memory: 1650 4GB
**Describe the problem**
Cant use tensorflow-io , if I try to install tensorflow-io with pip tensorflow downgrades please fix this dependency issue.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
pip install tensorflow
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
pip install tensorflow
Defaulting to user installation because normal site-packages is not writeable
Collecting tensorflow
Using cached tensorflow-2.3.0-cp38-cp38-manylinux2010_x86_64.whl (320.5 MB)
Requirement already satisfied: wrapt>=1.11.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.12.1)
Requirement already satisfied: gast==0.3.3 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.3.3)
Requirement already satisfied: numpy<1.19.0,>=1.16.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.18.4)
Requirement already satisfied: tensorflow-estimator<2.4.0,>=2.3.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.3.0)
Requirement already satisfied: termcolor>=1.1.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.1.0)
Requirement already satisfied: astunparse==1.6.3 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.6.3)
Requirement already satisfied: google-pasta>=0.1.8 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.2.0)
Requirement already satisfied: keras-preprocessing<1.2,>=1.1.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.1.2)
Requirement already satisfied: protobuf>=3.9.2 in ./.local/lib/python3.8/site-packages (from tensorflow) (3.12.2)
Requirement already satisfied: h5py<2.11.0,>=2.10.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.10.0)
Requirement already satisfied: tensorboard<3,>=2.3.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (2.3.0)
Requirement already satisfied: scipy==1.4.1 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.4.1)
Requirement already satisfied: grpcio>=1.8.6 in ./.local/lib/python3.8/site-packages (from tensorflow) (1.29.0)
Requirement already satisfied: six>=1.12.0 in /usr/lib/python3/dist-packages (from tensorflow) (1.14.0)
Requirement already satisfied: wheel>=0.26 in /usr/lib/python3/dist-packages (from tensorflow) (0.34.2)
Requirement already satisfied: opt-einsum>=2.3.2 in ./.local/lib/python3.8/site-packages (from tensorflow) (3.2.1)
Requirement already satisfied: absl-py>=0.7.0 in ./.local/lib/python3.8/site-packages (from tensorflow) (0.9.0)
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from protobuf>=3.9.2->tensorflow) (45.2.0)
Requirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (0.4.1)
Requirement already satisfied: google-auth<2,>=1.6.3 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.16.0)
Requirement already satisfied: markdown>=2.6.8 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (3.2.2)
Requirement already satisfied: requests<3,>=2.21.0 in /usr/lib/python3/dist-packages (from tensorboard<3,>=2.3.0->tensorflow) (2.22.0)
Requirement already satisfied: tensorboard-plugin-wit>=1.6.0 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.6.0.post3)
Requirement already satisfied: werkzeug>=0.11.15 in ./.local/lib/python3.8/site-packages (from tensorboard<3,>=2.3.0->tensorflow) (1.0.1)
Requirement already satisfied: requests-oauthlib>=0.7.0 in ./.local/lib/python3.8/site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard<3,>=2.3.0->tensorflow) (1.3.0)
Requirement already satisfied: rsa<4.1,>=3.1.4 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (4.0)
Requirement already satisfied: pyasn1-modules>=0.2.1 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (0.2.8)
Requirement already satisfied: cachetools<5.0,>=2.0.0 in ./.local/lib/python3.8/site-packages (from google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (4.1.0)
Requirement already satisfied: oauthlib>=3.0.0 in /usr/lib/python3/dist-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard<3,>=2.3.0->tensorflow) (3.1.0)
Requirement already satisfied: pyasn1>=0.1.3 in ./.local/lib/python3.8/site-packages (from rsa<4.1,>=3.1.4->google-auth<2,>=1.6.3->tensorboard<3,>=2.3.0->tensorflow) (0.4.8)
Installing collected packages: tensorflow
ERROR: After October 2020 you may experience errors when installing or updating packages. This is because pip will change the way that it resolves dependency conflicts.
We recommend you use --use-feature=2020-resolver to test your packages with the new resolver before it becomes the default.
tensorflow-io 0.14.0 requires tensorflow<2.3.0,>=2.2.0, but you'll have tensorflow 2.3.0 which is incompatible | non_infrastructure | tensorflow io requires tensorflow but you ll have tensorflow which is incompatible please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu linux ubuntu tensorflow installed from source or binary source tensorflow version python version installed using virtualenv pip conda pip cuda cudnn version gpu model and memory describe the problem cant use tensorflow io if i try to install tensorflow io with pip tensorflow downgrades please fix this dependency issue provide the exact sequence of commands steps that you executed before running into the problem pip install tensorflow any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached pip install tensorflow defaulting to user installation because normal site packages is not writeable collecting tensorflow using cached tensorflow whl mb requirement already satisfied wrapt in local lib site packages from tensorflow requirement already satisfied gast in local lib site packages from tensorflow requirement already satisfied numpy in local lib site packages from tensorflow requirement already satisfied tensorflow estimator in local lib site packages from tensorflow requirement already satisfied termcolor in local lib site packages from tensorflow requirement already satisfied astunparse in local lib site packages from tensorflow requirement already satisfied google pasta in local lib site packages from tensorflow requirement already satisfied keras preprocessing in local lib site packages from tensorflow requirement already satisfied protobuf in local lib site packages from tensorflow requirement already satisfied in local lib site packages from tensorflow requirement already satisfied tensorboard in local lib site packages from tensorflow requirement already satisfied scipy in local lib site packages from tensorflow requirement already satisfied grpcio in local lib site packages from tensorflow requirement already satisfied six in usr lib dist packages from tensorflow requirement already satisfied wheel in usr lib dist packages from tensorflow requirement already satisfied opt einsum in local lib site packages from tensorflow requirement already satisfied absl py in local lib site packages from tensorflow requirement already satisfied setuptools in usr lib dist packages from protobuf tensorflow requirement already satisfied google auth oauthlib in local lib site packages from tensorboard tensorflow requirement already satisfied google auth in local lib site packages from tensorboard tensorflow requirement already satisfied markdown in local lib site packages from tensorboard tensorflow requirement already satisfied requests in usr lib dist packages from tensorboard tensorflow requirement already satisfied tensorboard plugin wit in local lib site packages from tensorboard tensorflow requirement already satisfied werkzeug in local lib site packages from tensorboard tensorflow requirement already satisfied requests oauthlib in local lib site packages from google auth oauthlib tensorboard tensorflow requirement already satisfied rsa in local lib site packages from google auth tensorboard tensorflow requirement already satisfied modules in local lib site packages from google auth tensorboard tensorflow requirement already satisfied cachetools in local lib site packages from google auth tensorboard tensorflow requirement already satisfied oauthlib in usr lib dist packages from requests oauthlib google auth oauthlib tensorboard tensorflow requirement already satisfied in local lib site packages from rsa google auth tensorboard tensorflow installing collected packages tensorflow error after october you may experience errors when installing or updating packages this is because pip will change the way that it resolves dependency conflicts we recommend you use use feature resolver to test your packages with the new resolver before it becomes the default tensorflow io requires tensorflow but you ll have tensorflow which is incompatible | 0 |
29,635 | 24,126,647,272 | IssuesEvent | 2022-09-21 01:31:02 | astropy/pyvo | https://api.github.com/repos/astropy/pyvo | closed | BUG: test should NOT fail with missing optional dependencies | bug infrastructure | The docs and configs indicate that both `mimeparse` and `pillow` are optional dependencies, yet the test are not taking this into account (e.g. the test requiring these packages should be skipped when they are not installed rather than failing the test suite) | 1.0 | BUG: test should NOT fail with missing optional dependencies - The docs and configs indicate that both `mimeparse` and `pillow` are optional dependencies, yet the test are not taking this into account (e.g. the test requiring these packages should be skipped when they are not installed rather than failing the test suite) | infrastructure | bug test should not fail with missing optional dependencies the docs and configs indicate that both mimeparse and pillow are optional dependencies yet the test are not taking this into account e g the test requiring these packages should be skipped when they are not installed rather than failing the test suite | 1 |
13,630 | 10,351,386,767 | IssuesEvent | 2019-09-05 06:43:50 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Introduce initial data provider implementation | Priority: Essential Status: In implementation Type: Improvement Where: Infrastructure | This involves introducing an initial implementation of #6 for investigations and to test out different architectural strategies.
This is tracking implementing an initial iteration of the system, whereas #6 is tracking solving this generically with a thought-out design document. | 1.0 | Introduce initial data provider implementation - This involves introducing an initial implementation of #6 for investigations and to test out different architectural strategies.
This is tracking implementing an initial iteration of the system, whereas #6 is tracking solving this generically with a thought-out design document. | infrastructure | introduce initial data provider implementation this involves introducing an initial implementation of for investigations and to test out different architectural strategies this is tracking implementing an initial iteration of the system whereas is tracking solving this generically with a thought out design document | 1 |
23,354 | 16,088,129,448 | IssuesEvent | 2021-04-26 13:44:55 | airyhq/airy | https://api.github.com/repos/airyhq/airy | closed | AWS provider: Remove hardcoded availability zones | cli infrastructure | Currently our `createSubnet` function takes a hard-coded availability zones (us-east-1a and us-east-1b). Therefore, the provider will fail if the user has another default region.
We need to change this so that new vpc and subnets can be created in different regions. | 1.0 | AWS provider: Remove hardcoded availability zones - Currently our `createSubnet` function takes a hard-coded availability zones (us-east-1a and us-east-1b). Therefore, the provider will fail if the user has another default region.
We need to change this so that new vpc and subnets can be created in different regions. | infrastructure | aws provider remove hardcoded availability zones currently our createsubnet function takes a hard coded availability zones us east and us east therefore the provider will fail if the user has another default region we need to change this so that new vpc and subnets can be created in different regions | 1 |
52,264 | 12,907,685,988 | IssuesEvent | 2020-07-15 05:42:41 | vmware/singleton | https://api.github.com/repos/vmware/singleton | closed | [BUG] [Service] missing settings.gradle under tools\tool-cldr-extractor | area/build area/service kind/bug priority/medium | **Describe the bug**
When try to compile a formatting pattern jar package under *tools\tool-cldr-extractor*, it always fails, due to miss **settings.gradle** file.
**To Reproduce**
Steps to reproduce the behavior:
1. Git clone the latest Singleton Service code
2. Go to *g11n-ws\tools\tool-cldr-extractor*
3. Run `gradle build`
4. See error
```
Project directory '.\singleton\g11n-ws\tools\tool-cldr-extractor' is not part of the build defined by settings file 'D:\git\github\singleton\g11n-ws\settings.gradle'. If this is an unrelated build, it must have its own settings file.
```
**Expected behavior**
Everything runs well, and the formmating pattern jar package is generated well.
**Screenshots**
NA
**Desktop (please complete the following information):**
NA
**Additional context**
NA | 1.0 | [BUG] [Service] missing settings.gradle under tools\tool-cldr-extractor - **Describe the bug**
When try to compile a formatting pattern jar package under *tools\tool-cldr-extractor*, it always fails, due to miss **settings.gradle** file.
**To Reproduce**
Steps to reproduce the behavior:
1. Git clone the latest Singleton Service code
2. Go to *g11n-ws\tools\tool-cldr-extractor*
3. Run `gradle build`
4. See error
```
Project directory '.\singleton\g11n-ws\tools\tool-cldr-extractor' is not part of the build defined by settings file 'D:\git\github\singleton\g11n-ws\settings.gradle'. If this is an unrelated build, it must have its own settings file.
```
**Expected behavior**
Everything runs well, and the formmating pattern jar package is generated well.
**Screenshots**
NA
**Desktop (please complete the following information):**
NA
**Additional context**
NA | non_infrastructure | missing settings gradle under tools tool cldr extractor describe the bug when try to compile a formatting pattern jar package under tools tool cldr extractor it always fails due to miss settings gradle file to reproduce steps to reproduce the behavior git clone the latest singleton service code go to ws tools tool cldr extractor run gradle build see error project directory singleton ws tools tool cldr extractor is not part of the build defined by settings file d git github singleton ws settings gradle if this is an unrelated build it must have its own settings file expected behavior everything runs well and the formmating pattern jar package is generated well screenshots na desktop please complete the following information na additional context na | 0 |
87,578 | 10,549,506,708 | IssuesEvent | 2019-10-03 08:53:55 | ebinxavier/rubiksCube | https://api.github.com/repos/ebinxavier/rubiksCube | opened | Enhancements + Application Testing + Browser/Mobile Compatibility Testing | documentation enhancement good first issue help wanted | Hi Folks,
Let's make sure that this application runs in all platform smoothly. Also requesting your valuable comments and suggestions regarding the usability of this application.
Play with the [RubiksCube](https://ebinxavier.github.io/rubiksCube) and commet below. | 1.0 | Enhancements + Application Testing + Browser/Mobile Compatibility Testing - Hi Folks,
Let's make sure that this application runs in all platform smoothly. Also requesting your valuable comments and suggestions regarding the usability of this application.
Play with the [RubiksCube](https://ebinxavier.github.io/rubiksCube) and commet below. | non_infrastructure | enhancements application testing browser mobile compatibility testing hi folks let s make sure that this application runs in all platform smoothly also requesting your valuable comments and suggestions regarding the usability of this application play with the and commet below | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.