Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
47,055 | 10,023,163,741 | IssuesEvent | 2019-07-16 18:30:00 | phetsims/fluid-pressure-and-flow | https://api.github.com/repos/phetsims/fluid-pressure-and-flow | opened | factor out duplicated model code | dev:code-review | Noted during #323 evaluation.
There is a lot of duplicated code in the top-level model classes: `UnderPressureModel`, `FlowModel`, and `WaterTowerModel`. It's most obvious when examining the `reset` methods, but you can also see it in the constructors. Investigate factoring out a base class. | 1.0 | factor out duplicated model code - Noted during #323 evaluation.
There is a lot of duplicated code in the top-level model classes: `UnderPressureModel`, `FlowModel`, and `WaterTowerModel`. It's most obvious when examining the `reset` methods, but you can also see it in the constructors. Investigate factoring out a base class. | non_defect | factor out duplicated model code noted during evaluation there is a lot of duplicated code in the top level model classes underpressuremodel flowmodel and watertowermodel it s most obvious when examining the reset methods but you can also see it in the constructors investigate factoring out a base class | 0 |
77,173 | 26,822,369,712 | IssuesEvent | 2023-02-02 10:22:22 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Avatar overlays own message in threads panel | T-Defect X-Regression S-Minor O-Frequent A-Threads | ### Steps to reproduce
1. Reply in a thread
2. Open thread panel in the all threads view
Note this works fine if someone else has sent the last reply in the thread
### Outcome
#### What did you expect?
I can see my message next to my avatar, not under it.
#### What happened instead?
My avatar overlays my own message:

### Operating system
_No response_
### Browser information
Chromium 109.0.5414.74 (Official Build) Arch Linux (64-bit)
### URL for webapp
develop.element.io
### Application version
Element version: 50f2b532e907-react-5095bc0427a1-js-4f918f684e9d Olm version: 3.2.12
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | Avatar overlays own message in threads panel - ### Steps to reproduce
1. Reply in a thread
2. Open thread panel in the all threads view
Note this works fine if someone else has sent the last reply in the thread
### Outcome
#### What did you expect?
I can see my message next to my avatar, not under it.
#### What happened instead?
My avatar overlays my own message:

### Operating system
_No response_
### Browser information
Chromium 109.0.5414.74 (Official Build) Arch Linux (64-bit)
### URL for webapp
develop.element.io
### Application version
Element version: 50f2b532e907-react-5095bc0427a1-js-4f918f684e9d Olm version: 3.2.12
### Homeserver
matrix.org
### Will you send logs?
No | defect | avatar overlays own message in threads panel steps to reproduce reply in a thread open thread panel in the all threads view note this works fine if someone else has sent the last reply in the thread outcome what did you expect i can see my message next to my avatar not under it what happened instead my avatar overlays my own message operating system no response browser information chromium official build arch linux bit url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs no | 1 |
64,669 | 18,789,816,132 | IssuesEvent | 2021-11-08 15:42:53 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [🐛 Bug]: Java binding does not support calling quit in shutdown hook | C-java I-defect | ### What happened?
The Java binding does not support calling quit in shutdown hook since 4.0.0-beta-2.
It appears that the java binding has a shutdown hook that closes netty clients and this hook is executed prior to a user-registered hook designed to quit the driver.
### How can we reproduce the issue?
```shell
static void main(String[] args) {
WebDriver driver = new FirefoxDriver();
Runtime.getRuntime().addShutdownHook(new Thread(driver::quit));
driver.get("http://localhost/");
}
```
### Relevant log output
```shell
The browser remains open.
```
### Operating System
Windows 10
### Selenium version
4.0.0-beta-2 and later
### What are the browser(s) and version(s) where you see this issue?
Firefox
### What are the browser driver(s) and version(s) where you see this issue?
-
### Are you using Selenium Grid?
- | 1.0 | [🐛 Bug]: Java binding does not support calling quit in shutdown hook - ### What happened?
The Java binding does not support calling quit in shutdown hook since 4.0.0-beta-2.
It appears that the java binding has a shutdown hook that closes netty clients and this hook is executed prior to a user-registered hook designed to quit the driver.
### How can we reproduce the issue?
```shell
static void main(String[] args) {
WebDriver driver = new FirefoxDriver();
Runtime.getRuntime().addShutdownHook(new Thread(driver::quit));
driver.get("http://localhost/");
}
```
### Relevant log output
```shell
The browser remains open.
```
### Operating System
Windows 10
### Selenium version
4.0.0-beta-2 and later
### What are the browser(s) and version(s) where you see this issue?
Firefox
### What are the browser driver(s) and version(s) where you see this issue?
-
### Are you using Selenium Grid?
- | defect | java binding does not support calling quit in shutdown hook what happened the java binding does not support calling quit in shutdown hook since beta it appears that the java binding has a shutdown hook that closes netty clients and this hook is executed prior to a user registered hook designed to quit the driver how can we reproduce the issue shell static void main string args webdriver driver new firefoxdriver runtime getruntime addshutdownhook new thread driver quit driver get relevant log output shell the browser remains open operating system windows selenium version beta and later what are the browser s and version s where you see this issue firefox what are the browser driver s and version s where you see this issue are you using selenium grid | 1 |
24,882 | 4,117,872,459 | IssuesEvent | 2016-06-08 09:14:54 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | New Element class | breaking-change defect | 1. Create a new **Element** class
2. The members of **Element** class a mixed in with **HTMLElement**. They need to be separated.
3. **HTMLElement** should inherit from **Element**, not **Node**
Related to #1456 and #1371. | 1.0 | New Element class - 1. Create a new **Element** class
2. The members of **Element** class a mixed in with **HTMLElement**. They need to be separated.
3. **HTMLElement** should inherit from **Element**, not **Node**
Related to #1456 and #1371. | defect | new element class create a new element class the members of element class a mixed in with htmlelement they need to be separated htmlelement should inherit from element not node related to and | 1 |
49,185 | 10,327,453,144 | IssuesEvent | 2019-09-02 06:59:52 | eiffel-community/eiffel-intelligence | https://api.github.com/repos/eiffel-community/eiffel-intelligence | opened | Download endpoint changes | code improvements enhancement | ### Description
This issue is part of #291.
The download endpoint uses a verb as the root which is against RESTful conventions. The resources will now be grouped under the /templates endpoint.
### Motivation
Part of the endpoint restructure.
### Exemplification
These resources are now grouped under /templates and because of this the resources no longer need to append this to their names.
Method | Old | New
-- | -- | --
GET | /download | /templates
GET | /download/subscriptionsTemplate | /templates/subscriptions
GET | /download/rulesTemplate | /templates/rules
GET | /download/eventsTemplate | /templates/events
### Benefits
This makes all endpoints look and behave the same way.
### Possible Drawbacks
| 1.0 | Download endpoint changes - ### Description
This issue is part of #291.
The download endpoint uses a verb as the root which is against RESTful conventions. The resources will now be grouped under the /templates endpoint.
### Motivation
Part of the endpoint restructure.
### Exemplification
These resources are now grouped under /templates and because of this the resources no longer need to append this to their names.
Method | Old | New
-- | -- | --
GET | /download | /templates
GET | /download/subscriptionsTemplate | /templates/subscriptions
GET | /download/rulesTemplate | /templates/rules
GET | /download/eventsTemplate | /templates/events
### Benefits
This makes all endpoints look and behave the same way.
### Possible Drawbacks
| non_defect | download endpoint changes description this issue is part of the download endpoint uses a verb as the root which is against restful conventions the resources will now be grouped under the templates endpoint motivation part of the endpoint restructure exemplification these resources are now grouped under templates and because of this the resources no longer need to append this to their names method old new get download templates get download subscriptionstemplate templates subscriptions get download rulestemplate templates rules get download eventstemplate templates events benefits this makes all endpoints look and behave the same way possible drawbacks | 0 |
623,365 | 19,666,009,870 | IssuesEvent | 2022-01-10 22:37:32 | ScottUK/ladojrp-issues | https://api.github.com/repos/ScottUK/ladojrp-issues | closed | New Billboard at Sandy (east joshua!) | Class: enhancement Priority: medium Scope: maps | (SUGGESTION) For someone who has been playing Five M for a while, I think it would be a cool idea to add a new template in the Billboard at East Joshua. It will add some spice/touch to the server! There is lots of examples that can be used | 1.0 | New Billboard at Sandy (east joshua!) - (SUGGESTION) For someone who has been playing Five M for a while, I think it would be a cool idea to add a new template in the Billboard at East Joshua. It will add some spice/touch to the server! There is lots of examples that can be used | non_defect | new billboard at sandy east joshua suggestion for someone who has been playing five m for a while i think it would be a cool idea to add a new template in the billboard at east joshua it will add some spice touch to the server there is lots of examples that can be used | 0 |
85,520 | 10,442,881,424 | IssuesEvent | 2019-09-18 13:54:15 | legion-platform/legion | https://api.github.com/repos/legion-platform/legion | closed | Add detailed documentation about feedback gathering | 1.0.0 documentation feature | As a user of Legion platform I want to have information how I can use feedback mechanism and how it works under the hood.
Also examples should be presented. | 1.0 | Add detailed documentation about feedback gathering - As a user of Legion platform I want to have information how I can use feedback mechanism and how it works under the hood.
Also examples should be presented. | non_defect | add detailed documentation about feedback gathering as a user of legion platform i want to have information how i can use feedback mechanism and how it works under the hood also examples should be presented | 0 |
73,050 | 24,422,661,077 | IssuesEvent | 2022-10-05 22:00:01 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | progressBar not yielding to poll for long running process | :lady_beetle: defect :bangbang: needs-triage | ### Describe the bug
A page contains both a poll and progressBar.
On page load, poll works as expected.
When a long server process is started, progress bar shows the progression, but the polling stops.
### Reproducer
https://github.com/mores/primefaces-test/tree/progressBarYield
### Expected behavior
I expect the polling to continue to poll while the progressBar is progressing.
### PrimeFaces edition
Elite
### PrimeFaces version
7.0.29
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.9
### Java version
1.8
### Browser(s)
_No response_ | 1.0 | progressBar not yielding to poll for long running process - ### Describe the bug
A page contains both a poll and progressBar.
On page load, poll works as expected.
When a long server process is started, progress bar shows the progression, but the polling stops.
### Reproducer
https://github.com/mores/primefaces-test/tree/progressBarYield
### Expected behavior
I expect the polling to continue to poll while the progressBar is progressing.
### PrimeFaces edition
Elite
### PrimeFaces version
7.0.29
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.9
### Java version
1.8
### Browser(s)
_No response_ | defect | progressbar not yielding to poll for long running process describe the bug a page contains both a poll and progressbar on page load poll works as expected when a long server process is started progress bar shows the progression but the polling stops reproducer expected behavior i expect the polling to continue to poll while the progressbar is progressing primefaces edition elite primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response | 1 |
69,499 | 22,391,824,485 | IssuesEvent | 2022-06-17 08:28:50 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms | I-defect needs-triaging | ### What happened?
Following issue happens for FireFox and Google Chrome browsers/drivers for latest versions. I found same issues from past years, where it was recommended to update to latest versions of selenium/driver, however I already have latest versions.
**Test:**
```
import org.openqa.selenium.By;
import org.openqa.selenium.WebDriver;
import org.openqa.selenium.WebElement;
import org.openqa.selenium.firefox.FirefoxDriver;
public class testSel {
public static void main(String[] args){
System.out.println("1");
System.setProperty("webdriver.gecko.driver", "/Users/atkeshelashvili/Downloads/geckodriver");
WebDriver driver=new FirefoxDriver();
System.out.println("2");
driver.get("http://demo.guru99.com/");
System.out.println("3");
WebElement element=driver.findElement(By.xpath("//input[@name='emailid']"));
element.sendKeys("abc@gmail.com");
WebElement button=driver.findElement(By.xpath("//input[@name='btnLogin']"));
button.click();
}
}
```
### How can we reproduce the issue?
```shell
**Upload following versions:**
FireFox 101.0.1 (64 bit)
Selenium Web Driver 4.2.2 selenium-java-4.2.2.zip
Geckodriver 0.31.0
System macOS Monterey (M1)
Same happens for Google Chrome 102.0.5005.61.
```
### Relevant log output
```shell
**Logs:**
org.openqa.selenium.firefox.NotConnectedException: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms. Firefox console output:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:118)
at org.openqa.selenium.firefox.FirefoxDriver.startClient(FirefoxDriver.java:246)
at org.openqa.selenium.remote.RemoteWebDriver.<init>(RemoteWebDriver.java:114)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:191)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:186)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:182)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:95)
at testSel.main(testSel.java:10)
Exception in thread "main" org.openqa.selenium.WebDriverException: Failed to connect to binary FirefoxBinary(/Applications/Firefox.app/Contents/MacOS/firefox-bin) on port 7055; process output follows:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
Build info: version: '2.45.0', revision: '5017cb8e7ca8e37638dc3091b2440b90a1d8686f', time: '2015-02-27 09:10:26'
System info: host: 'ip-192-168-0-198.eu-central-1.compute.internal', ip: '192.168.0.198', os.name: 'Mac OS X', os.arch: 'aarch64', os.version: '12.1', java.version: '11.0.12'
Driver info: driver.version: FirefoxDriver
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:130)
at org.openqa.selenium.firefox.FirefoxDriver.startClient(FirefoxDriver.java:246)
at org.openqa.selenium.remote.RemoteWebDriver.<init>(RemoteWebDriver.java:114)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:191)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:186)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:182)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:95)
at testSel.main(testSel.java:10)
Caused by: org.openqa.selenium.firefox.NotConnectedException: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms. Firefox console output:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:118)
... 7 more
```
```
### Operating System
macOs
### Selenium version
4.2.2
### What are the browser(s) and version(s) where you see this issue?
FireFox 101.0.1 (64 bit)
### What are the browser driver(s) and version(s) where you see this issue?
Geckodriver 0.31.0
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms - ### What happened?
Following issue happens for FireFox and Google Chrome browsers/drivers for latest versions. I found same issues from past years, where it was recommended to update to latest versions of selenium/driver, however I already have latest versions.
**Test:**
```
import org.openqa.selenium.By;
import org.openqa.selenium.WebDriver;
import org.openqa.selenium.WebElement;
import org.openqa.selenium.firefox.FirefoxDriver;
public class testSel {
public static void main(String[] args){
System.out.println("1");
System.setProperty("webdriver.gecko.driver", "/Users/atkeshelashvili/Downloads/geckodriver");
WebDriver driver=new FirefoxDriver();
System.out.println("2");
driver.get("http://demo.guru99.com/");
System.out.println("3");
WebElement element=driver.findElement(By.xpath("//input[@name='emailid']"));
element.sendKeys("abc@gmail.com");
WebElement button=driver.findElement(By.xpath("//input[@name='btnLogin']"));
button.click();
}
}
```
### How can we reproduce the issue?
```shell
**Upload following versions:**
FireFox 101.0.1 (64 bit)
Selenium Web Driver 4.2.2 selenium-java-4.2.2.zip
Geckodriver 0.31.0
System macOS Monterey (M1)
Same happens for Google Chrome 102.0.5005.61.
```
### Relevant log output
```shell
**Logs:**
org.openqa.selenium.firefox.NotConnectedException: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms. Firefox console output:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:118)
at org.openqa.selenium.firefox.FirefoxDriver.startClient(FirefoxDriver.java:246)
at org.openqa.selenium.remote.RemoteWebDriver.<init>(RemoteWebDriver.java:114)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:191)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:186)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:182)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:95)
at testSel.main(testSel.java:10)
Exception in thread "main" org.openqa.selenium.WebDriverException: Failed to connect to binary FirefoxBinary(/Applications/Firefox.app/Contents/MacOS/firefox-bin) on port 7055; process output follows:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
Build info: version: '2.45.0', revision: '5017cb8e7ca8e37638dc3091b2440b90a1d8686f', time: '2015-02-27 09:10:26'
System info: host: 'ip-192-168-0-198.eu-central-1.compute.internal', ip: '192.168.0.198', os.name: 'Mac OS X', os.arch: 'aarch64', os.version: '12.1', java.version: '11.0.12'
Driver info: driver.version: FirefoxDriver
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:130)
at org.openqa.selenium.firefox.FirefoxDriver.startClient(FirefoxDriver.java:246)
at org.openqa.selenium.remote.RemoteWebDriver.<init>(RemoteWebDriver.java:114)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:191)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:186)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:182)
at org.openqa.selenium.firefox.FirefoxDriver.<init>(FirefoxDriver.java:95)
at testSel.main(testSel.java:10)
Caused by: org.openqa.selenium.firefox.NotConnectedException: Unable to connect to host 127.0.0.1 on port 7055 after 45000 ms. Firefox console output:
ce://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
getShutdownState@resource://gre/modules/TelemetrySend.jsm:1665:7
getShutdownState@resource://gre/modules/TelemetrySend.jsm:358:30
_getState@resource://gre/modules/TelemetryControllerParent.jsm:1073:33
setupTelemetry/<@resource://gre/modules/TelemetryControllerParent.jsm:913:18
safeGetState@resource://gre/modules/AsyncShutdown.jsm:257:17
Barrier/addBlocker/promise<@resource://gre/modules/AsyncShutdown.jsm:766:20
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
WARNING: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
WARNING: addBlocker@resource://gre/modules/AsyncShutdown.jsm:718:15
addBlocker@resource://gre/modules/AsyncShutdown.jsm:526:26
addBlocker@resource://gre/modules/AsyncShutdown.jsm:465:15
setupShutdown@resource://gre/modules/osfile/osfile_async_front.jsm:1564:28
@resource://gre/modules/osfile/osfile_async_front.jsm:1584:16
@resource://gre/modules/osfile.jsm:11:30
@resource://gre/modules/TelemetryStorage.jsm:18:28
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:717:5
removeAbortedSessionPing@resource://gre/modules/TelemetryControllerParent.jsm:246:17
shutdownChromeProcess/cleanup/<@resource://gre/modules/TelemetrySession.jsm:1231:37
observe@resource://gre/modules/AsyncShutdown.jsm:575:16
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
JavaScript error: resource://gre/modules/AsyncShutdown.jsm, line 718: Error: Phase "profile-before-change" is finished, it is too late to register completion condition "OS.File: flush I/O queued before profileBeforeChange"
2022-06-17 10:13:54.733 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.734 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972533] nil host used in call to allowsAnyHTTPSCertificateForHost:
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsSpecificHTTPSCertificateForHost
2022-06-17 10:13:54.735 plugin-container[84134:8972531] nil host used in call to allowsAnyHTTPSCertificateForHost:
UNSUPPORTED (log once): POSSIBLE ISSUE: unit 1 GLD_TEXTURE_INDEX_2D is unloadable and bound to sampler type (Float) - using zero texture because texture unloadable
console.error: BackgroundUpdate:
_reasonsToNotScheduleUpdates: Failed to check for Maintenance Service Registry Key: [Exception... "Component returned failure code: 0x80004001 (NS_ERROR_NOT_IMPLEMENTED) [nsIUpdateProcessor.getServiceRegKeyExists]" nsresult: "0x80004001 (NS_ERROR_NOT_IMPLEMENTED)" location: "JS frame :: resource://gre/modules/BackgroundUpdate.jsm :: _reasonsToNotScheduleUpdates :: line 241" data: no]
at org.openqa.selenium.firefox.internal.NewProfileExtensionConnection.start(NewProfileExtensionConnection.java:118)
... 7 more
```
```
### Operating System
macOs
### Selenium version
4.2.2
### What are the browser(s) and version(s) where you see this issue?
FireFox 101.0.1 (64 bit)
### What are the browser driver(s) and version(s) where you see this issue?
Geckodriver 0.31.0
### Are you using Selenium Grid?
_No response_ | defect | unable to connect to host on port after ms what happened following issue happens for firefox and google chrome browsers drivers for latest versions i found same issues from past years where it was recommended to update to latest versions of selenium driver however i already have latest versions test import org openqa selenium by import org openqa selenium webdriver import org openqa selenium webelement import org openqa selenium firefox firefoxdriver public class testsel public static void main string args system out println system setproperty webdriver gecko driver users atkeshelashvili downloads geckodriver webdriver driver new firefoxdriver system out println driver get system out println webelement element driver findelement by xpath input element sendkeys abc gmail com webelement button driver findelement by xpath input button click how can we reproduce the issue shell upload following versions firefox bit selenium web driver selenium java zip geckodriver system macos monterey same happens for google chrome relevant log output shell logs org openqa selenium firefox notconnectedexception unable to connect to host on port after ms firefox console output ce gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm getshutdownstate resource gre modules telemetrysend jsm getshutdownstate resource gre modules telemetrysend jsm getstate resource gre modules telemetrycontrollerparent jsm setuptelemetry resource gre modules telemetrycontrollerparent jsm safegetstate resource gre modules asyncshutdown jsm barrier addblocker promise resource gre modules asyncshutdown jsm observe resource gre modules asyncshutdown jsm warning error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange warning addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm shutdownchromeprocess cleanup resource gre modules telemetrysession jsm observe resource gre modules asyncshutdown jsm javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost unsupported log once possible issue unit gld texture index is unloadable and bound to sampler type float using zero texture because texture unloadable console error backgroundupdate reasonstonotscheduleupdates failed to check for maintenance service registry key nsresult ns error not implemented location js frame resource gre modules backgroundupdate jsm reasonstonotscheduleupdates line data no at org openqa selenium firefox internal newprofileextensionconnection start newprofileextensionconnection java at org openqa selenium firefox firefoxdriver startclient firefoxdriver java at org openqa selenium remote remotewebdriver remotewebdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at testsel main testsel java exception in thread main org openqa selenium webdriverexception failed to connect to binary firefoxbinary applications firefox app contents macos firefox bin on port process output follows ce gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm getshutdownstate resource gre modules telemetrysend jsm getshutdownstate resource gre modules telemetrysend jsm getstate resource gre modules telemetrycontrollerparent jsm setuptelemetry resource gre modules telemetrycontrollerparent jsm safegetstate resource gre modules asyncshutdown jsm barrier addblocker promise resource gre modules asyncshutdown jsm observe resource gre modules asyncshutdown jsm warning error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange warning addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm shutdownchromeprocess cleanup resource gre modules telemetrysession jsm observe resource gre modules asyncshutdown jsm javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost unsupported log once possible issue unit gld texture index is unloadable and bound to sampler type float using zero texture because texture unloadable console error backgroundupdate reasonstonotscheduleupdates failed to check for maintenance service registry key nsresult ns error not implemented location js frame resource gre modules backgroundupdate jsm reasonstonotscheduleupdates line data no build info version revision time system info host ip eu central compute internal ip os name mac os x os arch os version java version driver info driver version firefoxdriver at org openqa selenium firefox internal newprofileextensionconnection start newprofileextensionconnection java at org openqa selenium firefox firefoxdriver startclient firefoxdriver java at org openqa selenium remote remotewebdriver remotewebdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at org openqa selenium firefox firefoxdriver firefoxdriver java at testsel main testsel java caused by org openqa selenium firefox notconnectedexception unable to connect to host on port after ms firefox console output ce gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm getshutdownstate resource gre modules telemetrysend jsm getshutdownstate resource gre modules telemetrysend jsm getstate resource gre modules telemetrycontrollerparent jsm setuptelemetry resource gre modules telemetrycontrollerparent jsm safegetstate resource gre modules asyncshutdown jsm barrier addblocker promise resource gre modules asyncshutdown jsm observe resource gre modules asyncshutdown jsm warning error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange warning addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm addblocker resource gre modules asyncshutdown jsm setupshutdown resource gre modules osfile osfile async front jsm resource gre modules osfile osfile async front jsm resource gre modules osfile jsm resource gre modules telemetrystorage jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm removeabortedsessionping resource gre modules telemetrycontrollerparent jsm shutdownchromeprocess cleanup resource gre modules telemetrysession jsm observe resource gre modules asyncshutdown jsm javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange javascript error resource gre modules asyncshutdown jsm line error phase profile before change is finished it is too late to register completion condition os file flush i o queued before profilebeforechange plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost plugin container nil host used in call to allowsspecifichttpscertificateforhost plugin container nil host used in call to allowsanyhttpscertificateforhost unsupported log once possible issue unit gld texture index is unloadable and bound to sampler type float using zero texture because texture unloadable console error backgroundupdate reasonstonotscheduleupdates failed to check for maintenance service registry key nsresult ns error not implemented location js frame resource gre modules backgroundupdate jsm reasonstonotscheduleupdates line data no at org openqa selenium firefox internal newprofileextensionconnection start newprofileextensionconnection java more operating system macos selenium version what are the browser s and version s where you see this issue firefox bit what are the browser driver s and version s where you see this issue geckodriver are you using selenium grid no response | 1 |
110,164 | 11,693,671,670 | IssuesEvent | 2020-03-06 01:19:58 | hackforla/website | https://api.github.com/repos/hackforla/website | opened | Place your request for changes to your project homepage here | documentation | ### Overview
You are here because you were sent a link to your temp page on the hack for LA website (hackforla.org) and we want to hear about the information you want added or changed on your project homepage.
### Action Items
- [ ] Identify which team you are writing from (engage, tdm-calculator, etc.)
- [ ] Tell us what information needs to be changed, or what new information or resource you want us to link to.
### Resources/Instructions
Here is a list of all the active projects with project pages:
https://www.hackforla.org/projects/311-data.html
https://www.hackforla.org/projects/engage.html
https://www.hackforla.org/projects/equity-language.html
https://www.hackforla.org/projects/food-oasis.html
https://www.hackforla.org/projects/hellogov.html
https://www.hackforla.org/projects/lucky-parking.html
https://www.hackforla.org/projects/metro-ontime.html
https://www.hackforla.org/projects/new-schools-today.html
https://www.hackforla.org/projects/public-tree-map.html
https://www.hackforla.org/projects/record-clearance-project.html
https://www.hackforla.org/projects/tdm-calculator.html
https://www.hackforla.org/projects/undebate.html
https://www.hackforla.org/projects/vrms.html
https://www.hackforla.org/projects/website.html | 1.0 | Place your request for changes to your project homepage here - ### Overview
You are here because you were sent a link to your temp page on the hack for LA website (hackforla.org) and we want to hear about the information you want added or changed on your project homepage.
### Action Items
- [ ] Identify which team you are writing from (engage, tdm-calculator, etc.)
- [ ] Tell us what information needs to be changed, or what new information or resource you want us to link to.
### Resources/Instructions
Here is a list of all the active projects with project pages:
https://www.hackforla.org/projects/311-data.html
https://www.hackforla.org/projects/engage.html
https://www.hackforla.org/projects/equity-language.html
https://www.hackforla.org/projects/food-oasis.html
https://www.hackforla.org/projects/hellogov.html
https://www.hackforla.org/projects/lucky-parking.html
https://www.hackforla.org/projects/metro-ontime.html
https://www.hackforla.org/projects/new-schools-today.html
https://www.hackforla.org/projects/public-tree-map.html
https://www.hackforla.org/projects/record-clearance-project.html
https://www.hackforla.org/projects/tdm-calculator.html
https://www.hackforla.org/projects/undebate.html
https://www.hackforla.org/projects/vrms.html
https://www.hackforla.org/projects/website.html | non_defect | place your request for changes to your project homepage here overview you are here because you were sent a link to your temp page on the hack for la website hackforla org and we want to hear about the information you want added or changed on your project homepage action items identify which team you are writing from engage tdm calculator etc tell us what information needs to be changed or what new information or resource you want us to link to resources instructions here is a list of all the active projects with project pages | 0 |
64,781 | 6,922,037,817 | IssuesEvent | 2017-11-30 00:51:49 | Ekultek/Zeus-Scanner | https://api.github.com/repos/Ekultek/Zeus-Scanner | closed | Unhandled exception (DcCrHRj) | bug test tool issue | Zeus version:
`1.3.7.a96a04`
Firefox version:
`(57, 0)`
Geckodriver version:
`geckodriver-v0.19.0-linux64.tar.gz`
Error info:
```Traceback (most recent call last):
File "zeus.py", line 408, in <module>
__run_attacks_main()
File "zeus.py", line 305, in __run_attacks_main
threads=opt.amountOfThreads
File "/home/baal/bin/python/zeus-scanner/lib/core/settings.py", line 995, in run_attacks
verbose=verbose, do_threading=do_threading, batch=batch
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 204, in main
results = check_for_externals(url, robots=True, batch=batch)
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 38, in check_for_externals
conn = requests.get(full_url)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 72, in get
return request('get', url, params=params, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 58, in request
return session.request(method=method, url=url, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 508, in request
resp = self.send(prep, **send_kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 618, in send
r = adapter.send(request, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/adapters.py", line 508, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))
````
Running details:
`Linux-4.10.0-40-generic-x86_64-with-Ubuntu-17.04-zesty`
Commands used:
`zeus.py -d inurl:cp.php?cl=cp nowmenuid= --batch --verbose -a --random-agent --x-forward -M -L 3000`
Log file info:
```2017-11-29 14:47:07,768;zeus-log;DEBUG;trying 'http://www.gohontech.com/userlogin.%EXT%'...
2017-11-29 14:47:07,908;zeus-log;ERROR;failed to connect got error code 400:...
2017-11-29 14:47:07,908;zeus-log;INFO;found 0 possible connections(s) and 5 successful connection(s)...
2017-11-29 14:47:07,909;zeus-log;INFO;creating connection tree...
2017-11-29 14:47:07,909;zeus-log;WARNING;only writing successful connections to log file...
2017-11-29 14:47:07,910;zeus-log;INFO;successfully wrote found items to '/home/baal/bin/python/zeus-scanner/log/admin-page-log/www.gohontech.com-admin-page.log'...
2017-11-29 14:47:07,910;zeus-log;INFO;currently running on 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651' (target #10)...
2017-11-29 14:47:07,911;zeus-log;INFO;checking for HTTP headers...
2017-11-29 14:47:07,911;zeus-log;INFO;checking if target URL is protected by some kind of WAF/IPS/IDS...
2017-11-29 14:47:07,911;zeus-log;INFO;random IP address generated for header '106.103.231.249'...
2017-11-29 14:47:07,912;zeus-log;INFO;random IP address generated for header '63.170.25.74'...
2017-11-29 14:47:07,912;zeus-log;INFO;random IP address generated for header '48.166.214.189'...
2017-11-29 14:47:07,912;zeus-log;DEBUG;attempting connection to 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651 AND 1=1 UNION ALL SELECT 1,NULL,'<script>alert("XSS")</script>',table_name FROM information_schema.tables WHERE 2>1--/**/; EXEC xp_cmdshell('cat ../../../etc/passwd')#'...
2017-11-29 14:47:27,933;zeus-log;WARNING;detection request failed, assuming no protection and continuing...
2017-11-29 14:47:27,934;zeus-log;INFO;no WAF/IDS/IPS has been identified on target URL...
2017-11-29 14:47:27,934;zeus-log;DEBUG;loading XML data...
2017-11-29 14:47:27,935;zeus-log;INFO;attempting to get request headers for 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651'...
2017-11-29 14:47:27,936;zeus-log;INFO;random IP address generated for header '47.213.90.100'...
2017-11-29 14:47:27,936;zeus-log;INFO;random IP address generated for header '228.168.139.66'...
2017-11-29 14:47:27,937;zeus-log;INFO;random IP address generated for header '180.60.178.51'...
2017-11-29 14:47:47,944;zeus-log;ERROR;unable to retrieve headers for site 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651'...
2017-11-29 14:47:47,945;zeus-log;INFO;parsing robots.txt...
2017-11-29 14:48:07,961;zeus-log;ERROR;[7;31;31mZeus has hit an unexpected error and cannot continue, error code 'HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))'...
Traceback (most recent call last):
File "zeus.py", line 408, in <module>
__run_attacks_main()
File "zeus.py", line 305, in __run_attacks_main
threads=opt.amountOfThreads
File "/home/baal/bin/python/zeus-scanner/lib/core/settings.py", line 995, in run_attacks
verbose=verbose, do_threading=do_threading, batch=batch
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 204, in main
results = check_for_externals(url, robots=True, batch=batch)
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 38, in check_for_externals
conn = requests.get(full_url)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 72, in get
return request('get', url, params=params, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 58, in request
return session.request(method=method, url=url, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 508, in request
resp = self.send(prep, **send_kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 618, in send
r = adapter.send(request, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/adapters.py", line 508, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))
2017-11-29 14:48:08,477;zeus-log;INFO;[32mZeus got an unexpected error and will automatically create an issue for this error, please wait...[0m
2017-11-29 14:48:08,478;zeus-log;INFO;[32mgetting authorization...[0m
2017-11-29 14:48:08,488;zeus-log;INFO;[32mextracting traceback from log file...[0m
2017-11-29 14:48:08,492;zeus-log;INFO;[32mattempting to get firefox browser version...[0m
``` | 1.0 | Unhandled exception (DcCrHRj) - Zeus version:
`1.3.7.a96a04`
Firefox version:
`(57, 0)`
Geckodriver version:
`geckodriver-v0.19.0-linux64.tar.gz`
Error info:
```Traceback (most recent call last):
File "zeus.py", line 408, in <module>
__run_attacks_main()
File "zeus.py", line 305, in __run_attacks_main
threads=opt.amountOfThreads
File "/home/baal/bin/python/zeus-scanner/lib/core/settings.py", line 995, in run_attacks
verbose=verbose, do_threading=do_threading, batch=batch
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 204, in main
results = check_for_externals(url, robots=True, batch=batch)
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 38, in check_for_externals
conn = requests.get(full_url)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 72, in get
return request('get', url, params=params, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 58, in request
return session.request(method=method, url=url, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 508, in request
resp = self.send(prep, **send_kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 618, in send
r = adapter.send(request, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/adapters.py", line 508, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))
````
Running details:
`Linux-4.10.0-40-generic-x86_64-with-Ubuntu-17.04-zesty`
Commands used:
`zeus.py -d inurl:cp.php?cl=cp nowmenuid= --batch --verbose -a --random-agent --x-forward -M -L 3000`
Log file info:
```2017-11-29 14:47:07,768;zeus-log;DEBUG;trying 'http://www.gohontech.com/userlogin.%EXT%'...
2017-11-29 14:47:07,908;zeus-log;ERROR;failed to connect got error code 400:...
2017-11-29 14:47:07,908;zeus-log;INFO;found 0 possible connections(s) and 5 successful connection(s)...
2017-11-29 14:47:07,909;zeus-log;INFO;creating connection tree...
2017-11-29 14:47:07,909;zeus-log;WARNING;only writing successful connections to log file...
2017-11-29 14:47:07,910;zeus-log;INFO;successfully wrote found items to '/home/baal/bin/python/zeus-scanner/log/admin-page-log/www.gohontech.com-admin-page.log'...
2017-11-29 14:47:07,910;zeus-log;INFO;currently running on 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651' (target #10)...
2017-11-29 14:47:07,911;zeus-log;INFO;checking for HTTP headers...
2017-11-29 14:47:07,911;zeus-log;INFO;checking if target URL is protected by some kind of WAF/IPS/IDS...
2017-11-29 14:47:07,911;zeus-log;INFO;random IP address generated for header '106.103.231.249'...
2017-11-29 14:47:07,912;zeus-log;INFO;random IP address generated for header '63.170.25.74'...
2017-11-29 14:47:07,912;zeus-log;INFO;random IP address generated for header '48.166.214.189'...
2017-11-29 14:47:07,912;zeus-log;DEBUG;attempting connection to 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651 AND 1=1 UNION ALL SELECT 1,NULL,'<script>alert("XSS")</script>',table_name FROM information_schema.tables WHERE 2>1--/**/; EXEC xp_cmdshell('cat ../../../etc/passwd')#'...
2017-11-29 14:47:27,933;zeus-log;WARNING;detection request failed, assuming no protection and continuing...
2017-11-29 14:47:27,934;zeus-log;INFO;no WAF/IDS/IPS has been identified on target URL...
2017-11-29 14:47:27,934;zeus-log;DEBUG;loading XML data...
2017-11-29 14:47:27,935;zeus-log;INFO;attempting to get request headers for 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651'...
2017-11-29 14:47:27,936;zeus-log;INFO;random IP address generated for header '47.213.90.100'...
2017-11-29 14:47:27,936;zeus-log;INFO;random IP address generated for header '228.168.139.66'...
2017-11-29 14:47:27,937;zeus-log;INFO;random IP address generated for header '180.60.178.51'...
2017-11-29 14:47:47,944;zeus-log;ERROR;unable to retrieve headers for site 'http://www.kdsmed.com/cp.php?cl=cp&nowmenuid=20111651'...
2017-11-29 14:47:47,945;zeus-log;INFO;parsing robots.txt...
2017-11-29 14:48:07,961;zeus-log;ERROR;[7;31;31mZeus has hit an unexpected error and cannot continue, error code 'HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))'...
Traceback (most recent call last):
File "zeus.py", line 408, in <module>
__run_attacks_main()
File "zeus.py", line 305, in __run_attacks_main
threads=opt.amountOfThreads
File "/home/baal/bin/python/zeus-scanner/lib/core/settings.py", line 995, in run_attacks
verbose=verbose, do_threading=do_threading, batch=batch
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 204, in main
results = check_for_externals(url, robots=True, batch=batch)
File "/home/baal/bin/python/zeus-scanner/lib/attacks/admin_panel_finder/__init__.py", line 38, in check_for_externals
conn = requests.get(full_url)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 72, in get
return request('get', url, params=params, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/api.py", line 58, in request
return session.request(method=method, url=url, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 508, in request
resp = self.send(prep, **send_kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/sessions.py", line 618, in send
r = adapter.send(request, **kwargs)
File "/home/baal/.local/lib/python2.7/site-packages/requests/adapters.py", line 508, in send
raise ConnectionError(e, request=request)
ConnectionError: HTTPConnectionPool(host='www.kdsmed.com', port=80): Max retries exceeded with url: /robots.txt (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fbbfacb4a50>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution',))
2017-11-29 14:48:08,477;zeus-log;INFO;[32mZeus got an unexpected error and will automatically create an issue for this error, please wait...[0m
2017-11-29 14:48:08,478;zeus-log;INFO;[32mgetting authorization...[0m
2017-11-29 14:48:08,488;zeus-log;INFO;[32mextracting traceback from log file...[0m
2017-11-29 14:48:08,492;zeus-log;INFO;[32mattempting to get firefox browser version...[0m
``` | non_defect | unhandled exception dccrhrj zeus version firefox version geckodriver version geckodriver tar gz error info traceback most recent call last file zeus py line in run attacks main file zeus py line in run attacks main threads opt amountofthreads file home baal bin python zeus scanner lib core settings py line in run attacks verbose verbose do threading do threading batch batch file home baal bin python zeus scanner lib attacks admin panel finder init py line in main results check for externals url robots true batch batch file home baal bin python zeus scanner lib attacks admin panel finder init py line in check for externals conn requests get full url file home baal local lib site packages requests api py line in get return request get url params params kwargs file home baal local lib site packages requests api py line in request return session request method method url url kwargs file home baal local lib site packages requests sessions py line in request resp self send prep send kwargs file home baal local lib site packages requests sessions py line in send r adapter send request kwargs file home baal local lib site packages requests adapters py line in send raise connectionerror e request request connectionerror httpconnectionpool host port max retries exceeded with url robots txt caused by newconnectionerror failed to establish a new connection temporary failure in name resolution running details linux generic with ubuntu zesty commands used zeus py d inurl cp php cl cp nowmenuid batch verbose a random agent x forward m l log file info zeus log debug trying zeus log error failed to connect got error code zeus log info found possible connections s and successful connection s zeus log info creating connection tree zeus log warning only writing successful connections to log file zeus log info successfully wrote found items to home baal bin python zeus scanner log admin page log zeus log info currently running on target zeus log info checking for http headers zeus log info checking if target url is protected by some kind of waf ips ids zeus log info random ip address generated for header zeus log info random ip address generated for header zeus log info random ip address generated for header zeus log debug attempting connection to and union all select null alert xss table name from information schema tables where exec xp cmdshell cat etc passwd zeus log warning detection request failed assuming no protection and continuing zeus log info no waf ids ips has been identified on target url zeus log debug loading xml data zeus log info attempting to get request headers for zeus log info random ip address generated for header zeus log info random ip address generated for header zeus log info random ip address generated for header zeus log error unable to retrieve headers for site zeus log info parsing robots txt zeus log error temporary failure in name resolution traceback most recent call last file zeus py line in run attacks main file zeus py line in run attacks main threads opt amountofthreads file home baal bin python zeus scanner lib core settings py line in run attacks verbose verbose do threading do threading batch batch file home baal bin python zeus scanner lib attacks admin panel finder init py line in main results check for externals url robots true batch batch file home baal bin python zeus scanner lib attacks admin panel finder init py line in check for externals conn requests get full url file home baal local lib site packages requests api py line in get return request get url params params kwargs file home baal local lib site packages requests api py line in request return session request method method url url kwargs file home baal local lib site packages requests sessions py line in request resp self send prep send kwargs file home baal local lib site packages requests sessions py line in send r adapter send request kwargs file home baal local lib site packages requests adapters py line in send raise connectionerror e request request connectionerror httpconnectionpool host port max retries exceeded with url robots txt caused by newconnectionerror failed to establish a new connection temporary failure in name resolution zeus log info got an unexpected error and will automatically create an issue for this error please wait zeus log info authorization zeus log info traceback from log file zeus log info to get firefox browser version | 0 |
22,239 | 3,619,269,952 | IssuesEvent | 2016-02-08 15:26:05 | pavva94/snake-os | https://api.github.com/repos/pavva94/snake-os | closed | HDD heat up | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Use an HDD on the wlxkj652
2. Use the same HDD on a PC usb port.
3. See that the HDD with the wlxkj652 heat up much more than with the pc.
What is the expected output? What do you see instead?
I expected the same temperature in both cases. But the difference is huge.
What version of the product are you using? On what operating system?
SNAKE OS V1.3.2 (20111019), Windows 7 and 8.
Please provide any additional information below.
```
Original issue reported on code.google.com by `teodoro....@gmail.com` on 27 Dec 2013 at 1:42 | 1.0 | HDD heat up - ```
What steps will reproduce the problem?
1. Use an HDD on the wlxkj652
2. Use the same HDD on a PC usb port.
3. See that the HDD with the wlxkj652 heat up much more than with the pc.
What is the expected output? What do you see instead?
I expected the same temperature in both cases. But the difference is huge.
What version of the product are you using? On what operating system?
SNAKE OS V1.3.2 (20111019), Windows 7 and 8.
Please provide any additional information below.
```
Original issue reported on code.google.com by `teodoro....@gmail.com` on 27 Dec 2013 at 1:42 | defect | hdd heat up what steps will reproduce the problem use an hdd on the use the same hdd on a pc usb port see that the hdd with the heat up much more than with the pc what is the expected output what do you see instead i expected the same temperature in both cases but the difference is huge what version of the product are you using on what operating system snake os windows and please provide any additional information below original issue reported on code google com by teodoro gmail com on dec at | 1 |
22,124 | 3,602,416,853 | IssuesEvent | 2016-02-03 15:36:06 | vozdikililer/test | https://api.github.com/repos/vozdikililer/test | closed | df | auto-migrated HeadsUp Priority-Medium Type-Defect | ```
Chrome Version : Mozilla/5.0 (X11; U; Linux x86_64; en-US)
AppleWebKit/534.20 (KHTML, like Gecko) Chrome/11.0.672.2 Safari/534.20
URLs (if applicable) : http://www.google.ru/
Other browsers tested:
Add OK or FAIL after other browsers where you have tested this issue:
Safari 4: None
Firefox 3.x: None
IE 7: None
IE 8: None
What steps will reproduce the problem?
What is the expected result?
What happens instead?
Please provide any additional information below. Attach a screenshot if
possible.
---------------------------------------------------------
gThunder auto-generated comment.
---------------------------------------------------------
target_element:
{"tagName":"DIV","elementText":"","attributes":{"align":"left","style":"eight126
pxwidth364px","title":"Google","id":"hplogo","onload":"windowlollol"},"optimized
":true,"parentElem":{"tagName":"DIV","elementText":"","attributes":{"style":"pad
ding28px03px"},"optimized":true,"parentElem":{"tagName":"DIV","elementText":"","
attributes":{"id":"lga","style":"t171pxpaddingtop22px"},"optimized":true,"parent
Elem":{"tagName":"CENTER","elementText":"","optimized":true,"parentElem":null}}}
} :target_element
```
Original issue reported on code.google.com by `ekamensk...@google.com` on 28 Feb 2011 at 8:27 | 1.0 | df - ```
Chrome Version : Mozilla/5.0 (X11; U; Linux x86_64; en-US)
AppleWebKit/534.20 (KHTML, like Gecko) Chrome/11.0.672.2 Safari/534.20
URLs (if applicable) : http://www.google.ru/
Other browsers tested:
Add OK or FAIL after other browsers where you have tested this issue:
Safari 4: None
Firefox 3.x: None
IE 7: None
IE 8: None
What steps will reproduce the problem?
What is the expected result?
What happens instead?
Please provide any additional information below. Attach a screenshot if
possible.
---------------------------------------------------------
gThunder auto-generated comment.
---------------------------------------------------------
target_element:
{"tagName":"DIV","elementText":"","attributes":{"align":"left","style":"eight126
pxwidth364px","title":"Google","id":"hplogo","onload":"windowlollol"},"optimized
":true,"parentElem":{"tagName":"DIV","elementText":"","attributes":{"style":"pad
ding28px03px"},"optimized":true,"parentElem":{"tagName":"DIV","elementText":"","
attributes":{"id":"lga","style":"t171pxpaddingtop22px"},"optimized":true,"parent
Elem":{"tagName":"CENTER","elementText":"","optimized":true,"parentElem":null}}}
} :target_element
```
Original issue reported on code.google.com by `ekamensk...@google.com` on 28 Feb 2011 at 8:27 | defect | df chrome version mozilla u linux en us applewebkit khtml like gecko chrome safari urls if applicable other browsers tested add ok or fail after other browsers where you have tested this issue safari none firefox x none ie none ie none what steps will reproduce the problem what is the expected result what happens instead please provide any additional information below attach a screenshot if possible gthunder auto generated comment target element tagname div elementtext attributes align left style title google id hplogo onload windowlollol optimized true parentelem tagname div elementtext attributes style pad optimized true parentelem tagname div elementtext attributes id lga style optimized true parent elem tagname center elementtext optimized true parentelem null target element original issue reported on code google com by ekamensk google com on feb at | 1 |
62,616 | 17,092,453,998 | IssuesEvent | 2021-07-08 19:28:31 | SAP/fundamental-ngx | https://api.github.com/repos/SAP/fundamental-ngx | closed | build warning - deep imports into @angular/cdk/overlay/position/connected-position | Defect Hunting High bug core denoland-porto | #### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
clone and build https://github.com/SAP-samples/fundamental-ngx-sample-apps
the build shows this:
```
Warning: Entry point '@fundamental-ngx/core' contains deep imports into '/Users/i820906/develop/github/fundamental-ngx-sample-apps/node_modules/@angular/cdk/overlay/position/connected-position'. This is probably not a problem, but may cause the compilation of entry points to be out of order.
```
#### Which versions of Angular and Fundamental Library for Angular are affected? (If this is a feature request, use current version.)
`"@fundamental-ngx/core": "v0.26.0-rc.49"`
| 1.0 | build warning - deep imports into @angular/cdk/overlay/position/connected-position - #### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
clone and build https://github.com/SAP-samples/fundamental-ngx-sample-apps
the build shows this:
```
Warning: Entry point '@fundamental-ngx/core' contains deep imports into '/Users/i820906/develop/github/fundamental-ngx-sample-apps/node_modules/@angular/cdk/overlay/position/connected-position'. This is probably not a problem, but may cause the compilation of entry points to be out of order.
```
#### Which versions of Angular and Fundamental Library for Angular are affected? (If this is a feature request, use current version.)
`"@fundamental-ngx/core": "v0.26.0-rc.49"`
| defect | build warning deep imports into angular cdk overlay position connected position is this a bug enhancement or feature request bug briefly describe your proposal clone and build the build shows this warning entry point fundamental ngx core contains deep imports into users develop github fundamental ngx sample apps node modules angular cdk overlay position connected position this is probably not a problem but may cause the compilation of entry points to be out of order which versions of angular and fundamental library for angular are affected if this is a feature request use current version fundamental ngx core rc | 1 |
133,707 | 18,945,016,254 | IssuesEvent | 2021-11-18 09:15:19 | denoland/deno | https://api.github.com/repos/denoland/deno | closed | Deno "unboxes" primitive "this" while Chrome and Node do not | working as designed | **setup**
```js
function self() {
return this;
}
```
**Chrome** and **Node**
```console
> self.apply("hello")
String {'hello'}
> typeof self.apply("hello")
'object'
```
**Deno**
```console
> self.apply("hello")
"hello"
> typeof self.apply("hello")
"string"
```
I like how Deno appears to intelligently "unbox" the object to a primitive but as this behavior differs from Chrome and Node it makes me wonder if there is some ES specification issue here or not (or maybe Chrome and Node have it wrong?)
should Deno behave differently than Chrome and Node here? | 1.0 | Deno "unboxes" primitive "this" while Chrome and Node do not - **setup**
```js
function self() {
return this;
}
```
**Chrome** and **Node**
```console
> self.apply("hello")
String {'hello'}
> typeof self.apply("hello")
'object'
```
**Deno**
```console
> self.apply("hello")
"hello"
> typeof self.apply("hello")
"string"
```
I like how Deno appears to intelligently "unbox" the object to a primitive but as this behavior differs from Chrome and Node it makes me wonder if there is some ES specification issue here or not (or maybe Chrome and Node have it wrong?)
should Deno behave differently than Chrome and Node here? | non_defect | deno unboxes primitive this while chrome and node do not setup js function self return this chrome and node console self apply hello string hello typeof self apply hello object deno console self apply hello hello typeof self apply hello string i like how deno appears to intelligently unbox the object to a primitive but as this behavior differs from chrome and node it makes me wonder if there is some es specification issue here or not or maybe chrome and node have it wrong should deno behave differently than chrome and node here | 0 |
46,169 | 13,055,861,981 | IssuesEvent | 2020-07-30 02:57:24 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | SuiteSparse port needs makefile and portfile love (Trac #678) | Incomplete Migration Migrated from Trac defect tools/ports | Migrated from https://code.icecube.wisc.edu/ticket/678
```json
{
"status": "closed",
"changetime": "2012-06-10T19:35:02",
"description": "Makefile and Portfile need to be adjusted so that the user can select the compiler with environment variables",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1339356902000000",
"component": "tools/ports",
"summary": "SuiteSparse port needs makefile and portfile love",
"priority": "normal",
"keywords": "SuiteSparse",
"time": "2012-05-29T19:14:50",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | SuiteSparse port needs makefile and portfile love (Trac #678) - Migrated from https://code.icecube.wisc.edu/ticket/678
```json
{
"status": "closed",
"changetime": "2012-06-10T19:35:02",
"description": "Makefile and Portfile need to be adjusted so that the user can select the compiler with environment variables",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1339356902000000",
"component": "tools/ports",
"summary": "SuiteSparse port needs makefile and portfile love",
"priority": "normal",
"keywords": "SuiteSparse",
"time": "2012-05-29T19:14:50",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | suitesparse port needs makefile and portfile love trac migrated from json status closed changetime description makefile and portfile need to be adjusted so that the user can select the compiler with environment variables reporter nega cc resolution fixed ts component tools ports summary suitesparse port needs makefile and portfile love priority normal keywords suitesparse time milestone owner nega type defect | 1 |
25,940 | 12,324,408,909 | IssuesEvent | 2020-05-13 13:40:46 | dockstore/dockstore | https://api.github.com/repos/dockstore/dockstore | closed | Deleting a non-existent test parameter file does not return an error | bug review web-service | ## Bug Report
If you delete a non-existent test parameter file, the API returns a 200 status code.
It should return a 400 or 404 status code indicating that the file was not deleted.
This happens for both [tools](https://github.com/dockstore/dockstore/blob/469261112f80613393000b1d3fe51b46bfc15498/dockstore-webservice/src/main/java/io/dockstore/webservice/resources/DockerRepoResource.java#L887) and [workflows](https://github.com/dockstore/dockstore/blob/469261112f80613393000b1d3fe51b46bfc15498/dockstore-webservice/src/main/java/io/dockstore/webservice/resources/WorkflowResource.java#L1173) -- both methods just delete files they can find, but don't check if the files exist.
Not high priority.
┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-1039)
┆Issue Type: Story
┆Fix Versions: Dockstore 1.9
┆Sprint: Sprint 34 Isopods
┆Issue Number: DOCK-1039
| 1.0 | Deleting a non-existent test parameter file does not return an error - ## Bug Report
If you delete a non-existent test parameter file, the API returns a 200 status code.
It should return a 400 or 404 status code indicating that the file was not deleted.
This happens for both [tools](https://github.com/dockstore/dockstore/blob/469261112f80613393000b1d3fe51b46bfc15498/dockstore-webservice/src/main/java/io/dockstore/webservice/resources/DockerRepoResource.java#L887) and [workflows](https://github.com/dockstore/dockstore/blob/469261112f80613393000b1d3fe51b46bfc15498/dockstore-webservice/src/main/java/io/dockstore/webservice/resources/WorkflowResource.java#L1173) -- both methods just delete files they can find, but don't check if the files exist.
Not high priority.
┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-1039)
┆Issue Type: Story
┆Fix Versions: Dockstore 1.9
┆Sprint: Sprint 34 Isopods
┆Issue Number: DOCK-1039
| non_defect | deleting a non existent test parameter file does not return an error bug report if you delete a non existent test parameter file the api returns a status code it should return a or status code indicating that the file was not deleted this happens for both and both methods just delete files they can find but don t check if the files exist not high priority ┆issue is synchronized with this ┆issue type story ┆fix versions dockstore ┆sprint sprint isopods ┆issue number dock | 0 |
21,302 | 4,702,367,530 | IssuesEvent | 2016-10-13 01:46:45 | eris-ltd/coding | https://api.github.com/repos/eris-ltd/coding | closed | [Documentation] Detail porcelain docker throughout stack | documentation | **Problem:** There is a lot of reverse-engineering needed to tease out some failure with docker (e.g., docker-1.12-rc2 breaks eris, the switch to alpine, etc) ... the overall docker stack integrations are quite complex & IMO, becoming too distant from "here's how you'd do it with `docker run ...`". It's great that we can accomplish in 1-2 commands what would take 10-20 docker commands but _maybe_ things are getting a little out of hand...?
**Solution:** Dockument & maintain separate tests with porcelain docker for a handful of common commands. Some of these once existed in either eris-keys/mint-client...not sure where they went though. | 1.0 | [Documentation] Detail porcelain docker throughout stack - **Problem:** There is a lot of reverse-engineering needed to tease out some failure with docker (e.g., docker-1.12-rc2 breaks eris, the switch to alpine, etc) ... the overall docker stack integrations are quite complex & IMO, becoming too distant from "here's how you'd do it with `docker run ...`". It's great that we can accomplish in 1-2 commands what would take 10-20 docker commands but _maybe_ things are getting a little out of hand...?
**Solution:** Dockument & maintain separate tests with porcelain docker for a handful of common commands. Some of these once existed in either eris-keys/mint-client...not sure where they went though. | non_defect | detail porcelain docker throughout stack problem there is a lot of reverse engineering needed to tease out some failure with docker e g docker breaks eris the switch to alpine etc the overall docker stack integrations are quite complex imo becoming too distant from here s how you d do it with docker run it s great that we can accomplish in commands what would take docker commands but maybe things are getting a little out of hand solution dockument maintain separate tests with porcelain docker for a handful of common commands some of these once existed in either eris keys mint client not sure where they went though | 0 |
47,807 | 13,066,239,887 | IssuesEvent | 2020-07-30 21:16:52 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [DOMLauncher] fix/remove the TODOs (Trac #1201) | Migrated from Trac combo simulation defect | The TODOs should either be done or have tickets filed for them. We can track tickets, but not TODOs.
```text
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: store these constants somewhere else
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: it might be a good idea to turn this off and fall back on
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: use actual calibration data for the DOM in question
private/test/PMTResponseSimulatorTests.cxx: //TODO: should make a more deliberate choice of allowed ranges
private/test/PMTResponseSimulatorTests.cxx://TODO: Test jitter distribution with hit merging
private/test/PMTResponseSimulatorTests.cxx:/* TODO: rewrite this test
resources/docs/PMTRes.rst:TODO: Where does the earlyAfterPulseWeight parameterization come from?
resources/docs/PMTRes.rst:TODO: Is there an actual reference for Tom F.'s parameterization constants besides the pmt-simulator source code?
```
Migrated from https://code.icecube.wisc.edu/ticket/1201
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "The TODOs should either be done or have tickets filed for them. We can track tickets, but not TODOs.\n\n{{{\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t//TODO: store these constants somewhere else\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t\t//TODO: it might be a good idea to turn this off and fall back on\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t\t//TODO: use actual calibration data for the DOM in question\nprivate/test/PMTResponseSimulatorTests.cxx:\t//TODO: should make a more deliberate choice of allowed ranges\nprivate/test/PMTResponseSimulatorTests.cxx://TODO: Test jitter distribution with hit merging\nprivate/test/PMTResponseSimulatorTests.cxx:/* TODO: rewrite this test\nresources/docs/PMTRes.rst:TODO: Where does the earlyAfterPulseWeight parameterization come from?\nresources/docs/PMTRes.rst:TODO: Is there an actual reference for Tom F.'s parameterization constants besides the pmt-simulator source code?\n}}}",
"reporter": "david.schultz",
"cc": "cweaver",
"resolution": "invalid",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[DOMLauncher] fix/remove the TODOs",
"priority": "critical",
"keywords": "",
"time": "2015-08-19T18:59:39",
"milestone": "",
"owner": "sflis",
"type": "defect"
}
```
| 1.0 | [DOMLauncher] fix/remove the TODOs (Trac #1201) - The TODOs should either be done or have tickets filed for them. We can track tickets, but not TODOs.
```text
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: store these constants somewhere else
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: it might be a good idea to turn this off and fall back on
private/DOMLauncher/PMTResponseSimulator.cxx: //TODO: use actual calibration data for the DOM in question
private/test/PMTResponseSimulatorTests.cxx: //TODO: should make a more deliberate choice of allowed ranges
private/test/PMTResponseSimulatorTests.cxx://TODO: Test jitter distribution with hit merging
private/test/PMTResponseSimulatorTests.cxx:/* TODO: rewrite this test
resources/docs/PMTRes.rst:TODO: Where does the earlyAfterPulseWeight parameterization come from?
resources/docs/PMTRes.rst:TODO: Is there an actual reference for Tom F.'s parameterization constants besides the pmt-simulator source code?
```
Migrated from https://code.icecube.wisc.edu/ticket/1201
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "The TODOs should either be done or have tickets filed for them. We can track tickets, but not TODOs.\n\n{{{\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t//TODO: store these constants somewhere else\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t\t//TODO: it might be a good idea to turn this off and fall back on\nprivate/DOMLauncher/PMTResponseSimulator.cxx:\t\t//TODO: use actual calibration data for the DOM in question\nprivate/test/PMTResponseSimulatorTests.cxx:\t//TODO: should make a more deliberate choice of allowed ranges\nprivate/test/PMTResponseSimulatorTests.cxx://TODO: Test jitter distribution with hit merging\nprivate/test/PMTResponseSimulatorTests.cxx:/* TODO: rewrite this test\nresources/docs/PMTRes.rst:TODO: Where does the earlyAfterPulseWeight parameterization come from?\nresources/docs/PMTRes.rst:TODO: Is there an actual reference for Tom F.'s parameterization constants besides the pmt-simulator source code?\n}}}",
"reporter": "david.schultz",
"cc": "cweaver",
"resolution": "invalid",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[DOMLauncher] fix/remove the TODOs",
"priority": "critical",
"keywords": "",
"time": "2015-08-19T18:59:39",
"milestone": "",
"owner": "sflis",
"type": "defect"
}
```
| defect | fix remove the todos trac the todos should either be done or have tickets filed for them we can track tickets but not todos text private domlauncher pmtresponsesimulator cxx todo store these constants somewhere else private domlauncher pmtresponsesimulator cxx todo it might be a good idea to turn this off and fall back on private domlauncher pmtresponsesimulator cxx todo use actual calibration data for the dom in question private test pmtresponsesimulatortests cxx todo should make a more deliberate choice of allowed ranges private test pmtresponsesimulatortests cxx todo test jitter distribution with hit merging private test pmtresponsesimulatortests cxx todo rewrite this test resources docs pmtres rst todo where does the earlyafterpulseweight parameterization come from resources docs pmtres rst todo is there an actual reference for tom f s parameterization constants besides the pmt simulator source code migrated from json status closed changetime description the todos should either be done or have tickets filed for them we can track tickets but not todos n n nprivate domlauncher pmtresponsesimulator cxx t todo store these constants somewhere else nprivate domlauncher pmtresponsesimulator cxx t t todo it might be a good idea to turn this off and fall back on nprivate domlauncher pmtresponsesimulator cxx t t todo use actual calibration data for the dom in question nprivate test pmtresponsesimulatortests cxx t todo should make a more deliberate choice of allowed ranges nprivate test pmtresponsesimulatortests cxx todo test jitter distribution with hit merging nprivate test pmtresponsesimulatortests cxx todo rewrite this test nresources docs pmtres rst todo where does the earlyafterpulseweight parameterization come from nresources docs pmtres rst todo is there an actual reference for tom f s parameterization constants besides the pmt simulator source code n reporter david schultz cc cweaver resolution invalid ts component combo simulation summary fix remove the todos priority critical keywords time milestone owner sflis type defect | 1 |
218,868 | 17,027,660,306 | IssuesEvent | 2021-07-03 22:11:59 | FasterXML/jackson-dataformat-xml | https://api.github.com/repos/FasterXML/jackson-dataformat-xml | closed | XmlTokenStream does not respect FromXmlParser.Feature.EMPTY_ELEMENT_AS_NULL in 2.11 | test-needed | When deserializing the following XML, I would expect `<Element2>` to yield a `null` value in my POJO.
```xml
<MyObject proprietary-type="Array">
<Element1 proprietary-length="1" proprietary-type="String">Y</Element1>
<Element2 proprietary-length="2" proprietary-type="String"/>
<Element3 proprietary-length="1" proprietary-type="String">1</Element3>
<Element4 proprietary-length="2" proprietary-type="String">1</Element4>
</MyObject>
```
Given the following XML, I receive the results I'm expecting:
```xml
<MyObject proprietary-type="Array">
<Element1 proprietary-length="1" proprietary-type="String">Y</Element1>
<Element2/>
<Element3 proprietary-length="1" proprietary-type="String">1</Element3>
<Element4 proprietary-length="2" proprietary-type="String">1</Element4>
</MyObject>
```
It looks like this is due to #167 via [this code](https://github.com/FasterXML/jackson-dataformat-xml/blob/031aa5e10325929e16e24e2192a76442196fe74c/src/main/java/com/fasterxml/jackson/dataformat/xml/deser/XmlTokenStream.java#L319-L325), but I could be missing something. I would expect the linked code to perform a check for FromXmlParser.Feature.EMPTY_ELEMENT_AS_NULL rather than hard-default to empty String.
For context, I have no control over the XML I'm receiving and trying to deserialize—it's third-party. | 1.0 | XmlTokenStream does not respect FromXmlParser.Feature.EMPTY_ELEMENT_AS_NULL in 2.11 - When deserializing the following XML, I would expect `<Element2>` to yield a `null` value in my POJO.
```xml
<MyObject proprietary-type="Array">
<Element1 proprietary-length="1" proprietary-type="String">Y</Element1>
<Element2 proprietary-length="2" proprietary-type="String"/>
<Element3 proprietary-length="1" proprietary-type="String">1</Element3>
<Element4 proprietary-length="2" proprietary-type="String">1</Element4>
</MyObject>
```
Given the following XML, I receive the results I'm expecting:
```xml
<MyObject proprietary-type="Array">
<Element1 proprietary-length="1" proprietary-type="String">Y</Element1>
<Element2/>
<Element3 proprietary-length="1" proprietary-type="String">1</Element3>
<Element4 proprietary-length="2" proprietary-type="String">1</Element4>
</MyObject>
```
It looks like this is due to #167 via [this code](https://github.com/FasterXML/jackson-dataformat-xml/blob/031aa5e10325929e16e24e2192a76442196fe74c/src/main/java/com/fasterxml/jackson/dataformat/xml/deser/XmlTokenStream.java#L319-L325), but I could be missing something. I would expect the linked code to perform a check for FromXmlParser.Feature.EMPTY_ELEMENT_AS_NULL rather than hard-default to empty String.
For context, I have no control over the XML I'm receiving and trying to deserialize—it's third-party. | non_defect | xmltokenstream does not respect fromxmlparser feature empty element as null in when deserializing the following xml i would expect to yield a null value in my pojo xml y given the following xml i receive the results i m expecting xml y it looks like this is due to via but i could be missing something i would expect the linked code to perform a check for fromxmlparser feature empty element as null rather than hard default to empty string for context i have no control over the xml i m receiving and trying to deserialize—it s third party | 0 |
32,636 | 13,900,549,977 | IssuesEvent | 2020-10-20 00:37:00 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Review Vision Zero Zenhub estimates, velocity, etc. to inform Moped (& DTS) Product Ops | Product: Mobility Project Database Project: DTS Service Delivery Service: Product Type: Operations Workgroup: DTS | Having used [this estimate system](https://atd-dts.gitbook.io/wiki/product-ops/github-project-management#estimates) in Zenhub for VZ* work, let's
- Review velocity and other patterns over time
- Consider how this can inform Moped Sprint Planning going forward
- Consider whether we should modify our DTS-wide approach to estimates
| 2.0 | Review Vision Zero Zenhub estimates, velocity, etc. to inform Moped (& DTS) Product Ops - Having used [this estimate system](https://atd-dts.gitbook.io/wiki/product-ops/github-project-management#estimates) in Zenhub for VZ* work, let's
- Review velocity and other patterns over time
- Consider how this can inform Moped Sprint Planning going forward
- Consider whether we should modify our DTS-wide approach to estimates
| non_defect | review vision zero zenhub estimates velocity etc to inform moped dts product ops having used in zenhub for vz work let s review velocity and other patterns over time consider how this can inform moped sprint planning going forward consider whether we should modify our dts wide approach to estimates | 0 |
62,810 | 6,817,564,711 | IssuesEvent | 2017-11-07 00:02:54 | philiprbrenan/SlipsAndSpills | https://api.github.com/repos/philiprbrenan/SlipsAndSpills | closed | Psychedelic words on screen | waiting user test | This is already under discussion between us.
I find the flashing rainbow words really difficult to look at. They make me feel a bit dizzy and sick. I am worried about showing an app with the flashing words to anyone who is prone to migraine and epilepsy as flashing/strobing lights can trigger an attack. | 1.0 | Psychedelic words on screen - This is already under discussion between us.
I find the flashing rainbow words really difficult to look at. They make me feel a bit dizzy and sick. I am worried about showing an app with the flashing words to anyone who is prone to migraine and epilepsy as flashing/strobing lights can trigger an attack. | non_defect | psychedelic words on screen this is already under discussion between us i find the flashing rainbow words really difficult to look at they make me feel a bit dizzy and sick i am worried about showing an app with the flashing words to anyone who is prone to migraine and epilepsy as flashing strobing lights can trigger an attack | 0 |
578,969 | 17,169,385,792 | IssuesEvent | 2021-07-15 00:28:27 | googleapis/nodejs-spanner | https://api.github.com/repos/googleapis/nodejs-spanner | closed | Spanner Backups: "before all" hook for "should have completed a backup" failed | api: spanner flakybot: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 44895d2114ac4faa1b6304a5117d5a394de2bc48
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/51de58b0-2c9c-48c4-a4f9-68d266fbf414), [Sponge](http://sponge2/51de58b0-2c9c-48c4-a4f9-68d266fbf414)
status: failed
<details><summary>Test output</summary><br><pre>Timeout of 1600000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/build/system-test/spanner.js)
Error: Timeout of 1600000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/build/system-test/spanner.js)
at listOnTimeout (internal/timers.js:554:17)
at processTimers (internal/timers.js:497:7)</pre></details> | 1.0 | Spanner Backups: "before all" hook for "should have completed a backup" failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 44895d2114ac4faa1b6304a5117d5a394de2bc48
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/51de58b0-2c9c-48c4-a4f9-68d266fbf414), [Sponge](http://sponge2/51de58b0-2c9c-48c4-a4f9-68d266fbf414)
status: failed
<details><summary>Test output</summary><br><pre>Timeout of 1600000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/build/system-test/spanner.js)
Error: Timeout of 1600000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/build/system-test/spanner.js)
at listOnTimeout (internal/timers.js:554:17)
at processTimers (internal/timers.js:497:7)</pre></details> | non_defect | spanner backups before all hook for should have completed a backup failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves workspace build system test spanner js error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves workspace build system test spanner js at listontimeout internal timers js at processtimers internal timers js | 0 |
30,939 | 13,388,533,714 | IssuesEvent | 2020-09-02 17:30:47 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Meeting | Mobility Projects in Innovation Zones - Aug 31, 2020 | Service: Apps Type: Meeting Type: Open Data Workgroup: DTS | Meeting with Cap Metro to get a better understanding of Innovation Zones | 1.0 | Meeting | Mobility Projects in Innovation Zones - Aug 31, 2020 - Meeting with Cap Metro to get a better understanding of Innovation Zones | non_defect | meeting mobility projects in innovation zones aug meeting with cap metro to get a better understanding of innovation zones | 0 |
24,942 | 4,151,125,492 | IssuesEvent | 2016-06-15 19:34:17 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Can't use precomputed distances with `hierarchy.linkage` | defect scipy.cluster | [Here](http://nbviewer.jupyter.org/gist/olgabot/1341e73fd387db245f9ab1be3e65f2d6) is a minimal example using scipy v0.17.1 showing that when you use `hierarchy.linkage` to compute both the pairwise distances and the linkages, it works fine:
```python
ward_linkage = hierarchy.linkage(ica_components, method='ward', metric='euclidean')
```
(no error)
But if I precompute the pairwise distances:
```python
pairwise_dists = distance.pdist(ica_components, metric='euclidean')
hierarchy.linkage(pairwise_dists, method='ward', metric='euclidean')
```
Then I get this cryptic error about missing values. This seems especially weird because Ward clustering **does** use the euclidean metric, which is the default in the function.
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-3-b02b24ee06aa> in <module>()
1 pairwise_dists = distance.pdist(ica_components, metric='euclidean')
----> 2 hierarchy.linkage(pairwise_dists, method='ward', metric='euclidean')
/Users/olga/anaconda3/envs/single-cell-bioinformatics/lib/python3.5/site-packages/scipy/cluster/hierarchy.py in linkage(y, method, metric)
631 d = distance.num_obs_y(y)
632 if method not in _cpy_non_euclid_methods:
--> 633 raise ValueError("Valid methods when the raw observations are "
634 "omitted are 'single', 'complete', 'weighted', "
635 "and 'average'.")
ValueError: Valid methods when the raw observations are omitted are 'single', 'complete', 'weighted', and 'average'.
```
And it seems the [`hierarchy.linkage`](https://github.com/scipy/scipy/blob/v0.17.1/scipy/cluster/hierarchy.py#L633) code is supposed to check for whether this is already a distance array by checking the size, but maybe isn't working with exactly my input. However, [these lines](https://github.com/scipy/scipy/blob/master/scipy/cluster/hierarchy.py#L649) are changed in the `master` branch so maybe this has been fixed. Can you help with this?
Thanks,
Olga | 1.0 | Can't use precomputed distances with `hierarchy.linkage` - [Here](http://nbviewer.jupyter.org/gist/olgabot/1341e73fd387db245f9ab1be3e65f2d6) is a minimal example using scipy v0.17.1 showing that when you use `hierarchy.linkage` to compute both the pairwise distances and the linkages, it works fine:
```python
ward_linkage = hierarchy.linkage(ica_components, method='ward', metric='euclidean')
```
(no error)
But if I precompute the pairwise distances:
```python
pairwise_dists = distance.pdist(ica_components, metric='euclidean')
hierarchy.linkage(pairwise_dists, method='ward', metric='euclidean')
```
Then I get this cryptic error about missing values. This seems especially weird because Ward clustering **does** use the euclidean metric, which is the default in the function.
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-3-b02b24ee06aa> in <module>()
1 pairwise_dists = distance.pdist(ica_components, metric='euclidean')
----> 2 hierarchy.linkage(pairwise_dists, method='ward', metric='euclidean')
/Users/olga/anaconda3/envs/single-cell-bioinformatics/lib/python3.5/site-packages/scipy/cluster/hierarchy.py in linkage(y, method, metric)
631 d = distance.num_obs_y(y)
632 if method not in _cpy_non_euclid_methods:
--> 633 raise ValueError("Valid methods when the raw observations are "
634 "omitted are 'single', 'complete', 'weighted', "
635 "and 'average'.")
ValueError: Valid methods when the raw observations are omitted are 'single', 'complete', 'weighted', and 'average'.
```
And it seems the [`hierarchy.linkage`](https://github.com/scipy/scipy/blob/v0.17.1/scipy/cluster/hierarchy.py#L633) code is supposed to check for whether this is already a distance array by checking the size, but maybe isn't working with exactly my input. However, [these lines](https://github.com/scipy/scipy/blob/master/scipy/cluster/hierarchy.py#L649) are changed in the `master` branch so maybe this has been fixed. Can you help with this?
Thanks,
Olga | defect | can t use precomputed distances with hierarchy linkage is a minimal example using scipy showing that when you use hierarchy linkage to compute both the pairwise distances and the linkages it works fine python ward linkage hierarchy linkage ica components method ward metric euclidean no error but if i precompute the pairwise distances python pairwise dists distance pdist ica components metric euclidean hierarchy linkage pairwise dists method ward metric euclidean then i get this cryptic error about missing values this seems especially weird because ward clustering does use the euclidean metric which is the default in the function valueerror traceback most recent call last in pairwise dists distance pdist ica components metric euclidean hierarchy linkage pairwise dists method ward metric euclidean users olga envs single cell bioinformatics lib site packages scipy cluster hierarchy py in linkage y method metric d distance num obs y y if method not in cpy non euclid methods raise valueerror valid methods when the raw observations are omitted are single complete weighted and average valueerror valid methods when the raw observations are omitted are single complete weighted and average and it seems the code is supposed to check for whether this is already a distance array by checking the size but maybe isn t working with exactly my input however are changed in the master branch so maybe this has been fixed can you help with this thanks olga | 1 |
1,692 | 2,814,300,101 | IssuesEvent | 2015-05-18 19:14:32 | Semantic-Org/Semantic-UI | https://api.github.com/repos/Semantic-Org/Semantic-UI | closed | Custom fonts | Build Tools Usage Question | Hi,
I am wondering how to use custom fonts within a custom packaged theme, I have placed my fonts inside the `themes/mytheme/assets/fonts` however running gulp build doesn't output those files into the dist folder. Is there a solution to this? | 1.0 | Custom fonts - Hi,
I am wondering how to use custom fonts within a custom packaged theme, I have placed my fonts inside the `themes/mytheme/assets/fonts` however running gulp build doesn't output those files into the dist folder. Is there a solution to this? | non_defect | custom fonts hi i am wondering how to use custom fonts within a custom packaged theme i have placed my fonts inside the themes mytheme assets fonts however running gulp build doesn t output those files into the dist folder is there a solution to this | 0 |
21,510 | 3,512,814,954 | IssuesEvent | 2016-01-11 05:33:52 | anacrolix/imdbrt | https://api.github.com/repos/anacrolix/imdbrt | closed | Movie mismatch - Super K (2011) | auto-migrated Priority-Medium Type-Defect Usability | ```
What steps will reproduce the problem?
Go to http://www.imdb.com/title/tt2070831/
What is the expected output? What do you see instead?
It displays the score from the wrong movie:
http://www.rottentomatoes.com/m/supercop/
instead of 'movie not found' as I couldn't find the movie in RT website.
What version of the product are you using? On what operating system?
v 2.0 in chrome 15 on Windows XP
Please provide any additional information below.
This is a great extension, tks.
```
Original issue reported on code.google.com by `madgi...@gmail.com` on 1 Jan 2012 at 8:38 | 1.0 | Movie mismatch - Super K (2011) - ```
What steps will reproduce the problem?
Go to http://www.imdb.com/title/tt2070831/
What is the expected output? What do you see instead?
It displays the score from the wrong movie:
http://www.rottentomatoes.com/m/supercop/
instead of 'movie not found' as I couldn't find the movie in RT website.
What version of the product are you using? On what operating system?
v 2.0 in chrome 15 on Windows XP
Please provide any additional information below.
This is a great extension, tks.
```
Original issue reported on code.google.com by `madgi...@gmail.com` on 1 Jan 2012 at 8:38 | defect | movie mismatch super k what steps will reproduce the problem go to what is the expected output what do you see instead it displays the score from the wrong movie instead of movie not found as i couldn t find the movie in rt website what version of the product are you using on what operating system v in chrome on windows xp please provide any additional information below this is a great extension tks original issue reported on code google com by madgi gmail com on jan at | 1 |
14,164 | 24,569,055,124 | IssuesEvent | 2022-10-13 07:07:24 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | opened | Dependency update fails for yarrn > 1.x and binarysource=install | type:bug status:requirements priority-5-triage | ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
32.234.0
### If you're self-hosting Renovate, select which platform you are using.
Bitbucket Server
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
Starting with yarn 2.x, yarn can no longer be installed via NPM.
We are using the official renovate docker image and have it configured to use `binarysource=install`. This causes issues
for some projects, as they are using yarn 3.x, renovate detects this and tries to install it and fails when `install-tool.sh` cannot
install the requested yarn version.
I've opened a bug in the containerbase/base project, but they told me to open a bug with renovate.
https://github.com/containerbase/base/issues/558
### Relevant debug logs
<details><summary>Logs</summary>
```
/usr/local/buildpack/tools/v2/yarn-slim.sh: line 21: yarn: command not found
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository | 1.0 | Dependency update fails for yarrn > 1.x and binarysource=install - ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
32.234.0
### If you're self-hosting Renovate, select which platform you are using.
Bitbucket Server
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
Starting with yarn 2.x, yarn can no longer be installed via NPM.
We are using the official renovate docker image and have it configured to use `binarysource=install`. This causes issues
for some projects, as they are using yarn 3.x, renovate detects this and tries to install it and fails when `install-tool.sh` cannot
install the requested yarn version.
I've opened a bug in the containerbase/base project, but they told me to open a bug with renovate.
https://github.com/containerbase/base/issues/558
### Relevant debug logs
<details><summary>Logs</summary>
```
/usr/local/buildpack/tools/v2/yarn-slim.sh: line 21: yarn: command not found
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository | non_defect | dependency update fails for yarrn x and binarysource install how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run if you re self hosting renovate select which platform you are using bitbucket server if you re self hosting renovate tell us what version of the platform you run no response was this something which used to work for you and then stopped i never saw this working describe the bug starting with yarn x yarn can no longer be installed via npm we are using the official renovate docker image and have it configured to use binarysource install this causes issues for some projects as they are using yarn x renovate detects this and tries to install it and fails when install tool sh cannot install the requested yarn version i ve opened a bug in the containerbase base project but they told me to open a bug with renovate relevant debug logs logs usr local buildpack tools yarn slim sh line yarn command not found have you created a minimal reproduction repository no reproduction repository | 0 |
260,965 | 27,785,004,429 | IssuesEvent | 2023-03-17 01:56:08 | nlamirault/alan | https://api.github.com/repos/nlamirault/alan | opened | CVE-2023-28155 (Medium) detected in request-2.79.0.tgz, request-2.81.0.tgz | Mend: dependency security vulnerability | ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.79.0.tgz</b>, <b>request-2.81.0.tgz</b></p></summary>
<p>
<details><summary><b>request-2.79.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.79.0.tgz">https://registry.npmjs.org/request/-/request-2.79.0.tgz</a></p>
<p>Path to dependency file: /vendor/github.com/hashicorp/vault/ui/package.json</p>
<p>Path to vulnerable library: /vendor/github.com/hashicorp/vault/ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-favicon-1.0.0-beta.4.tgz (Root Library)
- broccoli-favicon-1.0.0.tgz
- favicons-4.8.6.tgz
- svg2png-3.0.1.tgz
- phantomjs-prebuilt-2.1.14.tgz
- :x: **request-2.79.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.81.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.81.0.tgz">https://registry.npmjs.org/request/-/request-2.81.0.tgz</a></p>
<p>Path to dependency file: /vendor/github.com/hashicorp/vault/ui/package.json</p>
<p>Path to vulnerable library: /vendor/github.com/hashicorp/vault/ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-qunit-4.0.0.tgz (Root Library)
- qunitjs-2.4.0.tgz
- chokidar-1.6.1.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- :x: **request-2.81.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/nlamirault/alan/commit/9060713df80212ee5546b36d1083fb607520eb0b">9060713df80212ee5546b36d1083fb607520eb0b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-28155 (Medium) detected in request-2.79.0.tgz, request-2.81.0.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.79.0.tgz</b>, <b>request-2.81.0.tgz</b></p></summary>
<p>
<details><summary><b>request-2.79.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.79.0.tgz">https://registry.npmjs.org/request/-/request-2.79.0.tgz</a></p>
<p>Path to dependency file: /vendor/github.com/hashicorp/vault/ui/package.json</p>
<p>Path to vulnerable library: /vendor/github.com/hashicorp/vault/ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-favicon-1.0.0-beta.4.tgz (Root Library)
- broccoli-favicon-1.0.0.tgz
- favicons-4.8.6.tgz
- svg2png-3.0.1.tgz
- phantomjs-prebuilt-2.1.14.tgz
- :x: **request-2.79.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.81.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.81.0.tgz">https://registry.npmjs.org/request/-/request-2.81.0.tgz</a></p>
<p>Path to dependency file: /vendor/github.com/hashicorp/vault/ui/package.json</p>
<p>Path to vulnerable library: /vendor/github.com/hashicorp/vault/ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-qunit-4.0.0.tgz (Root Library)
- qunitjs-2.4.0.tgz
- chokidar-1.6.1.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- :x: **request-2.81.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/nlamirault/alan/commit/9060713df80212ee5546b36d1083fb607520eb0b">9060713df80212ee5546b36d1083fb607520eb0b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in request tgz request tgz cve medium severity vulnerability vulnerable libraries request tgz request tgz request tgz simplified http request client library home page a href path to dependency file vendor github com hashicorp vault ui package json path to vulnerable library vendor github com hashicorp vault ui node modules request package json dependency hierarchy ember cli favicon beta tgz root library broccoli favicon tgz favicons tgz tgz phantomjs prebuilt tgz x request tgz vulnerable library request tgz simplified http request client library home page a href path to dependency file vendor github com hashicorp vault ui package json path to vulnerable library vendor github com hashicorp vault ui node modules request package json dependency hierarchy ember cli qunit tgz root library qunitjs tgz chokidar tgz fsevents tgz node pre gyp tgz x request tgz vulnerable library found in head commit a href found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
12,742 | 2,715,187,928 | IssuesEvent | 2015-04-10 11:09:47 | contao/core-bundle | https://api.github.com/repos/contao/core-bundle | closed | %locale% must not be in parameters.yml | defect | > <a href="https://github.com/aschempp"><img src="https://avatars.githubusercontent.com/u/1073273?v=3" align="left" width="42" height="42" hspace="10"></img></a> [Issue](https://github.com/contao/contao/issues/70) by @aschempp
Wednesday Mar 18, 2015 at 18:07 GMT
We have discussed this before. The `%locale%` is the system fallback for everything like language files and therefore should not be configurable by the user. Otherwise if I choose to use `de` as default, modules without a german translation would not provide any labels.
| 1.0 | %locale% must not be in parameters.yml - > <a href="https://github.com/aschempp"><img src="https://avatars.githubusercontent.com/u/1073273?v=3" align="left" width="42" height="42" hspace="10"></img></a> [Issue](https://github.com/contao/contao/issues/70) by @aschempp
Wednesday Mar 18, 2015 at 18:07 GMT
We have discussed this before. The `%locale%` is the system fallback for everything like language files and therefore should not be configurable by the user. Otherwise if I choose to use `de` as default, modules without a german translation would not provide any labels.
| defect | locale must not be in parameters yml by aschempp wednesday mar at gmt we have discussed this before the locale is the system fallback for everything like language files and therefore should not be configurable by the user otherwise if i choose to use de as default modules without a german translation would not provide any labels | 1 |
75,092 | 25,527,824,466 | IssuesEvent | 2022-11-29 05:00:32 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Caps lock blocks `ctrl+k` for search | T-Defect S-Minor A-Shortcuts O-Occasional A-New-Search-Experience | ### Steps to reproduce
caps lock on, `ctrl+k`
### Outcome
Without caps lock things work fine, I see no reason not to have both `ctrl+k` and `ctrl+K` both work for search, or if possible to detect specifically if caps lock is on so `ctrl + shift+ k` could be mapped to another command/
### Operating system
Pop!_OS
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Caps lock blocks `ctrl+k` for search - ### Steps to reproduce
caps lock on, `ctrl+k`
### Outcome
Without caps lock things work fine, I see no reason not to have both `ctrl+k` and `ctrl+K` both work for search, or if possible to detect specifically if caps lock is on so `ctrl + shift+ k` could be mapped to another command/
### Operating system
Pop!_OS
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No | defect | caps lock blocks ctrl k for search steps to reproduce caps lock on ctrl k outcome without caps lock things work fine i see no reason not to have both ctrl k and ctrl k both work for search or if possible to detect specifically if caps lock is on so ctrl shift k could be mapped to another command operating system pop os application version no response how did you install the app no response homeserver no response will you send logs no | 1 |
314,577 | 23,528,973,224 | IssuesEvent | 2022-08-19 13:37:44 | vegaprotocol/specs | https://api.github.com/repos/vegaprotocol/specs | opened | New spec to detail behaviour of the candles subscription | documentation specs | To help develop and test the candles data, we need to write a spec we can agree on and use as the reference for future work and testing. | 1.0 | New spec to detail behaviour of the candles subscription - To help develop and test the candles data, we need to write a spec we can agree on and use as the reference for future work and testing. | non_defect | new spec to detail behaviour of the candles subscription to help develop and test the candles data we need to write a spec we can agree on and use as the reference for future work and testing | 0 |
59,943 | 17,023,295,844 | IssuesEvent | 2021-07-03 01:17:13 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Windows 0.12 beta2 full zoom drawing bug yahoo images don't scale | Component: merkaartor Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 1.51am, Wednesday, 17th September 2008]**
If you zoom to a level more than yahoo supports, you get the previous image tiled to fill the space. This is closer to right than 0.11 had (squares of higher level images). Now what is needed is to scale the image to match the zoom of the map (what 0.11 did before it redrew with the squares)
This looks like a work in progress, and is really a variant of a previous 0.11 bug | 1.0 | Windows 0.12 beta2 full zoom drawing bug yahoo images don't scale - **[Submitted to the original trac issue database at 1.51am, Wednesday, 17th September 2008]**
If you zoom to a level more than yahoo supports, you get the previous image tiled to fill the space. This is closer to right than 0.11 had (squares of higher level images). Now what is needed is to scale the image to match the zoom of the map (what 0.11 did before it redrew with the squares)
This looks like a work in progress, and is really a variant of a previous 0.11 bug | defect | windows full zoom drawing bug yahoo images don t scale if you zoom to a level more than yahoo supports you get the previous image tiled to fill the space this is closer to right than had squares of higher level images now what is needed is to scale the image to match the zoom of the map what did before it redrew with the squares this looks like a work in progress and is really a variant of a previous bug | 1 |
127,177 | 17,197,162,355 | IssuesEvent | 2021-07-16 19:18:42 | willgarrett64/chess | https://api.github.com/repos/willgarrett64/chess | opened | Improve user input for Node | design enhancement nice to have | Currently playing the game through the console, running on Node.js, it is quite clunky and looks pretty poor - see image below.

Proposed changes include:
- Instructions printed before starting the game
- Option to input moves as algebraic notation or as start square/end square (as it is currently)
- Option to change selected piece to move (currently once a piece has been selected, it must be moved)
- Display captured pieces for each player
- Display turn number
- If possible, clear the terminal each move so the board appears to update rather than logging a new board
When this is tackled, it might be best to break down into separate issues.
| 1.0 | Improve user input for Node - Currently playing the game through the console, running on Node.js, it is quite clunky and looks pretty poor - see image below.

Proposed changes include:
- Instructions printed before starting the game
- Option to input moves as algebraic notation or as start square/end square (as it is currently)
- Option to change selected piece to move (currently once a piece has been selected, it must be moved)
- Display captured pieces for each player
- Display turn number
- If possible, clear the terminal each move so the board appears to update rather than logging a new board
When this is tackled, it might be best to break down into separate issues.
| non_defect | improve user input for node currently playing the game through the console running on node js it is quite clunky and looks pretty poor see image below proposed changes include instructions printed before starting the game option to input moves as algebraic notation or as start square end square as it is currently option to change selected piece to move currently once a piece has been selected it must be moved display captured pieces for each player display turn number if possible clear the terminal each move so the board appears to update rather than logging a new board when this is tackled it might be best to break down into separate issues | 0 |
393,705 | 27,005,058,718 | IssuesEvent | 2023-02-10 10:57:43 | HyphaApp/hypha | https://api.github.com/repos/HyphaApp/hypha | closed | Duplicate administrators and developers docs | Type: Documentation | ### Description
Hypha's documentation currently has sections for administrators and developers with overlapping topics. Could we please define administrators and developers in the documentation?
I think there is a great deal of value with what we are currently presenting. Perhaps both sets of documentation are accurately categorized, and there is little overlap?
Similar themes in administrators and developers:
- [ ] 2FA
administrators: https://docs.hypha.app/administrators/configuration/#if-hypha-should-enforce-2fa-for-all-users
developers: https://docs.hypha.app/developers/security2fa/
- [ ] Staff Lead permissions
administrators: https://docs.hypha.app/administrators/configuration/
developers: https://docs.hypha.app/developers/userrolespermissions/
- [ ] Deployment
administrators:
https://docs.hypha.app/administrators/deployment/
https://docs.hypha.app/administrators/configuration/#on-heroku-set-to-true-if-deploying-to-heroku
developers: https://docs.hypha.app/developers/localdeveldocker/
- [ ] Notifications
administrators:
https://docs.hypha.app/administrators/configuration/#send-out-e-mail-slack-messages-etc-from-hypha-set-to-true-for-production
https://docs.hypha.app/administrators/configuration/#slack-settings
https://docs.hypha.app/administrators/configuration/#staff-e-mail-domain-used-for-oauth2-whitelist-default-value-and-staff-account-creation
developers: https://docs.hypha.app/developers/notifications/
| 1.0 | Duplicate administrators and developers docs - ### Description
Hypha's documentation currently has sections for administrators and developers with overlapping topics. Could we please define administrators and developers in the documentation?
I think there is a great deal of value with what we are currently presenting. Perhaps both sets of documentation are accurately categorized, and there is little overlap?
Similar themes in administrators and developers:
- [ ] 2FA
administrators: https://docs.hypha.app/administrators/configuration/#if-hypha-should-enforce-2fa-for-all-users
developers: https://docs.hypha.app/developers/security2fa/
- [ ] Staff Lead permissions
administrators: https://docs.hypha.app/administrators/configuration/
developers: https://docs.hypha.app/developers/userrolespermissions/
- [ ] Deployment
administrators:
https://docs.hypha.app/administrators/deployment/
https://docs.hypha.app/administrators/configuration/#on-heroku-set-to-true-if-deploying-to-heroku
developers: https://docs.hypha.app/developers/localdeveldocker/
- [ ] Notifications
administrators:
https://docs.hypha.app/administrators/configuration/#send-out-e-mail-slack-messages-etc-from-hypha-set-to-true-for-production
https://docs.hypha.app/administrators/configuration/#slack-settings
https://docs.hypha.app/administrators/configuration/#staff-e-mail-domain-used-for-oauth2-whitelist-default-value-and-staff-account-creation
developers: https://docs.hypha.app/developers/notifications/
| non_defect | duplicate administrators and developers docs description hypha s documentation currently has sections for administrators and developers with overlapping topics could we please define administrators and developers in the documentation i think there is a great deal of value with what we are currently presenting perhaps both sets of documentation are accurately categorized and there is little overlap similar themes in administrators and developers administrators developers staff lead permissions administrators developers deployment administrators developers notifications administrators developers | 0 |
63,928 | 18,059,610,303 | IssuesEvent | 2021-09-20 12:38:21 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | /join explodes on develop | T-Defect S-Major X-Release-Blocker A-Slash-Commands O-Uncommon | ### Steps to reproduce
1. Type `/join #thirdroom-dev:matrix.org`
2. Press enter
### What happened?
### What did you expect?
Room Join spinner
### What happened?
Boom? As in full crash.
### Operating system
Windows
### Browser information
Firefox
### URL for webapp
develop.element.io
### Homeserver
nordgedanken.dev
### Have you submitted a rageshake?
Yes | 1.0 | /join explodes on develop - ### Steps to reproduce
1. Type `/join #thirdroom-dev:matrix.org`
2. Press enter
### What happened?
### What did you expect?
Room Join spinner
### What happened?
Boom? As in full crash.
### Operating system
Windows
### Browser information
Firefox
### URL for webapp
develop.element.io
### Homeserver
nordgedanken.dev
### Have you submitted a rageshake?
Yes | defect | join explodes on develop steps to reproduce type join thirdroom dev matrix org press enter what happened what did you expect room join spinner what happened boom as in full crash operating system windows browser information firefox url for webapp develop element io homeserver nordgedanken dev have you submitted a rageshake yes | 1 |
38,352 | 15,647,304,195 | IssuesEvent | 2021-03-23 02:57:57 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | [QUERY] Message processor stops processing messages. | Client Service Bus customer-reported needs-team-attention question | **Query/Question**
I am hosting a message receiver and session message receiver in a aspnet core hosted service. I am noticing the receivers going into long (very long) periods of inactivity where no messages are being processed. Restarting the consumers or adding new consumers has no effect. This issue effects multiple queues sometimes at the same time. Sometimes one queue will start processing again and deliver messages to another queue that won't process anything.
I am also periodically seeing this error:
Azure.Messaging.ServiceBus.ServiceBusException
The lock supplied is invalid. Either the lock expired, or the message has already been removed from the queue, or was received by a different receiver instance. (MessageLockLost)
This happens when dead lettering a message. Theres also a few instances of TaskCanceledException when trying to dead letter a message.
Below is the hosted service that runs the processor.
``` csharp
public class MessageReceiver : IHostedService
{
private readonly IContainer _container;
private readonly IDatabase _database;
private readonly ServiceBusProcessor _receiver;
private readonly ServiceBusProcessor _deadletterReceiver;
private readonly ILogger<MessageReceiver> _logger;
private const string MessageMimeType = "application/bson";
/// <summary>
/// Creates a message receiver hosted service for the desired queue and/or subscription.
/// </summary>
/// <param name="container">The dependency container. When a new message is received a new DI scope is created and the message and all dependencies are handled within that scope.</param>
/// <param name="serviceBusClient">The service bus client used to create a message processor.</param>
/// <param name="queueName">The queue to listen for new messages on.</param>
/// <param name="subscription">The optional subscription to target.</param>
public MessageReceiver(IContainer container, ServiceBusClient serviceBusClient, IDatabase database, string queueName, string subscription = null, int maxConcurrency = 1)
{
_container = container;
_database = database;
_logger = container.GetService<ILogger<MessageReceiver>>();
var options = new ServiceBusProcessorOptions
{
AutoCompleteMessages = false,
MaxConcurrentCalls = maxConcurrency,
//PrefetchCount = 10
};
if (subscription == null)
{
_receiver = serviceBusClient.CreateProcessor(queueName, options);
_deadletterReceiver = serviceBusClient.CreateProcessor($"{queueName}/$deadletterqueue");
}
else
{
_receiver = serviceBusClient.CreateProcessor(queueName, subscription, options);
_deadletterReceiver = serviceBusClient.CreateProcessor($"{queueName}/Subscriptions/{subscription}/$deadletterqueue");
}
}
private async Task DeadletterReceiverOnProcessMessageAsync(ProcessMessageEventArgs arg)
{
using var nestedContainer = _container.GetNestedContainer();
var logger = nestedContainer.GetInstance<ILogger<MessageSessionReceiver>>();
var correlationId = arg.Message.CorrelationId;
var jobId = arg.Message.ApplicationProperties["JobId"].ToString();
var fileId = arg.Message.ApplicationProperties["FileId"].ToString();
using (logger.BeginScope(new Dictionary<string, object>
{
{"CorrelationId", correlationId},
{"JobId", jobId},
{"FileId", fileId},
}))
{
try
{
logger.LogDebug("Processing failed message.");
arg.Message.ApplicationProperties.TryGetValue("PageNo", out var pageNo);
arg.Message.ApplicationProperties.TryGetValue("LogicalPageNo", out var logicalPageNo);
var context = nestedContainer.GetInstance<IContext>();
var failFile = arg.Message.ApplicationProperties.TryGetValue("DocumentType", out var documentType)
&& (string) documentType == "MultiPageDocument";
if (failFile || (pageNo == null && logicalPageNo == null))
{
await context.WriteAsync(jobId, new FailFileCommand() {FileId = fileId});
//file level processing has failed (probably the get pdf info/convert to pdf stage)
//so formally abort the file
await _database.StringSetAsync($"$Abort_{fileId}", "1", TimeSpan.FromDays(1));
}
else
{
await context.WriteAsync(jobId, new ChangePageStateCommand
{
FileId = fileId,
PageNumbers = new[] { int.Parse(logicalPageNo?.ToString() ?? pageNo?.ToString()) },
State = BatchFileProcessingEntity.OperationState.Failed
});
await _database.StringSetAsync($"$Abort_{fileId}_{logicalPageNo ?? pageNo}", "1",
TimeSpan.FromDays(1));
}
await arg.CompleteMessageAsync(arg.Message, arg.CancellationToken);
}
catch (Exception ex)
{
logger.LogError(ex, "Error occurred processing message failure.");
}
finally
{
//messages remaining in the DLQ should be investigated
}
}
}
public async Task StartAsync(CancellationToken cancellationToken)
{
_receiver.ProcessMessageAsync += ReceiverOnProcessMessageAsync;
_receiver.ProcessErrorAsync += ReceiverOnProcessErrorAsync;
_deadletterReceiver.ProcessMessageAsync += DeadletterReceiverOnProcessMessageAsync;
_deadletterReceiver.ProcessErrorAsync += ReceiverOnProcessErrorAsync;
await Task.WhenAll(_deadletterReceiver.StartProcessingAsync(cancellationToken), _receiver.StartProcessingAsync(cancellationToken));
_logger.LogInformation("Receiver listening to {Queue}", _receiver.EntityPath);
}
public async Task StopAsync(CancellationToken cancellationToken)
{
await _receiver.StopProcessingAsync(cancellationToken);
await _receiver.CloseAsync(cancellationToken);
}
private Task ReceiverOnProcessErrorAsync(ProcessErrorEventArgs arg)
{
_logger.LogError(arg.Exception, "SessionMessageProcessing Failed");
return Task.CompletedTask;
}
private async Task ReceiverOnProcessMessageAsync(ProcessMessageEventArgs arg)
{
try
{
var message = arg.Message;
var cancellationToken = arg.CancellationToken;
if (arg.Message.ApplicationProperties.TryGetValue("FileId", out var fileId))
{
var checkResult = false;
await Policy.Handle<RedisConnectionException>()
.WaitAndRetryAsync(10, (count) => TimeSpan.FromMilliseconds(count * 250))
.ExecuteAsync(async () =>
{
if (await _database.KeyExistsAsync($"$Abort_{fileId}"))
{
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
if (arg.Message.ApplicationProperties.TryGetValue("LogicalPageNo", out var logicalPageNo))
{
if (await _database.KeyExistsAsync($"$Abort_{fileId}_{logicalPageNo}"))
{
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
}
if (message.ApplicationProperties.ContainsKey("AbortFileId"))
{
await _database.StringSetAsync(
$"$Abort_{message.ApplicationProperties["AbortFileId"]}",
"1", TimeSpan.FromDays(1));
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
});
if (checkResult)
{
return;
}
}
arg.Message.ApplicationProperties.TryGetValue("OrganizationId", out var organisationId);
arg.Message.ApplicationProperties.TryGetValue("ProjectId", out var projectId);
var (payloadTypeName, userId, correlationId) = message.GetCommonMessageProperties();
var otherProperties = message.GetOtherMessageProperties();
var logValues = new Dictionary<string, object>
{
{"CorrelationId", correlationId}
};
foreach (var pair in otherProperties)
{
logValues[pair.Key] = pair.Value;
}
using (var nestedContainer = _container.GetNestedContainer())
{
nestedContainer.Inject(typeof(ICorrelationId), new CorrelationId(correlationId), true);
var logger = nestedContainer.GetInstance<ILogger<MessageReceiver>>();
using (logger.BeginScope(logValues))
{
try
{
//this validation is done here so all logging is scoped properly.
if (payloadTypeName == null || userId == null)
{
logger.LogDebug("A message was received with no specified payload type or no security token.");
await arg.DeadLetterMessageAsync(arg.Message, "Message is not in a format MessageReceiver can process.",
cancellationToken: cancellationToken);
return;
}
var claimList = new List<Claim>
{
new Claim(ClaimTypes.NameIdentifier, userId)
};
if (organisationId != null)
{
claimList.Add(new Claim("OrganizationId", organisationId.ToString()));
}
if (projectId != null)
{
claimList.Add(new Claim("ProjectId", projectId.ToString()));
}
var user = new ClaimsPrincipal(new ClaimsIdentity(claimList));
using (logger.BeginScope(new Dictionary<string, object>
{
{"UserId", userId}
}))
{
var messageBytes = message.Body.ToArray();
var request = RequestTypeAggregator.Deserialize(messageBytes, payloadTypeName, message.ContentType);
if (request == null)
{
logger.LogDebug("Failed to deserialize request payload. Type {payloadTypeName}", payloadTypeName);
await arg.DeadLetterMessageAsync(message, "Request payload was invalid.",
cancellationToken: cancellationToken);
return;
}
nestedContainer.Inject(typeof(IPrincipal), user, true);
nestedContainer.Inject(user, true);
var mediator = nestedContainer.GetInstance<IMediator>();
try
{
await mediator.Send(request, cancellationToken);
await arg.CompleteMessageAsync(message, cancellationToken);
}
catch
{
// any upstream errors in the mediator are logged by the logging behaviour and are not our concern here
await arg.DeadLetterMessageAsync(message, "Request execution failed with error.",
cancellationToken: cancellationToken);
}
}
}
catch (Exception ex)
{
logger.LogError(ex, "Exception with message handling.");
await arg.DeadLetterMessageAsync(message, "Unexpected error processing request.", cancellationToken: cancellationToken);
}
}
}
}
catch (Exception ex)
{
//this logger has no scope
_logger.LogError(ex, "Exception with message handling.");
await arg.DeadLetterMessageAsync(arg.Message, "Unexpected error processing request.", cancellationToken: arg.CancellationToken);
}
}
}
```
**Environment:**
- Azure.Messaging.ServiceBus 7.0.1
- Running in a azure k8s pod from linux aspnet core docker image
| 1.0 | [QUERY] Message processor stops processing messages. - **Query/Question**
I am hosting a message receiver and session message receiver in a aspnet core hosted service. I am noticing the receivers going into long (very long) periods of inactivity where no messages are being processed. Restarting the consumers or adding new consumers has no effect. This issue effects multiple queues sometimes at the same time. Sometimes one queue will start processing again and deliver messages to another queue that won't process anything.
I am also periodically seeing this error:
Azure.Messaging.ServiceBus.ServiceBusException
The lock supplied is invalid. Either the lock expired, or the message has already been removed from the queue, or was received by a different receiver instance. (MessageLockLost)
This happens when dead lettering a message. Theres also a few instances of TaskCanceledException when trying to dead letter a message.
Below is the hosted service that runs the processor.
``` csharp
public class MessageReceiver : IHostedService
{
private readonly IContainer _container;
private readonly IDatabase _database;
private readonly ServiceBusProcessor _receiver;
private readonly ServiceBusProcessor _deadletterReceiver;
private readonly ILogger<MessageReceiver> _logger;
private const string MessageMimeType = "application/bson";
/// <summary>
/// Creates a message receiver hosted service for the desired queue and/or subscription.
/// </summary>
/// <param name="container">The dependency container. When a new message is received a new DI scope is created and the message and all dependencies are handled within that scope.</param>
/// <param name="serviceBusClient">The service bus client used to create a message processor.</param>
/// <param name="queueName">The queue to listen for new messages on.</param>
/// <param name="subscription">The optional subscription to target.</param>
public MessageReceiver(IContainer container, ServiceBusClient serviceBusClient, IDatabase database, string queueName, string subscription = null, int maxConcurrency = 1)
{
_container = container;
_database = database;
_logger = container.GetService<ILogger<MessageReceiver>>();
var options = new ServiceBusProcessorOptions
{
AutoCompleteMessages = false,
MaxConcurrentCalls = maxConcurrency,
//PrefetchCount = 10
};
if (subscription == null)
{
_receiver = serviceBusClient.CreateProcessor(queueName, options);
_deadletterReceiver = serviceBusClient.CreateProcessor($"{queueName}/$deadletterqueue");
}
else
{
_receiver = serviceBusClient.CreateProcessor(queueName, subscription, options);
_deadletterReceiver = serviceBusClient.CreateProcessor($"{queueName}/Subscriptions/{subscription}/$deadletterqueue");
}
}
private async Task DeadletterReceiverOnProcessMessageAsync(ProcessMessageEventArgs arg)
{
using var nestedContainer = _container.GetNestedContainer();
var logger = nestedContainer.GetInstance<ILogger<MessageSessionReceiver>>();
var correlationId = arg.Message.CorrelationId;
var jobId = arg.Message.ApplicationProperties["JobId"].ToString();
var fileId = arg.Message.ApplicationProperties["FileId"].ToString();
using (logger.BeginScope(new Dictionary<string, object>
{
{"CorrelationId", correlationId},
{"JobId", jobId},
{"FileId", fileId},
}))
{
try
{
logger.LogDebug("Processing failed message.");
arg.Message.ApplicationProperties.TryGetValue("PageNo", out var pageNo);
arg.Message.ApplicationProperties.TryGetValue("LogicalPageNo", out var logicalPageNo);
var context = nestedContainer.GetInstance<IContext>();
var failFile = arg.Message.ApplicationProperties.TryGetValue("DocumentType", out var documentType)
&& (string) documentType == "MultiPageDocument";
if (failFile || (pageNo == null && logicalPageNo == null))
{
await context.WriteAsync(jobId, new FailFileCommand() {FileId = fileId});
//file level processing has failed (probably the get pdf info/convert to pdf stage)
//so formally abort the file
await _database.StringSetAsync($"$Abort_{fileId}", "1", TimeSpan.FromDays(1));
}
else
{
await context.WriteAsync(jobId, new ChangePageStateCommand
{
FileId = fileId,
PageNumbers = new[] { int.Parse(logicalPageNo?.ToString() ?? pageNo?.ToString()) },
State = BatchFileProcessingEntity.OperationState.Failed
});
await _database.StringSetAsync($"$Abort_{fileId}_{logicalPageNo ?? pageNo}", "1",
TimeSpan.FromDays(1));
}
await arg.CompleteMessageAsync(arg.Message, arg.CancellationToken);
}
catch (Exception ex)
{
logger.LogError(ex, "Error occurred processing message failure.");
}
finally
{
//messages remaining in the DLQ should be investigated
}
}
}
public async Task StartAsync(CancellationToken cancellationToken)
{
_receiver.ProcessMessageAsync += ReceiverOnProcessMessageAsync;
_receiver.ProcessErrorAsync += ReceiverOnProcessErrorAsync;
_deadletterReceiver.ProcessMessageAsync += DeadletterReceiverOnProcessMessageAsync;
_deadletterReceiver.ProcessErrorAsync += ReceiverOnProcessErrorAsync;
await Task.WhenAll(_deadletterReceiver.StartProcessingAsync(cancellationToken), _receiver.StartProcessingAsync(cancellationToken));
_logger.LogInformation("Receiver listening to {Queue}", _receiver.EntityPath);
}
public async Task StopAsync(CancellationToken cancellationToken)
{
await _receiver.StopProcessingAsync(cancellationToken);
await _receiver.CloseAsync(cancellationToken);
}
private Task ReceiverOnProcessErrorAsync(ProcessErrorEventArgs arg)
{
_logger.LogError(arg.Exception, "SessionMessageProcessing Failed");
return Task.CompletedTask;
}
private async Task ReceiverOnProcessMessageAsync(ProcessMessageEventArgs arg)
{
try
{
var message = arg.Message;
var cancellationToken = arg.CancellationToken;
if (arg.Message.ApplicationProperties.TryGetValue("FileId", out var fileId))
{
var checkResult = false;
await Policy.Handle<RedisConnectionException>()
.WaitAndRetryAsync(10, (count) => TimeSpan.FromMilliseconds(count * 250))
.ExecuteAsync(async () =>
{
if (await _database.KeyExistsAsync($"$Abort_{fileId}"))
{
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
if (arg.Message.ApplicationProperties.TryGetValue("LogicalPageNo", out var logicalPageNo))
{
if (await _database.KeyExistsAsync($"$Abort_{fileId}_{logicalPageNo}"))
{
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
}
if (message.ApplicationProperties.ContainsKey("AbortFileId"))
{
await _database.StringSetAsync(
$"$Abort_{message.ApplicationProperties["AbortFileId"]}",
"1", TimeSpan.FromDays(1));
await arg.CompleteMessageAsync(message, cancellationToken);
checkResult = true;
return;
}
});
if (checkResult)
{
return;
}
}
arg.Message.ApplicationProperties.TryGetValue("OrganizationId", out var organisationId);
arg.Message.ApplicationProperties.TryGetValue("ProjectId", out var projectId);
var (payloadTypeName, userId, correlationId) = message.GetCommonMessageProperties();
var otherProperties = message.GetOtherMessageProperties();
var logValues = new Dictionary<string, object>
{
{"CorrelationId", correlationId}
};
foreach (var pair in otherProperties)
{
logValues[pair.Key] = pair.Value;
}
using (var nestedContainer = _container.GetNestedContainer())
{
nestedContainer.Inject(typeof(ICorrelationId), new CorrelationId(correlationId), true);
var logger = nestedContainer.GetInstance<ILogger<MessageReceiver>>();
using (logger.BeginScope(logValues))
{
try
{
//this validation is done here so all logging is scoped properly.
if (payloadTypeName == null || userId == null)
{
logger.LogDebug("A message was received with no specified payload type or no security token.");
await arg.DeadLetterMessageAsync(arg.Message, "Message is not in a format MessageReceiver can process.",
cancellationToken: cancellationToken);
return;
}
var claimList = new List<Claim>
{
new Claim(ClaimTypes.NameIdentifier, userId)
};
if (organisationId != null)
{
claimList.Add(new Claim("OrganizationId", organisationId.ToString()));
}
if (projectId != null)
{
claimList.Add(new Claim("ProjectId", projectId.ToString()));
}
var user = new ClaimsPrincipal(new ClaimsIdentity(claimList));
using (logger.BeginScope(new Dictionary<string, object>
{
{"UserId", userId}
}))
{
var messageBytes = message.Body.ToArray();
var request = RequestTypeAggregator.Deserialize(messageBytes, payloadTypeName, message.ContentType);
if (request == null)
{
logger.LogDebug("Failed to deserialize request payload. Type {payloadTypeName}", payloadTypeName);
await arg.DeadLetterMessageAsync(message, "Request payload was invalid.",
cancellationToken: cancellationToken);
return;
}
nestedContainer.Inject(typeof(IPrincipal), user, true);
nestedContainer.Inject(user, true);
var mediator = nestedContainer.GetInstance<IMediator>();
try
{
await mediator.Send(request, cancellationToken);
await arg.CompleteMessageAsync(message, cancellationToken);
}
catch
{
// any upstream errors in the mediator are logged by the logging behaviour and are not our concern here
await arg.DeadLetterMessageAsync(message, "Request execution failed with error.",
cancellationToken: cancellationToken);
}
}
}
catch (Exception ex)
{
logger.LogError(ex, "Exception with message handling.");
await arg.DeadLetterMessageAsync(message, "Unexpected error processing request.", cancellationToken: cancellationToken);
}
}
}
}
catch (Exception ex)
{
//this logger has no scope
_logger.LogError(ex, "Exception with message handling.");
await arg.DeadLetterMessageAsync(arg.Message, "Unexpected error processing request.", cancellationToken: arg.CancellationToken);
}
}
}
```
**Environment:**
- Azure.Messaging.ServiceBus 7.0.1
- Running in a azure k8s pod from linux aspnet core docker image
| non_defect | message processor stops processing messages query question i am hosting a message receiver and session message receiver in a aspnet core hosted service i am noticing the receivers going into long very long periods of inactivity where no messages are being processed restarting the consumers or adding new consumers has no effect this issue effects multiple queues sometimes at the same time sometimes one queue will start processing again and deliver messages to another queue that won t process anything i am also periodically seeing this error azure messaging servicebus servicebusexception the lock supplied is invalid either the lock expired or the message has already been removed from the queue or was received by a different receiver instance messagelocklost this happens when dead lettering a message theres also a few instances of taskcanceledexception when trying to dead letter a message below is the hosted service that runs the processor csharp public class messagereceiver ihostedservice private readonly icontainer container private readonly idatabase database private readonly servicebusprocessor receiver private readonly servicebusprocessor deadletterreceiver private readonly ilogger logger private const string messagemimetype application bson creates a message receiver hosted service for the desired queue and or subscription the dependency container when a new message is received a new di scope is created and the message and all dependencies are handled within that scope the service bus client used to create a message processor the queue to listen for new messages on the optional subscription to target public messagereceiver icontainer container servicebusclient servicebusclient idatabase database string queuename string subscription null int maxconcurrency container container database database logger container getservice var options new servicebusprocessoroptions autocompletemessages false maxconcurrentcalls maxconcurrency prefetchcount if subscription null receiver servicebusclient createprocessor queuename options deadletterreceiver servicebusclient createprocessor queuename deadletterqueue else receiver servicebusclient createprocessor queuename subscription options deadletterreceiver servicebusclient createprocessor queuename subscriptions subscription deadletterqueue private async task deadletterreceiveronprocessmessageasync processmessageeventargs arg using var nestedcontainer container getnestedcontainer var logger nestedcontainer getinstance var correlationid arg message correlationid var jobid arg message applicationproperties tostring var fileid arg message applicationproperties tostring using logger beginscope new dictionary correlationid correlationid jobid jobid fileid fileid try logger logdebug processing failed message arg message applicationproperties trygetvalue pageno out var pageno arg message applicationproperties trygetvalue logicalpageno out var logicalpageno var context nestedcontainer getinstance var failfile arg message applicationproperties trygetvalue documenttype out var documenttype string documenttype multipagedocument if failfile pageno null logicalpageno null await context writeasync jobid new failfilecommand fileid fileid file level processing has failed probably the get pdf info convert to pdf stage so formally abort the file await database stringsetasync abort fileid timespan fromdays else await context writeasync jobid new changepagestatecommand fileid fileid pagenumbers new int parse logicalpageno tostring pageno tostring state batchfileprocessingentity operationstate failed await database stringsetasync abort fileid logicalpageno pageno timespan fromdays await arg completemessageasync arg message arg cancellationtoken catch exception ex logger logerror ex error occurred processing message failure finally messages remaining in the dlq should be investigated public async task startasync cancellationtoken cancellationtoken receiver processmessageasync receiveronprocessmessageasync receiver processerrorasync receiveronprocesserrorasync deadletterreceiver processmessageasync deadletterreceiveronprocessmessageasync deadletterreceiver processerrorasync receiveronprocesserrorasync await task whenall deadletterreceiver startprocessingasync cancellationtoken receiver startprocessingasync cancellationtoken logger loginformation receiver listening to queue receiver entitypath public async task stopasync cancellationtoken cancellationtoken await receiver stopprocessingasync cancellationtoken await receiver closeasync cancellationtoken private task receiveronprocesserrorasync processerroreventargs arg logger logerror arg exception sessionmessageprocessing failed return task completedtask private async task receiveronprocessmessageasync processmessageeventargs arg try var message arg message var cancellationtoken arg cancellationtoken if arg message applicationproperties trygetvalue fileid out var fileid var checkresult false await policy handle waitandretryasync count timespan frommilliseconds count executeasync async if await database keyexistsasync abort fileid await arg completemessageasync message cancellationtoken checkresult true return if arg message applicationproperties trygetvalue logicalpageno out var logicalpageno if await database keyexistsasync abort fileid logicalpageno await arg completemessageasync message cancellationtoken checkresult true return if message applicationproperties containskey abortfileid await database stringsetasync abort message applicationproperties timespan fromdays await arg completemessageasync message cancellationtoken checkresult true return if checkresult return arg message applicationproperties trygetvalue organizationid out var organisationid arg message applicationproperties trygetvalue projectid out var projectid var payloadtypename userid correlationid message getcommonmessageproperties var otherproperties message getothermessageproperties var logvalues new dictionary correlationid correlationid foreach var pair in otherproperties logvalues pair value using var nestedcontainer container getnestedcontainer nestedcontainer inject typeof icorrelationid new correlationid correlationid true var logger nestedcontainer getinstance using logger beginscope logvalues try this validation is done here so all logging is scoped properly if payloadtypename null userid null logger logdebug a message was received with no specified payload type or no security token await arg deadlettermessageasync arg message message is not in a format messagereceiver can process cancellationtoken cancellationtoken return var claimlist new list new claim claimtypes nameidentifier userid if organisationid null claimlist add new claim organizationid organisationid tostring if projectid null claimlist add new claim projectid projectid tostring var user new claimsprincipal new claimsidentity claimlist using logger beginscope new dictionary userid userid var messagebytes message body toarray var request requesttypeaggregator deserialize messagebytes payloadtypename message contenttype if request null logger logdebug failed to deserialize request payload type payloadtypename payloadtypename await arg deadlettermessageasync message request payload was invalid cancellationtoken cancellationtoken return nestedcontainer inject typeof iprincipal user true nestedcontainer inject user true var mediator nestedcontainer getinstance try await mediator send request cancellationtoken await arg completemessageasync message cancellationtoken catch any upstream errors in the mediator are logged by the logging behaviour and are not our concern here await arg deadlettermessageasync message request execution failed with error cancellationtoken cancellationtoken catch exception ex logger logerror ex exception with message handling await arg deadlettermessageasync message unexpected error processing request cancellationtoken cancellationtoken catch exception ex this logger has no scope logger logerror ex exception with message handling await arg deadlettermessageasync arg message unexpected error processing request cancellationtoken arg cancellationtoken environment azure messaging servicebus running in a azure pod from linux aspnet core docker image | 0 |
79,566 | 28,375,634,476 | IssuesEvent | 2023-04-12 20:37:00 | JohnAustinDev/xulsword | https://api.github.com/repos/JohnAustinDev/xulsword | closed | Menu:View: Combine tools for notes and cross-refs? | Type-Defect Priority-Medium auto-migrated | ```
In Menu:View, footnotes, cross references and personal notes are the only ones,
which have two tools
1. one to switch them on/off and
2. one to choose, where to display them (pop-ups or note-box)
What is the expected output? What do you see instead?
-> I find it easier, if those 'double'-tools for those 3 items would be in
close neighbourship instead on the oposite side of the menu list.
What version of the product are you using? On what operating system?
xs-3.3 (win/lin)
Please provide any additional information below.
Looking at this, three ideas come up:
1. Put the 6 tab tools at the beginning, so that the 2 groups of tools for
'additional text' come closer together,
or
2. Integrate the 3 'show tools' into the first list of tools for additional
text in the way:
- Footnotes
- Show Footnotes
- Cross References
- Show Cross Refernences
...
This 'combined list could also follow the tab list.
or
3. Combine the on/off tool and show tool for the three items in one, e.g. for
Footnotes:
- Footnotes: on/off,
and in case 'on' is chosen: Show in Popup/Note-box.
```
Original issue reported on code.google.com by `wolfgang...@wycliff.de` on 19 Apr 2012 at 9:37
| 1.0 | Menu:View: Combine tools for notes and cross-refs? - ```
In Menu:View, footnotes, cross references and personal notes are the only ones,
which have two tools
1. one to switch them on/off and
2. one to choose, where to display them (pop-ups or note-box)
What is the expected output? What do you see instead?
-> I find it easier, if those 'double'-tools for those 3 items would be in
close neighbourship instead on the oposite side of the menu list.
What version of the product are you using? On what operating system?
xs-3.3 (win/lin)
Please provide any additional information below.
Looking at this, three ideas come up:
1. Put the 6 tab tools at the beginning, so that the 2 groups of tools for
'additional text' come closer together,
or
2. Integrate the 3 'show tools' into the first list of tools for additional
text in the way:
- Footnotes
- Show Footnotes
- Cross References
- Show Cross Refernences
...
This 'combined list could also follow the tab list.
or
3. Combine the on/off tool and show tool for the three items in one, e.g. for
Footnotes:
- Footnotes: on/off,
and in case 'on' is chosen: Show in Popup/Note-box.
```
Original issue reported on code.google.com by `wolfgang...@wycliff.de` on 19 Apr 2012 at 9:37
| defect | menu view combine tools for notes and cross refs in menu view footnotes cross references and personal notes are the only ones which have two tools one to switch them on off and one to choose where to display them pop ups or note box what is the expected output what do you see instead i find it easier if those double tools for those items would be in close neighbourship instead on the oposite side of the menu list what version of the product are you using on what operating system xs win lin please provide any additional information below looking at this three ideas come up put the tab tools at the beginning so that the groups of tools for additional text come closer together or integrate the show tools into the first list of tools for additional text in the way footnotes show footnotes cross references show cross refernences this combined list could also follow the tab list or combine the on off tool and show tool for the three items in one e g for footnotes footnotes on off and in case on is chosen show in popup note box original issue reported on code google com by wolfgang wycliff de on apr at | 1 |
190,919 | 6,823,641,227 | IssuesEvent | 2017-11-08 01:05:48 | cilium/cilium | https://api.github.com/repos/cilium/cilium | opened | Access log: Support log rotation, backups and compression for access log | area/proxy priority/high project/1.0-gap | We need to support log rotation for production readiness.
A couple of things to keep in mind:
1. How often do we rotate ?
2. At what file size do we rotate? (100 MB)
3. How many backups do we keep?
4. We should compress the backup is possible
5. How often do we cleanup the backup files?
We should also support access logging by default on vagrant setups. | 1.0 | Access log: Support log rotation, backups and compression for access log - We need to support log rotation for production readiness.
A couple of things to keep in mind:
1. How often do we rotate ?
2. At what file size do we rotate? (100 MB)
3. How many backups do we keep?
4. We should compress the backup is possible
5. How often do we cleanup the backup files?
We should also support access logging by default on vagrant setups. | non_defect | access log support log rotation backups and compression for access log we need to support log rotation for production readiness a couple of things to keep in mind how often do we rotate at what file size do we rotate mb how many backups do we keep we should compress the backup is possible how often do we cleanup the backup files we should also support access logging by default on vagrant setups | 0 |
1,198 | 2,601,757,557 | IssuesEvent | 2015-02-24 00:33:42 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | Windowed mode does not support Battle.net | auto-migrated Priority-Low Type-Defect Usability | ```
As the issue states.
Windowed mode currently does not support the drawing of the Battle.net dialogs
or the display of the cinematics (Bink/Smacker Video).
```
-----
Original issue reported on code.google.com by `AHeinerm` on 14 Nov 2010 at 11:10 | 1.0 | Windowed mode does not support Battle.net - ```
As the issue states.
Windowed mode currently does not support the drawing of the Battle.net dialogs
or the display of the cinematics (Bink/Smacker Video).
```
-----
Original issue reported on code.google.com by `AHeinerm` on 14 Nov 2010 at 11:10 | defect | windowed mode does not support battle net as the issue states windowed mode currently does not support the drawing of the battle net dialogs or the display of the cinematics bink smacker video original issue reported on code google com by aheinerm on nov at | 1 |
17,254 | 2,986,952,683 | IssuesEvent | 2015-07-20 09:08:22 | wbsoft/frescobaldi | https://api.github.com/repos/wbsoft/frescobaldi | closed | Edit in place menu item has vanished | defect | For some reason the "Edit in place" item in the context menu of the manuscript viewer isn't there anymore.
In `viewers.contextmenu.ViewerContextMenu.addCursorLinksActions()` the `cursor` argument seems always `None`. So the issue must be in the stage when calling the context menu is prepared. | 1.0 | Edit in place menu item has vanished - For some reason the "Edit in place" item in the context menu of the manuscript viewer isn't there anymore.
In `viewers.contextmenu.ViewerContextMenu.addCursorLinksActions()` the `cursor` argument seems always `None`. So the issue must be in the stage when calling the context menu is prepared. | defect | edit in place menu item has vanished for some reason the edit in place item in the context menu of the manuscript viewer isn t there anymore in viewers contextmenu viewercontextmenu addcursorlinksactions the cursor argument seems always none so the issue must be in the stage when calling the context menu is prepared | 1 |
74,465 | 25,136,935,813 | IssuesEvent | 2022-11-09 19:25:17 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | BUG: "SystemError: initialization of beta_ufunc raised unreported exception" in >= 1.9.2 with Valgrind | defect | ### Describe your issue.
Hi all,
Our Valgrind CICD run broke after upgrading to Scipy >= 1.9.2 from 1.9.1.
The error does not occur without valgrind.
Due to the valgrind involvement making an MRE is though. If needed I can create a MRE repo and add a maintainer.
Some additional info:
* OS: Ubuntu-latest
* valgrind amd64 1:3.15.0-1ubuntu9.1
Bindings and generator:
* Pybind11 v2.10.1
* ninja v1.10.2.4
Extension does include a number of Numpy headers but nothing special.
Valgrind is run with:
```PYTHONMALLOC=malloc valgrind --leak-check=yes --track-origins=yes python -m pytest```
See [Workflow file](https://github.com/RUrlus/ModelMetricUncertainty/blob/stable/.github/workflows/valgrind.yml) for details.
### Reproducing Code Example
```python
import scipy.stats
```
### Error message
```shell
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/sklearn/utils/fixes.py:19: in <module>
import scipy.stats
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/__init__.py:467: in <module>
from ._stats_py import *
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_stats_py.py:46: in <module>
from . import distributions
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/distributions.py:10: in <module>
from . import _continuous_distns
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_continuous_distns.py:31: in <module>
import scipy.stats._boost as _boost
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_boost/__init__.py:1: in <module>
from scipy.stats._boost.beta_ufunc import (
E SystemError: initialization of beta_ufunc raised unreported exception
```
### SciPy/NumPy/Python version information
1.9.3 1.23.3 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0) | 1.0 | BUG: "SystemError: initialization of beta_ufunc raised unreported exception" in >= 1.9.2 with Valgrind - ### Describe your issue.
Hi all,
Our Valgrind CICD run broke after upgrading to Scipy >= 1.9.2 from 1.9.1.
The error does not occur without valgrind.
Due to the valgrind involvement making an MRE is though. If needed I can create a MRE repo and add a maintainer.
Some additional info:
* OS: Ubuntu-latest
* valgrind amd64 1:3.15.0-1ubuntu9.1
Bindings and generator:
* Pybind11 v2.10.1
* ninja v1.10.2.4
Extension does include a number of Numpy headers but nothing special.
Valgrind is run with:
```PYTHONMALLOC=malloc valgrind --leak-check=yes --track-origins=yes python -m pytest```
See [Workflow file](https://github.com/RUrlus/ModelMetricUncertainty/blob/stable/.github/workflows/valgrind.yml) for details.
### Reproducing Code Example
```python
import scipy.stats
```
### Error message
```shell
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/sklearn/utils/fixes.py:19: in <module>
import scipy.stats
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/__init__.py:467: in <module>
from ._stats_py import *
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_stats_py.py:46: in <module>
from . import distributions
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/distributions.py:10: in <module>
from . import _continuous_distns
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_continuous_distns.py:31: in <module>
import scipy.stats._boost as _boost
/opt/hostedtoolcache/Python/3.10.7/x64/lib/python3.10/site-packages/scipy/stats/_boost/__init__.py:1: in <module>
from scipy.stats._boost.beta_ufunc import (
E SystemError: initialization of beta_ufunc raised unreported exception
```
### SciPy/NumPy/Python version information
1.9.3 1.23.3 sys.version_info(major=3, minor=10, micro=7, releaselevel='final', serial=0) | defect | bug systemerror initialization of beta ufunc raised unreported exception in with valgrind describe your issue hi all our valgrind cicd run broke after upgrading to scipy from the error does not occur without valgrind due to the valgrind involvement making an mre is though if needed i can create a mre repo and add a maintainer some additional info os ubuntu latest valgrind bindings and generator ninja extension does include a number of numpy headers but nothing special valgrind is run with pythonmalloc malloc valgrind leak check yes track origins yes python m pytest see for details reproducing code example python import scipy stats error message shell opt hostedtoolcache python lib site packages sklearn utils fixes py in import scipy stats opt hostedtoolcache python lib site packages scipy stats init py in from stats py import opt hostedtoolcache python lib site packages scipy stats stats py py in from import distributions opt hostedtoolcache python lib site packages scipy stats distributions py in from import continuous distns opt hostedtoolcache python lib site packages scipy stats continuous distns py in import scipy stats boost as boost opt hostedtoolcache python lib site packages scipy stats boost init py in from scipy stats boost beta ufunc import e systemerror initialization of beta ufunc raised unreported exception scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
71,336 | 23,547,817,651 | IssuesEvent | 2022-08-21 11:38:39 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | closed | Clicking position on parchment map causes delta time camera jump | Priority-Medium Type-Defect | When you open the parchment map, you can zoom to a new location by clicking the spot you want to go to. Without delta time, you go there directly, but with the interpolation from delta_time, you see a camera glitch that moves you there. | 1.0 | Clicking position on parchment map causes delta time camera jump - When you open the parchment map, you can zoom to a new location by clicking the spot you want to go to. Without delta time, you go there directly, but with the interpolation from delta_time, you see a camera glitch that moves you there. | defect | clicking position on parchment map causes delta time camera jump when you open the parchment map you can zoom to a new location by clicking the spot you want to go to without delta time you go there directly but with the interpolation from delta time you see a camera glitch that moves you there | 1 |
80,039 | 29,927,087,958 | IssuesEvent | 2023-06-22 06:42:57 | matrix-org/matrix-public-archive | https://api.github.com/repos/matrix-org/matrix-public-archive | opened | Join event spam because of unstable URL in `reason` | T-Defect | Join event spam because of unstable URL in `reason`. Depending on how someone views the archive, the `reason` text will include a different URL.
Example:
Flip-flopping between `https://archive.matrix.org/roomid/wOlkWNmgkAZFxbTaqj:matrix.org` with the room ID and `https://archive.matrix.org/r/raspberrypi:matrix.org` and room alias depending on how the URL is visited.
```
{
"content": {
"displayname": "archive.matrix.org/faq",
"membership": "join",
"reason": "Joining room to check history visibility. If your room is public with shared or world readable history visibility, it will be accessible at https://archive.matrix.org/roomid/wOlkWNmgkAZFxbTaqj:matrix.org. See the FAQ for more details: https://github.com/matrix-org/matrix-public-archive/blob/main/docs/faq.md#why-did-the-archive-bot-join-my-room"
},
"origin_server_ts": 1687415330265,
"sender": "@archive:matrix.org",
"state_key": "@archive:matrix.org",
"type": "m.room.member",
"unsigned": {
"replaces_state": "$wlnRuzw6mTVYvUMkBmHD5jXGROSZglqnzG7ZrRQtmOA",
"prev_content": {
"displayname": "archive.matrix.org/faq",
"membership": "join",
"reason": "Joining room to check history visibility. If your room is public with shared or world readable history visibility, it will be accessible at https://archive.matrix.org/r/raspberrypi:matrix.org. See the FAQ for more details: https://github.com/matrix-org/matrix-public-archive/blob/main/docs/faq.md#why-did-the-archive-bot-join-my-room"
},
"prev_sender": "@archive:matrix.org",
"age": 2558
},
"event_id": "$cyfYrA3MW6w2c44z1_VWNfeRgUko2HkBlTM11EZqB1E",
"room_id": "!wOlkWNmgkAZFxbTaqj:matrix.org"
}
```
### Potential solutions
Always use room ID so it's consistent regardless of the alias someone visits from.
### Dev notes
The join `reason` logic was originally added in https://github.com/matrix-org/matrix-public-archive/pull/262
| 1.0 | Join event spam because of unstable URL in `reason` - Join event spam because of unstable URL in `reason`. Depending on how someone views the archive, the `reason` text will include a different URL.
Example:
Flip-flopping between `https://archive.matrix.org/roomid/wOlkWNmgkAZFxbTaqj:matrix.org` with the room ID and `https://archive.matrix.org/r/raspberrypi:matrix.org` and room alias depending on how the URL is visited.
```
{
"content": {
"displayname": "archive.matrix.org/faq",
"membership": "join",
"reason": "Joining room to check history visibility. If your room is public with shared or world readable history visibility, it will be accessible at https://archive.matrix.org/roomid/wOlkWNmgkAZFxbTaqj:matrix.org. See the FAQ for more details: https://github.com/matrix-org/matrix-public-archive/blob/main/docs/faq.md#why-did-the-archive-bot-join-my-room"
},
"origin_server_ts": 1687415330265,
"sender": "@archive:matrix.org",
"state_key": "@archive:matrix.org",
"type": "m.room.member",
"unsigned": {
"replaces_state": "$wlnRuzw6mTVYvUMkBmHD5jXGROSZglqnzG7ZrRQtmOA",
"prev_content": {
"displayname": "archive.matrix.org/faq",
"membership": "join",
"reason": "Joining room to check history visibility. If your room is public with shared or world readable history visibility, it will be accessible at https://archive.matrix.org/r/raspberrypi:matrix.org. See the FAQ for more details: https://github.com/matrix-org/matrix-public-archive/blob/main/docs/faq.md#why-did-the-archive-bot-join-my-room"
},
"prev_sender": "@archive:matrix.org",
"age": 2558
},
"event_id": "$cyfYrA3MW6w2c44z1_VWNfeRgUko2HkBlTM11EZqB1E",
"room_id": "!wOlkWNmgkAZFxbTaqj:matrix.org"
}
```
### Potential solutions
Always use room ID so it's consistent regardless of the alias someone visits from.
### Dev notes
The join `reason` logic was originally added in https://github.com/matrix-org/matrix-public-archive/pull/262
| defect | join event spam because of unstable url in reason join event spam because of unstable url in reason depending on how someone views the archive the reason text will include a different url example flip flopping between with the room id and and room alias depending on how the url is visited content displayname archive matrix org faq membership join reason joining room to check history visibility if your room is public with shared or world readable history visibility it will be accessible at see the faq for more details origin server ts sender archive matrix org state key archive matrix org type m room member unsigned replaces state prev content displayname archive matrix org faq membership join reason joining room to check history visibility if your room is public with shared or world readable history visibility it will be accessible at see the faq for more details prev sender archive matrix org age event id room id wolkwnmgkazfxbtaqj matrix org potential solutions always use room id so it s consistent regardless of the alias someone visits from dev notes the join reason logic was originally added in | 1 |
78,852 | 27,787,426,564 | IssuesEvent | 2023-03-17 05:23:00 | zed-industries/community | https://api.github.com/repos/zed-industries/community | closed | Garbled hover in .pre-commit-config.yaml | defect triage | ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
Garbled hover in [.pre-commit-config.yaml](https://github.com/seek-oss/aec/blob/9f94c4a/.pre-commit-config.yaml)
### Environment
Zed: v0.77.3 (stable)
OS: macOS 13.2.1
Memory: 32 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
<img width="849" alt="CleanShot 2023-03-17 at 14 07 05@2x" src="https://user-images.githubusercontent.com/125105/225802356-71881cae-6789-475a-8094-dfb8caad941c.png">
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_ | 1.0 | Garbled hover in .pre-commit-config.yaml - ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
Garbled hover in [.pre-commit-config.yaml](https://github.com/seek-oss/aec/blob/9f94c4a/.pre-commit-config.yaml)
### Environment
Zed: v0.77.3 (stable)
OS: macOS 13.2.1
Memory: 32 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
<img width="849" alt="CleanShot 2023-03-17 at 14 07 05@2x" src="https://user-images.githubusercontent.com/125105/225802356-71881cae-6789-475a-8094-dfb8caad941c.png">
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_ | defect | garbled hover in pre commit config yaml check for existing issues completed describe the bug provide steps to reproduce it garbled hover in environment zed stable os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature img width alt cleanshot at src if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response | 1 |
72,226 | 24,003,755,613 | IssuesEvent | 2022-09-14 13:26:45 | cf-convention/cf-conventions | https://api.github.com/repos/cf-convention/cf-conventions | closed | Automated build of documents not working with single-sourced version | defect GitHub Usage | Whilst building the documents for the public release of CF-1.10, I encountered some failures with the automated GitHub actions. These action were triggered by the creation of a new "Release". The details can be found at https://github.com/cf-convention/cf-conventions/actions/runs/2962652688
As a workaround, I ran the asciidoc commands (taken verbatim from in https://github.com/cf-convention/cf-conventions/blob/main/.github/workflows/adoc_build.yml) on the command line and successfully built the documents
The build of the PDF of the conformance document wouldn't work on the command line either, but this artefact is no currently posted to the web site, so that was not a problem for CF-1.10. However, we _should_ provide a PDF of the conformance document for future versions.
Many thanks for @zklaus for putting all this worthy improvement together in the first place! (https://github.com/cf-convention/cf-conventions/issues/343) | 1.0 | Automated build of documents not working with single-sourced version - Whilst building the documents for the public release of CF-1.10, I encountered some failures with the automated GitHub actions. These action were triggered by the creation of a new "Release". The details can be found at https://github.com/cf-convention/cf-conventions/actions/runs/2962652688
As a workaround, I ran the asciidoc commands (taken verbatim from in https://github.com/cf-convention/cf-conventions/blob/main/.github/workflows/adoc_build.yml) on the command line and successfully built the documents
The build of the PDF of the conformance document wouldn't work on the command line either, but this artefact is no currently posted to the web site, so that was not a problem for CF-1.10. However, we _should_ provide a PDF of the conformance document for future versions.
Many thanks for @zklaus for putting all this worthy improvement together in the first place! (https://github.com/cf-convention/cf-conventions/issues/343) | defect | automated build of documents not working with single sourced version whilst building the documents for the public release of cf i encountered some failures with the automated github actions these action were triggered by the creation of a new release the details can be found at as a workaround i ran the asciidoc commands taken verbatim from in on the command line and successfully built the documents the build of the pdf of the conformance document wouldn t work on the command line either but this artefact is no currently posted to the web site so that was not a problem for cf however we should provide a pdf of the conformance document for future versions many thanks for zklaus for putting all this worthy improvement together in the first place | 1 |
9,571 | 2,615,162,760 | IssuesEvent | 2015-03-01 06:42:10 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | receive timeout (0x02) with WG511T ( Atheros ath5k ) | auto-migrated Priority-Triage Type-Defect | ```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
reaver 1.4 and
1. What operating system are you using
BT4 R2 and WG511T 108 Mbps Wireless PC Card
wlan1 Atheros ath5k - [phy1]
(monitor mode enabled on mon0)
2. Is your wireless card in monitor mode (yes/no)? Yes. Injection is working,
Yes
3. What is the signal strength of the Access Point you are trying to crack?
Range tried varies between -67.
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b 00:01:02:03:04:05 -vv --dh-small
reaver -i mon0 -b 00:01:02:03:04:05 -vv -d 0
1. run: aireplay-ng mon0 -1 120 -a 68:7F:74:E2:4A:1C -e kitty-Home
2. then: reaver -i mon0 -A -b 68:7F:74:E2:4A:1C -c 6 -vv --no-nacks --win7
reaver -A -b 00:1C:DF:XX:XX:XX -c 6 -vv -i mon0 --dh-small --no-nacks
--ignore-locks --win7 -d 15
WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] 0.08% complete @ 2012-01-26 01:00:17 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[!] WARNING: 10 failed connections in a row
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
any help or points that would be very much apprecaited
```
Original issue reported on code.google.com by `zerr...@gmail.com` on 3 Jun 2012 at 10:04 | 1.0 | receive timeout (0x02) with WG511T ( Atheros ath5k ) - ```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
reaver 1.4 and
1. What operating system are you using
BT4 R2 and WG511T 108 Mbps Wireless PC Card
wlan1 Atheros ath5k - [phy1]
(monitor mode enabled on mon0)
2. Is your wireless card in monitor mode (yes/no)? Yes. Injection is working,
Yes
3. What is the signal strength of the Access Point you are trying to crack?
Range tried varies between -67.
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b 00:01:02:03:04:05 -vv --dh-small
reaver -i mon0 -b 00:01:02:03:04:05 -vv -d 0
1. run: aireplay-ng mon0 -1 120 -a 68:7F:74:E2:4A:1C -e kitty-Home
2. then: reaver -i mon0 -A -b 68:7F:74:E2:4A:1C -c 6 -vv --no-nacks --win7
reaver -A -b 00:1C:DF:XX:XX:XX -c 6 -vv -i mon0 --dh-small --no-nacks
--ignore-locks --win7 -d 15
WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] 0.08% complete @ 2012-01-26 01:00:17 (0 seconds/pin)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x02), re-trying last pin
[!] WARNING: 10 failed connections in a row
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
[!] WARNING: Failed to associate with XX:XX:XX:XX:XX:XX (ESSID: xxxx
any help or points that would be very much apprecaited
```
Original issue reported on code.google.com by `zerr...@gmail.com` on 3 Jun 2012 at 10:04 | defect | receive timeout with atheros a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered reaver and what operating system are you using and mbps wireless pc card atheros monitor mode enabled on is your wireless card in monitor mode yes no yes injection is working yes what is the signal strength of the access point you are trying to crack range tried varies between what is the entire command line string you are supplying to reaver reaver i b vv dh small reaver i b vv d run aireplay ng a e kitty home then reaver i a b c vv no nacks reaver a b df xx xx xx c vv i dh small no nacks ignore locks d wps transaction failed code re trying last pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received identity request sending identity response received identity request sending identity response warning receive timeout occurred sending wsc nack wps transaction failed code re trying last pin complete seconds pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending wsc nack sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received identity request sending identity response received identity request sending identity response warning receive timeout occurred sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response warning receive timeout occurred sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending wsc nack sending wsc nack wps transaction failed code re trying last pin warning failed connections in a row warning failed to associate with xx xx xx xx xx xx essid xxxx warning failed to associate with xx xx xx xx xx xx essid xxxx warning failed to associate with xx xx xx xx xx xx essid xxxx warning failed to associate with xx xx xx xx xx xx essid xxxx any help or points that would be very much apprecaited original issue reported on code google com by zerr gmail com on jun at | 1 |
223,157 | 17,105,408,945 | IssuesEvent | 2021-07-09 16:55:07 | typeorm/typeorm | https://api.github.com/repos/typeorm/typeorm | closed | Type '"abstract"' is not assignable to type 'TableType'. | bug documentation driver: postgres | The documentation states that the type 'abstract' can be used but when it is used the below error is thrown in TypeScript.
/**
* Table type. Tables can be abstract, closure, junction, embedded, etc.
*/
export type TableType = "regular"|"view"|"junction"|"closure"|"closure-junction"|"entity-child";
https://github.com/typeorm/typeorm/blob/63d993b1ef9f68a6e12bd535ac8ca8be39702502/src/metadata/types/TableTypes.ts
[ ] question
[x] bug report
[ ] feature request
[x] documentation issue
**Database system/driver:**
[ ] `cordova`
[ ] `mongodb`
[ ] `mssql`
[ ] `mysql` / `mariadb`
[ ] `oracle`
[x] `postgres`
[ ] `cockroachdb`
[ ] `sqlite`
[ ] `sqljs`
[ ] `react-native`
[ ] `expo`
**TypeORM version:**
[x] `latest`
[ ] `@next`
[ ] `0.x.x` (or put your version here)
**Steps to reproduce or a small repository showing the problem:**
```
import { EntitySchema } from 'typeorm';
import BaseEntity from '../models/base-entity.model';
export default new EntitySchema({
name: `BaseEntity`,
target: BaseEntity,
type: `abstract`,
columns: {
created_date: {
type: `timestamptz`,
nullable: false,
name: `created_date`,
},
modified_by: {
type: `character varying`,
nullable: true,
length: 254,
name: `modified_by`,
},
created_by: {
type: `character varying`,
nullable: false,
name: `created_by`,
},
modified_date: {
type: `timestamptz`,
nullable: true,
name: `modified_date`,
},
},
});
```
| 1.0 | Type '"abstract"' is not assignable to type 'TableType'. - The documentation states that the type 'abstract' can be used but when it is used the below error is thrown in TypeScript.
/**
* Table type. Tables can be abstract, closure, junction, embedded, etc.
*/
export type TableType = "regular"|"view"|"junction"|"closure"|"closure-junction"|"entity-child";
https://github.com/typeorm/typeorm/blob/63d993b1ef9f68a6e12bd535ac8ca8be39702502/src/metadata/types/TableTypes.ts
[ ] question
[x] bug report
[ ] feature request
[x] documentation issue
**Database system/driver:**
[ ] `cordova`
[ ] `mongodb`
[ ] `mssql`
[ ] `mysql` / `mariadb`
[ ] `oracle`
[x] `postgres`
[ ] `cockroachdb`
[ ] `sqlite`
[ ] `sqljs`
[ ] `react-native`
[ ] `expo`
**TypeORM version:**
[x] `latest`
[ ] `@next`
[ ] `0.x.x` (or put your version here)
**Steps to reproduce or a small repository showing the problem:**
```
import { EntitySchema } from 'typeorm';
import BaseEntity from '../models/base-entity.model';
export default new EntitySchema({
name: `BaseEntity`,
target: BaseEntity,
type: `abstract`,
columns: {
created_date: {
type: `timestamptz`,
nullable: false,
name: `created_date`,
},
modified_by: {
type: `character varying`,
nullable: true,
length: 254,
name: `modified_by`,
},
created_by: {
type: `character varying`,
nullable: false,
name: `created_by`,
},
modified_date: {
type: `timestamptz`,
nullable: true,
name: `modified_date`,
},
},
});
```
| non_defect | type abstract is not assignable to type tabletype the documentation states that the type abstract can be used but when it is used the below error is thrown in typescript table type tables can be abstract closure junction embedded etc export type tabletype regular view junction closure closure junction entity child question bug report feature request documentation issue database system driver cordova mongodb mssql mysql mariadb oracle postgres cockroachdb sqlite sqljs react native expo typeorm version latest next x x or put your version here steps to reproduce or a small repository showing the problem import entityschema from typeorm import baseentity from models base entity model export default new entityschema name baseentity target baseentity type abstract columns created date type timestamptz nullable false name created date modified by type character varying nullable true length name modified by created by type character varying nullable false name created by modified date type timestamptz nullable true name modified date | 0 |
290,409 | 21,877,139,216 | IssuesEvent | 2022-05-19 11:14:19 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | closed | [Docs] #12801 fix: Fix app crash on arrow down in where clause | Documentation User Education Pod | > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | BE Coders Pod
**Parent Ticket** | #12801
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment | | 1.0 | [Docs] #12801 fix: Fix app crash on arrow down in where clause - > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | BE Coders Pod
**Parent Ticket** | #12801
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment | | non_defect | fix fix app crash on arrow down in where clause todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod be coders pod parent ticket engineer release date live date first draft auto assign priority environment | 0 |
178,402 | 21,509,387,712 | IssuesEvent | 2022-04-28 01:35:38 | bsbtd/Teste | https://api.github.com/repos/bsbtd/Teste | closed | WS-2020-0163 (Medium) detected in marked-0.8.2.tgz, marked-0.7.0.tgz - autoclosed | security vulnerability | ## WS-2020-0163 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>marked-0.8.2.tgz</b>, <b>marked-0.7.0.tgz</b></p></summary>
<p>
<details><summary><b>marked-0.8.2.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.8.2.tgz">https://registry.npmjs.org/marked/-/marked-0.8.2.tgz</a></p>
<p>Path to dependency file: /api-umbrella/src/api-umbrella/admin-ui/package.json</p>
<p>Path to vulnerable library: /api-umbrella/src/api-umbrella/admin-ui/node_modules/marked/package.json</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.8.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>marked-0.7.0.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p>
<p>
Dependency Hierarchy:
- jsdoc-3.6.3.tgz (Root Library)
- :x: **marked-0.7.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 1.1.1 is vulnerable to Regular Expression Denial of Service (REDoS). rules.js have multiple unused capture groups which can lead to a Denial of Service.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://github.com/markedjs/marked/commit/bd4f8c464befad2b304d51e33e89e567326e62e0>WS-2020-0163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/releases/tag/v1.1.1">https://github.com/markedjs/marked/releases/tag/v1.1.1</a></p>
<p>Release Date: 2020-07-02</p>
<p>Fix Resolution (marked): 1.1.1</p>
<p>Direct dependency fix Resolution (jsdoc): 3.6.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0163 (Medium) detected in marked-0.8.2.tgz, marked-0.7.0.tgz - autoclosed - ## WS-2020-0163 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>marked-0.8.2.tgz</b>, <b>marked-0.7.0.tgz</b></p></summary>
<p>
<details><summary><b>marked-0.8.2.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.8.2.tgz">https://registry.npmjs.org/marked/-/marked-0.8.2.tgz</a></p>
<p>Path to dependency file: /api-umbrella/src/api-umbrella/admin-ui/package.json</p>
<p>Path to vulnerable library: /api-umbrella/src/api-umbrella/admin-ui/node_modules/marked/package.json</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.8.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>marked-0.7.0.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p>
<p>
Dependency Hierarchy:
- jsdoc-3.6.3.tgz (Root Library)
- :x: **marked-0.7.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 1.1.1 is vulnerable to Regular Expression Denial of Service (REDoS). rules.js have multiple unused capture groups which can lead to a Denial of Service.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://github.com/markedjs/marked/commit/bd4f8c464befad2b304d51e33e89e567326e62e0>WS-2020-0163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/releases/tag/v1.1.1">https://github.com/markedjs/marked/releases/tag/v1.1.1</a></p>
<p>Release Date: 2020-07-02</p>
<p>Fix Resolution (marked): 1.1.1</p>
<p>Direct dependency fix Resolution (jsdoc): 3.6.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in marked tgz marked tgz autoclosed ws medium severity vulnerability vulnerable libraries marked tgz marked tgz marked tgz a markdown parser built for speed library home page a href path to dependency file api umbrella src api umbrella admin ui package json path to vulnerable library api umbrella src api umbrella admin ui node modules marked package json dependency hierarchy x marked tgz vulnerable library marked tgz a markdown parser built for speed library home page a href dependency hierarchy jsdoc tgz root library x marked tgz vulnerable library found in head commit a href vulnerability details marked before is vulnerable to regular expression denial of service redos rules js have multiple unused capture groups which can lead to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution marked direct dependency fix resolution jsdoc step up your open source security game with whitesource | 0 |
808 | 2,546,153,674 | IssuesEvent | 2015-01-29 21:51:45 | nickvergessen/phpbb-ext-newspage | https://api.github.com/repos/nickvergessen/phpbb-ext-newspage | closed | nickvergessen/phpbb-tool-trimmessage dependency missing from download | Documentation | The option to change the number of characters in a news message has no effect. Could it be that this function is not working? I tried different values but it has no effect...
A user on phpbb.nl noticed it, when I tested it I noticed the same.
Source: http://www.phpbb.nl/forums/viewtopic.php?f=93&t=59796 | 1.0 | nickvergessen/phpbb-tool-trimmessage dependency missing from download - The option to change the number of characters in a news message has no effect. Could it be that this function is not working? I tried different values but it has no effect...
A user on phpbb.nl noticed it, when I tested it I noticed the same.
Source: http://www.phpbb.nl/forums/viewtopic.php?f=93&t=59796 | non_defect | nickvergessen phpbb tool trimmessage dependency missing from download the option to change the number of characters in a news message has no effect could it be that this function is not working i tried different values but it has no effect a user on phpbb nl noticed it when i tested it i noticed the same source | 0 |
199,580 | 15,773,371,712 | IssuesEvent | 2021-03-31 23:11:38 | scalameta/scalafmt | https://api.github.com/repos/scalameta/scalafmt | closed | How to ignore long imports? | documentation | ## Problem
Scalafmt formats code like this:
```diff
-import very.very.very.long.line.with.{lots, of, additional, items}
+import very.very.very.long.line.with.{
+ lots,
+ of,
+ additional,
+ items
+}
But Scalafix and Idea's import optimizers return them back to one-liner.
```
## Expectation
I would like to have a setting which would allow to ignore or don't break long imports into multiline. | 1.0 | How to ignore long imports? - ## Problem
Scalafmt formats code like this:
```diff
-import very.very.very.long.line.with.{lots, of, additional, items}
+import very.very.very.long.line.with.{
+ lots,
+ of,
+ additional,
+ items
+}
But Scalafix and Idea's import optimizers return them back to one-liner.
```
## Expectation
I would like to have a setting which would allow to ignore or don't break long imports into multiline. | non_defect | how to ignore long imports problem scalafmt formats code like this diff import very very very long line with lots of additional items import very very very long line with lots of additional items but scalafix and idea s import optimizers return them back to one liner expectation i would like to have a setting which would allow to ignore or don t break long imports into multiline | 0 |
60,806 | 17,023,527,111 | IssuesEvent | 2021-07-03 02:29:01 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Opened changeset shows wrong Closed at time | Component: website Priority: minor Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 7.51pm, Tuesday, 22nd December 2009]**
When a changeset is open the Created at time is correct. However the Closed at is not, it is an hour early.
At: Tue, 22 Dec 2009 19:50:24 +0000 changeset 3429794 shows:
Created at: Tue, 22 Dec 2009 19:45:34 +0000
Closed at: Tue, 22 Dec 2009 20:49:02 +0000
As you can see cloased at is an hour early.
Once the changeset is closed the time is correct. | 1.0 | Opened changeset shows wrong Closed at time - **[Submitted to the original trac issue database at 7.51pm, Tuesday, 22nd December 2009]**
When a changeset is open the Created at time is correct. However the Closed at is not, it is an hour early.
At: Tue, 22 Dec 2009 19:50:24 +0000 changeset 3429794 shows:
Created at: Tue, 22 Dec 2009 19:45:34 +0000
Closed at: Tue, 22 Dec 2009 20:49:02 +0000
As you can see cloased at is an hour early.
Once the changeset is closed the time is correct. | defect | opened changeset shows wrong closed at time when a changeset is open the created at time is correct however the closed at is not it is an hour early at tue dec changeset shows created at tue dec closed at tue dec as you can see cloased at is an hour early once the changeset is closed the time is correct | 1 |
289,462 | 8,871,117,357 | IssuesEvent | 2019-01-11 11:32:40 | InFact-coop/create-refresh | https://api.github.com/repos/InFact-coop/create-refresh | opened | Mobile menu dropdown background | T2h priority-4 | Not sure how to address this - media queries? But can take a look if time. | 1.0 | Mobile menu dropdown background - Not sure how to address this - media queries? But can take a look if time. | non_defect | mobile menu dropdown background not sure how to address this media queries but can take a look if time | 0 |
67,391 | 20,961,608,876 | IssuesEvent | 2022-03-27 21:48:46 | abedmaatalla/sipdroid | https://api.github.com/repos/abedmaatalla/sipdroid | closed | odec | Priority-Medium Type-Defect auto-migrated | ```
Before you begin please read the FAQ!
Attention!
Don't post doubles!
Why is it important to find doubles?
Sipdroid is a community project. This is no dropbox for your personal
issue. You need to find others having the same problem to get it fixed. If
the problem has already been reported you add more details and ideas to the
existing discussion. If it has not been reported yet you include as many
background info as possible to give the community a chance to follow up on
it. No one will just sit there and try to replicate your setup. The root
cause of the problem will need to be identified to get it fixed.
How can I find doubles?
When clicking on issues tab there is a search box. Enter keywords for your
issue to find previous reports matching yours. Choose find all to match
closed tickets as well.
How can I investigate an incompatiblity?
Just mentioning a SIP server or device that does not interoperate with
Sipdroid won't help much because the developers might not have that
particular setup on hands. So you will try to add a SIP or logcat trace
from your network/phone to make a fix at all possible.
Did you carefully read above?
If yes, please fill out the following form and provide all necessary
details.
If not, you may continue in the Forum where you can discuss your problem to
clear things up before filing an issue report.
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what device/operating system?
Which SIP server are you using? What happens with PBXes?
Which type of network are you using?
Please provide any additional information below.
```
Original issue reported on code.google.com by `stxnzda...@gmail.com` on 2 Apr 2013 at 5:29
| 1.0 | odec - ```
Before you begin please read the FAQ!
Attention!
Don't post doubles!
Why is it important to find doubles?
Sipdroid is a community project. This is no dropbox for your personal
issue. You need to find others having the same problem to get it fixed. If
the problem has already been reported you add more details and ideas to the
existing discussion. If it has not been reported yet you include as many
background info as possible to give the community a chance to follow up on
it. No one will just sit there and try to replicate your setup. The root
cause of the problem will need to be identified to get it fixed.
How can I find doubles?
When clicking on issues tab there is a search box. Enter keywords for your
issue to find previous reports matching yours. Choose find all to match
closed tickets as well.
How can I investigate an incompatiblity?
Just mentioning a SIP server or device that does not interoperate with
Sipdroid won't help much because the developers might not have that
particular setup on hands. So you will try to add a SIP or logcat trace
from your network/phone to make a fix at all possible.
Did you carefully read above?
If yes, please fill out the following form and provide all necessary
details.
If not, you may continue in the Forum where you can discuss your problem to
clear things up before filing an issue report.
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what device/operating system?
Which SIP server are you using? What happens with PBXes?
Which type of network are you using?
Please provide any additional information below.
```
Original issue reported on code.google.com by `stxnzda...@gmail.com` on 2 Apr 2013 at 5:29
| defect | odec before you begin please read the faq attention don t post doubles why is it important to find doubles sipdroid is a community project this is no dropbox for your personal issue you need to find others having the same problem to get it fixed if the problem has already been reported you add more details and ideas to the existing discussion if it has not been reported yet you include as many background info as possible to give the community a chance to follow up on it no one will just sit there and try to replicate your setup the root cause of the problem will need to be identified to get it fixed how can i find doubles when clicking on issues tab there is a search box enter keywords for your issue to find previous reports matching yours choose find all to match closed tickets as well how can i investigate an incompatiblity just mentioning a sip server or device that does not interoperate with sipdroid won t help much because the developers might not have that particular setup on hands so you will try to add a sip or logcat trace from your network phone to make a fix at all possible did you carefully read above if yes please fill out the following form and provide all necessary details if not you may continue in the forum where you can discuss your problem to clear things up before filing an issue report what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what device operating system which sip server are you using what happens with pbxes which type of network are you using please provide any additional information below original issue reported on code google com by stxnzda gmail com on apr at | 1 |
81,012 | 30,658,587,609 | IssuesEvent | 2023-07-25 13:42:08 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | Fix #editContentlet Macro on Detail Pages | Type : Defect Team : Lunik OKR : Core Features Priority : 3 Average |
### Problem Statement
The #editContentlet macro breaks the modal to edit in place or copy on detail pages like event and blog detail.
### Steps to Reproduce
1. Go to an event or blog detail page
2. Try to edit the content in place or copy using #editContentlet
### Acceptance Criteria
1. Research existing documentation to determine the cause of the issue
2. Identify possible solutions for the issue
3. Develop a fix for the issue
### External Links
N/A
### Assumptions & Initiation Needs
1. Developers should be familiar with #editContentlet and the modal to edit in place or copy.
2. Developers should have access to the relevant detail pages.
### Quality Assurance Notes & Workarounds
1. Test the fix against different browsers and devices.
2. As a workaround, users can edit the content from the Content tab of dotCMS. | 1.0 | Fix #editContentlet Macro on Detail Pages -
### Problem Statement
The #editContentlet macro breaks the modal to edit in place or copy on detail pages like event and blog detail.
### Steps to Reproduce
1. Go to an event or blog detail page
2. Try to edit the content in place or copy using #editContentlet
### Acceptance Criteria
1. Research existing documentation to determine the cause of the issue
2. Identify possible solutions for the issue
3. Develop a fix for the issue
### External Links
N/A
### Assumptions & Initiation Needs
1. Developers should be familiar with #editContentlet and the modal to edit in place or copy.
2. Developers should have access to the relevant detail pages.
### Quality Assurance Notes & Workarounds
1. Test the fix against different browsers and devices.
2. As a workaround, users can edit the content from the Content tab of dotCMS. | defect | fix editcontentlet macro on detail pages problem statement the editcontentlet macro breaks the modal to edit in place or copy on detail pages like event and blog detail steps to reproduce go to an event or blog detail page try to edit the content in place or copy using editcontentlet acceptance criteria research existing documentation to determine the cause of the issue identify possible solutions for the issue develop a fix for the issue external links n a assumptions initiation needs developers should be familiar with editcontentlet and the modal to edit in place or copy developers should have access to the relevant detail pages quality assurance notes workarounds test the fix against different browsers and devices as a workaround users can edit the content from the content tab of dotcms | 1 |
489,740 | 14,111,818,022 | IssuesEvent | 2020-11-07 01:50:45 | chingu-voyages/v25-geckos-team-01 | https://api.github.com/repos/chingu-voyages/v25-geckos-team-01 | opened | Display a contact & agreement checklist for Volunteers | UserStory priority:must_have | **User Story Description**
As a Volunteer
I want to see contact & agreement information when the Nonprofit accepts me for a task
So I can finalize arrangements with the Nonprofit to start work
**Steps to Follow (optional)**
- [ ] TBD
- [ ] Additional steps as necessary
**Additional Considerations**
Any supplemental information including unresolved questions, links to external resources, screenshots, etc.
| 1.0 | Display a contact & agreement checklist for Volunteers - **User Story Description**
As a Volunteer
I want to see contact & agreement information when the Nonprofit accepts me for a task
So I can finalize arrangements with the Nonprofit to start work
**Steps to Follow (optional)**
- [ ] TBD
- [ ] Additional steps as necessary
**Additional Considerations**
Any supplemental information including unresolved questions, links to external resources, screenshots, etc.
| non_defect | display a contact agreement checklist for volunteers user story description as a volunteer i want to see contact agreement information when the nonprofit accepts me for a task so i can finalize arrangements with the nonprofit to start work steps to follow optional tbd additional steps as necessary additional considerations any supplemental information including unresolved questions links to external resources screenshots etc | 0 |
39,875 | 9,708,797,523 | IssuesEvent | 2019-05-28 08:38:37 | PowerDNS/pdns | https://api.github.com/repos/PowerDNS/pdns | closed | dnsdist healthcheck with mustResolve doesn't check for answer | defect dnsdist docs | - Program: dnsdist
- Issue type: Bug report
### Short description
Creating a `newServer()` with `mustResolve=true` [according to the docs](https://dnsdist.org/reference/config.html#newServer) `MUST return a NOERROR RCODE and an answer`.
Currently it only checks for the `RCODE` to not be `NXDOMAIN` or `REFUSED`: https://github.com/PowerDNS/pdns/blob/18357a07c5b44366b0b3826b3ac66f4feea048be/pdns/dnsdist.cc#L1857
### Environment
- Operating system: Debian 9.9
- Software version: dnsdist 1.4.0-alpha2
- Software source: pdns repo
### Steps to reproduce
```lua
newServer({address="2a01:4f8:192:43a5::2", name="ns10.de.dns.opennic.glue", mustResolve=true})
```
This server does at the time of writing not send an ANSWER section (example at the bottom).
### Expected behaviour
Server should be detected as down.
### Actual behaviour
Server is detected mostly up.
Some extra output from `setVerboseHealthChecks(true)`
```
$ journalctl -u dnsdist.service -f | grep -F 2a01:4f8:192:43a5::2
May 24 18:46:13 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:46:13 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:46:17 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:46:28 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:46:28 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:46:33 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:07 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:07 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:11 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:30 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:30 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:34 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:39 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:39 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:43 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:47 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:47 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:50 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
```
### Other information
```
$ dig a a.root-servers.net. @2a01:4f8:192:43a5::2
; <<>> DiG 9.10.6 <<>> a a.root-servers.net. @2a01:4f8:192:43a5::2
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 19115
;; flags: qr rd; QUERY: 1, ANSWER: 0, AUTHORITY: 13, ADDITIONAL: 27
;; WARNING: recursion requested but not available
;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 4096
;; QUESTION SECTION:
;a.root-servers.net. IN A
;; AUTHORITY SECTION:
net. 172800 IN NS a.gtld-servers.net.
net. 172800 IN NS b.gtld-servers.net.
net. 172800 IN NS c.gtld-servers.net.
net. 172800 IN NS d.gtld-servers.net.
net. 172800 IN NS e.gtld-servers.net.
net. 172800 IN NS f.gtld-servers.net.
net. 172800 IN NS g.gtld-servers.net.
net. 172800 IN NS h.gtld-servers.net.
net. 172800 IN NS i.gtld-servers.net.
net. 172800 IN NS j.gtld-servers.net.
net. 172800 IN NS k.gtld-servers.net.
net. 172800 IN NS l.gtld-servers.net.
net. 172800 IN NS m.gtld-servers.net.
;; ADDITIONAL SECTION:
a.gtld-servers.net. 172800 IN AAAA 2001:503:a83e::2:30
b.gtld-servers.net. 172800 IN AAAA 2001:503:231d::2:30
c.gtld-servers.net. 172800 IN AAAA 2001:503:83eb::30
d.gtld-servers.net. 172800 IN AAAA 2001:500:856e::30
e.gtld-servers.net. 172800 IN AAAA 2001:502:1ca1::30
f.gtld-servers.net. 172800 IN AAAA 2001:503:d414::30
g.gtld-servers.net. 172800 IN AAAA 2001:503:eea3::30
h.gtld-servers.net. 172800 IN AAAA 2001:502:8cc::30
i.gtld-servers.net. 172800 IN AAAA 2001:503:39c1::30
j.gtld-servers.net. 172800 IN AAAA 2001:502:7094::30
k.gtld-servers.net. 172800 IN AAAA 2001:503:d2d::30
l.gtld-servers.net. 172800 IN AAAA 2001:500:d937::30
m.gtld-servers.net. 172800 IN AAAA 2001:501:b1f9::30
a.gtld-servers.net. 172800 IN A 192.5.6.30
b.gtld-servers.net. 172800 IN A 192.33.14.30
c.gtld-servers.net. 172800 IN A 192.26.92.30
d.gtld-servers.net. 172800 IN A 192.31.80.30
e.gtld-servers.net. 172800 IN A 192.12.94.30
f.gtld-servers.net. 172800 IN A 192.35.51.30
g.gtld-servers.net. 172800 IN A 192.42.93.30
h.gtld-servers.net. 172800 IN A 192.54.112.30
i.gtld-servers.net. 172800 IN A 192.43.172.30
j.gtld-servers.net. 172800 IN A 192.48.79.30
k.gtld-servers.net. 172800 IN A 192.52.178.30
l.gtld-servers.net. 172800 IN A 192.41.162.30
m.gtld-servers.net. 172800 IN A 192.55.83.30
;; Query time: 40 msec
;; SERVER: 2a01:4f8:192:43a5::2#53(2a01:4f8:192:43a5::2)
;; WHEN: Fri May 24 20:43:19 CEST 2019
;; MSG SIZE rcvd: 840
```
| 1.0 | dnsdist healthcheck with mustResolve doesn't check for answer - - Program: dnsdist
- Issue type: Bug report
### Short description
Creating a `newServer()` with `mustResolve=true` [according to the docs](https://dnsdist.org/reference/config.html#newServer) `MUST return a NOERROR RCODE and an answer`.
Currently it only checks for the `RCODE` to not be `NXDOMAIN` or `REFUSED`: https://github.com/PowerDNS/pdns/blob/18357a07c5b44366b0b3826b3ac66f4feea048be/pdns/dnsdist.cc#L1857
### Environment
- Operating system: Debian 9.9
- Software version: dnsdist 1.4.0-alpha2
- Software source: pdns repo
### Steps to reproduce
```lua
newServer({address="2a01:4f8:192:43a5::2", name="ns10.de.dns.opennic.glue", mustResolve=true})
```
This server does at the time of writing not send an ANSWER section (example at the bottom).
### Expected behaviour
Server should be detected as down.
### Actual behaviour
Server is detected mostly up.
Some extra output from `setVerboseHealthChecks(true)`
```
$ journalctl -u dnsdist.service -f | grep -F 2a01:4f8:192:43a5::2
May 24 18:46:13 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:46:13 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:46:17 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:46:28 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:46:28 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:46:33 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:07 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:07 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:11 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:30 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:30 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:34 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:39 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:39 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:43 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
May 24 18:47:47 thuban dnsdist[32726]: Timeout while waiting for the health check response from backend ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53)
May 24 18:47:47 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'down'
May 24 18:47:50 thuban dnsdist[32726]: Marking downstream ns10.de.dns.opennic.glue ([2a01:4f8:192:43a5::2]:53) as 'up'
```
### Other information
```
$ dig a a.root-servers.net. @2a01:4f8:192:43a5::2
; <<>> DiG 9.10.6 <<>> a a.root-servers.net. @2a01:4f8:192:43a5::2
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 19115
;; flags: qr rd; QUERY: 1, ANSWER: 0, AUTHORITY: 13, ADDITIONAL: 27
;; WARNING: recursion requested but not available
;; OPT PSEUDOSECTION:
; EDNS: version: 0, flags:; udp: 4096
;; QUESTION SECTION:
;a.root-servers.net. IN A
;; AUTHORITY SECTION:
net. 172800 IN NS a.gtld-servers.net.
net. 172800 IN NS b.gtld-servers.net.
net. 172800 IN NS c.gtld-servers.net.
net. 172800 IN NS d.gtld-servers.net.
net. 172800 IN NS e.gtld-servers.net.
net. 172800 IN NS f.gtld-servers.net.
net. 172800 IN NS g.gtld-servers.net.
net. 172800 IN NS h.gtld-servers.net.
net. 172800 IN NS i.gtld-servers.net.
net. 172800 IN NS j.gtld-servers.net.
net. 172800 IN NS k.gtld-servers.net.
net. 172800 IN NS l.gtld-servers.net.
net. 172800 IN NS m.gtld-servers.net.
;; ADDITIONAL SECTION:
a.gtld-servers.net. 172800 IN AAAA 2001:503:a83e::2:30
b.gtld-servers.net. 172800 IN AAAA 2001:503:231d::2:30
c.gtld-servers.net. 172800 IN AAAA 2001:503:83eb::30
d.gtld-servers.net. 172800 IN AAAA 2001:500:856e::30
e.gtld-servers.net. 172800 IN AAAA 2001:502:1ca1::30
f.gtld-servers.net. 172800 IN AAAA 2001:503:d414::30
g.gtld-servers.net. 172800 IN AAAA 2001:503:eea3::30
h.gtld-servers.net. 172800 IN AAAA 2001:502:8cc::30
i.gtld-servers.net. 172800 IN AAAA 2001:503:39c1::30
j.gtld-servers.net. 172800 IN AAAA 2001:502:7094::30
k.gtld-servers.net. 172800 IN AAAA 2001:503:d2d::30
l.gtld-servers.net. 172800 IN AAAA 2001:500:d937::30
m.gtld-servers.net. 172800 IN AAAA 2001:501:b1f9::30
a.gtld-servers.net. 172800 IN A 192.5.6.30
b.gtld-servers.net. 172800 IN A 192.33.14.30
c.gtld-servers.net. 172800 IN A 192.26.92.30
d.gtld-servers.net. 172800 IN A 192.31.80.30
e.gtld-servers.net. 172800 IN A 192.12.94.30
f.gtld-servers.net. 172800 IN A 192.35.51.30
g.gtld-servers.net. 172800 IN A 192.42.93.30
h.gtld-servers.net. 172800 IN A 192.54.112.30
i.gtld-servers.net. 172800 IN A 192.43.172.30
j.gtld-servers.net. 172800 IN A 192.48.79.30
k.gtld-servers.net. 172800 IN A 192.52.178.30
l.gtld-servers.net. 172800 IN A 192.41.162.30
m.gtld-servers.net. 172800 IN A 192.55.83.30
;; Query time: 40 msec
;; SERVER: 2a01:4f8:192:43a5::2#53(2a01:4f8:192:43a5::2)
;; WHEN: Fri May 24 20:43:19 CEST 2019
;; MSG SIZE rcvd: 840
```
| defect | dnsdist healthcheck with mustresolve doesn t check for answer program dnsdist issue type bug report short description creating a newserver with mustresolve true must return a noerror rcode and an answer currently it only checks for the rcode to not be nxdomain or refused environment operating system debian software version dnsdist software source pdns repo steps to reproduce lua newserver address name de dns opennic glue mustresolve true this server does at the time of writing not send an answer section example at the bottom expected behaviour server should be detected as down actual behaviour server is detected mostly up some extra output from setverbosehealthchecks true journalctl u dnsdist service f grep f may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up may thuban dnsdist timeout while waiting for the health check response from backend de dns opennic glue may thuban dnsdist marking downstream de dns opennic glue as down may thuban dnsdist marking downstream de dns opennic glue as up other information dig a a root servers net dig a a root servers net global options cmd got answer header opcode query status noerror id flags qr rd query answer authority additional warning recursion requested but not available opt pseudosection edns version flags udp question section a root servers net in a authority section net in ns a gtld servers net net in ns b gtld servers net net in ns c gtld servers net net in ns d gtld servers net net in ns e gtld servers net net in ns f gtld servers net net in ns g gtld servers net net in ns h gtld servers net net in ns i gtld servers net net in ns j gtld servers net net in ns k gtld servers net net in ns l gtld servers net net in ns m gtld servers net additional section a gtld servers net in aaaa b gtld servers net in aaaa c gtld servers net in aaaa d gtld servers net in aaaa e gtld servers net in aaaa f gtld servers net in aaaa g gtld servers net in aaaa h gtld servers net in aaaa i gtld servers net in aaaa j gtld servers net in aaaa k gtld servers net in aaaa l gtld servers net in aaaa m gtld servers net in aaaa a gtld servers net in a b gtld servers net in a c gtld servers net in a d gtld servers net in a e gtld servers net in a f gtld servers net in a g gtld servers net in a h gtld servers net in a i gtld servers net in a j gtld servers net in a k gtld servers net in a l gtld servers net in a m gtld servers net in a query time msec server when fri may cest msg size rcvd | 1 |
66,542 | 20,265,228,792 | IssuesEvent | 2022-02-15 11:23:07 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Custom cache engine is not available | defect | ### Description
I defined custom cache engine in `src/Cache/Engine/RedisCacheEngine`
In `config/app.php`, I defined the cache config for it.
```
'Cache' => [
'default' => [
'className' => 'File',
'path' => CACHE,
'url' => env('CACHE_DEFAULT_URL', null),
],
'my_session' => [
'className' => 'RedisCacheEngine',
'duration' => '+60 days',
'prefix' => 'my_session_',
'fallback' => 'default',
],
```
In `config/bootstrap.php`, I loaded the config
```
Cache::setConfig(Configure::consume('Cache'));
```
When I want to use `Cache::read('test', 'my_session')` to test it, it throws error:
```
Error: [BadMethodCallException] Cache engine RedisCacheEngine is not available. in /usr/local/rms/agency/releases/27/php/vendor/cakephp/cakephp/src/Cache/CacheRegistry.php on line 59
```
Is my cache configuration incorrect? What can I do to resolve this error? Thanks.
### CakePHP Version
4.2
### PHP Version
8.0 | 1.0 | Custom cache engine is not available - ### Description
I defined custom cache engine in `src/Cache/Engine/RedisCacheEngine`
In `config/app.php`, I defined the cache config for it.
```
'Cache' => [
'default' => [
'className' => 'File',
'path' => CACHE,
'url' => env('CACHE_DEFAULT_URL', null),
],
'my_session' => [
'className' => 'RedisCacheEngine',
'duration' => '+60 days',
'prefix' => 'my_session_',
'fallback' => 'default',
],
```
In `config/bootstrap.php`, I loaded the config
```
Cache::setConfig(Configure::consume('Cache'));
```
When I want to use `Cache::read('test', 'my_session')` to test it, it throws error:
```
Error: [BadMethodCallException] Cache engine RedisCacheEngine is not available. in /usr/local/rms/agency/releases/27/php/vendor/cakephp/cakephp/src/Cache/CacheRegistry.php on line 59
```
Is my cache configuration incorrect? What can I do to resolve this error? Thanks.
### CakePHP Version
4.2
### PHP Version
8.0 | defect | custom cache engine is not available description i defined custom cache engine in src cache engine rediscacheengine in config app php i defined the cache config for it cache default classname file path cache url env cache default url null my session classname rediscacheengine duration days prefix my session fallback default in config bootstrap php i loaded the config cache setconfig configure consume cache when i want to use cache read test my session to test it it throws error error cache engine rediscacheengine is not available in usr local rms agency releases php vendor cakephp cakephp src cache cacheregistry php on line is my cache configuration incorrect what can i do to resolve this error thanks cakephp version php version | 1 |
2,075 | 2,603,975,906 | IssuesEvent | 2015-02-24 19:01:30 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳阴茎疱疹好治好吗 | auto-migrated Priority-Medium Type-Defect | ```
沈阳阴茎疱疹好治好吗〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:19 | 1.0 | 沈阳阴茎疱疹好治好吗 - ```
沈阳阴茎疱疹好治好吗〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:19 | defect | 沈阳阴茎疱疹好治好吗 沈阳阴茎疱疹好治好吗〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at | 1 |
40,483 | 10,019,323,265 | IssuesEvent | 2019-07-16 09:50:22 | contao/contao | https://api.github.com/repos/contao/contao | closed | Language url parameter ignored, when requesting a page that does not exist | defect | **Affected version(s)**
4.7
**Description**
If you request a page that not exists in a multilanguage environment, e.g. `https://mydomain.tld/en/foobar`, the language parameter `en` is ignored and contao will deliver the 404 content based on your browser language.
`$_GET['language']` is not set here: https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/classes/Frontend.php#L312
In prior versions, this request would be routed via `FrontendIndex.php` and this would call `getPageIdFromUrl()`:
https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/controllers/FrontendIndex.php#L46
which itself sets the `language` property in `$_GET`:
https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/classes/Frontend.php#L259-L282
In 4.7 the Symfony CMF router throws a `NotFoundHttpException` which will be handled by the `PrettyErrorScreenListener` and the listener will call `getRootPageFromUrl()` directly (without executing `getPageIdFromUrl()` upfront).
This causes the language to be missed in `$_GET`.
**How to reproduce**
Set up a multilanguage installation with configured 404 pages and content.
Request a not existing page in a language that differs from your default browser language. You get the 404 content in your browser's default language instead of the target language. | 1.0 | Language url parameter ignored, when requesting a page that does not exist - **Affected version(s)**
4.7
**Description**
If you request a page that not exists in a multilanguage environment, e.g. `https://mydomain.tld/en/foobar`, the language parameter `en` is ignored and contao will deliver the 404 content based on your browser language.
`$_GET['language']` is not set here: https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/classes/Frontend.php#L312
In prior versions, this request would be routed via `FrontendIndex.php` and this would call `getPageIdFromUrl()`:
https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/controllers/FrontendIndex.php#L46
which itself sets the `language` property in `$_GET`:
https://github.com/contao/contao/blob/ed2fec5eaaa84875a9a14b5b90e803a65bccec7c/core-bundle/src/Resources/contao/classes/Frontend.php#L259-L282
In 4.7 the Symfony CMF router throws a `NotFoundHttpException` which will be handled by the `PrettyErrorScreenListener` and the listener will call `getRootPageFromUrl()` directly (without executing `getPageIdFromUrl()` upfront).
This causes the language to be missed in `$_GET`.
**How to reproduce**
Set up a multilanguage installation with configured 404 pages and content.
Request a not existing page in a language that differs from your default browser language. You get the 404 content in your browser's default language instead of the target language. | defect | language url parameter ignored when requesting a page that does not exist affected version s description if you request a page that not exists in a multilanguage environment e g the language parameter en is ignored and contao will deliver the content based on your browser language get is not set here in prior versions this request would be routed via frontendindex php and this would call getpageidfromurl which itself sets the language property in get in the symfony cmf router throws a notfoundhttpexception which will be handled by the prettyerrorscreenlistener and the listener will call getrootpagefromurl directly without executing getpageidfromurl upfront this causes the language to be missed in get how to reproduce set up a multilanguage installation with configured pages and content request a not existing page in a language that differs from your default browser language you get the content in your browser s default language instead of the target language | 1 |
13,610 | 2,772,593,269 | IssuesEvent | 2015-05-02 20:34:09 | agronholm/pythonfutures | https://api.github.com/repos/agronholm/pythonfutures | closed | release 2.1.4 regression. | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. acquire tarball from pypi, the only one I gather
2. unpack and count
3.
What is the expected output? What do you see instead?
CHANGES concurrent crawl.py docs futures futures.egg-info LICENSE
PKG-INFO primes.py setup.cfg setup.py test_futures.py tox.ini
concurrent futures futures.egg-info PKG-INFO setup.cfg setup.py
What version of the product are you using? On what operating system?
the latest release 2.1.4, gentoo
Please provide any additional information below.
Someone left out half the source code. ooopsie
```
Original issue reported on code.google.com by `del...@iinet.com.au` on 29 Jun 2013 at 7:25 | 1.0 | release 2.1.4 regression. - ```
What steps will reproduce the problem?
1. acquire tarball from pypi, the only one I gather
2. unpack and count
3.
What is the expected output? What do you see instead?
CHANGES concurrent crawl.py docs futures futures.egg-info LICENSE
PKG-INFO primes.py setup.cfg setup.py test_futures.py tox.ini
concurrent futures futures.egg-info PKG-INFO setup.cfg setup.py
What version of the product are you using? On what operating system?
the latest release 2.1.4, gentoo
Please provide any additional information below.
Someone left out half the source code. ooopsie
```
Original issue reported on code.google.com by `del...@iinet.com.au` on 29 Jun 2013 at 7:25 | defect | release regression what steps will reproduce the problem acquire tarball from pypi the only one i gather unpack and count what is the expected output what do you see instead changes concurrent crawl py docs futures futures egg info license pkg info primes py setup cfg setup py test futures py tox ini concurrent futures futures egg info pkg info setup cfg setup py what version of the product are you using on what operating system the latest release gentoo please provide any additional information below someone left out half the source code ooopsie original issue reported on code google com by del iinet com au on jun at | 1 |
74,778 | 25,320,889,696 | IssuesEvent | 2022-11-18 03:43:29 | AshleyYakeley/Truth | https://api.github.com/repos/AshleyYakeley/Truth | closed | Interpreter hangs instead of rejecting contravariant recursion | defect type system | For example:
`let f: rec r. (r -> Integer) = fn _ => 3 in 0`
Expected: reject
Found: hang | 1.0 | Interpreter hangs instead of rejecting contravariant recursion - For example:
`let f: rec r. (r -> Integer) = fn _ => 3 in 0`
Expected: reject
Found: hang | defect | interpreter hangs instead of rejecting contravariant recursion for example let f rec r r integer fn in expected reject found hang | 1 |
35,819 | 7,802,456,956 | IssuesEvent | 2018-06-10 12:49:54 | StrikeNP/trac_test | https://api.github.com/repos/StrikeNP/trac_test | closed | SILHS flexible paper: create mean and standard deviation plots for ensembles (Trac #776) | Migrated from Trac clubb_src defect raut@uwm.edu | '''Introduction'''
In #740, we created profile plots showing 12-member ensembles for each sampling method. A reviewer requested that we show mean and standard deviation plots for the ensemble members instead, as this would be easier to read. (https://github.com/larson-group/papers/issues/11#comment:130)
'''Technical spec'''
Let's do it!
Attachments:
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_mean.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_mean.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.2.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.2.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.3.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.3.pdf
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/776
```json
{
"status": "closed",
"changetime": "2015-12-22T02:17:06",
"description": "'''Introduction'''\n\nIn #740, we created profile plots showing 12-member ensembles for each sampling method. A reviewer requested that we show mean and standard deviation plots for the ensemble members instead, as this would be easier to read. (papers:ticket:11#comment:130)\n\n'''Technical spec'''\n\nLet's do it!",
"reporter": "raut@uwm.edu",
"cc": "vlarson@uwm.edu",
"resolution": "fixed",
"_ts": "1450750626862567",
"component": "clubb_src",
"summary": "SILHS flexible paper: create mean and standard deviation plots for ensembles",
"priority": "major",
"keywords": "",
"time": "2015-11-25T02:18:02",
"milestone": "",
"owner": "raut@uwm.edu",
"type": "defect"
}
```
| 1.0 | SILHS flexible paper: create mean and standard deviation plots for ensembles (Trac #776) - '''Introduction'''
In #740, we created profile plots showing 12-member ensembles for each sampling method. A reviewer requested that we show mean and standard deviation plots for the ensemble members instead, as this would be easier to read. (https://github.com/larson-group/papers/issues/11#comment:130)
'''Technical spec'''
Let's do it!
Attachments:
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_mean.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_mean.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.2.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.2.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/rico_lh_profiles_rmse.3.pdf
http://carson.math.uwm.edu/trac/clubb/attachment/ticket/776/dycoms2_rf02_do_profiles_rmse.3.pdf
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/776
```json
{
"status": "closed",
"changetime": "2015-12-22T02:17:06",
"description": "'''Introduction'''\n\nIn #740, we created profile plots showing 12-member ensembles for each sampling method. A reviewer requested that we show mean and standard deviation plots for the ensemble members instead, as this would be easier to read. (papers:ticket:11#comment:130)\n\n'''Technical spec'''\n\nLet's do it!",
"reporter": "raut@uwm.edu",
"cc": "vlarson@uwm.edu",
"resolution": "fixed",
"_ts": "1450750626862567",
"component": "clubb_src",
"summary": "SILHS flexible paper: create mean and standard deviation plots for ensembles",
"priority": "major",
"keywords": "",
"time": "2015-11-25T02:18:02",
"milestone": "",
"owner": "raut@uwm.edu",
"type": "defect"
}
```
| defect | silhs flexible paper create mean and standard deviation plots for ensembles trac introduction in we created profile plots showing member ensembles for each sampling method a reviewer requested that we show mean and standard deviation plots for the ensemble members instead as this would be easier to read technical spec let s do it attachments migrated from json status closed changetime description introduction n nin we created profile plots showing member ensembles for each sampling method a reviewer requested that we show mean and standard deviation plots for the ensemble members instead as this would be easier to read papers ticket comment n n technical spec n nlet s do it reporter raut uwm edu cc vlarson uwm edu resolution fixed ts component clubb src summary silhs flexible paper create mean and standard deviation plots for ensembles priority major keywords time milestone owner raut uwm edu type defect | 1 |
72,150 | 23,959,876,283 | IssuesEvent | 2022-09-12 18:04:55 | matrix-org/synapse | https://api.github.com/repos/matrix-org/synapse | opened | Unnecessary `device_lists.changed` entry in /sync after fetching device list | A-Device-List-Tracking A-Sync S-Tolerable T-Defect O-Occasional | While writing https://github.com/matrix-org/complement/pull/459, it was noticed that Synapse would sometimes emit an unnecessary `device_lists.changed` entry in the next sync after retrieving the device list for a remote user.
While it doesn't contravene the spirit of the spec, it's inefficient as well-behaved clients will ask the homeserver for device keys a second time. | 1.0 | Unnecessary `device_lists.changed` entry in /sync after fetching device list - While writing https://github.com/matrix-org/complement/pull/459, it was noticed that Synapse would sometimes emit an unnecessary `device_lists.changed` entry in the next sync after retrieving the device list for a remote user.
While it doesn't contravene the spirit of the spec, it's inefficient as well-behaved clients will ask the homeserver for device keys a second time. | defect | unnecessary device lists changed entry in sync after fetching device list while writing it was noticed that synapse would sometimes emit an unnecessary device lists changed entry in the next sync after retrieving the device list for a remote user while it doesn t contravene the spirit of the spec it s inefficient as well behaved clients will ask the homeserver for device keys a second time | 1 |
48,557 | 13,134,785,248 | IssuesEvent | 2020-08-07 00:37:21 | networkx/networkx | https://api.github.com/repos/networkx/networkx | closed | A better returning value for draw_networkx_edges | Defect Needs PR | `draw_networkx_edges` returns either `LineCollection` or `arrow_collection` (which is a list of `FancyArrowPatch`) depending on whether the given graph is `DiGraph` and `arrows=True`.
The documentation says
```
matplotlib.collection.LineCollection
`LineCollection` of the edges
list of matplotlib.patches.FancyArrowPatch
`FancyArrowPatch` instances of the directed edges
```
Internally, the function is short-circuited by checking whether the `edgelist` is empty:
```python
if not edgelist or len(edgelist) == 0: # no edges!
return None
```
Wouldn't it be better if
```python
if not edgelist or len(edgelist) == 0: # no edges!
if G.is_directed() and arrows:
return []
else:
return LineCollection(None)
```
Currently, the documentation does not mention that it returns `None` if no edge is drawn.
p.s. `draw_networkx_labels` returns an empty set if nothing is drawn. `draw_networkx_nodes` returns `None` but it may be able to return `PathCollection(None)` since documentation did not specify its behavior.
p.p.s. I am using `zip(edgelist, arcs)` after `arcs = draw_networkx_edges...`. It would be better to have an empty list than `None` which needs an additional check.
| 1.0 | A better returning value for draw_networkx_edges - `draw_networkx_edges` returns either `LineCollection` or `arrow_collection` (which is a list of `FancyArrowPatch`) depending on whether the given graph is `DiGraph` and `arrows=True`.
The documentation says
```
matplotlib.collection.LineCollection
`LineCollection` of the edges
list of matplotlib.patches.FancyArrowPatch
`FancyArrowPatch` instances of the directed edges
```
Internally, the function is short-circuited by checking whether the `edgelist` is empty:
```python
if not edgelist or len(edgelist) == 0: # no edges!
return None
```
Wouldn't it be better if
```python
if not edgelist or len(edgelist) == 0: # no edges!
if G.is_directed() and arrows:
return []
else:
return LineCollection(None)
```
Currently, the documentation does not mention that it returns `None` if no edge is drawn.
p.s. `draw_networkx_labels` returns an empty set if nothing is drawn. `draw_networkx_nodes` returns `None` but it may be able to return `PathCollection(None)` since documentation did not specify its behavior.
p.p.s. I am using `zip(edgelist, arcs)` after `arcs = draw_networkx_edges...`. It would be better to have an empty list than `None` which needs an additional check.
| defect | a better returning value for draw networkx edges draw networkx edges returns either linecollection or arrow collection which is a list of fancyarrowpatch depending on whether the given graph is digraph and arrows true the documentation says matplotlib collection linecollection linecollection of the edges list of matplotlib patches fancyarrowpatch fancyarrowpatch instances of the directed edges internally the function is short circuited by checking whether the edgelist is empty python if not edgelist or len edgelist no edges return none wouldn t it be better if python if not edgelist or len edgelist no edges if g is directed and arrows return else return linecollection none currently the documentation does not mention that it returns none if no edge is drawn p s draw networkx labels returns an empty set if nothing is drawn draw networkx nodes returns none but it may be able to return pathcollection none since documentation did not specify its behavior p p s i am using zip edgelist arcs after arcs draw networkx edges it would be better to have an empty list than none which needs an additional check | 1 |
38,366 | 8,789,029,398 | IssuesEvent | 2018-12-21 01:10:54 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Typehint errors in ViewBuilder | Defect | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.7.1.
* Platform and Target: N/A
### What you did
Ran PHPStan for a few things and encountered some more typehints that are missing
* `ViewBuilder::setTemplatePath()`
* `ViewBuilder::getTemplatePath()`
* `ViewBuilder::setLayoutPath()`
* `ViewBuilder::getLayoutPath()`
* `ViewBuilder::setTemplate()`
* `ViewBuilder::getTemplate()`
* `ViewBuilder::setLayout()`
* `ViewBuilder::getLayout()`
* `ViewBuilder::setName()`
* `ViewBuilder::getName()`
### What happened
PHPStan is failing due to invalid typehints
### What you expected to happen
PHPStan to run without errors
PR to follow
| 1.0 | Typehint errors in ViewBuilder - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.7.1.
* Platform and Target: N/A
### What you did
Ran PHPStan for a few things and encountered some more typehints that are missing
* `ViewBuilder::setTemplatePath()`
* `ViewBuilder::getTemplatePath()`
* `ViewBuilder::setLayoutPath()`
* `ViewBuilder::getLayoutPath()`
* `ViewBuilder::setTemplate()`
* `ViewBuilder::getTemplate()`
* `ViewBuilder::setLayout()`
* `ViewBuilder::getLayout()`
* `ViewBuilder::setName()`
* `ViewBuilder::getName()`
### What happened
PHPStan is failing due to invalid typehints
### What you expected to happen
PHPStan to run without errors
PR to follow
| defect | typehint errors in viewbuilder this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target n a what you did ran phpstan for a few things and encountered some more typehints that are missing viewbuilder settemplatepath viewbuilder gettemplatepath viewbuilder setlayoutpath viewbuilder getlayoutpath viewbuilder settemplate viewbuilder gettemplate viewbuilder setlayout viewbuilder getlayout viewbuilder setname viewbuilder getname what happened phpstan is failing due to invalid typehints what you expected to happen phpstan to run without errors pr to follow | 1 |
463,298 | 13,262,753,877 | IssuesEvent | 2020-08-20 22:31:40 | Poobslag/turbofat | https://api.github.com/repos/Poobslag/turbofat | closed | Customer camera should scale as customer grows in size | priority-4 | As customers grow in size, the camera should zoom out to keep more of them in frame. | 1.0 | Customer camera should scale as customer grows in size - As customers grow in size, the camera should zoom out to keep more of them in frame. | non_defect | customer camera should scale as customer grows in size as customers grow in size the camera should zoom out to keep more of them in frame | 0 |
158,506 | 13,734,831,092 | IssuesEvent | 2020-10-05 09:15:53 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | closed | Kubecf docs search doesn't work | Priority: Medium Type: Documentation | **Is your feature request related to a problem? Please describe.**
The search mechanism on [kubecf docs repo](https://kubecf.suse.dev/docs/) is terrible and/or even broken.
**Describe the solution you'd like**
We need to understand how hugo(?) search works and if we can make it better.
**Describe alternatives you've considered**
N/A
**Additional context**
N/A
| 1.0 | Kubecf docs search doesn't work - **Is your feature request related to a problem? Please describe.**
The search mechanism on [kubecf docs repo](https://kubecf.suse.dev/docs/) is terrible and/or even broken.
**Describe the solution you'd like**
We need to understand how hugo(?) search works and if we can make it better.
**Describe alternatives you've considered**
N/A
**Additional context**
N/A
| non_defect | kubecf docs search doesn t work is your feature request related to a problem please describe the search mechanism on is terrible and or even broken describe the solution you d like we need to understand how hugo search works and if we can make it better describe alternatives you ve considered n a additional context n a | 0 |
169,624 | 6,412,971,994 | IssuesEvent | 2017-08-08 05:58:02 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | closed | Bookmarks are sorted differently in qute:bookmarks and completion | component: completion priority: 2 - low | qute:bookmarks sorts by title/name, but completion by the (easily customizable) order from the files where they are stored. I would expect these to have the same order for easier orientation. | 1.0 | Bookmarks are sorted differently in qute:bookmarks and completion - qute:bookmarks sorts by title/name, but completion by the (easily customizable) order from the files where they are stored. I would expect these to have the same order for easier orientation. | non_defect | bookmarks are sorted differently in qute bookmarks and completion qute bookmarks sorts by title name but completion by the easily customizable order from the files where they are stored i would expect these to have the same order for easier orientation | 0 |
247,262 | 7,915,905,221 | IssuesEvent | 2018-07-04 02:31:50 | Automattic/mongoose | https://api.github.com/repos/Automattic/mongoose | closed | URGENT 5.2.0: _Cannot specify a custom index on `_id`, MongoDB does not allow overwriting the default `_id` index_ is not correctly implemented | confirmed-bug priority | 5.2.0 took a too restrictive measure to fix #6605:
- it makes sense not to create a duplicate index on _id
- it does not make sense to prevent any index designating _id as part of a list of indexed fields, e.g.:
- { _categoryId: 1, _id: 1}
- { _userId: 1, _id: 1 }
With 5.2.0, any index containing _id: 1 raises an error because of https://github.com/Automattic/mongoose/blob/master/lib/model.js#L1208
```
if (index[0]._id != null) {
```
should be replaced with something like
```
if (Object.keys(index[0]).length === 1 && index[0]._id != null) {
```
| 1.0 | URGENT 5.2.0: _Cannot specify a custom index on `_id`, MongoDB does not allow overwriting the default `_id` index_ is not correctly implemented - 5.2.0 took a too restrictive measure to fix #6605:
- it makes sense not to create a duplicate index on _id
- it does not make sense to prevent any index designating _id as part of a list of indexed fields, e.g.:
- { _categoryId: 1, _id: 1}
- { _userId: 1, _id: 1 }
With 5.2.0, any index containing _id: 1 raises an error because of https://github.com/Automattic/mongoose/blob/master/lib/model.js#L1208
```
if (index[0]._id != null) {
```
should be replaced with something like
```
if (Object.keys(index[0]).length === 1 && index[0]._id != null) {
```
| non_defect | urgent cannot specify a custom index on id mongodb does not allow overwriting the default id index is not correctly implemented took a too restrictive measure to fix it makes sense not to create a duplicate index on id it does not make sense to prevent any index designating id as part of a list of indexed fields e g categoryid id userid id with any index containing id raises an error because of if index id null should be replaced with something like if object keys index length index id null | 0 |
78,351 | 10,058,482,194 | IssuesEvent | 2019-07-22 14:00:44 | symuvia/symupy | https://api.github.com/repos/symuvia/symupy | closed | Package as PyPi | documentation question | This issue is just to elaborate the idea on packaging symupy as a standard python package
- [x] Recipe `setuptools`
- [x] Recipe `conda build`
- [ ] Integrate and distribute binaries (simulator)
- [ ] Adapt testing to the distributed binaries
- [x] Publish package on pypi first version | 1.0 | Package as PyPi - This issue is just to elaborate the idea on packaging symupy as a standard python package
- [x] Recipe `setuptools`
- [x] Recipe `conda build`
- [ ] Integrate and distribute binaries (simulator)
- [ ] Adapt testing to the distributed binaries
- [x] Publish package on pypi first version | non_defect | package as pypi this issue is just to elaborate the idea on packaging symupy as a standard python package recipe setuptools recipe conda build integrate and distribute binaries simulator adapt testing to the distributed binaries publish package on pypi first version | 0 |
389,819 | 11,517,594,378 | IssuesEvent | 2020-02-14 08:45:10 | DimensionDev/Maskbook | https://api.github.com/repos/DimensionDev/Maskbook | closed | Rationalise localization | Component: i18n Priority: P4 (Do when free) | Some users reports say that Maskbook is being used in Hong Kong too. Looks like we can't just settle with a single `zh`: we need to add some subdivisions. While we are at it, we should also rationalize our localisation system.
- [ ] Use a proper library that has:
- [ ] Plurals
- [ ] Grammatical Gender
- [ ] Put it on a platform with Git hooks
* * *
I recommend Gettext/Jed if we want to put the thing on some translation platform immediately; i18next is sort of decent at that too. If we want something really good and powerful, however, I strongly recommend using [MessageFormat](https://messageformat.github.io/). Few platforms support it for now, but Weblate (WeblateOrg/weblate#2967) might do soon. There's some precedence of its adoption in the JS community by Angular folks, so we aren't doing anything crazy here.
I don't really care about what platform we use, although again I am strongly leaning Weblate. If we are using Crowdin, however, be sure to review everything and don't let these damned stale votes decide what strings to use.
PS: don't change the spelling mix | 1.0 | Rationalise localization - Some users reports say that Maskbook is being used in Hong Kong too. Looks like we can't just settle with a single `zh`: we need to add some subdivisions. While we are at it, we should also rationalize our localisation system.
- [ ] Use a proper library that has:
- [ ] Plurals
- [ ] Grammatical Gender
- [ ] Put it on a platform with Git hooks
* * *
I recommend Gettext/Jed if we want to put the thing on some translation platform immediately; i18next is sort of decent at that too. If we want something really good and powerful, however, I strongly recommend using [MessageFormat](https://messageformat.github.io/). Few platforms support it for now, but Weblate (WeblateOrg/weblate#2967) might do soon. There's some precedence of its adoption in the JS community by Angular folks, so we aren't doing anything crazy here.
I don't really care about what platform we use, although again I am strongly leaning Weblate. If we are using Crowdin, however, be sure to review everything and don't let these damned stale votes decide what strings to use.
PS: don't change the spelling mix | non_defect | rationalise localization some users reports say that maskbook is being used in hong kong too looks like we can t just settle with a single zh we need to add some subdivisions while we are at it we should also rationalize our localisation system use a proper library that has plurals grammatical gender put it on a platform with git hooks i recommend gettext jed if we want to put the thing on some translation platform immediately is sort of decent at that too if we want something really good and powerful however i strongly recommend using few platforms support it for now but weblate weblateorg weblate might do soon there s some precedence of its adoption in the js community by angular folks so we aren t doing anything crazy here i don t really care about what platform we use although again i am strongly leaning weblate if we are using crowdin however be sure to review everything and don t let these damned stale votes decide what strings to use ps don t change the spelling mix | 0 |
143,073 | 21,938,393,448 | IssuesEvent | 2022-05-23 15:36:44 | vizhub-core/vizhub-issue-tracker | https://api.github.com/repos/vizhub-core/vizhub-issue-tracker | closed | Presence for Files | Design | Idea: In the files sidebar, show presence avatars to indicate what file each user is editing at the moment. | 1.0 | Presence for Files - Idea: In the files sidebar, show presence avatars to indicate what file each user is editing at the moment. | non_defect | presence for files idea in the files sidebar show presence avatars to indicate what file each user is editing at the moment | 0 |
20,339 | 29,870,061,029 | IssuesEvent | 2023-06-20 07:54:05 | mirage/mirage | https://api.github.com/repos/mirage/mirage | closed | should the type for interfaces be a numerical index, not a string? | breaks compatibility improvement | See discussion at https://github.com/mirage/mirage/pull/643 . Currently, the value is a string which some implementations then attempt to interpret as a number, which can be a runtime exception (see https://github.com/mirage/mirage-net-xen/blob/master/lib/frontend.ml#L238 ). | True | should the type for interfaces be a numerical index, not a string? - See discussion at https://github.com/mirage/mirage/pull/643 . Currently, the value is a string which some implementations then attempt to interpret as a number, which can be a runtime exception (see https://github.com/mirage/mirage-net-xen/blob/master/lib/frontend.ml#L238 ). | non_defect | should the type for interfaces be a numerical index not a string see discussion at currently the value is a string which some implementations then attempt to interpret as a number which can be a runtime exception see | 0 |
45,704 | 13,038,151,611 | IssuesEvent | 2020-07-28 14:48:53 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | p-autocomplete not aligned with ui-inputgroup-addon using ui-fluid p-formgrid p-grid | defect | The alignment of the input field of p-autoComplete is not consistent with the ui-inputgroup-addon. Expected behavior would be the same as input of pInputText.
Code to reproduce
```
<div class="ui-fluid p-formgrid p-grid">
<div class="p-field p-col-12 p-md-6">
<label for="owner">Owner</label>
<div class="ui-inputgroup">
<span class="ui-inputgroup-addon"><i class="pi pi-user"></i></span>
<p-autoComplete id="owner">
</p-autoComplete>
</div>
</div>
</div>
```
Version of modules
"@angular/animations": "~10.0.3",
"@angular/cdk": "^10.1.1",
"@angular/common": "~10.0.3",
"@angular/compiler": "~10.0.3",
"@angular/core": "~10.0.3",
"@angular/forms": "~10.0.3",
"@angular/platform-browser": "~10.0.3",
"@angular/platform-browser-dynamic": "~10.0.3",
"@angular/router": "~10.0.3",
"@fullcalendar/core": "^5.1.0",
"bootstrap": "^4.5.0",
"chart.js": "^2.9.3",
"primeflex": "^2.0.0-rc.1",
"primeicons": "^4.0.0",
"primeng": "^9.1.3",
"quill": "^1.3.7",
"rxjs": "~6.5.5",
"tslib": "^2.0.0",
"zone.js": "~0.10.3", | 1.0 | p-autocomplete not aligned with ui-inputgroup-addon using ui-fluid p-formgrid p-grid - The alignment of the input field of p-autoComplete is not consistent with the ui-inputgroup-addon. Expected behavior would be the same as input of pInputText.
Code to reproduce
```
<div class="ui-fluid p-formgrid p-grid">
<div class="p-field p-col-12 p-md-6">
<label for="owner">Owner</label>
<div class="ui-inputgroup">
<span class="ui-inputgroup-addon"><i class="pi pi-user"></i></span>
<p-autoComplete id="owner">
</p-autoComplete>
</div>
</div>
</div>
```
Version of modules
"@angular/animations": "~10.0.3",
"@angular/cdk": "^10.1.1",
"@angular/common": "~10.0.3",
"@angular/compiler": "~10.0.3",
"@angular/core": "~10.0.3",
"@angular/forms": "~10.0.3",
"@angular/platform-browser": "~10.0.3",
"@angular/platform-browser-dynamic": "~10.0.3",
"@angular/router": "~10.0.3",
"@fullcalendar/core": "^5.1.0",
"bootstrap": "^4.5.0",
"chart.js": "^2.9.3",
"primeflex": "^2.0.0-rc.1",
"primeicons": "^4.0.0",
"primeng": "^9.1.3",
"quill": "^1.3.7",
"rxjs": "~6.5.5",
"tslib": "^2.0.0",
"zone.js": "~0.10.3", | defect | p autocomplete not aligned with ui inputgroup addon using ui fluid p formgrid p grid the alignment of the input field of p autocomplete is not consistent with the ui inputgroup addon expected behavior would be the same as input of pinputtext code to reproduce owner version of modules angular animations angular cdk angular common angular compiler angular core angular forms angular platform browser angular platform browser dynamic angular router fullcalendar core bootstrap chart js primeflex rc primeicons primeng quill rxjs tslib zone js | 1 |
39,309 | 19,808,537,128 | IssuesEvent | 2022-01-19 09:42:48 | PostHog/posthog | https://api.github.com/repos/PostHog/posthog | closed | Related groups query is slow | performance persons group-analytics | ## In what situation are you experiencing subpar performance?
Opening the related groups tab takes a while
## How to reproduce
1. Go to a groups page
2. Open related groups
3. Wait ~5-10s
## Environment
- [x] PostHog Cloud
- [ ] self-hosted PostHog, version/commit: _please provide_
## Additional context
Left-over from groups project. Waiting to see if we'll get some usage for this feature as well.
#### *Thank you* for your performance issue report – we want PostHog to go supersonic!
| True | Related groups query is slow - ## In what situation are you experiencing subpar performance?
Opening the related groups tab takes a while
## How to reproduce
1. Go to a groups page
2. Open related groups
3. Wait ~5-10s
## Environment
- [x] PostHog Cloud
- [ ] self-hosted PostHog, version/commit: _please provide_
## Additional context
Left-over from groups project. Waiting to see if we'll get some usage for this feature as well.
#### *Thank you* for your performance issue report – we want PostHog to go supersonic!
| non_defect | related groups query is slow in what situation are you experiencing subpar performance opening the related groups tab takes a while how to reproduce go to a groups page open related groups wait environment posthog cloud self hosted posthog version commit please provide additional context left over from groups project waiting to see if we ll get some usage for this feature as well thank you for your performance issue report – we want posthog to go supersonic | 0 |
119,721 | 10,062,134,211 | IssuesEvent | 2019-07-22 23:47:16 | jahshaka/Studio | https://api.github.com/repos/jahshaka/Studio | closed | new scene desktop crash | fixed and waiting to be tested | create new scene crashes app in desktop when a scene is already loaded and open | 1.0 | new scene desktop crash - create new scene crashes app in desktop when a scene is already loaded and open | non_defect | new scene desktop crash create new scene crashes app in desktop when a scene is already loaded and open | 0 |
20,964 | 11,571,899,490 | IssuesEvent | 2020-02-20 22:36:01 | Azure/azure-sdk-for-js | https://api.github.com/repos/Azure/azure-sdk-for-js | closed | Q: looking the the Cognitive Services Speech SDK | Client Cognitive Services Question customer-reported | I'm trying to find the cognitive service speech sdk and typescript samples as I'm having issues using the currently published 'microsoft-cognitiveservices-speech-sdk' v1.9.0 | 1.0 | Q: looking the the Cognitive Services Speech SDK - I'm trying to find the cognitive service speech sdk and typescript samples as I'm having issues using the currently published 'microsoft-cognitiveservices-speech-sdk' v1.9.0 | non_defect | q looking the the cognitive services speech sdk i m trying to find the cognitive service speech sdk and typescript samples as i m having issues using the currently published microsoft cognitiveservices speech sdk | 0 |
20,591 | 3,385,817,123 | IssuesEvent | 2015-11-27 13:53:17 | relaxng/jing-trang | https://api.github.com/repos/relaxng/jing-trang | closed | NullPointerException running trang 20131210 | auto-migrated Priority-Medium Type-Defect | ```
The changes introduced in revision 2354
(http://code.google.com/p/jing-trang/source/detail?r=2354#) result in a
NullPointerException when trang is run without specifying a catalogue URI via
the "-C" argument. This occurs because 'resolver' is set to null in main() and
so the ResolverUriEntityManager object gets constructed with a null resolver.
Therefore a NullPointerException is generated when
ResolverUriEntityManager::open() attempts to call resolver.open(). The stack
trace is as follows:
Exception in thread "main" java.lang.NullPointerException
at com.thaiopensource.xml.em.ResolverUriEntityManager.open(ResolverUriEntityManager.java:56)
at com.thaiopensource.xml.em.ResolverUriEntityManager.open(ResolverUriEntityManager.java:27)
at com.thaiopensource.xml.dtd.parse.DtdParserImpl.parse(DtdParserImpl.java:14)
at com.thaiopensource.relaxng.input.dtd.DtdInputFormat.load(DtdInputFormat.java:144)
at com.thaiopensource.relaxng.translate.Driver.run(Driver.java:135)
at com.thaiopensource.relaxng.translate.Driver.main(Driver.java:44)
The attached patch fixes this problem by instantiating a UriEntityManager
instead of a ResolverUriEntityManager when no resolver is available.
```
Original issue reported on code.google.com by `zztop...@gmail.com` on 31 Jan 2014 at 1:18
Attachments:
* [trang.diff](https://storage.googleapis.com/google-code-attachments/jing-trang/issue-177/comment-0/trang.diff)
| 1.0 | NullPointerException running trang 20131210 - ```
The changes introduced in revision 2354
(http://code.google.com/p/jing-trang/source/detail?r=2354#) result in a
NullPointerException when trang is run without specifying a catalogue URI via
the "-C" argument. This occurs because 'resolver' is set to null in main() and
so the ResolverUriEntityManager object gets constructed with a null resolver.
Therefore a NullPointerException is generated when
ResolverUriEntityManager::open() attempts to call resolver.open(). The stack
trace is as follows:
Exception in thread "main" java.lang.NullPointerException
at com.thaiopensource.xml.em.ResolverUriEntityManager.open(ResolverUriEntityManager.java:56)
at com.thaiopensource.xml.em.ResolverUriEntityManager.open(ResolverUriEntityManager.java:27)
at com.thaiopensource.xml.dtd.parse.DtdParserImpl.parse(DtdParserImpl.java:14)
at com.thaiopensource.relaxng.input.dtd.DtdInputFormat.load(DtdInputFormat.java:144)
at com.thaiopensource.relaxng.translate.Driver.run(Driver.java:135)
at com.thaiopensource.relaxng.translate.Driver.main(Driver.java:44)
The attached patch fixes this problem by instantiating a UriEntityManager
instead of a ResolverUriEntityManager when no resolver is available.
```
Original issue reported on code.google.com by `zztop...@gmail.com` on 31 Jan 2014 at 1:18
Attachments:
* [trang.diff](https://storage.googleapis.com/google-code-attachments/jing-trang/issue-177/comment-0/trang.diff)
| defect | nullpointerexception running trang the changes introduced in revision result in a nullpointerexception when trang is run without specifying a catalogue uri via the c argument this occurs because resolver is set to null in main and so the resolverurientitymanager object gets constructed with a null resolver therefore a nullpointerexception is generated when resolverurientitymanager open attempts to call resolver open the stack trace is as follows exception in thread main java lang nullpointerexception at com thaiopensource xml em resolverurientitymanager open resolverurientitymanager java at com thaiopensource xml em resolverurientitymanager open resolverurientitymanager java at com thaiopensource xml dtd parse dtdparserimpl parse dtdparserimpl java at com thaiopensource relaxng input dtd dtdinputformat load dtdinputformat java at com thaiopensource relaxng translate driver run driver java at com thaiopensource relaxng translate driver main driver java the attached patch fixes this problem by instantiating a urientitymanager instead of a resolverurientitymanager when no resolver is available original issue reported on code google com by zztop gmail com on jan at attachments | 1 |
90,272 | 26,029,001,110 | IssuesEvent | 2022-12-21 19:06:28 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | bazel: remove reference to deprecated function `github.com/bazelbuild/rules_go/go/tools/bazel.Runfile` | C-bug A-build-system T-dev-inf | ```
pkg/build/bazel/bazel.go:38:9: inner.Runfile is deprecated: Use github.com/bazelbuild/rules_go/go/runfiles instead for cross-platform support matching the behavior of the Bazel-provided runfiles libraries. (SA1019)
``` | 1.0 | bazel: remove reference to deprecated function `github.com/bazelbuild/rules_go/go/tools/bazel.Runfile` - ```
pkg/build/bazel/bazel.go:38:9: inner.Runfile is deprecated: Use github.com/bazelbuild/rules_go/go/runfiles instead for cross-platform support matching the behavior of the Bazel-provided runfiles libraries. (SA1019)
``` | non_defect | bazel remove reference to deprecated function github com bazelbuild rules go go tools bazel runfile pkg build bazel bazel go inner runfile is deprecated use github com bazelbuild rules go go runfiles instead for cross platform support matching the behavior of the bazel provided runfiles libraries | 0 |
170,641 | 6,461,955,941 | IssuesEvent | 2017-08-16 09:26:51 | nathan-b/colortoggle | https://api.github.com/repos/nathan-b/colortoggle | closed | Add third color scheme | auto-migrated Priority-Medium Type-Enhancement | ```
Request: Allow toggle between 3 color schemes
Currently using: color toogle 0.15 on firefox 22
Usage scenario: Before I discover colortoggle I manually switch between 3
color schemes:
(1) default
(2) dark mode
javascript:(function(){N=document.createElement('link');S='*{background:black%20
!important;color:lightgrey%20!important}:link,:link%20*{color:#ddddff%20!importa
nt}:visited,:visited%20*{color:#ddffdd%20!important}';N.rel='stylesheet';N.href=
'data:text/css,'+escape(S);document.getElementsByTagName('head')[0].appendChild(
N)})()
(3) grey mode
javascript:(function(){N=document.createElement('link');S='*{background:grey%20!
important;color:black%20!important}:link,:link%20*{color:#ddddff%20!important}:v
isited,:visited%20*{color:#ddffdd%20!important}';N.rel='stylesheet';N.href='data
:text/css,'+escape(S);document.getElementsByTagName('head')[0].appendChild(N)})(
)
The reason for keeping three themes are as follow:
(a) some websites only work with their default color, whence (1)
(b) at night/in dark room the dark mode works very well
(c) during day time/in bright room I find the contrast between the dark mode
and the bright surrending a bit too much for my eyes, whence the grey mode.
At first I thought I was being picky, but when my coworks/friends saw these
they all want to do this, and (just like me) they wish that they could toggle
between the three modes. And no, no one, myself included, ask for a 4th scheme
:-)
Many THANKS for consideration!
```
Original issue reported on code.google.com by `partita1...@gmail.com` on 12 Jul 2013 at 6:26
| 1.0 | Add third color scheme - ```
Request: Allow toggle between 3 color schemes
Currently using: color toogle 0.15 on firefox 22
Usage scenario: Before I discover colortoggle I manually switch between 3
color schemes:
(1) default
(2) dark mode
javascript:(function(){N=document.createElement('link');S='*{background:black%20
!important;color:lightgrey%20!important}:link,:link%20*{color:#ddddff%20!importa
nt}:visited,:visited%20*{color:#ddffdd%20!important}';N.rel='stylesheet';N.href=
'data:text/css,'+escape(S);document.getElementsByTagName('head')[0].appendChild(
N)})()
(3) grey mode
javascript:(function(){N=document.createElement('link');S='*{background:grey%20!
important;color:black%20!important}:link,:link%20*{color:#ddddff%20!important}:v
isited,:visited%20*{color:#ddffdd%20!important}';N.rel='stylesheet';N.href='data
:text/css,'+escape(S);document.getElementsByTagName('head')[0].appendChild(N)})(
)
The reason for keeping three themes are as follow:
(a) some websites only work with their default color, whence (1)
(b) at night/in dark room the dark mode works very well
(c) during day time/in bright room I find the contrast between the dark mode
and the bright surrending a bit too much for my eyes, whence the grey mode.
At first I thought I was being picky, but when my coworks/friends saw these
they all want to do this, and (just like me) they wish that they could toggle
between the three modes. And no, no one, myself included, ask for a 4th scheme
:-)
Many THANKS for consideration!
```
Original issue reported on code.google.com by `partita1...@gmail.com` on 12 Jul 2013 at 6:26
| non_defect | add third color scheme request allow toggle between color schemes currently using color toogle on firefox usage scenario before i discover colortoggle i manually switch between color schemes default dark mode javascript function n document createelement link s background black important color lightgrey important link link color ddddff importa nt visited visited color ddffdd important n rel stylesheet n href data text css escape s document getelementsbytagname head appendchild n grey mode javascript function n document createelement link s background grey important color black important link link color ddddff important v isited visited color ddffdd important n rel stylesheet n href data text css escape s document getelementsbytagname head appendchild n the reason for keeping three themes are as follow a some websites only work with their default color whence b at night in dark room the dark mode works very well c during day time in bright room i find the contrast between the dark mode and the bright surrending a bit too much for my eyes whence the grey mode at first i thought i was being picky but when my coworks friends saw these they all want to do this and just like me they wish that they could toggle between the three modes and no no one myself included ask for a scheme many thanks for consideration original issue reported on code google com by gmail com on jul at | 0 |
328,267 | 28,110,364,659 | IssuesEvent | 2023-03-31 06:36:18 | CeresDB/ceresdb | https://api.github.com/repos/CeresDB/ceresdb | closed | Support integration tests with CeresMeta | enhancement A-cluster A-test | ### Describe This Problem
Currently, basic cluster mode of CeresDB has been supported after V0.4 is released. However, the test for that is all by manual, and we are looking forward to an automatic integration tests for the clustering features.
### Proposal
Just like the TSBS, let's choose a specific version of CeresMeta to build, and run multiple ceresdb-server with cluster configs under the control of CeresMeta. After all the service is ready, run a new harness test suites for the cluster.
### Additional Context
_No response_ | 1.0 | Support integration tests with CeresMeta - ### Describe This Problem
Currently, basic cluster mode of CeresDB has been supported after V0.4 is released. However, the test for that is all by manual, and we are looking forward to an automatic integration tests for the clustering features.
### Proposal
Just like the TSBS, let's choose a specific version of CeresMeta to build, and run multiple ceresdb-server with cluster configs under the control of CeresMeta. After all the service is ready, run a new harness test suites for the cluster.
### Additional Context
_No response_ | non_defect | support integration tests with ceresmeta describe this problem currently basic cluster mode of ceresdb has been supported after is released however the test for that is all by manual and we are looking forward to an automatic integration tests for the clustering features proposal just like the tsbs let s choose a specific version of ceresmeta to build and run multiple ceresdb server with cluster configs under the control of ceresmeta after all the service is ready run a new harness test suites for the cluster additional context no response | 0 |
772,918 | 27,141,679,520 | IssuesEvent | 2023-02-16 16:46:31 | googleapis/python-bigquery-sqlalchemy | https://api.github.com/repos/googleapis/python-bigquery-sqlalchemy | closed | tests.sqlalchemy_dialect_compliance.test_dialect_compliance.TimestampMicrosecondsTest_bigquery+bigquery: test_literal failed | type: bug priority: p1 flakybot: issue api: bigquery | Note: #698 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 074321ddaa10001773e7e6044f4a0df1bb530331
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call
result: Optional[TResult] = func()
^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda>
lambda: ihook(item=item, **kwds), when=when, reraise=reraise
^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall
return outcome.get_result()
^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result
raise ex[1].with_traceback(ex[2])
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall
res = hook_impl.function(*args)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown
item.session._setupstate.teardown_exact(item, nextitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact
self._teardown_towards(needed_collectors)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards
raise exc
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards
self._pop_and_teardown()
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown
self._teardown_with_finalization(colitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization
assert colitem in self.stack
^^^^^^^^^^^^^^^^^^^^^
AssertionError</pre></details> | 1.0 | tests.sqlalchemy_dialect_compliance.test_dialect_compliance.TimestampMicrosecondsTest_bigquery+bigquery: test_literal failed - Note: #698 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 074321ddaa10001773e7e6044f4a0df1bb530331
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call
result: Optional[TResult] = func()
^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda>
lambda: ihook(item=item, **kwds), when=when, reraise=reraise
^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall
return outcome.get_result()
^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result
raise ex[1].with_traceback(ex[2])
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall
res = hook_impl.function(*args)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown
item.session._setupstate.teardown_exact(item, nextitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact
self._teardown_towards(needed_collectors)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards
raise exc
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards
self._pop_and_teardown()
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown
self._teardown_with_finalization(colitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization
assert colitem in self.stack
^^^^^^^^^^^^^^^^^^^^^
AssertionError</pre></details> | non_defect | tests sqlalchemy dialect compliance test dialect compliance timestampmicrosecondstest bigquery bigquery test literal failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in from call result optional func file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in lambda ihook item item kwds when when reraise reraise file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy hooks py line in call return self hookexec self name self get hookimpls kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy manager py line in hookexec return self inner hookexec hook name methods kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall return outcome get result file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy result py line in get result raise ex with traceback ex file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall res hook impl function args file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pytest runtest teardown item session setupstate teardown exact item nextitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown exact self teardown towards needed collectors file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards raise exc file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards self pop and teardown file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pop and teardown self teardown with finalization colitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown with finalization assert colitem in self stack assertionerror | 0 |
19,876 | 3,268,100,513 | IssuesEvent | 2015-10-23 09:21:56 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Map Operations have uninformative toString | Team: Core Type: Defect | The IMap operations have uniformative toString implementations. The toString of the Operation is very informative and contains the call id and a lot of other information.
Probably it is best to add an additional method to Operation like toPartialString and instead of doing:
```
@Override
public String toString() {
return "BasePutOperation{" + name + "}";
}
```
You do e.g.
```
@Override
public String toString() {
return "BasePutOperation{name=" + name+", "+ toPartialString()+ "}";
}
```
This way it is easy to add additional field information to a toString.
If we do not fix these toString implementation, it could very well be that we can't analyze logging because a lot of relevant information has been dropped. | 1.0 | Map Operations have uninformative toString - The IMap operations have uniformative toString implementations. The toString of the Operation is very informative and contains the call id and a lot of other information.
Probably it is best to add an additional method to Operation like toPartialString and instead of doing:
```
@Override
public String toString() {
return "BasePutOperation{" + name + "}";
}
```
You do e.g.
```
@Override
public String toString() {
return "BasePutOperation{name=" + name+", "+ toPartialString()+ "}";
}
```
This way it is easy to add additional field information to a toString.
If we do not fix these toString implementation, it could very well be that we can't analyze logging because a lot of relevant information has been dropped. | defect | map operations have uninformative tostring the imap operations have uniformative tostring implementations the tostring of the operation is very informative and contains the call id and a lot of other information probably it is best to add an additional method to operation like topartialstring and instead of doing override public string tostring return baseputoperation name you do e g override public string tostring return baseputoperation name name topartialstring this way it is easy to add additional field information to a tostring if we do not fix these tostring implementation it could very well be that we can t analyze logging because a lot of relevant information has been dropped | 1 |
36,604 | 8,031,892,353 | IssuesEvent | 2018-07-28 08:11:16 | OpenMS/OpenMS | https://api.github.com/repos/OpenMS/OpenMS | closed | preferences in TOPPView broken | TOPPView defect minor | Global preferences such as `mapping of mz` are not stored and lost when re-starting `TOPPView`.

| 1.0 | preferences in TOPPView broken - Global preferences such as `mapping of mz` are not stored and lost when re-starting `TOPPView`.

| defect | preferences in toppview broken global preferences such as mapping of mz are not stored and lost when re starting toppview | 1 |
52,930 | 13,225,239,530 | IssuesEvent | 2020-08-17 20:46:30 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | cmake should reject project names with dashes (Trac #595) | Migrated from Trac cmake defect | this is because dashes make pybindings confusing, e.g.
phys-services vs phys_services, since dashes aren't legal in python
this will require a 'grandfathered-in' list
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/595">https://code.icecube.wisc.edu/projects/icecube/ticket/595</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-11T22:40:42",
"_ts": "1305153642000000",
"description": "this is because dashes make pybindings confusing, e.g.\nphys-services vs phys_services, since dashes aren't legal in python\n\nthis will require a 'grandfathered-in' list\n\n \n ",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2010-02-06T21:45:51",
"component": "cmake",
"summary": "cmake should reject project names with dashes",
"priority": "minor",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | cmake should reject project names with dashes (Trac #595) - this is because dashes make pybindings confusing, e.g.
phys-services vs phys_services, since dashes aren't legal in python
this will require a 'grandfathered-in' list
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/595">https://code.icecube.wisc.edu/projects/icecube/ticket/595</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-11T22:40:42",
"_ts": "1305153642000000",
"description": "this is because dashes make pybindings confusing, e.g.\nphys-services vs phys_services, since dashes aren't legal in python\n\nthis will require a 'grandfathered-in' list\n\n \n ",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"time": "2010-02-06T21:45:51",
"component": "cmake",
"summary": "cmake should reject project names with dashes",
"priority": "minor",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | cmake should reject project names with dashes trac this is because dashes make pybindings confusing e g phys services vs phys services since dashes aren t legal in python this will require a grandfathered in list migrated from json status closed changetime ts description this is because dashes make pybindings confusing e g nphys services vs phys services since dashes aren t legal in python n nthis will require a grandfathered in list n n n reporter troy cc resolution fixed time component cmake summary cmake should reject project names with dashes priority minor keywords milestone owner troy type defect | 1 |
64,142 | 18,243,838,436 | IssuesEvent | 2021-10-01 15:48:00 | SUI-Components/sui-components | https://api.github.com/repos/SUI-Components/sui-components | opened | Thumbnail - The width of the Caption is too wide | defect | Component [Thumbnail](https://sui-components.vercel.app/workbench/molecule/thumbnail/demo)
**Is your iteration request related to a problem?**
The width of the caption is too wide (100% + padding)

**Describe the solution you'd like**
The correct width should not be wider than the image

| 1.0 | Thumbnail - The width of the Caption is too wide - Component [Thumbnail](https://sui-components.vercel.app/workbench/molecule/thumbnail/demo)
**Is your iteration request related to a problem?**
The width of the caption is too wide (100% + padding)

**Describe the solution you'd like**
The correct width should not be wider than the image

| defect | thumbnail the width of the caption is too wide component is your iteration request related to a problem the width of the caption is too wide padding describe the solution you d like the correct width should not be wider than the image | 1 |
11,296 | 2,648,923,053 | IssuesEvent | 2015-03-14 12:01:34 | jancona/android-on-freerunner | https://api.github.com/repos/jancona/android-on-freerunner | closed | {{{{LIvE%%%%*!!!*world~*~golf}}}championships-accenture match play championship Live Streaming online pga tournament golf | auto-migrated invalid Priority-Medium Type-Defect | ```
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
Hello friend, I know you are a lover of golf & you are looking a tv who
telecast world golf championships-accenture match play championship-2013 Golf
live streaming. Please visit this streaming golf link and enjoy this pga
tournament.Watch live 2013 Golf stream online satellite TV, cable TV to watch
the pay per view online live and in HD on the PC over the Internet. Everyone
watch the live world golf championships-accenture match play championship-2013
stream Golf matches at Medinah Country Club and other sporting events live has
so easy. When you talk about a major sporting event, this World Golf world golf
championships-accenture match play championship-2013 game is one of the best.
Get instant access to the widest possible coverage of matches on the web
directly to your desktop from anywhere. maximum cheap rate From other Tv
channel. And also promised that if you faced any problem we will abide by to
refund your payments. So don’t hesitated use this tv
Click to watch world golf championships-accenture match play championship 2013
Golf liveOnline
undefined
Feb 20 - Feb 24 , 2013
World Golf Championships-Accenture Match Play Championship
Ritz-Carlton Golf Club, Marana, AZ
Click to watch world golf championships-accenture match play championship 2013
Golf liveOnline
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
watch world golf championships-accenture match play championship 2013
Watch world golf championships-accenture match play championship Golf Online,
Watch world golf
championships-accenture match play championship Golf Live, world golf
championships-accenture
match play championship Golf Live, world golf championships-accenture match
play championship
Golf Stream, world golf championships-accenture match play championship Golf
Streaming, world golf championships-accenture match play championship Live
Stream, world golf championships-accenture match play championship 2013 Live
Stream,World Golf Championships Live Stream world golf championships-accenture
match play championship Golf Live, world golf championships-accenture match
play championship Golf Live Stream, world golf championships-accenture match
play championship Golf Streaming, world golf championships-accenture match play
championship Live Stream, world golf championships-accenture match play
championship 2013 Live Stream.
watch world golf championships-accenture match play championship 2013
golf 2013 live stream free watch World Golf Championships-Accenture Match Play
Championship 2013 online tv World Golf Championships-Accenture Match Play
Championship 2013 live stream free watch World Golf Championships-Accenture
Match Play Championship 2013 online tv World Golf Championships-Accenture Match
Play Championship 2013 live stream free watch World Golf
Championships-Accenture Match Play Championship 2013 online tv World Golf
Championships-Accenture Match Play Championship 2013 live stream free watch
World Golf Championships-Accenture Match Play Championship 2013 online tv World
Golf Championships-Accenture Match Play Championship 2013 live stream free
watch World Golf Championships-Accenture Match Play Championship 2013 online tv
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship 2013
online tv World Golf Championships-Accenture Match Play Championship 2013 live
stream free watch World Golf Championships-Accenture Match Play Championship
2013 online tv.
World Golf Championships-Accenture Match Play Championship 2013 Live Streaming,
World Golf Championships-Accenture Match Play Championship 2013 Live Coverage, World Golf Championships-Accenture Match Play Championship 2013 Live Telecast, World Golf Championships-Accenture Match Play Championship 2013 Radio Coverage, World Golf Championships-Accenture Match Play Championship 2013 En Vivo, World Golf Championships-Accenture Match Play Championship 2013 Free Live Stream, World Golf Championships-Accenture Match Play Championship 2013 Live Feed, World Golf Championships-Accenture Match Play Championship 2013 Live Radio. World Golf Championships-Accenture Match Play Championship Concert Live.
hd World Golf Championships-Accenture Match Play Championship 2013 live stream
free
http://101golftv.blogspot.com/http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online
hd.World Golf Championships-Accenture Match Play Championship Nomination Live
On Direct Tv.WATCH LIVE STREAM: World Golf Championships-Accenture Match Play
Championship Red Carpet 2013 Live Stream.
```
Original issue reported on code.google.com by `usalives...@gmail.com` on 21 Feb 2013 at 2:53 | 1.0 | {{{{LIvE%%%%*!!!*world~*~golf}}}championships-accenture match play championship Live Streaming online pga tournament golf - ```
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
Hello friend, I know you are a lover of golf & you are looking a tv who
telecast world golf championships-accenture match play championship-2013 Golf
live streaming. Please visit this streaming golf link and enjoy this pga
tournament.Watch live 2013 Golf stream online satellite TV, cable TV to watch
the pay per view online live and in HD on the PC over the Internet. Everyone
watch the live world golf championships-accenture match play championship-2013
stream Golf matches at Medinah Country Club and other sporting events live has
so easy. When you talk about a major sporting event, this World Golf world golf
championships-accenture match play championship-2013 game is one of the best.
Get instant access to the widest possible coverage of matches on the web
directly to your desktop from anywhere. maximum cheap rate From other Tv
channel. And also promised that if you faced any problem we will abide by to
refund your payments. So don’t hesitated use this tv
Click to watch world golf championships-accenture match play championship 2013
Golf liveOnline
undefined
Feb 20 - Feb 24 , 2013
World Golf Championships-Accenture Match Play Championship
Ritz-Carlton Golf Club, Marana, AZ
Click to watch world golf championships-accenture match play championship 2013
Golf liveOnline
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
watch world golf championships-accenture match play championship 2013
Watch world golf championships-accenture match play championship Golf Online,
Watch world golf
championships-accenture match play championship Golf Live, world golf
championships-accenture
match play championship Golf Live, world golf championships-accenture match
play championship
Golf Stream, world golf championships-accenture match play championship Golf
Streaming, world golf championships-accenture match play championship Live
Stream, world golf championships-accenture match play championship 2013 Live
Stream,World Golf Championships Live Stream world golf championships-accenture
match play championship Golf Live, world golf championships-accenture match
play championship Golf Live Stream, world golf championships-accenture match
play championship Golf Streaming, world golf championships-accenture match play
championship Live Stream, world golf championships-accenture match play
championship 2013 Live Stream.
watch world golf championships-accenture match play championship 2013
golf 2013 live stream free watch World Golf Championships-Accenture Match Play
Championship 2013 online tv World Golf Championships-Accenture Match Play
Championship 2013 live stream free watch World Golf Championships-Accenture
Match Play Championship 2013 online tv World Golf Championships-Accenture Match
Play Championship 2013 live stream free watch World Golf
Championships-Accenture Match Play Championship 2013 online tv World Golf
Championships-Accenture Match Play Championship 2013 live stream free watch
World Golf Championships-Accenture Match Play Championship 2013 online tv World
Golf Championships-Accenture Match Play Championship 2013 live stream free
watch World Golf Championships-Accenture Match Play Championship 2013 online tv
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship 2013
online tv World Golf Championships-Accenture Match Play Championship 2013 live
stream free watch World Golf Championships-Accenture Match Play Championship
2013 online tv.
World Golf Championships-Accenture Match Play Championship 2013 Live Streaming,
World Golf Championships-Accenture Match Play Championship 2013 Live Coverage, World Golf Championships-Accenture Match Play Championship 2013 Live Telecast, World Golf Championships-Accenture Match Play Championship 2013 Radio Coverage, World Golf Championships-Accenture Match Play Championship 2013 En Vivo, World Golf Championships-Accenture Match Play Championship 2013 Free Live Stream, World Golf Championships-Accenture Match Play Championship 2013 Live Feed, World Golf Championships-Accenture Match Play Championship 2013 Live Radio. World Golf Championships-Accenture Match Play Championship Concert Live.
hd World Golf Championships-Accenture Match Play Championship 2013 live stream
free
http://101golftv.blogspot.com/http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
http://101golftv.blogspot.com/
watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online hd
World Golf Championships-Accenture Match Play Championship 2013 live stream
free watch World Golf Championships-Accenture Match Play Championship online
hd.World Golf Championships-Accenture Match Play Championship Nomination Live
On Direct Tv.WATCH LIVE STREAM: World Golf Championships-Accenture Match Play
Championship Red Carpet 2013 Live Stream.
```
Original issue reported on code.google.com by `usalives...@gmail.com` on 21 Feb 2013 at 2:53 | defect | live world golf championships accenture match play championship live streaming online pga tournament golf hello friend i know you are a lover of golf you are looking a tv who telecast world golf championships accenture match play championship golf live streaming please visit this streaming golf link and enjoy this pga tournament watch live golf stream online satellite tv cable tv to watch the pay per view online live and in hd on the pc over the internet everyone watch the live world golf championships accenture match play championship stream golf matches at medinah country club and other sporting events live has so easy when you talk about a major sporting event this world golf world golf championships accenture match play championship game is one of the best get instant access to the widest possible coverage of matches on the web directly to your desktop from anywhere maximum cheap rate from other tv channel and also promised that if you faced any problem we will abide by to refund your payments so don’t hesitated use this tv click to watch world golf championships accenture match play championship golf liveonline undefined feb feb world golf championships accenture match play championship ritz carlton golf club marana az click to watch world golf championships accenture match play championship golf liveonline watch world golf championships accenture match play championship watch world golf championships accenture match play championship golf online watch world golf championships accenture match play championship golf live world golf championships accenture match play championship golf live world golf championships accenture match play championship golf stream world golf championships accenture match play championship golf streaming world golf championships accenture match play championship live stream world golf championships accenture match play championship live stream world golf championships live stream world golf championships accenture match play championship golf live world golf championships accenture match play championship golf live stream world golf championships accenture match play championship golf streaming world golf championships accenture match play championship live stream world golf championships accenture match play championship live stream watch world golf championships accenture match play championship golf live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online tv world golf championships accenture match play championship live streaming world golf championships accenture match play championship live coverage world golf championships accenture match play championship live telecast world golf championships accenture match play championship radio coverage world golf championships accenture match play championship en vivo world golf championships accenture match play championship free live stream world golf championships accenture match play championship live feed world golf championships accenture match play championship live radio world golf championships accenture match play championship concert live hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship live stream free watch world golf championships accenture match play championship online hd world golf championships accenture match play championship nomination live on direct tv watch live stream world golf championships accenture match play championship red carpet live stream original issue reported on code google com by usalives gmail com on feb at | 1 |
76,139 | 26,259,316,612 | IssuesEvent | 2023-01-06 05:42:29 | scipy/scipy | https://api.github.com/repos/scipy/scipy | opened | BUG: dijkstra algorithm is returning different results in v1.10 compared to v1.9 | defect | ### Describe your issue.
A unit test is failing in my CI but not locally, which usually means something changed in a dependency. I have tracked it down to ``scipy.sparse.csgraph.dijkstra`` returning a different path in v1.10 (the CI version) compard to v1.9 (my local version).
### Reproducing Code Example
```python
import numpy as np
import scipy.sparse as sprs
from scipy.sparse import csgraph
am = \
[[0, 1, 0, 1, 0, 0, 0, 0, 0],
[1, 0, 1, 0, 1, 0, 0, 0, 0],
[0, 1, 0, 0, 0, 1, 0, 0, 0],
[1, 0, 0, 0, 1, 0, 1, 0, 0],
[0, 1, 0, 1, 0, 1, 0, 1, 0],
[0, 0, 1, 0, 1, 0, 0, 0, 1],
[0, 0, 0, 1, 0, 0, 0, 1, 0],
[0, 0, 0, 0, 1, 0, 1, 0, 1],
[0, 0, 0, 0, 0, 1, 0, 1, 0]]
am = sprs.coo_matrix(am)
paths = csgraph.dijkstra(csgraph=am,
directed=False,
indices=0,
return_predecessors=True,
min_only=False)[1]
print(paths)
```
### Error message
```shell
On v1.10 I get this: ``[-9999 0 1 0 1 4 3 4 7]``. Here is the full version info for this one: ``1.10.0 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)``
On v1.9 I get this: ``[-9999 0 1 0 1 2 3 6 7]``. Here is the full version number for this one: ``1.9.3 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)``
```
### SciPy/NumPy/Python version information
1.10.0 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0) | 1.0 | BUG: dijkstra algorithm is returning different results in v1.10 compared to v1.9 - ### Describe your issue.
A unit test is failing in my CI but not locally, which usually means something changed in a dependency. I have tracked it down to ``scipy.sparse.csgraph.dijkstra`` returning a different path in v1.10 (the CI version) compard to v1.9 (my local version).
### Reproducing Code Example
```python
import numpy as np
import scipy.sparse as sprs
from scipy.sparse import csgraph
am = \
[[0, 1, 0, 1, 0, 0, 0, 0, 0],
[1, 0, 1, 0, 1, 0, 0, 0, 0],
[0, 1, 0, 0, 0, 1, 0, 0, 0],
[1, 0, 0, 0, 1, 0, 1, 0, 0],
[0, 1, 0, 1, 0, 1, 0, 1, 0],
[0, 0, 1, 0, 1, 0, 0, 0, 1],
[0, 0, 0, 1, 0, 0, 0, 1, 0],
[0, 0, 0, 0, 1, 0, 1, 0, 1],
[0, 0, 0, 0, 0, 1, 0, 1, 0]]
am = sprs.coo_matrix(am)
paths = csgraph.dijkstra(csgraph=am,
directed=False,
indices=0,
return_predecessors=True,
min_only=False)[1]
print(paths)
```
### Error message
```shell
On v1.10 I get this: ``[-9999 0 1 0 1 4 3 4 7]``. Here is the full version info for this one: ``1.10.0 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)``
On v1.9 I get this: ``[-9999 0 1 0 1 2 3 6 7]``. Here is the full version number for this one: ``1.9.3 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)``
```
### SciPy/NumPy/Python version information
1.10.0 1.23.5 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0) | defect | bug dijkstra algorithm is returning different results in compared to describe your issue a unit test is failing in my ci but not locally which usually means something changed in a dependency i have tracked it down to scipy sparse csgraph dijkstra returning a different path in the ci version compard to my local version reproducing code example python import numpy as np import scipy sparse as sprs from scipy sparse import csgraph am am sprs coo matrix am paths csgraph dijkstra csgraph am directed false indices return predecessors true min only false print paths error message shell on i get this here is the full version info for this one sys version info major minor micro releaselevel final serial on i get this here is the full version number for this one sys version info major minor micro releaselevel final serial scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
369,278 | 10,894,978,098 | IssuesEvent | 2019-11-19 09:46:30 | OpenSRP/opensrp-client-chw-anc | https://api.github.com/repos/OpenSRP/opensrp-client-chw-anc | closed | Update the info icon for LMP for ANC Registration in BA in Swahili | BA-specific Medium priority | - [ ] Update the info icon for the `lmp` in the ANC Registration to look like this:
LNMP = Siku ya kwanza ya mwisho ya hedhi (LNMP). Kama tarehe halisi haijulikani, lakini kipindi cha mwezi kinajulikana;
- tumia tarehe 5 kwa mwanzo wa mwezi
- tumia tarehe 15 kwa katikati ya mwezi
- tumia tarehe 25 kwa mwisho wa mwezi.
Each bullet should be a new line for easier reading | 1.0 | Update the info icon for LMP for ANC Registration in BA in Swahili - - [ ] Update the info icon for the `lmp` in the ANC Registration to look like this:
LNMP = Siku ya kwanza ya mwisho ya hedhi (LNMP). Kama tarehe halisi haijulikani, lakini kipindi cha mwezi kinajulikana;
- tumia tarehe 5 kwa mwanzo wa mwezi
- tumia tarehe 15 kwa katikati ya mwezi
- tumia tarehe 25 kwa mwisho wa mwezi.
Each bullet should be a new line for easier reading | non_defect | update the info icon for lmp for anc registration in ba in swahili update the info icon for the lmp in the anc registration to look like this lnmp siku ya kwanza ya mwisho ya hedhi lnmp kama tarehe halisi haijulikani lakini kipindi cha mwezi kinajulikana tumia tarehe kwa mwanzo wa mwezi tumia tarehe kwa katikati ya mwezi tumia tarehe kwa mwisho wa mwezi each bullet should be a new line for easier reading | 0 |
75,565 | 7,477,395,846 | IssuesEvent | 2018-04-04 08:11:26 | LambdaFactory/Neptune-public | https://api.github.com/repos/LambdaFactory/Neptune-public | closed | Extraneous gutter icon with certain Expecto helpers | Expecto Test detection bug | Test cases which contain certain "test helpers" (e.g. `skipTest`, `failTest`) display an extra gutter icon
_Note: this might not be a bug, per se... but it's a bit confusing when "scanning" a file_
+ Work-around: none
+ Platform: Windows 10
+ Runtime: .NET Core 2.1
Minimal reproduction: https://github.com/pblasucci/neptune_expecto_repro | 1.0 | Extraneous gutter icon with certain Expecto helpers - Test cases which contain certain "test helpers" (e.g. `skipTest`, `failTest`) display an extra gutter icon
_Note: this might not be a bug, per se... but it's a bit confusing when "scanning" a file_
+ Work-around: none
+ Platform: Windows 10
+ Runtime: .NET Core 2.1
Minimal reproduction: https://github.com/pblasucci/neptune_expecto_repro | non_defect | extraneous gutter icon with certain expecto helpers test cases which contain certain test helpers e g skiptest failtest display an extra gutter icon note this might not be a bug per se but it s a bit confusing when scanning a file work around none platform windows runtime net core minimal reproduction | 0 |
20,759 | 3,411,765,986 | IssuesEvent | 2015-12-05 10:04:26 | Cockatrice/Cockatrice | https://api.github.com/repos/Cockatrice/Cockatrice | closed | An internal error has occurred, please try closing and reopening your client | App - Cockatrice Defect - Crash OS - Windows Troubleshooting | On trying to connect to any server, I get an error:
"An internal error has occurred, please try closing and reopening your client and try again. If the error persists try updating your client to the most recent build and if need be contact your software provider."
I'm using Windows Vista, Cockatrice is the client from http://www.woogerworks.com the XP/Vista version. | 1.0 | An internal error has occurred, please try closing and reopening your client - On trying to connect to any server, I get an error:
"An internal error has occurred, please try closing and reopening your client and try again. If the error persists try updating your client to the most recent build and if need be contact your software provider."
I'm using Windows Vista, Cockatrice is the client from http://www.woogerworks.com the XP/Vista version. | defect | an internal error has occurred please try closing and reopening your client on trying to connect to any server i get an error an internal error has occurred please try closing and reopening your client and try again if the error persists try updating your client to the most recent build and if need be contact your software provider i m using windows vista cockatrice is the client from the xp vista version | 1 |
46,889 | 19,539,589,212 | IssuesEvent | 2021-12-31 16:58:50 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | opened | Delete "csr_flex_notes" from AMD Data Tracker - Feb 2022 | Type: Data Service: Apps Product: AMD Data Tracker | Received automated reminder to delete flex notes in AMD Data Tracker PROD and reference the Flex Note Archiving wiki
Ran filter on: https://builder.knack.com/atd/amd/records/objects/object_95
Filter:
`ISSUE_STATUS_CODE` contains `closed`
`CREATED_DATE` is `before` 12/01/21
Total Records: {value}
To Delete: {value}
New Total: {value}
| 1.0 | Delete "csr_flex_notes" from AMD Data Tracker - Feb 2022 - Received automated reminder to delete flex notes in AMD Data Tracker PROD and reference the Flex Note Archiving wiki
Ran filter on: https://builder.knack.com/atd/amd/records/objects/object_95
Filter:
`ISSUE_STATUS_CODE` contains `closed`
`CREATED_DATE` is `before` 12/01/21
Total Records: {value}
To Delete: {value}
New Total: {value}
| non_defect | delete csr flex notes from amd data tracker feb received automated reminder to delete flex notes in amd data tracker prod and reference the flex note archiving wiki ran filter on filter issue status code contains closed created date is before total records value to delete value new total value | 0 |
51,652 | 13,207,559,310 | IssuesEvent | 2020-08-14 23:35:28 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | cmake - deal with cmake policy CMP0042 (Trac #792) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/792">https://code.icecube.wisc.edu/projects/icecube/ticket/792</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-23T18:59:11",
"_ts": "1414090751712739",
"description": "{{{\nCMP0042\n-------\n\n``MACOSX_RPATH`` is enabled by default.\n\nCMake 2.8.12 and newer has support for using ``@rpath`` in a target's install\nname. This was enabled by setting the target property\n``MACOSX_RPATH``. The ``@rpath`` in an install name is a more\nflexible and powerful mechanism than ``@executable_path`` or ``@loader_path``\nfor locating shared libraries.\n\nCMake 3.0 and later prefer this property to be ON by default. Projects\nwanting ``@rpath`` in a target's install name may remove any setting of\nthe ``INSTALL_NAME_DIR`` and ``CMAKE_INSTALL_NAME_DIR``\nvariables.\n\nThis policy was introduced in CMake version 3.0. CMake version\n3.0.0 warns when the policy is not set and uses OLD behavior. Use\nthe cmake_policy command to set it to OLD or NEW explicitly.\n}}}\n\nI think the best option for us is to set this to OLD, until we can guarantee wide spread adoption of newer cmake.\n\nAlso see #773\n\n",
"reporter": "nega",
"cc": "meike.dewith, olivas",
"resolution": "fixed",
"time": "2014-10-23T14:46:30",
"component": "cmake",
"summary": "cmake - deal with cmake policy CMP0042",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | cmake - deal with cmake policy CMP0042 (Trac #792) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/792">https://code.icecube.wisc.edu/projects/icecube/ticket/792</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-23T18:59:11",
"_ts": "1414090751712739",
"description": "{{{\nCMP0042\n-------\n\n``MACOSX_RPATH`` is enabled by default.\n\nCMake 2.8.12 and newer has support for using ``@rpath`` in a target's install\nname. This was enabled by setting the target property\n``MACOSX_RPATH``. The ``@rpath`` in an install name is a more\nflexible and powerful mechanism than ``@executable_path`` or ``@loader_path``\nfor locating shared libraries.\n\nCMake 3.0 and later prefer this property to be ON by default. Projects\nwanting ``@rpath`` in a target's install name may remove any setting of\nthe ``INSTALL_NAME_DIR`` and ``CMAKE_INSTALL_NAME_DIR``\nvariables.\n\nThis policy was introduced in CMake version 3.0. CMake version\n3.0.0 warns when the policy is not set and uses OLD behavior. Use\nthe cmake_policy command to set it to OLD or NEW explicitly.\n}}}\n\nI think the best option for us is to set this to OLD, until we can guarantee wide spread adoption of newer cmake.\n\nAlso see #773\n\n",
"reporter": "nega",
"cc": "meike.dewith, olivas",
"resolution": "fixed",
"time": "2014-10-23T14:46:30",
"component": "cmake",
"summary": "cmake - deal with cmake policy CMP0042",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | cmake deal with cmake policy trac migrated from json status closed changetime ts description n n n macosx rpath is enabled by default n ncmake and newer has support for using rpath in a target s install nname this was enabled by setting the target property n macosx rpath the rpath in an install name is a more nflexible and powerful mechanism than executable path or loader path nfor locating shared libraries n ncmake and later prefer this property to be on by default projects nwanting rpath in a target s install name may remove any setting of nthe install name dir and cmake install name dir nvariables n nthis policy was introduced in cmake version cmake version warns when the policy is not set and uses old behavior use nthe cmake policy command to set it to old or new explicitly n n ni think the best option for us is to set this to old until we can guarantee wide spread adoption of newer cmake n nalso see n n reporter nega cc meike dewith olivas resolution fixed time component cmake summary cmake deal with cmake policy priority normal keywords milestone owner nega type defect | 1 |
46,262 | 9,917,383,640 | IssuesEvent | 2019-06-29 00:04:31 | Grammarsalad/Casting_Attributes | https://api.github.com/repos/Grammarsalad/Casting_Attributes | opened | bg2fp_effect_batches_spell_saves | Code in process | ```
///// \\\\\
///// arrays for batch effects \\\\\
///// \\\\\
/*
The cd_apply_batch function, defined below, requires a macro name. That macro needs to define
three arrays for the function:
* cd_immunity_batches_key
If any effect in this array is present on the item/creature/spell being patches, it will proceed with
adding extra effects and deleting effects, as defined by the next two arrays. If you have mutiple
effects defined here, it will match any of them and add any missing.
* cd_immunity_batches_extras
This array contains all of the subsiduary effects that should accompany the effects in the key area.
* cd_immunity_batches_deletes
If a key is found, any effect matching this array will be deleted.
If a key is found, the function will then proceed back through the file being patched and add--if they
are not present--all of the effects not found in the _key and _extras arrays. It will also, only with
a key effect matched, delete any effects listed in the _deletes array.
Note that since it's a macro, you can conditionally define your arrays. The free action batch will
remove stun immunity unless a particular component from the Fixpack is installed, for example. Remember
that's it's being run in a patch context, so you'll need to use PATCH actions instead of ACTIONs.
An example use--the cd_full_dispel_arrays uses a dispel magic opcode (58) as the lone key element. If
found, the function will delete a variety of old item removal opcodes defined in the _deletes array,
and add (if not already present) the series of effects that cure feeblemind and deafness.
In general, try to avoid explicit spell blockages with 206 opcodes. If you can block the relevant
effects with other immunities, do so, as it's a more robust and extensible solution. In some cases
they are needed --e.g. the slow immunity batch blocks slow spells outright, as most come paired
with AC and THAC0 penalties which can't be blocked via opcode without serious collateral damage.
* Format of all arrays are opcode, parameter1, parameter2, resref, timing, duration
* Timing and duration are only used for new fx writes
* Use -1 in a field (or "same" in string fields) if the match value for that field doesn't matter.
* Always map the effect to zero.
*/
//dex save bonus for damaging evoc:
DEFINE_PATCH_MACRO ~b_dex_evoc_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
12, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//will save bonus for vs. mind affecting:
//sleep
DEFINE_PATCH_MACRO ~b_wis_sleep_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
2, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//unconsciousness
DEFINE_PATCH_MACRO ~b_wis_uncon_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
39, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Stun
DEFINE_PATCH_MACRO ~b_wis_stun_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
45, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Hold1
DEFINE_PATCH_MACRO ~b_wis_hold1_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
109, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Hold1
DEFINE_PATCH_MACRO ~b_wis_hold2_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
175, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
///// \\\\\
///// the batch... err... patch \\\\\
///// \\\\\
// batch effects
DEFINE_PATCH_FUNCTION ~cd_apply_batch~
INT_VAR debug = 0 // 1 - spit out C_E list, 2 same w/ additional info, 3 lists effects as they're added
force_cosmetic = "-1"
bonus_var_1 = "-1" // bonus vars for use in setting arrays
bonus_var_2 = "-1"
bonus_var_3 = "-1"
STR_VAR array_name = "same"
all_or = "or"
BEGIN
SET debug_message = 0
// formerly the immunity post array
READ_ASCII 0x00 type (3)
SET min_size = 0
PATCH_IF ("%type%" STRING_COMPARE_CASE "spl" = 0) BEGIN
READ_LONG 0x64 abil_off ELSE 0
READ_SHORT 0x68 abil_num ELSE 0
READ_LONG 0x6a fx_off ELSE 0
SET counter_offset = 0x70
SET abil_length = 0x28
SET global_loop = 0
SET fx_type = 0
SET min_size = 0x72
SET cosmetic = 1
END ELSE
PATCH_IF ("%type%" STRING_COMPARE_CASE "itm" = 0) BEGIN
READ_LONG 0x64 abil_off ELSE 0
READ_SHORT 0x68 abil_num ELSE 0
READ_LONG 0x6a fx_off ELSE 0
SET counter_offset = 0x70
SET abil_length = 0x38
SET global_loop = 1
SET fx_type = 0
SET min_size = 0x72
READ_BYTE 0x18 flags ELSE 0
PATCH_IF ((flags BAND BIT2) = BIT2) BEGIN // if droppable, add portrait icons and whatnot
SET cosmetic = 1
END ELSE BEGIN
SET cosmetic = 0
END
END ELSE
PATCH_IF ("%type%" STRING_COMPARE_CASE "cre" = 0) BEGIN
READ_LONG 0x2c4 fx_off ELSE 0
READ_BYTE 0x33 fx_type ELSE 2
SET abil_off = 0
SET abil_num = 0
SET counter_offset = 0x2c8
SET abil_length = 0
SET global_loop = 1
SET min_size = 0x2d4
SET cosmetic = 0
END ELSE BEGIN
PATCH_PRINT ~Warning: %macro_name% macro halting; file type not recognized (%type%)~
END
// if user overrides normal cosmetic stuff with function var
PATCH_IF force_cosmetic = 1 BEGIN
SET cosmetic = 1
END ELSE
PATCH_IF force_cosmetic = 0 BEGIN
SET cosmetic = 0
END
PATCH_IF ((SOURCE_SIZE >= min_size) AND (min_size != 0)) BEGIN // min_size must get set by file type detection
SET new_fx = 0
FOR (index = (0 - global_loop) ; index < abil_num ; ++index) BEGIN
SPRINT template "" // replaced if key match found
// always clear your arrays
CLEAR_ARRAY cd_immunity_batches_key
CLEAR_ARRAY cd_immunity_batches_extras
CLEAR_ARRAY cd_immunity_batches_deletes
LAUNCH_PATCH_MACRO ~%array_name%~ // loads up arrays with effects to match
PATCH_IF (index < 0) BEGIN // if loop through globals needed
SET abil_fx_idx = 0
END ELSE BEGIN // otherwise normal ability
SET counter_offset = (abil_off + 0x1e + (abil_length * index))
READ_SHORT (abil_off + 0x20 + (abil_length * index)) abil_fx_idx
SET abil_fx_idx += new_fx
WRITE_SHORT (abil_off + 0x20 + (abil_length * index)) (abil_fx_idx)
END
READ_SHORT counter_offset counter // fx_num on global loop, otherwise abil_fx_num
// run one pass purely looking for keys
FOR (index2 = 0 ; index2 < counter ; ++index2) BEGIN
READ_SHORT (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) opcode
READ_LONG (fx_off + 0x04 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param1
READ_LONG (fx_off + 0x08 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param2
READ_ASCII (fx_off + 0x14 + (0x14 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) resref
// look through key array for matches
PHP_EACH cd_immunity_batches_key AS matches => value BEGIN
PATCH_IF ((value != 0) AND
((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
SET value -= 1
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN "%matches_0%", "%matches_1%", "%matches_2%", "%matches_3%", "%matches_4%", "%matches_5%", "%matches_6%" => "%value%" END // update array to indicate match present
READ_ASCII (fx_off + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) template ((0x30 + (0xd8 * fx_type))) // read whole effect into a template
SET insert_point = index2
END
END
END
PATCH_IF ("%all_or%" STRING_COMPARE_CASE "all" = 0) BEGIN // if an all clause, clear template if any is not matched
PHP_EACH cd_immunity_batches_key AS matches => value BEGIN
PATCH_IF value != 0 BEGIN
SPRINT template "" // clear template, prevents advancement
END
END
END
// if we have a key match, run second pass looking for extras and deleting targeted effects
PATCH_IF ("%template%" STRING_COMPARE_CASE "") BEGIN // if key found, template is set and we look for associated/delete effects
SET last_adj = 0
FOR (index2 = 0 ; index2 < counter ; ++index2) BEGIN
READ_SHORT (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) opcode
READ_LONG (fx_off + 0x04 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param1
READ_LONG (fx_off + 0x08 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param2
READ_ASCII (fx_off + 0x14 + (0x14 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) resref
PATCH_IF index2 = (counter - 1) BEGIN // on final loop
PATCH_IF ((opcode = 206) AND ("%SOURCE_RES%" STRING_COMPARE_CASE "%resref%" = 0)) BEGIN // special check to make sure last effect isn't immunity to self
SET last_adj = "-1" // if it is, "last" effects need to go one effect before it
END
END
// look through associated array for matches
PHP_EACH cd_immunity_batches_extras AS matches => value BEGIN
PATCH_IF ((value != 0) AND
((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
SET value -= 1
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN "%matches_0%", "%matches_1%", "%matches_2%", "%matches_3%", "%matches_4%", "%matches_5%", "%matches_6%" => "%value%" END // update array to indicate match present
END
END
// look through delete array
PHP_EACH cd_immunity_batches_delete AS matches => value BEGIN
PATCH_IF (((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
DELETE_BYTES (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) ((0x30 + (0xd8 * fx_type))) // delete effect
PATCH_IF insert_point > index2 BEGIN SET insert_point -= 1 END // move insert point up if we're deleting an effect in front of it
SET new_fx -= 1
SET counter -= 1
SET index2 -= 1
SET debug_message = 1 // something's changed!
END
END
END // end second effects pass
// with a completed second pass, start adding needed effects
PATCH_FOR_EACH array IN cd_immunity_batches_key cd_immunity_batches_extras BEGIN
PHP_EACH ~%array%~ AS matches => value BEGIN
PATCH_IF debug > 2 BEGIN PATCH_PRINT ~%array% array: %matches_0%, %matches_1%, %matches_2%, %matches_3%, %matches_4%, %matches_5%, %matches_6% => %value%~ END
WHILE (value > 0) BEGIN
SET base = (fx_off + ((abil_fx_idx + insert_point) * (0x30 + (0xd8 * fx_type))))
PATCH_IF ("%matches_6%" STRING_COMPARE_CASE "first" = 0) BEGIN SET base = (fx_off + ((abil_fx_idx ) * (0x30 + (0xd8 * fx_type)))) END
PATCH_IF ("%matches_6%" STRING_COMPARE_CASE "last" = 0) BEGIN SET base = (fx_off + ((abil_fx_idx + counter + last_adj) * (0x30 + (0xd8 * fx_type)))) END
INSERT_BYTES (base) (0x30 + (0xd8 * fx_type))
WRITE_ASCIIE (base ) "%template%" // clones immunity effect
PATCH_IF (matches_0 >= "-9") BEGIN WRITE_SHORT (base + (0x08 * fx_type)) matches_0 END // opcode
PATCH_IF (matches_1 >= "-9") BEGIN WRITE_LONG (base + 0x04 + (0x10 * fx_type)) matches_1 END // parameter1
PATCH_IF (matches_2 >= "-9") BEGIN WRITE_LONG (base + 0x08 + (0x10 * fx_type)) matches_2 END // parameter2
PATCH_IF ("%matches_3%" STRING_COMPARE_CASE "same") BEGIN WRITE_ASCIIE (base + 0x14 + (0x14 * fx_type)) ~%matches_3%~ #8 END // resref
PATCH_IF (matches_4 >= "-9") BEGIN WRITE_BYTE (base + 0x0c + (0x10 * fx_type)) matches_4 END // timing
PATCH_IF (matches_5 >= "-9") BEGIN WRITE_BYTE (base + 0x0e + (0x13 * fx_type)) matches_5 END // duration
SET new_fx += 1
SET counter += 1
SET debug_message = 1 // something's changed!
SET value -= 1
END
END
END
END
WRITE_SHORT counter_offset counter
END
// formerly the immunity post array
PATCH_IF (("%type%" STRING_COMPARE_CASE "cre" = 0) AND (new_fx > 0)) BEGIN // fix offsets for cre files if fx inserted
PATCH_FOR_EACH offset IN 0x2a0 0x2a8 0x2b0 0x2b8 0x2bc BEGIN
READ_LONG offset curr_off
PATCH_IF (fx_off < curr_off) BEGIN
WRITE_LONG offset (THIS + ((0x30 + (0xd8 * fx_type)) * new_fx))
END
END
END
// debug message
PATCH_IF debug > 0 BEGIN
READ_LONG 0x0c name
PATCH_IF (name > 0) AND (name < 999999) BEGIN
READ_STRREF 0x0c name
END ELSE BEGIN
READ_STRREF 0x08 name
END
PATCH_IF (debug_message != 0) BEGIN
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^........$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^.......$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^......$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^.....$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^....$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^...$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^..$" = 0) BEGIN SPRINT spc " " END ELSE
BEGIN SPRINT spc " " END
PATCH_PRINT " ~%SOURCE_FILE%~%spc%~override~ // %name%, %new_fx% new effects from %array_name%"
END ELSE BEGIN
PATCH_IF (debug > 1) BEGIN
PATCH_PRINT " // %SOURCE_FILE% had no new effects from %array_name%"
END
END
END
END // end file size check
END
``` | 1.0 | bg2fp_effect_batches_spell_saves - ```
///// \\\\\
///// arrays for batch effects \\\\\
///// \\\\\
/*
The cd_apply_batch function, defined below, requires a macro name. That macro needs to define
three arrays for the function:
* cd_immunity_batches_key
If any effect in this array is present on the item/creature/spell being patches, it will proceed with
adding extra effects and deleting effects, as defined by the next two arrays. If you have mutiple
effects defined here, it will match any of them and add any missing.
* cd_immunity_batches_extras
This array contains all of the subsiduary effects that should accompany the effects in the key area.
* cd_immunity_batches_deletes
If a key is found, any effect matching this array will be deleted.
If a key is found, the function will then proceed back through the file being patched and add--if they
are not present--all of the effects not found in the _key and _extras arrays. It will also, only with
a key effect matched, delete any effects listed in the _deletes array.
Note that since it's a macro, you can conditionally define your arrays. The free action batch will
remove stun immunity unless a particular component from the Fixpack is installed, for example. Remember
that's it's being run in a patch context, so you'll need to use PATCH actions instead of ACTIONs.
An example use--the cd_full_dispel_arrays uses a dispel magic opcode (58) as the lone key element. If
found, the function will delete a variety of old item removal opcodes defined in the _deletes array,
and add (if not already present) the series of effects that cure feeblemind and deafness.
In general, try to avoid explicit spell blockages with 206 opcodes. If you can block the relevant
effects with other immunities, do so, as it's a more robust and extensible solution. In some cases
they are needed --e.g. the slow immunity batch blocks slow spells outright, as most come paired
with AC and THAC0 penalties which can't be blocked via opcode without serious collateral damage.
* Format of all arrays are opcode, parameter1, parameter2, resref, timing, duration
* Timing and duration are only used for new fx writes
* Use -1 in a field (or "same" in string fields) if the match value for that field doesn't matter.
* Always map the effect to zero.
*/
//dex save bonus for damaging evoc:
DEFINE_PATCH_MACRO ~b_dex_evoc_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
12, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//will save bonus for vs. mind affecting:
//sleep
DEFINE_PATCH_MACRO ~b_wis_sleep_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
2, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//unconsciousness
DEFINE_PATCH_MACRO ~b_wis_uncon_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
39, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Stun
DEFINE_PATCH_MACRO ~b_wis_stun_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
45, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Hold1
DEFINE_PATCH_MACRO ~b_wis_hold1_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
109, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
//Hold1
DEFINE_PATCH_MACRO ~b_wis_hold2_arrays_sr~ BEGIN
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN
175, "-10", "-10", "same", "-10", "-10", "same" => 1 // Some direct damage opcode (don't matter which)
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN //add 326 effects
318, 9, "-10", "ls_9", "-10", "-10", "first" => 1 // 318 less than 9 immune to base spell
318, 14, "-10", "gr_14", "-10", "-10", "first" => 1 // 318 greater than 14 immune to base spell
326, 6, "-10", "ls_6", "-10", "-10", "first" => 1 // 326 less than 6 -2 or auto fail
326, 6, "-10", "eq_6", "-10", "-10", "first" => 1 // 326 eq 6 -1
326, 7, "-10", "eq_7", "-10", "-10", "first" => 1 // 326 eq 7 -1
326, 8, "-10", "eq_8", "-10", "-10", "first" => 1 // 326 eq 8 -1
326, 19, "-10", "gr_19", "-10", "-10", "first" => 1 // 326 gr than 19 +5
326, 19, "-10", "eq_19", "-10", "-10", "first" => 1 // 326 = 19 +4
326, 18, "-10", "eq_18", "-10", "-10", "first" => 1 // 326 = 18 +4
326, 17, "-10", "eq_17", "-10", "-10", "first" => 1 // 326 is 17 +3
326, 16, "-10", "eq_16", "-10", "-10", "first" => 1 // 326 is 16 +2
326, 15, "-10", "eq_15", "-10", "-10", "first" => 1 // 326 is 15 +1
END
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_delete BEGIN //remove everything from base spell
END
END
///// \\\\\
///// the batch... err... patch \\\\\
///// \\\\\
// batch effects
DEFINE_PATCH_FUNCTION ~cd_apply_batch~
INT_VAR debug = 0 // 1 - spit out C_E list, 2 same w/ additional info, 3 lists effects as they're added
force_cosmetic = "-1"
bonus_var_1 = "-1" // bonus vars for use in setting arrays
bonus_var_2 = "-1"
bonus_var_3 = "-1"
STR_VAR array_name = "same"
all_or = "or"
BEGIN
SET debug_message = 0
// formerly the immunity post array
READ_ASCII 0x00 type (3)
SET min_size = 0
PATCH_IF ("%type%" STRING_COMPARE_CASE "spl" = 0) BEGIN
READ_LONG 0x64 abil_off ELSE 0
READ_SHORT 0x68 abil_num ELSE 0
READ_LONG 0x6a fx_off ELSE 0
SET counter_offset = 0x70
SET abil_length = 0x28
SET global_loop = 0
SET fx_type = 0
SET min_size = 0x72
SET cosmetic = 1
END ELSE
PATCH_IF ("%type%" STRING_COMPARE_CASE "itm" = 0) BEGIN
READ_LONG 0x64 abil_off ELSE 0
READ_SHORT 0x68 abil_num ELSE 0
READ_LONG 0x6a fx_off ELSE 0
SET counter_offset = 0x70
SET abil_length = 0x38
SET global_loop = 1
SET fx_type = 0
SET min_size = 0x72
READ_BYTE 0x18 flags ELSE 0
PATCH_IF ((flags BAND BIT2) = BIT2) BEGIN // if droppable, add portrait icons and whatnot
SET cosmetic = 1
END ELSE BEGIN
SET cosmetic = 0
END
END ELSE
PATCH_IF ("%type%" STRING_COMPARE_CASE "cre" = 0) BEGIN
READ_LONG 0x2c4 fx_off ELSE 0
READ_BYTE 0x33 fx_type ELSE 2
SET abil_off = 0
SET abil_num = 0
SET counter_offset = 0x2c8
SET abil_length = 0
SET global_loop = 1
SET min_size = 0x2d4
SET cosmetic = 0
END ELSE BEGIN
PATCH_PRINT ~Warning: %macro_name% macro halting; file type not recognized (%type%)~
END
// if user overrides normal cosmetic stuff with function var
PATCH_IF force_cosmetic = 1 BEGIN
SET cosmetic = 1
END ELSE
PATCH_IF force_cosmetic = 0 BEGIN
SET cosmetic = 0
END
PATCH_IF ((SOURCE_SIZE >= min_size) AND (min_size != 0)) BEGIN // min_size must get set by file type detection
SET new_fx = 0
FOR (index = (0 - global_loop) ; index < abil_num ; ++index) BEGIN
SPRINT template "" // replaced if key match found
// always clear your arrays
CLEAR_ARRAY cd_immunity_batches_key
CLEAR_ARRAY cd_immunity_batches_extras
CLEAR_ARRAY cd_immunity_batches_deletes
LAUNCH_PATCH_MACRO ~%array_name%~ // loads up arrays with effects to match
PATCH_IF (index < 0) BEGIN // if loop through globals needed
SET abil_fx_idx = 0
END ELSE BEGIN // otherwise normal ability
SET counter_offset = (abil_off + 0x1e + (abil_length * index))
READ_SHORT (abil_off + 0x20 + (abil_length * index)) abil_fx_idx
SET abil_fx_idx += new_fx
WRITE_SHORT (abil_off + 0x20 + (abil_length * index)) (abil_fx_idx)
END
READ_SHORT counter_offset counter // fx_num on global loop, otherwise abil_fx_num
// run one pass purely looking for keys
FOR (index2 = 0 ; index2 < counter ; ++index2) BEGIN
READ_SHORT (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) opcode
READ_LONG (fx_off + 0x04 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param1
READ_LONG (fx_off + 0x08 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param2
READ_ASCII (fx_off + 0x14 + (0x14 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) resref
// look through key array for matches
PHP_EACH cd_immunity_batches_key AS matches => value BEGIN
PATCH_IF ((value != 0) AND
((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
SET value -= 1
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_key BEGIN "%matches_0%", "%matches_1%", "%matches_2%", "%matches_3%", "%matches_4%", "%matches_5%", "%matches_6%" => "%value%" END // update array to indicate match present
READ_ASCII (fx_off + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) template ((0x30 + (0xd8 * fx_type))) // read whole effect into a template
SET insert_point = index2
END
END
END
PATCH_IF ("%all_or%" STRING_COMPARE_CASE "all" = 0) BEGIN // if an all clause, clear template if any is not matched
PHP_EACH cd_immunity_batches_key AS matches => value BEGIN
PATCH_IF value != 0 BEGIN
SPRINT template "" // clear template, prevents advancement
END
END
END
// if we have a key match, run second pass looking for extras and deleting targeted effects
PATCH_IF ("%template%" STRING_COMPARE_CASE "") BEGIN // if key found, template is set and we look for associated/delete effects
SET last_adj = 0
FOR (index2 = 0 ; index2 < counter ; ++index2) BEGIN
READ_SHORT (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) opcode
READ_LONG (fx_off + 0x04 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param1
READ_LONG (fx_off + 0x08 + (0x10 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) param2
READ_ASCII (fx_off + 0x14 + (0x14 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) resref
PATCH_IF index2 = (counter - 1) BEGIN // on final loop
PATCH_IF ((opcode = 206) AND ("%SOURCE_RES%" STRING_COMPARE_CASE "%resref%" = 0)) BEGIN // special check to make sure last effect isn't immunity to self
SET last_adj = "-1" // if it is, "last" effects need to go one effect before it
END
END
// look through associated array for matches
PHP_EACH cd_immunity_batches_extras AS matches => value BEGIN
PATCH_IF ((value != 0) AND
((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
SET value -= 1
DEFINE_ASSOCIATIVE_ARRAY cd_immunity_batches_extras BEGIN "%matches_0%", "%matches_1%", "%matches_2%", "%matches_3%", "%matches_4%", "%matches_5%", "%matches_6%" => "%value%" END // update array to indicate match present
END
END
// look through delete array
PHP_EACH cd_immunity_batches_delete AS matches => value BEGIN
PATCH_IF (((matches_0 = opcode) OR (matches_0 < "-9")) AND // opcode matches and/or isn't set
((matches_1 = param1) OR (matches_1 < "-9")) AND // param1 matches and/or isn't set
((matches_2 = param2) OR (matches_2 < "-9")) AND // param2 matches and/or isn't set
(("%matches_3%" STRING_COMPARE_CASE "%resref%" = 0) OR ("%matches_3%" STRING_COMPARE_CASE "same" = 0))) BEGIN // match found
DELETE_BYTES (fx_off + (0x08 * fx_type) + ((abil_fx_idx + index2) * (0x30 + (0xd8 * fx_type)))) ((0x30 + (0xd8 * fx_type))) // delete effect
PATCH_IF insert_point > index2 BEGIN SET insert_point -= 1 END // move insert point up if we're deleting an effect in front of it
SET new_fx -= 1
SET counter -= 1
SET index2 -= 1
SET debug_message = 1 // something's changed!
END
END
END // end second effects pass
// with a completed second pass, start adding needed effects
PATCH_FOR_EACH array IN cd_immunity_batches_key cd_immunity_batches_extras BEGIN
PHP_EACH ~%array%~ AS matches => value BEGIN
PATCH_IF debug > 2 BEGIN PATCH_PRINT ~%array% array: %matches_0%, %matches_1%, %matches_2%, %matches_3%, %matches_4%, %matches_5%, %matches_6% => %value%~ END
WHILE (value > 0) BEGIN
SET base = (fx_off + ((abil_fx_idx + insert_point) * (0x30 + (0xd8 * fx_type))))
PATCH_IF ("%matches_6%" STRING_COMPARE_CASE "first" = 0) BEGIN SET base = (fx_off + ((abil_fx_idx ) * (0x30 + (0xd8 * fx_type)))) END
PATCH_IF ("%matches_6%" STRING_COMPARE_CASE "last" = 0) BEGIN SET base = (fx_off + ((abil_fx_idx + counter + last_adj) * (0x30 + (0xd8 * fx_type)))) END
INSERT_BYTES (base) (0x30 + (0xd8 * fx_type))
WRITE_ASCIIE (base ) "%template%" // clones immunity effect
PATCH_IF (matches_0 >= "-9") BEGIN WRITE_SHORT (base + (0x08 * fx_type)) matches_0 END // opcode
PATCH_IF (matches_1 >= "-9") BEGIN WRITE_LONG (base + 0x04 + (0x10 * fx_type)) matches_1 END // parameter1
PATCH_IF (matches_2 >= "-9") BEGIN WRITE_LONG (base + 0x08 + (0x10 * fx_type)) matches_2 END // parameter2
PATCH_IF ("%matches_3%" STRING_COMPARE_CASE "same") BEGIN WRITE_ASCIIE (base + 0x14 + (0x14 * fx_type)) ~%matches_3%~ #8 END // resref
PATCH_IF (matches_4 >= "-9") BEGIN WRITE_BYTE (base + 0x0c + (0x10 * fx_type)) matches_4 END // timing
PATCH_IF (matches_5 >= "-9") BEGIN WRITE_BYTE (base + 0x0e + (0x13 * fx_type)) matches_5 END // duration
SET new_fx += 1
SET counter += 1
SET debug_message = 1 // something's changed!
SET value -= 1
END
END
END
END
WRITE_SHORT counter_offset counter
END
// formerly the immunity post array
PATCH_IF (("%type%" STRING_COMPARE_CASE "cre" = 0) AND (new_fx > 0)) BEGIN // fix offsets for cre files if fx inserted
PATCH_FOR_EACH offset IN 0x2a0 0x2a8 0x2b0 0x2b8 0x2bc BEGIN
READ_LONG offset curr_off
PATCH_IF (fx_off < curr_off) BEGIN
WRITE_LONG offset (THIS + ((0x30 + (0xd8 * fx_type)) * new_fx))
END
END
END
// debug message
PATCH_IF debug > 0 BEGIN
READ_LONG 0x0c name
PATCH_IF (name > 0) AND (name < 999999) BEGIN
READ_STRREF 0x0c name
END ELSE BEGIN
READ_STRREF 0x08 name
END
PATCH_IF (debug_message != 0) BEGIN
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^........$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^.......$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^......$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^.....$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^....$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^...$" = 0) BEGIN SPRINT spc " " END ELSE
PATCH_IF ("%SOURCE_RES%" STRING_COMPARE_REGEXP "^..$" = 0) BEGIN SPRINT spc " " END ELSE
BEGIN SPRINT spc " " END
PATCH_PRINT " ~%SOURCE_FILE%~%spc%~override~ // %name%, %new_fx% new effects from %array_name%"
END ELSE BEGIN
PATCH_IF (debug > 1) BEGIN
PATCH_PRINT " // %SOURCE_FILE% had no new effects from %array_name%"
END
END
END
END // end file size check
END
``` | non_defect | effect batches spell saves arrays for batch effects the cd apply batch function defined below requires a macro name that macro needs to define three arrays for the function cd immunity batches key if any effect in this array is present on the item creature spell being patches it will proceed with adding extra effects and deleting effects as defined by the next two arrays if you have mutiple effects defined here it will match any of them and add any missing cd immunity batches extras this array contains all of the subsiduary effects that should accompany the effects in the key area cd immunity batches deletes if a key is found any effect matching this array will be deleted if a key is found the function will then proceed back through the file being patched and add if they are not present all of the effects not found in the key and extras arrays it will also only with a key effect matched delete any effects listed in the deletes array note that since it s a macro you can conditionally define your arrays the free action batch will remove stun immunity unless a particular component from the fixpack is installed for example remember that s it s being run in a patch context so you ll need to use patch actions instead of actions an example use the cd full dispel arrays uses a dispel magic opcode as the lone key element if found the function will delete a variety of old item removal opcodes defined in the deletes array and add if not already present the series of effects that cure feeblemind and deafness in general try to avoid explicit spell blockages with opcodes if you can block the relevant effects with other immunities do so as it s a more robust and extensible solution in some cases they are needed e g the slow immunity batch blocks slow spells outright as most come paired with ac and penalties which can t be blocked via opcode without serious collateral damage format of all arrays are opcode resref timing duration timing and duration are only used for new fx writes use in a field or same in string fields if the match value for that field doesn t matter always map the effect to zero dex save bonus for damaging evoc define patch macro b dex evoc arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end will save bonus for vs mind affecting sleep define patch macro b wis sleep arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end unconsciousness define patch macro b wis uncon arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end stun define patch macro b wis stun arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end define patch macro b wis arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end define patch macro b wis arrays sr begin define associative array cd immunity batches key begin same same some direct damage opcode don t matter which end define associative array cd immunity batches extras begin add effects ls first less than immune to base spell gr first greater than immune to base spell ls first less than or auto fail eq first eq eq first eq eq first eq gr first gr than eq first eq first eq first is eq first is eq first is end define associative array cd immunity batches delete begin remove everything from base spell end end the batch err patch batch effects define patch function cd apply batch int var debug spit out c e list same w additional info lists effects as they re added force cosmetic bonus var bonus vars for use in setting arrays bonus var bonus var str var array name same all or or begin set debug message formerly the immunity post array read ascii type set min size patch if type string compare case spl begin read long abil off else read short abil num else read long fx off else set counter offset set abil length set global loop set fx type set min size set cosmetic end else patch if type string compare case itm begin read long abil off else read short abil num else read long fx off else set counter offset set abil length set global loop set fx type set min size read byte flags else patch if flags band begin if droppable add portrait icons and whatnot set cosmetic end else begin set cosmetic end end else patch if type string compare case cre begin read long fx off else read byte fx type else set abil off set abil num set counter offset set abil length set global loop set min size set cosmetic end else begin patch print warning macro name macro halting file type not recognized type end if user overrides normal cosmetic stuff with function var patch if force cosmetic begin set cosmetic end else patch if force cosmetic begin set cosmetic end patch if source size min size and min size begin min size must get set by file type detection set new fx for index global loop index abil num index begin sprint template replaced if key match found always clear your arrays clear array cd immunity batches key clear array cd immunity batches extras clear array cd immunity batches deletes launch patch macro array name loads up arrays with effects to match patch if index begin if loop through globals needed set abil fx idx end else begin otherwise normal ability set counter offset abil off abil length index read short abil off abil length index abil fx idx set abil fx idx new fx write short abil off abil length index abil fx idx end read short counter offset counter fx num on global loop otherwise abil fx num run one pass purely looking for keys for counter begin read short fx off fx type abil fx idx fx type opcode read long fx off fx type abil fx idx fx type read long fx off fx type abil fx idx fx type read ascii fx off fx type abil fx idx fx type resref look through key array for matches php each cd immunity batches key as matches value begin patch if value and matches opcode or matches and opcode matches and or isn t set matches or matches and matches and or isn t set matches or matches and matches and or isn t set matches string compare case resref or matches string compare case same begin match found set value define associative array cd immunity batches key begin matches matches matches matches matches matches matches value end update array to indicate match present read ascii fx off abil fx idx fx type template fx type read whole effect into a template set insert point end end end patch if all or string compare case all begin if an all clause clear template if any is not matched php each cd immunity batches key as matches value begin patch if value begin sprint template clear template prevents advancement end end end if we have a key match run second pass looking for extras and deleting targeted effects patch if template string compare case begin if key found template is set and we look for associated delete effects set last adj for counter begin read short fx off fx type abil fx idx fx type opcode read long fx off fx type abil fx idx fx type read long fx off fx type abil fx idx fx type read ascii fx off fx type abil fx idx fx type resref patch if counter begin on final loop patch if opcode and source res string compare case resref begin special check to make sure last effect isn t immunity to self set last adj if it is last effects need to go one effect before it end end look through associated array for matches php each cd immunity batches extras as matches value begin patch if value and matches opcode or matches and opcode matches and or isn t set matches or matches and matches and or isn t set matches or matches and matches and or isn t set matches string compare case resref or matches string compare case same begin match found set value define associative array cd immunity batches extras begin matches matches matches matches matches matches matches value end update array to indicate match present end end look through delete array php each cd immunity batches delete as matches value begin patch if matches opcode or matches and opcode matches and or isn t set matches or matches and matches and or isn t set matches or matches and matches and or isn t set matches string compare case resref or matches string compare case same begin match found delete bytes fx off fx type abil fx idx fx type fx type delete effect patch if insert point begin set insert point end move insert point up if we re deleting an effect in front of it set new fx set counter set set debug message something s changed end end end end second effects pass with a completed second pass start adding needed effects patch for each array in cd immunity batches key cd immunity batches extras begin php each array as matches value begin patch if debug begin patch print array array matches matches matches matches matches matches matches value end while value begin set base fx off abil fx idx insert point fx type patch if matches string compare case first begin set base fx off abil fx idx fx type end patch if matches string compare case last begin set base fx off abil fx idx counter last adj fx type end insert bytes base fx type write asciie base template clones immunity effect patch if matches begin write short base fx type matches end opcode patch if matches begin write long base fx type matches end patch if matches begin write long base fx type matches end patch if matches string compare case same begin write asciie base fx type matches end resref patch if matches begin write byte base fx type matches end timing patch if matches begin write byte base fx type matches end duration set new fx set counter set debug message something s changed set value end end end end write short counter offset counter end formerly the immunity post array patch if type string compare case cre and new fx begin fix offsets for cre files if fx inserted patch for each offset in begin read long offset curr off patch if fx off curr off begin write long offset this fx type new fx end end end debug message patch if debug begin read long name patch if name and name begin read strref name end else begin read strref name end patch if debug message begin patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else patch if source res string compare regexp begin sprint spc end else begin sprint spc end patch print source file spc override name new fx new effects from array name end else begin patch if debug begin patch print source file had no new effects from array name end end end end end file size check end | 0 |
43,028 | 11,438,573,708 | IssuesEvent | 2020-02-05 04:05:12 | zealdocs/zeal | https://api.github.com/repos/zealdocs/zeal | closed | High CPU usage on Xubuntu 19.10 without any docs installed | platform/linux resolution/fixed scope/ui/webview type/defect | As soon as I start Zeal, CPU usage for only Zeal shoots up to 100% instantly. I've uninstalled all docsets and the problem persists. As soon as I close Zeal everything is back to normal.
I'm using Zeal version 0.6.1 on Xubuntu 19.10.
I'm not sure when the problem started, but I have my computer's CPU information and temperature permanently displayed on the bottom of my screen, so I don't think it could have been that long. It's possible it started when I upgraded to Xubuntu 19.10 which was about a week ago.
I've looked at similar issues but it's not #965 or #211 as it's not Xorg that shoots up to 100% CPU usage but rather Zeal itself.
It's also not #999 because it also happens on the home screen without any docsets installed. | 1.0 | High CPU usage on Xubuntu 19.10 without any docs installed - As soon as I start Zeal, CPU usage for only Zeal shoots up to 100% instantly. I've uninstalled all docsets and the problem persists. As soon as I close Zeal everything is back to normal.
I'm using Zeal version 0.6.1 on Xubuntu 19.10.
I'm not sure when the problem started, but I have my computer's CPU information and temperature permanently displayed on the bottom of my screen, so I don't think it could have been that long. It's possible it started when I upgraded to Xubuntu 19.10 which was about a week ago.
I've looked at similar issues but it's not #965 or #211 as it's not Xorg that shoots up to 100% CPU usage but rather Zeal itself.
It's also not #999 because it also happens on the home screen without any docsets installed. | defect | high cpu usage on xubuntu without any docs installed as soon as i start zeal cpu usage for only zeal shoots up to instantly i ve uninstalled all docsets and the problem persists as soon as i close zeal everything is back to normal i m using zeal version on xubuntu i m not sure when the problem started but i have my computer s cpu information and temperature permanently displayed on the bottom of my screen so i don t think it could have been that long it s possible it started when i upgraded to xubuntu which was about a week ago i ve looked at similar issues but it s not or as it s not xorg that shoots up to cpu usage but rather zeal itself it s also not because it also happens on the home screen without any docsets installed | 1 |
382,420 | 11,305,995,278 | IssuesEvent | 2020-01-18 10:34:25 | BlueMap-Minecraft/BlueMap | https://api.github.com/repos/BlueMap-Minecraft/BlueMap | closed | [CLI] Only one world is checked for if it is done | bug high priority module: cli | Using the CLI to generate a map for a bunch of worlds.
## Steps taken
1. Rendered a rather small world.
2. Restarted the CLI once it was done. (It had started on the next world)
3. CLI informs me that the first map is done, and that nothing needs to be done.
4. At this point the CLI does not start rendering the next map.
If the map that was done rendering is commented out, it moves on to the next one. | 1.0 | [CLI] Only one world is checked for if it is done - Using the CLI to generate a map for a bunch of worlds.
## Steps taken
1. Rendered a rather small world.
2. Restarted the CLI once it was done. (It had started on the next world)
3. CLI informs me that the first map is done, and that nothing needs to be done.
4. At this point the CLI does not start rendering the next map.
If the map that was done rendering is commented out, it moves on to the next one. | non_defect | only one world is checked for if it is done using the cli to generate a map for a bunch of worlds steps taken rendered a rather small world restarted the cli once it was done it had started on the next world cli informs me that the first map is done and that nothing needs to be done at this point the cli does not start rendering the next map if the map that was done rendering is commented out it moves on to the next one | 0 |
57,350 | 14,099,391,979 | IssuesEvent | 2020-11-06 01:16:12 | storybookjs/storybook | https://api.github.com/repos/storybookjs/storybook | closed | Module not found error doesn't properly exit during build | P1 bug build-storybook has workaround mdx todo tracked | **Describe the bug**
When building storybook using `build-storybook` with module resolution errors in an `.mdx` file, a build error occurs but the build process hangs instead of exiting with a non-zero exit code. This is observed when using `@storybook/addons-docs` with a story using `.mdx` with an improper import.
This is problematic for our teams as we run this process during CI/CD pipelines and it hangs our builds if one of our developers makes an error in their `.mdx` files.
```
> storybook-build-hang@0.0.0 build-storybook M:\scratch\storybook-build-hang
> build-storybook
info @storybook/angular v5.3.18
info
info clean outputDir..
info => Copying prebuild dll's..
info => Building manager..
info => Loading manager config..
info => Loading presets
info => Compiling manager..
info => manager built (8.11 s)
info => Building preview..
info => Loading preview config..
info => Loading presets
info => Loading config/preview file in "./.storybook".
info => Adding stories defined in ".storybook\main.js".
info => Found custom tsconfig.json
info => Using default Webpack setup.
info => Using angular project 'storybook-build-hang' for configuring Storybook.
info => Loading angular-cli config.
info => Get angular-cli webpack config.
info => Compiling preview..
Starting type checking service...
Using 1 worker with 2048MB memory limit
70% building 40/90 modules 50 active M:\scratch\storybook-build-hang\node_modules\core-js\internals\set-to-string-tag.jsERR! => Failed to build the preview
ERR! Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
(node:39788) UnhandledPromiseRejectionWarning: ModuleNotFoundError: Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\Compilation.js:925:10
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:401:22
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:130:21
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:224:22
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:2830:7
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:6877:13
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:214:25
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:213:14
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\UnsafeCachePlugin.js:44:7
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:27:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\DescriptionFilePlugin.js:67:43
(node:39788) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:39788) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
70% building 1553/1553 modules 0 active
```
**To Reproduce**
Steps to reproduce the behavior:
1. Bootstrap a new Angular CLI Project
2. Run `npx -p @storybook/cli sb init` to bootstrap storybook for Angular
3. Install `@storybook/addon-docs` and create a story using MDX
4. Add a typo in a component import path
5. Build the project with `build-storybook`
**Expected behavior**
An error printed to the console and the process exiting with a non-zero exit code.
**Code snippets**
A repository has been created with this issue already reproduced:
https://github.com/esingletary/storybook-build-hang
**System:**
```
Environment Info:
System:
OS: Windows 10 10.0.17134
CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz
Binaries:
Node: 12.14.1 - C:\Program Files\nodejs\node.EXE
npm: 6.13.7 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: 42.17134.1098.0
``` | 1.0 | Module not found error doesn't properly exit during build - **Describe the bug**
When building storybook using `build-storybook` with module resolution errors in an `.mdx` file, a build error occurs but the build process hangs instead of exiting with a non-zero exit code. This is observed when using `@storybook/addons-docs` with a story using `.mdx` with an improper import.
This is problematic for our teams as we run this process during CI/CD pipelines and it hangs our builds if one of our developers makes an error in their `.mdx` files.
```
> storybook-build-hang@0.0.0 build-storybook M:\scratch\storybook-build-hang
> build-storybook
info @storybook/angular v5.3.18
info
info clean outputDir..
info => Copying prebuild dll's..
info => Building manager..
info => Loading manager config..
info => Loading presets
info => Compiling manager..
info => manager built (8.11 s)
info => Building preview..
info => Loading preview config..
info => Loading presets
info => Loading config/preview file in "./.storybook".
info => Adding stories defined in ".storybook\main.js".
info => Found custom tsconfig.json
info => Using default Webpack setup.
info => Using angular project 'storybook-build-hang' for configuring Storybook.
info => Loading angular-cli config.
info => Get angular-cli webpack config.
info => Compiling preview..
Starting type checking service...
Using 1 worker with 2048MB memory limit
70% building 40/90 modules 50 active M:\scratch\storybook-build-hang\node_modules\core-js\internals\set-to-string-tag.jsERR! => Failed to build the preview
ERR! Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
(node:39788) UnhandledPromiseRejectionWarning: ModuleNotFoundError: Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\Compilation.js:925:10
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:401:22
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:130:21
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:224:22
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:2830:7
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:6877:13
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:214:25
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:213:14
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\UnsafeCachePlugin.js:44:7
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:27:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\DescriptionFilePlugin.js:67:43
(node:39788) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:39788) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
70% building 1553/1553 modules 0 active
```
**To Reproduce**
Steps to reproduce the behavior:
1. Bootstrap a new Angular CLI Project
2. Run `npx -p @storybook/cli sb init` to bootstrap storybook for Angular
3. Install `@storybook/addon-docs` and create a story using MDX
4. Add a typo in a component import path
5. Build the project with `build-storybook`
**Expected behavior**
An error printed to the console and the process exiting with a non-zero exit code.
**Code snippets**
A repository has been created with this issue already reproduced:
https://github.com/esingletary/storybook-build-hang
**System:**
```
Environment Info:
System:
OS: Windows 10 10.0.17134
CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz
Binaries:
Node: 12.14.1 - C:\Program Files\nodejs\node.EXE
npm: 6.13.7 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: 42.17134.1098.0
``` | non_defect | module not found error doesn t properly exit during build describe the bug when building storybook using build storybook with module resolution errors in an mdx file a build error occurs but the build process hangs instead of exiting with a non zero exit code this is observed when using storybook addons docs with a story using mdx with an improper import this is problematic for our teams as we run this process during ci cd pipelines and it hangs our builds if one of our developers makes an error in their mdx files storybook build hang build storybook m scratch storybook build hang build storybook info storybook angular info info clean outputdir info copying prebuild dll s info building manager info loading manager config info loading presets info compiling manager info manager built s info building preview info loading preview config info loading presets info loading config preview file in storybook info adding stories defined in storybook main js info found custom tsconfig json info using default webpack setup info using angular project storybook build hang for configuring storybook info loading angular cli config info get angular cli webpack config info compiling preview starting type checking service using worker with memory limit building modules active m scratch storybook build hang node modules core js internals set to string tag jserr failed to build the preview err module not found error can t resolve app componen in m scratch storybook build hang src app node unhandledpromiserejectionwarning modulenotfounderror module not found error can t resolve app componen in m scratch storybook build hang src app at m scratch storybook build hang node modules webpack lib compilation js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules neo async async js at m scratch storybook build hang node modules neo async async js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules enhanced resolve lib resolver js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib unsafecacheplugin js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib descriptionfileplugin js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code building modules active to reproduce steps to reproduce the behavior bootstrap a new angular cli project run npx p storybook cli sb init to bootstrap storybook for angular install storybook addon docs and create a story using mdx add a typo in a component import path build the project with build storybook expected behavior an error printed to the console and the process exiting with a non zero exit code code snippets a repository has been created with this issue already reproduced system environment info system os windows cpu intel r core tm cpu binaries node c program files nodejs node exe npm c program files nodejs npm cmd browsers edge | 0 |
22,905 | 3,727,389,442 | IssuesEvent | 2016-03-06 08:05:05 | godfather1103/mentohust | https://api.github.com/repos/godfather1103/mentohust | closed | 在ubuntu 13.04下认证,显示无法找到服务器 | auto-migrated Priority-Medium Type-Defect | ```
mentohust_0.3.4-1_i386
我确认IP什么的设置正确,网线和接口也没问题(因为一回到
WIN就又能上网)
而且在WIN下使用WIN的mentohust也可以正常认证。
搞了很久也没搞定,在之前的ubuntu
12.10下也是一样的结果,改过组播方式,也是没用,我实在想
不出是哪里出问题了...
```
Original issue reported on code.google.com by `cpmsx...@gmail.com` on 9 May 2013 at 7:07 | 1.0 | 在ubuntu 13.04下认证,显示无法找到服务器 - ```
mentohust_0.3.4-1_i386
我确认IP什么的设置正确,网线和接口也没问题(因为一回到
WIN就又能上网)
而且在WIN下使用WIN的mentohust也可以正常认证。
搞了很久也没搞定,在之前的ubuntu
12.10下也是一样的结果,改过组播方式,也是没用,我实在想
不出是哪里出问题了...
```
Original issue reported on code.google.com by `cpmsx...@gmail.com` on 9 May 2013 at 7:07 | defect | 在ubuntu ,显示无法找到服务器 mentohust 我确认ip什么的设置正确,网线和接口也没问题(因为一回到 win就又能上网) 而且在win下使用win的mentohust也可以正常认证。 搞了很久也没搞定,在之前的ubuntu ,改过组播方式,也是没用,我实在想 不出是哪里出问题了 original issue reported on code google com by cpmsx gmail com on may at | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.