Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
625,974 | 19,783,535,242 | IssuesEvent | 2022-01-18 01:58:24 | tracer-protocol/pools-client | https://api.github.com/repos/tracer-protocol/pools-client | closed | Release 1.2 - Bugs from Testing | bug Priority: High | Based on current testing:
- [ ] Countdown timers don't match between toasts and table view

- [ ] Expected Price is usually $0, which affects the price through the whole process.

- [ ] 'View on Arbiscan' only appears on toast prior to confirming in web3 wallet, then changes to view order. It shouldn't disappear.

No toast notifications after queued toaster reaches 0 and disappears.
- [ ] No toast for minting in progress (user story is missing, I need to confirm with @KittyLomas)
- [ ] No toast for success

**To Reproduce**
1. Go to https://deploy-preview-380--tracer-pools.netlify.app
2. Follow gifs
**Desktop (please complete the following information):**
- OS: Windows 11
- Browser: Brave
- Version: 1.34.80
| 1.0 | Release 1.2 - Bugs from Testing - Based on current testing:
- [ ] Countdown timers don't match between toasts and table view

- [ ] Expected Price is usually $0, which affects the price through the whole process.

- [ ] 'View on Arbiscan' only appears on toast prior to confirming in web3 wallet, then changes to view order. It shouldn't disappear.

No toast notifications after queued toaster reaches 0 and disappears.
- [ ] No toast for minting in progress (user story is missing, I need to confirm with @KittyLomas)
- [ ] No toast for success

**To Reproduce**
1. Go to https://deploy-preview-380--tracer-pools.netlify.app
2. Follow gifs
**Desktop (please complete the following information):**
- OS: Windows 11
- Browser: Brave
- Version: 1.34.80
| priority | release bugs from testing based on current testing countdown timers don t match between toasts and table view expected price is usually which affects the price through the whole process view on arbiscan only appears on toast prior to confirming in wallet then changes to view order it shouldn t disappear no toast notifications after queued toaster reaches and disappears no toast for minting in progress user story is missing i need to confirm with kittylomas no toast for success to reproduce go to follow gifs desktop please complete the following information os windows browser brave version | 1 |
554,964 | 16,443,928,844 | IssuesEvent | 2021-05-20 17:13:23 | LBNL-ETA/BEDES-Manager | https://api.github.com/repos/LBNL-ETA/BEDES-Manager | closed | Warn user of duplicates when creating new composite term | bug high priority | Check if existing BEDES approved composite term exists and give an alert to the user. | 1.0 | Warn user of duplicates when creating new composite term - Check if existing BEDES approved composite term exists and give an alert to the user. | priority | warn user of duplicates when creating new composite term check if existing bedes approved composite term exists and give an alert to the user | 1 |
620,130 | 19,553,436,145 | IssuesEvent | 2022-01-03 04:05:05 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Resource path that precedes with / results in Missing identifier | Type/Improvement Priority/High Team/CompilerFE Area/Diagnostics Area/Parser | **Description:**
<!-- Give a brief description of the improvement -->
Then the resource path precedes with a `/`, it results in `Missing identifier` diagnostic error. This causes confusion in the Choreo resource form as we can't identify the issue. The error diagnostic should ne more meaningful to identify the exact error.
<img width="552" alt="Screenshot 2021-12-13 at 12 48 12" src="https://user-images.githubusercontent.com/5234623/145768863-a3beeede-6a46-4759-99ee-dad21d8dc56f.png">
**Describe your problem(s)**
```
import ballerina/http;
service / on new http:Listener(8080) {
resource function get /hello(string name) returns json|error? {
return error("");
}
}
```
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Resource path that precedes with / results in Missing identifier - **Description:**
<!-- Give a brief description of the improvement -->
Then the resource path precedes with a `/`, it results in `Missing identifier` diagnostic error. This causes confusion in the Choreo resource form as we can't identify the issue. The error diagnostic should ne more meaningful to identify the exact error.
<img width="552" alt="Screenshot 2021-12-13 at 12 48 12" src="https://user-images.githubusercontent.com/5234623/145768863-a3beeede-6a46-4759-99ee-dad21d8dc56f.png">
**Describe your problem(s)**
```
import ballerina/http;
service / on new http:Listener(8080) {
resource function get /hello(string name) returns json|error? {
return error("");
}
}
```
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | resource path that precedes with results in missing identifier description then the resource path precedes with a it results in missing identifier diagnostic error this causes confusion in the choreo resource form as we can t identify the issue the error diagnostic should ne more meaningful to identify the exact error img width alt screenshot at src describe your problem s import ballerina http service on new http listener resource function get hello string name returns json error return error describe your solution s related issues optional suggested labels optional suggested assignees optional | 1 |
450,298 | 13,001,671,303 | IssuesEvent | 2020-07-24 00:28:49 | UC-Davis-molecular-computing/scadnano-python-package | https://api.github.com/repos/UC-Davis-molecular-computing/scadnano-python-package | opened | add CI checks for API docs building and PyPI tar.gz file building | enhancement high priority | On each commit to dev, there should be a GitHub action that tries to build the API docs and the scadnano-x.x.x.tar.gz file that is uploaded to PyPI. This way, if either of these fails, we know that the docs or PyPI package action will fail when committed to master.
The checks could pass and the action on commit to master could still fail, but at least this way it would not be because the files simply could not be generated. | 1.0 | add CI checks for API docs building and PyPI tar.gz file building - On each commit to dev, there should be a GitHub action that tries to build the API docs and the scadnano-x.x.x.tar.gz file that is uploaded to PyPI. This way, if either of these fails, we know that the docs or PyPI package action will fail when committed to master.
The checks could pass and the action on commit to master could still fail, but at least this way it would not be because the files simply could not be generated. | priority | add ci checks for api docs building and pypi tar gz file building on each commit to dev there should be a github action that tries to build the api docs and the scadnano x x x tar gz file that is uploaded to pypi this way if either of these fails we know that the docs or pypi package action will fail when committed to master the checks could pass and the action on commit to master could still fail but at least this way it would not be because the files simply could not be generated | 1 |
34,766 | 2,787,472,776 | IssuesEvent | 2015-05-08 06:13:37 | CheckiO/checkio-empire-battle | https://api.github.com/repos/CheckiO/checkio-empire-battle | closed | Initial parameter "Size" should be a size of square, not a radius (2 time bigger) | complex:simple priority:high refactoring | Building sizes are not clear. | 1.0 | Initial parameter "Size" should be a size of square, not a radius (2 time bigger) - Building sizes are not clear. | priority | initial parameter size should be a size of square not a radius time bigger building sizes are not clear | 1 |
526,276 | 15,285,176,454 | IssuesEvent | 2021-02-23 13:13:52 | carbon-design-system/carbon-for-ibm-dotcom | https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom | closed | CTA section: Same height by CSS | Airtable Done dev package: react priority: high | ### The problem
Currently CTA section attempts to ensure all copy contents have the same height, by JavaScript code. It caused FOUC. FOUC causes false positives in Percy depending on when the screenshot is taken.
#### Additional Information
- scope includes content item horizontal, link list and button group
### The solution
Change it to CSS-based, e.g. `grid-auto-rows: 1fr` in CSS grid.
#### Acceptance Criteria
- [ ] No java script limitation for same height
- [ ] No user observable delay to apply the same height | 1.0 | CTA section: Same height by CSS - ### The problem
Currently CTA section attempts to ensure all copy contents have the same height, by JavaScript code. It caused FOUC. FOUC causes false positives in Percy depending on when the screenshot is taken.
#### Additional Information
- scope includes content item horizontal, link list and button group
### The solution
Change it to CSS-based, e.g. `grid-auto-rows: 1fr` in CSS grid.
#### Acceptance Criteria
- [ ] No java script limitation for same height
- [ ] No user observable delay to apply the same height | priority | cta section same height by css the problem currently cta section attempts to ensure all copy contents have the same height by javascript code it caused fouc fouc causes false positives in percy depending on when the screenshot is taken additional information scope includes content item horizontal link list and button group the solution change it to css based e g grid auto rows in css grid acceptance criteria no java script limitation for same height no user observable delay to apply the same height | 1 |
120,634 | 4,792,640,655 | IssuesEvent | 2016-10-31 16:02:13 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | classification creation | Enhancement Priority-High | When creating taxon names, species, scientific name, etc., are often not populated, and are somewhat painful to generate manually. Find interface magic.
| 1.0 | classification creation - When creating taxon names, species, scientific name, etc., are often not populated, and are somewhat painful to generate manually. Find interface magic.
| priority | classification creation when creating taxon names species scientific name etc are often not populated and are somewhat painful to generate manually find interface magic | 1 |
787,232 | 27,711,349,911 | IssuesEvent | 2023-03-14 14:26:06 | AY2223S2-CS2113-T13-1/tp | https://api.github.com/repos/AY2223S2-CS2113-T13-1/tp | closed | "create-account" command | type.Story priority.High | As a user, I can create multiple financial accounts so that I can better categorise my expenses and budgets.
### Acceptance Criteria
- Creates an account for the specified currency.
- For now, throw error if an account with the specified currency already exists
- Format: `create-account $/CURRENCY`
- Examples:
```java
>> create-account $/EUR
// Creates a $EUR account
``` | 1.0 | "create-account" command - As a user, I can create multiple financial accounts so that I can better categorise my expenses and budgets.
### Acceptance Criteria
- Creates an account for the specified currency.
- For now, throw error if an account with the specified currency already exists
- Format: `create-account $/CURRENCY`
- Examples:
```java
>> create-account $/EUR
// Creates a $EUR account
``` | priority | create account command as a user i can create multiple financial accounts so that i can better categorise my expenses and budgets acceptance criteria creates an account for the specified currency for now throw error if an account with the specified currency already exists format create account currency examples java create account eur creates a eur account | 1 |
352,632 | 10,544,332,336 | IssuesEvent | 2019-10-02 16:41:58 | eaudeweb/ozone | https://api.github.com/repos/eaudeweb/ozone | closed | Art7 - saving flags | Component: Backend Feature: Art7 Priority: High Status: In progress | related to #1354, copying last comment here, as this seems to be a backend issue:
When creating a submission as secretariat, before Submit there is a popup which allows changing the has_blanks flag:


Check some previously unchecked flags and submit:

After submit, the flags seem to be correctly updated:

The PUT request seems to have the right values:

but after refresh, the flags are reverted to the initial values:

| 1.0 | Art7 - saving flags - related to #1354, copying last comment here, as this seems to be a backend issue:
When creating a submission as secretariat, before Submit there is a popup which allows changing the has_blanks flag:


Check some previously unchecked flags and submit:

After submit, the flags seem to be correctly updated:

The PUT request seems to have the right values:

but after refresh, the flags are reverted to the initial values:

| priority | saving flags related to copying last comment here as this seems to be a backend issue when creating a submission as secretariat before submit there is a popup which allows changing the has blanks flag check some previously unchecked flags and submit after submit the flags seem to be correctly updated the put request seems to have the right values but after refresh the flags are reverted to the initial values | 1 |
606,278 | 18,759,032,629 | IssuesEvent | 2021-11-05 14:26:46 | wasmerio/wasmer | https://api.github.com/repos/wasmerio/wasmer | closed | Upgrade Inkwell to `0.1.0-beta.4` | 🎉 enhancement 📦 lib-compiler-llvm priority-high | Latest inkwell supports the metadata PR wasmer depends on. So we no longer need to depend on `wasmer-inkwell`
This will close the #2433 and allow us to release a new version of Wasmer supporting up to LLVM 13.
It should be a relatively easy change (just a version bump) | 1.0 | Upgrade Inkwell to `0.1.0-beta.4` - Latest inkwell supports the metadata PR wasmer depends on. So we no longer need to depend on `wasmer-inkwell`
This will close the #2433 and allow us to release a new version of Wasmer supporting up to LLVM 13.
It should be a relatively easy change (just a version bump) | priority | upgrade inkwell to beta latest inkwell supports the metadata pr wasmer depends on so we no longer need to depend on wasmer inkwell this will close the and allow us to release a new version of wasmer supporting up to llvm it should be a relatively easy change just a version bump | 1 |
755,722 | 26,437,827,103 | IssuesEvent | 2023-01-15 16:03:46 | Thorfusion/Mekanism-1.7.10-Community-Edition | https://api.github.com/repos/Thorfusion/Mekanism-1.7.10-Community-Edition | closed | [BUG]: crash report | TYPE: BUG PRIORITY: HIGH STATUS: FINISHED MC: 1.7.10 | ### Describe the bug
I don't know why the game won't start with the IC2.
### Expected behavior
Start the game normally.
### Mekanism Version
9.10.23-ALL
### Minecraft Version is this regarding?
1.7.10
### What OS are you seeing the problem on?
Windows
### Name of modpack if applicable
_No response_
### Version of said modpack if applicable
_No response_
### Screenshots

### The crash report in folder ./crash-reports (both server and client logs)
crash-2023-01-15_20.46.39-client.txt : [https://pastebin.com/qCr6SWw8](https://pastebin.com/qCr6SWw8)
### Please provide the following other files
Use default configuration file.
latest.txt : [https://pastebin.com/qFmSRLTV](https://pastebin.com/qFmSRLTV)
| 1.0 | [BUG]: crash report - ### Describe the bug
I don't know why the game won't start with the IC2.
### Expected behavior
Start the game normally.
### Mekanism Version
9.10.23-ALL
### Minecraft Version is this regarding?
1.7.10
### What OS are you seeing the problem on?
Windows
### Name of modpack if applicable
_No response_
### Version of said modpack if applicable
_No response_
### Screenshots

### The crash report in folder ./crash-reports (both server and client logs)
crash-2023-01-15_20.46.39-client.txt : [https://pastebin.com/qCr6SWw8](https://pastebin.com/qCr6SWw8)
### Please provide the following other files
Use default configuration file.
latest.txt : [https://pastebin.com/qFmSRLTV](https://pastebin.com/qFmSRLTV)
| priority | crash report describe the bug i don t know why the game won t start with the expected behavior start the game normally mekanism version all minecraft version is this regarding what os are you seeing the problem on windows name of modpack if applicable no response version of said modpack if applicable no response screenshots the crash report in folder crash reports both server and client logs crash client txt please provide the following other files use default configuration file latest txt | 1 |
326,828 | 9,961,591,617 | IssuesEvent | 2019-07-07 06:30:32 | orbs-network/orbs-network-go | https://api.github.com/repos/orbs-network/orbs-network-go | opened | TestSendSameTransactionFastToTwoNodes is flaky | flakiness high priority | ```
t.go:28: �[31;1minfo 2019-07-06T23:06:39.866969Z service sync node=a32884 service=block-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 function=servicesync.syncOneBlock source=services/blockstorage/servicesync/service_sync.go:62 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS
t.go:28: �[31;1minfo 2019-07-06T23:06:39.867052Z trying to commit state diff node=a32884 service=state-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 number-of-state-diffs=0 function=statestorage.(*service).CommitStateDiff source=services/statestorage/service.go:85 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS
require.go:157:
Error Trace: duplicate_tx_test.go:124
duplicate_tx_test.go:88
network_harness_builder.go:159
context.go:23
network_harness_builder.go:143
supervisor.go:60
supervisor.go:54
network_harness_builder.go:139
network_harness_builder.go:123
Error: Not equal:
expected: 1
actual : 0
Test: TestSendSameTransactionFastTwiceToSameNode/CONSENSUS_ALGO_TYPE_BENCHMARK_CONSENSUS
Messages: blocks should include tx exactly once
```
https://circleci.com/gh/orbs-network/orbs-network-go/16170#tests/containers/3
| 1.0 | TestSendSameTransactionFastToTwoNodes is flaky - ```
t.go:28: �[31;1minfo 2019-07-06T23:06:39.866969Z service sync node=a32884 service=block-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 function=servicesync.syncOneBlock source=services/blockstorage/servicesync/service_sync.go:62 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS
t.go:28: �[31;1minfo 2019-07-06T23:06:39.867052Z trying to commit state diff node=a32884 service=state-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 number-of-state-diffs=0 function=statestorage.(*service).CommitStateDiff source=services/statestorage/service.go:85 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS
require.go:157:
Error Trace: duplicate_tx_test.go:124
duplicate_tx_test.go:88
network_harness_builder.go:159
context.go:23
network_harness_builder.go:143
supervisor.go:60
supervisor.go:54
network_harness_builder.go:139
network_harness_builder.go:123
Error: Not equal:
expected: 1
actual : 0
Test: TestSendSameTransactionFastTwiceToSameNode/CONSENSUS_ALGO_TYPE_BENCHMARK_CONSENSUS
Messages: blocks should include tx exactly once
```
https://circleci.com/gh/orbs-network/orbs-network-go/16170#tests/containers/3
| priority | testsendsametransactionfasttotwonodes is flaky t go � service sync node service block storage entry point state storage sync request id state storage sync block height function servicesync synconeblock source services blockstorage servicesync service sync go test acceptance test id acc testsendsametransactionfasttwicetosamenode benchmark consensus t go � trying to commit state diff node service state storage entry point state storage sync request id state storage sync block height number of state diffs function statestorage service commitstatediff source services statestorage service go test acceptance test id acc testsendsametransactionfasttwicetosamenode benchmark consensus require go error trace duplicate tx test go duplicate tx test go network harness builder go context go network harness builder go supervisor go supervisor go network harness builder go network harness builder go error not equal expected actual test testsendsametransactionfasttwicetosamenode consensus algo type benchmark consensus messages blocks should include tx exactly once | 1 |
282,979 | 8,712,431,870 | IssuesEvent | 2018-12-06 22:12:34 | DaedalusGame/BetterWithAddons | https://api.github.com/repos/DaedalusGame/BetterWithAddons | closed | Book of Single example pictures broken after installation of BWA | high priority | Even the regular BWM example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links. Have tried rebooting, but to no avail. Were showing up just fine before inclusion of BWA into pack.
Using the following versions of the mods: BetterWithMods-1.12-2.3.16 & BetterWithLib-1.12-1.5 & Better+With+Addons-0.46

| 1.0 | Book of Single example pictures broken after installation of BWA - Even the regular BWM example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links. Have tried rebooting, but to no avail. Were showing up just fine before inclusion of BWA into pack.
Using the following versions of the mods: BetterWithMods-1.12-2.3.16 & BetterWithLib-1.12-1.5 & Better+With+Addons-0.46

| priority | book of single example pictures broken after installation of bwa even the regular bwm example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links have tried rebooting but to no avail were showing up just fine before inclusion of bwa into pack using the following versions of the mods betterwithmods betterwithlib better with addons | 1 |
474,485 | 13,670,906,422 | IssuesEvent | 2020-09-29 05:58:55 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.ebay-kleinanzeigen.de - site is not usable | browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/58983 -->
**URL**: https://www.ebay-kleinanzeigen.de/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
Verweist immer darauf eine App herunter zu laden. Daher muss der Desktopmodus aktiviert werden um die Seite anzeigen zu können
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/e9f97b56-8fdc-40d2-b0ec-7f56cd78a731.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/ccf949b2-368f-4520-9099-f2ade45c3e6a)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.ebay-kleinanzeigen.de - site is not usable - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/58983 -->
**URL**: https://www.ebay-kleinanzeigen.de/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
Verweist immer darauf eine App herunter zu laden. Daher muss der Desktopmodus aktiviert werden um die Seite anzeigen zu können
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/e9f97b56-8fdc-40d2-b0ec-7f56cd78a731.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/ccf949b2-368f-4520-9099-f2ade45c3e6a)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description missing items steps to reproduce verweist immer darauf eine app herunter zu laden daher muss der desktopmodus aktiviert werden um die seite anzeigen zu können view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
327,420 | 9,975,406,361 | IssuesEvent | 2019-07-09 13:01:39 | python/mypy | https://api.github.com/repos/python/mypy | closed | New semantic analyzer: crash on assignment to sqlalchemy @hybrid_property | crash new-semantic-analyzer priority-0-high | I'm testing out the new semantic analyzer, but it crashes on my codebase. Trying to assign to a [SQLAlchemy `@hybrid_property`](https://docs.sqlalchemy.org/en/13/orm/extensions/hybrid.html)) results in a `Cannot assign to a method` error followed by a crash.
Here's a fairly minimal reproduction for it:
```python3
from sqlalchemy import Base, Column, String
from sqlalchemy.ext.hybrid import hybrid_property
class FirstNameOnly(Base):
first_name = Column(String)
@hybrid_property
def name(self) -> str:
return self.first_name
@name.setter # type: ignore
def name(self, value: str) -> None:
self.first_name = value
def __init__(self, name: str):
self.name = name
```
The `# type: ignore` comment on the setter is a workaround for this issue I reported last year: https://github.com/python/mypy/issues/4430
Here's the full output:
```
(tildes) vagrant@ubuntu-xenial:/opt/tildes$ mypy --new-semantic-analyzer --show-traceback test_mypy.py
test_mypy.py:16: error: Cannot assign to a method
test_mypy.py:16: error: INTERNAL ERROR -- Please try using mypy master on Github:
https://mypy.rtfd.io/en/latest/common_issues.html#using-development-mypy-build
Please report a bug at https://github.com/python/mypy/issues
version: 0.720+dev.48916e63403645730a584d6898fbe925d513a841
Traceback (most recent call last):
File "/opt/venvs/tildes/bin/mypy", line 10, in <module>
sys.exit(console_entry())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/__main__.py", line 8, in console_entry
main(None, sys.stdout, sys.stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/main.py", line 83, in main
res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 164, in build
result = _build(sources, options, alt_lib_path, flush_errors, fscache, stdout, stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 224, in _build
graph = dispatch(sources, manager, stdout)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2567, in dispatch
process_graph(graph, manager)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2880, in process_graph
process_stale_scc(graph, scc, manager)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2987, in process_stale_scc
graph[id].type_check_first_pass()
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2096, in type_check_first_pass
self.type_checker().check_first_pass()
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 281, in check_first_pass
self.accept(d)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 913, in accept
return visitor.visit_class_def(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1596, in visit_class_def
self.accept(defn.defs)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept
return visitor.visit_block(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block
self.accept(s)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 655, in accept
return visitor.visit_func_def(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 703, in visit_func_def
self._visit_func_def(defn)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 707, in _visit_func_def
self.check_func_item(defn, name=defn.name())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 769, in check_func_item
self.check_func_def(defn, typ, name)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 935, in check_func_def
self.accept(item.body)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept
return visitor.visit_block(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block
self.accept(s)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 1036, in accept
return visitor.visit_assignment_stmt(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1793, in visit_assignment_stmt
self.check_assignment(s.lvalues[-1], s.rvalue, s.type is None, s.new_syntax)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1834, in check_assignment
lvalue_type, index_lvalue, inferred = self.check_lvalue(lvalue)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 2479, in check_lvalue
True)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkexpr.py", line 1766, in analyze_ordinary_member_access
in_literal_context=self.is_literal_context())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 103, in analyze_member_access
result = _analyze_member_access(name, typ, mx, override_info)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 117, in _analyze_member_access
return analyze_instance_member_access(name, typ, mx, override_info)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 179, in analyze_instance_member_access
signature = function_type(method, mx.builtin_type('builtins.function'))
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/types.py", line 2188, in function_type
assert isinstance(func, mypy.nodes.FuncItem), str(func)
AssertionError: OverloadedFuncDef:7(
Decorator:11(
Var(name)
MemberExpr:11(
NameExpr(name [test_mypy.FirstNameOnly.name])
setter)
FuncDef:12(
name
Args(
Var(self)
Var(value))
def (self: test_mypy.FirstNameOnly, value: builtins.str)
Block:12(
AssignmentStmt:13(
MemberExpr:13(
NameExpr(self [l])
first_name)
NameExpr(value [l]))))))
``` | 1.0 | New semantic analyzer: crash on assignment to sqlalchemy @hybrid_property - I'm testing out the new semantic analyzer, but it crashes on my codebase. Trying to assign to a [SQLAlchemy `@hybrid_property`](https://docs.sqlalchemy.org/en/13/orm/extensions/hybrid.html)) results in a `Cannot assign to a method` error followed by a crash.
Here's a fairly minimal reproduction for it:
```python3
from sqlalchemy import Base, Column, String
from sqlalchemy.ext.hybrid import hybrid_property
class FirstNameOnly(Base):
first_name = Column(String)
@hybrid_property
def name(self) -> str:
return self.first_name
@name.setter # type: ignore
def name(self, value: str) -> None:
self.first_name = value
def __init__(self, name: str):
self.name = name
```
The `# type: ignore` comment on the setter is a workaround for this issue I reported last year: https://github.com/python/mypy/issues/4430
Here's the full output:
```
(tildes) vagrant@ubuntu-xenial:/opt/tildes$ mypy --new-semantic-analyzer --show-traceback test_mypy.py
test_mypy.py:16: error: Cannot assign to a method
test_mypy.py:16: error: INTERNAL ERROR -- Please try using mypy master on Github:
https://mypy.rtfd.io/en/latest/common_issues.html#using-development-mypy-build
Please report a bug at https://github.com/python/mypy/issues
version: 0.720+dev.48916e63403645730a584d6898fbe925d513a841
Traceback (most recent call last):
File "/opt/venvs/tildes/bin/mypy", line 10, in <module>
sys.exit(console_entry())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/__main__.py", line 8, in console_entry
main(None, sys.stdout, sys.stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/main.py", line 83, in main
res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 164, in build
result = _build(sources, options, alt_lib_path, flush_errors, fscache, stdout, stderr)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 224, in _build
graph = dispatch(sources, manager, stdout)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2567, in dispatch
process_graph(graph, manager)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2880, in process_graph
process_stale_scc(graph, scc, manager)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2987, in process_stale_scc
graph[id].type_check_first_pass()
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2096, in type_check_first_pass
self.type_checker().check_first_pass()
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 281, in check_first_pass
self.accept(d)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 913, in accept
return visitor.visit_class_def(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1596, in visit_class_def
self.accept(defn.defs)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept
return visitor.visit_block(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block
self.accept(s)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 655, in accept
return visitor.visit_func_def(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 703, in visit_func_def
self._visit_func_def(defn)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 707, in _visit_func_def
self.check_func_item(defn, name=defn.name())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 769, in check_func_item
self.check_func_def(defn, typ, name)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 935, in check_func_def
self.accept(item.body)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept
return visitor.visit_block(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block
self.accept(s)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept
stmt.accept(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 1036, in accept
return visitor.visit_assignment_stmt(self)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1793, in visit_assignment_stmt
self.check_assignment(s.lvalues[-1], s.rvalue, s.type is None, s.new_syntax)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1834, in check_assignment
lvalue_type, index_lvalue, inferred = self.check_lvalue(lvalue)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 2479, in check_lvalue
True)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkexpr.py", line 1766, in analyze_ordinary_member_access
in_literal_context=self.is_literal_context())
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 103, in analyze_member_access
result = _analyze_member_access(name, typ, mx, override_info)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 117, in _analyze_member_access
return analyze_instance_member_access(name, typ, mx, override_info)
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 179, in analyze_instance_member_access
signature = function_type(method, mx.builtin_type('builtins.function'))
File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/types.py", line 2188, in function_type
assert isinstance(func, mypy.nodes.FuncItem), str(func)
AssertionError: OverloadedFuncDef:7(
Decorator:11(
Var(name)
MemberExpr:11(
NameExpr(name [test_mypy.FirstNameOnly.name])
setter)
FuncDef:12(
name
Args(
Var(self)
Var(value))
def (self: test_mypy.FirstNameOnly, value: builtins.str)
Block:12(
AssignmentStmt:13(
MemberExpr:13(
NameExpr(self [l])
first_name)
NameExpr(value [l]))))))
``` | priority | new semantic analyzer crash on assignment to sqlalchemy hybrid property i m testing out the new semantic analyzer but it crashes on my codebase trying to assign to a results in a cannot assign to a method error followed by a crash here s a fairly minimal reproduction for it from sqlalchemy import base column string from sqlalchemy ext hybrid import hybrid property class firstnameonly base first name column string hybrid property def name self str return self first name name setter type ignore def name self value str none self first name value def init self name str self name name the type ignore comment on the setter is a workaround for this issue i reported last year here s the full output tildes vagrant ubuntu xenial opt tildes mypy new semantic analyzer show traceback test mypy py test mypy py error cannot assign to a method test mypy py error internal error please try using mypy master on github please report a bug at version dev traceback most recent call last file opt venvs tildes bin mypy line in sys exit console entry file opt venvs tildes lib site packages mypy main py line in console entry main none sys stdout sys stderr file opt venvs tildes lib site packages mypy main py line in main res build build sources options none flush errors fscache stdout stderr file opt venvs tildes lib site packages mypy build py line in build result build sources options alt lib path flush errors fscache stdout stderr file opt venvs tildes lib site packages mypy build py line in build graph dispatch sources manager stdout file opt venvs tildes lib site packages mypy build py line in dispatch process graph graph manager file opt venvs tildes lib site packages mypy build py line in process graph process stale scc graph scc manager file opt venvs tildes lib site packages mypy build py line in process stale scc graph type check first pass file opt venvs tildes lib site packages mypy build py line in type check first pass self type checker check first pass file opt venvs tildes lib site packages mypy checker py line in check first pass self accept d file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit class def self file opt venvs tildes lib site packages mypy checker py line in visit class def self accept defn defs file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit block self file opt venvs tildes lib site packages mypy checker py line in visit block self accept s file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit func def self file opt venvs tildes lib site packages mypy checker py line in visit func def self visit func def defn file opt venvs tildes lib site packages mypy checker py line in visit func def self check func item defn name defn name file opt venvs tildes lib site packages mypy checker py line in check func item self check func def defn typ name file opt venvs tildes lib site packages mypy checker py line in check func def self accept item body file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit block self file opt venvs tildes lib site packages mypy checker py line in visit block self accept s file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit assignment stmt self file opt venvs tildes lib site packages mypy checker py line in visit assignment stmt self check assignment s lvalues s rvalue s type is none s new syntax file opt venvs tildes lib site packages mypy checker py line in check assignment lvalue type index lvalue inferred self check lvalue lvalue file opt venvs tildes lib site packages mypy checker py line in check lvalue true file opt venvs tildes lib site packages mypy checkexpr py line in analyze ordinary member access in literal context self is literal context file opt venvs tildes lib site packages mypy checkmember py line in analyze member access result analyze member access name typ mx override info file opt venvs tildes lib site packages mypy checkmember py line in analyze member access return analyze instance member access name typ mx override info file opt venvs tildes lib site packages mypy checkmember py line in analyze instance member access signature function type method mx builtin type builtins function file opt venvs tildes lib site packages mypy types py line in function type assert isinstance func mypy nodes funcitem str func assertionerror overloadedfuncdef decorator var name memberexpr nameexpr name setter funcdef name args var self var value def self test mypy firstnameonly value builtins str block assignmentstmt memberexpr nameexpr self first name nameexpr value | 1 |
697,789 | 23,952,898,564 | IssuesEvent | 2022-09-12 12:59:05 | benicamera/SupplyManager | https://api.github.com/repos/benicamera/SupplyManager | opened | Implement Item delete and Item create | good first issue Priority: High models business logic | # Tasks
- [ ] Create Item and add to list
- [ ] Remove Item from list
- [ ] Update view
## Create Item and add to list
- [ ] Creation form (maybe with image select)
- [ ] Add to list
## Remove Item from list
- [ ] Swipe or long press
- [ ] Confirm question | 1.0 | Implement Item delete and Item create - # Tasks
- [ ] Create Item and add to list
- [ ] Remove Item from list
- [ ] Update view
## Create Item and add to list
- [ ] Creation form (maybe with image select)
- [ ] Add to list
## Remove Item from list
- [ ] Swipe or long press
- [ ] Confirm question | priority | implement item delete and item create tasks create item and add to list remove item from list update view create item and add to list creation form maybe with image select add to list remove item from list swipe or long press confirm question | 1 |
347,216 | 10,426,653,488 | IssuesEvent | 2019-09-16 18:05:30 | jetrails/magento-cloudflare | https://api.github.com/repos/jetrails/magento-cloudflare | closed | Verify Zone ID Is Valid For Domain | priority: high request | Currently, only token is validated, but we should also check to see if zone id corresponds to the domain name in the current scope. | 1.0 | Verify Zone ID Is Valid For Domain - Currently, only token is validated, but we should also check to see if zone id corresponds to the domain name in the current scope. | priority | verify zone id is valid for domain currently only token is validated but we should also check to see if zone id corresponds to the domain name in the current scope | 1 |
239,841 | 7,800,088,230 | IssuesEvent | 2018-06-09 04:36:41 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0007142:
sometimes we filter to much html content | Bug Felamimail Mantis high priority | **Reported by pschuele on 25 Sep 2012 14:02**
**Version:** Joey (2012.10.1~beta2)
sometimes we filter to much html content -> empty mail
| 1.0 | 0007142:
sometimes we filter to much html content - **Reported by pschuele on 25 Sep 2012 14:02**
**Version:** Joey (2012.10.1~beta2)
sometimes we filter to much html content -> empty mail
| priority | sometimes we filter to much html content reported by pschuele on sep version joey sometimes we filter to much html content gt empty mail | 1 |
85,957 | 3,700,957,823 | IssuesEvent | 2016-02-29 10:57:08 | uds-datalab/PDBF | https://api.github.com/repos/uds-datalab/PDBF | closed | On Ubuntu with old version of TexLive mvn verify results in error | 1-high-priority bug wontfix | If you encounter an error message similar to this one:
> The file is not valid, error(s) :
> 1.2.1 : Body Syntax error, Single space expected [offset=2786901; key=2786901; line=5 0 obj <<; object=COSObject{5, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2786635; key=2786635; line=11 0 obj <<; object=COSObject{11, 0}]
> 1.2.1 : Body Syntax error, EOL expected before the 'endobj' keyword at offset 2786894
> 1.2.1 : Body Syntax error, Single space expected [offset=2761071; key=2761071; line=3 0 obj <<; object=COSObject{3, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2761433; key=2761433; line=8 0 obj <<; object=COSObject{8, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2787764; key=2787764; line=12 0 obj <<; object=COSObject{12, 0}]
then dont worry. This is a known issue with old versions of TexLive. You can safely ignore it or upgrade your TexLive to an up to date version (Should be fixed in TexLive 2015). | 1.0 | On Ubuntu with old version of TexLive mvn verify results in error - If you encounter an error message similar to this one:
> The file is not valid, error(s) :
> 1.2.1 : Body Syntax error, Single space expected [offset=2786901; key=2786901; line=5 0 obj <<; object=COSObject{5, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2786635; key=2786635; line=11 0 obj <<; object=COSObject{11, 0}]
> 1.2.1 : Body Syntax error, EOL expected before the 'endobj' keyword at offset 2786894
> 1.2.1 : Body Syntax error, Single space expected [offset=2761071; key=2761071; line=3 0 obj <<; object=COSObject{3, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2761433; key=2761433; line=8 0 obj <<; object=COSObject{8, 0}]
> 1.2.1 : Body Syntax error, Single space expected [offset=2787764; key=2787764; line=12 0 obj <<; object=COSObject{12, 0}]
then dont worry. This is a known issue with old versions of TexLive. You can safely ignore it or upgrade your TexLive to an up to date version (Should be fixed in TexLive 2015). | priority | on ubuntu with old version of texlive mvn verify results in error if you encounter an error message similar to this one the file is not valid error s body syntax error single space expected body syntax error single space expected body syntax error eol expected before the endobj keyword at offset body syntax error single space expected body syntax error single space expected body syntax error single space expected then dont worry this is a known issue with old versions of texlive you can safely ignore it or upgrade your texlive to an up to date version should be fixed in texlive | 1 |
84,180 | 3,654,789,606 | IssuesEvent | 2016-02-17 14:06:23 | emoncms/MyHomeEnergyPlanner | https://api.github.com/repos/emoncms/MyHomeEnergyPlanner | closed | Fabric Measures | feature High priority | Hi Carlos,
I've created a spreadsheet with some text measures in it as requested.
I've also had a think about the labels for the form fields - see notes in the spreadsheet.
Hopefully self-explanatory, and enough for you to be getting on with.
Thanks,
[20160208_Test Measures List.xlsx](https://github.com/emoncms/MyHomeEnergyPlanner/files/121593/20160208_Test.Measures.List.xlsx)
| 1.0 | Fabric Measures - Hi Carlos,
I've created a spreadsheet with some text measures in it as requested.
I've also had a think about the labels for the form fields - see notes in the spreadsheet.
Hopefully self-explanatory, and enough for you to be getting on with.
Thanks,
[20160208_Test Measures List.xlsx](https://github.com/emoncms/MyHomeEnergyPlanner/files/121593/20160208_Test.Measures.List.xlsx)
| priority | fabric measures hi carlos i ve created a spreadsheet with some text measures in it as requested i ve also had a think about the labels for the form fields see notes in the spreadsheet hopefully self explanatory and enough for you to be getting on with thanks | 1 |
710,806 | 24,435,502,110 | IssuesEvent | 2022-10-06 11:10:55 | hackforla/expunge-assist | https://api.github.com/repos/hackforla/expunge-assist | reopened | Review auto-generated text for repetition [from usability testing] | priority: high role: UX content writing feature: figma content writing size: 5pt | ### Overview
Auto-generated text needs to be reviewed and updated. For example, one user pointed out that several responses began with "Since my conviction…" Another user noticed repetitive sentences under the "Involvement: Job" section.
### Action Items
- [x] Review auto-generated text
- [x] Identify areas of repetition
- [x] Create new copy for those areas
- [x] Collaborate with Dev regarding the creation of randomly selected text (i.e. one sentence could be written in 3 ways and each user gets a randomly selected text inserted - this helps with creating variation in the letters that only 1-2 judges will see). (Answer from Dev: Cannot do right now)
- [x] Consider creating multiple sentence starters/fragments/etc. that users could choose from (personalize/make more authentic to each user). (Not for this iteration - revisit next)
- [x] Discuss in Content/iterate
- [x] Finalize
- [x] Link all appropriate documents/figma pages/etc. in the resource section below
- [ ] Hand over to Dev https://github.com/hackforla/expunge-assist/issues/705
### Resources/Instructions
This is for Form Fields Inconsistencies and Repetitiveness [Google Doc](https://docs.google.com/document/d/1UAjwLopUswtOleJrwB-oyuqyUR4mhOF08x8jR9AfJnk/edit?usp=sharing)
Continuing this work directly in Figma under the WIP LG page | 1.0 | Review auto-generated text for repetition [from usability testing] - ### Overview
Auto-generated text needs to be reviewed and updated. For example, one user pointed out that several responses began with "Since my conviction…" Another user noticed repetitive sentences under the "Involvement: Job" section.
### Action Items
- [x] Review auto-generated text
- [x] Identify areas of repetition
- [x] Create new copy for those areas
- [x] Collaborate with Dev regarding the creation of randomly selected text (i.e. one sentence could be written in 3 ways and each user gets a randomly selected text inserted - this helps with creating variation in the letters that only 1-2 judges will see). (Answer from Dev: Cannot do right now)
- [x] Consider creating multiple sentence starters/fragments/etc. that users could choose from (personalize/make more authentic to each user). (Not for this iteration - revisit next)
- [x] Discuss in Content/iterate
- [x] Finalize
- [x] Link all appropriate documents/figma pages/etc. in the resource section below
- [ ] Hand over to Dev https://github.com/hackforla/expunge-assist/issues/705
### Resources/Instructions
This is for Form Fields Inconsistencies and Repetitiveness [Google Doc](https://docs.google.com/document/d/1UAjwLopUswtOleJrwB-oyuqyUR4mhOF08x8jR9AfJnk/edit?usp=sharing)
Continuing this work directly in Figma under the WIP LG page | priority | review auto generated text for repetition overview auto generated text needs to be reviewed and updated for example one user pointed out that several responses began with since my conviction… another user noticed repetitive sentences under the involvement job section action items review auto generated text identify areas of repetition create new copy for those areas collaborate with dev regarding the creation of randomly selected text i e one sentence could be written in ways and each user gets a randomly selected text inserted this helps with creating variation in the letters that only judges will see answer from dev cannot do right now consider creating multiple sentence starters fragments etc that users could choose from personalize make more authentic to each user not for this iteration revisit next discuss in content iterate finalize link all appropriate documents figma pages etc in the resource section below hand over to dev resources instructions this is for form fields inconsistencies and repetitiveness continuing this work directly in figma under the wip lg page | 1 |
175,675 | 6,552,937,745 | IssuesEvent | 2017-09-05 20:22:08 | envistaInteractive/itagroup-ecommerce-template | https://api.github.com/repos/envistaInteractive/itagroup-ecommerce-template | opened | Events: Category / Search Page | High Priority Page Layout | ### Summary
Layout contents of Events: Category / Search page as specified on Events: Category / Search in Zeplin.
We do not have color mockups. The top bar is the blue that is also used on the checkout pages. Use those same classes and html. We will move that out of the checkout to be more generic later.
Use a mobile first approach to adjust the layout using responsive design as the screen gets larger.
**Use branch**: feature/events
**Layout file**: templates/events/list.liquid (file does not exist)
**Url for testing**: http://localhost:1337/events
**Delivery Date**: Sept 7th | 1.0 | Events: Category / Search Page - ### Summary
Layout contents of Events: Category / Search page as specified on Events: Category / Search in Zeplin.
We do not have color mockups. The top bar is the blue that is also used on the checkout pages. Use those same classes and html. We will move that out of the checkout to be more generic later.
Use a mobile first approach to adjust the layout using responsive design as the screen gets larger.
**Use branch**: feature/events
**Layout file**: templates/events/list.liquid (file does not exist)
**Url for testing**: http://localhost:1337/events
**Delivery Date**: Sept 7th | priority | events category search page summary layout contents of events category search page as specified on events category search in zeplin we do not have color mockups the top bar is the blue that is also used on the checkout pages use those same classes and html we will move that out of the checkout to be more generic later use a mobile first approach to adjust the layout using responsive design as the screen gets larger use branch feature events layout file templates events list liquid file does not exist url for testing delivery date sept | 1 |
347,031 | 10,423,479,679 | IssuesEvent | 2019-09-16 11:32:36 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Server always crashes/doesn't load | High Priority | When trying to start a server it always crashes and I get the message that errors ocurred loading users. Already tried to deinstall both Eco and Eco Server as well as restart multiple times. Nothing seems to help, can't play on the server with my friends at the moment.
The log with error message is attached. Might need some help!
[log_190915072601.log](https://github.com/StrangeLoopGames/EcoIssues/files/3615405/log_190915072601.log)
| 1.0 | Server always crashes/doesn't load - When trying to start a server it always crashes and I get the message that errors ocurred loading users. Already tried to deinstall both Eco and Eco Server as well as restart multiple times. Nothing seems to help, can't play on the server with my friends at the moment.
The log with error message is attached. Might need some help!
[log_190915072601.log](https://github.com/StrangeLoopGames/EcoIssues/files/3615405/log_190915072601.log)
| priority | server always crashes doesn t load when trying to start a server it always crashes and i get the message that errors ocurred loading users already tried to deinstall both eco and eco server as well as restart multiple times nothing seems to help can t play on the server with my friends at the moment the log with error message is attached might need some help | 1 |
831,773 | 32,060,525,832 | IssuesEvent | 2023-09-24 15:57:01 | oncokb/oncokb | https://api.github.com/repos/oncokb/oncokb | opened | Some OncoKB genes do not have ensembl gene curated | bug high priority | All OncoKB genes should have ensembl gene transcript, otherwise the genomic change will be filtered out for annotation(missing chromosome/start/end). Therefore the genomic change and hgvsg annotation will not work. | 1.0 | Some OncoKB genes do not have ensembl gene curated - All OncoKB genes should have ensembl gene transcript, otherwise the genomic change will be filtered out for annotation(missing chromosome/start/end). Therefore the genomic change and hgvsg annotation will not work. | priority | some oncokb genes do not have ensembl gene curated all oncokb genes should have ensembl gene transcript otherwise the genomic change will be filtered out for annotation missing chromosome start end therefore the genomic change and hgvsg annotation will not work | 1 |
345,629 | 10,370,688,866 | IssuesEvent | 2019-09-08 14:41:29 | byaka/VombatiDB | https://api.github.com/repos/byaka/VombatiDB | opened | Добавить режим хранения индекса `wide` при котором выделяется ячейка на хранение ссылки на данные | high-priority improvement optimization | В таком режиме все методы доступа к индексу будут разпаковывать ноду не на 2 обьекта (`props, childs`), а на 3 (`..,data`). При этом распакованный третий обьект станет передаваться в `_getData()` и аналоги.
**Похоже, реализация этого как отдельного режима работы ядра будет тяжелой изза того, что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок. В таком случае лучше выкинуть старый режим.** | 1.0 | Добавить режим хранения индекса `wide` при котором выделяется ячейка на хранение ссылки на данные - В таком режиме все методы доступа к индексу будут разпаковывать ноду не на 2 обьекта (`props, childs`), а на 3 (`..,data`). При этом распакованный третий обьект станет передаваться в `_getData()` и аналоги.
**Похоже, реализация этого как отдельного режима работы ядра будет тяжелой изза того, что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок. В таком случае лучше выкинуть старый режим.** | priority | добавить режим хранения индекса wide при котором выделяется ячейка на хранение ссылки на данные в таком режиме все методы доступа к индексу будут разпаковывать ноду не на обьекта props childs а на data при этом распакованный третий обьект станет передаваться в getdata и аналоги похоже реализация этого как отдельного режима работы ядра будет тяжелой изза того что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок в таком случае лучше выкинуть старый режим | 1 |
67,302 | 3,268,440,533 | IssuesEvent | 2015-10-23 11:30:51 | pakalbekim/armaldia | https://api.github.com/repos/pakalbekim/armaldia | opened | Rebalance gym once again | High Priority | When increasing energy limit - use a lot of stamina;
when increasing stamina limit - use a lot of energy;
when increasing health limit - use moderate amounts of both; | 1.0 | Rebalance gym once again - When increasing energy limit - use a lot of stamina;
when increasing stamina limit - use a lot of energy;
when increasing health limit - use moderate amounts of both; | priority | rebalance gym once again when increasing energy limit use a lot of stamina when increasing stamina limit use a lot of energy when increasing health limit use moderate amounts of both | 1 |
487,329 | 14,040,547,886 | IssuesEvent | 2020-11-01 03:18:34 | xournalpp/xournalpp | https://api.github.com/repos/xournalpp/xournalpp | closed | Recently used file with "no such device" causes Xournal++ to crash at start. | Crash bug priority: high | **Affects versions :**
- OS: Arch Linux
- Desktop environment: Gnome-Wayland
- Version of Xournal++: 45a619d83f97205c92ac146f13d5ae00af83af7e
- Installation method: AUR ([xournalpp-git](https://aur.archlinux.org/packages/xournalpp-git/))
**Describe the bug**
When a file exists in the "recently used files" where the device is not available (f.e. a Samba mount) Xournal++ crashes at start.
**To Reproduce**
Steps to reproduce the behavior:
1. Mount a samba drive using a VPN interface
2. Open the file in Xournal++, close it
3. Make sure the VPN interface is not available without unmounting the Samba drive (`no such device`)
4. Re-open Xournal++
**Expected behavior**
Xournal++ ignores the unavailable file.
**Additional context**
Crash:
```
terminate called after throwing an instance of 'std::filesystem::__cxx11::filesystem_error'
what(): filesystem error: status: No such device [/mnt/samba/mount/that/is/not/available/file.pdf]
** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.368: [Crash Handler] Crashed with signal 6
** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.369: [Crash Handler] Wrote crash log to: $HOME/.cache/com.github.xournalpp.xournalpp/errorlogs/errorlog.20200908-123929.log
```
Error log: https://fb.hash.works/HMqP0W
File path is from `./.local/share/recently-used.xbel`:
```
<bookmark href="file:///mnt/samba/mount/that/is/not/available/file.pdf" added="2020-09-02T23:54:55Z" modified="2020-09-03T00:00:03Z" visited="1969-12-31T23:59:59Z">
```
This sounds like something that should've been fixed with #1730, but as I stated above I'm using the latest version in master. | 1.0 | Recently used file with "no such device" causes Xournal++ to crash at start. - **Affects versions :**
- OS: Arch Linux
- Desktop environment: Gnome-Wayland
- Version of Xournal++: 45a619d83f97205c92ac146f13d5ae00af83af7e
- Installation method: AUR ([xournalpp-git](https://aur.archlinux.org/packages/xournalpp-git/))
**Describe the bug**
When a file exists in the "recently used files" where the device is not available (f.e. a Samba mount) Xournal++ crashes at start.
**To Reproduce**
Steps to reproduce the behavior:
1. Mount a samba drive using a VPN interface
2. Open the file in Xournal++, close it
3. Make sure the VPN interface is not available without unmounting the Samba drive (`no such device`)
4. Re-open Xournal++
**Expected behavior**
Xournal++ ignores the unavailable file.
**Additional context**
Crash:
```
terminate called after throwing an instance of 'std::filesystem::__cxx11::filesystem_error'
what(): filesystem error: status: No such device [/mnt/samba/mount/that/is/not/available/file.pdf]
** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.368: [Crash Handler] Crashed with signal 6
** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.369: [Crash Handler] Wrote crash log to: $HOME/.cache/com.github.xournalpp.xournalpp/errorlogs/errorlog.20200908-123929.log
```
Error log: https://fb.hash.works/HMqP0W
File path is from `./.local/share/recently-used.xbel`:
```
<bookmark href="file:///mnt/samba/mount/that/is/not/available/file.pdf" added="2020-09-02T23:54:55Z" modified="2020-09-03T00:00:03Z" visited="1969-12-31T23:59:59Z">
```
This sounds like something that should've been fixed with #1730, but as I stated above I'm using the latest version in master. | priority | recently used file with no such device causes xournal to crash at start affects versions os arch linux desktop environment gnome wayland version of xournal installation method aur describe the bug when a file exists in the recently used files where the device is not available f e a samba mount xournal crashes at start to reproduce steps to reproduce the behavior mount a samba drive using a vpn interface open the file in xournal close it make sure the vpn interface is not available without unmounting the samba drive no such device re open xournal expected behavior xournal ignores the unavailable file additional context crash terminate called after throwing an instance of std filesystem filesystem error what filesystem error status no such device com github xournalpp xournalpp warning crashed with signal com github xournalpp xournalpp warning wrote crash log to home cache com github xournalpp xournalpp errorlogs errorlog log error log file path is from local share recently used xbel this sounds like something that should ve been fixed with but as i stated above i m using the latest version in master | 1 |
2,438 | 2,525,857,317 | IssuesEvent | 2015-01-21 06:51:08 | graybeal/ont | https://api.github.com/repos/graybeal/ont | closed | Allow tab delimiter when importing data in voc2rdf | 1 star enhancement imported Milestone-Release1.2 Priority-High voc2rdf | _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on April 06, 2009 11:51:45_
(thanks John for this feedback) What capability do you want added or improved? Allow tab delimiter when importing data in voc2rdf Where do you want this capability to be accessible? voc2rdf What sort of input/command mechanism do you want? In the CSV dialog, have a checkbox or something to indicate that the
contents are tab-delimited columns
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=115_ | 1.0 | Allow tab delimiter when importing data in voc2rdf - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on April 06, 2009 11:51:45_
(thanks John for this feedback) What capability do you want added or improved? Allow tab delimiter when importing data in voc2rdf Where do you want this capability to be accessible? voc2rdf What sort of input/command mechanism do you want? In the CSV dialog, have a checkbox or something to indicate that the
contents are tab-delimited columns
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=115_ | priority | allow tab delimiter when importing data in from on april thanks john for this feedback what capability do you want added or improved allow tab delimiter when importing data in where do you want this capability to be accessible what sort of input command mechanism do you want in the csv dialog have a checkbox or something to indicate that the contents are tab delimited columns original issue | 1 |
421,761 | 12,261,138,679 | IssuesEvent | 2020-05-06 19:32:34 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | configurator: MariaDB doesn't use timezone setting until reboot | Priority: High Type: Bug | **Describe the bug**
When you set a timezone in configurator, MariaDB is not configured with that timezone until you reboot.
Consequently, admin user created in DB is created with a `valid_from` value based on **default** timezone. Depending on your timezone, it means that your account could be not valid after you reboot.
**To Reproduce**
Steps to reproduce the behavior:
1. Install a PacketFence ZEN
2. Set timezone to EST at step 2 of configurator
3. Set admin password
4. Check value in DB:
```sql
SELECT valid_from FROM password where pid='admin'\G;
```
5. Finish configurator
6. Log in on web admin with `admin` user
=> It works.
7. Reboot
8. Log in on web admin with `admin` user
=> It fails because account is not yet valid.
**Expected behavior**
MariaDB should be configured using timezone defined at step 2. | 1.0 | configurator: MariaDB doesn't use timezone setting until reboot - **Describe the bug**
When you set a timezone in configurator, MariaDB is not configured with that timezone until you reboot.
Consequently, admin user created in DB is created with a `valid_from` value based on **default** timezone. Depending on your timezone, it means that your account could be not valid after you reboot.
**To Reproduce**
Steps to reproduce the behavior:
1. Install a PacketFence ZEN
2. Set timezone to EST at step 2 of configurator
3. Set admin password
4. Check value in DB:
```sql
SELECT valid_from FROM password where pid='admin'\G;
```
5. Finish configurator
6. Log in on web admin with `admin` user
=> It works.
7. Reboot
8. Log in on web admin with `admin` user
=> It fails because account is not yet valid.
**Expected behavior**
MariaDB should be configured using timezone defined at step 2. | priority | configurator mariadb doesn t use timezone setting until reboot describe the bug when you set a timezone in configurator mariadb is not configured with that timezone until you reboot consequently admin user created in db is created with a valid from value based on default timezone depending on your timezone it means that your account could be not valid after you reboot to reproduce steps to reproduce the behavior install a packetfence zen set timezone to est at step of configurator set admin password check value in db sql select valid from from password where pid admin g finish configurator log in on web admin with admin user it works reboot log in on web admin with admin user it fails because account is not yet valid expected behavior mariadb should be configured using timezone defined at step | 1 |
636,792 | 20,609,351,789 | IssuesEvent | 2022-03-07 06:34:22 | harvester/harvester | https://api.github.com/repos/harvester/harvester | closed | [FEATURE] Soft reboot/shutdown | enhancement area/ui priority/1 highlight area/kubevirt | We should support a graceful soft reboot/shutdown from the UI, to allow the VM and filesystem on it to have a chance to shutdown properly.
Guest agent might be required. | 1.0 | [FEATURE] Soft reboot/shutdown - We should support a graceful soft reboot/shutdown from the UI, to allow the VM and filesystem on it to have a chance to shutdown properly.
Guest agent might be required. | priority | soft reboot shutdown we should support a graceful soft reboot shutdown from the ui to allow the vm and filesystem on it to have a chance to shutdown properly guest agent might be required | 1 |
129,076 | 5,088,229,403 | IssuesEvent | 2016-12-31 16:50:06 | zulip/zulip-electron | https://api.github.com/repos/zulip/zulip-electron | closed | Can not find module 'debug/browser' | bug help wanted Priority: High | Something is broken and I'm not able to figure it out 😭

Because of above error preload script can't be injected and hence spellchecker won't work 😢
| 1.0 | Can not find module 'debug/browser' - Something is broken and I'm not able to figure it out 😭

Because of above error preload script can't be injected and hence spellchecker won't work 😢
| priority | can not find module debug browser something is broken and i m not able to figure it out 😭 because of above error preload script can t be injected and hence spellchecker won t work 😢 | 1 |
283,027 | 8,713,246,379 | IssuesEvent | 2018-12-07 01:40:51 | aowen87/TicketTester | https://api.github.com/repos/aowen87/TicketTester | closed | visit2.9.2 release tarball[s?] are doubly-compressed | bug likelihood high priority reviewed severity medium wontfix | The release tarball:
http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz
is gzipped *twice*. That is, to extract one must:
$ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz
I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug?
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2307
Status: Rejected
Project: VisIt
Tracker: Bug
Priority: High
Subject: visit2.9.2 release tarball[s?] are doubly-compressed
Assigned to:
Category:
Target version: 2.10
Author: Tom Fogal
Start: 06/23/2015
Due date:
% Done: 0
Estimated time:
Created: 06/23/2015 09:34 pm
Updated: 08/26/2015 08:39 pm
Likelihood: 4 - Common
Severity: 3 - Major Irritation
Found in version: 2.9.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The release tarball:
http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz
is gzipped *twice*. That is, to extract one must:
$ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz
I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug?
Comments:
Hi Tom,I have double checked and the original tar file and when I do a "tar zxf visit2.9.2.linux-x86_64-rhel6.tar.gz" this uncompresses fine. I have run into issues where I download the file and doing a "tar zxf ..." gives an error. When I do it as 2 steps "gunzip visit2.9.2...; tar xf visit2.9.2..." it works fine. My thought was that my browser or the website was compressing it a second time. Can you let me know the md5sum of the file you downloaded?
I just verified that this behavior still exists with the 2.9.2 rhel6 download. When I look at the md5sum it is different from what is stored on the server at nersc. I am doing my download on Windows using IE. It also happens with Firefox on RHEL6. I gunzipped the file it downloaded and then the checksum matched. So, yes, it really is getting gzipped a second time. I will need to contact NERSC and see if they know why this is happening.
I just downloaded http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz via both firefox and in a terminal via wget. Via wget the file is fine. Via firefox I get the double-compressed issue.
We need to contact NERSC and find out how to turn this off.
This is no longer happening. I checked on Windows with IE and Linux with Firefox and Kevin checked on the Mac.
| 1.0 | visit2.9.2 release tarball[s?] are doubly-compressed - The release tarball:
http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz
is gzipped *twice*. That is, to extract one must:
$ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz
I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug?
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2307
Status: Rejected
Project: VisIt
Tracker: Bug
Priority: High
Subject: visit2.9.2 release tarball[s?] are doubly-compressed
Assigned to:
Category:
Target version: 2.10
Author: Tom Fogal
Start: 06/23/2015
Due date:
% Done: 0
Estimated time:
Created: 06/23/2015 09:34 pm
Updated: 08/26/2015 08:39 pm
Likelihood: 4 - Common
Severity: 3 - Major Irritation
Found in version: 2.9.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The release tarball:
http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz
is gzipped *twice*. That is, to extract one must:
$ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz
$ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz
I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug?
Comments:
Hi Tom,I have double checked and the original tar file and when I do a "tar zxf visit2.9.2.linux-x86_64-rhel6.tar.gz" this uncompresses fine. I have run into issues where I download the file and doing a "tar zxf ..." gives an error. When I do it as 2 steps "gunzip visit2.9.2...; tar xf visit2.9.2..." it works fine. My thought was that my browser or the website was compressing it a second time. Can you let me know the md5sum of the file you downloaded?
I just verified that this behavior still exists with the 2.9.2 rhel6 download. When I look at the md5sum it is different from what is stored on the server at nersc. I am doing my download on Windows using IE. It also happens with Firefox on RHEL6. I gunzipped the file it downloaded and then the checksum matched. So, yes, it really is getting gzipped a second time. I will need to contact NERSC and see if they know why this is happening.
I just downloaded http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz via both firefox and in a terminal via wget. Via wget the file is fine. Via firefox I get the double-compressed issue.
We need to contact NERSC and find out how to turn this off.
This is no longer happening. I checked on Windows with IE and Linux with Firefox and Kevin checked on the Mac.
| priority | release tarball are doubly compressed the release tarball is gzipped twice that is to extract one must gunzip linux tar gz mv linux tar linux tar gz tar zxvf linux tar gz i have a vague recollection that i hit this with as well perhaps there is a packaging script bug redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status rejected project visit tracker bug priority high subject release tarball are doubly compressed assigned to category target version author tom fogal start due date done estimated time created pm updated pm likelihood common severity major irritation found in version impact expected use os all support group any description the release tarball is gzipped twice that is to extract one must gunzip linux tar gz mv linux tar linux tar gz tar zxvf linux tar gz i have a vague recollection that i hit this with as well perhaps there is a packaging script bug comments hi tom i have double checked and the original tar file and when i do a tar zxf linux tar gz this uncompresses fine i have run into issues where i download the file and doing a tar zxf gives an error when i do it as steps gunzip tar xf it works fine my thought was that my browser or the website was compressing it a second time can you let me know the of the file you downloaded i just verified that this behavior still exists with the download when i look at the it is different from what is stored on the server at nersc i am doing my download on windows using ie it also happens with firefox on i gunzipped the file it downloaded and then the checksum matched so yes it really is getting gzipped a second time i will need to contact nersc and see if they know why this is happening i just downloaded via both firefox and in a terminal via wget via wget the file is fine via firefox i get the double compressed issue we need to contact nersc and find out how to turn this off this is no longer happening i checked on windows with ie and linux with firefox and kevin checked on the mac | 1 |
124,396 | 4,913,475,542 | IssuesEvent | 2016-11-23 12:41:30 | commercialhaskell/intero | https://api.github.com/repos/commercialhaskell/intero | closed | Not working for docker stack projects | component: emacs priority: high type: bug | I have a project using the "basic" stack template. When I build it with
``` yaml
docker:
enable: true
```
in my `stack.yaml` file intero stops working. Turning off docker support, deleting `.stack-work` and rebuilding and intero starts working again.
This is using the latest Docker for Mac with intero 0.1.18 for reference.
| 1.0 | Not working for docker stack projects - I have a project using the "basic" stack template. When I build it with
``` yaml
docker:
enable: true
```
in my `stack.yaml` file intero stops working. Turning off docker support, deleting `.stack-work` and rebuilding and intero starts working again.
This is using the latest Docker for Mac with intero 0.1.18 for reference.
| priority | not working for docker stack projects i have a project using the basic stack template when i build it with yaml docker enable true in my stack yaml file intero stops working turning off docker support deleting stack work and rebuilding and intero starts working again this is using the latest docker for mac with intero for reference | 1 |
403,667 | 11,844,572,962 | IssuesEvent | 2020-03-24 06:12:31 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Implement automatic cleanup of TRANSFER documents from ReqMgrAux | High Priority New Feature ReqMgr2MS Unified Porting | **Impact of the new feature**
ReqMgr2MS
**Is your feature request related to a problem? Please describe.**
Given that every request being handled by MSTransferor will result in one transfer info document in reqmgr_aux couch db, we need to have a mechanism that automatically deletes those documents, otherwise MSTransferor/MSMonitor might suffer of performance issues in the future.
**Describe the solution you'd like**
We need to decide at which status we should delete those files. For the moment, I think it will be useful to keep those documents for a bit longer in the system, such that we can get back to them if needed and check what were the transfer requests created.
Thus, I'd suggest to cleanup those documents once the workflow is done processing. I.e., if the workflow is NOT in `('staging', 'staged', 'acquired', 'running-open', 'running-closed')`, then the transfer info document can be deleted.
By the way, we should probably create a new cherrypy thread to take care of this cleanup. Running it twice a day is more than enough.
**Describe alternatives you've considered**
In the future, when this service is robust and stable, we can start deleting those transfer documents as soon as the workflow starts running.
**Additional context**
none | 1.0 | Implement automatic cleanup of TRANSFER documents from ReqMgrAux - **Impact of the new feature**
ReqMgr2MS
**Is your feature request related to a problem? Please describe.**
Given that every request being handled by MSTransferor will result in one transfer info document in reqmgr_aux couch db, we need to have a mechanism that automatically deletes those documents, otherwise MSTransferor/MSMonitor might suffer of performance issues in the future.
**Describe the solution you'd like**
We need to decide at which status we should delete those files. For the moment, I think it will be useful to keep those documents for a bit longer in the system, such that we can get back to them if needed and check what were the transfer requests created.
Thus, I'd suggest to cleanup those documents once the workflow is done processing. I.e., if the workflow is NOT in `('staging', 'staged', 'acquired', 'running-open', 'running-closed')`, then the transfer info document can be deleted.
By the way, we should probably create a new cherrypy thread to take care of this cleanup. Running it twice a day is more than enough.
**Describe alternatives you've considered**
In the future, when this service is robust and stable, we can start deleting those transfer documents as soon as the workflow starts running.
**Additional context**
none | priority | implement automatic cleanup of transfer documents from reqmgraux impact of the new feature is your feature request related to a problem please describe given that every request being handled by mstransferor will result in one transfer info document in reqmgr aux couch db we need to have a mechanism that automatically deletes those documents otherwise mstransferor msmonitor might suffer of performance issues in the future describe the solution you d like we need to decide at which status we should delete those files for the moment i think it will be useful to keep those documents for a bit longer in the system such that we can get back to them if needed and check what were the transfer requests created thus i d suggest to cleanup those documents once the workflow is done processing i e if the workflow is not in staging staged acquired running open running closed then the transfer info document can be deleted by the way we should probably create a new cherrypy thread to take care of this cleanup running it twice a day is more than enough describe alternatives you ve considered in the future when this service is robust and stable we can start deleting those transfer documents as soon as the workflow starts running additional context none | 1 |
168,489 | 6,376,540,156 | IssuesEvent | 2017-08-02 07:45:48 | aio-libs/aiokafka | https://api.github.com/repos/aio-libs/aiokafka | closed | Consumer and Coordinator should use a separate socket for connections | priority:high | Followup dev ticket for #128. See also https://issues.apache.org/jira/browse/KAFKA-1925 | 1.0 | Consumer and Coordinator should use a separate socket for connections - Followup dev ticket for #128. See also https://issues.apache.org/jira/browse/KAFKA-1925 | priority | consumer and coordinator should use a separate socket for connections followup dev ticket for see also | 1 |
353,787 | 10,558,799,694 | IssuesEvent | 2019-10-04 09:56:16 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | ID is shown as an object when fetching models with nested relationships | priority: high status: confirmed type: bug 🐛 | **Describe the bug**
If you create a group and one of the fields is a relationship with a model which have inside more relationships, when you fetch this data, it shows the id of the nested relationship as a object with the order of the character as key and the character as a value, in the same way as if you do {..."idString"}
**Steps to reproduce the behavior**
1. Create a group with a relation to a model with more relationships
2. Add the group to a model
**Expected behavior**
The nested relationship should show the id of the related content as a normal string
**Postman response**
```
[
{
"_id": "5d8db9bf0dc976285007b04d",
"items": [
{
"_id": "5d8db9bf0dc976285007b04e",
"notes": "qwe",
"qty": 2,
"createdAt": "2019-09-27T07:26:55.574Z",
"updatedAt": "2019-09-27T07:26:55.601Z",
"__v": 0,
"product": {
"_id": "5d6fbbe94e7c9554792b22b4",
"name": "product11",
"description": "tggyyfff77",
"createdAt": "2019-09-04T13:28:09.311Z",
"updatedAt": "2019-09-17T13:02:42.835Z",
"__v": 0,
"organization": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "2",
"5": "f",
"6": "e",
"7": "2",
"8": "8",
"9": "e",
"10": "6",
"11": "7",
"12": "5",
"13": "7",
"14": "3",
"15": "8",
"16": "7",
"17": "a",
"18": "c",
"19": "7",
"20": "6",
"21": "0",
"22": "3",
"23": "7"
},
"productgroup": {
"0": "5",
"1": "d",
"2": "6",
"3": "f",
"4": "b",
"5": "b",
"6": "d",
"7": "7",
"8": "4",
"9": "e",
"10": "7",
"11": "c",
"12": "9",
"13": "5",
"14": "5",
"15": "4",
"16": "7",
"17": "9",
"18": "2",
"19": "b",
"20": "2",
"21": "2",
"22": "b",
"23": "2"
},
"factory": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "4",
"5": "5",
"6": "7",
"7": "c",
"8": "7",
"9": "6",
"10": "2",
"11": "f",
"12": "2",
"13": "d",
"14": "4",
"15": "8",
"16": "f",
"17": "0",
"18": "6",
"19": "8",
"20": "8",
"21": "8",
"22": "6",
"23": "d"
},
"client": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "4",
"5": "2",
"6": "4",
"7": "0",
"8": "2",
"9": "8",
"10": "8",
"11": "e",
"12": "d",
"13": "e",
"14": "4",
"15": "7",
"16": "1",
"17": "e",
"18": "7",
"19": "2",
"20": "6",
"21": "1",
"22": "0",
"23": "3"
},
"sku": "12345",
"id": "5d6fbbe94e7c9554792b22b4"
},
"id": "5d8db9bf0dc976285007b04e"
}
]```
**System**
- Node.js version: v10.5
- NPM version: v6.7
- Strapi version: Beta 16.6
- Database: MongoDB
- Operating system: Windows 10
**Possible solution**
Now the id recovers its normal form with Object.values(strangeObjectID).reduce((a,b) => a+b, ""), maybe it lacks this in some point of the code. Or maybe there's a playful spread operator messing the string ID
| 1.0 | ID is shown as an object when fetching models with nested relationships - **Describe the bug**
If you create a group and one of the fields is a relationship with a model which have inside more relationships, when you fetch this data, it shows the id of the nested relationship as a object with the order of the character as key and the character as a value, in the same way as if you do {..."idString"}
**Steps to reproduce the behavior**
1. Create a group with a relation to a model with more relationships
2. Add the group to a model
**Expected behavior**
The nested relationship should show the id of the related content as a normal string
**Postman response**
```
[
{
"_id": "5d8db9bf0dc976285007b04d",
"items": [
{
"_id": "5d8db9bf0dc976285007b04e",
"notes": "qwe",
"qty": 2,
"createdAt": "2019-09-27T07:26:55.574Z",
"updatedAt": "2019-09-27T07:26:55.601Z",
"__v": 0,
"product": {
"_id": "5d6fbbe94e7c9554792b22b4",
"name": "product11",
"description": "tggyyfff77",
"createdAt": "2019-09-04T13:28:09.311Z",
"updatedAt": "2019-09-17T13:02:42.835Z",
"__v": 0,
"organization": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "2",
"5": "f",
"6": "e",
"7": "2",
"8": "8",
"9": "e",
"10": "6",
"11": "7",
"12": "5",
"13": "7",
"14": "3",
"15": "8",
"16": "7",
"17": "a",
"18": "c",
"19": "7",
"20": "6",
"21": "0",
"22": "3",
"23": "7"
},
"productgroup": {
"0": "5",
"1": "d",
"2": "6",
"3": "f",
"4": "b",
"5": "b",
"6": "d",
"7": "7",
"8": "4",
"9": "e",
"10": "7",
"11": "c",
"12": "9",
"13": "5",
"14": "5",
"15": "4",
"16": "7",
"17": "9",
"18": "2",
"19": "b",
"20": "2",
"21": "2",
"22": "b",
"23": "2"
},
"factory": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "4",
"5": "5",
"6": "7",
"7": "c",
"8": "7",
"9": "6",
"10": "2",
"11": "f",
"12": "2",
"13": "d",
"14": "4",
"15": "8",
"16": "f",
"17": "0",
"18": "6",
"19": "8",
"20": "8",
"21": "8",
"22": "6",
"23": "d"
},
"client": {
"0": "5",
"1": "d",
"2": "6",
"3": "e",
"4": "4",
"5": "2",
"6": "4",
"7": "0",
"8": "2",
"9": "8",
"10": "8",
"11": "e",
"12": "d",
"13": "e",
"14": "4",
"15": "7",
"16": "1",
"17": "e",
"18": "7",
"19": "2",
"20": "6",
"21": "1",
"22": "0",
"23": "3"
},
"sku": "12345",
"id": "5d6fbbe94e7c9554792b22b4"
},
"id": "5d8db9bf0dc976285007b04e"
}
]```
**System**
- Node.js version: v10.5
- NPM version: v6.7
- Strapi version: Beta 16.6
- Database: MongoDB
- Operating system: Windows 10
**Possible solution**
Now the id recovers its normal form with Object.values(strangeObjectID).reduce((a,b) => a+b, ""), maybe it lacks this in some point of the code. Or maybe there's a playful spread operator messing the string ID
| priority | id is shown as an object when fetching models with nested relationships describe the bug if you create a group and one of the fields is a relationship with a model which have inside more relationships when you fetch this data it shows the id of the nested relationship as a object with the order of the character as key and the character as a value in the same way as if you do idstring steps to reproduce the behavior create a group with a relation to a model with more relationships add the group to a model expected behavior the nested relationship should show the id of the related content as a normal string postman response id items id notes qwe qty createdat updatedat v product id name description createdat updatedat v organization d e f e e a c productgroup d f b b d e c b b factory d e c f d f d client d e e d e e sku id id system node js version npm version strapi version beta database mongodb operating system windows possible solution now the id recovers its normal form with object values strangeobjectid reduce a b a b maybe it lacks this in some point of the code or maybe there s a playful spread operator messing the string id | 1 |
796,272 | 28,104,684,476 | IssuesEvent | 2023-03-30 22:53:42 | medic/cht-core | https://api.github.com/repos/medic/cht-core | closed | Haproxy stops routing during stress test | Type: Bug Priority: 1 - High | **Describe the bug**
During e2e tests, I have frequently noticed the haproxy container reaching absurd CPU numbers on my machine (I have an AMD Ryzen 9 5900X). Very frequently, haproxy would take up to 100% of one CPU core.


Then, during a stress test on an AWS hosted distributed setup, everytime I loaded the instance with more than 100 users, the suite failed because haproxy had stopped responding. The container was not killed, or restarted, but the instance was not reachable and API reported not being able to connect to haproxy.
Upon inspecting haproxy logs, at first I noticed:
```
[NOTICE] (1) : haproxy version is 2.6.9-3a3700a
[NOTICE] (1) : path to executable is /usr/local/sbin/haproxy
[ALERT] (1) : Current worker (25) exited with code 137 (Killed)
[ALERT] (1) : exit-on-failure: killing every processes with SIGTERM
```
On a subsequent retry, I saw:
```
Thread 2 is about to kill the process.
Thread 1 : id=0x7fb8dce74f40 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=21 rqsz=760
1/1 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20789220579 now=21014641652 diff=225421073
curr_task=0x55705bb4b020 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x55709982b210
strm=0x55709982b210,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb7fbea2830,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x55705c4590f0,EST,20 scb=0x55705bb660e0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb85b606e50,80000300:H1(0x55705c46e3b0)/RAW((nil))/tcpv4(5635)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
*>Thread 2 : id=0x7fb8dce69700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=493
1/2 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=19770093994 now=86782386464 diff=67012292470
curr_task=0x7fb7d77d19d0 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb7d77d15e0
strm=0x7fb7d77d15e0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb7d77d1c30,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb813f12be0,EST,20 scb=0x7fb7d77d1af0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb893e46040,80000300:H1(0x7fb8da8cd700)/RAW((nil))/tcpv4(11218)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
call trace(22):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x7fb8dd600fe1 [48 8b 84 24 08 01 00 00]: libpthread:raise+0x141/0x179
| 0x55705a0ae3f7 [64 48 8b 53 10 64 48 8b]: main+0x165437
| 0x55705a0ae448 [0f 1f 84 00 00 00 00 00]: main+0x165488
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f82403 [48 85 d0 75 20 48 89 f0]: lua_take_global_lock+0x23/0x4c
| 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 3 : id=0x7fb8d3924700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=498
1/3 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20556386735 now=89985480430 diff=69429093695
curr_task=0x7fb833fd1390 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb833fd0fa0
strm=0x7fb833fd0fa0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb833fd15f0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb833fd0f30,EST,20 scb=0x7fb833fd14b0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb83741b110,80000300:H1(0x7fb85544fe40)/RAW((nil))/tcpv4(11845)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
Thread 4 : id=0x7fb8d3123700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=29 rqsz=778
1/4 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=21138813085 now=95953444002 diff=74814630917
curr_task=0x7fb86c80ae60 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb86c80aa70
strm=0x7fb86c80aa70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb86c80b0c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb86c80aa00,EST,20 scb=0x7fb86c80af80,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb853f53e20,80000300:H1(0x7fb8b12b27f0)/RAW((nil))/tcpv4(6201)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
>Thread 5 : id=0x7fb8d2922700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=10 rqsz=793
1/5 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=21689679358 now=99475683274 diff=77786003916
curr_task=0x7fb6d7fe4400 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb6d7fe4010
strm=0x7fb6d7fe4010,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb6d7fe4660,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb6d7fe3fa0,EST,20 scb=0x7fb6d7fe4520,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8730bcd20,80000300:H1(0x7fb8730bcf90)/RAW((nil))/tcpv4(6219)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f82428 [f3 90 eb d4 0f 1f 40 00]: lua_take_global_lock+0x48/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 6 : id=0x7fb8d2121700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=9 rqsz=706
1/6 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20990474533 now=100959756450 diff=79969281917
curr_task=0x7fb73bea7160 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb73bea6d70
strm=0x7fb73bea6d70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb73bea73c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb73bea6d00,EST,20 scb=0x7fb73bea7280,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb7e3f0bf90,80000300:H1(0x7fb89bf955e0)/RAW((nil))/tcpv4(6349)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 7 : id=0x7fb8d1920700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=3 rqsz=503
1/7 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20144821466 now=96965659318 diff=76820837852
curr_task=0x7fb74125de80 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb74125da90
strm=0x7fb74125da90,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb74125dff0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb7692a0e10,EST,20 scb=0x7fb74125df50,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8cf6f1b10,80000300:H1(0x7fb7692aa2f0)/RAW((nil))/tcpv4(10173)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
call trace(18):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
Thread 8 : id=0x7fb8d111f700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=25 rqsz=504
1/8 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20361036268 now=84727927263 diff=64366890995
curr_task=0x7fb8c0933120 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb8c0932d30
strm=0x7fb8c0932d30,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb8c3ce9a40,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb8c3cebee0,EST,20 scb=0x7fb8c3ce9930,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8abf12e60,80000300:H1(0x7fb8c3ce4af0)/RAW((nil))/tcpv4(8629)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
/usr/local/etc/haproxy/parse_basic.lua:12: in function line 9
[C]: in method 'gsub'
/usr/local/etc/haproxy/parse_basic.lua:9: in global 'dec'
/usr/local/etc/haproxy/parse_basic.lua:27: in function line 22
[NOTICE] (1) : haproxy version is 2.6.9-3a3700a
[NOTICE] (1) : path to executable is /usr/local/sbin/haproxy
[ALERT] (1) : Current worker (25) exited with code 134 (Aborted)
[ALERT] (1) : exit-on-failure: killing every processes with SIGTERM
```
Right before this happened, this is a snapshot of `docker stats` on the AWS instance:

Stats:
haproxy is using 282% cpu and 4.349G of memory.
**To Reproduce**
Steps to reproduce the behavior:
1. Run scalability suite with 100 concurrent users.
**Expected behavior**
Haproxy shouldn't take up as many resources. Also, the container should somehow restart automatically, instead of just hang when this issue happens.
**Environment**
- Instance: local & aws test instance
- App: haproxt
- Version: master
| 1.0 | Haproxy stops routing during stress test - **Describe the bug**
During e2e tests, I have frequently noticed the haproxy container reaching absurd CPU numbers on my machine (I have an AMD Ryzen 9 5900X). Very frequently, haproxy would take up to 100% of one CPU core.


Then, during a stress test on an AWS hosted distributed setup, everytime I loaded the instance with more than 100 users, the suite failed because haproxy had stopped responding. The container was not killed, or restarted, but the instance was not reachable and API reported not being able to connect to haproxy.
Upon inspecting haproxy logs, at first I noticed:
```
[NOTICE] (1) : haproxy version is 2.6.9-3a3700a
[NOTICE] (1) : path to executable is /usr/local/sbin/haproxy
[ALERT] (1) : Current worker (25) exited with code 137 (Killed)
[ALERT] (1) : exit-on-failure: killing every processes with SIGTERM
```
On a subsequent retry, I saw:
```
Thread 2 is about to kill the process.
Thread 1 : id=0x7fb8dce74f40 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=21 rqsz=760
1/1 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20789220579 now=21014641652 diff=225421073
curr_task=0x55705bb4b020 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x55709982b210
strm=0x55709982b210,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb7fbea2830,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x55705c4590f0,EST,20 scb=0x55705bb660e0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb85b606e50,80000300:H1(0x55705c46e3b0)/RAW((nil))/tcpv4(5635)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
*>Thread 2 : id=0x7fb8dce69700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=493
1/2 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=19770093994 now=86782386464 diff=67012292470
curr_task=0x7fb7d77d19d0 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb7d77d15e0
strm=0x7fb7d77d15e0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb7d77d1c30,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb813f12be0,EST,20 scb=0x7fb7d77d1af0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb893e46040,80000300:H1(0x7fb8da8cd700)/RAW((nil))/tcpv4(11218)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
call trace(22):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x7fb8dd600fe1 [48 8b 84 24 08 01 00 00]: libpthread:raise+0x141/0x179
| 0x55705a0ae3f7 [64 48 8b 53 10 64 48 8b]: main+0x165437
| 0x55705a0ae448 [0f 1f 84 00 00 00 00 00]: main+0x165488
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f82403 [48 85 d0 75 20 48 89 f0]: lua_take_global_lock+0x23/0x4c
| 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 3 : id=0x7fb8d3924700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=498
1/3 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20556386735 now=89985480430 diff=69429093695
curr_task=0x7fb833fd1390 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb833fd0fa0
strm=0x7fb833fd0fa0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb833fd15f0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb833fd0f30,EST,20 scb=0x7fb833fd14b0,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb83741b110,80000300:H1(0x7fb85544fe40)/RAW((nil))/tcpv4(11845)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
Thread 4 : id=0x7fb8d3123700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=29 rqsz=778
1/4 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=21138813085 now=95953444002 diff=74814630917
curr_task=0x7fb86c80ae60 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb86c80aa70
strm=0x7fb86c80aa70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb86c80b0c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb86c80aa00,EST,20 scb=0x7fb86c80af80,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb853f53e20,80000300:H1(0x7fb8b12b27f0)/RAW((nil))/tcpv4(6201)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
>Thread 5 : id=0x7fb8d2922700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=10 rqsz=793
1/5 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=21689679358 now=99475683274 diff=77786003916
curr_task=0x7fb6d7fe4400 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb6d7fe4010
strm=0x7fb6d7fe4010,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb6d7fe4660,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb6d7fe3fa0,EST,20 scb=0x7fb6d7fe4520,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8730bcd20,80000300:H1(0x7fb8730bcf90)/RAW((nil))/tcpv4(6219)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f82428 [f3 90 eb d4 0f 1f 40 00]: lua_take_global_lock+0x48/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 6 : id=0x7fb8d2121700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=9 rqsz=706
1/6 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20990474533 now=100959756450 diff=79969281917
curr_task=0x7fb73bea7160 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb73bea6d70
strm=0x7fb73bea6d70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb73bea73c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb73bea6d00,EST,20 scb=0x7fb73bea7280,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb7e3f0bf90,80000300:H1(0x7fb89bf955e0)/RAW((nil))/tcpv4(6349)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser --
call trace(19):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4
| 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
>Thread 7 : id=0x7fb8d1920700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=3 rqsz=503
1/7 stuck=1 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20144821466 now=96965659318 diff=76820837852
curr_task=0x7fb74125de80 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb74125da90
strm=0x7fb74125da90,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb74125dff0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb7692a0e10,EST,20 scb=0x7fb74125df50,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8cf6f1b10,80000300:H1(0x7fb7692aa2f0)/RAW((nil))/tcpv4(10173)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
call trace(18):
| 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0
| 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b
| 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140
| 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c
| 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f
| 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107
| 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1
| 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3
| 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034
| 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339
| 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035
| 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9
Thread 8 : id=0x7fb8d111f700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=25 rqsz=504
1/8 stuck=0 prof=0 harmless=0 wantrdv=0
cpu_ns: poll=20361036268 now=84727927263 diff=64366890995
curr_task=0x7fb8c0933120 (task) calls=1 last=0
fct=0x557059fe3950(process_stream) ctx=0x7fb8c0932d30
strm=0x7fb8c0932d30,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown
txn=0x7fb8c3ce9a40,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0
rqf=40d08002 rqa=30 rpf=80000000 rpa=0
scf=0x7fb8c3cebee0,EST,20 scb=0x7fb8c3ce9930,INI,21
af=(nil),0 sab=(nil),0
cof=0x7fb8abf12e60,80000300:H1(0x7fb8c3ce4af0)/RAW((nil))/tcpv4(8629)
cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1)
Current executing Lua from a stream analyser -- stack traceback:
/usr/local/etc/haproxy/parse_basic.lua:12: in function line 9
[C]: in method 'gsub'
/usr/local/etc/haproxy/parse_basic.lua:9: in global 'dec'
/usr/local/etc/haproxy/parse_basic.lua:27: in function line 22
[NOTICE] (1) : haproxy version is 2.6.9-3a3700a
[NOTICE] (1) : path to executable is /usr/local/sbin/haproxy
[ALERT] (1) : Current worker (25) exited with code 134 (Aborted)
[ALERT] (1) : exit-on-failure: killing every processes with SIGTERM
```
Right before this happened, this is a snapshot of `docker stats` on the AWS instance:

Stats:
haproxy is using 282% cpu and 4.349G of memory.
**To Reproduce**
Steps to reproduce the behavior:
1. Run scalability suite with 100 concurrent users.
**Expected behavior**
Haproxy shouldn't take up as many resources. Also, the container should somehow restart automatically, instead of just hang when this issue happens.
**Environment**
- Instance: local & aws test instance
- App: haproxt
- Version: master
| priority | haproxy stops routing during stress test describe the bug during tests i have frequently noticed the haproxy container reaching absurd cpu numbers on my machine i have an amd ryzen very frequently haproxy would take up to of one cpu core then during a stress test on an aws hosted distributed setup everytime i loaded the instance with more than users the suite failed because haproxy had stopped responding the container was not killed or restarted but the instance was not reachable and api reported not being able to connect to haproxy upon inspecting haproxy logs at first i noticed haproxy version is path to executable is usr local sbin haproxy current worker exited with code killed exit on failure killing every processes with sigterm on a subsequent retry i saw thread is about to kill the process thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback call trace ha dump backtrace debug handler libpthread libpthread raise main main libpthread lua take global lock main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback call trace ha dump backtrace debug handler libpthread lua take global lock main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback usr local etc haproxy parse basic lua in function line in method gsub usr local etc haproxy parse basic lua in global dec usr local etc haproxy parse basic lua in function line haproxy version is path to executable is usr local sbin haproxy current worker exited with code aborted exit on failure killing every processes with sigterm right before this happened this is a snapshot of docker stats on the aws instance stats haproxy is using cpu and of memory to reproduce steps to reproduce the behavior run scalability suite with concurrent users expected behavior haproxy shouldn t take up as many resources also the container should somehow restart automatically instead of just hang when this issue happens environment instance local aws test instance app haproxt version master | 1 |
383,854 | 11,363,266,132 | IssuesEvent | 2020-01-27 03:00:21 | wso2/docs-is | https://api.github.com/repos/wso2/docs-is | opened | Deprecate JWT Token Generation using SOAP service | Priority/Highest Severity/Critical | From IS 5.3.0 onward we have the user-info endpoint to retrieve user claims of a user in either plain-text or JWT format. Therefore JWT token generation using TokenValidationService [1] should be deprecated from all the IS docs from 5.3.0 version onwards.
[1] https://docs.wso2.com/display/IS530/JWT+Token+Generation | 1.0 | Deprecate JWT Token Generation using SOAP service - From IS 5.3.0 onward we have the user-info endpoint to retrieve user claims of a user in either plain-text or JWT format. Therefore JWT token generation using TokenValidationService [1] should be deprecated from all the IS docs from 5.3.0 version onwards.
[1] https://docs.wso2.com/display/IS530/JWT+Token+Generation | priority | deprecate jwt token generation using soap service from is onward we have the user info endpoint to retrieve user claims of a user in either plain text or jwt format therefore jwt token generation using tokenvalidationservice should be deprecated from all the is docs from version onwards | 1 |
114,346 | 4,629,241,142 | IssuesEvent | 2016-09-28 08:36:39 | VirtoCommerce/vc-platform | https://api.github.com/repos/VirtoCommerce/vc-platform | closed | "Object reference not set to an instance of an object." in CategoryBrowsingService.SearchCategories | bug high priority | can't start storefront on my localhost:

| 1.0 | "Object reference not set to an instance of an object." in CategoryBrowsingService.SearchCategories - can't start storefront on my localhost:

| priority | object reference not set to an instance of an object in categorybrowsingservice searchcategories can t start storefront on my localhost | 1 |
360,423 | 10,688,493,858 | IssuesEvent | 2019-10-22 18:21:09 | carbon-design-system/design-language-website | https://api.github.com/repos/carbon-design-system/design-language-website | opened | Update color palette specifications component to use v2 colors | priority: high | https://www.ibm.com/design/language/elements/color/#specifications
The `SwatchPalette` component on the Color page needs to get updated to show the v2 colors. Ideally this would pull directly from the colors package instead of using the hardcoded values. | 1.0 | Update color palette specifications component to use v2 colors - https://www.ibm.com/design/language/elements/color/#specifications
The `SwatchPalette` component on the Color page needs to get updated to show the v2 colors. Ideally this would pull directly from the colors package instead of using the hardcoded values. | priority | update color palette specifications component to use colors the swatchpalette component on the color page needs to get updated to show the colors ideally this would pull directly from the colors package instead of using the hardcoded values | 1 |
519,196 | 15,047,235,492 | IssuesEvent | 2021-02-03 08:38:32 | ossia/libossia | https://api.github.com/repos/ossia/libossia | closed | [ossia-max] ossia.remote outputs value twice (or more) | bug ossia-max priority::high | when value is changed from the bound parameter, ossia.remote outputs value twice (by both 1rst and 2nd outlets)
If value is changed by the inlet of the remote itself (or from another remote) : value is output 4 times.
if parameter / remote "repetitions" attribute is set to 0 : same issue
(this attribute seems to only filter repetition on the outlet of ossia.parameter)
<pre><code>
----------begin_max5_patcher----------
945.3oc0YssaiBCD84To9Of3Y1Hrwbak1ef8WX0pHGh2DWAFDXZS2p8ee8EH
MsEBNEm1lGRAaLLy4XOybr6S2dyB20k6IMtNe24WNKV7jnmEp9j8rnuiEtE3
8Y43F0.cyJKJHLtqW2C4j8b0CpIEkbhSYKupk23bONuk3fb3zBgI5GMqsPLf
bBW8w766ltQ8IJWe22f9GFbElmsix1tpljw0dIHBsLzyAjltz2S0RbA5uz24
2GYBJq2B.Um+61ajWEW7LFmqa47R1Q9RMtfvI0qHL75bxKb+WhJv.nB7Lpzi
j+XEQCI20X1V2m8+APMRhvCnFh5+q0AMi7fvae6baVYKSf8QlFQCAX33.lJW
9331+S1zD3CC.xKwApFvQge36G9sEqOBkmwbNbHJ.MNEHwswy6vvD4kP+I.t
8m2KaZn3kcQ1EOt5.iLxJgfgngvSSCpeFQCIcA8CvCYk4k05Q6uD.CQodhah
Si7iSj2kjh..wcfK.4cdYD4OPyHmQ9vjSlND.koCgQWWYCCmSxP.L4ZKYXj8
xE1g9qsbgAyMUXOtu3oBGMZ9GN0jJBleNZYhOUvKLNVhBjuZgbT3URvKDLmn
WHJQlwpCzeBQuMjbGe2ynJd5DvcxxWcPNHVAYzTgrvuJxWRlcLapF3fzOI4K
+M2IamvMImy7c7DxULBxZvFkdsLWO67yhJQBbmbomos7txlUpLMj0ZutJjgj
ZOYHZrqKFeIEgLB1qporypPr6zRK5RSgfWk6xBM2MY0Kv5K5lrNM2cfwLg9F
LAXzbx5qiF.PfdiInInNTpOHMxScWhefhECAwQn3KC2g475V5A.Ja1s+zJBm
xokrwNftA0+chS1YhkXI.sFd8RrKXYBKi3f2MhEUGDqLRAuO.qGlaNk8lSoU
YL4CdEMzT1Vm06b8NuyQFbCogSYXIGbznjq+kiZbF2bCBMwfHqYO.xH.FZOC
5ajAg1yfFwn.6QoQlXu.qYtPSLm8VfF7wt9L0nIO6wlFEN.sp4.SBuD6ELXD
eJOjBaYvDil.ismAiMhRSsmA+nKQ..lXvgAXWMQbU08j5lt2PaJgFf6zprR7
zsoLcasjOQ4+6o8uhdaBt3ZQMatnfcasxCc2G0Ik1snbColIjQzUYVBVooa3
Ol+l5wLgjS06+SbQYAd0N5VAUrcGmrYUY8QmAg.r+A2lyeEGgyxHL9QxDURF
7kpGz+u+VdDU30y20h2Q4NJS2agBwmW409ujyT5ijNZSEVOenzQc6Mhg7eP9
flpT
-----------end_max5_patcher-----------
</code></pre>
| 1.0 | [ossia-max] ossia.remote outputs value twice (or more) - when value is changed from the bound parameter, ossia.remote outputs value twice (by both 1rst and 2nd outlets)
If value is changed by the inlet of the remote itself (or from another remote) : value is output 4 times.
if parameter / remote "repetitions" attribute is set to 0 : same issue
(this attribute seems to only filter repetition on the outlet of ossia.parameter)
<pre><code>
----------begin_max5_patcher----------
945.3oc0YssaiBCD84To9Of3Y1Hrwbak1ef8WX0pHGh2DWAFDXZS2p8ee8EH
MsEBNEm1lGRAaLLy4XOybr6S2dyB20k6IMtNe24WNKV7jnmEp9j8rnuiEtE3
8Y43F0.cyJKJHLtqW2C4j8b0CpIEkbhSYKupk23bONuk3fb3zBgI5GMqsPLf
bBW8w766ltQ8IJWe22f9GFbElmsix1tpljw0dIHBsLzyAjltz2S0RbA5uz24
2GYBJq2B.Um+61ajWEW7LFmqa47R1Q9RMtfvI0qHL75bxKb+WhJv.nB7Lpzi
j+XEQCI20X1V2m8+APMRhvCnFh5+q0AMi7fvae6baVYKSf8QlFQCAX33.lJW
9331+S1zD3CC.xKwApFvQge36G9sEqOBkmwbNbHJ.MNEHwswy6vvD4kP+I.t
8m2KaZn3kcQ1EOt5.iLxJgfgngvSSCpeFQCIcA8CvCYk4k05Q6uD.CQodhah
Si7iSj2kjh..wcfK.4cdYD4OPyHmQ9vjSlND.koCgQWWYCCmSxP.L4ZKYXj8
xE1g9qsbgAyMUXOtu3oBGMZ9GN0jJBleNZYhOUvKLNVhBjuZgbT3URvKDLmn
WHJQlwpCzeBQuMjbGe2ynJd5DvcxxWcPNHVAYzTgrvuJxWRlcLapF3fzOI4K
+M2IamvMImy7c7DxULBxZvFkdsLWO67yhJQBbmbomos7txlUpLMj0ZutJjgj
ZOYHZrqKFeIEgLB1qporypPr6zRK5RSgfWk6xBM2MY0Kv5K5lrNM2cfwLg9F
LAXzbx5qiF.PfdiInInNTpOHMxScWhefhECAwQn3KC2g475V5A.Ja1s+zJBm
xokrwNftA0+chS1YhkXI.sFd8RrKXYBKi3f2MhEUGDqLRAuO.qGlaNk8lSoU
YL4CdEMzT1Vm06b8NuyQFbCogSYXIGbznjq+kiZbF2bCBMwfHqYO.xH.FZOC
5ajAg1yfFwn.6QoQlXu.qYtPSLm8VfF7wt9L0nIO6wlFEN.sp4.SBuD6ELXD
eJOjBaYvDil.ismAiMhRSsmA+nKQ..lXvgAXWMQbU08j5lt2PaJgFf6zprR7
zsoLcasjOQ4+6o8uhdaBt3ZQMatnfcasxCc2G0Ik1snbColIjQzUYVBVooa3
Ol+l5wLgjS06+SbQYAd0N5VAUrcGmrYUY8QmAg.r+A2lyeEGgyxHL9QxDURF
7kpGz+u+VdDU30y20h2Q4NJS2agBwmW409ujyT5ijNZSEVOenzQc6Mhg7eP9
flpT
-----------end_max5_patcher-----------
</code></pre>
| priority | ossia remote outputs value twice or more when value is changed from the bound parameter ossia remote outputs value twice by both and outlets if value is changed by the inlet of the remote itself or from another remote value is output times if parameter remote repetitions attribute is set to same issue this attribute seems to only filter repetition on the outlet of ossia parameter begin patcher um whjv j ljw i t cqodhah wcfk kocgqwwyccmsxp zjbm xh fzoc ejojbayvdil ismaimhrssma nkq ol r u flpt end patcher | 1 |
254,367 | 8,073,245,661 | IssuesEvent | 2018-08-06 18:36:32 | MusaMahmood/Android-HighSamplingRate-ECG | https://api.github.com/repos/MusaMahmood/Android-HighSamplingRate-ECG | opened | Combine all similar apps into single repository! | <HIGH PRIORITY> | Includes:
- https://github.com/MusaMahmood/Android-HighSamplingRate-ECG
- https://github.com/MusaMahmood/portable_ECG_Analysis (subdirectory)
- https://github.com/MusaMahmood/Android-SSVEP-Interface
- https://github.com/MusaMahmood/Android-EMG-DroneDemo
- https://github.com/MusaMahmood/Android-BLE-Motion-Control-RC-Car
- https://github.com/MusaMahmood/Android-BLE-Library-Kotlin-Sample
- https://github.com/MusaMahmood/Android-BLE-Library-Kotlin (??) | 1.0 | Combine all similar apps into single repository! - Includes:
- https://github.com/MusaMahmood/Android-HighSamplingRate-ECG
- https://github.com/MusaMahmood/portable_ECG_Analysis (subdirectory)
- https://github.com/MusaMahmood/Android-SSVEP-Interface
- https://github.com/MusaMahmood/Android-EMG-DroneDemo
- https://github.com/MusaMahmood/Android-BLE-Motion-Control-RC-Car
- https://github.com/MusaMahmood/Android-BLE-Library-Kotlin-Sample
- https://github.com/MusaMahmood/Android-BLE-Library-Kotlin (??) | priority | combine all similar apps into single repository includes subdirectory | 1 |
52,622 | 3,025,324,482 | IssuesEvent | 2015-08-03 07:43:12 | OpenSRP/opensrp-client | https://api.github.com/repos/OpenSRP/opensrp-client | closed | faulty logic on the opensrp ELCO forms | bug High Priority | @raihan-mpower, it seems the new .apk has faulty form logic around the questions asked to the MWRA since all my women have the tag: NOT In ELCO, see the screenshot below:

This has happened with the latest .apk, never observed it before. Also, none of the submission get couchDB and by extension OpenMRS, please look into this?
@julkarnain @sohelsarder @dimasciput @numeranazneen | 1.0 | faulty logic on the opensrp ELCO forms - @raihan-mpower, it seems the new .apk has faulty form logic around the questions asked to the MWRA since all my women have the tag: NOT In ELCO, see the screenshot below:

This has happened with the latest .apk, never observed it before. Also, none of the submission get couchDB and by extension OpenMRS, please look into this?
@julkarnain @sohelsarder @dimasciput @numeranazneen | priority | faulty logic on the opensrp elco forms raihan mpower it seems the new apk has faulty form logic around the questions asked to the mwra since all my women have the tag not in elco see the screenshot below this has happened with the latest apk never observed it before also none of the submission get couchdb and by extension openmrs please look into this julkarnain sohelsarder dimasciput numeranazneen | 1 |
793,533 | 28,001,264,310 | IssuesEvent | 2023-03-27 11:57:00 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | closed | Special characters in resource path caused issues in action invocation | Priority/High Type/Bug module/http Team/PCM Reason/Invalid | **Description:**
$subject refer to the sample code below.
**Steps to reproduce:**
```ballerina
import ballerina/http;
import ballerina/io;
public function main() returns error? {
http:Client openWeather = check new("https://api.openweathermap.org");
json|error res = openWeather->get("/data/2.5/forecast?lat=30.266666&lon=-97.733330&appid=<appid>");
io:println(res); //This works
io:println("1---------------");
json|error res2 = openWeather->/data/["2.5"]/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>");
io:println(res2);//This works
io:println("2---------------");//Following does not work
json|error res3 = openWeather->/data/\2\.5/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>");
io:println(res3);
}
```
Note : I used a free version of https://openweathermap.org/api to generate tokens.
Request hangs and it gives the following error.
```sh
error PayloadBindingClientError ("http:ApplicationResponseError creation failed: 404 response payload extraction failed",error GenericClientError ("Error occurred while retrieving the json payload from the response",error ParserError ("Error occurred while extracting content from message : Idle timeout triggered while reading inbound response headers")))
```
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Special characters in resource path caused issues in action invocation - **Description:**
$subject refer to the sample code below.
**Steps to reproduce:**
```ballerina
import ballerina/http;
import ballerina/io;
public function main() returns error? {
http:Client openWeather = check new("https://api.openweathermap.org");
json|error res = openWeather->get("/data/2.5/forecast?lat=30.266666&lon=-97.733330&appid=<appid>");
io:println(res); //This works
io:println("1---------------");
json|error res2 = openWeather->/data/["2.5"]/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>");
io:println(res2);//This works
io:println("2---------------");//Following does not work
json|error res3 = openWeather->/data/\2\.5/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>");
io:println(res3);
}
```
Note : I used a free version of https://openweathermap.org/api to generate tokens.
Request hangs and it gives the following error.
```sh
error PayloadBindingClientError ("http:ApplicationResponseError creation failed: 404 response payload extraction failed",error GenericClientError ("Error occurred while retrieving the json payload from the response",error ParserError ("Error occurred while extracting content from message : Idle timeout triggered while reading inbound response headers")))
```
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | special characters in resource path caused issues in action invocation description subject refer to the sample code below steps to reproduce ballerina import ballerina http import ballerina io public function main returns error http client openweather check new json error res openweather get data forecast lat lon appid io println res this works io println json error openweather data forecast get lat lon appid io println this works io println following does not work json error openweather data forecast get lat lon appid io println note i used a free version of to generate tokens request hangs and it gives the following error sh error payloadbindingclienterror http applicationresponseerror creation failed response payload extraction failed error genericclienterror error occurred while retrieving the json payload from the response error parsererror error occurred while extracting content from message idle timeout triggered while reading inbound response headers affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 1 |
438,554 | 12,641,128,364 | IssuesEvent | 2020-06-16 05:17:17 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | upgrade-zulip-from-git: Shut down server before upgrade if RAM minimal | area: production installer help wanted priority: high | Currently, if you run `upgrade-zulip-from-git` on a system with Zulip's minimum of 2GB of RAM, you're at risk of having it fail because 2GB of RAM is insufficient to both run the Zulip server and build static assets with webpack; the result is almost always an OOM kill.
This accounts for a significant portion of all reported problems upgrading Zulip servers, so is pretty important. We can fix this by having `upgrade-zulip-from-git` check the total RAM on the system, and if it is low (say, <3GB?), shut down the Zulip server early, before building static assets (maybe printing a warning to make clear what's happening).
I think the code for this belongs just before running `update-prod-static` in the `--from-git` part of `upgrade-zulip-stage-2`; it seems the failures happen just in that section, not the Python part.
This should be pretty easy to test -- just create a VM with 2GB of RAM and run `upgrade-zulip-from-git` until it works.
See for example #14361.
| 1.0 | upgrade-zulip-from-git: Shut down server before upgrade if RAM minimal - Currently, if you run `upgrade-zulip-from-git` on a system with Zulip's minimum of 2GB of RAM, you're at risk of having it fail because 2GB of RAM is insufficient to both run the Zulip server and build static assets with webpack; the result is almost always an OOM kill.
This accounts for a significant portion of all reported problems upgrading Zulip servers, so is pretty important. We can fix this by having `upgrade-zulip-from-git` check the total RAM on the system, and if it is low (say, <3GB?), shut down the Zulip server early, before building static assets (maybe printing a warning to make clear what's happening).
I think the code for this belongs just before running `update-prod-static` in the `--from-git` part of `upgrade-zulip-stage-2`; it seems the failures happen just in that section, not the Python part.
This should be pretty easy to test -- just create a VM with 2GB of RAM and run `upgrade-zulip-from-git` until it works.
See for example #14361.
| priority | upgrade zulip from git shut down server before upgrade if ram minimal currently if you run upgrade zulip from git on a system with zulip s minimum of of ram you re at risk of having it fail because of ram is insufficient to both run the zulip server and build static assets with webpack the result is almost always an oom kill this accounts for a significant portion of all reported problems upgrading zulip servers so is pretty important we can fix this by having upgrade zulip from git check the total ram on the system and if it is low say shut down the zulip server early before building static assets maybe printing a warning to make clear what s happening i think the code for this belongs just before running update prod static in the from git part of upgrade zulip stage it seems the failures happen just in that section not the python part this should be pretty easy to test just create a vm with of ram and run upgrade zulip from git until it works see for example | 1 |
708,049 | 24,328,473,456 | IssuesEvent | 2022-09-30 16:57:26 | hdmf-dev/hdmf-zarr | https://api.github.com/repos/hdmf-dev/hdmf-zarr | closed | Create release on Conda | priority: high | The Conda release is not a must but would be nice to have and will be good to do as learning experience | 1.0 | Create release on Conda - The Conda release is not a must but would be nice to have and will be good to do as learning experience | priority | create release on conda the conda release is not a must but would be nice to have and will be good to do as learning experience | 1 |
563,392 | 16,682,468,812 | IssuesEvent | 2021-06-08 02:44:39 | nlpsandbox/nlpsandbox-client | https://api.github.com/repos/nlpsandbox/nlpsandbox-client | closed | Add scoring code (evaluation) support for id and contact annotators | Priority: High | ### Is your proposal related to a problem?
Contact, Id, covid symptom annotators needs to be added to the evaluation code
- [x] id
- [x] contact | 1.0 | Add scoring code (evaluation) support for id and contact annotators - ### Is your proposal related to a problem?
Contact, Id, covid symptom annotators needs to be added to the evaluation code
- [x] id
- [x] contact | priority | add scoring code evaluation support for id and contact annotators is your proposal related to a problem contact id covid symptom annotators needs to be added to the evaluation code id contact | 1 |
33,362 | 2,764,296,158 | IssuesEvent | 2015-04-29 14:48:39 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | TaskArchiver subscription finishing logic "deadlocking" with multiple requests with the same input | High Priority WMAgent | In an agent we have 2 request reading on the same input dataset on the first step (it's TaskChain), now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running.
This is because the second part of the query which checks for parentage in l. 57 of GetAndMarkFinishedSubscriptions, this was added to prevent deletion of parents when shared between requests due to race conditions.
It was just in the code for the PromptSkimming case with Tier1PromptReco, for now in the affected agent I am removing that portion of the query while we come up with something better.
| 1.0 | TaskArchiver subscription finishing logic "deadlocking" with multiple requests with the same input - In an agent we have 2 request reading on the same input dataset on the first step (it's TaskChain), now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running.
This is because the second part of the query which checks for parentage in l. 57 of GetAndMarkFinishedSubscriptions, this was added to prevent deletion of parents when shared between requests due to race conditions.
It was just in the code for the PromptSkimming case with Tier1PromptReco, for now in the affected agent I am removing that portion of the query while we come up with something better.
| priority | taskarchiver subscription finishing logic deadlocking with multiple requests with the same input in an agent we have request reading on the same input dataset on the first step it s taskchain now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running this is because the second part of the query which checks for parentage in l of getandmarkfinishedsubscriptions this was added to prevent deletion of parents when shared between requests due to race conditions it was just in the code for the promptskimming case with for now in the affected agent i am removing that portion of the query while we come up with something better | 1 |
73,951 | 3,422,680,959 | IssuesEvent | 2015-12-09 00:23:23 | washingtontrails/vms | https://api.github.com/repos/washingtontrails/vms | opened | VMS: Addition to Location/Directions Tab | Enhancement High Priority Pyramid VMS BUDGET | 
Directly above the Google Map Image the following disclaimer should be added: "Warning: Always use the written directions (above) to get to a trailhead. The location of the pinpoint below is only approximate. Using mapping websites and gps navigation to get to trailheads could mislead you." | 1.0 | VMS: Addition to Location/Directions Tab - 
Directly above the Google Map Image the following disclaimer should be added: "Warning: Always use the written directions (above) to get to a trailhead. The location of the pinpoint below is only approximate. Using mapping websites and gps navigation to get to trailheads could mislead you." | priority | vms addition to location directions tab directly above the google map image the following disclaimer should be added warning always use the written directions above to get to a trailhead the location of the pinpoint below is only approximate using mapping websites and gps navigation to get to trailheads could mislead you | 1 |
825,063 | 31,241,464,088 | IssuesEvent | 2023-08-20 22:54:09 | MirageAegis/Sushi-Bot | https://api.github.com/repos/MirageAegis/Sushi-Bot | closed | Auto shout out and go-live | feature priority: high | # Feature
Automatically shout out streamers across all servers that has auto shout out enabled. This is a controlled way of promoting other streamers in your server!
Additionally, automatically post go-live posts in servers where it's enabled.
## Acceptance Criteria
- [x] Sushi Bot shall allow server administrators to configure auto shout outs.
- [x] Sushi Bot shall shout out users who go live.
- [x] This will only be done in servers that has the feature configured.
- [x] The shout out message may be set by the server administrators.
- [x] Sushi Bot shall allow server administrators to configure auto go-live posts.
- [x] Sushi Bot shall post a go-live message whenever the server owner goes live.
- [x] This will only be done in servers that has the feature configured.
- [x] The shout out message may be set by the server administrators.
| 1.0 | Auto shout out and go-live - # Feature
Automatically shout out streamers across all servers that has auto shout out enabled. This is a controlled way of promoting other streamers in your server!
Additionally, automatically post go-live posts in servers where it's enabled.
## Acceptance Criteria
- [x] Sushi Bot shall allow server administrators to configure auto shout outs.
- [x] Sushi Bot shall shout out users who go live.
- [x] This will only be done in servers that has the feature configured.
- [x] The shout out message may be set by the server administrators.
- [x] Sushi Bot shall allow server administrators to configure auto go-live posts.
- [x] Sushi Bot shall post a go-live message whenever the server owner goes live.
- [x] This will only be done in servers that has the feature configured.
- [x] The shout out message may be set by the server administrators.
| priority | auto shout out and go live feature automatically shout out streamers across all servers that has auto shout out enabled this is a controlled way of promoting other streamers in your server additionally automatically post go live posts in servers where it s enabled acceptance criteria sushi bot shall allow server administrators to configure auto shout outs sushi bot shall shout out users who go live this will only be done in servers that has the feature configured the shout out message may be set by the server administrators sushi bot shall allow server administrators to configure auto go live posts sushi bot shall post a go live message whenever the server owner goes live this will only be done in servers that has the feature configured the shout out message may be set by the server administrators | 1 |
510,482 | 14,791,317,621 | IssuesEvent | 2021-01-12 13:20:21 | blchelle/collabogreat | https://api.github.com/repos/blchelle/collabogreat | closed | Loading animation doesn't stop on successful send invite | Priority: High Status: In Progress Type: Bug | ## Description
When a project member sends an invite to another project member, if the invitation sends successfully the loading animation on the button doesn't stop | 1.0 | Loading animation doesn't stop on successful send invite - ## Description
When a project member sends an invite to another project member, if the invitation sends successfully the loading animation on the button doesn't stop | priority | loading animation doesn t stop on successful send invite description when a project member sends an invite to another project member if the invitation sends successfully the loading animation on the button doesn t stop | 1 |
699,651 | 24,025,790,441 | IssuesEvent | 2022-09-15 11:23:51 | CLOSER-Cohorts/archivist | https://api.github.com/repos/CLOSER-Cohorts/archivist | closed | REACT: questions in condition Else branch and loops are not showing in map view and can't be mapped via .txt files | bug High priority react | For example ncds_81_i question qc_currentjob_17_a_B is in the Else branch of c_currentjob_q17cA. https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/
However this question does not appear in the map view https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/map/
This is also true for questions in loops e.g. qc_familyincome_2_a_i which is inside l_familyincome_q2a_TEST
It will also not load mappings for these questions https://closer-archivist-staging.herokuapp.com/admin/instruments/1989/imports/546 | 1.0 | REACT: questions in condition Else branch and loops are not showing in map view and can't be mapped via .txt files - For example ncds_81_i question qc_currentjob_17_a_B is in the Else branch of c_currentjob_q17cA. https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/
However this question does not appear in the map view https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/map/
This is also true for questions in loops e.g. qc_familyincome_2_a_i which is inside l_familyincome_q2a_TEST
It will also not load mappings for these questions https://closer-archivist-staging.herokuapp.com/admin/instruments/1989/imports/546 | priority | react questions in condition else branch and loops are not showing in map view and can t be mapped via txt files for example ncds i question qc currentjob a b is in the else branch of c currentjob however this question does not appear in the map view this is also true for questions in loops e g qc familyincome a i which is inside l familyincome test it will also not load mappings for these questions | 1 |
211,113 | 7,198,487,847 | IssuesEvent | 2018-02-05 13:01:15 | hpi-swt2/sport-portal | https://api.github.com/repos/hpi-swt2/sport-portal | closed | Unlogged user can't see team members on any team page | epic team management po-review priority high team kickerz | Table with team members on any team page should be invisible for unsigned users.
- Go to `Teams` page as unlogged user
- Open any team
- You can see team members, though you should not

| 1.0 | Unlogged user can't see team members on any team page - Table with team members on any team page should be invisible for unsigned users.
- Go to `Teams` page as unlogged user
- Open any team
- You can see team members, though you should not

| priority | unlogged user can t see team members on any team page table with team members on any team page should be invisible for unsigned users go to teams page as unlogged user open any team you can see team members though you should not | 1 |
822,307 | 30,864,530,512 | IssuesEvent | 2023-08-03 07:03:30 | doubleopen-project/dos | https://api.github.com/repos/doubleopen-project/dos | closed | [DO-109] Stress test issues | bug High priority | This is a top level issue to mark down all issues which have arisen from stress testing DOS with the DOS software itself.
<sub>From [SyncLinear.com](https://synclinear.com) | [DO-109](https://linear.app/double-open/issue/DO-109/stress-test-issues)</sub> | 1.0 | [DO-109] Stress test issues - This is a top level issue to mark down all issues which have arisen from stress testing DOS with the DOS software itself.
<sub>From [SyncLinear.com](https://synclinear.com) | [DO-109](https://linear.app/double-open/issue/DO-109/stress-test-issues)</sub> | priority | stress test issues this is a top level issue to mark down all issues which have arisen from stress testing dos with the dos software itself from | 1 |
212,441 | 7,237,130,143 | IssuesEvent | 2018-02-13 09:57:46 | cyberFund/cyber-design | https://api.github.com/repos/cyberFund/cyber-design | closed | Setup design workflow | Priority: High Status: Completed Type: Analytics | In https://github.com/cyberFund/cyber-ui/issues/37#issuecomment-361559171 we've decided that concepts (images, tables, drawings) are added to `cyber-design` repo, where they are processed and after mockups (psd, sketch) and finalized. After that, @Olga584 creates an issue in cyber-browser repository that new design is ready and @Pechalka converts mockups to layouts (html+js). | 1.0 | Setup design workflow - In https://github.com/cyberFund/cyber-ui/issues/37#issuecomment-361559171 we've decided that concepts (images, tables, drawings) are added to `cyber-design` repo, where they are processed and after mockups (psd, sketch) and finalized. After that, @Olga584 creates an issue in cyber-browser repository that new design is ready and @Pechalka converts mockups to layouts (html+js). | priority | setup design workflow in we ve decided that concepts images tables drawings are added to cyber design repo where they are processed and after mockups psd sketch and finalized after that creates an issue in cyber browser repository that new design is ready and pechalka converts mockups to layouts html js | 1 |
281,319 | 8,693,752,207 | IssuesEvent | 2018-12-04 10:37:03 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | combineGenotypes does not store strain name or taxon ID | PHI-Canto bug high priority | (Sub-issue of #1700)
The function `combineGenotypes` in `genotypeListViewCtrl` makes a call to `CursGenotypeList.storeGenotype`, but doesn't specify values for the last two arguments `taxonid` and `strain_name`.
https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L4689-L4690
Here's the full list of expected arguments, for reference:
https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L382-L383
@kimrutherford can this be fixed just by passing in the strain of one of the genotypes in `$scope.genotypeList?` A lot of the existing code seems to be using `$.map` and `$.grep` to collect values from every genotype, but since we're planning to enforce the strain being equal for every allele in a multi-allele genotype, we shouldn't need to check them all.
Now that I understand that this strain equality check should probably be in `combineGenotypes`, I might as well fix all these issues at once. I've opened an issue for this at #1704.
Also, shouldn't we be passing in the taxon ID as well? | 1.0 | combineGenotypes does not store strain name or taxon ID - (Sub-issue of #1700)
The function `combineGenotypes` in `genotypeListViewCtrl` makes a call to `CursGenotypeList.storeGenotype`, but doesn't specify values for the last two arguments `taxonid` and `strain_name`.
https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L4689-L4690
Here's the full list of expected arguments, for reference:
https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L382-L383
@kimrutherford can this be fixed just by passing in the strain of one of the genotypes in `$scope.genotypeList?` A lot of the existing code seems to be using `$.map` and `$.grep` to collect values from every genotype, but since we're planning to enforce the strain being equal for every allele in a multi-allele genotype, we shouldn't need to check them all.
Now that I understand that this strain equality check should probably be in `combineGenotypes`, I might as well fix all these issues at once. I've opened an issue for this at #1704.
Also, shouldn't we be passing in the taxon ID as well? | priority | combinegenotypes does not store strain name or taxon id sub issue of the function combinegenotypes in genotypelistviewctrl makes a call to cursgenotypelist storegenotype but doesn t specify values for the last two arguments taxonid and strain name here s the full list of expected arguments for reference kimrutherford can this be fixed just by passing in the strain of one of the genotypes in scope genotypelist a lot of the existing code seems to be using map and grep to collect values from every genotype but since we re planning to enforce the strain being equal for every allele in a multi allele genotype we shouldn t need to check them all now that i understand that this strain equality check should probably be in combinegenotypes i might as well fix all these issues at once i ve opened an issue for this at also shouldn t we be passing in the taxon id as well | 1 |
612,644 | 19,027,475,083 | IssuesEvent | 2021-11-24 06:34:42 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Creating Same-Name API after deleting the original one casuses gateway artifact retrieval API to fail on MySQL. | Type/Bug Priority/High WUM Affected/3.2.0 | Related Test Case [GatewayRestAPITestCase](https://github.com/wso2-support/product-apim/blob/support-3.2.0.x-full/modules/integration/tests-integration/tests-backend/src/test/java/org/wso2/am/integration/tests/restapi/GatewayRestAPITestCase.java)
GatewayRestAPITestCase is failing when it is run against MySQL DB. Following is the steps to replicate it in a local environment
**Steps to reproduce**
1. Setup APIM 3.2 with MySQL DB
2. Create API and publish. (Say TestAPI)
3. Delete this API and again create the API with the same name (TestAPI)
4. Invoke /gateway/v1/api-artifact API (ex: https://localhost:9443/api/am/gateway/v1/api-artifact?apiName=TestAPI&version=1.0.0&tenantDomain=carbon.super)
Above invocation would return in 400 status code with `TestAPI is not deployed in the Gateway` message.
**Reason**.
When /api-artifac is called, it calls SynapseAttributesApiServiceImpl[1] service impl and here it queries for the APIId using apiname, version and tenant domain from the AM_GW_PUBLISHED_API_DETAILS table [2]. When an API is deleted, entries are not removed from AM_GW_PUBLISHED_API_DETAILS table. As a result, when same-name APIs are created after deleting the first one, there are two entries AM_GW_PUBLISHED_API_DETAILS in this table. In MySQL scenario, when apiId is retrieved from[2], it returns the deleted API's one. This behavior is dependant on the database (This test is passing in H2)
**Fix**
Remove old entries from the AM_GW_PUBLISHED_API_DETAILS table (and check whether we could remove entries from AM_GW_API_ARTIFACTS as well)
[1] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.internal.service/src/main/java/org/wso2/carbon/apimgt/internal/service/impl/SynapseAttributesApiServiceImpl.java#L29
[2] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/GatewayArtifactsMgtDAO.java#L287 | 1.0 | Creating Same-Name API after deleting the original one casuses gateway artifact retrieval API to fail on MySQL. - Related Test Case [GatewayRestAPITestCase](https://github.com/wso2-support/product-apim/blob/support-3.2.0.x-full/modules/integration/tests-integration/tests-backend/src/test/java/org/wso2/am/integration/tests/restapi/GatewayRestAPITestCase.java)
GatewayRestAPITestCase is failing when it is run against MySQL DB. Following is the steps to replicate it in a local environment
**Steps to reproduce**
1. Setup APIM 3.2 with MySQL DB
2. Create API and publish. (Say TestAPI)
3. Delete this API and again create the API with the same name (TestAPI)
4. Invoke /gateway/v1/api-artifact API (ex: https://localhost:9443/api/am/gateway/v1/api-artifact?apiName=TestAPI&version=1.0.0&tenantDomain=carbon.super)
Above invocation would return in 400 status code with `TestAPI is not deployed in the Gateway` message.
**Reason**.
When /api-artifac is called, it calls SynapseAttributesApiServiceImpl[1] service impl and here it queries for the APIId using apiname, version and tenant domain from the AM_GW_PUBLISHED_API_DETAILS table [2]. When an API is deleted, entries are not removed from AM_GW_PUBLISHED_API_DETAILS table. As a result, when same-name APIs are created after deleting the first one, there are two entries AM_GW_PUBLISHED_API_DETAILS in this table. In MySQL scenario, when apiId is retrieved from[2], it returns the deleted API's one. This behavior is dependant on the database (This test is passing in H2)
**Fix**
Remove old entries from the AM_GW_PUBLISHED_API_DETAILS table (and check whether we could remove entries from AM_GW_API_ARTIFACTS as well)
[1] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.internal.service/src/main/java/org/wso2/carbon/apimgt/internal/service/impl/SynapseAttributesApiServiceImpl.java#L29
[2] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/GatewayArtifactsMgtDAO.java#L287 | priority | creating same name api after deleting the original one casuses gateway artifact retrieval api to fail on mysql related test case gatewayrestapitestcase is failing when it is run against mysql db following is the steps to replicate it in a local environment steps to reproduce setup apim with mysql db create api and publish say testapi delete this api and again create the api with the same name testapi invoke gateway api artifact api ex above invocation would return in status code with testapi is not deployed in the gateway message reason when api artifac is called it calls synapseattributesapiserviceimpl service impl and here it queries for the apiid using apiname version and tenant domain from the am gw published api details table when an api is deleted entries are not removed from am gw published api details table as a result when same name apis are created after deleting the first one there are two entries am gw published api details in this table in mysql scenario when apiid is retrieved from it returns the deleted api s one this behavior is dependant on the database this test is passing in fix remove old entries from the am gw published api details table and check whether we could remove entries from am gw api artifacts as well | 1 |
433,947 | 12,512,765,973 | IssuesEvent | 2020-06-02 23:47:43 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | kernel: k_queue_get return NULL before timeout | area: Kernel bug priority: high | **Describe the bug**
The API call `k_queue_get` return no data element before timeout has expired.
Calling either `k_queue_get(queue, K_FOREVER);` or `k_queue_get(queue, K_SECONDS(20))` return `NULL` within one second.
This appears to occur because two threads are both waiting for an element in the queue.
Once an element is posted to the queue, both threads are woken, one will retrieve the new element, while the other will discover an empty queue.
This is a regression from: 7832738ae985a63febb8f82e7c4e34824f48486e
Specifically this change:
```
-static void *k_queue_poll(struct k_queue *queue, s32_t timeout)
+static void *k_queue_poll(struct k_queue *queue, k_timeout_t timeout)
{
struct k_poll_event event;
- int err, elapsed = 0, done = 0;
+ int err;
k_spinlock_key_t key;
void *val;
- u32_t start;
k_poll_event_init(&event, K_POLL_TYPE_FIFO_DATA_AVAILABLE,
K_POLL_MODE_NOTIFY_ONLY, queue);
- if (timeout != K_FOREVER) {
- start = k_uptime_get_32();
- }
-
- do {
- event.state = K_POLL_STATE_NOT_READY;
-
- err = k_poll(&event, 1, timeout - elapsed);
-
- if (err && err != -EAGAIN) {
- return NULL;
- }
+ event.state = K_POLL_STATE_NOT_READY;
+ err = k_poll(&event, 1, timeout);
- key = k_spin_lock(&queue->lock);
- val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true);
- k_spin_unlock(&queue->lock, key);
+ if (err && err != -EAGAIN) {
+ return NULL;
+ }
- if ((val == NULL) && (timeout != K_FOREVER)) {
- elapsed = k_uptime_get_32() - start;
- done = elapsed > timeout;
- }
- } while (!val && !done);
+ key = k_spin_lock(&queue->lock);
+ val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true);
+ k_spin_unlock(&queue->lock, key);
return val;
```
Reverting this change (using legacy timeout API) fixes the issue.
The commit message says this:
```
Also in queue.c, a (when POLL was
enabled) a similar loop was needlessly used to try to retry the
k_poll() call after a spurious failure. But k_poll() does not fail
spuriously, so the loop was removed.
```
The loop appears to have been removed for the wrong reason.
**To Reproduce**
The current steps involves 2 nRF52 dev-kits and a few manual steps.
I can try to make a more minimal failing test if needed, otherwise I can verify using my current setup.
Use branch: https://github.com/joerchan/zephyr/tree/bt-recv-deadlock-debug
```
export BOARD=nrf52840dk
west build test/bluetooth/shell
west flash --erase
west build samples/bluetooth/peripheral
west flash --erase
# Attach one terminal to each
# shell terminal:
bt init
bt connect <id addr>
bt gatt exchange-mtu
```
**Console output**
peripheral sample:
```
Advertising successfully started
Connected: 4e:1a:6f:9b:b8:00 (random)
Write: count= 122, len= 2440, rate= 19523 bps.
Write: count= 1266, len= 25320, rate= 202578 bps.
Write: count= 1815, len= 36300, rate= 290408 bps.
Write: count= 1768, len= 35360, rate= 282880 bps.
Write cmd failed (-55).
Write: count= 1222, len= 37964, rate= 303758 bps.
[00:00:13.090,026] <err> bt_conn: k_is_in_isr 0
[00:00:13.090,057] <err> bt_conn: conn_tx_alloc timeout 0 secs.
[00:00:13.090,087] <err> bt_conn: Unable to allocate TX context
Write: count= 603, len= 37386, rate= 299444 bps.
LE conn param updated: int 0x0028 lat 0 to 42
Write: count= 529, len= 32798, rate= 262680 bps.
Write: count= 594, len= 36828, rate= 294957 bps.
Write: count= 654, len= 40548, rate= 324750 bps.
Write: count= 671, len= 41602, rate= 333049 bps.
Disconnected (reason 0x13)
Write cmd failed (-57).
```
**Expected behavior**
k_queue_get should not return NULL before timeout has passed.
**Impact**
This breaks the current flow-control behavior of Bluetooth, the attempted k_queue_get from BT RX thread will drop the attempt to answer the ATT request and will result in a disconnected ATT channel.
**Additional context**
CONFIG_POLL is enable
Blocker to fix: https://github.com/zephyrproject-rtos/zephyr/issues/23364
| 1.0 | kernel: k_queue_get return NULL before timeout - **Describe the bug**
The API call `k_queue_get` return no data element before timeout has expired.
Calling either `k_queue_get(queue, K_FOREVER);` or `k_queue_get(queue, K_SECONDS(20))` return `NULL` within one second.
This appears to occur because two threads are both waiting for an element in the queue.
Once an element is posted to the queue, both threads are woken, one will retrieve the new element, while the other will discover an empty queue.
This is a regression from: 7832738ae985a63febb8f82e7c4e34824f48486e
Specifically this change:
```
-static void *k_queue_poll(struct k_queue *queue, s32_t timeout)
+static void *k_queue_poll(struct k_queue *queue, k_timeout_t timeout)
{
struct k_poll_event event;
- int err, elapsed = 0, done = 0;
+ int err;
k_spinlock_key_t key;
void *val;
- u32_t start;
k_poll_event_init(&event, K_POLL_TYPE_FIFO_DATA_AVAILABLE,
K_POLL_MODE_NOTIFY_ONLY, queue);
- if (timeout != K_FOREVER) {
- start = k_uptime_get_32();
- }
-
- do {
- event.state = K_POLL_STATE_NOT_READY;
-
- err = k_poll(&event, 1, timeout - elapsed);
-
- if (err && err != -EAGAIN) {
- return NULL;
- }
+ event.state = K_POLL_STATE_NOT_READY;
+ err = k_poll(&event, 1, timeout);
- key = k_spin_lock(&queue->lock);
- val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true);
- k_spin_unlock(&queue->lock, key);
+ if (err && err != -EAGAIN) {
+ return NULL;
+ }
- if ((val == NULL) && (timeout != K_FOREVER)) {
- elapsed = k_uptime_get_32() - start;
- done = elapsed > timeout;
- }
- } while (!val && !done);
+ key = k_spin_lock(&queue->lock);
+ val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true);
+ k_spin_unlock(&queue->lock, key);
return val;
```
Reverting this change (using legacy timeout API) fixes the issue.
The commit message says this:
```
Also in queue.c, a (when POLL was
enabled) a similar loop was needlessly used to try to retry the
k_poll() call after a spurious failure. But k_poll() does not fail
spuriously, so the loop was removed.
```
The loop appears to have been removed for the wrong reason.
**To Reproduce**
The current steps involves 2 nRF52 dev-kits and a few manual steps.
I can try to make a more minimal failing test if needed, otherwise I can verify using my current setup.
Use branch: https://github.com/joerchan/zephyr/tree/bt-recv-deadlock-debug
```
export BOARD=nrf52840dk
west build test/bluetooth/shell
west flash --erase
west build samples/bluetooth/peripheral
west flash --erase
# Attach one terminal to each
# shell terminal:
bt init
bt connect <id addr>
bt gatt exchange-mtu
```
**Console output**
peripheral sample:
```
Advertising successfully started
Connected: 4e:1a:6f:9b:b8:00 (random)
Write: count= 122, len= 2440, rate= 19523 bps.
Write: count= 1266, len= 25320, rate= 202578 bps.
Write: count= 1815, len= 36300, rate= 290408 bps.
Write: count= 1768, len= 35360, rate= 282880 bps.
Write cmd failed (-55).
Write: count= 1222, len= 37964, rate= 303758 bps.
[00:00:13.090,026] <err> bt_conn: k_is_in_isr 0
[00:00:13.090,057] <err> bt_conn: conn_tx_alloc timeout 0 secs.
[00:00:13.090,087] <err> bt_conn: Unable to allocate TX context
Write: count= 603, len= 37386, rate= 299444 bps.
LE conn param updated: int 0x0028 lat 0 to 42
Write: count= 529, len= 32798, rate= 262680 bps.
Write: count= 594, len= 36828, rate= 294957 bps.
Write: count= 654, len= 40548, rate= 324750 bps.
Write: count= 671, len= 41602, rate= 333049 bps.
Disconnected (reason 0x13)
Write cmd failed (-57).
```
**Expected behavior**
k_queue_get should not return NULL before timeout has passed.
**Impact**
This breaks the current flow-control behavior of Bluetooth, the attempted k_queue_get from BT RX thread will drop the attempt to answer the ATT request and will result in a disconnected ATT channel.
**Additional context**
CONFIG_POLL is enable
Blocker to fix: https://github.com/zephyrproject-rtos/zephyr/issues/23364
| priority | kernel k queue get return null before timeout describe the bug the api call k queue get return no data element before timeout has expired calling either k queue get queue k forever or k queue get queue k seconds return null within one second this appears to occur because two threads are both waiting for an element in the queue once an element is posted to the queue both threads are woken one will retrieve the new element while the other will discover an empty queue this is a regression from specifically this change static void k queue poll struct k queue queue t timeout static void k queue poll struct k queue queue k timeout t timeout struct k poll event event int err elapsed done int err k spinlock key t key void val t start k poll event init event k poll type fifo data available k poll mode notify only queue if timeout k forever start k uptime get do event state k poll state not ready err k poll event timeout elapsed if err err eagain return null event state k poll state not ready err k poll event timeout key k spin lock queue lock val z queue node peek sys sflist get queue data q true k spin unlock queue lock key if err err eagain return null if val null timeout k forever elapsed k uptime get start done elapsed timeout while val done key k spin lock queue lock val z queue node peek sys sflist get queue data q true k spin unlock queue lock key return val reverting this change using legacy timeout api fixes the issue the commit message says this also in queue c a when poll was enabled a similar loop was needlessly used to try to retry the k poll call after a spurious failure but k poll does not fail spuriously so the loop was removed the loop appears to have been removed for the wrong reason to reproduce the current steps involves dev kits and a few manual steps i can try to make a more minimal failing test if needed otherwise i can verify using my current setup use branch export board west build test bluetooth shell west flash erase west build samples bluetooth peripheral west flash erase attach one terminal to each shell terminal bt init bt connect bt gatt exchange mtu console output peripheral sample advertising successfully started connected random write count len rate bps write count len rate bps write count len rate bps write count len rate bps write cmd failed write count len rate bps bt conn k is in isr bt conn conn tx alloc timeout secs bt conn unable to allocate tx context write count len rate bps le conn param updated int lat to write count len rate bps write count len rate bps write count len rate bps write count len rate bps disconnected reason write cmd failed expected behavior k queue get should not return null before timeout has passed impact this breaks the current flow control behavior of bluetooth the attempted k queue get from bt rx thread will drop the attempt to answer the att request and will result in a disconnected att channel additional context config poll is enable blocker to fix | 1 |
74,688 | 3,442,288,270 | IssuesEvent | 2015-12-14 21:58:26 | cytoscape/cytoscape.js | https://api.github.com/repos/cytoscape/cytoscape.js | opened | Calculate rendered styles dynamically | 1-high-priority performance renderer | Rather than diffing on each frame to check for a calculation, just use events to determine when to recalculate the rendered style | 1.0 | Calculate rendered styles dynamically - Rather than diffing on each frame to check for a calculation, just use events to determine when to recalculate the rendered style | priority | calculate rendered styles dynamically rather than diffing on each frame to check for a calculation just use events to determine when to recalculate the rendered style | 1 |
656,240 | 21,724,195,225 | IssuesEvent | 2022-05-11 05:39:29 | jordan-sullivan/flashcards-2.5 | https://api.github.com/repos/jordan-sullivan/flashcards-2.5 | opened | Fix last should statement on Turn test | bug high priority | The last 'should' statement won't allow tests to pass , please fix. | 1.0 | Fix last should statement on Turn test - The last 'should' statement won't allow tests to pass , please fix. | priority | fix last should statement on turn test the last should statement won t allow tests to pass please fix | 1 |
339,469 | 10,254,900,796 | IssuesEvent | 2019-08-21 14:26:00 | wso2/docker-apim | https://api.github.com/repos/wso2/docker-apim | opened | Support for artifact synchronization between API Manager nodes | Priority/Highest Type/Improvement | **Description:**
Currently, WSO2 API Manager highly available deployments require us to persist and share runtime artifacts [1]. For this purpose, we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations. Hence, these content need to be persisted in the image.
[1]: [Common Runtime and Configuration Artifacts](https://docs.wso2.com/display/AM260/Common+Runtime+and+Configuration+Artifacts)
**Affected Product Version:**
Docker resources for WSO2 API Manager v2.6.x or below | 1.0 | Support for artifact synchronization between API Manager nodes - **Description:**
Currently, WSO2 API Manager highly available deployments require us to persist and share runtime artifacts [1]. For this purpose, we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations. Hence, these content need to be persisted in the image.
[1]: [Common Runtime and Configuration Artifacts](https://docs.wso2.com/display/AM260/Common+Runtime+and+Configuration+Artifacts)
**Affected Product Version:**
Docker resources for WSO2 API Manager v2.6.x or below | priority | support for artifact synchronization between api manager nodes description currently api manager highly available deployments require us to persist and share runtime artifacts for this purpose we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations hence these content need to be persisted in the image affected product version docker resources for api manager x or below | 1 |
197,837 | 6,965,107,202 | IssuesEvent | 2017-12-09 01:52:20 | FireflyArtsCollective/ffagc | https://api.github.com/repos/FireflyArtsCollective/ffagc | closed | grant submission creation dying when I submit invalid grant dollar amount | bug High Priority | ```
Started POST "/grant_submissions" for 127.0.0.1 at 2017-12-08 20:44:39 -0500
Processing by GrantSubmissionsController#create as HTML
Parameters: {"utf8"=>"✓", "authenticity_token"=>"9h3AOLavJs25WE9Gx496yngHuYz9rv//+4Q6bVM+hOGkzNKcYzGU+RYzPav+bp+3UauHWqnFSKGDjBcsyxGieQ==", "grant_submission"=>{"name"=>"again?", "grant_id"=>"1", "requested_funding_dollars"=>"1004", "proposal"=>#<ActionDispatch::Http::UploadedFile:0x007f13b252f7e8 @tempfile=#<Tempfile:/tmp/RackMultipart20171208-22992-9nwb5u.pdf>, @original_filename="3dmark03_audit_report.pdf", @content_type="application/pdf", @headers="Content-Disposition: form-data; name=\"grant_submission[proposal]\"; filename=\"3dmark03_audit_report.pdf\"\r\nContent-Type: application/pdf\r\n">}, "commit"=>"Create Grant submission"}
Admin Load (0.1ms) SELECT "admins".* FROM "admins" WHERE "admins"."id" = ? LIMIT 1 [["id", nil]]
Artist Load (0.2ms) SELECT "artists".* FROM "artists" WHERE "artists"."id" = ? LIMIT 1 [["id", 3]]
Admin Exists (0.2ms) SELECT 1 AS one FROM "admins" LIMIT 1
(0.1ms) SELECT "grant_submissions".id FROM "grant_submissions" WHERE "grant_submissions"."artist_id" = ? [["artist_id", 3]]
(0.1ms) begin transaction
Grant Load (0.5ms) SELECT "grants".* FROM "grants" WHERE "grants"."id" = ? LIMIT 1 [["id", 1]]
(0.1ms) rollback transaction
Completed 500 Internal Server Error in 28ms (ActiveRecord: 1.2ms)
ActionView::MissingTemplate (Missing template grant_submissions/create, application/create with {:locale=>[:en], :formats=>[:html], :variants=>[], :handlers=>[:erb, :builder, :raw, :ruby, :jbuilder]}. Searched in:
* "/home/owen/src/github/ffagc/app/views"
):
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/path_set.rb:46:in `find'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/lookup_context.rb:121:in `find'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/abstract_renderer.rb:18:in `find_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:40:in `determine_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:8:in `render'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:46:in `render_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:27:in `render'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:100:in `_render_template'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/streaming.rb:217:in `_render_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:83:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:32:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/renderers.rb:37:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/rendering.rb:25:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:16:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block (2 levels) in render'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `block in ms'
/usr/lib/ruby/2.3.0/benchmark.rb:308:in `realtime'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `ms'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block in render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:87:in `cleanup_view_runtime'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:25:in `cleanup_view_runtime'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:43:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:10:in `default_render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:5:in `send_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:198:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:10:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:20:in `block in process_action'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:117:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:555:in `block (2 levels) in compile'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:505:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:92:in `__run_callbacks__'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_process_action_callbacks'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:19:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rescue.rb:29:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:32:in `block in process_action'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `block in instrument'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications/instrumenter.rb:20:in `instrument'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `instrument'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:30:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/params_wrapper.rb:250:in `process_action'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:18:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:137:in `process'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:30:in `process'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:196:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rack_delegation.rb:13:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:237:in `block in action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:74:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:43:in `serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:43:in `block in serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `each'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:817:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/etag.rb:24:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/conditionalget.rb:38:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/head.rb:13:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/params_parser.rb:27:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/flash.rb:260:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:225:in `context'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:220:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/cookies.rb:560:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/query_cache.rb:36:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/connection_adapters/abstract/connection_pool.rb:653:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/migration.rb:377:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:29:in `block in call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:88:in `__run_callbacks__'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_call_callbacks'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:27:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/reloader.rb:73:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/remote_ip.rb:78:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/debug_exceptions.rb:17:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/show_exceptions.rb:30:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:38:in `call_app'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `block in call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `block in tagged'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:26:in `tagged'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `tagged'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/request_id.rb:21:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/methodoverride.rb:22:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/runtime.rb:18:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/cache/strategy/local_cache_middleware.rb:28:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/static.rb:120:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/sendfile.rb:113:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/engine.rb:518:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/application.rb:165:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/content_length.rb:15:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/handler/webrick.rb:88:in `service'
/usr/lib/ruby/2.3.0/webrick/httpserver.rb:140:in `service'
/usr/lib/ruby/2.3.0/webrick/httpserver.rb:96:in `run'
/usr/lib/ruby/2.3.0/webrick/server.rb:296:in `block in start_thread'
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_source.erb (18.9ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_trace.html.erb (3.7ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_request_and_response.html.erb (1.4ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/missing_template.html.erb within rescues/layout (36.8ms)
``` | 1.0 | grant submission creation dying when I submit invalid grant dollar amount - ```
Started POST "/grant_submissions" for 127.0.0.1 at 2017-12-08 20:44:39 -0500
Processing by GrantSubmissionsController#create as HTML
Parameters: {"utf8"=>"✓", "authenticity_token"=>"9h3AOLavJs25WE9Gx496yngHuYz9rv//+4Q6bVM+hOGkzNKcYzGU+RYzPav+bp+3UauHWqnFSKGDjBcsyxGieQ==", "grant_submission"=>{"name"=>"again?", "grant_id"=>"1", "requested_funding_dollars"=>"1004", "proposal"=>#<ActionDispatch::Http::UploadedFile:0x007f13b252f7e8 @tempfile=#<Tempfile:/tmp/RackMultipart20171208-22992-9nwb5u.pdf>, @original_filename="3dmark03_audit_report.pdf", @content_type="application/pdf", @headers="Content-Disposition: form-data; name=\"grant_submission[proposal]\"; filename=\"3dmark03_audit_report.pdf\"\r\nContent-Type: application/pdf\r\n">}, "commit"=>"Create Grant submission"}
Admin Load (0.1ms) SELECT "admins".* FROM "admins" WHERE "admins"."id" = ? LIMIT 1 [["id", nil]]
Artist Load (0.2ms) SELECT "artists".* FROM "artists" WHERE "artists"."id" = ? LIMIT 1 [["id", 3]]
Admin Exists (0.2ms) SELECT 1 AS one FROM "admins" LIMIT 1
(0.1ms) SELECT "grant_submissions".id FROM "grant_submissions" WHERE "grant_submissions"."artist_id" = ? [["artist_id", 3]]
(0.1ms) begin transaction
Grant Load (0.5ms) SELECT "grants".* FROM "grants" WHERE "grants"."id" = ? LIMIT 1 [["id", 1]]
(0.1ms) rollback transaction
Completed 500 Internal Server Error in 28ms (ActiveRecord: 1.2ms)
ActionView::MissingTemplate (Missing template grant_submissions/create, application/create with {:locale=>[:en], :formats=>[:html], :variants=>[], :handlers=>[:erb, :builder, :raw, :ruby, :jbuilder]}. Searched in:
* "/home/owen/src/github/ffagc/app/views"
):
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/path_set.rb:46:in `find'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/lookup_context.rb:121:in `find'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/abstract_renderer.rb:18:in `find_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:40:in `determine_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:8:in `render'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:46:in `render_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:27:in `render'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:100:in `_render_template'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/streaming.rb:217:in `_render_template'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:83:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:32:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/renderers.rb:37:in `render_to_body'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/rendering.rb:25:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:16:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block (2 levels) in render'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `block in ms'
/usr/lib/ruby/2.3.0/benchmark.rb:308:in `realtime'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `ms'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block in render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:87:in `cleanup_view_runtime'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:25:in `cleanup_view_runtime'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:43:in `render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:10:in `default_render'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:5:in `send_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:198:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:10:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:20:in `block in process_action'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:117:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:555:in `block (2 levels) in compile'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:505:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:92:in `__run_callbacks__'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_process_action_callbacks'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:19:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rescue.rb:29:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:32:in `block in process_action'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `block in instrument'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications/instrumenter.rb:20:in `instrument'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `instrument'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:30:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/params_wrapper.rb:250:in `process_action'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:18:in `process_action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:137:in `process'
vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:30:in `process'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:196:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rack_delegation.rb:13:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:237:in `block in action'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:74:in `dispatch'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:43:in `serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:43:in `block in serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `each'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `serve'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:817:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/etag.rb:24:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/conditionalget.rb:38:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/head.rb:13:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/params_parser.rb:27:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/flash.rb:260:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:225:in `context'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:220:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/cookies.rb:560:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/query_cache.rb:36:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/connection_adapters/abstract/connection_pool.rb:653:in `call'
vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/migration.rb:377:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:29:in `block in call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:88:in `__run_callbacks__'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_call_callbacks'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:27:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/reloader.rb:73:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/remote_ip.rb:78:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/debug_exceptions.rb:17:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/show_exceptions.rb:30:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:38:in `call_app'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `block in call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `block in tagged'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:26:in `tagged'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `tagged'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/request_id.rb:21:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/methodoverride.rb:22:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/runtime.rb:18:in `call'
vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/cache/strategy/local_cache_middleware.rb:28:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call'
vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/static.rb:120:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/sendfile.rb:113:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/engine.rb:518:in `call'
vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/application.rb:165:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/content_length.rb:15:in `call'
vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/handler/webrick.rb:88:in `service'
/usr/lib/ruby/2.3.0/webrick/httpserver.rb:140:in `service'
/usr/lib/ruby/2.3.0/webrick/httpserver.rb:96:in `run'
/usr/lib/ruby/2.3.0/webrick/server.rb:296:in `block in start_thread'
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_source.erb (18.9ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_trace.html.erb (3.7ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_request_and_response.html.erb (1.4ms)
Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/missing_template.html.erb within rescues/layout (36.8ms)
``` | priority | grant submission creation dying when i submit invalid grant dollar amount started post grant submissions for at processing by grantsubmissionscontroller create as html parameters ✓ authenticity token hogkznkcyzgu ryzpav bp grant submission name again grant id requested funding dollars proposal original filename audit report pdf content type application pdf headers content disposition form data name grant submission filename audit report pdf r ncontent type application pdf r n commit create grant submission admin load select admins from admins where admins id limit artist load select artists from artists where artists id limit admin exists select as one from admins limit select grant submissions id from grant submissions where grant submissions artist id begin transaction grant load select grants from grants where grants id limit rollback transaction completed internal server error in activerecord actionview missingtemplate missing template grant submissions create application create with locale formats variants handlers searched in home owen src github ffagc app views vendor bundle ruby lib action view path set rb in find vendor bundle ruby lib action view lookup context rb in find vendor bundle ruby lib action view renderer abstract renderer rb in find template vendor bundle ruby lib action view renderer template renderer rb in determine template vendor bundle ruby lib action view renderer template renderer rb in render vendor bundle ruby lib action view renderer renderer rb in render template vendor bundle ruby lib action view renderer renderer rb in render vendor bundle ruby lib action view rendering rb in render template vendor bundle ruby lib action controller metal streaming rb in render template vendor bundle ruby lib action view rendering rb in render to body vendor bundle ruby lib action controller metal rendering rb in render to body vendor bundle ruby lib action controller metal renderers rb in render to body vendor bundle ruby lib abstract controller rendering rb in render vendor bundle ruby lib action controller metal rendering rb in render vendor bundle ruby lib action controller metal instrumentation rb in block levels in render vendor bundle ruby lib active support core ext benchmark rb in block in ms usr lib ruby benchmark rb in realtime vendor bundle ruby lib active support core ext benchmark rb in ms vendor bundle ruby lib action controller metal instrumentation rb in block in render vendor bundle ruby lib action controller metal instrumentation rb in cleanup view runtime vendor bundle ruby lib active record railties controller runtime rb in cleanup view runtime vendor bundle ruby lib action controller metal instrumentation rb in render vendor bundle ruby lib action controller metal implicit render rb in default render vendor bundle ruby lib action controller metal implicit render rb in send action vendor bundle ruby lib abstract controller base rb in process action vendor bundle ruby lib action controller metal rendering rb in process action vendor bundle ruby lib abstract controller callbacks rb in block in process action vendor bundle ruby lib active support callbacks rb in call vendor bundle ruby lib active support callbacks rb in block levels in compile vendor bundle ruby lib active support callbacks rb in call vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib active support callbacks rb in run process action callbacks vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib abstract controller callbacks rb in process action vendor bundle ruby lib action controller metal rescue rb in process action vendor bundle ruby lib action controller metal instrumentation rb in block in process action vendor bundle ruby lib active support notifications rb in block in instrument vendor bundle ruby lib active support notifications instrumenter rb in instrument vendor bundle ruby lib active support notifications rb in instrument vendor bundle ruby lib action controller metal instrumentation rb in process action vendor bundle ruby lib action controller metal params wrapper rb in process action vendor bundle ruby lib active record railties controller runtime rb in process action vendor bundle ruby lib abstract controller base rb in process vendor bundle ruby lib action view rendering rb in process vendor bundle ruby lib action controller metal rb in dispatch vendor bundle ruby lib action controller metal rack delegation rb in dispatch vendor bundle ruby lib action controller metal rb in block in action vendor bundle ruby lib action dispatch routing route set rb in dispatch vendor bundle ruby lib action dispatch routing route set rb in serve vendor bundle ruby lib action dispatch journey router rb in block in serve vendor bundle ruby lib action dispatch journey router rb in each vendor bundle ruby lib action dispatch journey router rb in serve vendor bundle ruby lib action dispatch routing route set rb in call vendor bundle ruby lib rack etag rb in call vendor bundle ruby lib rack conditionalget rb in call vendor bundle ruby lib rack head rb in call vendor bundle ruby lib action dispatch middleware params parser rb in call vendor bundle ruby lib action dispatch middleware flash rb in call vendor bundle ruby lib rack session abstract id rb in context vendor bundle ruby lib rack session abstract id rb in call vendor bundle ruby lib action dispatch middleware cookies rb in call vendor bundle ruby lib active record query cache rb in call vendor bundle ruby lib active record connection adapters abstract connection pool rb in call vendor bundle ruby lib active record migration rb in call vendor bundle ruby lib action dispatch middleware callbacks rb in block in call vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib active support callbacks rb in run call callbacks vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib action dispatch middleware callbacks rb in call vendor bundle ruby lib action dispatch middleware reloader rb in call vendor bundle ruby lib action dispatch middleware remote ip rb in call vendor bundle ruby lib action dispatch middleware debug exceptions rb in call vendor bundle ruby lib action dispatch middleware show exceptions rb in call vendor bundle ruby lib rails rack logger rb in call app vendor bundle ruby lib rails rack logger rb in block in call vendor bundle ruby lib active support tagged logging rb in block in tagged vendor bundle ruby lib active support tagged logging rb in tagged vendor bundle ruby lib active support tagged logging rb in tagged vendor bundle ruby lib rails rack logger rb in call vendor bundle ruby lib action dispatch middleware request id rb in call vendor bundle ruby lib rack methodoverride rb in call vendor bundle ruby lib rack runtime rb in call vendor bundle ruby lib active support cache strategy local cache middleware rb in call vendor bundle ruby lib rack lock rb in call vendor bundle ruby lib action dispatch middleware static rb in call vendor bundle ruby lib rack sendfile rb in call vendor bundle ruby lib rails engine rb in call vendor bundle ruby lib rails application rb in call vendor bundle ruby lib rack lock rb in call vendor bundle ruby lib rack content length rb in call vendor bundle ruby lib rack handler webrick rb in service usr lib ruby webrick httpserver rb in service usr lib ruby webrick httpserver rb in run usr lib ruby webrick server rb in block in start thread rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues source erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues trace html erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues request and response html erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues missing template html erb within rescues layout | 1 |
185,929 | 6,732,010,213 | IssuesEvent | 2017-10-18 09:47:29 | ballerinalang/composer | https://api.github.com/repos/ballerinalang/composer | closed | [Firefox] Package declaration doesn't take effect | 0.94-pre-release Priority/Highest Severity/Critical | Browser: Firefox 56.0 (64-bit)
**Steps**
1. Add a package
2. Open split view
3. Reopen the package declaration box after refreshing the page
The package dissapears from the package declaration ui component.

| 1.0 | [Firefox] Package declaration doesn't take effect - Browser: Firefox 56.0 (64-bit)
**Steps**
1. Add a package
2. Open split view
3. Reopen the package declaration box after refreshing the page
The package dissapears from the package declaration ui component.

| priority | package declaration doesn t take effect browser firefox bit steps add a package open split view reopen the package declaration box after refreshing the page the package dissapears from the package declaration ui component | 1 |
710,225 | 24,411,467,636 | IssuesEvent | 2022-10-05 12:45:00 | HiAvatar/backend | https://api.github.com/repos/HiAvatar/backend | closed | 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결 | Type: Bug Priority: High | ### Description
오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다. | 1.0 | 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결 - ### Description
오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다. | priority | 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결 description 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다 | 1 |
62,426 | 3,185,250,908 | IssuesEvent | 2015-09-28 01:56:42 | babel/babel | https://api.github.com/repos/babel/babel | closed | Improve require speed | discussion good first bug help wanted high priority | This is an issue to track possible require optimisations. Not sure if there's anything that can be done but it's worth investigating.
/cc @Globegitter @stefanpenner | 1.0 | Improve require speed - This is an issue to track possible require optimisations. Not sure if there's anything that can be done but it's worth investigating.
/cc @Globegitter @stefanpenner | priority | improve require speed this is an issue to track possible require optimisations not sure if there s anything that can be done but it s worth investigating cc globegitter stefanpenner | 1 |
252,412 | 8,036,068,595 | IssuesEvent | 2018-07-30 07:52:16 | systers/powerup-android | https://api.github.com/repos/systers/powerup-android | opened | Has to select the final answer twice to close the game | Priority: HIGH Program: GSoC Type: Bug | ## Description
As a user, when I'm selecting the answers to questions in different scenarios when I select the last answer, the screen should change to the game points activity.
## Mocks

To recreate:
open any scenario
keep answering questions until you come to the last question
the last answer should be selected twice in order to move to scenario over screen.
This issue is caused by the asynchronous call used in Gameactivity:
<img width="708" alt="screen shot 2018-07-30 at 7 49 29 pm" src="https://user-images.githubusercontent.com/1425395/43384301-b5c576e0-9431-11e8-9dca-4db66e103795.png">
prevScene is checked before the presenter is loaded with the previous screen.
## Acceptance Criteria
### Update [Required]
- [ ] the last answer should be selected only once
- [ ] asynchronous calls should be properly handled.
## Definition of Done
- [ ] All of the required items are completed.
- [ ] Approval by 1 mentor.
## Estimation
2 hours
| 1.0 | Has to select the final answer twice to close the game - ## Description
As a user, when I'm selecting the answers to questions in different scenarios when I select the last answer, the screen should change to the game points activity.
## Mocks

To recreate:
open any scenario
keep answering questions until you come to the last question
the last answer should be selected twice in order to move to scenario over screen.
This issue is caused by the asynchronous call used in Gameactivity:
<img width="708" alt="screen shot 2018-07-30 at 7 49 29 pm" src="https://user-images.githubusercontent.com/1425395/43384301-b5c576e0-9431-11e8-9dca-4db66e103795.png">
prevScene is checked before the presenter is loaded with the previous screen.
## Acceptance Criteria
### Update [Required]
- [ ] the last answer should be selected only once
- [ ] asynchronous calls should be properly handled.
## Definition of Done
- [ ] All of the required items are completed.
- [ ] Approval by 1 mentor.
## Estimation
2 hours
| priority | has to select the final answer twice to close the game description as a user when i m selecting the answers to questions in different scenarios when i select the last answer the screen should change to the game points activity mocks to recreate open any scenario keep answering questions until you come to the last question the last answer should be selected twice in order to move to scenario over screen this issue is caused by the asynchronous call used in gameactivity img width alt screen shot at pm src prevscene is checked before the presenter is loaded with the previous screen acceptance criteria update the last answer should be selected only once asynchronous calls should be properly handled definition of done all of the required items are completed approval by mentor estimation hours | 1 |
579,747 | 17,197,984,544 | IssuesEvent | 2021-07-16 20:43:49 | mikemerin/FM-layouts | https://api.github.com/repos/mikemerin/FM-layouts | closed | Announcement upgrade | Category - Graphics Priority - 1 (high) Type - Upgrade / Fix | Adding to dashboard panel (https://github.com/mikemerin/FM2020-layouts/issues/11) and if another genre, linking to that community | 1.0 | Announcement upgrade - Adding to dashboard panel (https://github.com/mikemerin/FM2020-layouts/issues/11) and if another genre, linking to that community | priority | announcement upgrade adding to dashboard panel and if another genre linking to that community | 1 |
104,329 | 4,209,474,737 | IssuesEvent | 2016-06-29 05:37:38 | japanesemediamanager/jmmserver | https://api.github.com/repos/japanesemediamanager/jmmserver | closed | Group Filters are not working | Bug - High Priority | How I reproduced
- Completely new installation using latest source
- DB : SQL Server
- Import files into new database
- Mark a file as watched from the Metro series view
- Nothing is shown in the 'Continue Watching (System)' group filter even though there are more files to watch
NOTE - Actually nothing is shown in any of the group filters | 1.0 | Group Filters are not working - How I reproduced
- Completely new installation using latest source
- DB : SQL Server
- Import files into new database
- Mark a file as watched from the Metro series view
- Nothing is shown in the 'Continue Watching (System)' group filter even though there are more files to watch
NOTE - Actually nothing is shown in any of the group filters | priority | group filters are not working how i reproduced completely new installation using latest source db sql server import files into new database mark a file as watched from the metro series view nothing is shown in the continue watching system group filter even though there are more files to watch note actually nothing is shown in any of the group filters | 1 |
372,160 | 11,009,917,859 | IssuesEvent | 2019-12-04 13:39:11 | boi123212321/porn-manager | https://api.github.com/repos/boi123212321/porn-manager | closed | Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN | bug high priority | Identified that I have 10 videos in my directory but didn't extract any.
`Generating thumbnails...`
```
Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN
```
Pointed the app at a different folder and created a new 'collection' (clearing the previous video queue) and it worked well.
I suspect when the app doesn't catch the exception when a video doesn't have a correct duration. Probably very easy to fix -- but I'm not up for creating a PR at this moment.
Awesome app! Keep it up 👍 | 1.0 | Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN - Identified that I have 10 videos in my directory but didn't extract any.
`Generating thumbnails...`
```
Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN
```
Pointed the app at a different folder and created a new 'collection' (clearing the previous video queue) and it worked well.
I suspect when the app doesn't catch the exception when a video doesn't have a correct duration. Probably very easy to fix -- but I'm not up for creating a PR at this moment.
Awesome app! Keep it up 👍 | priority | error ffmpeg exited with code invalid duration during specification for ss nan identified that i have videos in my directory but didn t extract any generating thumbnails error ffmpeg exited with code invalid duration during specification for ss nan pointed the app at a different folder and created a new collection clearing the previous video queue and it worked well i suspect when the app doesn t catch the exception when a video doesn t have a correct duration probably very easy to fix but i m not up for creating a pr at this moment awesome app keep it up 👍 | 1 |
106,529 | 4,273,920,676 | IssuesEvent | 2016-07-13 18:49:37 | Octanis1/Octanis1-Mainboard-Firmware_MSP_EXP432P401RLP | https://api.github.com/repos/Octanis1/Octanis1-Mainboard-Firmware_MSP_EXP432P401RLP | closed | Navigation/Driving not correctly working | bug enhancement high priority in progress | rover does not follow correct direction as entered in apm planner. Also, curve radii are too small and give too high mechanical stress. limit difference to 100% vs. 80% motor speed between left and right.
bug must be in navigation.c, since IMU information is correct (x-direction on mainboard = forward) | 1.0 | Navigation/Driving not correctly working - rover does not follow correct direction as entered in apm planner. Also, curve radii are too small and give too high mechanical stress. limit difference to 100% vs. 80% motor speed between left and right.
bug must be in navigation.c, since IMU information is correct (x-direction on mainboard = forward) | priority | navigation driving not correctly working rover does not follow correct direction as entered in apm planner also curve radii are too small and give too high mechanical stress limit difference to vs motor speed between left and right bug must be in navigation c since imu information is correct x direction on mainboard forward | 1 |
674,418 | 23,050,276,972 | IssuesEvent | 2022-07-24 14:21:09 | FTBTeam/FTB-App | https://api.github.com/repos/FTBTeam/FTB-App | closed | [Bug]: locked down to java 17 and fail to start downloaded runtime | bug app installer os/linux priority/high | ### What Operating System
Linux (Arch)
### App Version
202205171618-14abefffbf-release
### UI Version
_No response_
### Log Files
_No response_
### Debug Code
-
### Describe the bug
installer is locked down specific to java 17 and the script fails to run the downloaded runtime
### Steps to reproduce
download current linux "SFX" script (for real - why not just use curl or wget to load the archive but construct some obscure "rar-sfx style" hybrid to not have to serve a binary? pointless)
try to run
note that it downloads a runtime of java 17 because arch has already advanced to java 18
note further an error message that the script can't find the java binary of the just downloaded and extract runtime
### Expected behaviour
just a simple script to scan for ANY version of java by just executing "java" command instead of search in a couple dozen paths and use it
### Screenshots
_No response_
### Additional information
there's no need to fix down to java 17 for the installer - just by its default look and feel it's clearly a swing application without even bother looking into the source
also: why to try to search thru couple of dozen of directories instead of rely on the user has java proper "installed" by the distributions package manager? you surely don't go as far to support gentoo or even LFS to even bother for manual scanning - so why not just rely on "java" be an available command thanks to package manager magic? if it's not installed just give a hint to install java runtime via package manager
this also solves the requirement of need to download a specific runtime (which btw doesn't work cause for whatever reason I not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway) - and anything version specific can also be done within java - eliminating the need for checking specific version by debug output
but - as ftb being ftb I know for years again I had to tinker around to get it even working - as the legacy launcher somehow fails to find the authlib thanks to some NOT SUPPOSED TO USED classloader hacking of the systemclassloader ... man, why noone back then ever read recommendations about how NOT to use sun internals?
as some genius had that bright idea to pack a script and a binary in an incompatible way I had to painly do a hexdump - edit the version detection on a hex level - and rebuild the mess to even get it running
why is FTB so damn bad at these most basics of java programming over and over again?
also: why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available - heck, it's even the basic tools on even gentoo and LFS - and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed - so it can be taken as given same as gzip and tar
oh - btw - why do gunzip and tar in two steps? you do know that tar also supports gzip decompression by either provide -z specific for gzip or just -a for auto detection based on filename
and please - just for the sake - just provide a simple bashscript that just downloads the app pack and runs it with local java runtime - why even the need for "an installer"? linux != windows - there's no need for such crap - or just ship some form of archive that contains everything instead of rely to curl/wget anything afterwards ... oh, wait, right, you do have to cheap out on bandwidth ... then how about use git as mirror?
there's so much wrong with this new "app" crap the same as back with the old .jar launcher - let alone the literally THOUSANDS of ERROR level messages when booting up any of the packs - it's not just infos or warnings - but ERRORS - which, again as the same as the past years, just noone bothers to pay attention
why the F* I used ftb again - right, just to see that immersive petroleum is not up to 1.18 yet ... geez, was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows - instead of just keep everything contained within that one directory I downloaded th script to - as it should be done on linux and with java applications in general
### Information
- [X] I have provided as much information as possible | 1.0 | [Bug]: locked down to java 17 and fail to start downloaded runtime - ### What Operating System
Linux (Arch)
### App Version
202205171618-14abefffbf-release
### UI Version
_No response_
### Log Files
_No response_
### Debug Code
-
### Describe the bug
installer is locked down specific to java 17 and the script fails to run the downloaded runtime
### Steps to reproduce
download current linux "SFX" script (for real - why not just use curl or wget to load the archive but construct some obscure "rar-sfx style" hybrid to not have to serve a binary? pointless)
try to run
note that it downloads a runtime of java 17 because arch has already advanced to java 18
note further an error message that the script can't find the java binary of the just downloaded and extract runtime
### Expected behaviour
just a simple script to scan for ANY version of java by just executing "java" command instead of search in a couple dozen paths and use it
### Screenshots
_No response_
### Additional information
there's no need to fix down to java 17 for the installer - just by its default look and feel it's clearly a swing application without even bother looking into the source
also: why to try to search thru couple of dozen of directories instead of rely on the user has java proper "installed" by the distributions package manager? you surely don't go as far to support gentoo or even LFS to even bother for manual scanning - so why not just rely on "java" be an available command thanks to package manager magic? if it's not installed just give a hint to install java runtime via package manager
this also solves the requirement of need to download a specific runtime (which btw doesn't work cause for whatever reason I not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway) - and anything version specific can also be done within java - eliminating the need for checking specific version by debug output
but - as ftb being ftb I know for years again I had to tinker around to get it even working - as the legacy launcher somehow fails to find the authlib thanks to some NOT SUPPOSED TO USED classloader hacking of the systemclassloader ... man, why noone back then ever read recommendations about how NOT to use sun internals?
as some genius had that bright idea to pack a script and a binary in an incompatible way I had to painly do a hexdump - edit the version detection on a hex level - and rebuild the mess to even get it running
why is FTB so damn bad at these most basics of java programming over and over again?
also: why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available - heck, it's even the basic tools on even gentoo and LFS - and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed - so it can be taken as given same as gzip and tar
oh - btw - why do gunzip and tar in two steps? you do know that tar also supports gzip decompression by either provide -z specific for gzip or just -a for auto detection based on filename
and please - just for the sake - just provide a simple bashscript that just downloads the app pack and runs it with local java runtime - why even the need for "an installer"? linux != windows - there's no need for such crap - or just ship some form of archive that contains everything instead of rely to curl/wget anything afterwards ... oh, wait, right, you do have to cheap out on bandwidth ... then how about use git as mirror?
there's so much wrong with this new "app" crap the same as back with the old .jar launcher - let alone the literally THOUSANDS of ERROR level messages when booting up any of the packs - it's not just infos or warnings - but ERRORS - which, again as the same as the past years, just noone bothers to pay attention
why the F* I used ftb again - right, just to see that immersive petroleum is not up to 1.18 yet ... geez, was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows - instead of just keep everything contained within that one directory I downloaded th script to - as it should be done on linux and with java applications in general
### Information
- [X] I have provided as much information as possible | priority | locked down to java and fail to start downloaded runtime what operating system linux arch app version release ui version no response log files no response debug code describe the bug installer is locked down specific to java and the script fails to run the downloaded runtime steps to reproduce download current linux sfx script for real why not just use curl or wget to load the archive but construct some obscure rar sfx style hybrid to not have to serve a binary pointless try to run note that it downloads a runtime of java because arch has already advanced to java note further an error message that the script can t find the java binary of the just downloaded and extract runtime expected behaviour just a simple script to scan for any version of java by just executing java command instead of search in a couple dozen paths and use it screenshots no response additional information there s no need to fix down to java for the installer just by its default look and feel it s clearly a swing application without even bother looking into the source also why to try to search thru couple of dozen of directories instead of rely on the user has java proper installed by the distributions package manager you surely don t go as far to support gentoo or even lfs to even bother for manual scanning so why not just rely on java be an available command thanks to package manager magic if it s not installed just give a hint to install java runtime via package manager this also solves the requirement of need to download a specific runtime which btw doesn t work cause for whatever reason i not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway and anything version specific can also be done within java eliminating the need for checking specific version by debug output but as ftb being ftb i know for years again i had to tinker around to get it even working as the legacy launcher somehow fails to find the authlib thanks to some not supposed to used classloader hacking of the systemclassloader man why noone back then ever read recommendations about how not to use sun internals as some genius had that bright idea to pack a script and a binary in an incompatible way i had to painly do a hexdump edit the version detection on a hex level and rebuild the mess to even get it running why is ftb so damn bad at these most basics of java programming over and over again also why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available heck it s even the basic tools on even gentoo and lfs and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed so it can be taken as given same as gzip and tar oh btw why do gunzip and tar in two steps you do know that tar also supports gzip decompression by either provide z specific for gzip or just a for auto detection based on filename and please just for the sake just provide a simple bashscript that just downloads the app pack and runs it with local java runtime why even the need for an installer linux windows there s no need for such crap or just ship some form of archive that contains everything instead of rely to curl wget anything afterwards oh wait right you do have to cheap out on bandwidth then how about use git as mirror there s so much wrong with this new app crap the same as back with the old jar launcher let alone the literally thousands of error level messages when booting up any of the packs it s not just infos or warnings but errors which again as the same as the past years just noone bothers to pay attention why the f i used ftb again right just to see that immersive petroleum is not up to yet geez was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows instead of just keep everything contained within that one directory i downloaded th script to as it should be done on linux and with java applications in general information i have provided as much information as possible | 1 |
156,823 | 5,989,636,293 | IssuesEvent | 2017-06-02 09:09:17 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [New xCAT prototype]build ansible playbook to setup xcat3 service | priority:high sprint1 type:feature | Acceptance:
1. git clone xcat3 codes onto hosts, start and configure xcat3 service
2. configure apache and dhcp, tftp, mysql/mariaDB service | 1.0 | [New xCAT prototype]build ansible playbook to setup xcat3 service - Acceptance:
1. git clone xcat3 codes onto hosts, start and configure xcat3 service
2. configure apache and dhcp, tftp, mysql/mariaDB service | priority | build ansible playbook to setup service acceptance git clone codes onto hosts start and configure service configure apache and dhcp tftp mysql mariadb service | 1 |
711,240 | 24,455,103,816 | IssuesEvent | 2022-10-07 05:41:54 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | [Bug]: InvalidPathException during the code coverage generation in Windows | Type/Bug Priority/High Team/CompilerFE Team/jBallerina Points/2 | ### Description
```
[2022-10-04 09:25:07,184] SEVERE {b7a.log.crash} - Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class
java.nio.file.InvalidPathException: Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class
```
Found in HTTP PR build on windows.
```
at org.ballerinalang.test.runtime.util.CodeCoverageUtils.unzipCompiledSource(CodeCoverageUtils.java:116)
at org.ballerinalang.test.runtime.entity.CoverageReport.addCompiledSources(CoverageReport.java:302)
at org.ballerinalang.test.runtime.entity.CoverageReport.generateReport(CoverageReport.java:139)
```
### Steps to Reproduce
Build the HTTP package in windows
Checkout to : https://github.com/chamil321/module-ballerina-http/tree/exit
Cmd :
```
./gradlew clean build -x http-native:test -x :http-ballerina-tests:startLdapServe -x :http-ballerina-tests:stopLdapServer -x :http-compiler-plugin-tests:test -x :http-ballerina-tests:commitTomlFiles -x :http-ballerina:commitTomlFiles
```
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Runtime
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_ | 1.0 | [Bug]: InvalidPathException during the code coverage generation in Windows - ### Description
```
[2022-10-04 09:25:07,184] SEVERE {b7a.log.crash} - Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class
java.nio.file.InvalidPathException: Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class
```
Found in HTTP PR build on windows.
```
at org.ballerinalang.test.runtime.util.CodeCoverageUtils.unzipCompiledSource(CodeCoverageUtils.java:116)
at org.ballerinalang.test.runtime.entity.CoverageReport.addCompiledSources(CoverageReport.java:302)
at org.ballerinalang.test.runtime.entity.CoverageReport.generateReport(CoverageReport.java:139)
```
### Steps to Reproduce
Build the HTTP package in windows
Checkout to : https://github.com/chamil321/module-ballerina-http/tree/exit
Cmd :
```
./gradlew clean build -x http-native:test -x :http-ballerina-tests:startLdapServe -x :http-ballerina-tests:stopLdapServer -x :http-compiler-plugin-tests:test -x :http-ballerina-tests:commitTomlFiles -x :http-ballerina:commitTomlFiles
```
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Runtime
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_ | priority | invalidpathexception during the code coverage generation in windows description severe log crash illegal char at index d a module ballerina http module ballerina http ballerina tests target cache tests cache coverage bin ballerina http value ballerina anontype method delete readonly class java nio file invalidpathexception illegal char at index d a module ballerina http module ballerina http ballerina tests target cache tests cache coverage bin ballerina http value ballerina anontype method delete readonly class found in http pr build on windows at org ballerinalang test runtime util codecoverageutils unzipcompiledsource codecoverageutils java at org ballerinalang test runtime entity coveragereport addcompiledsources coveragereport java at org ballerinalang test runtime entity coveragereport generatereport coveragereport java steps to reproduce build the http package in windows checkout to cmd gradlew clean build x http native test x http ballerina tests startldapserve x http ballerina tests stopldapserver x http compiler plugin tests test x http ballerina tests committomlfiles x http ballerina committomlfiles affected version s no response os db other environment details and versions no response related area runtime related issue s optional no response suggested label s optional no response suggested assignee s optional no response | 1 |
685,448 | 23,456,345,705 | IssuesEvent | 2022-08-16 09:12:30 | nkhanh44/ic-survey-ios | https://api.github.com/repos/nkhanh44/ic-survey-ios | opened | [Chore] setup CD pineline | type : chore @0.4.0 priority: high | ## Why
- No manual execution of releases
## Who Benefits?
- CD pipeline is run on Github Action
- CD pipeline run to build a staging version and upload to Firebase Distribution
| 1.0 | [Chore] setup CD pineline - ## Why
- No manual execution of releases
## Who Benefits?
- CD pipeline is run on Github Action
- CD pipeline run to build a staging version and upload to Firebase Distribution
| priority | setup cd pineline why no manual execution of releases who benefits cd pipeline is run on github action cd pipeline run to build a staging version and upload to firebase distribution | 1 |
694,268 | 23,808,295,849 | IssuesEvent | 2022-09-04 11:33:33 | ctm/mb2-doc | https://api.github.com/repos/ctm/mb2-doc | closed | PLO button issue | wontfix can't reproduce high priority easy request | Look into the PLO button issue reported below.
> The Pot button in PLO is exhibiting weird behavior – I tried to bet pot of 1500 but it only allowed 900. I can’t reproduce yet as that was the last hand of the PLO round. I’ll try to get you more details. I was able to use the slider – didn’t try the Max button.
> If you give me a hand number (which you might be able to still get, by scrolling back), that'll help.
> I am semi-sureish that it’s 284669.
This was reported to me via email. I didn't make an issue out of it at the time and then I forgot.
I'm starting with this as `high priority` and `easy`, because I have a hand number, which will also give me the table and event. I can poke around and see if there were any error messages reported from the server to the client. If there were, then it should be trivial to reproduce and fix. If there are no error messages, I'll have to ask some more questions and it may no longer be easy.
| 1.0 | PLO button issue - Look into the PLO button issue reported below.
> The Pot button in PLO is exhibiting weird behavior – I tried to bet pot of 1500 but it only allowed 900. I can’t reproduce yet as that was the last hand of the PLO round. I’ll try to get you more details. I was able to use the slider – didn’t try the Max button.
> If you give me a hand number (which you might be able to still get, by scrolling back), that'll help.
> I am semi-sureish that it’s 284669.
This was reported to me via email. I didn't make an issue out of it at the time and then I forgot.
I'm starting with this as `high priority` and `easy`, because I have a hand number, which will also give me the table and event. I can poke around and see if there were any error messages reported from the server to the client. If there were, then it should be trivial to reproduce and fix. If there are no error messages, I'll have to ask some more questions and it may no longer be easy.
| priority | plo button issue look into the plo button issue reported below the pot button in plo is exhibiting weird behavior – i tried to bet pot of but it only allowed i can’t reproduce yet as that was the last hand of the plo round i’ll try to get you more details i was able to use the slider – didn’t try the max button if you give me a hand number which you might be able to still get by scrolling back that ll help i am semi sureish that it’s this was reported to me via email i didn t make an issue out of it at the time and then i forgot i m starting with this as high priority and easy because i have a hand number which will also give me the table and event i can poke around and see if there were any error messages reported from the server to the client if there were then it should be trivial to reproduce and fix if there are no error messages i ll have to ask some more questions and it may no longer be easy | 1 |
664,406 | 22,269,345,555 | IssuesEvent | 2022-06-10 10:38:50 | opencrvs/opencrvs-core | https://api.github.com/repos/opencrvs/opencrvs-core | closed | If user search a birth/death application and open it, Date of death,Place of death and Informant do not show | 👹Bug Priority: high | **Bug Description:**
If a user searches a birth/death application and then opens it, the Following information shows blank:
- Date of death
- Place of death
- Informant
**Steps:**
1. Log in as a field agent/Registration Clerk/Register
2. Click on any application which has the Date of death, Place of death, and Informant from any tab
3. Download the application(if logged in with Registration clerk/Register)
4. Observe that the Date of death, Place of death, and Informant are showing
5. Copy the tracking id or name or phone number
6. Search the application with the tracking id or name or phone number
7. Click on the application
**Actual Result:**
- Date of death, Place of death, and Informant do not show any data
**Expected Result:**
- Date of death, Place of death, and Informant should show if the user opens the application after searching
**Screen record:**
https://www.screencast.com/t/3T0jmT70bl
**Tested on:**
https://login.farajaland-qa.opencrvs.org/
**Username & Password Used:**
- Username: felix.katongo
- password: test
**Desktop:**
- OS: Windows 10
- Browser: Chrome | 1.0 | If user search a birth/death application and open it, Date of death,Place of death and Informant do not show - **Bug Description:**
If a user searches a birth/death application and then opens it, the Following information shows blank:
- Date of death
- Place of death
- Informant
**Steps:**
1. Log in as a field agent/Registration Clerk/Register
2. Click on any application which has the Date of death, Place of death, and Informant from any tab
3. Download the application(if logged in with Registration clerk/Register)
4. Observe that the Date of death, Place of death, and Informant are showing
5. Copy the tracking id or name or phone number
6. Search the application with the tracking id or name or phone number
7. Click on the application
**Actual Result:**
- Date of death, Place of death, and Informant do not show any data
**Expected Result:**
- Date of death, Place of death, and Informant should show if the user opens the application after searching
**Screen record:**
https://www.screencast.com/t/3T0jmT70bl
**Tested on:**
https://login.farajaland-qa.opencrvs.org/
**Username & Password Used:**
- Username: felix.katongo
- password: test
**Desktop:**
- OS: Windows 10
- Browser: Chrome | priority | if user search a birth death application and open it date of death place of death and informant do not show bug description if a user searches a birth death application and then opens it the following information shows blank date of death place of death informant steps log in as a field agent registration clerk register click on any application which has the date of death place of death and informant from any tab download the application if logged in with registration clerk register observe that the date of death place of death and informant are showing copy the tracking id or name or phone number search the application with the tracking id or name or phone number click on the application actual result date of death place of death and informant do not show any data expected result date of death place of death and informant should show if the user opens the application after searching screen record tested on username password used username felix katongo password test desktop os windows browser chrome | 1 |
621,757 | 19,595,953,455 | IssuesEvent | 2022-01-05 17:52:13 | vaticle/typedb | https://api.github.com/repos/vaticle/typedb | closed | Investigate OOM caused using TypeDB Loader | type: bug priority: high | ## Description
Bayer has reported an occasional OOM when doing heavy data loading.
## Environment
TypeDB 2.2.0 on a cloud-based linux VM:
```
Host: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz, 4 cores, 15G, Amazon Linux release 2 (Karoo)
```
## Reproducible Steps
- see private dropbox folder for dataset and reproducible steps
## Expected Output
- data loading runs indefinitely without fail
## Actual Output
One failure occurs with the following error, unsure if others occur:
```
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 58720256 bytes for committing reserved memory.
# Possible reasons:
# The system is out of physical RAM or swap space
# The process is running with CompressedOops enabled, and the Java Heap may be blocking the growth of the native heap
# Possible solutions:
# Reduce memory load on the system
# Increase physical memory or swap space
# Check if swap backing store is full
# Decrease Java heap size (-Xmx/-Xms)
# Decrease number of Java threads
# Decrease Java thread stack sizes (-Xss)
# Set larger code cache with -XX:ReservedCodeCacheSize=
# JVM is running with Zero Based Compressed Oops mode in which the Java heap is
# placed in the first 32GB address space. The Java Heap base address is the
# maximum limit for the native heap growth. Please use -XX:HeapBaseMinAddress
# to set the Java Heap base and to place the Java Heap above 32GB virtual address.
# This output file may be truncated or incomplete.
#
# Out of Memory Error (os_linux.cpp:2977), pid=17756, tid=17768
#
# JRE version: OpenJDK Runtime Environment 20.9 (15.0.2+7) (build 15.0.2+7)
# Java VM: OpenJDK 64-Bit Server VM 20.9 (15.0.2+7, mixed mode, sharing, tiered, compressed oops, g1 gc, linux-amd64)
# No core dump will be written. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
```
## Additional information
| 1.0 | Investigate OOM caused using TypeDB Loader - ## Description
Bayer has reported an occasional OOM when doing heavy data loading.
## Environment
TypeDB 2.2.0 on a cloud-based linux VM:
```
Host: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz, 4 cores, 15G, Amazon Linux release 2 (Karoo)
```
## Reproducible Steps
- see private dropbox folder for dataset and reproducible steps
## Expected Output
- data loading runs indefinitely without fail
## Actual Output
One failure occurs with the following error, unsure if others occur:
```
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 58720256 bytes for committing reserved memory.
# Possible reasons:
# The system is out of physical RAM or swap space
# The process is running with CompressedOops enabled, and the Java Heap may be blocking the growth of the native heap
# Possible solutions:
# Reduce memory load on the system
# Increase physical memory or swap space
# Check if swap backing store is full
# Decrease Java heap size (-Xmx/-Xms)
# Decrease number of Java threads
# Decrease Java thread stack sizes (-Xss)
# Set larger code cache with -XX:ReservedCodeCacheSize=
# JVM is running with Zero Based Compressed Oops mode in which the Java heap is
# placed in the first 32GB address space. The Java Heap base address is the
# maximum limit for the native heap growth. Please use -XX:HeapBaseMinAddress
# to set the Java Heap base and to place the Java Heap above 32GB virtual address.
# This output file may be truncated or incomplete.
#
# Out of Memory Error (os_linux.cpp:2977), pid=17756, tid=17768
#
# JRE version: OpenJDK Runtime Environment 20.9 (15.0.2+7) (build 15.0.2+7)
# Java VM: OpenJDK 64-Bit Server VM 20.9 (15.0.2+7, mixed mode, sharing, tiered, compressed oops, g1 gc, linux-amd64)
# No core dump will be written. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
```
## Additional information
| priority | investigate oom caused using typedb loader description bayer has reported an occasional oom when doing heavy data loading environment typedb on a cloud based linux vm host intel r xeon r platinum cpu cores amazon linux release karoo reproducible steps see private dropbox folder for dataset and reproducible steps expected output data loading runs indefinitely without fail actual output one failure occurs with the following error unsure if others occur there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory possible reasons the system is out of physical ram or swap space the process is running with compressedoops enabled and the java heap may be blocking the growth of the native heap possible solutions reduce memory load on the system increase physical memory or swap space check if swap backing store is full decrease java heap size xmx xms decrease number of java threads decrease java thread stack sizes xss set larger code cache with xx reservedcodecachesize jvm is running with zero based compressed oops mode in which the java heap is placed in the first address space the java heap base address is the maximum limit for the native heap growth please use xx heapbaseminaddress to set the java heap base and to place the java heap above virtual address this output file may be truncated or incomplete out of memory error os linux cpp pid tid jre version openjdk runtime environment build java vm openjdk bit server vm mixed mode sharing tiered compressed oops gc linux no core dump will be written core dumps have been disabled to enable core dumping try ulimit c unlimited before starting java again additional information | 1 |
22,067 | 2,644,998,602 | IssuesEvent | 2015-03-12 20:04:06 | starteam/starcellbio_html | https://api.github.com/repos/starteam/starcellbio_html | closed | SUTD Exercise #2 - Flow Cytometry: clicking and dragging for drawing gates not working properly | bug High Priority | The gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy (for both the single and the bisector gate tools) | 1.0 | SUTD Exercise #2 - Flow Cytometry: clicking and dragging for drawing gates not working properly - The gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy (for both the single and the bisector gate tools) | priority | sutd exercise flow cytometry clicking and dragging for drawing gates not working properly the gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy for both the single and the bisector gate tools | 1 |
796,839 | 28,128,869,293 | IssuesEvent | 2023-03-31 20:24:59 | mila-iqia/mila-docs | https://api.github.com/repos/mila-iqia/mila-docs | reopened | Search box seems broken. | bug priority:high | The search box (implemented within the browser space) seems to spin endlessly. Doesn't provide results even after a long while.
Is it broken ?
This is on firefox / linux. | 1.0 | Search box seems broken. - The search box (implemented within the browser space) seems to spin endlessly. Doesn't provide results even after a long while.
Is it broken ?
This is on firefox / linux. | priority | search box seems broken the search box implemented within the browser space seems to spin endlessly doesn t provide results even after a long while is it broken this is on firefox linux | 1 |
305,576 | 9,371,448,465 | IssuesEvent | 2019-04-03 15:24:56 | CS2103-AY1819S2-T12-4/main | https://api.github.com/repos/CS2103-AY1819S2-T12-4/main | closed | Update of UG/DG to fit v1.3 | priority.High | **For DG** _(under Implementation)_:
- [x] Decrypt
- [x] Encrypt
- [x] Filter
- [x] Find
- [x] Label
- [x] Merge
_Might or must need more details/ illustrations:_
- [x] Clear
- [x] Deadline
- [x] Delete
- [x] Exit
- [x] Help
- [x] History
- [x] List
- [x] Select
- [x] Sort
- [x] Tag
All hands on deck boys | 1.0 | Update of UG/DG to fit v1.3 - **For DG** _(under Implementation)_:
- [x] Decrypt
- [x] Encrypt
- [x] Filter
- [x] Find
- [x] Label
- [x] Merge
_Might or must need more details/ illustrations:_
- [x] Clear
- [x] Deadline
- [x] Delete
- [x] Exit
- [x] Help
- [x] History
- [x] List
- [x] Select
- [x] Sort
- [x] Tag
All hands on deck boys | priority | update of ug dg to fit for dg under implementation decrypt encrypt filter find label merge might or must need more details illustrations clear deadline delete exit help history list select sort tag all hands on deck boys | 1 |
554,798 | 16,439,102,889 | IssuesEvent | 2021-05-20 12:36:25 | TEIC/TEI | https://api.github.com/repos/TEIC/TEI | closed | allow <schemaSpec> in <front> and <back> | Priority: High TEI: Schema Type: Bug | In ticket #1921 we are removing `<schemaSpec>` from a variety of silly places (like `<figure>`, `<license>`, and `<scriptNote>`), and we now explicitly recommend “It should be a child of front, body, back, encodingDesc, div, div1, …, or div7.”.
HOWEVER, `<schemaSpec>` is not allowed as a child of `<front>` or `<back>`.
This, I think is a corrigible schema error that should just be fixed.
Alternatively, of course, we could remove the explicit recommendation that it go in `<front>` or `<back>`. | 1.0 | allow <schemaSpec> in <front> and <back> - In ticket #1921 we are removing `<schemaSpec>` from a variety of silly places (like `<figure>`, `<license>`, and `<scriptNote>`), and we now explicitly recommend “It should be a child of front, body, back, encodingDesc, div, div1, …, or div7.”.
HOWEVER, `<schemaSpec>` is not allowed as a child of `<front>` or `<back>`.
This, I think is a corrigible schema error that should just be fixed.
Alternatively, of course, we could remove the explicit recommendation that it go in `<front>` or `<back>`. | priority | allow in and in ticket we are removing from a variety of silly places like and and we now explicitly recommend “it should be a child of front body back encodingdesc div … or ” however is not allowed as a child of or this i think is a corrigible schema error that should just be fixed alternatively of course we could remove the explicit recommendation that it go in or | 1 |
499,850 | 14,480,799,167 | IssuesEvent | 2020-12-10 11:41:35 | mpagels/collectoPod | https://api.github.com/repos/mpagels/collectoPod | opened | Podcast sort function | Priority: High Type: New Feature | ## Value statement - a requirement, not a solution
As a **user**
I need **a sort function**
so that **I can sort the order of the podcasts**
## Description (Text, Scribble, Wireframe, Design)
## Acceptance criteria
- [ ] I see a list of the podcasts
- [ ] when I move the first podcast under the second podcast, the second podcast moves up and the first podcast takes the second place
- [ ] I hit the save button to save the new list to localStorage
## Tasks
## Size
(medium)
| 1.0 | Podcast sort function - ## Value statement - a requirement, not a solution
As a **user**
I need **a sort function**
so that **I can sort the order of the podcasts**
## Description (Text, Scribble, Wireframe, Design)
## Acceptance criteria
- [ ] I see a list of the podcasts
- [ ] when I move the first podcast under the second podcast, the second podcast moves up and the first podcast takes the second place
- [ ] I hit the save button to save the new list to localStorage
## Tasks
## Size
(medium)
| priority | podcast sort function value statement a requirement not a solution as a user i need a sort function so that i can sort the order of the podcasts description text scribble wireframe design acceptance criteria i see a list of the podcasts when i move the first podcast under the second podcast the second podcast moves up and the first podcast takes the second place i hit the save button to save the new list to localstorage tasks size medium | 1 |
171,736 | 6,493,927,926 | IssuesEvent | 2017-08-21 19:26:23 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Preview tools panel and tooltips does not use Open Sans for the font | bug Priority: High | Using the Website_editorial bp, click on the wrench at the top right of studio to open the preview tools panel. Notice the font in the dialog is not Open Sans:

Click on **Page Components**, notice that the font used in the panel is also not Open Sans:

Click on **Browse Features**, after selecting the component you want to add, notice the notification on the top right of the screen does not use Open Sans for the font:

In the **Sidebar**, hover on any of the folders/items, notice the font used in the tooltip is not Open Sans

Clicn on **Approve & Publish** in the context nav, then click on **Items go live on a specific date & time**, notice that the font used by the datetime picker is not Open Sans

| 1.0 | [studio-ui] Preview tools panel and tooltips does not use Open Sans for the font - Using the Website_editorial bp, click on the wrench at the top right of studio to open the preview tools panel. Notice the font in the dialog is not Open Sans:

Click on **Page Components**, notice that the font used in the panel is also not Open Sans:

Click on **Browse Features**, after selecting the component you want to add, notice the notification on the top right of the screen does not use Open Sans for the font:

In the **Sidebar**, hover on any of the folders/items, notice the font used in the tooltip is not Open Sans

Clicn on **Approve & Publish** in the context nav, then click on **Items go live on a specific date & time**, notice that the font used by the datetime picker is not Open Sans

| priority | preview tools panel and tooltips does not use open sans for the font using the website editorial bp click on the wrench at the top right of studio to open the preview tools panel notice the font in the dialog is not open sans click on page components notice that the font used in the panel is also not open sans click on browse features after selecting the component you want to add notice the notification on the top right of the screen does not use open sans for the font in the sidebar hover on any of the folders items notice the font used in the tooltip is not open sans clicn on approve publish in the context nav then click on items go live on a specific date time notice that the font used by the datetime picker is not open sans | 1 |
802,385 | 28,959,738,227 | IssuesEvent | 2023-05-10 00:43:36 | steedos/steedos-platform | https://api.github.com/repos/steedos/steedos-platform | closed | [Bug]: 远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白 | bug done priority: High | ### Description
<img width="1440" alt="image" src="https://user-images.githubusercontent.com/26241897/236722162-99928ac5-0b0c-40cb-87d9-9928c02c590d.png">
控制台报错信息:
steedos_1 | memory, and will not scale past a single process.
steedos_1 | service ~packages-standard-objects started: 1.479s
steedos_1 | service ~packages-@steedos/standard-space started: 307.312ms
steedos_1 | service ~packages-@steedos/standard-cms started: 105.359ms
steedos_1 | service ~packages-@steedos/standard-object-database started: 512.988ms
steedos_1 | service ~packages-@steedos/standard-process-approval started: 338.997ms
steedos_1 | service ~packages-@steedos/standard-collaboration started: 179.123ms
steedos_1 | service ~packages-@steedos/standard-ui started: 210.011ms
steedos_1 | service ~packages-@steedos/standard-permission started: 190.842ms
steedos_1 | service ~packages-@steedos/workflow started: 45.182ms
steedos_1 | service ~packages-@steedos/accounts started
steedos_1 | service ~packages-@steedos/plugin-company started: 221.344ms
steedos_1 | service ~packages-@steedos/metadata-api started: 54.139ms
steedos_1 |
steedos_1 | Project is running at https://5000-tomato-finch-em8hx9yz.ws.vscode.steedos.cn
steedos_1 |
steedos_1 | service ~packages-@steedos/data-import started: 61.041ms
steedos_1 | service ~packages-@steedos/unpkg started: 1.302s
steedos_1 | service ~packages-@steedos/webapp-accounts started: 1.276s
steedos_1 | service ~packages-@steedos/service-analytics started: 1.236s
steedos_1 | service ~packages-@steedos/service-ui started: 1.184s
steedos_1 | service ~packages-@steedos/webapp-public started: 492.923ms
steedos_1 | service ~packages-@steedos/service-files started: 253.453ms
steedos_1 | service ~packages-@steedos-labs/flows started: 1.059s
steedos_1 | service ~packages-@steedos/service-i18n started: 274.951ms
steedos_1 | service ~packages-@steedos-labs/workflow started: 942.123ms
steedos_1 | service ~packages-@steedos/workflow_time_trigger started: 1.538s
steedos_1 | service ~packages-@steedos-labs/oa started: 1.278s
steedos_1 | service ~packages-@steedos/service-package-registry started: 637.725ms
steedos_1 | service ~packages-@steedos/service-charts started: 1.162s
steedos_1 | service ~packages-@steedos/steedos-plugin-schema-builder started: 2.095s
steedos_1 | service ~packages-@steedos/service-pages started: 1.675s
steedos_1 | service ~packages-@steedos-labs/meeting started: 2.234s
steedos_1 | service ~packages-@steedos/service-plugin-amis started: 265.047ms
steedos_1 | service ~packages-@steedos-labs/okr started: 2.562s
steedos_1 | service ~packages-@steedos-labs/budget started: 3.193s
steedos_1 | service ~packages-@steedos-labs/order started: 3.474s
steedos_1 | service ~packages-@steedos-labs/purchase started: 3.537s
steedos_1 | service ~packages-@steedos/service-package-tool started: 3.579s
steedos_1 | service ~packages-@steedos-labs/project started: 4.404s
steedos_1 | service ~packages-@steedos-labs/master started: 4.822s
steedos_1 | service ~packages-@steedos-labs/cost started: 5.548s
steedos_1 | service ~packages-@steedos-labs/project-ee started: 699.506ms
steedos_1 | service ~packages-@steedos-labs/contract started: 5.927s
steedos_1 | service ~packages-@steedos-labs/crm started: 5.646s
steedos_1 | service ~packages-@steedos-labs/demo-ee started: 497.628ms
steedos_1 | service ~packages-@steedos-labs/contract-ee started: 1.031s
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos/plugin-qywx started: 1.343s
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos/plugin-dingtalk started: 947.906ms
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos-labs/finance started: 3.395s
steedos_1 | service ~packages-@steedos-labs/asset started: 8.400s
steedos_1 | service ~packages-@steedos-labs/cost-ee started: 2.719s
steedos_1 | service ~packages-@steedos-labs/finance-ee started: 42.196ms
steedos_1 | install package @steedos-labs/project-ee successful.
steedos_1 | service ~packages-@steedos/service-workflow started: 8.870s
steedos_1 | service ~packages-@steedos-labs/project-ee stopped
steedos_1 | service ~packages-@steedos-labs/project-ee started: 217.276ms
steedos_1 | install package @steedos-labs/demo-ee successful.
steedos_1 | service ~packages-@steedos-labs/demo-ee stopped
steedos_1 | service ~packages-@steedos-labs/demo-ee started: 54.886ms
steedos_1 | install package @steedos-labs/flows successful.
steedos_1 | service ~packages-@steedos-labs/flows stopped
steedos_1 | service ~packages-@steedos-labs/flows started: 59.969ms
steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined
steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90)
steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1439:47)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 3)
steedos_1 | (Use `node --trace-warnings ...` to show where the warning was created)
steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined
steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90)
steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1814:47)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.processUnreadableField (/app/node_modules/@steedos/objectql/lib/types/object.js:2240:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1072:24)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.find (/app/node_modules/@steedos/objectql/lib/types/object.js:1182:24)
steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:158:64
steedos_1 | at Generator.next (<anonymous>)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at getHiddenTabNames (/app/node_modules/@steedos/service-metadata-apps/lib/actionsHandler.js:135:20)
steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:91:34
steedos_1 | at Generator.next (<anonymous>)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 5)
### Steps To Reproduce 重现步骤
1. 重开template项目,版本升级到2.5.0-beta.20;
2. 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息
### Version 版本
2.5.0-beta.20 | 1.0 | [Bug]: 远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白 - ### Description
<img width="1440" alt="image" src="https://user-images.githubusercontent.com/26241897/236722162-99928ac5-0b0c-40cb-87d9-9928c02c590d.png">
控制台报错信息:
steedos_1 | memory, and will not scale past a single process.
steedos_1 | service ~packages-standard-objects started: 1.479s
steedos_1 | service ~packages-@steedos/standard-space started: 307.312ms
steedos_1 | service ~packages-@steedos/standard-cms started: 105.359ms
steedos_1 | service ~packages-@steedos/standard-object-database started: 512.988ms
steedos_1 | service ~packages-@steedos/standard-process-approval started: 338.997ms
steedos_1 | service ~packages-@steedos/standard-collaboration started: 179.123ms
steedos_1 | service ~packages-@steedos/standard-ui started: 210.011ms
steedos_1 | service ~packages-@steedos/standard-permission started: 190.842ms
steedos_1 | service ~packages-@steedos/workflow started: 45.182ms
steedos_1 | service ~packages-@steedos/accounts started
steedos_1 | service ~packages-@steedos/plugin-company started: 221.344ms
steedos_1 | service ~packages-@steedos/metadata-api started: 54.139ms
steedos_1 |
steedos_1 | Project is running at https://5000-tomato-finch-em8hx9yz.ws.vscode.steedos.cn
steedos_1 |
steedos_1 | service ~packages-@steedos/data-import started: 61.041ms
steedos_1 | service ~packages-@steedos/unpkg started: 1.302s
steedos_1 | service ~packages-@steedos/webapp-accounts started: 1.276s
steedos_1 | service ~packages-@steedos/service-analytics started: 1.236s
steedos_1 | service ~packages-@steedos/service-ui started: 1.184s
steedos_1 | service ~packages-@steedos/webapp-public started: 492.923ms
steedos_1 | service ~packages-@steedos/service-files started: 253.453ms
steedos_1 | service ~packages-@steedos-labs/flows started: 1.059s
steedos_1 | service ~packages-@steedos/service-i18n started: 274.951ms
steedos_1 | service ~packages-@steedos-labs/workflow started: 942.123ms
steedos_1 | service ~packages-@steedos/workflow_time_trigger started: 1.538s
steedos_1 | service ~packages-@steedos-labs/oa started: 1.278s
steedos_1 | service ~packages-@steedos/service-package-registry started: 637.725ms
steedos_1 | service ~packages-@steedos/service-charts started: 1.162s
steedos_1 | service ~packages-@steedos/steedos-plugin-schema-builder started: 2.095s
steedos_1 | service ~packages-@steedos/service-pages started: 1.675s
steedos_1 | service ~packages-@steedos-labs/meeting started: 2.234s
steedos_1 | service ~packages-@steedos/service-plugin-amis started: 265.047ms
steedos_1 | service ~packages-@steedos-labs/okr started: 2.562s
steedos_1 | service ~packages-@steedos-labs/budget started: 3.193s
steedos_1 | service ~packages-@steedos-labs/order started: 3.474s
steedos_1 | service ~packages-@steedos-labs/purchase started: 3.537s
steedos_1 | service ~packages-@steedos/service-package-tool started: 3.579s
steedos_1 | service ~packages-@steedos-labs/project started: 4.404s
steedos_1 | service ~packages-@steedos-labs/master started: 4.822s
steedos_1 | service ~packages-@steedos-labs/cost started: 5.548s
steedos_1 | service ~packages-@steedos-labs/project-ee started: 699.506ms
steedos_1 | service ~packages-@steedos-labs/contract started: 5.927s
steedos_1 | service ~packages-@steedos-labs/crm started: 5.646s
steedos_1 | service ~packages-@steedos-labs/demo-ee started: 497.628ms
steedos_1 | service ~packages-@steedos-labs/contract-ee started: 1.031s
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos/plugin-qywx started: 1.343s
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos/plugin-dingtalk started: 947.906ms
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | Error: Invalid Session
steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | service ~packages-@steedos-labs/finance started: 3.395s
steedos_1 | service ~packages-@steedos-labs/asset started: 8.400s
steedos_1 | service ~packages-@steedos-labs/cost-ee started: 2.719s
steedos_1 | service ~packages-@steedos-labs/finance-ee started: 42.196ms
steedos_1 | install package @steedos-labs/project-ee successful.
steedos_1 | service ~packages-@steedos/service-workflow started: 8.870s
steedos_1 | service ~packages-@steedos-labs/project-ee stopped
steedos_1 | service ~packages-@steedos-labs/project-ee started: 217.276ms
steedos_1 | install package @steedos-labs/demo-ee successful.
steedos_1 | service ~packages-@steedos-labs/demo-ee stopped
steedos_1 | service ~packages-@steedos-labs/demo-ee started: 54.886ms
steedos_1 | install package @steedos-labs/flows successful.
steedos_1 | service ~packages-@steedos-labs/flows stopped
steedos_1 | service ~packages-@steedos-labs/flows started: 59.969ms
steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined
steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90)
steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1439:47)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 3)
steedos_1 | (Use `node --trace-warnings ...` to show where the warning was created)
steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined
steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90)
steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1814:47)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.processUnreadableField (/app/node_modules/@steedos/objectql/lib/types/object.js:2240:24)
steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1072:24)
steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27)
steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at SteedosObjectType.find (/app/node_modules/@steedos/objectql/lib/types/object.js:1182:24)
steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:158:64
steedos_1 | at Generator.next (<anonymous>)
steedos_1 | at /app/node_modules/tslib/tslib.js:167:75
steedos_1 | at new Promise (<anonymous>)
steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16)
steedos_1 | at getHiddenTabNames (/app/node_modules/@steedos/service-metadata-apps/lib/actionsHandler.js:135:20)
steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:91:34
steedos_1 | at Generator.next (<anonymous>)
steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62)
steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39
steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 5)
### Steps To Reproduce 重现步骤
1. 重开template项目,版本升级到2.5.0-beta.20;
2. 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息
### Version 版本
2.5.0-beta.20 | priority | 远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白 description img width alt image src 控制台报错信息: steedos memory and will not scale past a single process steedos service packages standard objects started steedos service packages steedos standard space started steedos service packages steedos standard cms started steedos service packages steedos standard object database started steedos service packages steedos standard process approval started steedos service packages steedos standard collaboration started steedos service packages steedos standard ui started steedos service packages steedos standard permission started steedos service packages steedos workflow started steedos service packages steedos accounts started steedos service packages steedos plugin company started steedos service packages steedos metadata api started steedos steedos project is running at steedos steedos service packages steedos data import started steedos service packages steedos unpkg started steedos service packages steedos webapp accounts started steedos service packages steedos service analytics started steedos service packages steedos service ui started steedos service packages steedos webapp public started steedos service packages steedos service files started steedos service packages steedos labs flows started steedos service packages steedos service started steedos service packages steedos labs workflow started steedos service packages steedos workflow time trigger started steedos service packages steedos labs oa started steedos service packages steedos service package registry started steedos service packages steedos service charts started steedos service packages steedos steedos plugin schema builder started steedos service packages steedos service pages started steedos service packages steedos labs meeting started steedos service packages steedos service plugin amis started steedos service packages steedos labs okr started steedos service packages steedos labs budget started steedos service packages steedos labs order started steedos service packages steedos labs purchase started steedos service packages steedos service package tool started steedos service packages steedos labs project started steedos service packages steedos labs master started steedos service packages steedos labs cost started steedos service packages steedos labs project ee started steedos service packages steedos labs contract started steedos service packages steedos labs crm started steedos service packages steedos labs demo ee started steedos service packages steedos labs contract ee started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos plugin qywx started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos plugin dingtalk started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos labs finance started steedos service packages steedos labs asset started steedos service packages steedos labs cost ee started steedos service packages steedos labs finance ee started steedos install package steedos labs project ee successful steedos service packages steedos service workflow started steedos service packages steedos labs project ee stopped steedos service packages steedos labs project ee started steedos install package steedos labs demo ee successful steedos service packages steedos labs demo ee stopped steedos service packages steedos labs demo ee started steedos install package steedos labs flows successful steedos service packages steedos labs flows stopped steedos service packages steedos labs flows started steedos this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch the promise rejected with the reason steedos typeerror cannot read property of undefined steedos at steedosdatasourcetype getobjectspacerolespermission app node modules steedos objectql src types datasource ts steedos at steedosobjecttype getobjectrolespermission app node modules steedos objectql src types object ts steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype getuserobjectpermission app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos node promiserejectionhandledwarning promise rejection was handled asynchronously rejection id steedos use node trace warnings to show where the warning was created steedos this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch the promise rejected with the reason steedos typeerror cannot read property of undefined steedos at steedosdatasourcetype getobjectspacerolespermission app node modules steedos objectql src types datasource ts steedos at steedosobjecttype getobjectrolespermission app node modules steedos objectql src types object ts steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype getuserobjectpermission app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype processunreadablefield app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype find app node modules steedos objectql lib types object js steedos at app node modules steedos service metadata apps src actionshandler ts steedos at generator next steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at gethiddentabnames app node modules steedos service metadata apps lib actionshandler js steedos at app node modules steedos service metadata apps src actionshandler ts steedos at generator next steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos node promiserejectionhandledwarning promise rejection was handled asynchronously rejection id steps to reproduce 重现步骤 重开template项目, beta ; 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息 version 版本 beta | 1 |
330,153 | 10,035,478,451 | IssuesEvent | 2019-07-18 08:26:59 | zdnscloud/singlecloud | https://api.github.com/repos/zdnscloud/singlecloud | closed | error "does not generate config" returns when creating ceph storagecluster | bug priority: High | workaround:
delete and redeploy ceph will be OK.

| 1.0 | error "does not generate config" returns when creating ceph storagecluster - workaround:
delete and redeploy ceph will be OK.

| priority | error does not generate config returns when creating ceph storagecluster workaround delete and redeploy ceph will be ok | 1 |
474,916 | 13,684,409,842 | IssuesEvent | 2020-09-30 04:48:49 | rkulkar5/OnlineTAT | https://api.github.com/repos/rkulkar5/OnlineTAT | opened | Account level sme: Calculate final score is not working | Issues after Sprint-9 high priority | Calculate final score in Initiate technical interview screen of sme queue is not working

| 1.0 | Account level sme: Calculate final score is not working - Calculate final score in Initiate technical interview screen of sme queue is not working

| priority | account level sme calculate final score is not working calculate final score in initiate technical interview screen of sme queue is not working | 1 |
429,006 | 12,419,599,454 | IssuesEvent | 2020-05-23 07:14:04 | dailydotdev/daily-api | https://api.github.com/repos/dailydotdev/daily-api | opened | Add seed data to docker | Priority: High Status: Accepted Type: Maintenance | To complete https://github.com/dailydotdev/daily-apps/issues/110, we must copy seed data to docker so it will be available to `docker-compose` based setup. | 1.0 | Add seed data to docker - To complete https://github.com/dailydotdev/daily-apps/issues/110, we must copy seed data to docker so it will be available to `docker-compose` based setup. | priority | add seed data to docker to complete we must copy seed data to docker so it will be available to docker compose based setup | 1 |
469,404 | 13,507,960,720 | IssuesEvent | 2020-09-14 06:57:00 | TerriaJS/terriajs | https://api.github.com/repos/TerriaJS/terriajs | closed | V8: Icons in hamburger menu on StoryBuilder scenes are wrong | High priority Version 8 | The view and edit icons in the hamburger menu on StoryBuilder scenes are incorrect. Pictured here:

See https://github.com/TerriaJS/terriajs/pull/4483#issuecomment-690915337 for icon files | 1.0 | V8: Icons in hamburger menu on StoryBuilder scenes are wrong - The view and edit icons in the hamburger menu on StoryBuilder scenes are incorrect. Pictured here:

See https://github.com/TerriaJS/terriajs/pull/4483#issuecomment-690915337 for icon files | priority | icons in hamburger menu on storybuilder scenes are wrong the view and edit icons in the hamburger menu on storybuilder scenes are incorrect pictured here see for icon files | 1 |
240,335 | 7,801,047,762 | IssuesEvent | 2018-06-09 16:27:34 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0012544:
Missing packages expressomail, requesttracker and sipgate available | Bug Mantis Packaging high priority | **Reported by weberho on 20 Jan 2017 08:49**
**Version:** 2016.09.7 Community Edition
Those packages are no longer available. What's the reason? Is this intended?
| 1.0 | 0012544:
Missing packages expressomail, requesttracker and sipgate available - **Reported by weberho on 20 Jan 2017 08:49**
**Version:** 2016.09.7 Community Edition
Those packages are no longer available. What's the reason? Is this intended?
| priority | missing packages expressomail requesttracker and sipgate available reported by weberho on jan version community edition those packages are no longer available what s the reason is this intended | 1 |
535,464 | 15,688,749,496 | IssuesEvent | 2021-03-25 14:59:29 | itslupus/gamersnet | https://api.github.com/repos/itslupus/gamersnet | closed | Database for chat messages | backend dev task high priority | **Description**:
Create database method to fetch and store chat messages
| 1.0 | Database for chat messages - **Description**:
Create database method to fetch and store chat messages
| priority | database for chat messages description create database method to fetch and store chat messages | 1 |
47,023 | 2,971,374,977 | IssuesEvent | 2015-07-14 06:13:46 | mcdermott-scholars/mcdermott | https://api.github.com/repos/mcdermott-scholars/mcdermott | opened | Class year should also be a drop down | high priority | Starting from 2001, I believe? (or maybe 2000 if you want to count the special class?) | 1.0 | Class year should also be a drop down - Starting from 2001, I believe? (or maybe 2000 if you want to count the special class?) | priority | class year should also be a drop down starting from i believe or maybe if you want to count the special class | 1 |
340,581 | 10,273,523,747 | IssuesEvent | 2019-08-23 19:33:30 | byu-animation/dccpipe | https://api.github.com/repos/byu-animation/dccpipe | opened | Change all references to "character" to "actor" | Houdini Maya enhancement priority: high | Because a character is rather limited in definition, but an actor can be an inanimate object that is more involved than a prop, or it can also be a character.
i.e. a rug with a hair sim, a clothesline with cloth, a scythe with a 'tude, etc... | 1.0 | Change all references to "character" to "actor" - Because a character is rather limited in definition, but an actor can be an inanimate object that is more involved than a prop, or it can also be a character.
i.e. a rug with a hair sim, a clothesline with cloth, a scythe with a 'tude, etc... | priority | change all references to character to actor because a character is rather limited in definition but an actor can be an inanimate object that is more involved than a prop or it can also be a character i e a rug with a hair sim a clothesline with cloth a scythe with a tude etc | 1 |
242,396 | 7,841,782,789 | IssuesEvent | 2018-06-18 20:46:05 | stats4sd/Stats4SD-Resources-Site | https://api.github.com/repos/stats4sd/Stats4SD-Resources-Site | closed | Editing always leads to Binomial Distribution | 3 - In Progress Priority-High Type-bug | When I click 'edit resource', the active resource changes to `resources/binomial-distribution`, so I can only ever edit that resource.
From the console log, clicking the edit button causes the "active resource" to be reset.
Oddly, the keywords and resource type are carried over from which-ever resource I clicked through from.
E.g. Going to edit [this resource](https://resources.stats4sd.org/resources/experiments-with-farmers-ii---what-happens-to-designs) takes me back to editing the binomial resource, but with the 3 keywords and resource type= video. | 1.0 | Editing always leads to Binomial Distribution - When I click 'edit resource', the active resource changes to `resources/binomial-distribution`, so I can only ever edit that resource.
From the console log, clicking the edit button causes the "active resource" to be reset.
Oddly, the keywords and resource type are carried over from which-ever resource I clicked through from.
E.g. Going to edit [this resource](https://resources.stats4sd.org/resources/experiments-with-farmers-ii---what-happens-to-designs) takes me back to editing the binomial resource, but with the 3 keywords and resource type= video. | priority | editing always leads to binomial distribution when i click edit resource the active resource changes to resources binomial distribution so i can only ever edit that resource from the console log clicking the edit button causes the active resource to be reset oddly the keywords and resource type are carried over from which ever resource i clicked through from e g going to edit takes me back to editing the binomial resource but with the keywords and resource type video | 1 |
111,712 | 4,481,763,900 | IssuesEvent | 2016-08-29 00:40:18 | influxdata/docs.influxdata.com | https://api.github.com/repos/influxdata/docs.influxdata.com | closed | remove support for negative timestamps | Fix underway Priority: high | Query engine refactor in 0.11 appears to have disabled negative timestamps in InfluxQL. They can still be written, but not queried. The WHERE clause won't accept negative timestamps, relative or absolute, and the default lower bound is epoch 0.
###### URL for relevant page?
https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range
https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range
https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v0.13/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v1.0/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
possibly others
| 1.0 | remove support for negative timestamps - Query engine refactor in 0.11 appears to have disabled negative timestamps in InfluxQL. They can still be written, but not queried. The WHERE clause won't accept negative timestamps, relative or absolute, and the default lower bound is epoch 0.
###### URL for relevant page?
https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range
https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range
https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v0.13/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
https://docs.influxdata.com/influxdb/v1.0/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0
possibly others
| priority | remove support for negative timestamps query engine refactor in appears to have disabled negative timestamps in influxql they can still be written but not queried the where clause won t accept negative timestamps relative or absolute and the default lower bound is epoch url for relevant page possibly others | 1 |
264,740 | 8,319,139,077 | IssuesEvent | 2018-09-25 16:22:15 | Zenika/FAQ | https://api.github.com/repos/Zenika/FAQ | opened | Setup demo.faq.zenika.com | priority/high | In order to showcase the FAQ, we need to have a demo instance.
Needed:
* [ ] Check auth0
* [ ] Automatic deployment multi-instances
* [ ] Create domain + ssl | 1.0 | Setup demo.faq.zenika.com - In order to showcase the FAQ, we need to have a demo instance.
Needed:
* [ ] Check auth0
* [ ] Automatic deployment multi-instances
* [ ] Create domain + ssl | priority | setup demo faq zenika com in order to showcase the faq we need to have a demo instance needed check automatic deployment multi instances create domain ssl | 1 |
520,873 | 15,096,250,418 | IssuesEvent | 2021-02-07 14:23:47 | uah-cpe-sd-2020/myportion_app | https://api.github.com/repos/uah-cpe-sd-2020/myportion_app | opened | Create Database Collections | HIGH PRIORITY | We need to have pet data and feeder data stored in collections. @GuassianFlux Had some ideas on how we might organize this. | 1.0 | Create Database Collections - We need to have pet data and feeder data stored in collections. @GuassianFlux Had some ideas on how we might organize this. | priority | create database collections we need to have pet data and feeder data stored in collections guassianflux had some ideas on how we might organize this | 1 |
786,568 | 27,658,909,962 | IssuesEvent | 2023-03-12 09:44:27 | AY2223S2-CS2113-T14-3/tp | https://api.github.com/repos/AY2223S2-CS2113-T14-3/tp | closed | As a user I can delete an expenditure record | type.Story priority.High | As a user I can delete an expenditure record so that
I can get rid of expenditure that I no longer plan to use
| 1.0 | As a user I can delete an expenditure record - As a user I can delete an expenditure record so that
I can get rid of expenditure that I no longer plan to use
| priority | as a user i can delete an expenditure record as a user i can delete an expenditure record so that i can get rid of expenditure that i no longer plan to use | 1 |
620,607 | 19,565,671,702 | IssuesEvent | 2022-01-03 23:39:45 | bounswe/2021SpringGroup12 | https://api.github.com/repos/bounswe/2021SpringGroup12 | closed | Design of the Search in Market Place | priority: high android | **Description**
- UI design of searching market place should be implemented | 1.0 | Design of the Search in Market Place - **Description**
- UI design of searching market place should be implemented | priority | design of the search in market place description ui design of searching market place should be implemented | 1 |
721,861 | 24,840,842,348 | IssuesEvent | 2022-10-26 12:36:55 | owncloud/ocis | https://api.github.com/repos/owncloud/ocis | closed | Uploading large files fails when using the desktop client | Status:Bug-Analysis Interaction:Needs-help Priority:p2-high | Hi! I am unable to upload larger files to my ocis instance using the desktop client.
It works perfectly fine for files <~100mb but for larger ones, it seems to upload about 100-150MB and then skip the file. After a second of wait it usually retries and fails with a "File is currently in use" type of error.
Uploading those files via the web ui works without any issues. An oc10 instance also will process them just fine (mostly).
I am using the desktop client 2.11.1 (build 8438) on Windows 11 (22621.675) with the ocis beta 8 running inside the official docker container on another system.
As the logs may contain some data which I probably shouldnt share to the public I replaced some names and shorted it down a little.
The log of the ocis instance
[server_log.txt](https://github.com/owncloud/ocis/files/9863568/server_log.txt)
And two from the client. Due to it retrying to upload it created about 8 of them.
The "File is currently in use" type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones.
[Client log 0, 1, 2](https://github.com/owncloud/ocis/files/9863652/20221025_1637_1_owncloud.log)
[Client log 6](https://github.com/owncloud/ocis/files/9863619/20221025_1637_owncloud.log)
[Client log 7](https://github.com/owncloud/ocis/files/9863607/20221025_1638_owncloud.log) | 1.0 | Uploading large files fails when using the desktop client - Hi! I am unable to upload larger files to my ocis instance using the desktop client.
It works perfectly fine for files <~100mb but for larger ones, it seems to upload about 100-150MB and then skip the file. After a second of wait it usually retries and fails with a "File is currently in use" type of error.
Uploading those files via the web ui works without any issues. An oc10 instance also will process them just fine (mostly).
I am using the desktop client 2.11.1 (build 8438) on Windows 11 (22621.675) with the ocis beta 8 running inside the official docker container on another system.
As the logs may contain some data which I probably shouldnt share to the public I replaced some names and shorted it down a little.
The log of the ocis instance
[server_log.txt](https://github.com/owncloud/ocis/files/9863568/server_log.txt)
And two from the client. Due to it retrying to upload it created about 8 of them.
The "File is currently in use" type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones.
[Client log 0, 1, 2](https://github.com/owncloud/ocis/files/9863652/20221025_1637_1_owncloud.log)
[Client log 6](https://github.com/owncloud/ocis/files/9863619/20221025_1637_owncloud.log)
[Client log 7](https://github.com/owncloud/ocis/files/9863607/20221025_1638_owncloud.log) | priority | uploading large files fails when using the desktop client hi i am unable to upload larger files to my ocis instance using the desktop client it works perfectly fine for files but for larger ones it seems to upload about and then skip the file after a second of wait it usually retries and fails with a file is currently in use type of error uploading those files via the web ui works without any issues an instance also will process them just fine mostly i am using the desktop client build on windows with the ocis beta running inside the official docker container on another system as the logs may contain some data which i probably shouldnt share to the public i replaced some names and shorted it down a little the log of the ocis instance and two from the client due to it retrying to upload it created about of them the file is currently in use type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones | 1 |
461,187 | 13,224,837,217 | IssuesEvent | 2020-08-17 19:57:04 | processing/p5.js-web-editor | https://api.github.com/repos/processing/p5.js-web-editor | closed | Console doesn't scroll to bottom to reveal new messages | good first issue help wanted priority:high type:bug | <!--
Hi there! If you are here to report a bug, or to discuss a feature (new or existing), you can use the below template to get started quickly. Fill out all those parts which you're comfortable with, and delete the remaining ones.
-->
#### Nature of issue?
<!-- Select any one issue and delete the other two -->
- Found a bug
<!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. -->
#### Details about the bug:
- Web browser and version: <!-- On Chrome/FireFox/Opera you can enter "about:" in the address bar to find out the version --> Chrome
- Operating System: <!-- Ex: Windows/MacOSX/Linux along with version --> Mac OSX
- Steps to reproduce this bug:
1. Create a sketch that prints lots of messages to the console (i.e. `print()` in the `draw()` loop)
2. Open the console
3. Notice that the console doesn't scroll to the bottom, so it's hard to see new messages.
<!-- Include a simple code snippet that demonstrates the problem, along with any console errors produced. If this isn't possible, then simply describe the issue as best you can! Feel free to link to the web editor or include pictures or a video. -->
I believe this was introduced when refactoring the `<Console />` to use `connect()` (see #1502 and #1507). Not sure why it's not working, as I doubled checked and the code that scrolled the console to the bottom is still in there. | 1.0 | Console doesn't scroll to bottom to reveal new messages - <!--
Hi there! If you are here to report a bug, or to discuss a feature (new or existing), you can use the below template to get started quickly. Fill out all those parts which you're comfortable with, and delete the remaining ones.
-->
#### Nature of issue?
<!-- Select any one issue and delete the other two -->
- Found a bug
<!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. -->
#### Details about the bug:
- Web browser and version: <!-- On Chrome/FireFox/Opera you can enter "about:" in the address bar to find out the version --> Chrome
- Operating System: <!-- Ex: Windows/MacOSX/Linux along with version --> Mac OSX
- Steps to reproduce this bug:
1. Create a sketch that prints lots of messages to the console (i.e. `print()` in the `draw()` loop)
2. Open the console
3. Notice that the console doesn't scroll to the bottom, so it's hard to see new messages.
<!-- Include a simple code snippet that demonstrates the problem, along with any console errors produced. If this isn't possible, then simply describe the issue as best you can! Feel free to link to the web editor or include pictures or a video. -->
I believe this was introduced when refactoring the `<Console />` to use `connect()` (see #1502 and #1507). Not sure why it's not working, as I doubled checked and the code that scrolled the console to the bottom is still in there. | priority | console doesn t scroll to bottom to reveal new messages hi there if you are here to report a bug or to discuss a feature new or existing you can use the below template to get started quickly fill out all those parts which you re comfortable with and delete the remaining ones nature of issue found a bug details about the bug web browser and version chrome operating system mac osx steps to reproduce this bug create a sketch that prints lots of messages to the console i e print in the draw loop open the console notice that the console doesn t scroll to the bottom so it s hard to see new messages i believe this was introduced when refactoring the to use connect see and not sure why it s not working as i doubled checked and the code that scrolled the console to the bottom is still in there | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.