Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
625,974
19,783,535,242
IssuesEvent
2022-01-18 01:58:24
tracer-protocol/pools-client
https://api.github.com/repos/tracer-protocol/pools-client
closed
Release 1.2 - Bugs from Testing
bug Priority: High
Based on current testing: - [ ] Countdown timers don't match between toasts and table view ![countdown-timer](https://user-images.githubusercontent.com/8314736/149742827-f6a4a3b1-3eb5-4241-ad98-df6ff64136f6.gif) - [ ] Expected Price is usually $0, which affects the price through the whole process. ![expected](https://user-images.githubusercontent.com/8314736/149741172-87b310e5-36e7-4b20-a4e2-abd17a7fc4ed.png) - [ ] 'View on Arbiscan' only appears on toast prior to confirming in web3 wallet, then changes to view order. It shouldn't disappear. ![view-token-on-arbiscan](https://user-images.githubusercontent.com/8314736/149741524-af4fbb68-29e7-433b-8d74-7328acbed221.gif) No toast notifications after queued toaster reaches 0 and disappears. - [ ] No toast for minting in progress (user story is missing, I need to confirm with @KittyLomas) - [ ] No toast for success ![no-more-toasts](https://user-images.githubusercontent.com/8314736/149743596-04a68407-f1fe-4a52-8b29-052a0f53bf92.gif) **To Reproduce** 1. Go to https://deploy-preview-380--tracer-pools.netlify.app 2. Follow gifs **Desktop (please complete the following information):** - OS: Windows 11 - Browser: Brave - Version: 1.34.80
1.0
Release 1.2 - Bugs from Testing - Based on current testing: - [ ] Countdown timers don't match between toasts and table view ![countdown-timer](https://user-images.githubusercontent.com/8314736/149742827-f6a4a3b1-3eb5-4241-ad98-df6ff64136f6.gif) - [ ] Expected Price is usually $0, which affects the price through the whole process. ![expected](https://user-images.githubusercontent.com/8314736/149741172-87b310e5-36e7-4b20-a4e2-abd17a7fc4ed.png) - [ ] 'View on Arbiscan' only appears on toast prior to confirming in web3 wallet, then changes to view order. It shouldn't disappear. ![view-token-on-arbiscan](https://user-images.githubusercontent.com/8314736/149741524-af4fbb68-29e7-433b-8d74-7328acbed221.gif) No toast notifications after queued toaster reaches 0 and disappears. - [ ] No toast for minting in progress (user story is missing, I need to confirm with @KittyLomas) - [ ] No toast for success ![no-more-toasts](https://user-images.githubusercontent.com/8314736/149743596-04a68407-f1fe-4a52-8b29-052a0f53bf92.gif) **To Reproduce** 1. Go to https://deploy-preview-380--tracer-pools.netlify.app 2. Follow gifs **Desktop (please complete the following information):** - OS: Windows 11 - Browser: Brave - Version: 1.34.80
priority
release bugs from testing based on current testing countdown timers don t match between toasts and table view expected price is usually which affects the price through the whole process view on arbiscan only appears on toast prior to confirming in wallet then changes to view order it shouldn t disappear no toast notifications after queued toaster reaches and disappears no toast for minting in progress user story is missing i need to confirm with kittylomas no toast for success to reproduce go to follow gifs desktop please complete the following information os windows browser brave version
1
554,964
16,443,928,844
IssuesEvent
2021-05-20 17:13:23
LBNL-ETA/BEDES-Manager
https://api.github.com/repos/LBNL-ETA/BEDES-Manager
closed
Warn user of duplicates when creating new composite term
bug high priority
Check if existing BEDES approved composite term exists and give an alert to the user.
1.0
Warn user of duplicates when creating new composite term - Check if existing BEDES approved composite term exists and give an alert to the user.
priority
warn user of duplicates when creating new composite term check if existing bedes approved composite term exists and give an alert to the user
1
620,130
19,553,436,145
IssuesEvent
2022-01-03 04:05:05
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Resource path that precedes with / results in Missing identifier
Type/Improvement Priority/High Team/CompilerFE Area/Diagnostics Area/Parser
**Description:** <!-- Give a brief description of the improvement --> Then the resource path precedes with a `/`, it results in `Missing identifier` diagnostic error. This causes confusion in the Choreo resource form as we can't identify the issue. The error diagnostic should ne more meaningful to identify the exact error. <img width="552" alt="Screenshot 2021-12-13 at 12 48 12" src="https://user-images.githubusercontent.com/5234623/145768863-a3beeede-6a46-4759-99ee-dad21d8dc56f.png"> **Describe your problem(s)** ``` import ballerina/http; service / on new http:Listener(8080) { resource function get /hello(string name) returns json|error? { return error(""); } } ``` **Describe your solution(s)** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Resource path that precedes with / results in Missing identifier - **Description:** <!-- Give a brief description of the improvement --> Then the resource path precedes with a `/`, it results in `Missing identifier` diagnostic error. This causes confusion in the Choreo resource form as we can't identify the issue. The error diagnostic should ne more meaningful to identify the exact error. <img width="552" alt="Screenshot 2021-12-13 at 12 48 12" src="https://user-images.githubusercontent.com/5234623/145768863-a3beeede-6a46-4759-99ee-dad21d8dc56f.png"> **Describe your problem(s)** ``` import ballerina/http; service / on new http:Listener(8080) { resource function get /hello(string name) returns json|error? { return error(""); } } ``` **Describe your solution(s)** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
priority
resource path that precedes with results in missing identifier description then the resource path precedes with a it results in missing identifier diagnostic error this causes confusion in the choreo resource form as we can t identify the issue the error diagnostic should ne more meaningful to identify the exact error img width alt screenshot at src describe your problem s import ballerina http service on new http listener resource function get hello string name returns json error return error describe your solution s related issues optional suggested labels optional suggested assignees optional
1
450,298
13,001,671,303
IssuesEvent
2020-07-24 00:28:49
UC-Davis-molecular-computing/scadnano-python-package
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano-python-package
opened
add CI checks for API docs building and PyPI tar.gz file building
enhancement high priority
On each commit to dev, there should be a GitHub action that tries to build the API docs and the scadnano-x.x.x.tar.gz file that is uploaded to PyPI. This way, if either of these fails, we know that the docs or PyPI package action will fail when committed to master. The checks could pass and the action on commit to master could still fail, but at least this way it would not be because the files simply could not be generated.
1.0
add CI checks for API docs building and PyPI tar.gz file building - On each commit to dev, there should be a GitHub action that tries to build the API docs and the scadnano-x.x.x.tar.gz file that is uploaded to PyPI. This way, if either of these fails, we know that the docs or PyPI package action will fail when committed to master. The checks could pass and the action on commit to master could still fail, but at least this way it would not be because the files simply could not be generated.
priority
add ci checks for api docs building and pypi tar gz file building on each commit to dev there should be a github action that tries to build the api docs and the scadnano x x x tar gz file that is uploaded to pypi this way if either of these fails we know that the docs or pypi package action will fail when committed to master the checks could pass and the action on commit to master could still fail but at least this way it would not be because the files simply could not be generated
1
34,766
2,787,472,776
IssuesEvent
2015-05-08 06:13:37
CheckiO/checkio-empire-battle
https://api.github.com/repos/CheckiO/checkio-empire-battle
closed
Initial parameter "Size" should be a size of square, not a radius (2 time bigger)
complex:simple priority:high refactoring
Building sizes are not clear.
1.0
Initial parameter "Size" should be a size of square, not a radius (2 time bigger) - Building sizes are not clear.
priority
initial parameter size should be a size of square not a radius time bigger building sizes are not clear
1
526,276
15,285,176,454
IssuesEvent
2021-02-23 13:13:52
carbon-design-system/carbon-for-ibm-dotcom
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
closed
CTA section: Same height by CSS
Airtable Done dev package: react priority: high
### The problem Currently CTA section attempts to ensure all copy contents have the same height, by JavaScript code. It caused FOUC. FOUC causes false positives in Percy depending on when the screenshot is taken. #### Additional Information - scope includes content item horizontal, link list and button group ### The solution Change it to CSS-based, e.g. `grid-auto-rows: 1fr` in CSS grid. #### Acceptance Criteria - [ ] No java script limitation for same height - [ ] No user observable delay to apply the same height
1.0
CTA section: Same height by CSS - ### The problem Currently CTA section attempts to ensure all copy contents have the same height, by JavaScript code. It caused FOUC. FOUC causes false positives in Percy depending on when the screenshot is taken. #### Additional Information - scope includes content item horizontal, link list and button group ### The solution Change it to CSS-based, e.g. `grid-auto-rows: 1fr` in CSS grid. #### Acceptance Criteria - [ ] No java script limitation for same height - [ ] No user observable delay to apply the same height
priority
cta section same height by css the problem currently cta section attempts to ensure all copy contents have the same height by javascript code it caused fouc fouc causes false positives in percy depending on when the screenshot is taken additional information scope includes content item horizontal link list and button group the solution change it to css based e g grid auto rows in css grid acceptance criteria no java script limitation for same height no user observable delay to apply the same height
1
120,634
4,792,640,655
IssuesEvent
2016-10-31 16:02:13
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
classification creation
Enhancement Priority-High
When creating taxon names, species, scientific name, etc., are often not populated, and are somewhat painful to generate manually. Find interface magic.
1.0
classification creation - When creating taxon names, species, scientific name, etc., are often not populated, and are somewhat painful to generate manually. Find interface magic.
priority
classification creation when creating taxon names species scientific name etc are often not populated and are somewhat painful to generate manually find interface magic
1
787,232
27,711,349,911
IssuesEvent
2023-03-14 14:26:06
AY2223S2-CS2113-T13-1/tp
https://api.github.com/repos/AY2223S2-CS2113-T13-1/tp
closed
"create-account" command
type.Story priority.High
As a user, I can create multiple financial accounts so that I can better categorise my expenses and budgets. ### Acceptance Criteria - Creates an account for the specified currency. - For now, throw error if an account with the specified currency already exists - Format: `create-account $/CURRENCY` - Examples: ```java >> create-account $/EUR // Creates a $EUR account ```
1.0
"create-account" command - As a user, I can create multiple financial accounts so that I can better categorise my expenses and budgets. ### Acceptance Criteria - Creates an account for the specified currency. - For now, throw error if an account with the specified currency already exists - Format: `create-account $/CURRENCY` - Examples: ```java >> create-account $/EUR // Creates a $EUR account ```
priority
create account command as a user i can create multiple financial accounts so that i can better categorise my expenses and budgets acceptance criteria creates an account for the specified currency for now throw error if an account with the specified currency already exists format create account currency examples java create account eur creates a eur account
1
352,632
10,544,332,336
IssuesEvent
2019-10-02 16:41:58
eaudeweb/ozone
https://api.github.com/repos/eaudeweb/ozone
closed
Art7 - saving flags
Component: Backend Feature: Art7 Priority: High Status: In progress
related to #1354, copying last comment here, as this seems to be a backend issue: When creating a submission as secretariat, before Submit there is a popup which allows changing the has_blanks flag: ![image](https://user-images.githubusercontent.com/1218168/66038569-28a09100-e51b-11e9-914f-df622997f8c0.png) ![image](https://user-images.githubusercontent.com/1218168/66038572-2dfddb80-e51b-11e9-81d0-22b9a35521ff.png) Check some previously unchecked flags and submit: ![image](https://user-images.githubusercontent.com/1218168/66038582-33f3bc80-e51b-11e9-8009-453c3281dd34.png) After submit, the flags seem to be correctly updated: ![image](https://user-images.githubusercontent.com/1218168/66038612-4bcb4080-e51b-11e9-8734-a16cd5428e94.png) The PUT request seems to have the right values: ![image](https://user-images.githubusercontent.com/1218168/66038697-8c2abe80-e51b-11e9-9212-f6ae81bf5cd6.png) but after refresh, the flags are reverted to the initial values: ![image](https://user-images.githubusercontent.com/1218168/66038634-600f3d80-e51b-11e9-9228-f2a1e56151fe.png)
1.0
Art7 - saving flags - related to #1354, copying last comment here, as this seems to be a backend issue: When creating a submission as secretariat, before Submit there is a popup which allows changing the has_blanks flag: ![image](https://user-images.githubusercontent.com/1218168/66038569-28a09100-e51b-11e9-914f-df622997f8c0.png) ![image](https://user-images.githubusercontent.com/1218168/66038572-2dfddb80-e51b-11e9-81d0-22b9a35521ff.png) Check some previously unchecked flags and submit: ![image](https://user-images.githubusercontent.com/1218168/66038582-33f3bc80-e51b-11e9-8009-453c3281dd34.png) After submit, the flags seem to be correctly updated: ![image](https://user-images.githubusercontent.com/1218168/66038612-4bcb4080-e51b-11e9-8734-a16cd5428e94.png) The PUT request seems to have the right values: ![image](https://user-images.githubusercontent.com/1218168/66038697-8c2abe80-e51b-11e9-9212-f6ae81bf5cd6.png) but after refresh, the flags are reverted to the initial values: ![image](https://user-images.githubusercontent.com/1218168/66038634-600f3d80-e51b-11e9-9228-f2a1e56151fe.png)
priority
saving flags related to copying last comment here as this seems to be a backend issue when creating a submission as secretariat before submit there is a popup which allows changing the has blanks flag check some previously unchecked flags and submit after submit the flags seem to be correctly updated the put request seems to have the right values but after refresh the flags are reverted to the initial values
1
606,278
18,759,032,629
IssuesEvent
2021-11-05 14:26:46
wasmerio/wasmer
https://api.github.com/repos/wasmerio/wasmer
closed
Upgrade Inkwell to `0.1.0-beta.4`
🎉 enhancement 📦 lib-compiler-llvm priority-high
Latest inkwell supports the metadata PR wasmer depends on. So we no longer need to depend on `wasmer-inkwell` This will close the #2433 and allow us to release a new version of Wasmer supporting up to LLVM 13. It should be a relatively easy change (just a version bump)
1.0
Upgrade Inkwell to `0.1.0-beta.4` - Latest inkwell supports the metadata PR wasmer depends on. So we no longer need to depend on `wasmer-inkwell` This will close the #2433 and allow us to release a new version of Wasmer supporting up to LLVM 13. It should be a relatively easy change (just a version bump)
priority
upgrade inkwell to beta latest inkwell supports the metadata pr wasmer depends on so we no longer need to depend on wasmer inkwell this will close the and allow us to release a new version of wasmer supporting up to llvm it should be a relatively easy change just a version bump
1
755,722
26,437,827,103
IssuesEvent
2023-01-15 16:03:46
Thorfusion/Mekanism-1.7.10-Community-Edition
https://api.github.com/repos/Thorfusion/Mekanism-1.7.10-Community-Edition
closed
[BUG]: crash report
TYPE: BUG PRIORITY: HIGH STATUS: FINISHED MC: 1.7.10
### Describe the bug I don't know why the game won't start with the IC2. ### Expected behavior Start the game normally. ### Mekanism Version 9.10.23-ALL ### Minecraft Version is this regarding? 1.7.10 ### What OS are you seeing the problem on? Windows ### Name of modpack if applicable _No response_ ### Version of said modpack if applicable _No response_ ### Screenshots ![image](https://user-images.githubusercontent.com/12579533/212541321-88e102ad-c8b8-41e8-9b1d-2dd668e1ae71.png) ### The crash report in folder ./crash-reports (both server and client logs) crash-2023-01-15_20.46.39-client.txt : [https://pastebin.com/qCr6SWw8](https://pastebin.com/qCr6SWw8) ### Please provide the following other files Use default configuration file. latest.txt : [https://pastebin.com/qFmSRLTV](https://pastebin.com/qFmSRLTV)
1.0
[BUG]: crash report - ### Describe the bug I don't know why the game won't start with the IC2. ### Expected behavior Start the game normally. ### Mekanism Version 9.10.23-ALL ### Minecraft Version is this regarding? 1.7.10 ### What OS are you seeing the problem on? Windows ### Name of modpack if applicable _No response_ ### Version of said modpack if applicable _No response_ ### Screenshots ![image](https://user-images.githubusercontent.com/12579533/212541321-88e102ad-c8b8-41e8-9b1d-2dd668e1ae71.png) ### The crash report in folder ./crash-reports (both server and client logs) crash-2023-01-15_20.46.39-client.txt : [https://pastebin.com/qCr6SWw8](https://pastebin.com/qCr6SWw8) ### Please provide the following other files Use default configuration file. latest.txt : [https://pastebin.com/qFmSRLTV](https://pastebin.com/qFmSRLTV)
priority
crash report describe the bug i don t know why the game won t start with the expected behavior start the game normally mekanism version all minecraft version is this regarding what os are you seeing the problem on windows name of modpack if applicable no response version of said modpack if applicable no response screenshots the crash report in folder crash reports both server and client logs crash client txt please provide the following other files use default configuration file latest txt
1
326,828
9,961,591,617
IssuesEvent
2019-07-07 06:30:32
orbs-network/orbs-network-go
https://api.github.com/repos/orbs-network/orbs-network-go
opened
TestSendSameTransactionFastToTwoNodes is flaky
flakiness high priority
``` t.go:28: �[31;1minfo 2019-07-06T23:06:39.866969Z service sync node=a32884 service=block-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 function=servicesync.syncOneBlock source=services/blockstorage/servicesync/service_sync.go:62 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS t.go:28: �[31;1minfo 2019-07-06T23:06:39.867052Z trying to commit state diff node=a32884 service=state-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 number-of-state-diffs=0 function=statestorage.(*service).CommitStateDiff source=services/statestorage/service.go:85 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS require.go:157: Error Trace: duplicate_tx_test.go:124 duplicate_tx_test.go:88 network_harness_builder.go:159 context.go:23 network_harness_builder.go:143 supervisor.go:60 supervisor.go:54 network_harness_builder.go:139 network_harness_builder.go:123 Error: Not equal: expected: 1 actual : 0 Test: TestSendSameTransactionFastTwiceToSameNode/CONSENSUS_ALGO_TYPE_BENCHMARK_CONSENSUS Messages: blocks should include tx exactly once ``` https://circleci.com/gh/orbs-network/orbs-network-go/16170#tests/containers/3
1.0
TestSendSameTransactionFastToTwoNodes is flaky - ``` t.go:28: �[31;1minfo 2019-07-06T23:06:39.866969Z service sync node=a32884 service=block-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 function=servicesync.syncOneBlock source=services/blockstorage/servicesync/service_sync.go:62 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS t.go:28: �[31;1minfo 2019-07-06T23:06:39.867052Z trying to commit state diff node=a32884 service=state-storage entry-point=state-storage-sync request-id=state-storage-sync-1562454398690812451 block-height=488 number-of-state-diffs=0 function=statestorage.(*service).CommitStateDiff source=services/statestorage/service.go:85 _test=acceptance _test-id=acc-TestSendSameTransactionFastTwiceToSameNode-1562454398-65-BENCHMARK_CONSENSUS require.go:157: Error Trace: duplicate_tx_test.go:124 duplicate_tx_test.go:88 network_harness_builder.go:159 context.go:23 network_harness_builder.go:143 supervisor.go:60 supervisor.go:54 network_harness_builder.go:139 network_harness_builder.go:123 Error: Not equal: expected: 1 actual : 0 Test: TestSendSameTransactionFastTwiceToSameNode/CONSENSUS_ALGO_TYPE_BENCHMARK_CONSENSUS Messages: blocks should include tx exactly once ``` https://circleci.com/gh/orbs-network/orbs-network-go/16170#tests/containers/3
priority
testsendsametransactionfasttotwonodes is flaky t go � service sync node service block storage entry point state storage sync request id state storage sync block height function servicesync synconeblock source services blockstorage servicesync service sync go test acceptance test id acc testsendsametransactionfasttwicetosamenode benchmark consensus t go � trying to commit state diff node service state storage entry point state storage sync request id state storage sync block height number of state diffs function statestorage service commitstatediff source services statestorage service go test acceptance test id acc testsendsametransactionfasttwicetosamenode benchmark consensus require go error trace duplicate tx test go duplicate tx test go network harness builder go context go network harness builder go supervisor go supervisor go network harness builder go network harness builder go error not equal expected actual test testsendsametransactionfasttwicetosamenode consensus algo type benchmark consensus messages blocks should include tx exactly once
1
282,979
8,712,431,870
IssuesEvent
2018-12-06 22:12:34
DaedalusGame/BetterWithAddons
https://api.github.com/repos/DaedalusGame/BetterWithAddons
closed
Book of Single example pictures broken after installation of BWA
high priority
Even the regular BWM example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links. Have tried rebooting, but to no avail. Were showing up just fine before inclusion of BWA into pack. Using the following versions of the mods: BetterWithMods-1.12-2.3.16 & BetterWithLib-1.12-1.5 & Better+With+Addons-0.46 ![oddity](https://user-images.githubusercontent.com/4361068/48102291-622c2780-e1f8-11e8-887f-0a4479a486b6.png)
1.0
Book of Single example pictures broken after installation of BWA - Even the regular BWM example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links. Have tried rebooting, but to no avail. Were showing up just fine before inclusion of BWA into pack. Using the following versions of the mods: BetterWithMods-1.12-2.3.16 & BetterWithLib-1.12-1.5 & Better+With+Addons-0.46 ![oddity](https://user-images.githubusercontent.com/4361068/48102291-622c2780-e1f8-11e8-887f-0a4479a486b6.png)
priority
book of single example pictures broken after installation of bwa even the regular bwm example pics of the kiln structures or the crash course manual pics of the windmills end up as dead red links have tried rebooting but to no avail were showing up just fine before inclusion of bwa into pack using the following versions of the mods betterwithmods betterwithlib better with addons
1
474,485
13,670,906,422
IssuesEvent
2020-09-29 05:58:55
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay-kleinanzeigen.de - site is not usable
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58983 --> **URL**: https://www.ebay-kleinanzeigen.de/ **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: Verweist immer darauf eine App herunter zu laden. Daher muss der Desktopmodus aktiviert werden um die Seite anzeigen zu können <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/e9f97b56-8fdc-40d2-b0ec-7f56cd78a731.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/ccf949b2-368f-4520-9099-f2ade45c3e6a) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay-kleinanzeigen.de - site is not usable - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58983 --> **URL**: https://www.ebay-kleinanzeigen.de/ **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: Verweist immer darauf eine App herunter zu laden. Daher muss der Desktopmodus aktiviert werden um die Seite anzeigen zu können <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/e9f97b56-8fdc-40d2-b0ec-7f56cd78a731.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/ccf949b2-368f-4520-9099-f2ade45c3e6a) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description missing items steps to reproduce verweist immer darauf eine app herunter zu laden daher muss der desktopmodus aktiviert werden um die seite anzeigen zu können view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
327,420
9,975,406,361
IssuesEvent
2019-07-09 13:01:39
python/mypy
https://api.github.com/repos/python/mypy
closed
New semantic analyzer: crash on assignment to sqlalchemy @hybrid_property
crash new-semantic-analyzer priority-0-high
I'm testing out the new semantic analyzer, but it crashes on my codebase. Trying to assign to a [SQLAlchemy `@hybrid_property`](https://docs.sqlalchemy.org/en/13/orm/extensions/hybrid.html)) results in a `Cannot assign to a method` error followed by a crash. Here's a fairly minimal reproduction for it: ```python3 from sqlalchemy import Base, Column, String from sqlalchemy.ext.hybrid import hybrid_property class FirstNameOnly(Base): first_name = Column(String) @hybrid_property def name(self) -> str: return self.first_name @name.setter # type: ignore def name(self, value: str) -> None: self.first_name = value def __init__(self, name: str): self.name = name ``` The `# type: ignore` comment on the setter is a workaround for this issue I reported last year: https://github.com/python/mypy/issues/4430 Here's the full output: ``` (tildes) vagrant@ubuntu-xenial:/opt/tildes$ mypy --new-semantic-analyzer --show-traceback test_mypy.py test_mypy.py:16: error: Cannot assign to a method test_mypy.py:16: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.rtfd.io/en/latest/common_issues.html#using-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.720+dev.48916e63403645730a584d6898fbe925d513a841 Traceback (most recent call last): File "/opt/venvs/tildes/bin/mypy", line 10, in <module> sys.exit(console_entry()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/__main__.py", line 8, in console_entry main(None, sys.stdout, sys.stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/main.py", line 83, in main res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 164, in build result = _build(sources, options, alt_lib_path, flush_errors, fscache, stdout, stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 224, in _build graph = dispatch(sources, manager, stdout) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2567, in dispatch process_graph(graph, manager) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2880, in process_graph process_stale_scc(graph, scc, manager) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2987, in process_stale_scc graph[id].type_check_first_pass() File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2096, in type_check_first_pass self.type_checker().check_first_pass() File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 281, in check_first_pass self.accept(d) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 913, in accept return visitor.visit_class_def(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1596, in visit_class_def self.accept(defn.defs) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept return visitor.visit_block(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block self.accept(s) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 655, in accept return visitor.visit_func_def(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 703, in visit_func_def self._visit_func_def(defn) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 707, in _visit_func_def self.check_func_item(defn, name=defn.name()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 769, in check_func_item self.check_func_def(defn, typ, name) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 935, in check_func_def self.accept(item.body) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept return visitor.visit_block(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block self.accept(s) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 1036, in accept return visitor.visit_assignment_stmt(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1793, in visit_assignment_stmt self.check_assignment(s.lvalues[-1], s.rvalue, s.type is None, s.new_syntax) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1834, in check_assignment lvalue_type, index_lvalue, inferred = self.check_lvalue(lvalue) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 2479, in check_lvalue True) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkexpr.py", line 1766, in analyze_ordinary_member_access in_literal_context=self.is_literal_context()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 103, in analyze_member_access result = _analyze_member_access(name, typ, mx, override_info) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 117, in _analyze_member_access return analyze_instance_member_access(name, typ, mx, override_info) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 179, in analyze_instance_member_access signature = function_type(method, mx.builtin_type('builtins.function')) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/types.py", line 2188, in function_type assert isinstance(func, mypy.nodes.FuncItem), str(func) AssertionError: OverloadedFuncDef:7( Decorator:11( Var(name) MemberExpr:11( NameExpr(name [test_mypy.FirstNameOnly.name]) setter) FuncDef:12( name Args( Var(self) Var(value)) def (self: test_mypy.FirstNameOnly, value: builtins.str) Block:12( AssignmentStmt:13( MemberExpr:13( NameExpr(self [l]) first_name) NameExpr(value [l])))))) ```
1.0
New semantic analyzer: crash on assignment to sqlalchemy @hybrid_property - I'm testing out the new semantic analyzer, but it crashes on my codebase. Trying to assign to a [SQLAlchemy `@hybrid_property`](https://docs.sqlalchemy.org/en/13/orm/extensions/hybrid.html)) results in a `Cannot assign to a method` error followed by a crash. Here's a fairly minimal reproduction for it: ```python3 from sqlalchemy import Base, Column, String from sqlalchemy.ext.hybrid import hybrid_property class FirstNameOnly(Base): first_name = Column(String) @hybrid_property def name(self) -> str: return self.first_name @name.setter # type: ignore def name(self, value: str) -> None: self.first_name = value def __init__(self, name: str): self.name = name ``` The `# type: ignore` comment on the setter is a workaround for this issue I reported last year: https://github.com/python/mypy/issues/4430 Here's the full output: ``` (tildes) vagrant@ubuntu-xenial:/opt/tildes$ mypy --new-semantic-analyzer --show-traceback test_mypy.py test_mypy.py:16: error: Cannot assign to a method test_mypy.py:16: error: INTERNAL ERROR -- Please try using mypy master on Github: https://mypy.rtfd.io/en/latest/common_issues.html#using-development-mypy-build Please report a bug at https://github.com/python/mypy/issues version: 0.720+dev.48916e63403645730a584d6898fbe925d513a841 Traceback (most recent call last): File "/opt/venvs/tildes/bin/mypy", line 10, in <module> sys.exit(console_entry()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/__main__.py", line 8, in console_entry main(None, sys.stdout, sys.stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/main.py", line 83, in main res = build.build(sources, options, None, flush_errors, fscache, stdout, stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 164, in build result = _build(sources, options, alt_lib_path, flush_errors, fscache, stdout, stderr) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 224, in _build graph = dispatch(sources, manager, stdout) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2567, in dispatch process_graph(graph, manager) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2880, in process_graph process_stale_scc(graph, scc, manager) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2987, in process_stale_scc graph[id].type_check_first_pass() File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/build.py", line 2096, in type_check_first_pass self.type_checker().check_first_pass() File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 281, in check_first_pass self.accept(d) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 913, in accept return visitor.visit_class_def(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1596, in visit_class_def self.accept(defn.defs) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept return visitor.visit_block(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block self.accept(s) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 655, in accept return visitor.visit_func_def(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 703, in visit_func_def self._visit_func_def(defn) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 707, in _visit_func_def self.check_func_item(defn, name=defn.name()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 769, in check_func_item self.check_func_def(defn, typ, name) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 935, in check_func_def self.accept(item.body) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 978, in accept return visitor.visit_block(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1785, in visit_block self.accept(s) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 392, in accept stmt.accept(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/nodes.py", line 1036, in accept return visitor.visit_assignment_stmt(self) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1793, in visit_assignment_stmt self.check_assignment(s.lvalues[-1], s.rvalue, s.type is None, s.new_syntax) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 1834, in check_assignment lvalue_type, index_lvalue, inferred = self.check_lvalue(lvalue) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checker.py", line 2479, in check_lvalue True) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkexpr.py", line 1766, in analyze_ordinary_member_access in_literal_context=self.is_literal_context()) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 103, in analyze_member_access result = _analyze_member_access(name, typ, mx, override_info) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 117, in _analyze_member_access return analyze_instance_member_access(name, typ, mx, override_info) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/checkmember.py", line 179, in analyze_instance_member_access signature = function_type(method, mx.builtin_type('builtins.function')) File "/opt/venvs/tildes/lib/python3.7/site-packages/mypy/types.py", line 2188, in function_type assert isinstance(func, mypy.nodes.FuncItem), str(func) AssertionError: OverloadedFuncDef:7( Decorator:11( Var(name) MemberExpr:11( NameExpr(name [test_mypy.FirstNameOnly.name]) setter) FuncDef:12( name Args( Var(self) Var(value)) def (self: test_mypy.FirstNameOnly, value: builtins.str) Block:12( AssignmentStmt:13( MemberExpr:13( NameExpr(self [l]) first_name) NameExpr(value [l])))))) ```
priority
new semantic analyzer crash on assignment to sqlalchemy hybrid property i m testing out the new semantic analyzer but it crashes on my codebase trying to assign to a results in a cannot assign to a method error followed by a crash here s a fairly minimal reproduction for it from sqlalchemy import base column string from sqlalchemy ext hybrid import hybrid property class firstnameonly base first name column string hybrid property def name self str return self first name name setter type ignore def name self value str none self first name value def init self name str self name name the type ignore comment on the setter is a workaround for this issue i reported last year here s the full output tildes vagrant ubuntu xenial opt tildes mypy new semantic analyzer show traceback test mypy py test mypy py error cannot assign to a method test mypy py error internal error please try using mypy master on github please report a bug at version dev traceback most recent call last file opt venvs tildes bin mypy line in sys exit console entry file opt venvs tildes lib site packages mypy main py line in console entry main none sys stdout sys stderr file opt venvs tildes lib site packages mypy main py line in main res build build sources options none flush errors fscache stdout stderr file opt venvs tildes lib site packages mypy build py line in build result build sources options alt lib path flush errors fscache stdout stderr file opt venvs tildes lib site packages mypy build py line in build graph dispatch sources manager stdout file opt venvs tildes lib site packages mypy build py line in dispatch process graph graph manager file opt venvs tildes lib site packages mypy build py line in process graph process stale scc graph scc manager file opt venvs tildes lib site packages mypy build py line in process stale scc graph type check first pass file opt venvs tildes lib site packages mypy build py line in type check first pass self type checker check first pass file opt venvs tildes lib site packages mypy checker py line in check first pass self accept d file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit class def self file opt venvs tildes lib site packages mypy checker py line in visit class def self accept defn defs file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit block self file opt venvs tildes lib site packages mypy checker py line in visit block self accept s file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit func def self file opt venvs tildes lib site packages mypy checker py line in visit func def self visit func def defn file opt venvs tildes lib site packages mypy checker py line in visit func def self check func item defn name defn name file opt venvs tildes lib site packages mypy checker py line in check func item self check func def defn typ name file opt venvs tildes lib site packages mypy checker py line in check func def self accept item body file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit block self file opt venvs tildes lib site packages mypy checker py line in visit block self accept s file opt venvs tildes lib site packages mypy checker py line in accept stmt accept self file opt venvs tildes lib site packages mypy nodes py line in accept return visitor visit assignment stmt self file opt venvs tildes lib site packages mypy checker py line in visit assignment stmt self check assignment s lvalues s rvalue s type is none s new syntax file opt venvs tildes lib site packages mypy checker py line in check assignment lvalue type index lvalue inferred self check lvalue lvalue file opt venvs tildes lib site packages mypy checker py line in check lvalue true file opt venvs tildes lib site packages mypy checkexpr py line in analyze ordinary member access in literal context self is literal context file opt venvs tildes lib site packages mypy checkmember py line in analyze member access result analyze member access name typ mx override info file opt venvs tildes lib site packages mypy checkmember py line in analyze member access return analyze instance member access name typ mx override info file opt venvs tildes lib site packages mypy checkmember py line in analyze instance member access signature function type method mx builtin type builtins function file opt venvs tildes lib site packages mypy types py line in function type assert isinstance func mypy nodes funcitem str func assertionerror overloadedfuncdef decorator var name memberexpr nameexpr name setter funcdef name args var self var value def self test mypy firstnameonly value builtins str block assignmentstmt memberexpr nameexpr self first name nameexpr value
1
697,789
23,952,898,564
IssuesEvent
2022-09-12 12:59:05
benicamera/SupplyManager
https://api.github.com/repos/benicamera/SupplyManager
opened
Implement Item delete and Item create
good first issue Priority: High models business logic
# Tasks - [ ] Create Item and add to list - [ ] Remove Item from list - [ ] Update view ## Create Item and add to list - [ ] Creation form (maybe with image select) - [ ] Add to list ## Remove Item from list - [ ] Swipe or long press - [ ] Confirm question
1.0
Implement Item delete and Item create - # Tasks - [ ] Create Item and add to list - [ ] Remove Item from list - [ ] Update view ## Create Item and add to list - [ ] Creation form (maybe with image select) - [ ] Add to list ## Remove Item from list - [ ] Swipe or long press - [ ] Confirm question
priority
implement item delete and item create tasks create item and add to list remove item from list update view create item and add to list creation form maybe with image select add to list remove item from list swipe or long press confirm question
1
347,216
10,426,653,488
IssuesEvent
2019-09-16 18:05:30
jetrails/magento-cloudflare
https://api.github.com/repos/jetrails/magento-cloudflare
closed
Verify Zone ID Is Valid For Domain
priority: high request
Currently, only token is validated, but we should also check to see if zone id corresponds to the domain name in the current scope.
1.0
Verify Zone ID Is Valid For Domain - Currently, only token is validated, but we should also check to see if zone id corresponds to the domain name in the current scope.
priority
verify zone id is valid for domain currently only token is validated but we should also check to see if zone id corresponds to the domain name in the current scope
1
239,841
7,800,088,230
IssuesEvent
2018-06-09 04:36:41
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0007142: sometimes we filter to much html content
Bug Felamimail Mantis high priority
**Reported by pschuele on 25 Sep 2012 14:02** **Version:** Joey (2012.10.1~beta2) sometimes we filter to much html content -&gt; empty mail
1.0
0007142: sometimes we filter to much html content - **Reported by pschuele on 25 Sep 2012 14:02** **Version:** Joey (2012.10.1~beta2) sometimes we filter to much html content -&gt; empty mail
priority
sometimes we filter to much html content reported by pschuele on sep version joey sometimes we filter to much html content gt empty mail
1
85,957
3,700,957,823
IssuesEvent
2016-02-29 10:57:08
uds-datalab/PDBF
https://api.github.com/repos/uds-datalab/PDBF
closed
On Ubuntu with old version of TexLive mvn verify results in error
1-high-priority bug wontfix
If you encounter an error message similar to this one: > The file is not valid, error(s) : > 1.2.1 : Body Syntax error, Single space expected [offset=2786901; key=2786901; line=5 0 obj <<; object=COSObject{5, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2786635; key=2786635; line=11 0 obj <<; object=COSObject{11, 0}] > 1.2.1 : Body Syntax error, EOL expected before the 'endobj' keyword at offset 2786894 > 1.2.1 : Body Syntax error, Single space expected [offset=2761071; key=2761071; line=3 0 obj <<; object=COSObject{3, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2761433; key=2761433; line=8 0 obj <<; object=COSObject{8, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2787764; key=2787764; line=12 0 obj <<; object=COSObject{12, 0}] then dont worry. This is a known issue with old versions of TexLive. You can safely ignore it or upgrade your TexLive to an up to date version (Should be fixed in TexLive 2015).
1.0
On Ubuntu with old version of TexLive mvn verify results in error - If you encounter an error message similar to this one: > The file is not valid, error(s) : > 1.2.1 : Body Syntax error, Single space expected [offset=2786901; key=2786901; line=5 0 obj <<; object=COSObject{5, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2786635; key=2786635; line=11 0 obj <<; object=COSObject{11, 0}] > 1.2.1 : Body Syntax error, EOL expected before the 'endobj' keyword at offset 2786894 > 1.2.1 : Body Syntax error, Single space expected [offset=2761071; key=2761071; line=3 0 obj <<; object=COSObject{3, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2761433; key=2761433; line=8 0 obj <<; object=COSObject{8, 0}] > 1.2.1 : Body Syntax error, Single space expected [offset=2787764; key=2787764; line=12 0 obj <<; object=COSObject{12, 0}] then dont worry. This is a known issue with old versions of TexLive. You can safely ignore it or upgrade your TexLive to an up to date version (Should be fixed in TexLive 2015).
priority
on ubuntu with old version of texlive mvn verify results in error if you encounter an error message similar to this one the file is not valid error s body syntax error single space expected body syntax error single space expected body syntax error eol expected before the endobj keyword at offset body syntax error single space expected body syntax error single space expected body syntax error single space expected then dont worry this is a known issue with old versions of texlive you can safely ignore it or upgrade your texlive to an up to date version should be fixed in texlive
1
84,180
3,654,789,606
IssuesEvent
2016-02-17 14:06:23
emoncms/MyHomeEnergyPlanner
https://api.github.com/repos/emoncms/MyHomeEnergyPlanner
closed
Fabric Measures
feature High priority
Hi Carlos, I've created a spreadsheet with some text measures in it as requested. I've also had a think about the labels for the form fields - see notes in the spreadsheet. Hopefully self-explanatory, and enough for you to be getting on with. Thanks, [20160208_Test Measures List.xlsx](https://github.com/emoncms/MyHomeEnergyPlanner/files/121593/20160208_Test.Measures.List.xlsx)
1.0
Fabric Measures - Hi Carlos, I've created a spreadsheet with some text measures in it as requested. I've also had a think about the labels for the form fields - see notes in the spreadsheet. Hopefully self-explanatory, and enough for you to be getting on with. Thanks, [20160208_Test Measures List.xlsx](https://github.com/emoncms/MyHomeEnergyPlanner/files/121593/20160208_Test.Measures.List.xlsx)
priority
fabric measures hi carlos i ve created a spreadsheet with some text measures in it as requested i ve also had a think about the labels for the form fields see notes in the spreadsheet hopefully self explanatory and enough for you to be getting on with thanks
1
710,806
24,435,502,110
IssuesEvent
2022-10-06 11:10:55
hackforla/expunge-assist
https://api.github.com/repos/hackforla/expunge-assist
reopened
Review auto-generated text for repetition [from usability testing]
priority: high role: UX content writing feature: figma content writing size: 5pt
### Overview Auto-generated text needs to be reviewed and updated. For example, one user pointed out that several responses began with "Since my conviction…" Another user noticed repetitive sentences under the "Involvement: Job" section. ### Action Items - [x] Review auto-generated text - [x] Identify areas of repetition - [x] Create new copy for those areas - [x] Collaborate with Dev regarding the creation of randomly selected text (i.e. one sentence could be written in 3 ways and each user gets a randomly selected text inserted - this helps with creating variation in the letters that only 1-2 judges will see). (Answer from Dev: Cannot do right now) - [x] Consider creating multiple sentence starters/fragments/etc. that users could choose from (personalize/make more authentic to each user). (Not for this iteration - revisit next) - [x] Discuss in Content/iterate - [x] Finalize - [x] Link all appropriate documents/figma pages/etc. in the resource section below - [ ] Hand over to Dev https://github.com/hackforla/expunge-assist/issues/705 ### Resources/Instructions This is for Form Fields Inconsistencies and Repetitiveness [Google Doc](https://docs.google.com/document/d/1UAjwLopUswtOleJrwB-oyuqyUR4mhOF08x8jR9AfJnk/edit?usp=sharing) Continuing this work directly in Figma under the WIP LG page
1.0
Review auto-generated text for repetition [from usability testing] - ### Overview Auto-generated text needs to be reviewed and updated. For example, one user pointed out that several responses began with "Since my conviction…" Another user noticed repetitive sentences under the "Involvement: Job" section. ### Action Items - [x] Review auto-generated text - [x] Identify areas of repetition - [x] Create new copy for those areas - [x] Collaborate with Dev regarding the creation of randomly selected text (i.e. one sentence could be written in 3 ways and each user gets a randomly selected text inserted - this helps with creating variation in the letters that only 1-2 judges will see). (Answer from Dev: Cannot do right now) - [x] Consider creating multiple sentence starters/fragments/etc. that users could choose from (personalize/make more authentic to each user). (Not for this iteration - revisit next) - [x] Discuss in Content/iterate - [x] Finalize - [x] Link all appropriate documents/figma pages/etc. in the resource section below - [ ] Hand over to Dev https://github.com/hackforla/expunge-assist/issues/705 ### Resources/Instructions This is for Form Fields Inconsistencies and Repetitiveness [Google Doc](https://docs.google.com/document/d/1UAjwLopUswtOleJrwB-oyuqyUR4mhOF08x8jR9AfJnk/edit?usp=sharing) Continuing this work directly in Figma under the WIP LG page
priority
review auto generated text for repetition overview auto generated text needs to be reviewed and updated for example one user pointed out that several responses began with since my conviction… another user noticed repetitive sentences under the involvement job section action items review auto generated text identify areas of repetition create new copy for those areas collaborate with dev regarding the creation of randomly selected text i e one sentence could be written in ways and each user gets a randomly selected text inserted this helps with creating variation in the letters that only judges will see answer from dev cannot do right now consider creating multiple sentence starters fragments etc that users could choose from personalize make more authentic to each user not for this iteration revisit next discuss in content iterate finalize link all appropriate documents figma pages etc in the resource section below hand over to dev resources instructions this is for form fields inconsistencies and repetitiveness continuing this work directly in figma under the wip lg page
1
175,675
6,552,937,745
IssuesEvent
2017-09-05 20:22:08
envistaInteractive/itagroup-ecommerce-template
https://api.github.com/repos/envistaInteractive/itagroup-ecommerce-template
opened
Events: Category / Search Page
High Priority Page Layout
### Summary Layout contents of Events: Category / Search page as specified on Events: Category / Search in Zeplin. We do not have color mockups. The top bar is the blue that is also used on the checkout pages. Use those same classes and html. We will move that out of the checkout to be more generic later. Use a mobile first approach to adjust the layout using responsive design as the screen gets larger. **Use branch**: feature/events **Layout file**: templates/events/list.liquid (file does not exist) **Url for testing**: http://localhost:1337/events **Delivery Date**: Sept 7th
1.0
Events: Category / Search Page - ### Summary Layout contents of Events: Category / Search page as specified on Events: Category / Search in Zeplin. We do not have color mockups. The top bar is the blue that is also used on the checkout pages. Use those same classes and html. We will move that out of the checkout to be more generic later. Use a mobile first approach to adjust the layout using responsive design as the screen gets larger. **Use branch**: feature/events **Layout file**: templates/events/list.liquid (file does not exist) **Url for testing**: http://localhost:1337/events **Delivery Date**: Sept 7th
priority
events category search page summary layout contents of events category search page as specified on events category search in zeplin we do not have color mockups the top bar is the blue that is also used on the checkout pages use those same classes and html we will move that out of the checkout to be more generic later use a mobile first approach to adjust the layout using responsive design as the screen gets larger use branch feature events layout file templates events list liquid file does not exist url for testing delivery date sept
1
347,031
10,423,479,679
IssuesEvent
2019-09-16 11:32:36
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Server always crashes/doesn't load
High Priority
When trying to start a server it always crashes and I get the message that errors ocurred loading users. Already tried to deinstall both Eco and Eco Server as well as restart multiple times. Nothing seems to help, can't play on the server with my friends at the moment. The log with error message is attached. Might need some help! [log_190915072601.log](https://github.com/StrangeLoopGames/EcoIssues/files/3615405/log_190915072601.log)
1.0
Server always crashes/doesn't load - When trying to start a server it always crashes and I get the message that errors ocurred loading users. Already tried to deinstall both Eco and Eco Server as well as restart multiple times. Nothing seems to help, can't play on the server with my friends at the moment. The log with error message is attached. Might need some help! [log_190915072601.log](https://github.com/StrangeLoopGames/EcoIssues/files/3615405/log_190915072601.log)
priority
server always crashes doesn t load when trying to start a server it always crashes and i get the message that errors ocurred loading users already tried to deinstall both eco and eco server as well as restart multiple times nothing seems to help can t play on the server with my friends at the moment the log with error message is attached might need some help
1
831,773
32,060,525,832
IssuesEvent
2023-09-24 15:57:01
oncokb/oncokb
https://api.github.com/repos/oncokb/oncokb
opened
Some OncoKB genes do not have ensembl gene curated
bug high priority
All OncoKB genes should have ensembl gene transcript, otherwise the genomic change will be filtered out for annotation(missing chromosome/start/end). Therefore the genomic change and hgvsg annotation will not work.
1.0
Some OncoKB genes do not have ensembl gene curated - All OncoKB genes should have ensembl gene transcript, otherwise the genomic change will be filtered out for annotation(missing chromosome/start/end). Therefore the genomic change and hgvsg annotation will not work.
priority
some oncokb genes do not have ensembl gene curated all oncokb genes should have ensembl gene transcript otherwise the genomic change will be filtered out for annotation missing chromosome start end therefore the genomic change and hgvsg annotation will not work
1
345,629
10,370,688,866
IssuesEvent
2019-09-08 14:41:29
byaka/VombatiDB
https://api.github.com/repos/byaka/VombatiDB
opened
Добавить режим хранения индекса `wide` при котором выделяется ячейка на хранение ссылки на данные
high-priority improvement optimization
В таком режиме все методы доступа к индексу будут разпаковывать ноду не на 2 обьекта (`props, childs`), а на 3 (`..,data`). При этом распакованный третий обьект станет передаваться в `_getData()` и аналоги. **Похоже, реализация этого как отдельного режима работы ядра будет тяжелой изза того, что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок. В таком случае лучше выкинуть старый режим.**
1.0
Добавить режим хранения индекса `wide` при котором выделяется ячейка на хранение ссылки на данные - В таком режиме все методы доступа к индексу будут разпаковывать ноду не на 2 обьекта (`props, childs`), а на 3 (`..,data`). При этом распакованный третий обьект станет передаваться в `_getData()` и аналоги. **Похоже, реализация этого как отдельного режима работы ядра будет тяжелой изза того, что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок. В таком случае лучше выкинуть старый режим.**
priority
добавить режим хранения индекса wide при котором выделяется ячейка на хранение ссылки на данные в таком режиме все методы доступа к индексу будут разпаковывать ноду не на обьекта props childs а на data при этом распакованный третий обьект станет передаваться в getdata и аналоги похоже реализация этого как отдельного режима работы ядра будет тяжелой изза того что придется усложнять код передачи аргументов и в целом раздувать кодовую базу кучей однотипных проверок в таком случае лучше выкинуть старый режим
1
67,302
3,268,440,533
IssuesEvent
2015-10-23 11:30:51
pakalbekim/armaldia
https://api.github.com/repos/pakalbekim/armaldia
opened
Rebalance gym once again
High Priority
When increasing energy limit - use a lot of stamina; when increasing stamina limit - use a lot of energy; when increasing health limit - use moderate amounts of both;
1.0
Rebalance gym once again - When increasing energy limit - use a lot of stamina; when increasing stamina limit - use a lot of energy; when increasing health limit - use moderate amounts of both;
priority
rebalance gym once again when increasing energy limit use a lot of stamina when increasing stamina limit use a lot of energy when increasing health limit use moderate amounts of both
1
487,329
14,040,547,886
IssuesEvent
2020-11-01 03:18:34
xournalpp/xournalpp
https://api.github.com/repos/xournalpp/xournalpp
closed
Recently used file with "no such device" causes Xournal++ to crash at start.
Crash bug priority: high
**Affects versions :** - OS: Arch Linux - Desktop environment: Gnome-Wayland - Version of Xournal++: 45a619d83f97205c92ac146f13d5ae00af83af7e - Installation method: AUR ([xournalpp-git](https://aur.archlinux.org/packages/xournalpp-git/)) **Describe the bug** When a file exists in the "recently used files" where the device is not available (f.e. a Samba mount) Xournal++ crashes at start. **To Reproduce** Steps to reproduce the behavior: 1. Mount a samba drive using a VPN interface 2. Open the file in Xournal++, close it 3. Make sure the VPN interface is not available without unmounting the Samba drive (`no such device`) 4. Re-open Xournal++ **Expected behavior** Xournal++ ignores the unavailable file. **Additional context** Crash: ``` terminate called after throwing an instance of 'std::filesystem::__cxx11::filesystem_error' what(): filesystem error: status: No such device [/mnt/samba/mount/that/is/not/available/file.pdf] ** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.368: [Crash Handler] Crashed with signal 6 ** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.369: [Crash Handler] Wrote crash log to: $HOME/.cache/com.github.xournalpp.xournalpp/errorlogs/errorlog.20200908-123929.log ``` Error log: https://fb.hash.works/HMqP0W File path is from `./.local/share/recently-used.xbel`: ``` <bookmark href="file:///mnt/samba/mount/that/is/not/available/file.pdf" added="2020-09-02T23:54:55Z" modified="2020-09-03T00:00:03Z" visited="1969-12-31T23:59:59Z"> ``` This sounds like something that should've been fixed with #1730, but as I stated above I'm using the latest version in master.
1.0
Recently used file with "no such device" causes Xournal++ to crash at start. - **Affects versions :** - OS: Arch Linux - Desktop environment: Gnome-Wayland - Version of Xournal++: 45a619d83f97205c92ac146f13d5ae00af83af7e - Installation method: AUR ([xournalpp-git](https://aur.archlinux.org/packages/xournalpp-git/)) **Describe the bug** When a file exists in the "recently used files" where the device is not available (f.e. a Samba mount) Xournal++ crashes at start. **To Reproduce** Steps to reproduce the behavior: 1. Mount a samba drive using a VPN interface 2. Open the file in Xournal++, close it 3. Make sure the VPN interface is not available without unmounting the Samba drive (`no such device`) 4. Re-open Xournal++ **Expected behavior** Xournal++ ignores the unavailable file. **Additional context** Crash: ``` terminate called after throwing an instance of 'std::filesystem::__cxx11::filesystem_error' what(): filesystem error: status: No such device [/mnt/samba/mount/that/is/not/available/file.pdf] ** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.368: [Crash Handler] Crashed with signal 6 ** (com.github.xournalpp.xournalpp:26191): WARNING **: 12:39:29.369: [Crash Handler] Wrote crash log to: $HOME/.cache/com.github.xournalpp.xournalpp/errorlogs/errorlog.20200908-123929.log ``` Error log: https://fb.hash.works/HMqP0W File path is from `./.local/share/recently-used.xbel`: ``` <bookmark href="file:///mnt/samba/mount/that/is/not/available/file.pdf" added="2020-09-02T23:54:55Z" modified="2020-09-03T00:00:03Z" visited="1969-12-31T23:59:59Z"> ``` This sounds like something that should've been fixed with #1730, but as I stated above I'm using the latest version in master.
priority
recently used file with no such device causes xournal to crash at start affects versions os arch linux desktop environment gnome wayland version of xournal installation method aur describe the bug when a file exists in the recently used files where the device is not available f e a samba mount xournal crashes at start to reproduce steps to reproduce the behavior mount a samba drive using a vpn interface open the file in xournal close it make sure the vpn interface is not available without unmounting the samba drive no such device re open xournal expected behavior xournal ignores the unavailable file additional context crash terminate called after throwing an instance of std filesystem filesystem error what filesystem error status no such device com github xournalpp xournalpp warning crashed with signal com github xournalpp xournalpp warning wrote crash log to home cache com github xournalpp xournalpp errorlogs errorlog log error log file path is from local share recently used xbel this sounds like something that should ve been fixed with but as i stated above i m using the latest version in master
1
2,438
2,525,857,317
IssuesEvent
2015-01-21 06:51:08
graybeal/ont
https://api.github.com/repos/graybeal/ont
closed
Allow tab delimiter when importing data in voc2rdf
1 star enhancement imported Milestone-Release1.2 Priority-High voc2rdf
_From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on April 06, 2009 11:51:45_ (thanks John for this feedback) What capability do you want added or improved? Allow tab delimiter when importing data in voc2rdf Where do you want this capability to be accessible? voc2rdf What sort of input/command mechanism do you want? In the CSV dialog, have a checkbox or something to indicate that the contents are tab-delimited columns _Original issue: http://code.google.com/p/mmisw/issues/detail?id=115_
1.0
Allow tab delimiter when importing data in voc2rdf - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on April 06, 2009 11:51:45_ (thanks John for this feedback) What capability do you want added or improved? Allow tab delimiter when importing data in voc2rdf Where do you want this capability to be accessible? voc2rdf What sort of input/command mechanism do you want? In the CSV dialog, have a checkbox or something to indicate that the contents are tab-delimited columns _Original issue: http://code.google.com/p/mmisw/issues/detail?id=115_
priority
allow tab delimiter when importing data in from on april thanks john for this feedback what capability do you want added or improved allow tab delimiter when importing data in where do you want this capability to be accessible what sort of input command mechanism do you want in the csv dialog have a checkbox or something to indicate that the contents are tab delimited columns original issue
1
421,761
12,261,138,679
IssuesEvent
2020-05-06 19:32:34
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
configurator: MariaDB doesn't use timezone setting until reboot
Priority: High Type: Bug
**Describe the bug** When you set a timezone in configurator, MariaDB is not configured with that timezone until you reboot. Consequently, admin user created in DB is created with a `valid_from` value based on **default** timezone. Depending on your timezone, it means that your account could be not valid after you reboot. **To Reproduce** Steps to reproduce the behavior: 1. Install a PacketFence ZEN 2. Set timezone to EST at step 2 of configurator 3. Set admin password 4. Check value in DB: ```sql SELECT valid_from FROM password where pid='admin'\G; ``` 5. Finish configurator 6. Log in on web admin with `admin` user => It works. 7. Reboot 8. Log in on web admin with `admin` user => It fails because account is not yet valid. **Expected behavior** MariaDB should be configured using timezone defined at step 2.
1.0
configurator: MariaDB doesn't use timezone setting until reboot - **Describe the bug** When you set a timezone in configurator, MariaDB is not configured with that timezone until you reboot. Consequently, admin user created in DB is created with a `valid_from` value based on **default** timezone. Depending on your timezone, it means that your account could be not valid after you reboot. **To Reproduce** Steps to reproduce the behavior: 1. Install a PacketFence ZEN 2. Set timezone to EST at step 2 of configurator 3. Set admin password 4. Check value in DB: ```sql SELECT valid_from FROM password where pid='admin'\G; ``` 5. Finish configurator 6. Log in on web admin with `admin` user => It works. 7. Reboot 8. Log in on web admin with `admin` user => It fails because account is not yet valid. **Expected behavior** MariaDB should be configured using timezone defined at step 2.
priority
configurator mariadb doesn t use timezone setting until reboot describe the bug when you set a timezone in configurator mariadb is not configured with that timezone until you reboot consequently admin user created in db is created with a valid from value based on default timezone depending on your timezone it means that your account could be not valid after you reboot to reproduce steps to reproduce the behavior install a packetfence zen set timezone to est at step of configurator set admin password check value in db sql select valid from from password where pid admin g finish configurator log in on web admin with admin user it works reboot log in on web admin with admin user it fails because account is not yet valid expected behavior mariadb should be configured using timezone defined at step
1
636,792
20,609,351,789
IssuesEvent
2022-03-07 06:34:22
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[FEATURE] Soft reboot/shutdown
enhancement area/ui priority/1 highlight area/kubevirt
We should support a graceful soft reboot/shutdown from the UI, to allow the VM and filesystem on it to have a chance to shutdown properly. Guest agent might be required.
1.0
[FEATURE] Soft reboot/shutdown - We should support a graceful soft reboot/shutdown from the UI, to allow the VM and filesystem on it to have a chance to shutdown properly. Guest agent might be required.
priority
soft reboot shutdown we should support a graceful soft reboot shutdown from the ui to allow the vm and filesystem on it to have a chance to shutdown properly guest agent might be required
1
129,076
5,088,229,403
IssuesEvent
2016-12-31 16:50:06
zulip/zulip-electron
https://api.github.com/repos/zulip/zulip-electron
closed
Can not find module 'debug/browser'
bug help wanted Priority: High
Something is broken and I'm not able to figure it out 😭 ![image](https://cloud.githubusercontent.com/assets/2263909/21577415/af9f0434-cf7f-11e6-817f-bbc88a7b6657.png) Because of above error preload script can't be injected and hence spellchecker won't work 😢
1.0
Can not find module 'debug/browser' - Something is broken and I'm not able to figure it out 😭 ![image](https://cloud.githubusercontent.com/assets/2263909/21577415/af9f0434-cf7f-11e6-817f-bbc88a7b6657.png) Because of above error preload script can't be injected and hence spellchecker won't work 😢
priority
can not find module debug browser something is broken and i m not able to figure it out 😭 because of above error preload script can t be injected and hence spellchecker won t work 😢
1
283,027
8,713,246,379
IssuesEvent
2018-12-07 01:40:51
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
visit2.9.2 release tarball[s?] are doubly-compressed
bug likelihood high priority reviewed severity medium wontfix
The release tarball: http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz is gzipped *twice*. That is, to extract one must: $ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug? -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2307 Status: Rejected Project: VisIt Tracker: Bug Priority: High Subject: visit2.9.2 release tarball[s?] are doubly-compressed Assigned to: Category: Target version: 2.10 Author: Tom Fogal Start: 06/23/2015 Due date: % Done: 0 Estimated time: Created: 06/23/2015 09:34 pm Updated: 08/26/2015 08:39 pm Likelihood: 4 - Common Severity: 3 - Major Irritation Found in version: 2.9.2 Impact: Expected Use: OS: All Support Group: Any Description: The release tarball: http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz is gzipped *twice*. That is, to extract one must: $ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug? Comments: Hi Tom,I have double checked and the original tar file and when I do a "tar zxf visit2.9.2.linux-x86_64-rhel6.tar.gz" this uncompresses fine. I have run into issues where I download the file and doing a "tar zxf ..." gives an error. When I do it as 2 steps "gunzip visit2.9.2...; tar xf visit2.9.2..." it works fine. My thought was that my browser or the website was compressing it a second time. Can you let me know the md5sum of the file you downloaded? I just verified that this behavior still exists with the 2.9.2 rhel6 download. When I look at the md5sum it is different from what is stored on the server at nersc. I am doing my download on Windows using IE. It also happens with Firefox on RHEL6. I gunzipped the file it downloaded and then the checksum matched. So, yes, it really is getting gzipped a second time. I will need to contact NERSC and see if they know why this is happening. I just downloaded http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz via both firefox and in a terminal via wget. Via wget the file is fine. Via firefox I get the double-compressed issue. We need to contact NERSC and find out how to turn this off. This is no longer happening. I checked on Windows with IE and Linux with Firefox and Kevin checked on the Mac.
1.0
visit2.9.2 release tarball[s?] are doubly-compressed - The release tarball: http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz is gzipped *twice*. That is, to extract one must: $ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug? -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2307 Status: Rejected Project: VisIt Tracker: Bug Priority: High Subject: visit2.9.2 release tarball[s?] are doubly-compressed Assigned to: Category: Target version: 2.10 Author: Tom Fogal Start: 06/23/2015 Due date: % Done: 0 Estimated time: Created: 06/23/2015 09:34 pm Updated: 08/26/2015 08:39 pm Likelihood: 4 - Common Severity: 3 - Major Irritation Found in version: 2.9.2 Impact: Expected Use: OS: All Support Group: Any Description: The release tarball: http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz is gzipped *twice*. That is, to extract one must: $ gunzip visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ mv visit2_9_2.linux-x86_64-ubuntu14.tar visit2_9_2.linux-x86_64-ubuntu14.tar.gz $ tar zxvf visit2_9_2.linux-x86_64-ubuntu14.tar.gz I have a vague recollection that I hit this with 2.9.1 as well. Perhaps there is a packaging script bug? Comments: Hi Tom,I have double checked and the original tar file and when I do a "tar zxf visit2.9.2.linux-x86_64-rhel6.tar.gz" this uncompresses fine. I have run into issues where I download the file and doing a "tar zxf ..." gives an error. When I do it as 2 steps "gunzip visit2.9.2...; tar xf visit2.9.2..." it works fine. My thought was that my browser or the website was compressing it a second time. Can you let me know the md5sum of the file you downloaded? I just verified that this behavior still exists with the 2.9.2 rhel6 download. When I look at the md5sum it is different from what is stored on the server at nersc. I am doing my download on Windows using IE. It also happens with Firefox on RHEL6. I gunzipped the file it downloaded and then the checksum matched. So, yes, it really is getting gzipped a second time. I will need to contact NERSC and see if they know why this is happening. I just downloaded http://portal.nersc.gov/project/visit/releases/2.9.2/visit2_9_2.linux-x86_64-ubuntu14.tar.gz via both firefox and in a terminal via wget. Via wget the file is fine. Via firefox I get the double-compressed issue. We need to contact NERSC and find out how to turn this off. This is no longer happening. I checked on Windows with IE and Linux with Firefox and Kevin checked on the Mac.
priority
release tarball are doubly compressed the release tarball is gzipped twice that is to extract one must gunzip linux tar gz mv linux tar linux tar gz tar zxvf linux tar gz i have a vague recollection that i hit this with as well perhaps there is a packaging script bug redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status rejected project visit tracker bug priority high subject release tarball are doubly compressed assigned to category target version author tom fogal start due date done estimated time created pm updated pm likelihood common severity major irritation found in version impact expected use os all support group any description the release tarball is gzipped twice that is to extract one must gunzip linux tar gz mv linux tar linux tar gz tar zxvf linux tar gz i have a vague recollection that i hit this with as well perhaps there is a packaging script bug comments hi tom i have double checked and the original tar file and when i do a tar zxf linux tar gz this uncompresses fine i have run into issues where i download the file and doing a tar zxf gives an error when i do it as steps gunzip tar xf it works fine my thought was that my browser or the website was compressing it a second time can you let me know the of the file you downloaded i just verified that this behavior still exists with the download when i look at the it is different from what is stored on the server at nersc i am doing my download on windows using ie it also happens with firefox on i gunzipped the file it downloaded and then the checksum matched so yes it really is getting gzipped a second time i will need to contact nersc and see if they know why this is happening i just downloaded via both firefox and in a terminal via wget via wget the file is fine via firefox i get the double compressed issue we need to contact nersc and find out how to turn this off this is no longer happening i checked on windows with ie and linux with firefox and kevin checked on the mac
1
124,396
4,913,475,542
IssuesEvent
2016-11-23 12:41:30
commercialhaskell/intero
https://api.github.com/repos/commercialhaskell/intero
closed
Not working for docker stack projects
component: emacs priority: high type: bug
I have a project using the "basic" stack template. When I build it with ``` yaml docker: enable: true ``` in my `stack.yaml` file intero stops working. Turning off docker support, deleting `.stack-work` and rebuilding and intero starts working again. This is using the latest Docker for Mac with intero 0.1.18 for reference.
1.0
Not working for docker stack projects - I have a project using the "basic" stack template. When I build it with ``` yaml docker: enable: true ``` in my `stack.yaml` file intero stops working. Turning off docker support, deleting `.stack-work` and rebuilding and intero starts working again. This is using the latest Docker for Mac with intero 0.1.18 for reference.
priority
not working for docker stack projects i have a project using the basic stack template when i build it with yaml docker enable true in my stack yaml file intero stops working turning off docker support deleting stack work and rebuilding and intero starts working again this is using the latest docker for mac with intero for reference
1
403,667
11,844,572,962
IssuesEvent
2020-03-24 06:12:31
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
Implement automatic cleanup of TRANSFER documents from ReqMgrAux
High Priority New Feature ReqMgr2MS Unified Porting
**Impact of the new feature** ReqMgr2MS **Is your feature request related to a problem? Please describe.** Given that every request being handled by MSTransferor will result in one transfer info document in reqmgr_aux couch db, we need to have a mechanism that automatically deletes those documents, otherwise MSTransferor/MSMonitor might suffer of performance issues in the future. **Describe the solution you'd like** We need to decide at which status we should delete those files. For the moment, I think it will be useful to keep those documents for a bit longer in the system, such that we can get back to them if needed and check what were the transfer requests created. Thus, I'd suggest to cleanup those documents once the workflow is done processing. I.e., if the workflow is NOT in `('staging', 'staged', 'acquired', 'running-open', 'running-closed')`, then the transfer info document can be deleted. By the way, we should probably create a new cherrypy thread to take care of this cleanup. Running it twice a day is more than enough. **Describe alternatives you've considered** In the future, when this service is robust and stable, we can start deleting those transfer documents as soon as the workflow starts running. **Additional context** none
1.0
Implement automatic cleanup of TRANSFER documents from ReqMgrAux - **Impact of the new feature** ReqMgr2MS **Is your feature request related to a problem? Please describe.** Given that every request being handled by MSTransferor will result in one transfer info document in reqmgr_aux couch db, we need to have a mechanism that automatically deletes those documents, otherwise MSTransferor/MSMonitor might suffer of performance issues in the future. **Describe the solution you'd like** We need to decide at which status we should delete those files. For the moment, I think it will be useful to keep those documents for a bit longer in the system, such that we can get back to them if needed and check what were the transfer requests created. Thus, I'd suggest to cleanup those documents once the workflow is done processing. I.e., if the workflow is NOT in `('staging', 'staged', 'acquired', 'running-open', 'running-closed')`, then the transfer info document can be deleted. By the way, we should probably create a new cherrypy thread to take care of this cleanup. Running it twice a day is more than enough. **Describe alternatives you've considered** In the future, when this service is robust and stable, we can start deleting those transfer documents as soon as the workflow starts running. **Additional context** none
priority
implement automatic cleanup of transfer documents from reqmgraux impact of the new feature is your feature request related to a problem please describe given that every request being handled by mstransferor will result in one transfer info document in reqmgr aux couch db we need to have a mechanism that automatically deletes those documents otherwise mstransferor msmonitor might suffer of performance issues in the future describe the solution you d like we need to decide at which status we should delete those files for the moment i think it will be useful to keep those documents for a bit longer in the system such that we can get back to them if needed and check what were the transfer requests created thus i d suggest to cleanup those documents once the workflow is done processing i e if the workflow is not in staging staged acquired running open running closed then the transfer info document can be deleted by the way we should probably create a new cherrypy thread to take care of this cleanup running it twice a day is more than enough describe alternatives you ve considered in the future when this service is robust and stable we can start deleting those transfer documents as soon as the workflow starts running additional context none
1
168,489
6,376,540,156
IssuesEvent
2017-08-02 07:45:48
aio-libs/aiokafka
https://api.github.com/repos/aio-libs/aiokafka
closed
Consumer and Coordinator should use a separate socket for connections
priority:high
Followup dev ticket for #128. See also https://issues.apache.org/jira/browse/KAFKA-1925
1.0
Consumer and Coordinator should use a separate socket for connections - Followup dev ticket for #128. See also https://issues.apache.org/jira/browse/KAFKA-1925
priority
consumer and coordinator should use a separate socket for connections followup dev ticket for see also
1
353,787
10,558,799,694
IssuesEvent
2019-10-04 09:56:16
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
ID is shown as an object when fetching models with nested relationships
priority: high status: confirmed type: bug 🐛
**Describe the bug** If you create a group and one of the fields is a relationship with a model which have inside more relationships, when you fetch this data, it shows the id of the nested relationship as a object with the order of the character as key and the character as a value, in the same way as if you do {..."idString"} **Steps to reproduce the behavior** 1. Create a group with a relation to a model with more relationships 2. Add the group to a model **Expected behavior** The nested relationship should show the id of the related content as a normal string **Postman response** ``` [ { "_id": "5d8db9bf0dc976285007b04d", "items": [ { "_id": "5d8db9bf0dc976285007b04e", "notes": "qwe", "qty": 2, "createdAt": "2019-09-27T07:26:55.574Z", "updatedAt": "2019-09-27T07:26:55.601Z", "__v": 0, "product": { "_id": "5d6fbbe94e7c9554792b22b4", "name": "product11", "description": "tggyyfff77", "createdAt": "2019-09-04T13:28:09.311Z", "updatedAt": "2019-09-17T13:02:42.835Z", "__v": 0, "organization": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "2", "5": "f", "6": "e", "7": "2", "8": "8", "9": "e", "10": "6", "11": "7", "12": "5", "13": "7", "14": "3", "15": "8", "16": "7", "17": "a", "18": "c", "19": "7", "20": "6", "21": "0", "22": "3", "23": "7" }, "productgroup": { "0": "5", "1": "d", "2": "6", "3": "f", "4": "b", "5": "b", "6": "d", "7": "7", "8": "4", "9": "e", "10": "7", "11": "c", "12": "9", "13": "5", "14": "5", "15": "4", "16": "7", "17": "9", "18": "2", "19": "b", "20": "2", "21": "2", "22": "b", "23": "2" }, "factory": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "4", "5": "5", "6": "7", "7": "c", "8": "7", "9": "6", "10": "2", "11": "f", "12": "2", "13": "d", "14": "4", "15": "8", "16": "f", "17": "0", "18": "6", "19": "8", "20": "8", "21": "8", "22": "6", "23": "d" }, "client": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "4", "5": "2", "6": "4", "7": "0", "8": "2", "9": "8", "10": "8", "11": "e", "12": "d", "13": "e", "14": "4", "15": "7", "16": "1", "17": "e", "18": "7", "19": "2", "20": "6", "21": "1", "22": "0", "23": "3" }, "sku": "12345", "id": "5d6fbbe94e7c9554792b22b4" }, "id": "5d8db9bf0dc976285007b04e" } ]``` **System** - Node.js version: v10.5 - NPM version: v6.7 - Strapi version: Beta 16.6 - Database: MongoDB - Operating system: Windows 10 **Possible solution** Now the id recovers its normal form with Object.values(strangeObjectID).reduce((a,b) => a+b, ""), maybe it lacks this in some point of the code. Or maybe there's a playful spread operator messing the string ID
1.0
ID is shown as an object when fetching models with nested relationships - **Describe the bug** If you create a group and one of the fields is a relationship with a model which have inside more relationships, when you fetch this data, it shows the id of the nested relationship as a object with the order of the character as key and the character as a value, in the same way as if you do {..."idString"} **Steps to reproduce the behavior** 1. Create a group with a relation to a model with more relationships 2. Add the group to a model **Expected behavior** The nested relationship should show the id of the related content as a normal string **Postman response** ``` [ { "_id": "5d8db9bf0dc976285007b04d", "items": [ { "_id": "5d8db9bf0dc976285007b04e", "notes": "qwe", "qty": 2, "createdAt": "2019-09-27T07:26:55.574Z", "updatedAt": "2019-09-27T07:26:55.601Z", "__v": 0, "product": { "_id": "5d6fbbe94e7c9554792b22b4", "name": "product11", "description": "tggyyfff77", "createdAt": "2019-09-04T13:28:09.311Z", "updatedAt": "2019-09-17T13:02:42.835Z", "__v": 0, "organization": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "2", "5": "f", "6": "e", "7": "2", "8": "8", "9": "e", "10": "6", "11": "7", "12": "5", "13": "7", "14": "3", "15": "8", "16": "7", "17": "a", "18": "c", "19": "7", "20": "6", "21": "0", "22": "3", "23": "7" }, "productgroup": { "0": "5", "1": "d", "2": "6", "3": "f", "4": "b", "5": "b", "6": "d", "7": "7", "8": "4", "9": "e", "10": "7", "11": "c", "12": "9", "13": "5", "14": "5", "15": "4", "16": "7", "17": "9", "18": "2", "19": "b", "20": "2", "21": "2", "22": "b", "23": "2" }, "factory": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "4", "5": "5", "6": "7", "7": "c", "8": "7", "9": "6", "10": "2", "11": "f", "12": "2", "13": "d", "14": "4", "15": "8", "16": "f", "17": "0", "18": "6", "19": "8", "20": "8", "21": "8", "22": "6", "23": "d" }, "client": { "0": "5", "1": "d", "2": "6", "3": "e", "4": "4", "5": "2", "6": "4", "7": "0", "8": "2", "9": "8", "10": "8", "11": "e", "12": "d", "13": "e", "14": "4", "15": "7", "16": "1", "17": "e", "18": "7", "19": "2", "20": "6", "21": "1", "22": "0", "23": "3" }, "sku": "12345", "id": "5d6fbbe94e7c9554792b22b4" }, "id": "5d8db9bf0dc976285007b04e" } ]``` **System** - Node.js version: v10.5 - NPM version: v6.7 - Strapi version: Beta 16.6 - Database: MongoDB - Operating system: Windows 10 **Possible solution** Now the id recovers its normal form with Object.values(strangeObjectID).reduce((a,b) => a+b, ""), maybe it lacks this in some point of the code. Or maybe there's a playful spread operator messing the string ID
priority
id is shown as an object when fetching models with nested relationships describe the bug if you create a group and one of the fields is a relationship with a model which have inside more relationships when you fetch this data it shows the id of the nested relationship as a object with the order of the character as key and the character as a value in the same way as if you do idstring steps to reproduce the behavior create a group with a relation to a model with more relationships add the group to a model expected behavior the nested relationship should show the id of the related content as a normal string postman response id items id notes qwe qty createdat updatedat v product id name description createdat updatedat v organization d e f e e a c productgroup d f b b d e c b b factory d e c f d f d client d e e d e e sku id id system node js version npm version strapi version beta database mongodb operating system windows possible solution now the id recovers its normal form with object values strangeobjectid reduce a b a b maybe it lacks this in some point of the code or maybe there s a playful spread operator messing the string id
1
796,272
28,104,684,476
IssuesEvent
2023-03-30 22:53:42
medic/cht-core
https://api.github.com/repos/medic/cht-core
closed
Haproxy stops routing during stress test
Type: Bug Priority: 1 - High
**Describe the bug** During e2e tests, I have frequently noticed the haproxy container reaching absurd CPU numbers on my machine (I have an AMD Ryzen 9 5900X). Very frequently, haproxy would take up to 100% of one CPU core. ![image](https://user-images.githubusercontent.com/35681649/225641452-00f0f0b5-56cc-4d35-a751-ee257a645b57.png) ![image](https://user-images.githubusercontent.com/35681649/225641506-7118d8ae-9609-4fe1-9e72-de9f9577ba6f.png) Then, during a stress test on an AWS hosted distributed setup, everytime I loaded the instance with more than 100 users, the suite failed because haproxy had stopped responding. The container was not killed, or restarted, but the instance was not reachable and API reported not being able to connect to haproxy. Upon inspecting haproxy logs, at first I noticed: ``` [NOTICE] (1) : haproxy version is 2.6.9-3a3700a [NOTICE] (1) : path to executable is /usr/local/sbin/haproxy [ALERT] (1) : Current worker (25) exited with code 137 (Killed) [ALERT] (1) : exit-on-failure: killing every processes with SIGTERM ``` On a subsequent retry, I saw: ``` Thread 2 is about to kill the process. Thread 1 : id=0x7fb8dce74f40 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=21 rqsz=760 1/1 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20789220579 now=21014641652 diff=225421073 curr_task=0x55705bb4b020 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x55709982b210 strm=0x55709982b210,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb7fbea2830,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x55705c4590f0,EST,20 scb=0x55705bb660e0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb85b606e50,80000300:H1(0x55705c46e3b0)/RAW((nil))/tcpv4(5635) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: *>Thread 2 : id=0x7fb8dce69700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=493 1/2 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=19770093994 now=86782386464 diff=67012292470 curr_task=0x7fb7d77d19d0 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb7d77d15e0 strm=0x7fb7d77d15e0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb7d77d1c30,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb813f12be0,EST,20 scb=0x7fb7d77d1af0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb893e46040,80000300:H1(0x7fb8da8cd700)/RAW((nil))/tcpv4(11218) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: call trace(22): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x7fb8dd600fe1 [48 8b 84 24 08 01 00 00]: libpthread:raise+0x141/0x179 | 0x55705a0ae3f7 [64 48 8b 53 10 64 48 8b]: main+0x165437 | 0x55705a0ae448 [0f 1f 84 00 00 00 00 00]: main+0x165488 | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f82403 [48 85 d0 75 20 48 89 f0]: lua_take_global_lock+0x23/0x4c | 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 3 : id=0x7fb8d3924700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=498 1/3 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20556386735 now=89985480430 diff=69429093695 curr_task=0x7fb833fd1390 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb833fd0fa0 strm=0x7fb833fd0fa0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb833fd15f0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb833fd0f30,EST,20 scb=0x7fb833fd14b0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb83741b110,80000300:H1(0x7fb85544fe40)/RAW((nil))/tcpv4(11845) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 Thread 4 : id=0x7fb8d3123700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=29 rqsz=778 1/4 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=21138813085 now=95953444002 diff=74814630917 curr_task=0x7fb86c80ae60 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb86c80aa70 strm=0x7fb86c80aa70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb86c80b0c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb86c80aa00,EST,20 scb=0x7fb86c80af80,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb853f53e20,80000300:H1(0x7fb8b12b27f0)/RAW((nil))/tcpv4(6201) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- >Thread 5 : id=0x7fb8d2922700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=10 rqsz=793 1/5 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=21689679358 now=99475683274 diff=77786003916 curr_task=0x7fb6d7fe4400 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb6d7fe4010 strm=0x7fb6d7fe4010,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb6d7fe4660,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb6d7fe3fa0,EST,20 scb=0x7fb6d7fe4520,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8730bcd20,80000300:H1(0x7fb8730bcf90)/RAW((nil))/tcpv4(6219) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f82428 [f3 90 eb d4 0f 1f 40 00]: lua_take_global_lock+0x48/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 6 : id=0x7fb8d2121700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=9 rqsz=706 1/6 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20990474533 now=100959756450 diff=79969281917 curr_task=0x7fb73bea7160 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb73bea6d70 strm=0x7fb73bea6d70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb73bea73c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb73bea6d00,EST,20 scb=0x7fb73bea7280,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb7e3f0bf90,80000300:H1(0x7fb89bf955e0)/RAW((nil))/tcpv4(6349) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 7 : id=0x7fb8d1920700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=3 rqsz=503 1/7 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20144821466 now=96965659318 diff=76820837852 curr_task=0x7fb74125de80 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb74125da90 strm=0x7fb74125da90,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb74125dff0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb7692a0e10,EST,20 scb=0x7fb74125df50,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8cf6f1b10,80000300:H1(0x7fb7692aa2f0)/RAW((nil))/tcpv4(10173) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: call trace(18): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 Thread 8 : id=0x7fb8d111f700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=25 rqsz=504 1/8 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20361036268 now=84727927263 diff=64366890995 curr_task=0x7fb8c0933120 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb8c0932d30 strm=0x7fb8c0932d30,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb8c3ce9a40,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb8c3cebee0,EST,20 scb=0x7fb8c3ce9930,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8abf12e60,80000300:H1(0x7fb8c3ce4af0)/RAW((nil))/tcpv4(8629) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: /usr/local/etc/haproxy/parse_basic.lua:12: in function line 9 [C]: in method 'gsub' /usr/local/etc/haproxy/parse_basic.lua:9: in global 'dec' /usr/local/etc/haproxy/parse_basic.lua:27: in function line 22 [NOTICE] (1) : haproxy version is 2.6.9-3a3700a [NOTICE] (1) : path to executable is /usr/local/sbin/haproxy [ALERT] (1) : Current worker (25) exited with code 134 (Aborted) [ALERT] (1) : exit-on-failure: killing every processes with SIGTERM ``` Right before this happened, this is a snapshot of `docker stats` on the AWS instance: ![image](https://user-images.githubusercontent.com/35681649/225642684-970c75d3-887f-43b4-a1b1-3454670e25ae.png) Stats: haproxy is using 282% cpu and 4.349G of memory. **To Reproduce** Steps to reproduce the behavior: 1. Run scalability suite with 100 concurrent users. **Expected behavior** Haproxy shouldn't take up as many resources. Also, the container should somehow restart automatically, instead of just hang when this issue happens. **Environment** - Instance: local & aws test instance - App: haproxt - Version: master
1.0
Haproxy stops routing during stress test - **Describe the bug** During e2e tests, I have frequently noticed the haproxy container reaching absurd CPU numbers on my machine (I have an AMD Ryzen 9 5900X). Very frequently, haproxy would take up to 100% of one CPU core. ![image](https://user-images.githubusercontent.com/35681649/225641452-00f0f0b5-56cc-4d35-a751-ee257a645b57.png) ![image](https://user-images.githubusercontent.com/35681649/225641506-7118d8ae-9609-4fe1-9e72-de9f9577ba6f.png) Then, during a stress test on an AWS hosted distributed setup, everytime I loaded the instance with more than 100 users, the suite failed because haproxy had stopped responding. The container was not killed, or restarted, but the instance was not reachable and API reported not being able to connect to haproxy. Upon inspecting haproxy logs, at first I noticed: ``` [NOTICE] (1) : haproxy version is 2.6.9-3a3700a [NOTICE] (1) : path to executable is /usr/local/sbin/haproxy [ALERT] (1) : Current worker (25) exited with code 137 (Killed) [ALERT] (1) : exit-on-failure: killing every processes with SIGTERM ``` On a subsequent retry, I saw: ``` Thread 2 is about to kill the process. Thread 1 : id=0x7fb8dce74f40 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=21 rqsz=760 1/1 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20789220579 now=21014641652 diff=225421073 curr_task=0x55705bb4b020 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x55709982b210 strm=0x55709982b210,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb7fbea2830,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x55705c4590f0,EST,20 scb=0x55705bb660e0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb85b606e50,80000300:H1(0x55705c46e3b0)/RAW((nil))/tcpv4(5635) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: *>Thread 2 : id=0x7fb8dce69700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=493 1/2 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=19770093994 now=86782386464 diff=67012292470 curr_task=0x7fb7d77d19d0 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb7d77d15e0 strm=0x7fb7d77d15e0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb7d77d1c30,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb813f12be0,EST,20 scb=0x7fb7d77d1af0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb893e46040,80000300:H1(0x7fb8da8cd700)/RAW((nil))/tcpv4(11218) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: call trace(22): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x7fb8dd600fe1 [48 8b 84 24 08 01 00 00]: libpthread:raise+0x141/0x179 | 0x55705a0ae3f7 [64 48 8b 53 10 64 48 8b]: main+0x165437 | 0x55705a0ae448 [0f 1f 84 00 00 00 00 00]: main+0x165488 | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f82403 [48 85 d0 75 20 48 89 f0]: lua_take_global_lock+0x23/0x4c | 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 3 : id=0x7fb8d3924700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=2 rqsz=498 1/3 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20556386735 now=89985480430 diff=69429093695 curr_task=0x7fb833fd1390 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb833fd0fa0 strm=0x7fb833fd0fa0,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb833fd15f0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb833fd0f30,EST,20 scb=0x7fb833fd14b0,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb83741b110,80000300:H1(0x7fb85544fe40)/RAW((nil))/tcpv4(11845) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 Thread 4 : id=0x7fb8d3123700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=29 rqsz=778 1/4 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=21138813085 now=95953444002 diff=74814630917 curr_task=0x7fb86c80ae60 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb86c80aa70 strm=0x7fb86c80aa70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb86c80b0c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb86c80aa00,EST,20 scb=0x7fb86c80af80,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb853f53e20,80000300:H1(0x7fb8b12b27f0)/RAW((nil))/tcpv4(6201) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- >Thread 5 : id=0x7fb8d2922700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=10 rqsz=793 1/5 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=21689679358 now=99475683274 diff=77786003916 curr_task=0x7fb6d7fe4400 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb6d7fe4010 strm=0x7fb6d7fe4010,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb6d7fe4660,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb6d7fe3fa0,EST,20 scb=0x7fb6d7fe4520,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8730bcd20,80000300:H1(0x7fb8730bcf90)/RAW((nil))/tcpv4(6219) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f82428 [f3 90 eb d4 0f 1f 40 00]: lua_take_global_lock+0x48/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 6 : id=0x7fb8d2121700 act=1 glob=1 wq=1 rq=1 tl=1 tlsz=9 rqsz=706 1/6 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20990474533 now=100959756450 diff=79969281917 curr_task=0x7fb73bea7160 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb73bea6d70 strm=0x7fb73bea6d70,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb73bea73c0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb73bea6d00,EST,20 scb=0x7fb73bea7280,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb7e3f0bf90,80000300:H1(0x7fb89bf955e0)/RAW((nil))/tcpv4(6349) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- call trace(19): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f85d97 [e9 f2 fe ff ff 0f 1f 40]: hlua_ctx_init+0x167/0x1e4 | 0x557059f8840c [85 c0 0f 84 bc 00 00 00]: main+0x3f44c | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 >Thread 7 : id=0x7fb8d1920700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=3 rqsz=503 1/7 stuck=1 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20144821466 now=96965659318 diff=76820837852 curr_task=0x7fb74125de80 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb74125da90 strm=0x7fb74125da90,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb74125dff0,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb7692a0e10,EST,20 scb=0x7fb74125df50,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8cf6f1b10,80000300:H1(0x7fb7692aa2f0)/RAW((nil))/tcpv4(10173) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: call trace(18): | 0x55705a0aefcb [89 44 24 04 85 c0 75 15]: ha_dump_backtrace+0x2b/0x2d0 | 0x55705a0afa66 [48 8b 05 e3 aa 1d 00 48]: debug_handler+0x66/0x10b | 0x7fb8dd601140 [48 c7 c0 0f 00 00 00 0f]: libpthread:+0x13140 | 0x557059f8242a [eb d4 0f 1f 40 00 41 54]: lua_take_global_lock+0x4a/0x4c | 0x557059f8842f [e9 f3 f9 ff ff 0f 1f 40]: main+0x3f46f | 0x55705a01e534 [85 c0 74 69 48 8b 5d 20]: sample_process+0x54/0x107 | 0x55705a01f342 [48 85 c0 74 59 64 48 63]: sample_fetch_as_type+0x42/0xc1 | 0x55705a03b50b [48 89 c6 48 85 c0 0f 84]: sess_build_logline+0x248b/0x40f3 | 0x55705a098ff4 [4c 8b a3 88 00 00 00 48]: main+0x150034 | 0x557059ff92f9 [4c 8b 0c 24 48 8b 4c 24]: main+0xb0339 | 0x557059ffb9d6 [4c 8b 44 24 18 83 f8 07]: http_process_req_common+0xa6/0x1035 | 0x557059fe58b8 [85 c0 0f 85 fc f7 ff ff]: process_stream+0x1f68/0x35e9 Thread 8 : id=0x7fb8d111f700 act=1 glob=0 wq=1 rq=1 tl=1 tlsz=25 rqsz=504 1/8 stuck=0 prof=0 harmless=0 wantrdv=0 cpu_ns: poll=20361036268 now=84727927263 diff=64366890995 curr_task=0x7fb8c0933120 (task) calls=1 last=0 fct=0x557059fe3950(process_stream) ctx=0x7fb8c0932d30 strm=0x7fb8c0932d30,c00 src=10.0.1.4 fe=http-in be=http-in dst=unknown txn=0x7fb8c3ce9a40,3000 txn.req=MSG_BODY,4c txn.rsp=MSG_RPBEFORE,0 rqf=40d08002 rqa=30 rpf=80000000 rpa=0 scf=0x7fb8c3cebee0,EST,20 scb=0x7fb8c3ce9930,INI,21 af=(nil),0 sab=(nil),0 cof=0x7fb8abf12e60,80000300:H1(0x7fb8c3ce4af0)/RAW((nil))/tcpv4(8629) cob=(nil),0:NONE((nil))/NONE((nil))/NONE(-1) Current executing Lua from a stream analyser -- stack traceback: /usr/local/etc/haproxy/parse_basic.lua:12: in function line 9 [C]: in method 'gsub' /usr/local/etc/haproxy/parse_basic.lua:9: in global 'dec' /usr/local/etc/haproxy/parse_basic.lua:27: in function line 22 [NOTICE] (1) : haproxy version is 2.6.9-3a3700a [NOTICE] (1) : path to executable is /usr/local/sbin/haproxy [ALERT] (1) : Current worker (25) exited with code 134 (Aborted) [ALERT] (1) : exit-on-failure: killing every processes with SIGTERM ``` Right before this happened, this is a snapshot of `docker stats` on the AWS instance: ![image](https://user-images.githubusercontent.com/35681649/225642684-970c75d3-887f-43b4-a1b1-3454670e25ae.png) Stats: haproxy is using 282% cpu and 4.349G of memory. **To Reproduce** Steps to reproduce the behavior: 1. Run scalability suite with 100 concurrent users. **Expected behavior** Haproxy shouldn't take up as many resources. Also, the container should somehow restart automatically, instead of just hang when this issue happens. **Environment** - Instance: local & aws test instance - App: haproxt - Version: master
priority
haproxy stops routing during stress test describe the bug during tests i have frequently noticed the haproxy container reaching absurd cpu numbers on my machine i have an amd ryzen very frequently haproxy would take up to of one cpu core then during a stress test on an aws hosted distributed setup everytime i loaded the instance with more than users the suite failed because haproxy had stopped responding the container was not killed or restarted but the instance was not reachable and api reported not being able to connect to haproxy upon inspecting haproxy logs at first i noticed haproxy version is path to executable is usr local sbin haproxy current worker exited with code killed exit on failure killing every processes with sigterm on a subsequent retry i saw thread is about to kill the process thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback call trace ha dump backtrace debug handler libpthread libpthread raise main main libpthread lua take global lock main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser call trace ha dump backtrace debug handler libpthread lua take global lock hlua ctx init main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback call trace ha dump backtrace debug handler libpthread lua take global lock main sample process sample fetch as type sess build logline main main http process req common process stream thread id act glob wq rq tl tlsz rqsz stuck prof harmless wantrdv cpu ns poll now diff curr task task calls last fct process stream ctx strm src fe http in be http in dst unknown txn txn req msg body txn rsp msg rpbefore rqf rqa rpf rpa scf est scb ini af nil sab nil cof raw nil cob nil none nil none nil none current executing lua from a stream analyser stack traceback usr local etc haproxy parse basic lua in function line in method gsub usr local etc haproxy parse basic lua in global dec usr local etc haproxy parse basic lua in function line haproxy version is path to executable is usr local sbin haproxy current worker exited with code aborted exit on failure killing every processes with sigterm right before this happened this is a snapshot of docker stats on the aws instance stats haproxy is using cpu and of memory to reproduce steps to reproduce the behavior run scalability suite with concurrent users expected behavior haproxy shouldn t take up as many resources also the container should somehow restart automatically instead of just hang when this issue happens environment instance local aws test instance app haproxt version master
1
383,854
11,363,266,132
IssuesEvent
2020-01-27 03:00:21
wso2/docs-is
https://api.github.com/repos/wso2/docs-is
opened
Deprecate JWT Token Generation using SOAP service
Priority/Highest Severity/Critical
From IS 5.3.0 onward we have the user-info endpoint to retrieve user claims of a user in either plain-text or JWT format. Therefore JWT token generation using TokenValidationService [1] should be deprecated from all the IS docs from 5.3.0 version onwards. [1] https://docs.wso2.com/display/IS530/JWT+Token+Generation
1.0
Deprecate JWT Token Generation using SOAP service - From IS 5.3.0 onward we have the user-info endpoint to retrieve user claims of a user in either plain-text or JWT format. Therefore JWT token generation using TokenValidationService [1] should be deprecated from all the IS docs from 5.3.0 version onwards. [1] https://docs.wso2.com/display/IS530/JWT+Token+Generation
priority
deprecate jwt token generation using soap service from is onward we have the user info endpoint to retrieve user claims of a user in either plain text or jwt format therefore jwt token generation using tokenvalidationservice should be deprecated from all the is docs from version onwards
1
114,346
4,629,241,142
IssuesEvent
2016-09-28 08:36:39
VirtoCommerce/vc-platform
https://api.github.com/repos/VirtoCommerce/vc-platform
closed
"Object reference not set to an instance of an object." in CategoryBrowsingService.SearchCategories
bug high priority
can't start storefront on my localhost: ![image](https://cloud.githubusercontent.com/assets/5801549/18903569/9803eb8a-8562-11e6-8bbb-e4453a0120b9.png)
1.0
"Object reference not set to an instance of an object." in CategoryBrowsingService.SearchCategories - can't start storefront on my localhost: ![image](https://cloud.githubusercontent.com/assets/5801549/18903569/9803eb8a-8562-11e6-8bbb-e4453a0120b9.png)
priority
object reference not set to an instance of an object in categorybrowsingservice searchcategories can t start storefront on my localhost
1
360,423
10,688,493,858
IssuesEvent
2019-10-22 18:21:09
carbon-design-system/design-language-website
https://api.github.com/repos/carbon-design-system/design-language-website
opened
Update color palette specifications component to use v2 colors
priority: high
https://www.ibm.com/design/language/elements/color/#specifications The `SwatchPalette` component on the Color page needs to get updated to show the v2 colors. Ideally this would pull directly from the colors package instead of using the hardcoded values.
1.0
Update color palette specifications component to use v2 colors - https://www.ibm.com/design/language/elements/color/#specifications The `SwatchPalette` component on the Color page needs to get updated to show the v2 colors. Ideally this would pull directly from the colors package instead of using the hardcoded values.
priority
update color palette specifications component to use colors the swatchpalette component on the color page needs to get updated to show the colors ideally this would pull directly from the colors package instead of using the hardcoded values
1
519,196
15,047,235,492
IssuesEvent
2021-02-03 08:38:32
ossia/libossia
https://api.github.com/repos/ossia/libossia
closed
[ossia-max] ossia.remote outputs value twice (or more)
bug ossia-max priority::high
when value is changed from the bound parameter, ossia.remote outputs value twice (by both 1rst and 2nd outlets) If value is changed by the inlet of the remote itself (or from another remote) : value is output 4 times. if parameter / remote "repetitions" attribute is set to 0 : same issue (this attribute seems to only filter repetition on the outlet of ossia.parameter) <pre><code> ----------begin_max5_patcher---------- 945.3oc0YssaiBCD84To9Of3Y1Hrwbak1ef8WX0pHGh2DWAFDXZS2p8ee8EH MsEBNEm1lGRAaLLy4XOybr6S2dyB20k6IMtNe24WNKV7jnmEp9j8rnuiEtE3 8Y43F0.cyJKJHLtqW2C4j8b0CpIEkbhSYKupk23bONuk3fb3zBgI5GMqsPLf bBW8w766ltQ8IJWe22f9GFbElmsix1tpljw0dIHBsLzyAjltz2S0RbA5uz24 2GYBJq2B.Um+61ajWEW7LFmqa47R1Q9RMtfvI0qHL75bxKb+WhJv.nB7Lpzi j+XEQCI20X1V2m8+APMRhvCnFh5+q0AMi7fvae6baVYKSf8QlFQCAX33.lJW 9331+S1zD3CC.xKwApFvQge36G9sEqOBkmwbNbHJ.MNEHwswy6vvD4kP+I.t 8m2KaZn3kcQ1EOt5.iLxJgfgngvSSCpeFQCIcA8CvCYk4k05Q6uD.CQodhah Si7iSj2kjh..wcfK.4cdYD4OPyHmQ9vjSlND.koCgQWWYCCmSxP.L4ZKYXj8 xE1g9qsbgAyMUXOtu3oBGMZ9GN0jJBleNZYhOUvKLNVhBjuZgbT3URvKDLmn WHJQlwpCzeBQuMjbGe2ynJd5DvcxxWcPNHVAYzTgrvuJxWRlcLapF3fzOI4K +M2IamvMImy7c7DxULBxZvFkdsLWO67yhJQBbmbomos7txlUpLMj0ZutJjgj ZOYHZrqKFeIEgLB1qporypPr6zRK5RSgfWk6xBM2MY0Kv5K5lrNM2cfwLg9F LAXzbx5qiF.PfdiInInNTpOHMxScWhefhECAwQn3KC2g475V5A.Ja1s+zJBm xokrwNftA0+chS1YhkXI.sFd8RrKXYBKi3f2MhEUGDqLRAuO.qGlaNk8lSoU YL4CdEMzT1Vm06b8NuyQFbCogSYXIGbznjq+kiZbF2bCBMwfHqYO.xH.FZOC 5ajAg1yfFwn.6QoQlXu.qYtPSLm8VfF7wt9L0nIO6wlFEN.sp4.SBuD6ELXD eJOjBaYvDil.ismAiMhRSsmA+nKQ..lXvgAXWMQbU08j5lt2PaJgFf6zprR7 zsoLcasjOQ4+6o8uhdaBt3ZQMatnfcasxCc2G0Ik1snbColIjQzUYVBVooa3 Ol+l5wLgjS06+SbQYAd0N5VAUrcGmrYUY8QmAg.r+A2lyeEGgyxHL9QxDURF 7kpGz+u+VdDU30y20h2Q4NJS2agBwmW409ujyT5ijNZSEVOenzQc6Mhg7eP9 flpT -----------end_max5_patcher----------- </code></pre>
1.0
[ossia-max] ossia.remote outputs value twice (or more) - when value is changed from the bound parameter, ossia.remote outputs value twice (by both 1rst and 2nd outlets) If value is changed by the inlet of the remote itself (or from another remote) : value is output 4 times. if parameter / remote "repetitions" attribute is set to 0 : same issue (this attribute seems to only filter repetition on the outlet of ossia.parameter) <pre><code> ----------begin_max5_patcher---------- 945.3oc0YssaiBCD84To9Of3Y1Hrwbak1ef8WX0pHGh2DWAFDXZS2p8ee8EH MsEBNEm1lGRAaLLy4XOybr6S2dyB20k6IMtNe24WNKV7jnmEp9j8rnuiEtE3 8Y43F0.cyJKJHLtqW2C4j8b0CpIEkbhSYKupk23bONuk3fb3zBgI5GMqsPLf bBW8w766ltQ8IJWe22f9GFbElmsix1tpljw0dIHBsLzyAjltz2S0RbA5uz24 2GYBJq2B.Um+61ajWEW7LFmqa47R1Q9RMtfvI0qHL75bxKb+WhJv.nB7Lpzi j+XEQCI20X1V2m8+APMRhvCnFh5+q0AMi7fvae6baVYKSf8QlFQCAX33.lJW 9331+S1zD3CC.xKwApFvQge36G9sEqOBkmwbNbHJ.MNEHwswy6vvD4kP+I.t 8m2KaZn3kcQ1EOt5.iLxJgfgngvSSCpeFQCIcA8CvCYk4k05Q6uD.CQodhah Si7iSj2kjh..wcfK.4cdYD4OPyHmQ9vjSlND.koCgQWWYCCmSxP.L4ZKYXj8 xE1g9qsbgAyMUXOtu3oBGMZ9GN0jJBleNZYhOUvKLNVhBjuZgbT3URvKDLmn WHJQlwpCzeBQuMjbGe2ynJd5DvcxxWcPNHVAYzTgrvuJxWRlcLapF3fzOI4K +M2IamvMImy7c7DxULBxZvFkdsLWO67yhJQBbmbomos7txlUpLMj0ZutJjgj ZOYHZrqKFeIEgLB1qporypPr6zRK5RSgfWk6xBM2MY0Kv5K5lrNM2cfwLg9F LAXzbx5qiF.PfdiInInNTpOHMxScWhefhECAwQn3KC2g475V5A.Ja1s+zJBm xokrwNftA0+chS1YhkXI.sFd8RrKXYBKi3f2MhEUGDqLRAuO.qGlaNk8lSoU YL4CdEMzT1Vm06b8NuyQFbCogSYXIGbznjq+kiZbF2bCBMwfHqYO.xH.FZOC 5ajAg1yfFwn.6QoQlXu.qYtPSLm8VfF7wt9L0nIO6wlFEN.sp4.SBuD6ELXD eJOjBaYvDil.ismAiMhRSsmA+nKQ..lXvgAXWMQbU08j5lt2PaJgFf6zprR7 zsoLcasjOQ4+6o8uhdaBt3ZQMatnfcasxCc2G0Ik1snbColIjQzUYVBVooa3 Ol+l5wLgjS06+SbQYAd0N5VAUrcGmrYUY8QmAg.r+A2lyeEGgyxHL9QxDURF 7kpGz+u+VdDU30y20h2Q4NJS2agBwmW409ujyT5ijNZSEVOenzQc6Mhg7eP9 flpT -----------end_max5_patcher----------- </code></pre>
priority
ossia remote outputs value twice or more when value is changed from the bound parameter ossia remote outputs value twice by both and outlets if value is changed by the inlet of the remote itself or from another remote value is output times if parameter remote repetitions attribute is set to same issue this attribute seems to only filter repetition on the outlet of ossia parameter begin patcher um whjv j ljw i t cqodhah wcfk kocgqwwyccmsxp zjbm xh fzoc ejojbayvdil ismaimhrssma nkq ol r u flpt end patcher
1
254,367
8,073,245,661
IssuesEvent
2018-08-06 18:36:32
MusaMahmood/Android-HighSamplingRate-ECG
https://api.github.com/repos/MusaMahmood/Android-HighSamplingRate-ECG
opened
Combine all similar apps into single repository!
<HIGH PRIORITY>
Includes: - https://github.com/MusaMahmood/Android-HighSamplingRate-ECG - https://github.com/MusaMahmood/portable_ECG_Analysis (subdirectory) - https://github.com/MusaMahmood/Android-SSVEP-Interface - https://github.com/MusaMahmood/Android-EMG-DroneDemo - https://github.com/MusaMahmood/Android-BLE-Motion-Control-RC-Car - https://github.com/MusaMahmood/Android-BLE-Library-Kotlin-Sample - https://github.com/MusaMahmood/Android-BLE-Library-Kotlin (??)
1.0
Combine all similar apps into single repository! - Includes: - https://github.com/MusaMahmood/Android-HighSamplingRate-ECG - https://github.com/MusaMahmood/portable_ECG_Analysis (subdirectory) - https://github.com/MusaMahmood/Android-SSVEP-Interface - https://github.com/MusaMahmood/Android-EMG-DroneDemo - https://github.com/MusaMahmood/Android-BLE-Motion-Control-RC-Car - https://github.com/MusaMahmood/Android-BLE-Library-Kotlin-Sample - https://github.com/MusaMahmood/Android-BLE-Library-Kotlin (??)
priority
combine all similar apps into single repository includes subdirectory
1
52,622
3,025,324,482
IssuesEvent
2015-08-03 07:43:12
OpenSRP/opensrp-client
https://api.github.com/repos/OpenSRP/opensrp-client
closed
faulty logic on the opensrp ELCO forms
bug High Priority
@raihan-mpower, it seems the new .apk has faulty form logic around the questions asked to the MWRA since all my women have the tag: NOT In ELCO, see the screenshot below: ![screenshot_2015-07-27-05-43-30](https://cloud.githubusercontent.com/assets/801106/8903257/4c8731da-345f-11e5-865f-430ae70bd08c.png) This has happened with the latest .apk, never observed it before. Also, none of the submission get couchDB and by extension OpenMRS, please look into this? @julkarnain @sohelsarder @dimasciput @numeranazneen
1.0
faulty logic on the opensrp ELCO forms - @raihan-mpower, it seems the new .apk has faulty form logic around the questions asked to the MWRA since all my women have the tag: NOT In ELCO, see the screenshot below: ![screenshot_2015-07-27-05-43-30](https://cloud.githubusercontent.com/assets/801106/8903257/4c8731da-345f-11e5-865f-430ae70bd08c.png) This has happened with the latest .apk, never observed it before. Also, none of the submission get couchDB and by extension OpenMRS, please look into this? @julkarnain @sohelsarder @dimasciput @numeranazneen
priority
faulty logic on the opensrp elco forms raihan mpower it seems the new apk has faulty form logic around the questions asked to the mwra since all my women have the tag not in elco see the screenshot below this has happened with the latest apk never observed it before also none of the submission get couchdb and by extension openmrs please look into this julkarnain sohelsarder dimasciput numeranazneen
1
793,533
28,001,264,310
IssuesEvent
2023-03-27 11:57:00
ballerina-platform/ballerina-standard-library
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
closed
Special characters in resource path caused issues in action invocation
Priority/High Type/Bug module/http Team/PCM Reason/Invalid
**Description:** $subject refer to the sample code below. **Steps to reproduce:** ```ballerina import ballerina/http; import ballerina/io; public function main() returns error? { http:Client openWeather = check new("https://api.openweathermap.org"); json|error res = openWeather->get("/data/2.5/forecast?lat=30.266666&lon=-97.733330&appid=<appid>"); io:println(res); //This works io:println("1---------------"); json|error res2 = openWeather->/data/["2.5"]/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>"); io:println(res2);//This works io:println("2---------------");//Following does not work json|error res3 = openWeather->/data/\2\.5/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>"); io:println(res3); } ``` Note : I used a free version of https://openweathermap.org/api to generate tokens. Request hangs and it gives the following error. ```sh error PayloadBindingClientError ("http:ApplicationResponseError creation failed: 404 response payload extraction failed",error GenericClientError ("Error occurred while retrieving the json payload from the response",error ParserError ("Error occurred while extracting content from message : Idle timeout triggered while reading inbound response headers"))) ``` **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Special characters in resource path caused issues in action invocation - **Description:** $subject refer to the sample code below. **Steps to reproduce:** ```ballerina import ballerina/http; import ballerina/io; public function main() returns error? { http:Client openWeather = check new("https://api.openweathermap.org"); json|error res = openWeather->get("/data/2.5/forecast?lat=30.266666&lon=-97.733330&appid=<appid>"); io:println(res); //This works io:println("1---------------"); json|error res2 = openWeather->/data/["2.5"]/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>"); io:println(res2);//This works io:println("2---------------");//Following does not work json|error res3 = openWeather->/data/\2\.5/forecast.get(lat=30.266666, lon=-97.733330, appid="<appid>"); io:println(res3); } ``` Note : I used a free version of https://openweathermap.org/api to generate tokens. Request hangs and it gives the following error. ```sh error PayloadBindingClientError ("http:ApplicationResponseError creation failed: 404 response payload extraction failed",error GenericClientError ("Error occurred while retrieving the json payload from the response",error ParserError ("Error occurred while extracting content from message : Idle timeout triggered while reading inbound response headers"))) ``` **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
priority
special characters in resource path caused issues in action invocation description subject refer to the sample code below steps to reproduce ballerina import ballerina http import ballerina io public function main returns error http client openweather check new json error res openweather get data forecast lat lon appid io println res this works io println json error openweather data forecast get lat lon appid io println this works io println following does not work json error openweather data forecast get lat lon appid io println note i used a free version of to generate tokens request hangs and it gives the following error sh error payloadbindingclienterror http applicationresponseerror creation failed response payload extraction failed error genericclienterror error occurred while retrieving the json payload from the response error parsererror error occurred while extracting content from message idle timeout triggered while reading inbound response headers affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
1
438,554
12,641,128,364
IssuesEvent
2020-06-16 05:17:17
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
upgrade-zulip-from-git: Shut down server before upgrade if RAM minimal
area: production installer help wanted priority: high
Currently, if you run `upgrade-zulip-from-git` on a system with Zulip's minimum of 2GB of RAM, you're at risk of having it fail because 2GB of RAM is insufficient to both run the Zulip server and build static assets with webpack; the result is almost always an OOM kill. This accounts for a significant portion of all reported problems upgrading Zulip servers, so is pretty important. We can fix this by having `upgrade-zulip-from-git` check the total RAM on the system, and if it is low (say, <3GB?), shut down the Zulip server early, before building static assets (maybe printing a warning to make clear what's happening). I think the code for this belongs just before running `update-prod-static` in the `--from-git` part of `upgrade-zulip-stage-2`; it seems the failures happen just in that section, not the Python part. This should be pretty easy to test -- just create a VM with 2GB of RAM and run `upgrade-zulip-from-git` until it works. See for example #14361.
1.0
upgrade-zulip-from-git: Shut down server before upgrade if RAM minimal - Currently, if you run `upgrade-zulip-from-git` on a system with Zulip's minimum of 2GB of RAM, you're at risk of having it fail because 2GB of RAM is insufficient to both run the Zulip server and build static assets with webpack; the result is almost always an OOM kill. This accounts for a significant portion of all reported problems upgrading Zulip servers, so is pretty important. We can fix this by having `upgrade-zulip-from-git` check the total RAM on the system, and if it is low (say, <3GB?), shut down the Zulip server early, before building static assets (maybe printing a warning to make clear what's happening). I think the code for this belongs just before running `update-prod-static` in the `--from-git` part of `upgrade-zulip-stage-2`; it seems the failures happen just in that section, not the Python part. This should be pretty easy to test -- just create a VM with 2GB of RAM and run `upgrade-zulip-from-git` until it works. See for example #14361.
priority
upgrade zulip from git shut down server before upgrade if ram minimal currently if you run upgrade zulip from git on a system with zulip s minimum of of ram you re at risk of having it fail because of ram is insufficient to both run the zulip server and build static assets with webpack the result is almost always an oom kill this accounts for a significant portion of all reported problems upgrading zulip servers so is pretty important we can fix this by having upgrade zulip from git check the total ram on the system and if it is low say shut down the zulip server early before building static assets maybe printing a warning to make clear what s happening i think the code for this belongs just before running update prod static in the from git part of upgrade zulip stage it seems the failures happen just in that section not the python part this should be pretty easy to test just create a vm with of ram and run upgrade zulip from git until it works see for example
1
708,049
24,328,473,456
IssuesEvent
2022-09-30 16:57:26
hdmf-dev/hdmf-zarr
https://api.github.com/repos/hdmf-dev/hdmf-zarr
closed
Create release on Conda
priority: high
The Conda release is not a must but would be nice to have and will be good to do as learning experience
1.0
Create release on Conda - The Conda release is not a must but would be nice to have and will be good to do as learning experience
priority
create release on conda the conda release is not a must but would be nice to have and will be good to do as learning experience
1
563,392
16,682,468,812
IssuesEvent
2021-06-08 02:44:39
nlpsandbox/nlpsandbox-client
https://api.github.com/repos/nlpsandbox/nlpsandbox-client
closed
Add scoring code (evaluation) support for id and contact annotators
Priority: High
### Is your proposal related to a problem? Contact, Id, covid symptom annotators needs to be added to the evaluation code - [x] id - [x] contact
1.0
Add scoring code (evaluation) support for id and contact annotators - ### Is your proposal related to a problem? Contact, Id, covid symptom annotators needs to be added to the evaluation code - [x] id - [x] contact
priority
add scoring code evaluation support for id and contact annotators is your proposal related to a problem contact id covid symptom annotators needs to be added to the evaluation code id contact
1
33,362
2,764,296,158
IssuesEvent
2015-04-29 14:48:39
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
TaskArchiver subscription finishing logic "deadlocking" with multiple requests with the same input
High Priority WMAgent
In an agent we have 2 request reading on the same input dataset on the first step (it's TaskChain), now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running. This is because the second part of the query which checks for parentage in l. 57 of GetAndMarkFinishedSubscriptions, this was added to prevent deletion of parents when shared between requests due to race conditions. It was just in the code for the PromptSkimming case with Tier1PromptReco, for now in the affected agent I am removing that portion of the query while we come up with something better.
1.0
TaskArchiver subscription finishing logic "deadlocking" with multiple requests with the same input - In an agent we have 2 request reading on the same input dataset on the first step (it's TaskChain), now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running. This is because the second part of the query which checks for parentage in l. 57 of GetAndMarkFinishedSubscriptions, this was added to prevent deletion of parents when shared between requests due to race conditions. It was just in the code for the PromptSkimming case with Tier1PromptReco, for now in the affected agent I am removing that portion of the query while we come up with something better.
priority
taskarchiver subscription finishing logic deadlocking with multiple requests with the same input in an agent we have request reading on the same input dataset on the first step it s taskchain now both have finished processing everything but some subscriptions are not marked as finished and therefore the request is stuck at running this is because the second part of the query which checks for parentage in l of getandmarkfinishedsubscriptions this was added to prevent deletion of parents when shared between requests due to race conditions it was just in the code for the promptskimming case with for now in the affected agent i am removing that portion of the query while we come up with something better
1
73,951
3,422,680,959
IssuesEvent
2015-12-09 00:23:23
washingtontrails/vms
https://api.github.com/repos/washingtontrails/vms
opened
VMS: Addition to Location/Directions Tab
Enhancement High Priority Pyramid VMS BUDGET
![image](https://cloud.githubusercontent.com/assets/14843712/11672638/2e89eb92-9dc7-11e5-8782-6639cddd5f8c.png) Directly above the Google Map Image the following disclaimer should be added: "Warning: Always use the written directions (above) to get to a trailhead. The location of the pinpoint below is only approximate. Using mapping websites and gps navigation to get to trailheads could mislead you."
1.0
VMS: Addition to Location/Directions Tab - ![image](https://cloud.githubusercontent.com/assets/14843712/11672638/2e89eb92-9dc7-11e5-8782-6639cddd5f8c.png) Directly above the Google Map Image the following disclaimer should be added: "Warning: Always use the written directions (above) to get to a trailhead. The location of the pinpoint below is only approximate. Using mapping websites and gps navigation to get to trailheads could mislead you."
priority
vms addition to location directions tab directly above the google map image the following disclaimer should be added warning always use the written directions above to get to a trailhead the location of the pinpoint below is only approximate using mapping websites and gps navigation to get to trailheads could mislead you
1
825,063
31,241,464,088
IssuesEvent
2023-08-20 22:54:09
MirageAegis/Sushi-Bot
https://api.github.com/repos/MirageAegis/Sushi-Bot
closed
Auto shout out and go-live
feature priority: high
# Feature Automatically shout out streamers across all servers that has auto shout out enabled. This is a controlled way of promoting other streamers in your server! Additionally, automatically post go-live posts in servers where it's enabled. ## Acceptance Criteria - [x] Sushi Bot shall allow server administrators to configure auto shout outs. - [x] Sushi Bot shall shout out users who go live. - [x] This will only be done in servers that has the feature configured. - [x] The shout out message may be set by the server administrators. - [x] Sushi Bot shall allow server administrators to configure auto go-live posts. - [x] Sushi Bot shall post a go-live message whenever the server owner goes live. - [x] This will only be done in servers that has the feature configured. - [x] The shout out message may be set by the server administrators.
1.0
Auto shout out and go-live - # Feature Automatically shout out streamers across all servers that has auto shout out enabled. This is a controlled way of promoting other streamers in your server! Additionally, automatically post go-live posts in servers where it's enabled. ## Acceptance Criteria - [x] Sushi Bot shall allow server administrators to configure auto shout outs. - [x] Sushi Bot shall shout out users who go live. - [x] This will only be done in servers that has the feature configured. - [x] The shout out message may be set by the server administrators. - [x] Sushi Bot shall allow server administrators to configure auto go-live posts. - [x] Sushi Bot shall post a go-live message whenever the server owner goes live. - [x] This will only be done in servers that has the feature configured. - [x] The shout out message may be set by the server administrators.
priority
auto shout out and go live feature automatically shout out streamers across all servers that has auto shout out enabled this is a controlled way of promoting other streamers in your server additionally automatically post go live posts in servers where it s enabled acceptance criteria sushi bot shall allow server administrators to configure auto shout outs sushi bot shall shout out users who go live this will only be done in servers that has the feature configured the shout out message may be set by the server administrators sushi bot shall allow server administrators to configure auto go live posts sushi bot shall post a go live message whenever the server owner goes live this will only be done in servers that has the feature configured the shout out message may be set by the server administrators
1
510,482
14,791,317,621
IssuesEvent
2021-01-12 13:20:21
blchelle/collabogreat
https://api.github.com/repos/blchelle/collabogreat
closed
Loading animation doesn't stop on successful send invite
Priority: High Status: In Progress Type: Bug
## Description When a project member sends an invite to another project member, if the invitation sends successfully the loading animation on the button doesn't stop
1.0
Loading animation doesn't stop on successful send invite - ## Description When a project member sends an invite to another project member, if the invitation sends successfully the loading animation on the button doesn't stop
priority
loading animation doesn t stop on successful send invite description when a project member sends an invite to another project member if the invitation sends successfully the loading animation on the button doesn t stop
1
699,651
24,025,790,441
IssuesEvent
2022-09-15 11:23:51
CLOSER-Cohorts/archivist
https://api.github.com/repos/CLOSER-Cohorts/archivist
closed
REACT: questions in condition Else branch and loops are not showing in map view and can't be mapped via .txt files
bug High priority react
For example ncds_81_i question qc_currentjob_17_a_B is in the Else branch of c_currentjob_q17cA. https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/ However this question does not appear in the map view https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/map/ This is also true for questions in loops e.g. qc_familyincome_2_a_i which is inside l_familyincome_q2a_TEST It will also not load mappings for these questions https://closer-archivist-staging.herokuapp.com/admin/instruments/1989/imports/546
1.0
REACT: questions in condition Else branch and loops are not showing in map view and can't be mapped via .txt files - For example ncds_81_i question qc_currentjob_17_a_B is in the Else branch of c_currentjob_q17cA. https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/ However this question does not appear in the map view https://closer-archivist-staging.herokuapp.com/instruments/ncds_81_i/map/ This is also true for questions in loops e.g. qc_familyincome_2_a_i which is inside l_familyincome_q2a_TEST It will also not load mappings for these questions https://closer-archivist-staging.herokuapp.com/admin/instruments/1989/imports/546
priority
react questions in condition else branch and loops are not showing in map view and can t be mapped via txt files for example ncds i question qc currentjob a b is in the else branch of c currentjob however this question does not appear in the map view this is also true for questions in loops e g qc familyincome a i which is inside l familyincome test it will also not load mappings for these questions
1
211,113
7,198,487,847
IssuesEvent
2018-02-05 13:01:15
hpi-swt2/sport-portal
https://api.github.com/repos/hpi-swt2/sport-portal
closed
Unlogged user can't see team members on any team page
epic team management po-review priority high team kickerz
Table with team members on any team page should be invisible for unsigned users. - Go to `Teams` page as unlogged user - Open any team - You can see team members, though you should not ![sportify](https://user-images.githubusercontent.com/27727717/35574661-b39bc8e6-05da-11e8-9e28-e81a2eaf4b24.png)
1.0
Unlogged user can't see team members on any team page - Table with team members on any team page should be invisible for unsigned users. - Go to `Teams` page as unlogged user - Open any team - You can see team members, though you should not ![sportify](https://user-images.githubusercontent.com/27727717/35574661-b39bc8e6-05da-11e8-9e28-e81a2eaf4b24.png)
priority
unlogged user can t see team members on any team page table with team members on any team page should be invisible for unsigned users go to teams page as unlogged user open any team you can see team members though you should not
1
822,307
30,864,530,512
IssuesEvent
2023-08-03 07:03:30
doubleopen-project/dos
https://api.github.com/repos/doubleopen-project/dos
closed
[DO-109] Stress test issues
bug High priority
This is a top level issue to mark down all issues which have arisen from stress testing DOS with the DOS software itself. <sub>From [SyncLinear.com](https://synclinear.com) | [DO-109](https://linear.app/double-open/issue/DO-109/stress-test-issues)</sub>
1.0
[DO-109] Stress test issues - This is a top level issue to mark down all issues which have arisen from stress testing DOS with the DOS software itself. <sub>From [SyncLinear.com](https://synclinear.com) | [DO-109](https://linear.app/double-open/issue/DO-109/stress-test-issues)</sub>
priority
stress test issues this is a top level issue to mark down all issues which have arisen from stress testing dos with the dos software itself from
1
212,441
7,237,130,143
IssuesEvent
2018-02-13 09:57:46
cyberFund/cyber-design
https://api.github.com/repos/cyberFund/cyber-design
closed
Setup design workflow
Priority: High Status: Completed Type: Analytics
In https://github.com/cyberFund/cyber-ui/issues/37#issuecomment-361559171 we've decided that concepts (images, tables, drawings) are added to `cyber-design` repo, where they are processed and after mockups (psd, sketch) and finalized. After that, @Olga584 creates an issue in cyber-browser repository that new design is ready and @Pechalka converts mockups to layouts (html+js).
1.0
Setup design workflow - In https://github.com/cyberFund/cyber-ui/issues/37#issuecomment-361559171 we've decided that concepts (images, tables, drawings) are added to `cyber-design` repo, where they are processed and after mockups (psd, sketch) and finalized. After that, @Olga584 creates an issue in cyber-browser repository that new design is ready and @Pechalka converts mockups to layouts (html+js).
priority
setup design workflow in we ve decided that concepts images tables drawings are added to cyber design repo where they are processed and after mockups psd sketch and finalized after that creates an issue in cyber browser repository that new design is ready and pechalka converts mockups to layouts html js
1
281,319
8,693,752,207
IssuesEvent
2018-12-04 10:37:03
pombase/canto
https://api.github.com/repos/pombase/canto
closed
combineGenotypes does not store strain name or taxon ID
PHI-Canto bug high priority
(Sub-issue of #1700) The function `combineGenotypes` in `genotypeListViewCtrl` makes a call to `CursGenotypeList.storeGenotype`, but doesn't specify values for the last two arguments `taxonid` and `strain_name`. https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L4689-L4690 Here's the full list of expected arguments, for reference: https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L382-L383 @kimrutherford can this be fixed just by passing in the strain of one of the genotypes in `$scope.genotypeList?` A lot of the existing code seems to be using `$.map` and `$.grep` to collect values from every genotype, but since we're planning to enforce the strain being equal for every allele in a multi-allele genotype, we shouldn't need to check them all. Now that I understand that this strain equality check should probably be in `combineGenotypes`, I might as well fix all these issues at once. I've opened an issue for this at #1704. Also, shouldn't we be passing in the taxon ID as well?
1.0
combineGenotypes does not store strain name or taxon ID - (Sub-issue of #1700) The function `combineGenotypes` in `genotypeListViewCtrl` makes a call to `CursGenotypeList.storeGenotype`, but doesn't specify values for the last two arguments `taxonid` and `strain_name`. https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L4689-L4690 Here's the full list of expected arguments, for reference: https://github.com/pombase/canto/blob/f161098a78ed069410c95ecb4816140479d8fe3f/root/static/js/canto-modules.js#L382-L383 @kimrutherford can this be fixed just by passing in the strain of one of the genotypes in `$scope.genotypeList?` A lot of the existing code seems to be using `$.map` and `$.grep` to collect values from every genotype, but since we're planning to enforce the strain being equal for every allele in a multi-allele genotype, we shouldn't need to check them all. Now that I understand that this strain equality check should probably be in `combineGenotypes`, I might as well fix all these issues at once. I've opened an issue for this at #1704. Also, shouldn't we be passing in the taxon ID as well?
priority
combinegenotypes does not store strain name or taxon id sub issue of the function combinegenotypes in genotypelistviewctrl makes a call to cursgenotypelist storegenotype but doesn t specify values for the last two arguments taxonid and strain name here s the full list of expected arguments for reference kimrutherford can this be fixed just by passing in the strain of one of the genotypes in scope genotypelist a lot of the existing code seems to be using map and grep to collect values from every genotype but since we re planning to enforce the strain being equal for every allele in a multi allele genotype we shouldn t need to check them all now that i understand that this strain equality check should probably be in combinegenotypes i might as well fix all these issues at once i ve opened an issue for this at also shouldn t we be passing in the taxon id as well
1
612,644
19,027,475,083
IssuesEvent
2021-11-24 06:34:42
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Creating Same-Name API after deleting the original one casuses gateway artifact retrieval API to fail on MySQL.
Type/Bug Priority/High WUM Affected/3.2.0
Related Test Case [GatewayRestAPITestCase](https://github.com/wso2-support/product-apim/blob/support-3.2.0.x-full/modules/integration/tests-integration/tests-backend/src/test/java/org/wso2/am/integration/tests/restapi/GatewayRestAPITestCase.java) GatewayRestAPITestCase is failing when it is run against MySQL DB. Following is the steps to replicate it in a local environment **Steps to reproduce** 1. Setup APIM 3.2 with MySQL DB 2. Create API and publish. (Say TestAPI) 3. Delete this API and again create the API with the same name (TestAPI) 4. Invoke /gateway/v1/api-artifact API (ex: https://localhost:9443/api/am/gateway/v1/api-artifact?apiName=TestAPI&version=1.0.0&tenantDomain=carbon.super) Above invocation would return in 400 status code with `TestAPI is not deployed in the Gateway` message. **Reason**. When /api-artifac is called, it calls SynapseAttributesApiServiceImpl[1] service impl and here it queries for the APIId using apiname, version and tenant domain from the AM_GW_PUBLISHED_API_DETAILS table [2]. When an API is deleted, entries are not removed from AM_GW_PUBLISHED_API_DETAILS table. As a result, when same-name APIs are created after deleting the first one, there are two entries AM_GW_PUBLISHED_API_DETAILS in this table. In MySQL scenario, when apiId is retrieved from[2], it returns the deleted API's one. This behavior is dependant on the database (This test is passing in H2) **Fix** Remove old entries from the AM_GW_PUBLISHED_API_DETAILS table (and check whether we could remove entries from AM_GW_API_ARTIFACTS as well) [1] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.internal.service/src/main/java/org/wso2/carbon/apimgt/internal/service/impl/SynapseAttributesApiServiceImpl.java#L29 [2] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/GatewayArtifactsMgtDAO.java#L287
1.0
Creating Same-Name API after deleting the original one casuses gateway artifact retrieval API to fail on MySQL. - Related Test Case [GatewayRestAPITestCase](https://github.com/wso2-support/product-apim/blob/support-3.2.0.x-full/modules/integration/tests-integration/tests-backend/src/test/java/org/wso2/am/integration/tests/restapi/GatewayRestAPITestCase.java) GatewayRestAPITestCase is failing when it is run against MySQL DB. Following is the steps to replicate it in a local environment **Steps to reproduce** 1. Setup APIM 3.2 with MySQL DB 2. Create API and publish. (Say TestAPI) 3. Delete this API and again create the API with the same name (TestAPI) 4. Invoke /gateway/v1/api-artifact API (ex: https://localhost:9443/api/am/gateway/v1/api-artifact?apiName=TestAPI&version=1.0.0&tenantDomain=carbon.super) Above invocation would return in 400 status code with `TestAPI is not deployed in the Gateway` message. **Reason**. When /api-artifac is called, it calls SynapseAttributesApiServiceImpl[1] service impl and here it queries for the APIId using apiname, version and tenant domain from the AM_GW_PUBLISHED_API_DETAILS table [2]. When an API is deleted, entries are not removed from AM_GW_PUBLISHED_API_DETAILS table. As a result, when same-name APIs are created after deleting the first one, there are two entries AM_GW_PUBLISHED_API_DETAILS in this table. In MySQL scenario, when apiId is retrieved from[2], it returns the deleted API's one. This behavior is dependant on the database (This test is passing in H2) **Fix** Remove old entries from the AM_GW_PUBLISHED_API_DETAILS table (and check whether we could remove entries from AM_GW_API_ARTIFACTS as well) [1] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.internal.service/src/main/java/org/wso2/carbon/apimgt/internal/service/impl/SynapseAttributesApiServiceImpl.java#L29 [2] https://github.com/wso2-support/carbon-apimgt/blob/v6.7.206.186/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/GatewayArtifactsMgtDAO.java#L287
priority
creating same name api after deleting the original one casuses gateway artifact retrieval api to fail on mysql related test case gatewayrestapitestcase is failing when it is run against mysql db following is the steps to replicate it in a local environment steps to reproduce setup apim with mysql db create api and publish say testapi delete this api and again create the api with the same name testapi invoke gateway api artifact api ex above invocation would return in status code with testapi is not deployed in the gateway message reason when api artifac is called it calls synapseattributesapiserviceimpl service impl and here it queries for the apiid using apiname version and tenant domain from the am gw published api details table when an api is deleted entries are not removed from am gw published api details table as a result when same name apis are created after deleting the first one there are two entries am gw published api details in this table in mysql scenario when apiid is retrieved from it returns the deleted api s one this behavior is dependant on the database this test is passing in fix remove old entries from the am gw published api details table and check whether we could remove entries from am gw api artifacts as well
1
433,947
12,512,765,973
IssuesEvent
2020-06-02 23:47:43
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
kernel: k_queue_get return NULL before timeout
area: Kernel bug priority: high
**Describe the bug** The API call `k_queue_get` return no data element before timeout has expired. Calling either `k_queue_get(queue, K_FOREVER);` or `k_queue_get(queue, K_SECONDS(20))` return `NULL` within one second. This appears to occur because two threads are both waiting for an element in the queue. Once an element is posted to the queue, both threads are woken, one will retrieve the new element, while the other will discover an empty queue. This is a regression from: 7832738ae985a63febb8f82e7c4e34824f48486e Specifically this change: ``` -static void *k_queue_poll(struct k_queue *queue, s32_t timeout) +static void *k_queue_poll(struct k_queue *queue, k_timeout_t timeout) { struct k_poll_event event; - int err, elapsed = 0, done = 0; + int err; k_spinlock_key_t key; void *val; - u32_t start; k_poll_event_init(&event, K_POLL_TYPE_FIFO_DATA_AVAILABLE, K_POLL_MODE_NOTIFY_ONLY, queue); - if (timeout != K_FOREVER) { - start = k_uptime_get_32(); - } - - do { - event.state = K_POLL_STATE_NOT_READY; - - err = k_poll(&event, 1, timeout - elapsed); - - if (err && err != -EAGAIN) { - return NULL; - } + event.state = K_POLL_STATE_NOT_READY; + err = k_poll(&event, 1, timeout); - key = k_spin_lock(&queue->lock); - val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true); - k_spin_unlock(&queue->lock, key); + if (err && err != -EAGAIN) { + return NULL; + } - if ((val == NULL) && (timeout != K_FOREVER)) { - elapsed = k_uptime_get_32() - start; - done = elapsed > timeout; - } - } while (!val && !done); + key = k_spin_lock(&queue->lock); + val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true); + k_spin_unlock(&queue->lock, key); return val; ``` Reverting this change (using legacy timeout API) fixes the issue. The commit message says this: ``` Also in queue.c, a (when POLL was enabled) a similar loop was needlessly used to try to retry the k_poll() call after a spurious failure. But k_poll() does not fail spuriously, so the loop was removed. ``` The loop appears to have been removed for the wrong reason. **To Reproduce** The current steps involves 2 nRF52 dev-kits and a few manual steps. I can try to make a more minimal failing test if needed, otherwise I can verify using my current setup. Use branch: https://github.com/joerchan/zephyr/tree/bt-recv-deadlock-debug ``` export BOARD=nrf52840dk west build test/bluetooth/shell west flash --erase west build samples/bluetooth/peripheral west flash --erase # Attach one terminal to each # shell terminal: bt init bt connect <id addr> bt gatt exchange-mtu ``` **Console output** peripheral sample: ``` Advertising successfully started Connected: 4e:1a:6f:9b:b8:00 (random) Write: count= 122, len= 2440, rate= 19523 bps. Write: count= 1266, len= 25320, rate= 202578 bps. Write: count= 1815, len= 36300, rate= 290408 bps. Write: count= 1768, len= 35360, rate= 282880 bps. Write cmd failed (-55). Write: count= 1222, len= 37964, rate= 303758 bps. [00:00:13.090,026] <err> bt_conn: k_is_in_isr 0 [00:00:13.090,057] <err> bt_conn: conn_tx_alloc timeout 0 secs. [00:00:13.090,087] <err> bt_conn: Unable to allocate TX context Write: count= 603, len= 37386, rate= 299444 bps. LE conn param updated: int 0x0028 lat 0 to 42 Write: count= 529, len= 32798, rate= 262680 bps. Write: count= 594, len= 36828, rate= 294957 bps. Write: count= 654, len= 40548, rate= 324750 bps. Write: count= 671, len= 41602, rate= 333049 bps. Disconnected (reason 0x13) Write cmd failed (-57). ``` **Expected behavior** k_queue_get should not return NULL before timeout has passed. **Impact** This breaks the current flow-control behavior of Bluetooth, the attempted k_queue_get from BT RX thread will drop the attempt to answer the ATT request and will result in a disconnected ATT channel. **Additional context** CONFIG_POLL is enable Blocker to fix: https://github.com/zephyrproject-rtos/zephyr/issues/23364
1.0
kernel: k_queue_get return NULL before timeout - **Describe the bug** The API call `k_queue_get` return no data element before timeout has expired. Calling either `k_queue_get(queue, K_FOREVER);` or `k_queue_get(queue, K_SECONDS(20))` return `NULL` within one second. This appears to occur because two threads are both waiting for an element in the queue. Once an element is posted to the queue, both threads are woken, one will retrieve the new element, while the other will discover an empty queue. This is a regression from: 7832738ae985a63febb8f82e7c4e34824f48486e Specifically this change: ``` -static void *k_queue_poll(struct k_queue *queue, s32_t timeout) +static void *k_queue_poll(struct k_queue *queue, k_timeout_t timeout) { struct k_poll_event event; - int err, elapsed = 0, done = 0; + int err; k_spinlock_key_t key; void *val; - u32_t start; k_poll_event_init(&event, K_POLL_TYPE_FIFO_DATA_AVAILABLE, K_POLL_MODE_NOTIFY_ONLY, queue); - if (timeout != K_FOREVER) { - start = k_uptime_get_32(); - } - - do { - event.state = K_POLL_STATE_NOT_READY; - - err = k_poll(&event, 1, timeout - elapsed); - - if (err && err != -EAGAIN) { - return NULL; - } + event.state = K_POLL_STATE_NOT_READY; + err = k_poll(&event, 1, timeout); - key = k_spin_lock(&queue->lock); - val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true); - k_spin_unlock(&queue->lock, key); + if (err && err != -EAGAIN) { + return NULL; + } - if ((val == NULL) && (timeout != K_FOREVER)) { - elapsed = k_uptime_get_32() - start; - done = elapsed > timeout; - } - } while (!val && !done); + key = k_spin_lock(&queue->lock); + val = z_queue_node_peek(sys_sflist_get(&queue->data_q), true); + k_spin_unlock(&queue->lock, key); return val; ``` Reverting this change (using legacy timeout API) fixes the issue. The commit message says this: ``` Also in queue.c, a (when POLL was enabled) a similar loop was needlessly used to try to retry the k_poll() call after a spurious failure. But k_poll() does not fail spuriously, so the loop was removed. ``` The loop appears to have been removed for the wrong reason. **To Reproduce** The current steps involves 2 nRF52 dev-kits and a few manual steps. I can try to make a more minimal failing test if needed, otherwise I can verify using my current setup. Use branch: https://github.com/joerchan/zephyr/tree/bt-recv-deadlock-debug ``` export BOARD=nrf52840dk west build test/bluetooth/shell west flash --erase west build samples/bluetooth/peripheral west flash --erase # Attach one terminal to each # shell terminal: bt init bt connect <id addr> bt gatt exchange-mtu ``` **Console output** peripheral sample: ``` Advertising successfully started Connected: 4e:1a:6f:9b:b8:00 (random) Write: count= 122, len= 2440, rate= 19523 bps. Write: count= 1266, len= 25320, rate= 202578 bps. Write: count= 1815, len= 36300, rate= 290408 bps. Write: count= 1768, len= 35360, rate= 282880 bps. Write cmd failed (-55). Write: count= 1222, len= 37964, rate= 303758 bps. [00:00:13.090,026] <err> bt_conn: k_is_in_isr 0 [00:00:13.090,057] <err> bt_conn: conn_tx_alloc timeout 0 secs. [00:00:13.090,087] <err> bt_conn: Unable to allocate TX context Write: count= 603, len= 37386, rate= 299444 bps. LE conn param updated: int 0x0028 lat 0 to 42 Write: count= 529, len= 32798, rate= 262680 bps. Write: count= 594, len= 36828, rate= 294957 bps. Write: count= 654, len= 40548, rate= 324750 bps. Write: count= 671, len= 41602, rate= 333049 bps. Disconnected (reason 0x13) Write cmd failed (-57). ``` **Expected behavior** k_queue_get should not return NULL before timeout has passed. **Impact** This breaks the current flow-control behavior of Bluetooth, the attempted k_queue_get from BT RX thread will drop the attempt to answer the ATT request and will result in a disconnected ATT channel. **Additional context** CONFIG_POLL is enable Blocker to fix: https://github.com/zephyrproject-rtos/zephyr/issues/23364
priority
kernel k queue get return null before timeout describe the bug the api call k queue get return no data element before timeout has expired calling either k queue get queue k forever or k queue get queue k seconds return null within one second this appears to occur because two threads are both waiting for an element in the queue once an element is posted to the queue both threads are woken one will retrieve the new element while the other will discover an empty queue this is a regression from specifically this change static void k queue poll struct k queue queue t timeout static void k queue poll struct k queue queue k timeout t timeout struct k poll event event int err elapsed done int err k spinlock key t key void val t start k poll event init event k poll type fifo data available k poll mode notify only queue if timeout k forever start k uptime get do event state k poll state not ready err k poll event timeout elapsed if err err eagain return null event state k poll state not ready err k poll event timeout key k spin lock queue lock val z queue node peek sys sflist get queue data q true k spin unlock queue lock key if err err eagain return null if val null timeout k forever elapsed k uptime get start done elapsed timeout while val done key k spin lock queue lock val z queue node peek sys sflist get queue data q true k spin unlock queue lock key return val reverting this change using legacy timeout api fixes the issue the commit message says this also in queue c a when poll was enabled a similar loop was needlessly used to try to retry the k poll call after a spurious failure but k poll does not fail spuriously so the loop was removed the loop appears to have been removed for the wrong reason to reproduce the current steps involves dev kits and a few manual steps i can try to make a more minimal failing test if needed otherwise i can verify using my current setup use branch export board west build test bluetooth shell west flash erase west build samples bluetooth peripheral west flash erase attach one terminal to each shell terminal bt init bt connect bt gatt exchange mtu console output peripheral sample advertising successfully started connected random write count len rate bps write count len rate bps write count len rate bps write count len rate bps write cmd failed write count len rate bps bt conn k is in isr bt conn conn tx alloc timeout secs bt conn unable to allocate tx context write count len rate bps le conn param updated int lat to write count len rate bps write count len rate bps write count len rate bps write count len rate bps disconnected reason write cmd failed expected behavior k queue get should not return null before timeout has passed impact this breaks the current flow control behavior of bluetooth the attempted k queue get from bt rx thread will drop the attempt to answer the att request and will result in a disconnected att channel additional context config poll is enable blocker to fix
1
74,688
3,442,288,270
IssuesEvent
2015-12-14 21:58:26
cytoscape/cytoscape.js
https://api.github.com/repos/cytoscape/cytoscape.js
opened
Calculate rendered styles dynamically
1-high-priority performance renderer
Rather than diffing on each frame to check for a calculation, just use events to determine when to recalculate the rendered style
1.0
Calculate rendered styles dynamically - Rather than diffing on each frame to check for a calculation, just use events to determine when to recalculate the rendered style
priority
calculate rendered styles dynamically rather than diffing on each frame to check for a calculation just use events to determine when to recalculate the rendered style
1
656,240
21,724,195,225
IssuesEvent
2022-05-11 05:39:29
jordan-sullivan/flashcards-2.5
https://api.github.com/repos/jordan-sullivan/flashcards-2.5
opened
Fix last should statement on Turn test
bug high priority
The last 'should' statement won't allow tests to pass , please fix.
1.0
Fix last should statement on Turn test - The last 'should' statement won't allow tests to pass , please fix.
priority
fix last should statement on turn test the last should statement won t allow tests to pass please fix
1
339,469
10,254,900,796
IssuesEvent
2019-08-21 14:26:00
wso2/docker-apim
https://api.github.com/repos/wso2/docker-apim
opened
Support for artifact synchronization between API Manager nodes
Priority/Highest Type/Improvement
**Description:** Currently, WSO2 API Manager highly available deployments require us to persist and share runtime artifacts [1]. For this purpose, we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations. Hence, these content need to be persisted in the image. [1]: [Common Runtime and Configuration Artifacts](https://docs.wso2.com/display/AM260/Common+Runtime+and+Configuration+Artifacts) **Affected Product Version:** Docker resources for WSO2 API Manager v2.6.x or below
1.0
Support for artifact synchronization between API Manager nodes - **Description:** Currently, WSO2 API Manager highly available deployments require us to persist and share runtime artifacts [1]. For this purpose, we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations. Hence, these content need to be persisted in the image. [1]: [Common Runtime and Configuration Artifacts](https://docs.wso2.com/display/AM260/Common+Runtime+and+Configuration+Artifacts) **Affected Product Version:** Docker resources for WSO2 API Manager v2.6.x or below
priority
support for artifact synchronization between api manager nodes description currently api manager highly available deployments require us to persist and share runtime artifacts for this purpose we need to persist the default content available in these shared directories as directory mounts may erase of the original content in these locations hence these content need to be persisted in the image affected product version docker resources for api manager x or below
1
197,837
6,965,107,202
IssuesEvent
2017-12-09 01:52:20
FireflyArtsCollective/ffagc
https://api.github.com/repos/FireflyArtsCollective/ffagc
closed
grant submission creation dying when I submit invalid grant dollar amount
bug High Priority
``` Started POST "/grant_submissions" for 127.0.0.1 at 2017-12-08 20:44:39 -0500 Processing by GrantSubmissionsController#create as HTML Parameters: {"utf8"=>"✓", "authenticity_token"=>"9h3AOLavJs25WE9Gx496yngHuYz9rv//+4Q6bVM+hOGkzNKcYzGU+RYzPav+bp+3UauHWqnFSKGDjBcsyxGieQ==", "grant_submission"=>{"name"=>"again?", "grant_id"=>"1", "requested_funding_dollars"=>"1004", "proposal"=>#<ActionDispatch::Http::UploadedFile:0x007f13b252f7e8 @tempfile=#<Tempfile:/tmp/RackMultipart20171208-22992-9nwb5u.pdf>, @original_filename="3dmark03_audit_report.pdf", @content_type="application/pdf", @headers="Content-Disposition: form-data; name=\"grant_submission[proposal]\"; filename=\"3dmark03_audit_report.pdf\"\r\nContent-Type: application/pdf\r\n">}, "commit"=>"Create Grant submission"} Admin Load (0.1ms) SELECT "admins".* FROM "admins" WHERE "admins"."id" = ? LIMIT 1 [["id", nil]] Artist Load (0.2ms) SELECT "artists".* FROM "artists" WHERE "artists"."id" = ? LIMIT 1 [["id", 3]] Admin Exists (0.2ms) SELECT 1 AS one FROM "admins" LIMIT 1 (0.1ms) SELECT "grant_submissions".id FROM "grant_submissions" WHERE "grant_submissions"."artist_id" = ? [["artist_id", 3]] (0.1ms) begin transaction Grant Load (0.5ms) SELECT "grants".* FROM "grants" WHERE "grants"."id" = ? LIMIT 1 [["id", 1]] (0.1ms) rollback transaction Completed 500 Internal Server Error in 28ms (ActiveRecord: 1.2ms) ActionView::MissingTemplate (Missing template grant_submissions/create, application/create with {:locale=>[:en], :formats=>[:html], :variants=>[], :handlers=>[:erb, :builder, :raw, :ruby, :jbuilder]}. Searched in: * "/home/owen/src/github/ffagc/app/views" ): vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/path_set.rb:46:in `find' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/lookup_context.rb:121:in `find' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/abstract_renderer.rb:18:in `find_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:40:in `determine_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:8:in `render' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:46:in `render_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:27:in `render' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:100:in `_render_template' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/streaming.rb:217:in `_render_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:83:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:32:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/renderers.rb:37:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/rendering.rb:25:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:16:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block (2 levels) in render' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `block in ms' /usr/lib/ruby/2.3.0/benchmark.rb:308:in `realtime' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `ms' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block in render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:87:in `cleanup_view_runtime' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:25:in `cleanup_view_runtime' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:43:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:10:in `default_render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:5:in `send_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:198:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:10:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:20:in `block in process_action' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:117:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:555:in `block (2 levels) in compile' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:505:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:92:in `__run_callbacks__' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_process_action_callbacks' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:19:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rescue.rb:29:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:32:in `block in process_action' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `block in instrument' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications/instrumenter.rb:20:in `instrument' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `instrument' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:30:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/params_wrapper.rb:250:in `process_action' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:18:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:137:in `process' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:30:in `process' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:196:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rack_delegation.rb:13:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:237:in `block in action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:74:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:43:in `serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:43:in `block in serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `each' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:817:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/etag.rb:24:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/conditionalget.rb:38:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/head.rb:13:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/params_parser.rb:27:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/flash.rb:260:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:225:in `context' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:220:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/cookies.rb:560:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/query_cache.rb:36:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/connection_adapters/abstract/connection_pool.rb:653:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/migration.rb:377:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:29:in `block in call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:88:in `__run_callbacks__' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_call_callbacks' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:27:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/reloader.rb:73:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/remote_ip.rb:78:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/debug_exceptions.rb:17:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/show_exceptions.rb:30:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:38:in `call_app' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `block in call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `block in tagged' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:26:in `tagged' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `tagged' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/request_id.rb:21:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/methodoverride.rb:22:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/runtime.rb:18:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/cache/strategy/local_cache_middleware.rb:28:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/static.rb:120:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/sendfile.rb:113:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/engine.rb:518:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/application.rb:165:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/content_length.rb:15:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/handler/webrick.rb:88:in `service' /usr/lib/ruby/2.3.0/webrick/httpserver.rb:140:in `service' /usr/lib/ruby/2.3.0/webrick/httpserver.rb:96:in `run' /usr/lib/ruby/2.3.0/webrick/server.rb:296:in `block in start_thread' Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_source.erb (18.9ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_trace.html.erb (3.7ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_request_and_response.html.erb (1.4ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/missing_template.html.erb within rescues/layout (36.8ms) ```
1.0
grant submission creation dying when I submit invalid grant dollar amount - ``` Started POST "/grant_submissions" for 127.0.0.1 at 2017-12-08 20:44:39 -0500 Processing by GrantSubmissionsController#create as HTML Parameters: {"utf8"=>"✓", "authenticity_token"=>"9h3AOLavJs25WE9Gx496yngHuYz9rv//+4Q6bVM+hOGkzNKcYzGU+RYzPav+bp+3UauHWqnFSKGDjBcsyxGieQ==", "grant_submission"=>{"name"=>"again?", "grant_id"=>"1", "requested_funding_dollars"=>"1004", "proposal"=>#<ActionDispatch::Http::UploadedFile:0x007f13b252f7e8 @tempfile=#<Tempfile:/tmp/RackMultipart20171208-22992-9nwb5u.pdf>, @original_filename="3dmark03_audit_report.pdf", @content_type="application/pdf", @headers="Content-Disposition: form-data; name=\"grant_submission[proposal]\"; filename=\"3dmark03_audit_report.pdf\"\r\nContent-Type: application/pdf\r\n">}, "commit"=>"Create Grant submission"} Admin Load (0.1ms) SELECT "admins".* FROM "admins" WHERE "admins"."id" = ? LIMIT 1 [["id", nil]] Artist Load (0.2ms) SELECT "artists".* FROM "artists" WHERE "artists"."id" = ? LIMIT 1 [["id", 3]] Admin Exists (0.2ms) SELECT 1 AS one FROM "admins" LIMIT 1 (0.1ms) SELECT "grant_submissions".id FROM "grant_submissions" WHERE "grant_submissions"."artist_id" = ? [["artist_id", 3]] (0.1ms) begin transaction Grant Load (0.5ms) SELECT "grants".* FROM "grants" WHERE "grants"."id" = ? LIMIT 1 [["id", 1]] (0.1ms) rollback transaction Completed 500 Internal Server Error in 28ms (ActiveRecord: 1.2ms) ActionView::MissingTemplate (Missing template grant_submissions/create, application/create with {:locale=>[:en], :formats=>[:html], :variants=>[], :handlers=>[:erb, :builder, :raw, :ruby, :jbuilder]}. Searched in: * "/home/owen/src/github/ffagc/app/views" ): vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/path_set.rb:46:in `find' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/lookup_context.rb:121:in `find' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/abstract_renderer.rb:18:in `find_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:40:in `determine_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/template_renderer.rb:8:in `render' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:46:in `render_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/renderer/renderer.rb:27:in `render' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:100:in `_render_template' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/streaming.rb:217:in `_render_template' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:83:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:32:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/renderers.rb:37:in `render_to_body' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/rendering.rb:25:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:16:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block (2 levels) in render' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `block in ms' /usr/lib/ruby/2.3.0/benchmark.rb:308:in `realtime' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/core_ext/benchmark.rb:12:in `ms' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:44:in `block in render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:87:in `cleanup_view_runtime' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:25:in `cleanup_view_runtime' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:43:in `render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:10:in `default_render' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/implicit_render.rb:5:in `send_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:198:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rendering.rb:10:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:20:in `block in process_action' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:117:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:555:in `block (2 levels) in compile' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:505:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:92:in `__run_callbacks__' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_process_action_callbacks' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/callbacks.rb:19:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rescue.rb:29:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:32:in `block in process_action' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `block in instrument' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications/instrumenter.rb:20:in `instrument' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/notifications.rb:164:in `instrument' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/instrumentation.rb:30:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/params_wrapper.rb:250:in `process_action' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/railties/controller_runtime.rb:18:in `process_action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/abstract_controller/base.rb:137:in `process' vendor/bundle/ruby/2.3.0actionview (4.2.7.1) lib/action_view/rendering.rb:30:in `process' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:196:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal/rack_delegation.rb:13:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_controller/metal.rb:237:in `block in action' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:74:in `dispatch' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:43:in `serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:43:in `block in serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `each' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/journey/router.rb:30:in `serve' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/routing/route_set.rb:817:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/etag.rb:24:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/conditionalget.rb:38:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/head.rb:13:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/params_parser.rb:27:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/flash.rb:260:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:225:in `context' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/session/abstract/id.rb:220:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/cookies.rb:560:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/query_cache.rb:36:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/connection_adapters/abstract/connection_pool.rb:653:in `call' vendor/bundle/ruby/2.3.0activerecord (4.2.7.1) lib/active_record/migration.rb:377:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:29:in `block in call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:88:in `__run_callbacks__' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:778:in `_run_call_callbacks' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/callbacks.rb:81:in `run_callbacks' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/callbacks.rb:27:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/reloader.rb:73:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/remote_ip.rb:78:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/debug_exceptions.rb:17:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/show_exceptions.rb:30:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:38:in `call_app' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `block in call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `block in tagged' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:26:in `tagged' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/tagged_logging.rb:68:in `tagged' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/rack/logger.rb:20:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/request_id.rb:21:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/methodoverride.rb:22:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/runtime.rb:18:in `call' vendor/bundle/ruby/2.3.0activesupport (4.2.7.1) lib/active_support/cache/strategy/local_cache_middleware.rb:28:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call' vendor/bundle/ruby/2.3.0actionpack (4.2.7.1) lib/action_dispatch/middleware/static.rb:120:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/sendfile.rb:113:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/engine.rb:518:in `call' vendor/bundle/ruby/2.3.0railties (4.2.7.1) lib/rails/application.rb:165:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/lock.rb:17:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/content_length.rb:15:in `call' vendor/bundle/ruby/2.3.0rack (1.6.5) lib/rack/handler/webrick.rb:88:in `service' /usr/lib/ruby/2.3.0/webrick/httpserver.rb:140:in `service' /usr/lib/ruby/2.3.0/webrick/httpserver.rb:96:in `run' /usr/lib/ruby/2.3.0/webrick/server.rb:296:in `block in start_thread' Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_source.erb (18.9ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_trace.html.erb (3.7ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/_request_and_response.html.erb (1.4ms) Rendered vendor/bundle/ruby/2.3.0/gems/actionpack-4.2.7.1/lib/action_dispatch/middleware/templates/rescues/missing_template.html.erb within rescues/layout (36.8ms) ```
priority
grant submission creation dying when i submit invalid grant dollar amount started post grant submissions for at processing by grantsubmissionscontroller create as html parameters ✓ authenticity token hogkznkcyzgu ryzpav bp grant submission name again grant id requested funding dollars proposal original filename audit report pdf content type application pdf headers content disposition form data name grant submission filename audit report pdf r ncontent type application pdf r n commit create grant submission admin load select admins from admins where admins id limit artist load select artists from artists where artists id limit admin exists select as one from admins limit select grant submissions id from grant submissions where grant submissions artist id begin transaction grant load select grants from grants where grants id limit rollback transaction completed internal server error in activerecord actionview missingtemplate missing template grant submissions create application create with locale formats variants handlers searched in home owen src github ffagc app views vendor bundle ruby lib action view path set rb in find vendor bundle ruby lib action view lookup context rb in find vendor bundle ruby lib action view renderer abstract renderer rb in find template vendor bundle ruby lib action view renderer template renderer rb in determine template vendor bundle ruby lib action view renderer template renderer rb in render vendor bundle ruby lib action view renderer renderer rb in render template vendor bundle ruby lib action view renderer renderer rb in render vendor bundle ruby lib action view rendering rb in render template vendor bundle ruby lib action controller metal streaming rb in render template vendor bundle ruby lib action view rendering rb in render to body vendor bundle ruby lib action controller metal rendering rb in render to body vendor bundle ruby lib action controller metal renderers rb in render to body vendor bundle ruby lib abstract controller rendering rb in render vendor bundle ruby lib action controller metal rendering rb in render vendor bundle ruby lib action controller metal instrumentation rb in block levels in render vendor bundle ruby lib active support core ext benchmark rb in block in ms usr lib ruby benchmark rb in realtime vendor bundle ruby lib active support core ext benchmark rb in ms vendor bundle ruby lib action controller metal instrumentation rb in block in render vendor bundle ruby lib action controller metal instrumentation rb in cleanup view runtime vendor bundle ruby lib active record railties controller runtime rb in cleanup view runtime vendor bundle ruby lib action controller metal instrumentation rb in render vendor bundle ruby lib action controller metal implicit render rb in default render vendor bundle ruby lib action controller metal implicit render rb in send action vendor bundle ruby lib abstract controller base rb in process action vendor bundle ruby lib action controller metal rendering rb in process action vendor bundle ruby lib abstract controller callbacks rb in block in process action vendor bundle ruby lib active support callbacks rb in call vendor bundle ruby lib active support callbacks rb in block levels in compile vendor bundle ruby lib active support callbacks rb in call vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib active support callbacks rb in run process action callbacks vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib abstract controller callbacks rb in process action vendor bundle ruby lib action controller metal rescue rb in process action vendor bundle ruby lib action controller metal instrumentation rb in block in process action vendor bundle ruby lib active support notifications rb in block in instrument vendor bundle ruby lib active support notifications instrumenter rb in instrument vendor bundle ruby lib active support notifications rb in instrument vendor bundle ruby lib action controller metal instrumentation rb in process action vendor bundle ruby lib action controller metal params wrapper rb in process action vendor bundle ruby lib active record railties controller runtime rb in process action vendor bundle ruby lib abstract controller base rb in process vendor bundle ruby lib action view rendering rb in process vendor bundle ruby lib action controller metal rb in dispatch vendor bundle ruby lib action controller metal rack delegation rb in dispatch vendor bundle ruby lib action controller metal rb in block in action vendor bundle ruby lib action dispatch routing route set rb in dispatch vendor bundle ruby lib action dispatch routing route set rb in serve vendor bundle ruby lib action dispatch journey router rb in block in serve vendor bundle ruby lib action dispatch journey router rb in each vendor bundle ruby lib action dispatch journey router rb in serve vendor bundle ruby lib action dispatch routing route set rb in call vendor bundle ruby lib rack etag rb in call vendor bundle ruby lib rack conditionalget rb in call vendor bundle ruby lib rack head rb in call vendor bundle ruby lib action dispatch middleware params parser rb in call vendor bundle ruby lib action dispatch middleware flash rb in call vendor bundle ruby lib rack session abstract id rb in context vendor bundle ruby lib rack session abstract id rb in call vendor bundle ruby lib action dispatch middleware cookies rb in call vendor bundle ruby lib active record query cache rb in call vendor bundle ruby lib active record connection adapters abstract connection pool rb in call vendor bundle ruby lib active record migration rb in call vendor bundle ruby lib action dispatch middleware callbacks rb in block in call vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib active support callbacks rb in run call callbacks vendor bundle ruby lib active support callbacks rb in run callbacks vendor bundle ruby lib action dispatch middleware callbacks rb in call vendor bundle ruby lib action dispatch middleware reloader rb in call vendor bundle ruby lib action dispatch middleware remote ip rb in call vendor bundle ruby lib action dispatch middleware debug exceptions rb in call vendor bundle ruby lib action dispatch middleware show exceptions rb in call vendor bundle ruby lib rails rack logger rb in call app vendor bundle ruby lib rails rack logger rb in block in call vendor bundle ruby lib active support tagged logging rb in block in tagged vendor bundle ruby lib active support tagged logging rb in tagged vendor bundle ruby lib active support tagged logging rb in tagged vendor bundle ruby lib rails rack logger rb in call vendor bundle ruby lib action dispatch middleware request id rb in call vendor bundle ruby lib rack methodoverride rb in call vendor bundle ruby lib rack runtime rb in call vendor bundle ruby lib active support cache strategy local cache middleware rb in call vendor bundle ruby lib rack lock rb in call vendor bundle ruby lib action dispatch middleware static rb in call vendor bundle ruby lib rack sendfile rb in call vendor bundle ruby lib rails engine rb in call vendor bundle ruby lib rails application rb in call vendor bundle ruby lib rack lock rb in call vendor bundle ruby lib rack content length rb in call vendor bundle ruby lib rack handler webrick rb in service usr lib ruby webrick httpserver rb in service usr lib ruby webrick httpserver rb in run usr lib ruby webrick server rb in block in start thread rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues source erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues trace html erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues request and response html erb rendered vendor bundle ruby gems actionpack lib action dispatch middleware templates rescues missing template html erb within rescues layout
1
185,929
6,732,010,213
IssuesEvent
2017-10-18 09:47:29
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
closed
[Firefox] Package declaration doesn't take effect
0.94-pre-release Priority/Highest Severity/Critical
Browser: Firefox 56.0 (64-bit) **Steps** 1. Add a package 2. Open split view 3. Reopen the package declaration box after refreshing the page The package dissapears from the package declaration ui component. ![com_0 94_package_1](https://user-images.githubusercontent.com/15624590/31648776-34ee8b28-b32d-11e7-91f0-ad17d91bf994.png)
1.0
[Firefox] Package declaration doesn't take effect - Browser: Firefox 56.0 (64-bit) **Steps** 1. Add a package 2. Open split view 3. Reopen the package declaration box after refreshing the page The package dissapears from the package declaration ui component. ![com_0 94_package_1](https://user-images.githubusercontent.com/15624590/31648776-34ee8b28-b32d-11e7-91f0-ad17d91bf994.png)
priority
package declaration doesn t take effect browser firefox bit steps add a package open split view reopen the package declaration box after refreshing the page the package dissapears from the package declaration ui component
1
710,225
24,411,467,636
IssuesEvent
2022-10-05 12:45:00
HiAvatar/backend
https://api.github.com/repos/HiAvatar/backend
closed
오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결
Type: Bug Priority: High
### Description 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다.
1.0
오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결 - ### Description 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다.
priority
오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그 해결 description 오디오 파일을 데이터베이스에 저장할 때 파일 확장자만 반영되는 버그를 해결한다
1
62,426
3,185,250,908
IssuesEvent
2015-09-28 01:56:42
babel/babel
https://api.github.com/repos/babel/babel
closed
Improve require speed
discussion good first bug help wanted high priority
This is an issue to track possible require optimisations. Not sure if there's anything that can be done but it's worth investigating. /cc @Globegitter @stefanpenner
1.0
Improve require speed - This is an issue to track possible require optimisations. Not sure if there's anything that can be done but it's worth investigating. /cc @Globegitter @stefanpenner
priority
improve require speed this is an issue to track possible require optimisations not sure if there s anything that can be done but it s worth investigating cc globegitter stefanpenner
1
252,412
8,036,068,595
IssuesEvent
2018-07-30 07:52:16
systers/powerup-android
https://api.github.com/repos/systers/powerup-android
opened
Has to select the final answer twice to close the game
Priority: HIGH Program: GSoC Type: Bug
## Description As a user, when I'm selecting the answers to questions in different scenarios when I select the last answer, the screen should change to the game points activity. ## Mocks ![20180730_194156](https://user-images.githubusercontent.com/1425395/43384176-53a5668c-9431-11e8-9d8e-3a334b1ee4d5.gif) To recreate: open any scenario keep answering questions until you come to the last question the last answer should be selected twice in order to move to scenario over screen. This issue is caused by the asynchronous call used in Gameactivity: <img width="708" alt="screen shot 2018-07-30 at 7 49 29 pm" src="https://user-images.githubusercontent.com/1425395/43384301-b5c576e0-9431-11e8-9dca-4db66e103795.png"> prevScene is checked before the presenter is loaded with the previous screen. ## Acceptance Criteria ### Update [Required] - [ ] the last answer should be selected only once - [ ] asynchronous calls should be properly handled. ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 2 hours
1.0
Has to select the final answer twice to close the game - ## Description As a user, when I'm selecting the answers to questions in different scenarios when I select the last answer, the screen should change to the game points activity. ## Mocks ![20180730_194156](https://user-images.githubusercontent.com/1425395/43384176-53a5668c-9431-11e8-9d8e-3a334b1ee4d5.gif) To recreate: open any scenario keep answering questions until you come to the last question the last answer should be selected twice in order to move to scenario over screen. This issue is caused by the asynchronous call used in Gameactivity: <img width="708" alt="screen shot 2018-07-30 at 7 49 29 pm" src="https://user-images.githubusercontent.com/1425395/43384301-b5c576e0-9431-11e8-9dca-4db66e103795.png"> prevScene is checked before the presenter is loaded with the previous screen. ## Acceptance Criteria ### Update [Required] - [ ] the last answer should be selected only once - [ ] asynchronous calls should be properly handled. ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 2 hours
priority
has to select the final answer twice to close the game description as a user when i m selecting the answers to questions in different scenarios when i select the last answer the screen should change to the game points activity mocks to recreate open any scenario keep answering questions until you come to the last question the last answer should be selected twice in order to move to scenario over screen this issue is caused by the asynchronous call used in gameactivity img width alt screen shot at pm src prevscene is checked before the presenter is loaded with the previous screen acceptance criteria update the last answer should be selected only once asynchronous calls should be properly handled definition of done all of the required items are completed approval by mentor estimation hours
1
579,747
17,197,984,544
IssuesEvent
2021-07-16 20:43:49
mikemerin/FM-layouts
https://api.github.com/repos/mikemerin/FM-layouts
closed
Announcement upgrade
Category - Graphics Priority - 1 (high) Type - Upgrade / Fix
Adding to dashboard panel (https://github.com/mikemerin/FM2020-layouts/issues/11) and if another genre, linking to that community
1.0
Announcement upgrade - Adding to dashboard panel (https://github.com/mikemerin/FM2020-layouts/issues/11) and if another genre, linking to that community
priority
announcement upgrade adding to dashboard panel and if another genre linking to that community
1
104,329
4,209,474,737
IssuesEvent
2016-06-29 05:37:38
japanesemediamanager/jmmserver
https://api.github.com/repos/japanesemediamanager/jmmserver
closed
Group Filters are not working
Bug - High Priority
How I reproduced - Completely new installation using latest source - DB : SQL Server - Import files into new database - Mark a file as watched from the Metro series view - Nothing is shown in the 'Continue Watching (System)' group filter even though there are more files to watch NOTE - Actually nothing is shown in any of the group filters
1.0
Group Filters are not working - How I reproduced - Completely new installation using latest source - DB : SQL Server - Import files into new database - Mark a file as watched from the Metro series view - Nothing is shown in the 'Continue Watching (System)' group filter even though there are more files to watch NOTE - Actually nothing is shown in any of the group filters
priority
group filters are not working how i reproduced completely new installation using latest source db sql server import files into new database mark a file as watched from the metro series view nothing is shown in the continue watching system group filter even though there are more files to watch note actually nothing is shown in any of the group filters
1
372,160
11,009,917,859
IssuesEvent
2019-12-04 13:39:11
boi123212321/porn-manager
https://api.github.com/repos/boi123212321/porn-manager
closed
Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN
bug high priority
Identified that I have 10 videos in my directory but didn't extract any. `Generating thumbnails...` ``` Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN ``` Pointed the app at a different folder and created a new 'collection' (clearing the previous video queue) and it worked well. I suspect when the app doesn't catch the exception when a video doesn't have a correct duration. Probably very easy to fix -- but I'm not up for creating a PR at this moment. Awesome app! Keep it up 👍
1.0
Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN - Identified that I have 10 videos in my directory but didn't extract any. `Generating thumbnails...` ``` Error: ffmpeg exited with code 1: Invalid duration during specification for ss: NaN ``` Pointed the app at a different folder and created a new 'collection' (clearing the previous video queue) and it worked well. I suspect when the app doesn't catch the exception when a video doesn't have a correct duration. Probably very easy to fix -- but I'm not up for creating a PR at this moment. Awesome app! Keep it up 👍
priority
error ffmpeg exited with code invalid duration during specification for ss nan identified that i have videos in my directory but didn t extract any generating thumbnails error ffmpeg exited with code invalid duration during specification for ss nan pointed the app at a different folder and created a new collection clearing the previous video queue and it worked well i suspect when the app doesn t catch the exception when a video doesn t have a correct duration probably very easy to fix but i m not up for creating a pr at this moment awesome app keep it up 👍
1
106,529
4,273,920,676
IssuesEvent
2016-07-13 18:49:37
Octanis1/Octanis1-Mainboard-Firmware_MSP_EXP432P401RLP
https://api.github.com/repos/Octanis1/Octanis1-Mainboard-Firmware_MSP_EXP432P401RLP
closed
Navigation/Driving not correctly working
bug enhancement high priority in progress
rover does not follow correct direction as entered in apm planner. Also, curve radii are too small and give too high mechanical stress. limit difference to 100% vs. 80% motor speed between left and right. bug must be in navigation.c, since IMU information is correct (x-direction on mainboard = forward)
1.0
Navigation/Driving not correctly working - rover does not follow correct direction as entered in apm planner. Also, curve radii are too small and give too high mechanical stress. limit difference to 100% vs. 80% motor speed between left and right. bug must be in navigation.c, since IMU information is correct (x-direction on mainboard = forward)
priority
navigation driving not correctly working rover does not follow correct direction as entered in apm planner also curve radii are too small and give too high mechanical stress limit difference to vs motor speed between left and right bug must be in navigation c since imu information is correct x direction on mainboard forward
1
674,418
23,050,276,972
IssuesEvent
2022-07-24 14:21:09
FTBTeam/FTB-App
https://api.github.com/repos/FTBTeam/FTB-App
closed
[Bug]: locked down to java 17 and fail to start downloaded runtime
bug app installer os/linux priority/high
### What Operating System Linux (Arch) ### App Version 202205171618-14abefffbf-release ### UI Version _No response_ ### Log Files _No response_ ### Debug Code - ### Describe the bug installer is locked down specific to java 17 and the script fails to run the downloaded runtime ### Steps to reproduce download current linux "SFX" script (for real - why not just use curl or wget to load the archive but construct some obscure "rar-sfx style" hybrid to not have to serve a binary? pointless) try to run note that it downloads a runtime of java 17 because arch has already advanced to java 18 note further an error message that the script can't find the java binary of the just downloaded and extract runtime ### Expected behaviour just a simple script to scan for ANY version of java by just executing "java" command instead of search in a couple dozen paths and use it ### Screenshots _No response_ ### Additional information there's no need to fix down to java 17 for the installer - just by its default look and feel it's clearly a swing application without even bother looking into the source also: why to try to search thru couple of dozen of directories instead of rely on the user has java proper "installed" by the distributions package manager? you surely don't go as far to support gentoo or even LFS to even bother for manual scanning - so why not just rely on "java" be an available command thanks to package manager magic? if it's not installed just give a hint to install java runtime via package manager this also solves the requirement of need to download a specific runtime (which btw doesn't work cause for whatever reason I not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway) - and anything version specific can also be done within java - eliminating the need for checking specific version by debug output but - as ftb being ftb I know for years again I had to tinker around to get it even working - as the legacy launcher somehow fails to find the authlib thanks to some NOT SUPPOSED TO USED classloader hacking of the systemclassloader ... man, why noone back then ever read recommendations about how NOT to use sun internals? as some genius had that bright idea to pack a script and a binary in an incompatible way I had to painly do a hexdump - edit the version detection on a hex level - and rebuild the mess to even get it running why is FTB so damn bad at these most basics of java programming over and over again? also: why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available - heck, it's even the basic tools on even gentoo and LFS - and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed - so it can be taken as given same as gzip and tar oh - btw - why do gunzip and tar in two steps? you do know that tar also supports gzip decompression by either provide -z specific for gzip or just -a for auto detection based on filename and please - just for the sake - just provide a simple bashscript that just downloads the app pack and runs it with local java runtime - why even the need for "an installer"? linux != windows - there's no need for such crap - or just ship some form of archive that contains everything instead of rely to curl/wget anything afterwards ... oh, wait, right, you do have to cheap out on bandwidth ... then how about use git as mirror? there's so much wrong with this new "app" crap the same as back with the old .jar launcher - let alone the literally THOUSANDS of ERROR level messages when booting up any of the packs - it's not just infos or warnings - but ERRORS - which, again as the same as the past years, just noone bothers to pay attention why the F* I used ftb again - right, just to see that immersive petroleum is not up to 1.18 yet ... geez, was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows - instead of just keep everything contained within that one directory I downloaded th script to - as it should be done on linux and with java applications in general ### Information - [X] I have provided as much information as possible
1.0
[Bug]: locked down to java 17 and fail to start downloaded runtime - ### What Operating System Linux (Arch) ### App Version 202205171618-14abefffbf-release ### UI Version _No response_ ### Log Files _No response_ ### Debug Code - ### Describe the bug installer is locked down specific to java 17 and the script fails to run the downloaded runtime ### Steps to reproduce download current linux "SFX" script (for real - why not just use curl or wget to load the archive but construct some obscure "rar-sfx style" hybrid to not have to serve a binary? pointless) try to run note that it downloads a runtime of java 17 because arch has already advanced to java 18 note further an error message that the script can't find the java binary of the just downloaded and extract runtime ### Expected behaviour just a simple script to scan for ANY version of java by just executing "java" command instead of search in a couple dozen paths and use it ### Screenshots _No response_ ### Additional information there's no need to fix down to java 17 for the installer - just by its default look and feel it's clearly a swing application without even bother looking into the source also: why to try to search thru couple of dozen of directories instead of rely on the user has java proper "installed" by the distributions package manager? you surely don't go as far to support gentoo or even LFS to even bother for manual scanning - so why not just rely on "java" be an available command thanks to package manager magic? if it's not installed just give a hint to install java runtime via package manager this also solves the requirement of need to download a specific runtime (which btw doesn't work cause for whatever reason I not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway) - and anything version specific can also be done within java - eliminating the need for checking specific version by debug output but - as ftb being ftb I know for years again I had to tinker around to get it even working - as the legacy launcher somehow fails to find the authlib thanks to some NOT SUPPOSED TO USED classloader hacking of the systemclassloader ... man, why noone back then ever read recommendations about how NOT to use sun internals? as some genius had that bright idea to pack a script and a binary in an incompatible way I had to painly do a hexdump - edit the version detection on a hex level - and rebuild the mess to even get it running why is FTB so damn bad at these most basics of java programming over and over again? also: why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available - heck, it's even the basic tools on even gentoo and LFS - and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed - so it can be taken as given same as gzip and tar oh - btw - why do gunzip and tar in two steps? you do know that tar also supports gzip decompression by either provide -z specific for gzip or just -a for auto detection based on filename and please - just for the sake - just provide a simple bashscript that just downloads the app pack and runs it with local java runtime - why even the need for "an installer"? linux != windows - there's no need for such crap - or just ship some form of archive that contains everything instead of rely to curl/wget anything afterwards ... oh, wait, right, you do have to cheap out on bandwidth ... then how about use git as mirror? there's so much wrong with this new "app" crap the same as back with the old .jar launcher - let alone the literally THOUSANDS of ERROR level messages when booting up any of the packs - it's not just infos or warnings - but ERRORS - which, again as the same as the past years, just noone bothers to pay attention why the F* I used ftb again - right, just to see that immersive petroleum is not up to 1.18 yet ... geez, was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows - instead of just keep everything contained within that one directory I downloaded th script to - as it should be done on linux and with java applications in general ### Information - [X] I have provided as much information as possible
priority
locked down to java and fail to start downloaded runtime what operating system linux arch app version release ui version no response log files no response debug code describe the bug installer is locked down specific to java and the script fails to run the downloaded runtime steps to reproduce download current linux sfx script for real why not just use curl or wget to load the archive but construct some obscure rar sfx style hybrid to not have to serve a binary pointless try to run note that it downloads a runtime of java because arch has already advanced to java note further an error message that the script can t find the java binary of the just downloaded and extract runtime expected behaviour just a simple script to scan for any version of java by just executing java command instead of search in a couple dozen paths and use it screenshots no response additional information there s no need to fix down to java for the installer just by its default look and feel it s clearly a swing application without even bother looking into the source also why to try to search thru couple of dozen of directories instead of rely on the user has java proper installed by the distributions package manager you surely don t go as far to support gentoo or even lfs to even bother for manual scanning so why not just rely on java be an available command thanks to package manager magic if it s not installed just give a hint to install java runtime via package manager this also solves the requirement of need to download a specific runtime which btw doesn t work cause for whatever reason i not even bothered to analyze the script fails to find the binary of the downloaded and extracted runtime anyway and anything version specific can also be done within java eliminating the need for checking specific version by debug output but as ftb being ftb i know for years again i had to tinker around to get it even working as the legacy launcher somehow fails to find the authlib thanks to some not supposed to used classloader hacking of the systemclassloader man why noone back then ever read recommendations about how not to use sun internals as some genius had that bright idea to pack a script and a binary in an incompatible way i had to painly do a hexdump edit the version detection on a hex level and rebuild the mess to even get it running why is ftb so damn bad at these most basics of java programming over and over again also why to bother for this sophisticated manual search for a runtime but then rely on curl or wget to be available heck it s even the basic tools on even gentoo and lfs and even on arch although not directly mentioned in the install guide one comes rather fast to a point where one of them is needed so it can be taken as given same as gzip and tar oh btw why do gunzip and tar in two steps you do know that tar also supports gzip decompression by either provide z specific for gzip or just a for auto detection based on filename and please just for the sake just provide a simple bashscript that just downloads the app pack and runs it with local java runtime why even the need for an installer linux windows there s no need for such crap or just ship some form of archive that contains everything instead of rely to curl wget anything afterwards oh wait right you do have to cheap out on bandwidth then how about use git as mirror there s so much wrong with this new app crap the same as back with the old jar launcher let alone the literally thousands of error level messages when booting up any of the packs it s not just infos or warnings but errors which again as the same as the past years just noone bothers to pay attention why the f i used ftb again right just to see that immersive petroleum is not up to yet geez was a mess to clean everything up again cause it had to be blasted in several different directories same as on windows instead of just keep everything contained within that one directory i downloaded th script to as it should be done on linux and with java applications in general information i have provided as much information as possible
1
156,823
5,989,636,293
IssuesEvent
2017-06-02 09:09:17
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[New xCAT prototype]build ansible playbook to setup xcat3 service
priority:high sprint1 type:feature
Acceptance: 1. git clone xcat3 codes onto hosts, start and configure xcat3 service 2. configure apache and dhcp, tftp, mysql/mariaDB service
1.0
[New xCAT prototype]build ansible playbook to setup xcat3 service - Acceptance: 1. git clone xcat3 codes onto hosts, start and configure xcat3 service 2. configure apache and dhcp, tftp, mysql/mariaDB service
priority
build ansible playbook to setup service acceptance git clone codes onto hosts start and configure service configure apache and dhcp tftp mysql mariadb service
1
711,240
24,455,103,816
IssuesEvent
2022-10-07 05:41:54
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Bug]: InvalidPathException during the code coverage generation in Windows
Type/Bug Priority/High Team/CompilerFE Team/jBallerina Points/2
### Description ``` [2022-10-04 09:25:07,184] SEVERE {b7a.log.crash} - Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class java.nio.file.InvalidPathException: Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class ``` Found in HTTP PR build on windows. ``` at org.ballerinalang.test.runtime.util.CodeCoverageUtils.unzipCompiledSource(CodeCoverageUtils.java:116) at org.ballerinalang.test.runtime.entity.CoverageReport.addCompiledSources(CoverageReport.java:302) at org.ballerinalang.test.runtime.entity.CoverageReport.generateReport(CoverageReport.java:139) ``` ### Steps to Reproduce Build the HTTP package in windows Checkout to : https://github.com/chamil321/module-ballerina-http/tree/exit Cmd : ``` ./gradlew clean build -x http-native:test -x :http-ballerina-tests:startLdapServe -x :http-ballerina-tests:stopLdapServer -x :http-compiler-plugin-tests:test -x :http-ballerina-tests:commitTomlFiles -x :http-ballerina:commitTomlFiles ``` ### Affected Version(s) _No response_ ### OS, DB, other environment details and versions _No response_ ### Related area -> Runtime ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
1.0
[Bug]: InvalidPathException during the code coverage generation in Windows - ### Description ``` [2022-10-04 09:25:07,184] SEVERE {b7a.log.crash} - Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class java.nio.file.InvalidPathException: Illegal char <*> at index 182: D:\a\module-ballerina-http\module-ballerina-http\ballerina-tests\target\cache\tests_cache\coverage\bin\ballerina\http\2\$value$(ballerina$0047http$00582$0058$anonType$Method$$delete$**$_2 & readonly).class ``` Found in HTTP PR build on windows. ``` at org.ballerinalang.test.runtime.util.CodeCoverageUtils.unzipCompiledSource(CodeCoverageUtils.java:116) at org.ballerinalang.test.runtime.entity.CoverageReport.addCompiledSources(CoverageReport.java:302) at org.ballerinalang.test.runtime.entity.CoverageReport.generateReport(CoverageReport.java:139) ``` ### Steps to Reproduce Build the HTTP package in windows Checkout to : https://github.com/chamil321/module-ballerina-http/tree/exit Cmd : ``` ./gradlew clean build -x http-native:test -x :http-ballerina-tests:startLdapServe -x :http-ballerina-tests:stopLdapServer -x :http-compiler-plugin-tests:test -x :http-ballerina-tests:commitTomlFiles -x :http-ballerina:commitTomlFiles ``` ### Affected Version(s) _No response_ ### OS, DB, other environment details and versions _No response_ ### Related area -> Runtime ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
priority
invalidpathexception during the code coverage generation in windows description severe log crash illegal char at index d a module ballerina http module ballerina http ballerina tests target cache tests cache coverage bin ballerina http value ballerina anontype method delete readonly class java nio file invalidpathexception illegal char at index d a module ballerina http module ballerina http ballerina tests target cache tests cache coverage bin ballerina http value ballerina anontype method delete readonly class found in http pr build on windows at org ballerinalang test runtime util codecoverageutils unzipcompiledsource codecoverageutils java at org ballerinalang test runtime entity coveragereport addcompiledsources coveragereport java at org ballerinalang test runtime entity coveragereport generatereport coveragereport java steps to reproduce build the http package in windows checkout to cmd gradlew clean build x http native test x http ballerina tests startldapserve x http ballerina tests stopldapserver x http compiler plugin tests test x http ballerina tests committomlfiles x http ballerina committomlfiles affected version s no response os db other environment details and versions no response related area runtime related issue s optional no response suggested label s optional no response suggested assignee s optional no response
1
685,448
23,456,345,705
IssuesEvent
2022-08-16 09:12:30
nkhanh44/ic-survey-ios
https://api.github.com/repos/nkhanh44/ic-survey-ios
opened
[Chore] setup CD pineline
type : chore @0.4.0 priority: high
## Why - No manual execution of releases ## Who Benefits? - CD pipeline is run on Github Action - CD pipeline run to build a staging version and upload to Firebase Distribution
1.0
[Chore] setup CD pineline - ## Why - No manual execution of releases ## Who Benefits? - CD pipeline is run on Github Action - CD pipeline run to build a staging version and upload to Firebase Distribution
priority
setup cd pineline why no manual execution of releases who benefits cd pipeline is run on github action cd pipeline run to build a staging version and upload to firebase distribution
1
694,268
23,808,295,849
IssuesEvent
2022-09-04 11:33:33
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
PLO button issue
wontfix can't reproduce high priority easy request
Look into the PLO button issue reported below. > The Pot button in PLO is exhibiting weird behavior – I tried to bet pot of 1500 but it only allowed 900. I can’t reproduce yet as that was the last hand of the PLO round. I’ll try to get you more details. I was able to use the slider – didn’t try the Max button. > If you give me a hand number (which you might be able to still get, by scrolling back), that'll help. > I am semi-sureish that it’s 284669. This was reported to me via email. I didn't make an issue out of it at the time and then I forgot. I'm starting with this as `high priority` and `easy`, because I have a hand number, which will also give me the table and event. I can poke around and see if there were any error messages reported from the server to the client. If there were, then it should be trivial to reproduce and fix. If there are no error messages, I'll have to ask some more questions and it may no longer be easy.
1.0
PLO button issue - Look into the PLO button issue reported below. > The Pot button in PLO is exhibiting weird behavior – I tried to bet pot of 1500 but it only allowed 900. I can’t reproduce yet as that was the last hand of the PLO round. I’ll try to get you more details. I was able to use the slider – didn’t try the Max button. > If you give me a hand number (which you might be able to still get, by scrolling back), that'll help. > I am semi-sureish that it’s 284669. This was reported to me via email. I didn't make an issue out of it at the time and then I forgot. I'm starting with this as `high priority` and `easy`, because I have a hand number, which will also give me the table and event. I can poke around and see if there were any error messages reported from the server to the client. If there were, then it should be trivial to reproduce and fix. If there are no error messages, I'll have to ask some more questions and it may no longer be easy.
priority
plo button issue look into the plo button issue reported below the pot button in plo is exhibiting weird behavior – i tried to bet pot of but it only allowed i can’t reproduce yet as that was the last hand of the plo round i’ll try to get you more details i was able to use the slider – didn’t try the max button if you give me a hand number which you might be able to still get by scrolling back that ll help i am semi sureish that it’s this was reported to me via email i didn t make an issue out of it at the time and then i forgot i m starting with this as high priority and easy because i have a hand number which will also give me the table and event i can poke around and see if there were any error messages reported from the server to the client if there were then it should be trivial to reproduce and fix if there are no error messages i ll have to ask some more questions and it may no longer be easy
1
664,406
22,269,345,555
IssuesEvent
2022-06-10 10:38:50
opencrvs/opencrvs-core
https://api.github.com/repos/opencrvs/opencrvs-core
closed
If user search a birth/death application and open it, Date of death,Place of death and Informant do not show
👹Bug Priority: high
**Bug Description:** If a user searches a birth/death application and then opens it, the Following information shows blank: - Date of death - Place of death - Informant **Steps:** 1. Log in as a field agent/Registration Clerk/Register 2. Click on any application which has the Date of death, Place of death, and Informant from any tab 3. Download the application(if logged in with Registration clerk/Register) 4. Observe that the Date of death, Place of death, and Informant are showing 5. Copy the tracking id or name or phone number 6. Search the application with the tracking id or name or phone number 7. Click on the application **Actual Result:** - Date of death, Place of death, and Informant do not show any data **Expected Result:** - Date of death, Place of death, and Informant should show if the user opens the application after searching **Screen record:** https://www.screencast.com/t/3T0jmT70bl **Tested on:** https://login.farajaland-qa.opencrvs.org/ **Username & Password Used:** - Username: felix.katongo - password: test **Desktop:** - OS: Windows 10 - Browser: Chrome
1.0
If user search a birth/death application and open it, Date of death,Place of death and Informant do not show - **Bug Description:** If a user searches a birth/death application and then opens it, the Following information shows blank: - Date of death - Place of death - Informant **Steps:** 1. Log in as a field agent/Registration Clerk/Register 2. Click on any application which has the Date of death, Place of death, and Informant from any tab 3. Download the application(if logged in with Registration clerk/Register) 4. Observe that the Date of death, Place of death, and Informant are showing 5. Copy the tracking id or name or phone number 6. Search the application with the tracking id or name or phone number 7. Click on the application **Actual Result:** - Date of death, Place of death, and Informant do not show any data **Expected Result:** - Date of death, Place of death, and Informant should show if the user opens the application after searching **Screen record:** https://www.screencast.com/t/3T0jmT70bl **Tested on:** https://login.farajaland-qa.opencrvs.org/ **Username & Password Used:** - Username: felix.katongo - password: test **Desktop:** - OS: Windows 10 - Browser: Chrome
priority
if user search a birth death application and open it date of death place of death and informant do not show bug description if a user searches a birth death application and then opens it the following information shows blank date of death place of death informant steps log in as a field agent registration clerk register click on any application which has the date of death place of death and informant from any tab download the application if logged in with registration clerk register observe that the date of death place of death and informant are showing copy the tracking id or name or phone number search the application with the tracking id or name or phone number click on the application actual result date of death place of death and informant do not show any data expected result date of death place of death and informant should show if the user opens the application after searching screen record tested on username password used username felix katongo password test desktop os windows browser chrome
1
621,757
19,595,953,455
IssuesEvent
2022-01-05 17:52:13
vaticle/typedb
https://api.github.com/repos/vaticle/typedb
closed
Investigate OOM caused using TypeDB Loader
type: bug priority: high
## Description Bayer has reported an occasional OOM when doing heavy data loading. ## Environment TypeDB 2.2.0 on a cloud-based linux VM: ``` Host: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz, 4 cores, 15G, Amazon Linux release 2 (Karoo) ``` ## Reproducible Steps - see private dropbox folder for dataset and reproducible steps ## Expected Output - data loading runs indefinitely without fail ## Actual Output One failure occurs with the following error, unsure if others occur: ``` # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 58720256 bytes for committing reserved memory. # Possible reasons: # The system is out of physical RAM or swap space # The process is running with CompressedOops enabled, and the Java Heap may be blocking the growth of the native heap # Possible solutions: # Reduce memory load on the system # Increase physical memory or swap space # Check if swap backing store is full # Decrease Java heap size (-Xmx/-Xms) # Decrease number of Java threads # Decrease Java thread stack sizes (-Xss) # Set larger code cache with -XX:ReservedCodeCacheSize= # JVM is running with Zero Based Compressed Oops mode in which the Java heap is # placed in the first 32GB address space. The Java Heap base address is the # maximum limit for the native heap growth. Please use -XX:HeapBaseMinAddress # to set the Java Heap base and to place the Java Heap above 32GB virtual address. # This output file may be truncated or incomplete. # # Out of Memory Error (os_linux.cpp:2977), pid=17756, tid=17768 # # JRE version: OpenJDK Runtime Environment 20.9 (15.0.2+7) (build 15.0.2+7) # Java VM: OpenJDK 64-Bit Server VM 20.9 (15.0.2+7, mixed mode, sharing, tiered, compressed oops, g1 gc, linux-amd64) # No core dump will be written. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again # ``` ## Additional information
1.0
Investigate OOM caused using TypeDB Loader - ## Description Bayer has reported an occasional OOM when doing heavy data loading. ## Environment TypeDB 2.2.0 on a cloud-based linux VM: ``` Host: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz, 4 cores, 15G, Amazon Linux release 2 (Karoo) ``` ## Reproducible Steps - see private dropbox folder for dataset and reproducible steps ## Expected Output - data loading runs indefinitely without fail ## Actual Output One failure occurs with the following error, unsure if others occur: ``` # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 58720256 bytes for committing reserved memory. # Possible reasons: # The system is out of physical RAM or swap space # The process is running with CompressedOops enabled, and the Java Heap may be blocking the growth of the native heap # Possible solutions: # Reduce memory load on the system # Increase physical memory or swap space # Check if swap backing store is full # Decrease Java heap size (-Xmx/-Xms) # Decrease number of Java threads # Decrease Java thread stack sizes (-Xss) # Set larger code cache with -XX:ReservedCodeCacheSize= # JVM is running with Zero Based Compressed Oops mode in which the Java heap is # placed in the first 32GB address space. The Java Heap base address is the # maximum limit for the native heap growth. Please use -XX:HeapBaseMinAddress # to set the Java Heap base and to place the Java Heap above 32GB virtual address. # This output file may be truncated or incomplete. # # Out of Memory Error (os_linux.cpp:2977), pid=17756, tid=17768 # # JRE version: OpenJDK Runtime Environment 20.9 (15.0.2+7) (build 15.0.2+7) # Java VM: OpenJDK 64-Bit Server VM 20.9 (15.0.2+7, mixed mode, sharing, tiered, compressed oops, g1 gc, linux-amd64) # No core dump will be written. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again # ``` ## Additional information
priority
investigate oom caused using typedb loader description bayer has reported an occasional oom when doing heavy data loading environment typedb on a cloud based linux vm host intel r xeon r platinum cpu cores amazon linux release karoo reproducible steps see private dropbox folder for dataset and reproducible steps expected output data loading runs indefinitely without fail actual output one failure occurs with the following error unsure if others occur there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory possible reasons the system is out of physical ram or swap space the process is running with compressedoops enabled and the java heap may be blocking the growth of the native heap possible solutions reduce memory load on the system increase physical memory or swap space check if swap backing store is full decrease java heap size xmx xms decrease number of java threads decrease java thread stack sizes xss set larger code cache with xx reservedcodecachesize jvm is running with zero based compressed oops mode in which the java heap is placed in the first address space the java heap base address is the maximum limit for the native heap growth please use xx heapbaseminaddress to set the java heap base and to place the java heap above virtual address this output file may be truncated or incomplete out of memory error os linux cpp pid tid jre version openjdk runtime environment build java vm openjdk bit server vm mixed mode sharing tiered compressed oops gc linux no core dump will be written core dumps have been disabled to enable core dumping try ulimit c unlimited before starting java again additional information
1
22,067
2,644,998,602
IssuesEvent
2015-03-12 20:04:06
starteam/starcellbio_html
https://api.github.com/repos/starteam/starcellbio_html
closed
SUTD Exercise #2 - Flow Cytometry: clicking and dragging for drawing gates not working properly
bug High Priority
The gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy (for both the single and the bisector gate tools)
1.0
SUTD Exercise #2 - Flow Cytometry: clicking and dragging for drawing gates not working properly - The gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy (for both the single and the bisector gate tools)
priority
sutd exercise flow cytometry clicking and dragging for drawing gates not working properly the gates appear in a location that is slightly off from where you actually click and drag to demarcate the area they occupy for both the single and the bisector gate tools
1
796,839
28,128,869,293
IssuesEvent
2023-03-31 20:24:59
mila-iqia/mila-docs
https://api.github.com/repos/mila-iqia/mila-docs
reopened
Search box seems broken.
bug priority:high
The search box (implemented within the browser space) seems to spin endlessly. Doesn't provide results even after a long while. Is it broken ? This is on firefox / linux.
1.0
Search box seems broken. - The search box (implemented within the browser space) seems to spin endlessly. Doesn't provide results even after a long while. Is it broken ? This is on firefox / linux.
priority
search box seems broken the search box implemented within the browser space seems to spin endlessly doesn t provide results even after a long while is it broken this is on firefox linux
1
305,576
9,371,448,465
IssuesEvent
2019-04-03 15:24:56
CS2103-AY1819S2-T12-4/main
https://api.github.com/repos/CS2103-AY1819S2-T12-4/main
closed
Update of UG/DG to fit v1.3
priority.High
**For DG** _(under Implementation)_: - [x] Decrypt - [x] Encrypt - [x] Filter - [x] Find - [x] Label - [x] Merge _Might or must need more details/ illustrations:_ - [x] Clear - [x] Deadline - [x] Delete - [x] Exit - [x] Help - [x] History - [x] List - [x] Select - [x] Sort - [x] Tag All hands on deck boys
1.0
Update of UG/DG to fit v1.3 - **For DG** _(under Implementation)_: - [x] Decrypt - [x] Encrypt - [x] Filter - [x] Find - [x] Label - [x] Merge _Might or must need more details/ illustrations:_ - [x] Clear - [x] Deadline - [x] Delete - [x] Exit - [x] Help - [x] History - [x] List - [x] Select - [x] Sort - [x] Tag All hands on deck boys
priority
update of ug dg to fit for dg under implementation decrypt encrypt filter find label merge might or must need more details illustrations clear deadline delete exit help history list select sort tag all hands on deck boys
1
554,798
16,439,102,889
IssuesEvent
2021-05-20 12:36:25
TEIC/TEI
https://api.github.com/repos/TEIC/TEI
closed
allow <schemaSpec> in <front> and <back>
Priority: High TEI: Schema Type: Bug
In ticket #1921 we are removing `<schemaSpec>` from a variety of silly places (like `<figure>`, `<license>`, and `<scriptNote>`), and we now explicitly recommend “It should be a child of front, body, back, encodingDesc, div, div1, …, or div7.”. HOWEVER, `<schemaSpec>` is not allowed as a child of `<front>` or `<back>`. This, I think is a corrigible schema error that should just be fixed. Alternatively, of course, we could remove the explicit recommendation that it go in `<front>` or `<back>`.
1.0
allow <schemaSpec> in <front> and <back> - In ticket #1921 we are removing `<schemaSpec>` from a variety of silly places (like `<figure>`, `<license>`, and `<scriptNote>`), and we now explicitly recommend “It should be a child of front, body, back, encodingDesc, div, div1, …, or div7.”. HOWEVER, `<schemaSpec>` is not allowed as a child of `<front>` or `<back>`. This, I think is a corrigible schema error that should just be fixed. Alternatively, of course, we could remove the explicit recommendation that it go in `<front>` or `<back>`.
priority
allow in and in ticket we are removing from a variety of silly places like and and we now explicitly recommend “it should be a child of front body back encodingdesc div … or ” however is not allowed as a child of or this i think is a corrigible schema error that should just be fixed alternatively of course we could remove the explicit recommendation that it go in or
1
499,850
14,480,799,167
IssuesEvent
2020-12-10 11:41:35
mpagels/collectoPod
https://api.github.com/repos/mpagels/collectoPod
opened
Podcast sort function
Priority: High Type: New Feature
## Value statement - a requirement, not a solution As a **user** I need **a sort function** so that **I can sort the order of the podcasts** ## Description (Text, Scribble, Wireframe, Design) ## Acceptance criteria - [ ] I see a list of the podcasts - [ ] when I move the first podcast under the second podcast, the second podcast moves up and the first podcast takes the second place - [ ] I hit the save button to save the new list to localStorage ## Tasks ## Size (medium)
1.0
Podcast sort function - ## Value statement - a requirement, not a solution As a **user** I need **a sort function** so that **I can sort the order of the podcasts** ## Description (Text, Scribble, Wireframe, Design) ## Acceptance criteria - [ ] I see a list of the podcasts - [ ] when I move the first podcast under the second podcast, the second podcast moves up and the first podcast takes the second place - [ ] I hit the save button to save the new list to localStorage ## Tasks ## Size (medium)
priority
podcast sort function value statement a requirement not a solution as a user i need a sort function so that i can sort the order of the podcasts description text scribble wireframe design acceptance criteria i see a list of the podcasts when i move the first podcast under the second podcast the second podcast moves up and the first podcast takes the second place i hit the save button to save the new list to localstorage tasks size medium
1
171,736
6,493,927,926
IssuesEvent
2017-08-21 19:26:23
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Preview tools panel and tooltips does not use Open Sans for the font
bug Priority: High
Using the Website_editorial bp, click on the wrench at the top right of studio to open the preview tools panel. Notice the font in the dialog is not Open Sans: ![screen shot 2017-08-18 at 10 59 15 am](https://user-images.githubusercontent.com/25483966/29464586-507545c4-8404-11e7-84b5-91b2b3aa892d.png) Click on **Page Components**, notice that the font used in the panel is also not Open Sans: ![screen shot 2017-08-18 at 10 27 50 am](https://user-images.githubusercontent.com/25483966/29464530-24a3f166-8404-11e7-8c63-8edb34722b1d.png) Click on **Browse Features**, after selecting the component you want to add, notice the notification on the top right of the screen does not use Open Sans for the font: ![screen shot 2017-08-18 at 11 01 29 am](https://user-images.githubusercontent.com/25483966/29464716-b4f91f8e-8404-11e7-963b-254bb3aa9329.png) In the **Sidebar**, hover on any of the folders/items, notice the font used in the tooltip is not Open Sans ![screen shot 2017-08-18 at 11 32 33 am](https://user-images.githubusercontent.com/25483966/29466082-92025b94-8409-11e7-81dc-153ffd1d4553.png) Clicn on **Approve & Publish** in the context nav, then click on **Items go live on a specific date & time**, notice that the font used by the datetime picker is not Open Sans ![screen shot 2017-08-18 at 11 38 45 am](https://user-images.githubusercontent.com/25483966/29466229-107f20a6-840a-11e7-9307-b7d88f8b1df1.png)
1.0
[studio-ui] Preview tools panel and tooltips does not use Open Sans for the font - Using the Website_editorial bp, click on the wrench at the top right of studio to open the preview tools panel. Notice the font in the dialog is not Open Sans: ![screen shot 2017-08-18 at 10 59 15 am](https://user-images.githubusercontent.com/25483966/29464586-507545c4-8404-11e7-84b5-91b2b3aa892d.png) Click on **Page Components**, notice that the font used in the panel is also not Open Sans: ![screen shot 2017-08-18 at 10 27 50 am](https://user-images.githubusercontent.com/25483966/29464530-24a3f166-8404-11e7-8c63-8edb34722b1d.png) Click on **Browse Features**, after selecting the component you want to add, notice the notification on the top right of the screen does not use Open Sans for the font: ![screen shot 2017-08-18 at 11 01 29 am](https://user-images.githubusercontent.com/25483966/29464716-b4f91f8e-8404-11e7-963b-254bb3aa9329.png) In the **Sidebar**, hover on any of the folders/items, notice the font used in the tooltip is not Open Sans ![screen shot 2017-08-18 at 11 32 33 am](https://user-images.githubusercontent.com/25483966/29466082-92025b94-8409-11e7-81dc-153ffd1d4553.png) Clicn on **Approve & Publish** in the context nav, then click on **Items go live on a specific date & time**, notice that the font used by the datetime picker is not Open Sans ![screen shot 2017-08-18 at 11 38 45 am](https://user-images.githubusercontent.com/25483966/29466229-107f20a6-840a-11e7-9307-b7d88f8b1df1.png)
priority
preview tools panel and tooltips does not use open sans for the font using the website editorial bp click on the wrench at the top right of studio to open the preview tools panel notice the font in the dialog is not open sans click on page components notice that the font used in the panel is also not open sans click on browse features after selecting the component you want to add notice the notification on the top right of the screen does not use open sans for the font in the sidebar hover on any of the folders items notice the font used in the tooltip is not open sans clicn on approve publish in the context nav then click on items go live on a specific date time notice that the font used by the datetime picker is not open sans
1
802,385
28,959,738,227
IssuesEvent
2023-05-10 00:43:36
steedos/steedos-platform
https://api.github.com/repos/steedos/steedos-platform
closed
[Bug]: 远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白
bug done priority: High
### Description <img width="1440" alt="image" src="https://user-images.githubusercontent.com/26241897/236722162-99928ac5-0b0c-40cb-87d9-9928c02c590d.png"> 控制台报错信息: steedos_1 | memory, and will not scale past a single process. steedos_1 | service ~packages-standard-objects started: 1.479s steedos_1 | service ~packages-@steedos/standard-space started: 307.312ms steedos_1 | service ~packages-@steedos/standard-cms started: 105.359ms steedos_1 | service ~packages-@steedos/standard-object-database started: 512.988ms steedos_1 | service ~packages-@steedos/standard-process-approval started: 338.997ms steedos_1 | service ~packages-@steedos/standard-collaboration started: 179.123ms steedos_1 | service ~packages-@steedos/standard-ui started: 210.011ms steedos_1 | service ~packages-@steedos/standard-permission started: 190.842ms steedos_1 | service ~packages-@steedos/workflow started: 45.182ms steedos_1 | service ~packages-@steedos/accounts started steedos_1 | service ~packages-@steedos/plugin-company started: 221.344ms steedos_1 | service ~packages-@steedos/metadata-api started: 54.139ms steedos_1 | steedos_1 | Project is running at https://5000-tomato-finch-em8hx9yz.ws.vscode.steedos.cn steedos_1 | steedos_1 | service ~packages-@steedos/data-import started: 61.041ms steedos_1 | service ~packages-@steedos/unpkg started: 1.302s steedos_1 | service ~packages-@steedos/webapp-accounts started: 1.276s steedos_1 | service ~packages-@steedos/service-analytics started: 1.236s steedos_1 | service ~packages-@steedos/service-ui started: 1.184s steedos_1 | service ~packages-@steedos/webapp-public started: 492.923ms steedos_1 | service ~packages-@steedos/service-files started: 253.453ms steedos_1 | service ~packages-@steedos-labs/flows started: 1.059s steedos_1 | service ~packages-@steedos/service-i18n started: 274.951ms steedos_1 | service ~packages-@steedos-labs/workflow started: 942.123ms steedos_1 | service ~packages-@steedos/workflow_time_trigger started: 1.538s steedos_1 | service ~packages-@steedos-labs/oa started: 1.278s steedos_1 | service ~packages-@steedos/service-package-registry started: 637.725ms steedos_1 | service ~packages-@steedos/service-charts started: 1.162s steedos_1 | service ~packages-@steedos/steedos-plugin-schema-builder started: 2.095s steedos_1 | service ~packages-@steedos/service-pages started: 1.675s steedos_1 | service ~packages-@steedos-labs/meeting started: 2.234s steedos_1 | service ~packages-@steedos/service-plugin-amis started: 265.047ms steedos_1 | service ~packages-@steedos-labs/okr started: 2.562s steedos_1 | service ~packages-@steedos-labs/budget started: 3.193s steedos_1 | service ~packages-@steedos-labs/order started: 3.474s steedos_1 | service ~packages-@steedos-labs/purchase started: 3.537s steedos_1 | service ~packages-@steedos/service-package-tool started: 3.579s steedos_1 | service ~packages-@steedos-labs/project started: 4.404s steedos_1 | service ~packages-@steedos-labs/master started: 4.822s steedos_1 | service ~packages-@steedos-labs/cost started: 5.548s steedos_1 | service ~packages-@steedos-labs/project-ee started: 699.506ms steedos_1 | service ~packages-@steedos-labs/contract started: 5.927s steedos_1 | service ~packages-@steedos-labs/crm started: 5.646s steedos_1 | service ~packages-@steedos-labs/demo-ee started: 497.628ms steedos_1 | service ~packages-@steedos-labs/contract-ee started: 1.031s steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos/plugin-qywx started: 1.343s steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos/plugin-dingtalk started: 947.906ms steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos-labs/finance started: 3.395s steedos_1 | service ~packages-@steedos-labs/asset started: 8.400s steedos_1 | service ~packages-@steedos-labs/cost-ee started: 2.719s steedos_1 | service ~packages-@steedos-labs/finance-ee started: 42.196ms steedos_1 | install package @steedos-labs/project-ee successful. steedos_1 | service ~packages-@steedos/service-workflow started: 8.870s steedos_1 | service ~packages-@steedos-labs/project-ee stopped steedos_1 | service ~packages-@steedos-labs/project-ee started: 217.276ms steedos_1 | install package @steedos-labs/demo-ee successful. steedos_1 | service ~packages-@steedos-labs/demo-ee stopped steedos_1 | service ~packages-@steedos-labs/demo-ee started: 54.886ms steedos_1 | install package @steedos-labs/flows successful. steedos_1 | service ~packages-@steedos-labs/flows stopped steedos_1 | service ~packages-@steedos-labs/flows started: 59.969ms steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90) steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1439:47) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 3) steedos_1 | (Use `node --trace-warnings ...` to show where the warning was created) steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90) steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1814:47) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.processUnreadableField (/app/node_modules/@steedos/objectql/lib/types/object.js:2240:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1072:24) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.find (/app/node_modules/@steedos/objectql/lib/types/object.js:1182:24) steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:158:64 steedos_1 | at Generator.next (<anonymous>) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at getHiddenTabNames (/app/node_modules/@steedos/service-metadata-apps/lib/actionsHandler.js:135:20) steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:91:34 steedos_1 | at Generator.next (<anonymous>) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 5) ### Steps To Reproduce 重现步骤 1. 重开template项目,版本升级到2.5.0-beta.20; 2. 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息 ### Version 版本 2.5.0-beta.20
1.0
[Bug]: 远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白 - ### Description <img width="1440" alt="image" src="https://user-images.githubusercontent.com/26241897/236722162-99928ac5-0b0c-40cb-87d9-9928c02c590d.png"> 控制台报错信息: steedos_1 | memory, and will not scale past a single process. steedos_1 | service ~packages-standard-objects started: 1.479s steedos_1 | service ~packages-@steedos/standard-space started: 307.312ms steedos_1 | service ~packages-@steedos/standard-cms started: 105.359ms steedos_1 | service ~packages-@steedos/standard-object-database started: 512.988ms steedos_1 | service ~packages-@steedos/standard-process-approval started: 338.997ms steedos_1 | service ~packages-@steedos/standard-collaboration started: 179.123ms steedos_1 | service ~packages-@steedos/standard-ui started: 210.011ms steedos_1 | service ~packages-@steedos/standard-permission started: 190.842ms steedos_1 | service ~packages-@steedos/workflow started: 45.182ms steedos_1 | service ~packages-@steedos/accounts started steedos_1 | service ~packages-@steedos/plugin-company started: 221.344ms steedos_1 | service ~packages-@steedos/metadata-api started: 54.139ms steedos_1 | steedos_1 | Project is running at https://5000-tomato-finch-em8hx9yz.ws.vscode.steedos.cn steedos_1 | steedos_1 | service ~packages-@steedos/data-import started: 61.041ms steedos_1 | service ~packages-@steedos/unpkg started: 1.302s steedos_1 | service ~packages-@steedos/webapp-accounts started: 1.276s steedos_1 | service ~packages-@steedos/service-analytics started: 1.236s steedos_1 | service ~packages-@steedos/service-ui started: 1.184s steedos_1 | service ~packages-@steedos/webapp-public started: 492.923ms steedos_1 | service ~packages-@steedos/service-files started: 253.453ms steedos_1 | service ~packages-@steedos-labs/flows started: 1.059s steedos_1 | service ~packages-@steedos/service-i18n started: 274.951ms steedos_1 | service ~packages-@steedos-labs/workflow started: 942.123ms steedos_1 | service ~packages-@steedos/workflow_time_trigger started: 1.538s steedos_1 | service ~packages-@steedos-labs/oa started: 1.278s steedos_1 | service ~packages-@steedos/service-package-registry started: 637.725ms steedos_1 | service ~packages-@steedos/service-charts started: 1.162s steedos_1 | service ~packages-@steedos/steedos-plugin-schema-builder started: 2.095s steedos_1 | service ~packages-@steedos/service-pages started: 1.675s steedos_1 | service ~packages-@steedos-labs/meeting started: 2.234s steedos_1 | service ~packages-@steedos/service-plugin-amis started: 265.047ms steedos_1 | service ~packages-@steedos-labs/okr started: 2.562s steedos_1 | service ~packages-@steedos-labs/budget started: 3.193s steedos_1 | service ~packages-@steedos-labs/order started: 3.474s steedos_1 | service ~packages-@steedos-labs/purchase started: 3.537s steedos_1 | service ~packages-@steedos/service-package-tool started: 3.579s steedos_1 | service ~packages-@steedos-labs/project started: 4.404s steedos_1 | service ~packages-@steedos-labs/master started: 4.822s steedos_1 | service ~packages-@steedos-labs/cost started: 5.548s steedos_1 | service ~packages-@steedos-labs/project-ee started: 699.506ms steedos_1 | service ~packages-@steedos-labs/contract started: 5.927s steedos_1 | service ~packages-@steedos-labs/crm started: 5.646s steedos_1 | service ~packages-@steedos-labs/demo-ee started: 497.628ms steedos_1 | service ~packages-@steedos-labs/contract-ee started: 1.031s steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos/plugin-qywx started: 1.343s steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos/plugin-dingtalk started: 947.906ms steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | Error: Invalid Session steedos_1 | at AccountsServer.<anonymous> (/app/node_modules/@steedos/accounts/src/server/accounts-server.ts:588:13) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | service ~packages-@steedos-labs/finance started: 3.395s steedos_1 | service ~packages-@steedos-labs/asset started: 8.400s steedos_1 | service ~packages-@steedos-labs/cost-ee started: 2.719s steedos_1 | service ~packages-@steedos-labs/finance-ee started: 42.196ms steedos_1 | install package @steedos-labs/project-ee successful. steedos_1 | service ~packages-@steedos/service-workflow started: 8.870s steedos_1 | service ~packages-@steedos-labs/project-ee stopped steedos_1 | service ~packages-@steedos-labs/project-ee started: 217.276ms steedos_1 | install package @steedos-labs/demo-ee successful. steedos_1 | service ~packages-@steedos-labs/demo-ee stopped steedos_1 | service ~packages-@steedos-labs/demo-ee started: 54.886ms steedos_1 | install package @steedos-labs/flows successful. steedos_1 | service ~packages-@steedos-labs/flows stopped steedos_1 | service ~packages-@steedos-labs/flows started: 59.969ms steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90) steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1439:47) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 3) steedos_1 | (Use `node --trace-warnings ...` to show where the warning was created) steedos_1 | This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: steedos_1 | TypeError: Cannot read property '64586162353161002f6c6b2e' of undefined steedos_1 | at SteedosDataSourceType.getObjectSpaceRolesPermission (/app/node_modules/@steedos/objectql/src/types/datasource.ts:382:90) steedos_1 | at SteedosObjectType.getObjectRolesPermission (/app/node_modules/@steedos/objectql/src/types/object.ts:869:47) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:891:42) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.getUserObjectPermission (/app/node_modules/@steedos/objectql/lib/types/object.js:949:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1814:47) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.processUnreadableField (/app/node_modules/@steedos/objectql/lib/types/object.js:2240:24) steedos_1 | at SteedosObjectType.<anonymous> (/app/node_modules/@steedos/objectql/src/types/object.ts:1072:24) steedos_1 | at step (/app/node_modules/tslib/tslib.js:193:27) steedos_1 | at Object.next (/app/node_modules/tslib/tslib.js:174:57) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at SteedosObjectType.find (/app/node_modules/@steedos/objectql/lib/types/object.js:1182:24) steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:158:64 steedos_1 | at Generator.next (<anonymous>) steedos_1 | at /app/node_modules/tslib/tslib.js:167:75 steedos_1 | at new Promise (<anonymous>) steedos_1 | at Object.__awaiter (/app/node_modules/tslib/tslib.js:163:16) steedos_1 | at getHiddenTabNames (/app/node_modules/@steedos/service-metadata-apps/lib/actionsHandler.js:135:20) steedos_1 | at /app/node_modules/@steedos/service-metadata-apps/src/actionsHandler.ts:91:34 steedos_1 | at Generator.next (<anonymous>) steedos_1 | at fulfilled (/app/node_modules/tslib/tslib.js:164:62) steedos_1 | at /app/node_modules/meteor-promise/fiber_pool.js:43:39 steedos_1 | (node:47) PromiseRejectionHandledWarning: Promise rejection was handled asynchronously (rejection id: 5) ### Steps To Reproduce 重现步骤 1. 重开template项目,版本升级到2.5.0-beta.20; 2. 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息 ### Version 版本 2.5.0-beta.20
priority
远程新环境注册工作区后进入系统,顶部弹出红色警告框,系统页面空白 description img width alt image src 控制台报错信息: steedos memory and will not scale past a single process steedos service packages standard objects started steedos service packages steedos standard space started steedos service packages steedos standard cms started steedos service packages steedos standard object database started steedos service packages steedos standard process approval started steedos service packages steedos standard collaboration started steedos service packages steedos standard ui started steedos service packages steedos standard permission started steedos service packages steedos workflow started steedos service packages steedos accounts started steedos service packages steedos plugin company started steedos service packages steedos metadata api started steedos steedos project is running at steedos steedos service packages steedos data import started steedos service packages steedos unpkg started steedos service packages steedos webapp accounts started steedos service packages steedos service analytics started steedos service packages steedos service ui started steedos service packages steedos webapp public started steedos service packages steedos service files started steedos service packages steedos labs flows started steedos service packages steedos service started steedos service packages steedos labs workflow started steedos service packages steedos workflow time trigger started steedos service packages steedos labs oa started steedos service packages steedos service package registry started steedos service packages steedos service charts started steedos service packages steedos steedos plugin schema builder started steedos service packages steedos service pages started steedos service packages steedos labs meeting started steedos service packages steedos service plugin amis started steedos service packages steedos labs okr started steedos service packages steedos labs budget started steedos service packages steedos labs order started steedos service packages steedos labs purchase started steedos service packages steedos service package tool started steedos service packages steedos labs project started steedos service packages steedos labs master started steedos service packages steedos labs cost started steedos service packages steedos labs project ee started steedos service packages steedos labs contract started steedos service packages steedos labs crm started steedos service packages steedos labs demo ee started steedos service packages steedos labs contract ee started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos plugin qywx started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos plugin dingtalk started steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos error invalid session steedos at accountsserver app node modules steedos accounts src server accounts server ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos service packages steedos labs finance started steedos service packages steedos labs asset started steedos service packages steedos labs cost ee started steedos service packages steedos labs finance ee started steedos install package steedos labs project ee successful steedos service packages steedos service workflow started steedos service packages steedos labs project ee stopped steedos service packages steedos labs project ee started steedos install package steedos labs demo ee successful steedos service packages steedos labs demo ee stopped steedos service packages steedos labs demo ee started steedos install package steedos labs flows successful steedos service packages steedos labs flows stopped steedos service packages steedos labs flows started steedos this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch the promise rejected with the reason steedos typeerror cannot read property of undefined steedos at steedosdatasourcetype getobjectspacerolespermission app node modules steedos objectql src types datasource ts steedos at steedosobjecttype getobjectrolespermission app node modules steedos objectql src types object ts steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype getuserobjectpermission app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos node promiserejectionhandledwarning promise rejection was handled asynchronously rejection id steedos use node trace warnings to show where the warning was created steedos this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch the promise rejected with the reason steedos typeerror cannot read property of undefined steedos at steedosdatasourcetype getobjectspacerolespermission app node modules steedos objectql src types datasource ts steedos at steedosobjecttype getobjectrolespermission app node modules steedos objectql src types object ts steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype getuserobjectpermission app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype processunreadablefield app node modules steedos objectql lib types object js steedos at steedosobjecttype app node modules steedos objectql src types object ts steedos at step app node modules tslib tslib js steedos at object next app node modules tslib tslib js steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at steedosobjecttype find app node modules steedos objectql lib types object js steedos at app node modules steedos service metadata apps src actionshandler ts steedos at generator next steedos at app node modules tslib tslib js steedos at new promise steedos at object awaiter app node modules tslib tslib js steedos at gethiddentabnames app node modules steedos service metadata apps lib actionshandler js steedos at app node modules steedos service metadata apps src actionshandler ts steedos at generator next steedos at fulfilled app node modules tslib tslib js steedos at app node modules meteor promise fiber pool js steedos node promiserejectionhandledwarning promise rejection was handled asynchronously rejection id steps to reproduce 重现步骤 重开template项目, beta ; 服务启动后注册工作区后,跳转到登陆后系统页面,可以看到截图报错信息 version 版本 beta
1
330,153
10,035,478,451
IssuesEvent
2019-07-18 08:26:59
zdnscloud/singlecloud
https://api.github.com/repos/zdnscloud/singlecloud
closed
error "does not generate config" returns when creating ceph storagecluster
bug priority: High
workaround: delete and redeploy ceph will be OK. ![noconfig](https://user-images.githubusercontent.com/47511655/61193768-09e6e700-a6f0-11e9-8220-c8d1e1bc7b8a.png)
1.0
error "does not generate config" returns when creating ceph storagecluster - workaround: delete and redeploy ceph will be OK. ![noconfig](https://user-images.githubusercontent.com/47511655/61193768-09e6e700-a6f0-11e9-8220-c8d1e1bc7b8a.png)
priority
error does not generate config returns when creating ceph storagecluster workaround delete and redeploy ceph will be ok
1
474,916
13,684,409,842
IssuesEvent
2020-09-30 04:48:49
rkulkar5/OnlineTAT
https://api.github.com/repos/rkulkar5/OnlineTAT
opened
Account level sme: Calculate final score is not working
Issues after Sprint-9 high priority
Calculate final score in Initiate technical interview screen of sme queue is not working ![image](https://user-images.githubusercontent.com/61975055/94644169-35aab880-0306-11eb-86d7-41bf04266ae0.png)
1.0
Account level sme: Calculate final score is not working - Calculate final score in Initiate technical interview screen of sme queue is not working ![image](https://user-images.githubusercontent.com/61975055/94644169-35aab880-0306-11eb-86d7-41bf04266ae0.png)
priority
account level sme calculate final score is not working calculate final score in initiate technical interview screen of sme queue is not working
1
429,006
12,419,599,454
IssuesEvent
2020-05-23 07:14:04
dailydotdev/daily-api
https://api.github.com/repos/dailydotdev/daily-api
opened
Add seed data to docker
Priority: High Status: Accepted Type: Maintenance
To complete https://github.com/dailydotdev/daily-apps/issues/110, we must copy seed data to docker so it will be available to `docker-compose` based setup.
1.0
Add seed data to docker - To complete https://github.com/dailydotdev/daily-apps/issues/110, we must copy seed data to docker so it will be available to `docker-compose` based setup.
priority
add seed data to docker to complete we must copy seed data to docker so it will be available to docker compose based setup
1
469,404
13,507,960,720
IssuesEvent
2020-09-14 06:57:00
TerriaJS/terriajs
https://api.github.com/repos/TerriaJS/terriajs
closed
V8: Icons in hamburger menu on StoryBuilder scenes are wrong
High priority Version 8
The view and edit icons in the hamburger menu on StoryBuilder scenes are incorrect. Pictured here: ![image](https://user-images.githubusercontent.com/13863530/92880351-47104b80-f451-11ea-9a7a-b0eff9a5a979.png) See https://github.com/TerriaJS/terriajs/pull/4483#issuecomment-690915337 for icon files
1.0
V8: Icons in hamburger menu on StoryBuilder scenes are wrong - The view and edit icons in the hamburger menu on StoryBuilder scenes are incorrect. Pictured here: ![image](https://user-images.githubusercontent.com/13863530/92880351-47104b80-f451-11ea-9a7a-b0eff9a5a979.png) See https://github.com/TerriaJS/terriajs/pull/4483#issuecomment-690915337 for icon files
priority
icons in hamburger menu on storybuilder scenes are wrong the view and edit icons in the hamburger menu on storybuilder scenes are incorrect pictured here see for icon files
1
240,335
7,801,047,762
IssuesEvent
2018-06-09 16:27:34
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0012544: Missing packages expressomail, requesttracker and sipgate available
Bug Mantis Packaging high priority
**Reported by weberho on 20 Jan 2017 08:49** **Version:** 2016.09.7 Community Edition Those packages are no longer available. What&#039;s the reason? Is this intended?
1.0
0012544: Missing packages expressomail, requesttracker and sipgate available - **Reported by weberho on 20 Jan 2017 08:49** **Version:** 2016.09.7 Community Edition Those packages are no longer available. What&#039;s the reason? Is this intended?
priority
missing packages expressomail requesttracker and sipgate available reported by weberho on jan version community edition those packages are no longer available what s the reason is this intended
1
535,464
15,688,749,496
IssuesEvent
2021-03-25 14:59:29
itslupus/gamersnet
https://api.github.com/repos/itslupus/gamersnet
closed
Database for chat messages
backend dev task high priority
**Description**: Create database method to fetch and store chat messages
1.0
Database for chat messages - **Description**: Create database method to fetch and store chat messages
priority
database for chat messages description create database method to fetch and store chat messages
1
47,023
2,971,374,977
IssuesEvent
2015-07-14 06:13:46
mcdermott-scholars/mcdermott
https://api.github.com/repos/mcdermott-scholars/mcdermott
opened
Class year should also be a drop down
high priority
Starting from 2001, I believe? (or maybe 2000 if you want to count the special class?)
1.0
Class year should also be a drop down - Starting from 2001, I believe? (or maybe 2000 if you want to count the special class?)
priority
class year should also be a drop down starting from i believe or maybe if you want to count the special class
1
340,581
10,273,523,747
IssuesEvent
2019-08-23 19:33:30
byu-animation/dccpipe
https://api.github.com/repos/byu-animation/dccpipe
opened
Change all references to "character" to "actor"
Houdini Maya enhancement priority: high
Because a character is rather limited in definition, but an actor can be an inanimate object that is more involved than a prop, or it can also be a character. i.e. a rug with a hair sim, a clothesline with cloth, a scythe with a 'tude, etc...
1.0
Change all references to "character" to "actor" - Because a character is rather limited in definition, but an actor can be an inanimate object that is more involved than a prop, or it can also be a character. i.e. a rug with a hair sim, a clothesline with cloth, a scythe with a 'tude, etc...
priority
change all references to character to actor because a character is rather limited in definition but an actor can be an inanimate object that is more involved than a prop or it can also be a character i e a rug with a hair sim a clothesline with cloth a scythe with a tude etc
1
242,396
7,841,782,789
IssuesEvent
2018-06-18 20:46:05
stats4sd/Stats4SD-Resources-Site
https://api.github.com/repos/stats4sd/Stats4SD-Resources-Site
closed
Editing always leads to Binomial Distribution
3 - In Progress Priority-High Type-bug
When I click 'edit resource', the active resource changes to `resources/binomial-distribution`, so I can only ever edit that resource. From the console log, clicking the edit button causes the "active resource" to be reset. Oddly, the keywords and resource type are carried over from which-ever resource I clicked through from. E.g. Going to edit [this resource](https://resources.stats4sd.org/resources/experiments-with-farmers-ii---what-happens-to-designs) takes me back to editing the binomial resource, but with the 3 keywords and resource type= video.
1.0
Editing always leads to Binomial Distribution - When I click 'edit resource', the active resource changes to `resources/binomial-distribution`, so I can only ever edit that resource. From the console log, clicking the edit button causes the "active resource" to be reset. Oddly, the keywords and resource type are carried over from which-ever resource I clicked through from. E.g. Going to edit [this resource](https://resources.stats4sd.org/resources/experiments-with-farmers-ii---what-happens-to-designs) takes me back to editing the binomial resource, but with the 3 keywords and resource type= video.
priority
editing always leads to binomial distribution when i click edit resource the active resource changes to resources binomial distribution so i can only ever edit that resource from the console log clicking the edit button causes the active resource to be reset oddly the keywords and resource type are carried over from which ever resource i clicked through from e g going to edit takes me back to editing the binomial resource but with the keywords and resource type video
1
111,712
4,481,763,900
IssuesEvent
2016-08-29 00:40:18
influxdata/docs.influxdata.com
https://api.github.com/repos/influxdata/docs.influxdata.com
closed
remove support for negative timestamps
Fix underway Priority: high
Query engine refactor in 0.11 appears to have disabled negative timestamps in InfluxQL. They can still be written, but not queried. The WHERE clause won't accept negative timestamps, relative or absolute, and the default lower bound is epoch 0. ###### URL for relevant page? https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v0.13/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v1.0/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 possibly others
1.0
remove support for negative timestamps - Query engine refactor in 0.11 appears to have disabled negative timestamps in InfluxQL. They can still be written, but not queried. The WHERE clause won't accept negative timestamps, relative or absolute, and the default lower bound is epoch 0. ###### URL for relevant page? https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range https://docs.influxdata.com/influxdb/v0.11/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-outside-the-min-max-time-range https://docs.influxdata.com/influxdb/v0.12/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v0.13/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 https://docs.influxdata.com/influxdb/v1.0/troubleshooting/frequently_encountered_issues/#querying-a-time-range-that-spans-epoch-0 possibly others
priority
remove support for negative timestamps query engine refactor in appears to have disabled negative timestamps in influxql they can still be written but not queried the where clause won t accept negative timestamps relative or absolute and the default lower bound is epoch url for relevant page possibly others
1
264,740
8,319,139,077
IssuesEvent
2018-09-25 16:22:15
Zenika/FAQ
https://api.github.com/repos/Zenika/FAQ
opened
Setup demo.faq.zenika.com
priority/high
In order to showcase the FAQ, we need to have a demo instance. Needed: * [ ] Check auth0 * [ ] Automatic deployment multi-instances * [ ] Create domain + ssl
1.0
Setup demo.faq.zenika.com - In order to showcase the FAQ, we need to have a demo instance. Needed: * [ ] Check auth0 * [ ] Automatic deployment multi-instances * [ ] Create domain + ssl
priority
setup demo faq zenika com in order to showcase the faq we need to have a demo instance needed check automatic deployment multi instances create domain ssl
1
520,873
15,096,250,418
IssuesEvent
2021-02-07 14:23:47
uah-cpe-sd-2020/myportion_app
https://api.github.com/repos/uah-cpe-sd-2020/myportion_app
opened
Create Database Collections
HIGH PRIORITY
We need to have pet data and feeder data stored in collections. @GuassianFlux Had some ideas on how we might organize this.
1.0
Create Database Collections - We need to have pet data and feeder data stored in collections. @GuassianFlux Had some ideas on how we might organize this.
priority
create database collections we need to have pet data and feeder data stored in collections guassianflux had some ideas on how we might organize this
1
786,568
27,658,909,962
IssuesEvent
2023-03-12 09:44:27
AY2223S2-CS2113-T14-3/tp
https://api.github.com/repos/AY2223S2-CS2113-T14-3/tp
closed
As a user I can delete an expenditure record
type.Story priority.High
As a user I can delete an expenditure record so that I can get rid of expenditure that I no longer plan to use
1.0
As a user I can delete an expenditure record - As a user I can delete an expenditure record so that I can get rid of expenditure that I no longer plan to use
priority
as a user i can delete an expenditure record as a user i can delete an expenditure record so that i can get rid of expenditure that i no longer plan to use
1
620,607
19,565,671,702
IssuesEvent
2022-01-03 23:39:45
bounswe/2021SpringGroup12
https://api.github.com/repos/bounswe/2021SpringGroup12
closed
Design of the Search in Market Place
priority: high android
**Description** - UI design of searching market place should be implemented
1.0
Design of the Search in Market Place - **Description** - UI design of searching market place should be implemented
priority
design of the search in market place description ui design of searching market place should be implemented
1
721,861
24,840,842,348
IssuesEvent
2022-10-26 12:36:55
owncloud/ocis
https://api.github.com/repos/owncloud/ocis
closed
Uploading large files fails when using the desktop client
Status:Bug-Analysis Interaction:Needs-help Priority:p2-high
Hi! I am unable to upload larger files to my ocis instance using the desktop client. It works perfectly fine for files <~100mb but for larger ones, it seems to upload about 100-150MB and then skip the file. After a second of wait it usually retries and fails with a "File is currently in use" type of error. Uploading those files via the web ui works without any issues. An oc10 instance also will process them just fine (mostly). I am using the desktop client 2.11.1 (build 8438) on Windows 11 (22621.675) with the ocis beta 8 running inside the official docker container on another system. As the logs may contain some data which I probably shouldnt share to the public I replaced some names and shorted it down a little. The log of the ocis instance [server_log.txt](https://github.com/owncloud/ocis/files/9863568/server_log.txt) And two from the client. Due to it retrying to upload it created about 8 of them. The "File is currently in use" type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones. [Client log 0, 1, 2](https://github.com/owncloud/ocis/files/9863652/20221025_1637_1_owncloud.log) [Client log 6](https://github.com/owncloud/ocis/files/9863619/20221025_1637_owncloud.log) [Client log 7](https://github.com/owncloud/ocis/files/9863607/20221025_1638_owncloud.log)
1.0
Uploading large files fails when using the desktop client - Hi! I am unable to upload larger files to my ocis instance using the desktop client. It works perfectly fine for files <~100mb but for larger ones, it seems to upload about 100-150MB and then skip the file. After a second of wait it usually retries and fails with a "File is currently in use" type of error. Uploading those files via the web ui works without any issues. An oc10 instance also will process them just fine (mostly). I am using the desktop client 2.11.1 (build 8438) on Windows 11 (22621.675) with the ocis beta 8 running inside the official docker container on another system. As the logs may contain some data which I probably shouldnt share to the public I replaced some names and shorted it down a little. The log of the ocis instance [server_log.txt](https://github.com/owncloud/ocis/files/9863568/server_log.txt) And two from the client. Due to it retrying to upload it created about 8 of them. The "File is currently in use" type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones. [Client log 0, 1, 2](https://github.com/owncloud/ocis/files/9863652/20221025_1637_1_owncloud.log) [Client log 6](https://github.com/owncloud/ocis/files/9863619/20221025_1637_owncloud.log) [Client log 7](https://github.com/owncloud/ocis/files/9863607/20221025_1638_owncloud.log)
priority
uploading large files fails when using the desktop client hi i am unable to upload larger files to my ocis instance using the desktop client it works perfectly fine for files but for larger ones it seems to upload about and then skip the file after a second of wait it usually retries and fails with a file is currently in use type of error uploading those files via the web ui works without any issues an instance also will process them just fine mostly i am using the desktop client build on windows with the ocis beta running inside the official docker container on another system as the logs may contain some data which i probably shouldnt share to the public i replaced some names and shorted it down a little the log of the ocis instance and two from the client due to it retrying to upload it created about of them the file is currently in use type of error seems to have only appeared in the last log while it may have automatically retried the upload in the other ones
1
461,187
13,224,837,217
IssuesEvent
2020-08-17 19:57:04
processing/p5.js-web-editor
https://api.github.com/repos/processing/p5.js-web-editor
closed
Console doesn't scroll to bottom to reveal new messages
good first issue help wanted priority:high type:bug
<!-- Hi there! If you are here to report a bug, or to discuss a feature (new or existing), you can use the below template to get started quickly. Fill out all those parts which you're comfortable with, and delete the remaining ones. --> #### Nature of issue? <!-- Select any one issue and delete the other two --> - Found a bug <!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. --> #### Details about the bug: - Web browser and version: <!-- On Chrome/FireFox/Opera you can enter "about:" in the address bar to find out the version --> Chrome - Operating System: <!-- Ex: Windows/MacOSX/Linux along with version --> Mac OSX - Steps to reproduce this bug: 1. Create a sketch that prints lots of messages to the console (i.e. `print()` in the `draw()` loop) 2. Open the console 3. Notice that the console doesn't scroll to the bottom, so it's hard to see new messages. <!-- Include a simple code snippet that demonstrates the problem, along with any console errors produced. If this isn't possible, then simply describe the issue as best you can! Feel free to link to the web editor or include pictures or a video. --> I believe this was introduced when refactoring the `<Console />` to use `connect()` (see #1502 and #1507). Not sure why it's not working, as I doubled checked and the code that scrolled the console to the bottom is still in there.
1.0
Console doesn't scroll to bottom to reveal new messages - <!-- Hi there! If you are here to report a bug, or to discuss a feature (new or existing), you can use the below template to get started quickly. Fill out all those parts which you're comfortable with, and delete the remaining ones. --> #### Nature of issue? <!-- Select any one issue and delete the other two --> - Found a bug <!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. --> #### Details about the bug: - Web browser and version: <!-- On Chrome/FireFox/Opera you can enter "about:" in the address bar to find out the version --> Chrome - Operating System: <!-- Ex: Windows/MacOSX/Linux along with version --> Mac OSX - Steps to reproduce this bug: 1. Create a sketch that prints lots of messages to the console (i.e. `print()` in the `draw()` loop) 2. Open the console 3. Notice that the console doesn't scroll to the bottom, so it's hard to see new messages. <!-- Include a simple code snippet that demonstrates the problem, along with any console errors produced. If this isn't possible, then simply describe the issue as best you can! Feel free to link to the web editor or include pictures or a video. --> I believe this was introduced when refactoring the `<Console />` to use `connect()` (see #1502 and #1507). Not sure why it's not working, as I doubled checked and the code that scrolled the console to the bottom is still in there.
priority
console doesn t scroll to bottom to reveal new messages hi there if you are here to report a bug or to discuss a feature new or existing you can use the below template to get started quickly fill out all those parts which you re comfortable with and delete the remaining ones nature of issue found a bug details about the bug web browser and version chrome operating system mac osx steps to reproduce this bug create a sketch that prints lots of messages to the console i e print in the draw loop open the console notice that the console doesn t scroll to the bottom so it s hard to see new messages i believe this was introduced when refactoring the to use connect see and not sure why it s not working as i doubled checked and the code that scrolled the console to the bottom is still in there
1