Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
303,594
| 9,308,731,476
|
IssuesEvent
|
2019-03-25 15:06:22
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[craftercms] Provide a consistent monitoring API across all modules
|
enhancement priority: high
|
Provide a consistent set of monitoring APIs that return the same information across all Crafter CMS modules:
* Studio
* Engine
* Deployer
* Profile
* Social
There should be 3 different monitoring APIs, one for each of the following monitoring aspects:
* Memory (used, max)
* Status (startup date time, uptime)
* Version (Crafter version, OS version, JVM version, etc)
You should use the models provided in commons under https://github.com/craftercms/commons/tree/master/utilities/src/main/java/org/craftercms/commons/monitoring, and refactor/enhance them as necessary. OAS documentation is expected for the Studio APIs.
|
1.0
|
[craftercms] Provide a consistent monitoring API across all modules - Provide a consistent set of monitoring APIs that return the same information across all Crafter CMS modules:
* Studio
* Engine
* Deployer
* Profile
* Social
There should be 3 different monitoring APIs, one for each of the following monitoring aspects:
* Memory (used, max)
* Status (startup date time, uptime)
* Version (Crafter version, OS version, JVM version, etc)
You should use the models provided in commons under https://github.com/craftercms/commons/tree/master/utilities/src/main/java/org/craftercms/commons/monitoring, and refactor/enhance them as necessary. OAS documentation is expected for the Studio APIs.
|
priority
|
provide a consistent monitoring api across all modules provide a consistent set of monitoring apis that return the same information across all crafter cms modules studio engine deployer profile social there should be different monitoring apis one for each of the following monitoring aspects memory used max status startup date time uptime version crafter version os version jvm version etc you should use the models provided in commons under and refactor enhance them as necessary oas documentation is expected for the studio apis
| 1
|
406,767
| 11,902,665,655
|
IssuesEvent
|
2020-03-30 14:17:00
|
department-of-veterans-affairs/caseflow
|
https://api.github.com/repos/department-of-veterans-affairs/caseflow
|
closed
|
Run through end-to-end virtual hearings flow
|
Priority: High Product: caseflow-hearings Stakeholder: BVA Team: Tango 💃
|
In advance of pilot, run through the end-to-end virtual hearings flow in UAT to ensure seamless connectivity after adjustments to timezones, Granicus, Pexip, and other integrations.
- [x] Set up UAT
- [x] Run through test scripts
Test email
- [x] Confirm "from" address
- [x] Confirm email is sending with new key
- [x] Confirm timezones by testing hearing scheduled for another (non-east coast) timezone
- [x] Forward emails to stakeholders for content review
Test Pexip
- [x] Set up UAT
- [x] Live test that the the emails send from Caseflow also allow the judge, user, and POA (using our email addresses for testing) to join the hearing
- [ ] Email Pexip with screenshot of the conferences before the jobs run
- [x] Check portal to ensure the jobs have deleted the conferences
- [x] Wait for Pexip's email confirmation the following day that the Caseflow job has deleted the hearing
## Background
Test script from our last round of dogfooding / end-to-end testing, for context:
https://github.com/department-of-veterans-affairs/appeals-team/blob/master/qa/test-plans/virtualhearingsDogfooding-2020-01-13.md
|
1.0
|
Run through end-to-end virtual hearings flow - In advance of pilot, run through the end-to-end virtual hearings flow in UAT to ensure seamless connectivity after adjustments to timezones, Granicus, Pexip, and other integrations.
- [x] Set up UAT
- [x] Run through test scripts
Test email
- [x] Confirm "from" address
- [x] Confirm email is sending with new key
- [x] Confirm timezones by testing hearing scheduled for another (non-east coast) timezone
- [x] Forward emails to stakeholders for content review
Test Pexip
- [x] Set up UAT
- [x] Live test that the the emails send from Caseflow also allow the judge, user, and POA (using our email addresses for testing) to join the hearing
- [ ] Email Pexip with screenshot of the conferences before the jobs run
- [x] Check portal to ensure the jobs have deleted the conferences
- [x] Wait for Pexip's email confirmation the following day that the Caseflow job has deleted the hearing
## Background
Test script from our last round of dogfooding / end-to-end testing, for context:
https://github.com/department-of-veterans-affairs/appeals-team/blob/master/qa/test-plans/virtualhearingsDogfooding-2020-01-13.md
|
priority
|
run through end to end virtual hearings flow in advance of pilot run through the end to end virtual hearings flow in uat to ensure seamless connectivity after adjustments to timezones granicus pexip and other integrations set up uat run through test scripts test email confirm from address confirm email is sending with new key confirm timezones by testing hearing scheduled for another non east coast timezone forward emails to stakeholders for content review test pexip set up uat live test that the the emails send from caseflow also allow the judge user and poa using our email addresses for testing to join the hearing email pexip with screenshot of the conferences before the jobs run check portal to ensure the jobs have deleted the conferences wait for pexip s email confirmation the following day that the caseflow job has deleted the hearing background test script from our last round of dogfooding end to end testing for context
| 1
|
602,472
| 18,469,796,752
|
IssuesEvent
|
2021-10-17 14:41:02
|
buithuytien/GeneCloudOmics
|
https://api.github.com/repos/buithuytien/GeneCloudOmics
|
opened
|
Scatter plot collage download error
|
bug GSOC'21 Priority:High
|
When I download to pdf, it generates html and inside the file says “An error has occurred”
|
1.0
|
Scatter plot collage download error - When I download to pdf, it generates html and inside the file says “An error has occurred”
|
priority
|
scatter plot collage download error when i download to pdf it generates html and inside the file says “an error has occurred”
| 1
|
69,554
| 3,305,466,931
|
IssuesEvent
|
2015-11-04 05:05:08
|
falling-sky/source
|
https://api.github.com/repos/falling-sky/source
|
opened
|
Need tool: "rename text"
|
enhancement Priority-High
|
Need ability to:
Identify text string currently in the project
Identify new text
Insert new text on crowdin
For each language, copy old text translation to new text
|
1.0
|
Need tool: "rename text" - Need ability to:
Identify text string currently in the project
Identify new text
Insert new text on crowdin
For each language, copy old text translation to new text
|
priority
|
need tool rename text need ability to identify text string currently in the project identify new text insert new text on crowdin for each language copy old text translation to new text
| 1
|
188,517
| 6,777,162,912
|
IssuesEvent
|
2017-10-27 20:52:54
|
ngageoint/hootenanny
|
https://api.github.com/repos/ngageoint/hootenanny
|
opened
|
api db bulk inserter changes - 10/27/17
|
Category: Core Priority: High Status: Defined Type: Bug
|
* make the setval statement output optional
* make sure the ID's stored are 64 bit
* reproduce bug with starting ids and fix
|
1.0
|
api db bulk inserter changes - 10/27/17 - * make the setval statement output optional
* make sure the ID's stored are 64 bit
* reproduce bug with starting ids and fix
|
priority
|
api db bulk inserter changes make the setval statement output optional make sure the id s stored are bit reproduce bug with starting ids and fix
| 1
|
432,189
| 12,490,293,878
|
IssuesEvent
|
2020-05-31 23:15:12
|
philburk/simpidemic
|
https://api.github.com/repos/philburk/simpidemic
|
closed
|
slider thumb not updated when query applied
|
bug high priority
|
REPRODUCE
file:///Users/phil/Work/simpidemic/simpidemic.html?ver=10002&cpd=15.00&tpk=25&nds=378&pcd=4.00&ctg=0.200&mtr=2.0&mut=4.0&dtb=7&tdr=14&iml=2.000
EXPECT
ImmunityLoss is at maximum so slider should be at far right.
ACTUAL
Slider thumb is near left side.
Move the slider slightly and the result jumps.
The numeric text is correct but does not match the position.
|
1.0
|
slider thumb not updated when query applied - REPRODUCE
file:///Users/phil/Work/simpidemic/simpidemic.html?ver=10002&cpd=15.00&tpk=25&nds=378&pcd=4.00&ctg=0.200&mtr=2.0&mut=4.0&dtb=7&tdr=14&iml=2.000
EXPECT
ImmunityLoss is at maximum so slider should be at far right.
ACTUAL
Slider thumb is near left side.
Move the slider slightly and the result jumps.
The numeric text is correct but does not match the position.
|
priority
|
slider thumb not updated when query applied reproduce file users phil work simpidemic simpidemic html ver cpd tpk nds pcd ctg mtr mut dtb tdr iml expect immunityloss is at maximum so slider should be at far right actual slider thumb is near left side move the slider slightly and the result jumps the numeric text is correct but does not match the position
| 1
|
47,802
| 2,985,357,174
|
IssuesEvent
|
2015-07-18 23:43:09
|
Efreak/node-steam-chat-bot
|
https://api.github.com/repos/Efreak/node-steam-chat-bot
|
opened
|
Migrate to steam@1.0.0 (version 2.2.0) (meta)
|
High Priority Internal
|
1. #88 :ballot_box_with_check: Get the bot logging in again
2. #89 :ballot_box_with_check: Get chatBot.js working
3. #90 :ballot_box_with_check: Fix triggers and api calls.
5. #91 :ballot_box_with_check: update package.json & release
Notes:
- [this](https://github.com/seishun/node-steam/wiki/Protobufs#gameplayed) may help with custom game names.
- Eventually, I want to actually have a helper script that can generate a config file for the less coding-inclined, as it shouldn't be too hard (just dump `#!/bin/node -` at the top).
- If you want to see my todo/ideas file, please contact me on steam. It's basically got everything I'm working on, planning to work on, want to work on, have vaguely thought about working on regarding steam-chat-bot.
- I want a 2.1.1 release unless we can get 2.2.0 out *very* soon, partly because the npm release 2.1.0 incorrectly identifies itself as 2.0.1.
|
1.0
|
Migrate to steam@1.0.0 (version 2.2.0) (meta) - 1. #88 :ballot_box_with_check: Get the bot logging in again
2. #89 :ballot_box_with_check: Get chatBot.js working
3. #90 :ballot_box_with_check: Fix triggers and api calls.
5. #91 :ballot_box_with_check: update package.json & release
Notes:
- [this](https://github.com/seishun/node-steam/wiki/Protobufs#gameplayed) may help with custom game names.
- Eventually, I want to actually have a helper script that can generate a config file for the less coding-inclined, as it shouldn't be too hard (just dump `#!/bin/node -` at the top).
- If you want to see my todo/ideas file, please contact me on steam. It's basically got everything I'm working on, planning to work on, want to work on, have vaguely thought about working on regarding steam-chat-bot.
- I want a 2.1.1 release unless we can get 2.2.0 out *very* soon, partly because the npm release 2.1.0 incorrectly identifies itself as 2.0.1.
|
priority
|
migrate to steam version meta ballot box with check get the bot logging in again ballot box with check get chatbot js working ballot box with check fix triggers and api calls ballot box with check update package json release notes may help with custom game names eventually i want to actually have a helper script that can generate a config file for the less coding inclined as it shouldn t be too hard just dump bin node at the top if you want to see my todo ideas file please contact me on steam it s basically got everything i m working on planning to work on want to work on have vaguely thought about working on regarding steam chat bot i want a release unless we can get out very soon partly because the npm release incorrectly identifies itself as
| 1
|
283,513
| 8,719,745,357
|
IssuesEvent
|
2018-12-08 03:55:08
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
opened
|
Get XAR working to massively optimize Zulip process startup times
|
area: production difficult priority: high
|
Even with the optimizations we've done in #9953, a basic Zulip process with all Python dependencies takes 1.5s to startup, and most of that is unavoidable given the set of dependencies we have (principally, Django itself).
The main idea we haven't tried that could potentially make a huge difference is integrating Facebook's Xar tool (https://code.fb.com/data-infrastructure/xars-a-more-efficient-open-source-system-for-self-contained-executables/) for optimizing the Python startup time of Zulip. Based on the benchmarks they list there, one could easily imagine making Zulip process startup time more like a few hundred milliseconds, which is fast enough that we could basically declare victory here.
However, I'm not sure how to get Xar working with Zulip's virtualenv, and further investigation is required.
E.g. we don't have a `setup.py` file and it's not clear how to use Xar without one (I opened https://github.com/facebookincubator/xar/issues/27 for this). It's also not clear it supports all the Python versions we need, and how we'd do the production toolchain work given that Zulip supports a range of architectures. But I think we can figure that out once we have a basic prototype of e.g. running `manage.py help` via a Xar archive. Eventually we'll want to make `manage.py` and `uwsgi` work with this, but even just `manage.py` would be huge for our production server restart load situation.
Working on this may be difficult just because it involves using a Facebook open source project that has only 13 commits of public history, but I don't think it requires deep context on how Zulip works.
|
1.0
|
Get XAR working to massively optimize Zulip process startup times - Even with the optimizations we've done in #9953, a basic Zulip process with all Python dependencies takes 1.5s to startup, and most of that is unavoidable given the set of dependencies we have (principally, Django itself).
The main idea we haven't tried that could potentially make a huge difference is integrating Facebook's Xar tool (https://code.fb.com/data-infrastructure/xars-a-more-efficient-open-source-system-for-self-contained-executables/) for optimizing the Python startup time of Zulip. Based on the benchmarks they list there, one could easily imagine making Zulip process startup time more like a few hundred milliseconds, which is fast enough that we could basically declare victory here.
However, I'm not sure how to get Xar working with Zulip's virtualenv, and further investigation is required.
E.g. we don't have a `setup.py` file and it's not clear how to use Xar without one (I opened https://github.com/facebookincubator/xar/issues/27 for this). It's also not clear it supports all the Python versions we need, and how we'd do the production toolchain work given that Zulip supports a range of architectures. But I think we can figure that out once we have a basic prototype of e.g. running `manage.py help` via a Xar archive. Eventually we'll want to make `manage.py` and `uwsgi` work with this, but even just `manage.py` would be huge for our production server restart load situation.
Working on this may be difficult just because it involves using a Facebook open source project that has only 13 commits of public history, but I don't think it requires deep context on how Zulip works.
|
priority
|
get xar working to massively optimize zulip process startup times even with the optimizations we ve done in a basic zulip process with all python dependencies takes to startup and most of that is unavoidable given the set of dependencies we have principally django itself the main idea we haven t tried that could potentially make a huge difference is integrating facebook s xar tool for optimizing the python startup time of zulip based on the benchmarks they list there one could easily imagine making zulip process startup time more like a few hundred milliseconds which is fast enough that we could basically declare victory here however i m not sure how to get xar working with zulip s virtualenv and further investigation is required e g we don t have a setup py file and it s not clear how to use xar without one i opened for this it s also not clear it supports all the python versions we need and how we d do the production toolchain work given that zulip supports a range of architectures but i think we can figure that out once we have a basic prototype of e g running manage py help via a xar archive eventually we ll want to make manage py and uwsgi work with this but even just manage py would be huge for our production server restart load situation working on this may be difficult just because it involves using a facebook open source project that has only commits of public history but i don t think it requires deep context on how zulip works
| 1
|
825,243
| 31,280,990,485
|
IssuesEvent
|
2023-08-22 09:34:14
|
hotosm/tasking-manager
|
https://api.github.com/repos/hotosm/tasking-manager
|
opened
|
Displaying images in email content
|
Component: Frontend Priority: High Component: UI/UX
|
**Is your feature request related to a problem? Please describe.**
From slack -
<img width="971" alt="image" src="https://github.com/hotosm/tasking-manager/assets/12103383/6941049e-b312-4e08-8df6-cddaece1b817">
**Describe the solution you'd like**
Ideally an image should appear in the email body instead of markdown content
|
1.0
|
Displaying images in email content - **Is your feature request related to a problem? Please describe.**
From slack -
<img width="971" alt="image" src="https://github.com/hotosm/tasking-manager/assets/12103383/6941049e-b312-4e08-8df6-cddaece1b817">
**Describe the solution you'd like**
Ideally an image should appear in the email body instead of markdown content
|
priority
|
displaying images in email content is your feature request related to a problem please describe from slack img width alt image src describe the solution you d like ideally an image should appear in the email body instead of markdown content
| 1
|
372,984
| 11,031,366,782
|
IssuesEvent
|
2019-12-06 17:35:11
|
fossasia/open-event-frontend
|
https://api.github.com/repos/fossasia/open-event-frontend
|
opened
|
Byuing Ticket with Credit Card with insufficient funds yields "Unexpected Error"
|
Priority: High bug
|
When a user buys a ticket with insufficient funds on their credit card, the error message that comes up is "Unexpected Error". Instead the correct error should appear. Please implement the correct error messages of failed stripe payments.
|
1.0
|
Byuing Ticket with Credit Card with insufficient funds yields "Unexpected Error" - When a user buys a ticket with insufficient funds on their credit card, the error message that comes up is "Unexpected Error". Instead the correct error should appear. Please implement the correct error messages of failed stripe payments.
|
priority
|
byuing ticket with credit card with insufficient funds yields unexpected error when a user buys a ticket with insufficient funds on their credit card the error message that comes up is unexpected error instead the correct error should appear please implement the correct error messages of failed stripe payments
| 1
|
220,466
| 7,360,133,162
|
IssuesEvent
|
2018-03-10 15:27:19
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
closed
|
Running API Manager as a windows service is failed
|
2.2.0 Priority/Highest Type/Bug
|
**Description:**
I am following documentation [1].
The current version available in soureforge for "Yet Another Java Service Wrapper" is 12.12 in [2].
However after setting JAVA_HOME, CARBON_HOME properly,
when I issue **runConsole.bat** command within <YAJSW_HOME>/bat/ directory, it gives following error.
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>runConsole.b
> at
>
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>pushd C:\Use
> rs\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\
>
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>call setenv.
> bat
> "java" -Xmx30m -Dwrapper_home="C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajs
> w-stable-12.12\bat\/.." -Djna_tmpdir="C:\Users\chamalee\Desktop\yajsw-stable-12.
> 12\yajsw-stable-12.12\bat\/../tmp" -Djava.net.preferIPv4Stack=true -jar "C:\User
> s\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\/../wrapper.jar" -c
> "C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\/../conf/w
> rapper.conf"
> YAJSW: yajsw-stable-12.12
> OS : Windows 7/6.1/amd64
> JVM : Oracle Corporation/1.8.0_161/C:\Program Files\Java\jre1.8.0_161/64
> Mar 10, 2018 3:48:18 PM org.apache.commons.vfs2.VfsLog info
> INFO: Using "C:\Users\chamalee\AppData\Local\Temp\vfs_cache" as temporary files
> store.
> WARNING|wrapper||18-03-10 15:48:18|YAJSW: yajsw-stable-12.12
> WARNING|wrapper||18-03-10 15:48:18|OS : Windows 7/6.1/amd64
> WARNING|wrapper||18-03-10 15:48:18|JVM : Oracle Corporation/1.8.0_161/C:\Progra
> m Files\Java\jre1.8.0_161/64
> SEVERE|wrapper||18-03-10 15:48:19|ERROR: could not get java command
> system.env 52
> INFO|wrapper||18-03-10 15:48:19|could not start process 2
> INFO|wrapper||18-03-10 15:48:19|The system cannot find the file specified.
> INFO|wrapper||18-03-10 15:48:19|null/null/null
> SEVERE|wrapper||18-03-10 15:48:19|failed to spawn wrapped process
> INFO|wrapper||18-03-10 15:48:19|Error in GetExitCodeProcess OS Error #6
> INFO|wrapper||18-03-10 15:48:19|Shutting down Wrapper
> INFO|wrapper||18-03-10 15:48:19|Error in GetExitCodeProcess OS Error #6
[1] https://docs.wso2.com/display/AM2xx/Installing+as+a+Windows+Service
[2] https://sourceforge.net/projects/yajsw/
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
1.0
|
Running API Manager as a windows service is failed - **Description:**
I am following documentation [1].
The current version available in soureforge for "Yet Another Java Service Wrapper" is 12.12 in [2].
However after setting JAVA_HOME, CARBON_HOME properly,
when I issue **runConsole.bat** command within <YAJSW_HOME>/bat/ directory, it gives following error.
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>runConsole.b
> at
>
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>pushd C:\Use
> rs\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\
>
> C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat>call setenv.
> bat
> "java" -Xmx30m -Dwrapper_home="C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajs
> w-stable-12.12\bat\/.." -Djna_tmpdir="C:\Users\chamalee\Desktop\yajsw-stable-12.
> 12\yajsw-stable-12.12\bat\/../tmp" -Djava.net.preferIPv4Stack=true -jar "C:\User
> s\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\/../wrapper.jar" -c
> "C:\Users\chamalee\Desktop\yajsw-stable-12.12\yajsw-stable-12.12\bat\/../conf/w
> rapper.conf"
> YAJSW: yajsw-stable-12.12
> OS : Windows 7/6.1/amd64
> JVM : Oracle Corporation/1.8.0_161/C:\Program Files\Java\jre1.8.0_161/64
> Mar 10, 2018 3:48:18 PM org.apache.commons.vfs2.VfsLog info
> INFO: Using "C:\Users\chamalee\AppData\Local\Temp\vfs_cache" as temporary files
> store.
> WARNING|wrapper||18-03-10 15:48:18|YAJSW: yajsw-stable-12.12
> WARNING|wrapper||18-03-10 15:48:18|OS : Windows 7/6.1/amd64
> WARNING|wrapper||18-03-10 15:48:18|JVM : Oracle Corporation/1.8.0_161/C:\Progra
> m Files\Java\jre1.8.0_161/64
> SEVERE|wrapper||18-03-10 15:48:19|ERROR: could not get java command
> system.env 52
> INFO|wrapper||18-03-10 15:48:19|could not start process 2
> INFO|wrapper||18-03-10 15:48:19|The system cannot find the file specified.
> INFO|wrapper||18-03-10 15:48:19|null/null/null
> SEVERE|wrapper||18-03-10 15:48:19|failed to spawn wrapped process
> INFO|wrapper||18-03-10 15:48:19|Error in GetExitCodeProcess OS Error #6
> INFO|wrapper||18-03-10 15:48:19|Shutting down Wrapper
> INFO|wrapper||18-03-10 15:48:19|Error in GetExitCodeProcess OS Error #6
[1] https://docs.wso2.com/display/AM2xx/Installing+as+a+Windows+Service
[2] https://sourceforge.net/projects/yajsw/
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
priority
|
running api manager as a windows service is failed description i am following documentation the current version available in soureforge for yet another java service wrapper is in however after setting java home carbon home properly when i issue runconsole bat command within bat directory it gives following error c users chamalee desktop yajsw stable yajsw stable bat runconsole b at c users chamalee desktop yajsw stable yajsw stable bat pushd c use rs chamalee desktop yajsw stable yajsw stable bat c users chamalee desktop yajsw stable yajsw stable bat call setenv bat java dwrapper home c users chamalee desktop yajsw stable yajs w stable bat djna tmpdir c users chamalee desktop yajsw stable yajsw stable bat tmp djava net true jar c user s chamalee desktop yajsw stable yajsw stable bat wrapper jar c c users chamalee desktop yajsw stable yajsw stable bat conf w rapper conf yajsw yajsw stable os windows jvm oracle corporation c program files java mar pm org apache commons vfslog info info using c users chamalee appdata local temp vfs cache as temporary files store warning wrapper yajsw yajsw stable warning wrapper os windows warning wrapper jvm oracle corporation c progra m files java severe wrapper error could not get java command system env info wrapper could not start process info wrapper the system cannot find the file specified info wrapper null null null severe wrapper failed to spawn wrapped process info wrapper error in getexitcodeprocess os error info wrapper shutting down wrapper info wrapper error in getexitcodeprocess os error suggested labels suggested assignees affected product version os db other environment details and versions steps to reproduce related issues
| 1
|
724,949
| 24,946,106,690
|
IssuesEvent
|
2022-11-01 00:24:40
|
okTurtles/group-income
|
https://api.github.com/repos/okTurtles/group-income
|
closed
|
Gameification designs for stories, streaks, and badges
|
Kind:Enhancement App:Frontend Priority:High Note:UI/UX
|
### Problem
There are several major problems:
1. The app has no built-in way for informing pledgers of the impact of their $
2. Recievers aren't incentivized to share their progress
3. Pledgers aren't incentivized to make timely payments
### Solution
We need to use streaks, badges, stories, and timelines to incentivize the following behavior:
- Pledgers: completing TODOs
- completion streak
- one additional idea is to include a separate state for "first sender", e.g. you can get ⭐ 's for each time you're the first person in a distribution period to make the payment. And there's a running counter on each pledger's profile showing how many of these they've gotten (if they've gotten at least one).
- Receivers: completing stories
We need a visual display of success rate:
- TODOs graph over time success rate (already an open issue: #1286)
- Timeline of stories/events from receivers
And prominently feature a share story button for receivers.
|
1.0
|
Gameification designs for stories, streaks, and badges - ### Problem
There are several major problems:
1. The app has no built-in way for informing pledgers of the impact of their $
2. Recievers aren't incentivized to share their progress
3. Pledgers aren't incentivized to make timely payments
### Solution
We need to use streaks, badges, stories, and timelines to incentivize the following behavior:
- Pledgers: completing TODOs
- completion streak
- one additional idea is to include a separate state for "first sender", e.g. you can get ⭐ 's for each time you're the first person in a distribution period to make the payment. And there's a running counter on each pledger's profile showing how many of these they've gotten (if they've gotten at least one).
- Receivers: completing stories
We need a visual display of success rate:
- TODOs graph over time success rate (already an open issue: #1286)
- Timeline of stories/events from receivers
And prominently feature a share story button for receivers.
|
priority
|
gameification designs for stories streaks and badges problem there are several major problems the app has no built in way for informing pledgers of the impact of their recievers aren t incentivized to share their progress pledgers aren t incentivized to make timely payments solution we need to use streaks badges stories and timelines to incentivize the following behavior pledgers completing todos completion streak one additional idea is to include a separate state for first sender e g you can get ⭐ s for each time you re the first person in a distribution period to make the payment and there s a running counter on each pledger s profile showing how many of these they ve gotten if they ve gotten at least one receivers completing stories we need a visual display of success rate todos graph over time success rate already an open issue timeline of stories events from receivers and prominently feature a share story button for receivers
| 1
|
463,233
| 13,262,043,801
|
IssuesEvent
|
2020-08-20 21:00:10
|
returntocorp/semgrep
|
https://api.github.com/repos/returntocorp/semgrep
|
closed
|
--skip-unknown-extensions skips all files
|
bug priority:high
|
**Describe the bug**
The new `--skip-unknown-extensions` flag from 0.20.0 seems to just skip all files.
**To Reproduce**
```
echo "eval(eval)" > /tmp/test.py
semgrep --config "https://semgrep.dev/r/python.lang.security.audit.eval-detected.eval-detected" --skip-unknown-extensions /tmp/test.py
```
**Expected behavior**
A finding should be returned, just like when the flag is not given.
The flag is intended to skip files with unknown extensions, so the `.py` file above is meant to be scanned.
**Environment**
semgrep 0.20.0
|
1.0
|
--skip-unknown-extensions skips all files - **Describe the bug**
The new `--skip-unknown-extensions` flag from 0.20.0 seems to just skip all files.
**To Reproduce**
```
echo "eval(eval)" > /tmp/test.py
semgrep --config "https://semgrep.dev/r/python.lang.security.audit.eval-detected.eval-detected" --skip-unknown-extensions /tmp/test.py
```
**Expected behavior**
A finding should be returned, just like when the flag is not given.
The flag is intended to skip files with unknown extensions, so the `.py` file above is meant to be scanned.
**Environment**
semgrep 0.20.0
|
priority
|
skip unknown extensions skips all files describe the bug the new skip unknown extensions flag from seems to just skip all files to reproduce echo eval eval tmp test py semgrep config skip unknown extensions tmp test py expected behavior a finding should be returned just like when the flag is not given the flag is intended to skip files with unknown extensions so the py file above is meant to be scanned environment semgrep
| 1
|
346,485
| 10,412,833,894
|
IssuesEvent
|
2019-09-13 16:57:28
|
RADAR-base/radar-upload-source-connector
|
https://api.github.com/repos/RADAR-base/radar-upload-source-connector
|
closed
|
Add project-id to the route when opening a project-view
|
high-priority upload-backend upload-frontend
|
Currently there is a `projects` route for both project-list and when we select a project. It can be improved by adding project-name to the route when opening a project-view
|
1.0
|
Add project-id to the route when opening a project-view - Currently there is a `projects` route for both project-list and when we select a project. It can be improved by adding project-name to the route when opening a project-view
|
priority
|
add project id to the route when opening a project view currently there is a projects route for both project list and when we select a project it can be improved by adding project name to the route when opening a project view
| 1
|
362,407
| 10,727,132,298
|
IssuesEvent
|
2019-10-28 10:56:30
|
Javacord/Javacord
|
https://api.github.com/repos/Javacord/Javacord
|
opened
|
Auto-Reconnect not working in rare cases
|
bug high priority
|
Sometimes the bot wents down an does not automatically reconnect.
So far, we only have a thread dump when this happened, but no trace log: https://gist.github.com/Vampire/46daaa1fba73c95ae93e8767cd9fc332
This might be related to https://github.com/Javacord/Javacord/issues/274.
|
1.0
|
Auto-Reconnect not working in rare cases - Sometimes the bot wents down an does not automatically reconnect.
So far, we only have a thread dump when this happened, but no trace log: https://gist.github.com/Vampire/46daaa1fba73c95ae93e8767cd9fc332
This might be related to https://github.com/Javacord/Javacord/issues/274.
|
priority
|
auto reconnect not working in rare cases sometimes the bot wents down an does not automatically reconnect so far we only have a thread dump when this happened but no trace log this might be related to
| 1
|
108,270
| 4,330,411,781
|
IssuesEvent
|
2016-07-26 19:56:00
|
meumobi/sitebuilder
|
https://api.github.com/repos/meumobi/sitebuilder
|
closed
|
/dev/root partition on Disk reach 100% on elefante
|
high priority server application platform
|
```bash
root@ks387594[ELEFANTE]:/home/meumobi/PROJECTS/meumobi.com/current# df -k
Filesystem 1K-blocks Used Available Use% Mounted on
rootfs 20424792 19356440 44404 100% /
/dev/root 20424792 19356440 44404 100% /
tmpfs 202880 224 202656 1% /run
tmpfs 5120 0 5120 0% /run/lock
tmpfs 10240 0 10240 0% /dev
tmpfs 510400 0 510400 0% /dev/shm
/dev/sda2 955266072 104856324 802622024 12% /home
```
|
1.0
|
/dev/root partition on Disk reach 100% on elefante -
```bash
root@ks387594[ELEFANTE]:/home/meumobi/PROJECTS/meumobi.com/current# df -k
Filesystem 1K-blocks Used Available Use% Mounted on
rootfs 20424792 19356440 44404 100% /
/dev/root 20424792 19356440 44404 100% /
tmpfs 202880 224 202656 1% /run
tmpfs 5120 0 5120 0% /run/lock
tmpfs 10240 0 10240 0% /dev
tmpfs 510400 0 510400 0% /dev/shm
/dev/sda2 955266072 104856324 802622024 12% /home
```
|
priority
|
dev root partition on disk reach on elefante bash root home meumobi projects meumobi com current df k filesystem blocks used available use mounted on rootfs dev root tmpfs run tmpfs run lock tmpfs dev tmpfs dev shm dev home
| 1
|
452,987
| 13,062,823,654
|
IssuesEvent
|
2020-07-30 15:40:46
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[engine] Authentication conflicts
|
CI bug priority: high
|
## Describe the bug
In preview there is a conflict with targeting that prevents headers authentication to work
## To Reproduce
Steps to reproduce the behavior:
1. Setup Crafter Profile & headers authentication
2. Add a url restriction to trigger the authentication
3. Try to preview the site
## Expected behavior
All authentication methods should work as expected in preview
## Logs
```
java.lang.IllegalArgumentException: Object of class [org.springframework.security.web.authentication.WebAuthenticationDetails] must be an instance of interface org.springframework.security.core.authority.GrantedAuthoritiesContainer
at org.springframework.util.Assert.instanceCheckFailed(Assert.java:389) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.util.Assert.isInstanceOf(Assert.java:327) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.util.Assert.isInstanceOf(Assert.java:339) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedGrantedAuthoritiesUserDetailsService.loadUserDetails(PreAuthenticatedGrantedAuthoritiesUserDetailsService.java:58) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedGrantedAuthoritiesUserDetailsService.loadUserDetails(PreAuthenticatedGrantedAuthoritiesUserDetailsService.java:48) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedAuthenticationProvider.authenticate(PreAuthenticatedAuthenticationProvider.java:103) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.authentication.ProviderManager.authenticate(ProviderManager.java:174) ~[spring-security-core-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.AbstractPreAuthenticatedProcessingFilter.doAuthenticate(AbstractPreAuthenticatedProcessingFilter.java:184) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.AbstractPreAuthenticatedProcessingFilter.doFilter(AbstractPreAuthenticatedProcessingFilter.java:118) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
```
## Specs
### Version
3.1.8
|
1.0
|
[engine] Authentication conflicts - ## Describe the bug
In preview there is a conflict with targeting that prevents headers authentication to work
## To Reproduce
Steps to reproduce the behavior:
1. Setup Crafter Profile & headers authentication
2. Add a url restriction to trigger the authentication
3. Try to preview the site
## Expected behavior
All authentication methods should work as expected in preview
## Logs
```
java.lang.IllegalArgumentException: Object of class [org.springframework.security.web.authentication.WebAuthenticationDetails] must be an instance of interface org.springframework.security.core.authority.GrantedAuthoritiesContainer
at org.springframework.util.Assert.instanceCheckFailed(Assert.java:389) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.util.Assert.isInstanceOf(Assert.java:327) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.util.Assert.isInstanceOf(Assert.java:339) ~[spring-core-4.3.18.RELEASE.jar:4.3.18.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedGrantedAuthoritiesUserDetailsService.loadUserDetails(PreAuthenticatedGrantedAuthoritiesUserDetailsService.java:58) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedGrantedAuthoritiesUserDetailsService.loadUserDetails(PreAuthenticatedGrantedAuthoritiesUserDetailsService.java:48) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.PreAuthenticatedAuthenticationProvider.authenticate(PreAuthenticatedAuthenticationProvider.java:103) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.authentication.ProviderManager.authenticate(ProviderManager.java:174) ~[spring-security-core-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.AbstractPreAuthenticatedProcessingFilter.doAuthenticate(AbstractPreAuthenticatedProcessingFilter.java:184) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
at org.springframework.security.web.authentication.preauth.AbstractPreAuthenticatedProcessingFilter.doFilter(AbstractPreAuthenticatedProcessingFilter.java:118) ~[spring-security-web-4.2.13.RELEASE.jar:4.2.13.RELEASE]
```
## Specs
### Version
3.1.8
|
priority
|
authentication conflicts describe the bug in preview there is a conflict with targeting that prevents headers authentication to work to reproduce steps to reproduce the behavior setup crafter profile headers authentication add a url restriction to trigger the authentication try to preview the site expected behavior all authentication methods should work as expected in preview logs java lang illegalargumentexception object of class must be an instance of interface org springframework security core authority grantedauthoritiescontainer at org springframework util assert instancecheckfailed assert java at org springframework util assert isinstanceof assert java at org springframework util assert isinstanceof assert java at org springframework security web authentication preauth preauthenticatedgrantedauthoritiesuserdetailsservice loaduserdetails preauthenticatedgrantedauthoritiesuserdetailsservice java at org springframework security web authentication preauth preauthenticatedgrantedauthoritiesuserdetailsservice loaduserdetails preauthenticatedgrantedauthoritiesuserdetailsservice java at org springframework security web authentication preauth preauthenticatedauthenticationprovider authenticate preauthenticatedauthenticationprovider java at org springframework security authentication providermanager authenticate providermanager java at org springframework security web authentication preauth abstractpreauthenticatedprocessingfilter doauthenticate abstractpreauthenticatedprocessingfilter java at org springframework security web authentication preauth abstractpreauthenticatedprocessingfilter dofilter abstractpreauthenticatedprocessingfilter java specs version
| 1
|
546,902
| 16,021,014,970
|
IssuesEvent
|
2021-04-20 23:20:25
|
localstack/localstack
|
https://api.github.com/repos/localstack/localstack
|
closed
|
AWS::ApiGatewayV2::Authorizer: Parameter validation failed: Missing required parameter in input: "ApiId"
|
priority-high
|
<!-- Love localstack? Please consider supporting our collective:
👉 https://opencollective.com/localstack/donate -->
# Type of request: This is a ...
- [x] bug report
- [ ] feature request
# Detailed description
Unable to add routes to an ApiGatewayV2 with Authorizer.
## Expected behavior
ApiGatewayV2 routed added with Authorizer
## Actual behavior
```bash
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.services.cloudformation.service_models: Unable to fetch state for resource <localstack_ext.services.cloudformation.service_models.ApiGatewayV2Authorizer object at 0x7fed1ee35eb0>: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId"
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Extract resource attribute: ApiGatewayV2::Authorizer Ref
localstack_1 | 2021-04-20T18:29:14:WARNING:localstack.utils.cloudformation.template_deployer: Unexpected error retrieving details for resource AWS::ApiGatewayV2::Authorizer: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId" File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
localstack_1 | self._bootstrap_inner()
localstack_1 | File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
localstack_1 | self.run()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/utils/aws/aws_utils.py", line 85, in thread_run
localstack_1 | File "/opt/code/localstack/localstack/utils/bootstrap.py", line 681, in run
localstack_1 | result = self.func(self.params)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1936, in _run
localstack_1 | self.do_apply_changes_in_loop(changes, stack, stack_name)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1977, in do_apply_changes_in_loop
localstack_1 | should_deploy = self.prepare_should_deploy_change(
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 2017, in prepare_should_deploy_change
localstack_1 | resolve_refs_recursively(stack.stack_name, resource, new_resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 943, in resolve_refs_recursively
localstack_1 | value[key] = resolve_refs_recursively(stack_name, val, resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 943, in resolve_refs_recursively
localstack_1 | value[key] = resolve_refs_recursively(stack_name, val, resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 846, in resolve_refs_recursively
localstack_1 | ref = resolve_ref(stack_name, value['Ref'], resources, attribute='Ref')
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 812, in resolve_ref
localstack_1 | return determine_resource_physical_id(resource_id=ref,
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1500, in determine_resource_physical_id
localstack_1 | result = extract_resource_attribute(resource_type, {}, attribute or 'PhysicalResourceId',
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py", line 630, in extract_resource_attribute
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 659, in extract_resource_attribute
localstack_1 | resource_state = retrieve_resource_details(resource_id, {}, resources, stack_name) or {}
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py", line 384, in retrieve_resource_details
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 613, in retrieve_resource_details
localstack_1 | state = instance.fetch_and_update_state(stack_name=stack_name, resources=resources)
localstack_1 | File "/opt/code/localstack/localstack/services/cloudformation/service_models.py", line 121, in fetch_and_update_state
localstack_1 | if not template_deployer.check_not_found_exception(e, self.resource_type, self.properties):
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 644, in check_not_found_exception
localstack_1 | (resource_type, e, ''.join(traceback.format_stack()), resource, resource_status))
localstack_1 | - {'ApiId': 'd7b630a1', 'AuthorizerType': 'JWT', 'IdentitySource': ['$request.header.Authorization'], 'Name': 'UserPoolAuthorizer', 'JwtConfiguration': {'Audience': ['jhjvtqpvwf6gftpljddy45urfc'], 'Issuer': 'https://cognito-idp.eu-west-2.amazonaws.com/eu-west-2_3beab2207d6d4ba28b106187aff81524'}} None
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.services.cloudformation.service_models: Unable to fetch state for resource <localstack_ext.services.cloudformation.service_models.ApiGatewayV2Authorizer object at 0x7fed1e7addc0>: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId"
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Dependencies for "testappapiGETprofileAD843515" not yet satisfied, retrying in next loop: Unable to fetch details for resource "testappapiGETprofileUserPoolAuthorizer3B35C8BC" (attribute "Ref")
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Error applying changes for CloudFormation stack "test": Resource deployment loop completed, pending resource changes: [{'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileUserPoolAuthorizer3B35C8BC', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Authorizer', 'Replacement': 'False', 'ChangeSetId': None}}, {'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileAD843515', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Route', 'Replacement': 'False', 'ChangeSetId': None}}] Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1936, in _run
localstack_1 | self.do_apply_changes_in_loop(changes, stack, stack_name)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1997, in do_apply_changes_in_loop
localstack_1 | raise Exception('Resource deployment loop completed, pending resource changes: %s' % changes)
localstack_1 | Exception: Resource deployment loop completed, pending resource changes: [{'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileUserPoolAuthorizer3B35C8BC', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Authorizer', 'Replacement': 'False', 'ChangeSetId': None}}, {'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileAD843515', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Route', 'Replacement': 'False', 'ChangeSetId': None}}]
```
# Steps to reproduce
## Command used to start LocalStack
```bash
docker-compose up -d localstack
```
### docker-compose.yml
```yml
localstack:
image: 'localstack/localstack'
ports:
- "4510-4520:4510-4520"
- "4566-4620:4566-4620"
environment:
- LOCALSTACK_API_KEY=${LOCALSTACK_API_KEY}
- SERVICES=s3,sqs,rds,sns,cloudformation,ec2,iam,lambda,kms,cognito,apigatewayv2,secretsmanager,route53,apigateway
- DEBUG=true
- DATA_DIR=/tmp/localstack/data
- PORT_WEB_UI=8080
- LAMBDA_EXECUTOR=docker
- DOCKER_HOST=unix:///var/run/docker.sock
- DEFAULT_REGION=eu-west-2
- LOCALSTACK_HOSTNAME=localhost
- HOSTNAME=localhost
volumes:
- "/tmp/localstack:/tmp/localstack"
- "/var/run/docker.sock:/var/run/docker.sock"
```
## Client code (AWS SDK code snippet, or sequence of "awslocal" commands)
Stack definition:
```typescript
import * as cdk from "@aws-cdk/core";
import * as cognito from "@aws-cdk/aws-cognito";
import * as apigatewayv2 from "@aws-cdk/aws-apigatewayv2";
import * as apigatewayv2Authorizers from "@aws-cdk/aws-apigatewayv2-authorizers";
import * as apigatewayv2Integrations from "@aws-cdk/aws-apigatewayv2-integrations";
import * as lambda from "@aws-cdk/aws-lambda";
import * as lambdaNodeJs from "@aws-cdk/aws-lambda-nodejs";
export class TestStack extends cdk.Stack {
constructor(scope: cdk.App, id: string, props?: cdk.StackProps) {
super(scope, id, props);
const userPool = new cognito.UserPool(this, `${id}-user-pool`, {
userPoolName: `${id}-user-pool`,
autoVerify: {
email: true,
phone: false
},
signInAliases: {
email: true,
phone: false,
username: false,
preferredUsername: false
}
});
const userPoolClient = new cognito.UserPoolClient(
this,
`${id}-user-pool-client`,
{
userPoolClientName: `${id}-user-pool-client`,
userPool: userPool,
generateSecret: false
}
);
const httpApi = new apigatewayv2.HttpApi(this, `${id}-app-api`, {
apiName: `${id}-app-api`,
corsPreflight: {
allowOrigins: ["*"],
allowHeaders: [
"content-type",
"authorization",
"x-amz-date",
"x-api-key",
"x-amz-security-token",
"x-amz-user-agent"
],
allowMethods: [
apigatewayv2.CorsHttpMethod.GET,
apigatewayv2.CorsHttpMethod.POST,
apigatewayv2.CorsHttpMethod.PUT,
apigatewayv2.CorsHttpMethod.DELETE,
apigatewayv2.CorsHttpMethod.OPTIONS
]
},
defaultAuthorizer: new apigatewayv2Authorizers.HttpUserPoolAuthorizer({
userPool: userPool,
userPoolClient: userPoolClient,
identitySource: ["$request.header.Authorization"]
})
});
const handler = new lambdaNodeJs.NodejsFunction(this, `${id}-profile`, {
functionName: `${id}-profile`,
awsSdkConnectionReuse: true,
handler: "read",
entry: "services/api/routes/profile.js",
memorySize: 128,
bundling: {
externalModules: ["pg-native"]
},
timeout: cdk.Duration.seconds(10),
runtime: lambda.Runtime.NODEJS_12_X
});
const proxy = new apigatewayv2Integrations.LambdaProxyIntegration({
handler
});
httpApi.addRoutes({
integration: proxy,
path: "/profile",
methods: [apigatewayv2.HttpMethod.GET]
});
}
}
```
|
1.0
|
AWS::ApiGatewayV2::Authorizer: Parameter validation failed: Missing required parameter in input: "ApiId" - <!-- Love localstack? Please consider supporting our collective:
👉 https://opencollective.com/localstack/donate -->
# Type of request: This is a ...
- [x] bug report
- [ ] feature request
# Detailed description
Unable to add routes to an ApiGatewayV2 with Authorizer.
## Expected behavior
ApiGatewayV2 routed added with Authorizer
## Actual behavior
```bash
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.services.cloudformation.service_models: Unable to fetch state for resource <localstack_ext.services.cloudformation.service_models.ApiGatewayV2Authorizer object at 0x7fed1ee35eb0>: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId"
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Extract resource attribute: ApiGatewayV2::Authorizer Ref
localstack_1 | 2021-04-20T18:29:14:WARNING:localstack.utils.cloudformation.template_deployer: Unexpected error retrieving details for resource AWS::ApiGatewayV2::Authorizer: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId" File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
localstack_1 | self._bootstrap_inner()
localstack_1 | File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
localstack_1 | self.run()
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/utils/aws/aws_utils.py", line 85, in thread_run
localstack_1 | File "/opt/code/localstack/localstack/utils/bootstrap.py", line 681, in run
localstack_1 | result = self.func(self.params)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1936, in _run
localstack_1 | self.do_apply_changes_in_loop(changes, stack, stack_name)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1977, in do_apply_changes_in_loop
localstack_1 | should_deploy = self.prepare_should_deploy_change(
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 2017, in prepare_should_deploy_change
localstack_1 | resolve_refs_recursively(stack.stack_name, resource, new_resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 943, in resolve_refs_recursively
localstack_1 | value[key] = resolve_refs_recursively(stack_name, val, resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 943, in resolve_refs_recursively
localstack_1 | value[key] = resolve_refs_recursively(stack_name, val, resources)
localstack_1 | File "/opt/code/localstack/localstack/utils/common.py", line 400, in func
localstack_1 | return wrapped(*args, **kwargs)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 846, in resolve_refs_recursively
localstack_1 | ref = resolve_ref(stack_name, value['Ref'], resources, attribute='Ref')
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 812, in resolve_ref
localstack_1 | return determine_resource_physical_id(resource_id=ref,
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1500, in determine_resource_physical_id
localstack_1 | result = extract_resource_attribute(resource_type, {}, attribute or 'PhysicalResourceId',
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py", line 630, in extract_resource_attribute
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 659, in extract_resource_attribute
localstack_1 | resource_state = retrieve_resource_details(resource_id, {}, resources, stack_name) or {}
localstack_1 | File "/opt/code/localstack/.venv/lib/python3.8/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py", line 384, in retrieve_resource_details
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 613, in retrieve_resource_details
localstack_1 | state = instance.fetch_and_update_state(stack_name=stack_name, resources=resources)
localstack_1 | File "/opt/code/localstack/localstack/services/cloudformation/service_models.py", line 121, in fetch_and_update_state
localstack_1 | if not template_deployer.check_not_found_exception(e, self.resource_type, self.properties):
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 644, in check_not_found_exception
localstack_1 | (resource_type, e, ''.join(traceback.format_stack()), resource, resource_status))
localstack_1 | - {'ApiId': 'd7b630a1', 'AuthorizerType': 'JWT', 'IdentitySource': ['$request.header.Authorization'], 'Name': 'UserPoolAuthorizer', 'JwtConfiguration': {'Audience': ['jhjvtqpvwf6gftpljddy45urfc'], 'Issuer': 'https://cognito-idp.eu-west-2.amazonaws.com/eu-west-2_3beab2207d6d4ba28b106187aff81524'}} None
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.services.cloudformation.service_models: Unable to fetch state for resource <localstack_ext.services.cloudformation.service_models.ApiGatewayV2Authorizer object at 0x7fed1e7addc0>: Parameter validation failed:
localstack_1 | Missing required parameter in input: "ApiId"
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Dependencies for "testappapiGETprofileAD843515" not yet satisfied, retrying in next loop: Unable to fetch details for resource "testappapiGETprofileUserPoolAuthorizer3B35C8BC" (attribute "Ref")
localstack_1 | 2021-04-20T18:29:14:DEBUG:localstack.utils.cloudformation.template_deployer: Error applying changes for CloudFormation stack "test": Resource deployment loop completed, pending resource changes: [{'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileUserPoolAuthorizer3B35C8BC', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Authorizer', 'Replacement': 'False', 'ChangeSetId': None}}, {'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileAD843515', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Route', 'Replacement': 'False', 'ChangeSetId': None}}] Traceback (most recent call last):
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1936, in _run
localstack_1 | self.do_apply_changes_in_loop(changes, stack, stack_name)
localstack_1 | File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 1997, in do_apply_changes_in_loop
localstack_1 | raise Exception('Resource deployment loop completed, pending resource changes: %s' % changes)
localstack_1 | Exception: Resource deployment loop completed, pending resource changes: [{'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileUserPoolAuthorizer3B35C8BC', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Authorizer', 'Replacement': 'False', 'ChangeSetId': None}}, {'Type': 'Resource', 'ResourceChange': {'Action': 'Add', 'LogicalResourceId': 'testappapiGETprofileAD843515', 'PhysicalResourceId': None, 'ResourceType': 'AWS::ApiGatewayV2::Route', 'Replacement': 'False', 'ChangeSetId': None}}]
```
# Steps to reproduce
## Command used to start LocalStack
```bash
docker-compose up -d localstack
```
### docker-compose.yml
```yml
localstack:
image: 'localstack/localstack'
ports:
- "4510-4520:4510-4520"
- "4566-4620:4566-4620"
environment:
- LOCALSTACK_API_KEY=${LOCALSTACK_API_KEY}
- SERVICES=s3,sqs,rds,sns,cloudformation,ec2,iam,lambda,kms,cognito,apigatewayv2,secretsmanager,route53,apigateway
- DEBUG=true
- DATA_DIR=/tmp/localstack/data
- PORT_WEB_UI=8080
- LAMBDA_EXECUTOR=docker
- DOCKER_HOST=unix:///var/run/docker.sock
- DEFAULT_REGION=eu-west-2
- LOCALSTACK_HOSTNAME=localhost
- HOSTNAME=localhost
volumes:
- "/tmp/localstack:/tmp/localstack"
- "/var/run/docker.sock:/var/run/docker.sock"
```
## Client code (AWS SDK code snippet, or sequence of "awslocal" commands)
Stack definition:
```typescript
import * as cdk from "@aws-cdk/core";
import * as cognito from "@aws-cdk/aws-cognito";
import * as apigatewayv2 from "@aws-cdk/aws-apigatewayv2";
import * as apigatewayv2Authorizers from "@aws-cdk/aws-apigatewayv2-authorizers";
import * as apigatewayv2Integrations from "@aws-cdk/aws-apigatewayv2-integrations";
import * as lambda from "@aws-cdk/aws-lambda";
import * as lambdaNodeJs from "@aws-cdk/aws-lambda-nodejs";
export class TestStack extends cdk.Stack {
constructor(scope: cdk.App, id: string, props?: cdk.StackProps) {
super(scope, id, props);
const userPool = new cognito.UserPool(this, `${id}-user-pool`, {
userPoolName: `${id}-user-pool`,
autoVerify: {
email: true,
phone: false
},
signInAliases: {
email: true,
phone: false,
username: false,
preferredUsername: false
}
});
const userPoolClient = new cognito.UserPoolClient(
this,
`${id}-user-pool-client`,
{
userPoolClientName: `${id}-user-pool-client`,
userPool: userPool,
generateSecret: false
}
);
const httpApi = new apigatewayv2.HttpApi(this, `${id}-app-api`, {
apiName: `${id}-app-api`,
corsPreflight: {
allowOrigins: ["*"],
allowHeaders: [
"content-type",
"authorization",
"x-amz-date",
"x-api-key",
"x-amz-security-token",
"x-amz-user-agent"
],
allowMethods: [
apigatewayv2.CorsHttpMethod.GET,
apigatewayv2.CorsHttpMethod.POST,
apigatewayv2.CorsHttpMethod.PUT,
apigatewayv2.CorsHttpMethod.DELETE,
apigatewayv2.CorsHttpMethod.OPTIONS
]
},
defaultAuthorizer: new apigatewayv2Authorizers.HttpUserPoolAuthorizer({
userPool: userPool,
userPoolClient: userPoolClient,
identitySource: ["$request.header.Authorization"]
})
});
const handler = new lambdaNodeJs.NodejsFunction(this, `${id}-profile`, {
functionName: `${id}-profile`,
awsSdkConnectionReuse: true,
handler: "read",
entry: "services/api/routes/profile.js",
memorySize: 128,
bundling: {
externalModules: ["pg-native"]
},
timeout: cdk.Duration.seconds(10),
runtime: lambda.Runtime.NODEJS_12_X
});
const proxy = new apigatewayv2Integrations.LambdaProxyIntegration({
handler
});
httpApi.addRoutes({
integration: proxy,
path: "/profile",
methods: [apigatewayv2.HttpMethod.GET]
});
}
}
```
|
priority
|
aws authorizer parameter validation failed missing required parameter in input apiid love localstack please consider supporting our collective 👉 type of request this is a bug report feature request detailed description unable to add routes to an with authorizer expected behavior routed added with authorizer actual behavior bash localstack debug localstack services cloudformation service models unable to fetch state for resource parameter validation failed localstack missing required parameter in input apiid localstack debug localstack utils cloudformation template deployer extract resource attribute authorizer ref localstack warning localstack utils cloudformation template deployer unexpected error retrieving details for resource aws authorizer parameter validation failed localstack missing required parameter in input apiid file usr lib threading py line in bootstrap localstack self bootstrap inner localstack file usr lib threading py line in bootstrap inner localstack self run localstack file opt code localstack venv lib site packages localstack ext utils aws aws utils py line in thread run localstack file opt code localstack localstack utils bootstrap py line in run localstack result self func self params localstack file opt code localstack localstack utils cloudformation template deployer py line in run localstack self do apply changes in loop changes stack stack name localstack file opt code localstack localstack utils cloudformation template deployer py line in do apply changes in loop localstack should deploy self prepare should deploy change localstack file opt code localstack localstack utils cloudformation template deployer py line in prepare should deploy change localstack resolve refs recursively stack stack name resource new resources localstack file opt code localstack localstack utils common py line in func localstack return wrapped args kwargs localstack file opt code localstack localstack utils cloudformation template deployer py line in resolve refs recursively localstack value resolve refs recursively stack name val resources localstack file opt code localstack localstack utils common py line in func localstack return wrapped args kwargs localstack file opt code localstack localstack utils cloudformation template deployer py line in resolve refs recursively localstack value resolve refs recursively stack name val resources localstack file opt code localstack localstack utils common py line in func localstack return wrapped args kwargs localstack file opt code localstack localstack utils cloudformation template deployer py line in resolve refs recursively localstack ref resolve ref stack name value resources attribute ref localstack file opt code localstack localstack utils cloudformation template deployer py line in resolve ref localstack return determine resource physical id resource id ref localstack file opt code localstack localstack utils cloudformation template deployer py line in determine resource physical id localstack result extract resource attribute resource type attribute or physicalresourceid localstack file opt code localstack venv lib site packages localstack ext services cloudformation cloudformation extended py line in extract resource attribute localstack file opt code localstack localstack utils cloudformation template deployer py line in extract resource attribute localstack resource state retrieve resource details resource id resources stack name or localstack file opt code localstack venv lib site packages localstack ext services cloudformation cloudformation extended py line in retrieve resource details localstack file opt code localstack localstack utils cloudformation template deployer py line in retrieve resource details localstack state instance fetch and update state stack name stack name resources resources localstack file opt code localstack localstack services cloudformation service models py line in fetch and update state localstack if not template deployer check not found exception e self resource type self properties localstack file opt code localstack localstack utils cloudformation template deployer py line in check not found exception localstack resource type e join traceback format stack resource resource status localstack apiid authorizertype jwt identitysource name userpoolauthorizer jwtconfiguration audience issuer none localstack debug localstack services cloudformation service models unable to fetch state for resource parameter validation failed localstack missing required parameter in input apiid localstack debug localstack utils cloudformation template deployer dependencies for not yet satisfied retrying in next loop unable to fetch details for resource attribute ref localstack debug localstack utils cloudformation template deployer error applying changes for cloudformation stack test resource deployment loop completed pending resource changes traceback most recent call last localstack file opt code localstack localstack utils cloudformation template deployer py line in run localstack self do apply changes in loop changes stack stack name localstack file opt code localstack localstack utils cloudformation template deployer py line in do apply changes in loop localstack raise exception resource deployment loop completed pending resource changes s changes localstack exception resource deployment loop completed pending resource changes steps to reproduce command used to start localstack bash docker compose up d localstack docker compose yml yml localstack image localstack localstack ports environment localstack api key localstack api key services sqs rds sns cloudformation iam lambda kms cognito secretsmanager apigateway debug true data dir tmp localstack data port web ui lambda executor docker docker host unix var run docker sock default region eu west localstack hostname localhost hostname localhost volumes tmp localstack tmp localstack var run docker sock var run docker sock client code aws sdk code snippet or sequence of awslocal commands stack definition typescript import as cdk from aws cdk core import as cognito from aws cdk aws cognito import as from aws cdk aws import as from aws cdk aws authorizers import as from aws cdk aws integrations import as lambda from aws cdk aws lambda import as lambdanodejs from aws cdk aws lambda nodejs export class teststack extends cdk stack constructor scope cdk app id string props cdk stackprops super scope id props const userpool new cognito userpool this id user pool userpoolname id user pool autoverify email true phone false signinaliases email true phone false username false preferredusername false const userpoolclient new cognito userpoolclient this id user pool client userpoolclientname id user pool client userpool userpool generatesecret false const httpapi new httpapi this id app api apiname id app api corspreflight alloworigins allowheaders content type authorization x amz date x api key x amz security token x amz user agent allowmethods corshttpmethod get corshttpmethod post corshttpmethod put corshttpmethod delete corshttpmethod options defaultauthorizer new httpuserpoolauthorizer userpool userpool userpoolclient userpoolclient identitysource const handler new lambdanodejs nodejsfunction this id profile functionname id profile awssdkconnectionreuse true handler read entry services api routes profile js memorysize bundling externalmodules timeout cdk duration seconds runtime lambda runtime nodejs x const proxy new lambdaproxyintegration handler httpapi addroutes integration proxy path profile methods
| 1
|
823,226
| 30,960,815,970
|
IssuesEvent
|
2023-08-08 03:42:04
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
opened
|
[DocDB] Wait on Conflict - 99%+ transactions are failing with "Unexpected SQL Errors" with featurebench causing the variation in throughput and latency
|
area/docdb priority/high status/awaiting-triage
|
### Description
For the report shared by @robertsami [http://perf.dev.yugabyte.com/report/view/W3siaXNCYXNlbGluZSI6ZmFsc2UsIm5hbWUiOiJTZWxlY[…]oiTGF0ZXN0IE1hc3RlciBCcmFuY2giLCJ0ZXN0X2lkIjoyMjU5NjAyfV0=](http://perf.dev.yugabyte.com/report/view/W3siaXNCYXNlbGluZSI6ZmFsc2UsIm5hbWUiOiJTZWxlY3RlZCBUZXN0LWlkIiwidGVzdF9pZCI6MjIzNjcwMn0seyJpc0Jhc2VsaW5lIjp0cnVlLCJuYW1lIjoiTGF0ZXN0IE1hc3RlciBCcmFuY2giLCJ0ZXN0X2lkIjoyMjU5NjAyfV0=)
Currently in featurebench the Throughput achieved for a given workload is "Total measured requests" achieved in the total execution time (does not include warmup and cooldown period. 120sec). Also "Total measured requests" is addition of all the transaction status like Completed, Rejected, Zero, Aborted and Unexpected SQL Errors.
**Reason for variations:**
For the test LATEST MASTER BRANCH (TEST_ID: 2259602) workload update_key_from_multiple_threads_with_weights_* with 60/100 weights, 99.5%+ of "Total measured requests" are of "Unexpected SQL Errors" state which completed within very less latencies compared to completed transactions i.e. for 100 weight test out of 27683795 "Total measured requests" transactions 27634531 transaction are of "Unexpected SQL Errors" which is why the overall throughput is absurdly high where as the test with 20 weight none of the transactions are of "Unexpected SQL Errors" status and the "Total measured requests" achieved are 176642 which comprise of completed and rejected transactions.
So for same test I tried to calculate the throughput by removing the "Unexpected SQL Errors" from "Total measured requests" which essentially leaves with completed and rejected transactions achieved in complete test duration:
Throughput Test with 100 weight: (27683795 - 27634531)/120(execution time) = 410.53
Throughput Test with 20 weight: (176642-2)/120 = 1472
Post removal of Unexpected SQL Error, the test with absurdly high throughput and low latencies are in ballpark range of corresponding workload of SELECTED TEST-ID (TEST_ID: 2236702).
Slack thread for detailed discussion:
https://yugabyte.slack.com/archives/C027QFMTPA8/p1690705598703589?thread_ts=1690397439.597299&cid=C027QFMTPA8
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
|
1.0
|
[DocDB] Wait on Conflict - 99%+ transactions are failing with "Unexpected SQL Errors" with featurebench causing the variation in throughput and latency - ### Description
For the report shared by @robertsami [http://perf.dev.yugabyte.com/report/view/W3siaXNCYXNlbGluZSI6ZmFsc2UsIm5hbWUiOiJTZWxlY[…]oiTGF0ZXN0IE1hc3RlciBCcmFuY2giLCJ0ZXN0X2lkIjoyMjU5NjAyfV0=](http://perf.dev.yugabyte.com/report/view/W3siaXNCYXNlbGluZSI6ZmFsc2UsIm5hbWUiOiJTZWxlY3RlZCBUZXN0LWlkIiwidGVzdF9pZCI6MjIzNjcwMn0seyJpc0Jhc2VsaW5lIjp0cnVlLCJuYW1lIjoiTGF0ZXN0IE1hc3RlciBCcmFuY2giLCJ0ZXN0X2lkIjoyMjU5NjAyfV0=)
Currently in featurebench the Throughput achieved for a given workload is "Total measured requests" achieved in the total execution time (does not include warmup and cooldown period. 120sec). Also "Total measured requests" is addition of all the transaction status like Completed, Rejected, Zero, Aborted and Unexpected SQL Errors.
**Reason for variations:**
For the test LATEST MASTER BRANCH (TEST_ID: 2259602) workload update_key_from_multiple_threads_with_weights_* with 60/100 weights, 99.5%+ of "Total measured requests" are of "Unexpected SQL Errors" state which completed within very less latencies compared to completed transactions i.e. for 100 weight test out of 27683795 "Total measured requests" transactions 27634531 transaction are of "Unexpected SQL Errors" which is why the overall throughput is absurdly high where as the test with 20 weight none of the transactions are of "Unexpected SQL Errors" status and the "Total measured requests" achieved are 176642 which comprise of completed and rejected transactions.
So for same test I tried to calculate the throughput by removing the "Unexpected SQL Errors" from "Total measured requests" which essentially leaves with completed and rejected transactions achieved in complete test duration:
Throughput Test with 100 weight: (27683795 - 27634531)/120(execution time) = 410.53
Throughput Test with 20 weight: (176642-2)/120 = 1472
Post removal of Unexpected SQL Error, the test with absurdly high throughput and low latencies are in ballpark range of corresponding workload of SELECTED TEST-ID (TEST_ID: 2236702).
Slack thread for detailed discussion:
https://yugabyte.slack.com/archives/C027QFMTPA8/p1690705598703589?thread_ts=1690397439.597299&cid=C027QFMTPA8
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
|
priority
|
wait on conflict transactions are failing with unexpected sql errors with featurebench causing the variation in throughput and latency description for the report shared by robertsami currently in featurebench the throughput achieved for a given workload is total measured requests achieved in the total execution time does not include warmup and cooldown period also total measured requests is addition of all the transaction status like completed rejected zero aborted and unexpected sql errors reason for variations for the test latest master branch test id workload update key from multiple threads with weights with weights of total measured requests are of unexpected sql errors state which completed within very less latencies compared to completed transactions i e for weight test out of total measured requests transactions transaction are of unexpected sql errors which is why the overall throughput is absurdly high where as the test with weight none of the transactions are of unexpected sql errors status and the total measured requests achieved are which comprise of completed and rejected transactions so for same test i tried to calculate the throughput by removing the unexpected sql errors from total measured requests which essentially leaves with completed and rejected transactions achieved in complete test duration throughput test with weight execution time throughput test with weight post removal of unexpected sql error the test with absurdly high throughput and low latencies are in ballpark range of corresponding workload of selected test id test id slack thread for detailed discussion warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information
| 1
|
242,871
| 7,849,726,145
|
IssuesEvent
|
2018-06-20 05:36:04
|
getslash/slash
|
https://api.github.com/repos/getslash/slash
|
closed
|
Remove backport support for `log.errors_subpath` configuration
|
Priority: High
|
This configuration was changed in https://github.com/getslash/slash/issues/692 to `log.highlights_subpath`
|
1.0
|
Remove backport support for `log.errors_subpath` configuration - This configuration was changed in https://github.com/getslash/slash/issues/692 to `log.highlights_subpath`
|
priority
|
remove backport support for log errors subpath configuration this configuration was changed in to log highlights subpath
| 1
|
149,529
| 5,720,391,685
|
IssuesEvent
|
2017-04-20 01:34:19
|
tootsuite/mastodon
|
https://api.github.com/repos/tootsuite/mastodon
|
closed
|
reference error on login
|
bug priority - high ui
|
ReferenceError: Can't find variable: en
I'm stumped. can reach the login screen but once logged in site goes blank dark grey.
* * * *
- [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
|
1.0
|
reference error on login - ReferenceError: Can't find variable: en
I'm stumped. can reach the login screen but once logged in site goes blank dark grey.
* * * *
- [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
|
priority
|
reference error on login referenceerror can t find variable en i m stumped can reach the login screen but once logged in site goes blank dark grey i searched or browsed the repo’s other issues to ensure this is not a duplicate
| 1
|
479,841
| 13,806,276,136
|
IssuesEvent
|
2020-10-11 17:01:32
|
AY2021S1-CS2103T-W15-3/tp
|
https://api.github.com/repos/AY2021S1-CS2103T-W15-3/tp
|
closed
|
Edit help functionality
|
priority.High type.Task
|
Edit help functionality to display command table and updated link to user guide.
|
1.0
|
Edit help functionality - Edit help functionality to display command table and updated link to user guide.
|
priority
|
edit help functionality edit help functionality to display command table and updated link to user guide
| 1
|
273,026
| 8,519,810,416
|
IssuesEvent
|
2018-11-01 15:37:08
|
CS2103-AY1819S1-F10-4/main
|
https://api.github.com/repos/CS2103-AY1819S1-F10-4/main
|
closed
|
Refactor all mentions of "Address Book"
|
priority.high type.task
|
Anything that says "Address Book" or relevant must be refactored to our product name. That can be discussed in our 2nd meeting.
|
1.0
|
Refactor all mentions of "Address Book" - Anything that says "Address Book" or relevant must be refactored to our product name. That can be discussed in our 2nd meeting.
|
priority
|
refactor all mentions of address book anything that says address book or relevant must be refactored to our product name that can be discussed in our meeting
| 1
|
459,669
| 13,196,858,930
|
IssuesEvent
|
2020-08-13 21:33:30
|
WarEmu/WarBugs
|
https://api.github.com/repos/WarEmu/WarBugs
|
closed
|
[Gunbad] Boss not dead, no loot, cant progress.
|
Dungeon: Mount Gunbad High Priority
|
We were fighting one of the gunbad bosses on left wing, Garrolath the Poxbeare, we killed him, but apparently game decided he cannot be killed, and hes dead body was following us.
No chest, no loot and no PQ contribution which means we cannot go through barrier to last boss on left wing.
Video below:
https://youtu.be/2ke_ISIGXO8
|
1.0
|
[Gunbad] Boss not dead, no loot, cant progress. - We were fighting one of the gunbad bosses on left wing, Garrolath the Poxbeare, we killed him, but apparently game decided he cannot be killed, and hes dead body was following us.
No chest, no loot and no PQ contribution which means we cannot go through barrier to last boss on left wing.
Video below:
https://youtu.be/2ke_ISIGXO8
|
priority
|
boss not dead no loot cant progress we were fighting one of the gunbad bosses on left wing garrolath the poxbeare we killed him but apparently game decided he cannot be killed and hes dead body was following us no chest no loot and no pq contribution which means we cannot go through barrier to last boss on left wing video below
| 1
|
515,712
| 14,968,009,645
|
IssuesEvent
|
2021-01-27 16:20:24
|
ga4gh/task-execution-schemas
|
https://api.github.com/repos/ga4gh/task-execution-schemas
|
closed
|
Add CONTRIBUTING.md
|
Priority: High Project: TES
|
Need to add a `CONTRIBUTING.md` file to outline the contribution process. In particular the move to using [HubFlow](http://datasift.github.io/gitflow/) for repo development. Probably easiest to copy over the one from the DRS repo.
|
1.0
|
Add CONTRIBUTING.md - Need to add a `CONTRIBUTING.md` file to outline the contribution process. In particular the move to using [HubFlow](http://datasift.github.io/gitflow/) for repo development. Probably easiest to copy over the one from the DRS repo.
|
priority
|
add contributing md need to add a contributing md file to outline the contribution process in particular the move to using for repo development probably easiest to copy over the one from the drs repo
| 1
|
274,752
| 8,564,954,040
|
IssuesEvent
|
2018-11-09 18:18:34
|
Con-Shoc/auto
|
https://api.github.com/repos/Con-Shoc/auto
|
opened
|
Average Time to First Byte is 1.4s for PDP Pages
|
Beta High Priority PDP Pagespeed
|
**Current Behaviour**
The PDP templates time to first byte is 1.4s.
**Expected Behaviour**
Googles best practice is for time to first byte to be under 1.3s.
|
1.0
|
Average Time to First Byte is 1.4s for PDP Pages - **Current Behaviour**
The PDP templates time to first byte is 1.4s.
**Expected Behaviour**
Googles best practice is for time to first byte to be under 1.3s.
|
priority
|
average time to first byte is for pdp pages current behaviour the pdp templates time to first byte is expected behaviour googles best practice is for time to first byte to be under
| 1
|
619,987
| 19,541,872,871
|
IssuesEvent
|
2022-01-01 03:19:57
|
Pycord-Development/pycord
|
https://api.github.com/repos/Pycord-Development/pycord
|
closed
|
Add support for `guild events`
|
priority: high status: in progress feature
|
### Summary
Add support for `guild events`
### What is the feature request for?
The core library
### The Problem
Add support for `guild events`. Refer https://github.com/discord/discord-api-docs/pull/3586
### The Ideal Solution
Add support for `guild events`.
### The Current Solution
_No response_
### Additional Context
https://github.com/discord/discord-api-docs/pull/3586
|
1.0
|
Add support for `guild events` - ### Summary
Add support for `guild events`
### What is the feature request for?
The core library
### The Problem
Add support for `guild events`. Refer https://github.com/discord/discord-api-docs/pull/3586
### The Ideal Solution
Add support for `guild events`.
### The Current Solution
_No response_
### Additional Context
https://github.com/discord/discord-api-docs/pull/3586
|
priority
|
add support for guild events summary add support for guild events what is the feature request for the core library the problem add support for guild events refer the ideal solution add support for guild events the current solution no response additional context
| 1
|
165,549
| 6,278,030,648
|
IssuesEvent
|
2017-07-18 13:36:31
|
end222/lineage_device_omega
|
https://api.github.com/repos/end222/lineage_device_omega
|
closed
|
Missing fp related file
|
high priority
|
Missing .kl file that has important information regarding the location of the fp sensor.
_Sent from my Xiaomi Redmi Pro using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_
|
1.0
|
Missing fp related file - Missing .kl file that has important information regarding the location of the fp sensor.
_Sent from my Xiaomi Redmi Pro using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_
|
priority
|
missing fp related file missing kl file that has important information regarding the location of the fp sensor sent from my xiaomi redmi pro using
| 1
|
67,023
| 3,265,300,659
|
IssuesEvent
|
2015-10-22 15:40:51
|
tborres/angular-unit-testing
|
https://api.github.com/repos/tborres/angular-unit-testing
|
closed
|
item-count directive
|
priority: 2-high status: 4-review type: feature
|
Create an item-count directive that displays a label and the number of times that label has been used. This will be used within the Language column of the Tally page.
|
1.0
|
item-count directive - Create an item-count directive that displays a label and the number of times that label has been used. This will be used within the Language column of the Tally page.
|
priority
|
item count directive create an item count directive that displays a label and the number of times that label has been used this will be used within the language column of the tally page
| 1
|
146,304
| 5,615,115,057
|
IssuesEvent
|
2017-04-03 13:58:26
|
fossasia/loklak_search
|
https://api.github.com/repos/fossasia/loklak_search
|
closed
|
Start page loading slowly
|
bug Priority: HIGH Priority: URGENT
|
The start page of http://loklak.net should be loading pretty quickly as there is not much on it. Please check why it loads slowly and fix this.
|
2.0
|
Start page loading slowly - The start page of http://loklak.net should be loading pretty quickly as there is not much on it. Please check why it loads slowly and fix this.
|
priority
|
start page loading slowly the start page of should be loading pretty quickly as there is not much on it please check why it loads slowly and fix this
| 1
|
97,825
| 4,006,925,335
|
IssuesEvent
|
2016-05-12 16:22:05
|
w3c/browser-payment-api
|
https://api.github.com/repos/w3c/browser-payment-api
|
closed
|
What is the format for payment method identifiers for distributed extensibility
|
Cat: Extensibility Doc:PaymentMethodIdentifiers in progress Priority: High proposal - needs discussion
|
This issue comes from WICG/paymentrequest#34 and was [discussed at the F2F](https://www.w3.org/2016/02/24-wpwg-minutes#item03).
|
1.0
|
What is the format for payment method identifiers for distributed extensibility - This issue comes from WICG/paymentrequest#34 and was [discussed at the F2F](https://www.w3.org/2016/02/24-wpwg-minutes#item03).
|
priority
|
what is the format for payment method identifiers for distributed extensibility this issue comes from wicg paymentrequest and was
| 1
|
233,299
| 7,696,352,188
|
IssuesEvent
|
2018-05-18 15:04:03
|
Connexions/nebuchadnezzar
|
https://api.github.com/repos/Connexions/nebuchadnezzar
|
closed
|
neb publish on qa returns 502 error
|
bug priority:high
|
Using Nebuchadnezzar 2.0.0:
```
$ neb publish qa col11740 'change "agrarian" to "antiquarian"'
unknown response:
status: 502
contents: <html><body><h1>502 Bad Gateway</h1>
The server returned an invalid or incomplete response.
</body></html>
Traceback (most recent call last):
File "/usr/local/bin/neb", line 11, in <module>
sys.exit(cli())
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/nebu/cli/main.py", line 298, in publish
has_published = _publish(base_url, struct, publication_message)
File "/usr/local/lib/python3.6/site-packages/nebu/cli/main.py", line 280, in _publish
raise RuntimeError('unknown response, see output above')
RuntimeError: unknown response, see output above
```
|
1.0
|
neb publish on qa returns 502 error - Using Nebuchadnezzar 2.0.0:
```
$ neb publish qa col11740 'change "agrarian" to "antiquarian"'
unknown response:
status: 502
contents: <html><body><h1>502 Bad Gateway</h1>
The server returned an invalid or incomplete response.
</body></html>
Traceback (most recent call last):
File "/usr/local/bin/neb", line 11, in <module>
sys.exit(cli())
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/nebu/cli/main.py", line 298, in publish
has_published = _publish(base_url, struct, publication_message)
File "/usr/local/lib/python3.6/site-packages/nebu/cli/main.py", line 280, in _publish
raise RuntimeError('unknown response, see output above')
RuntimeError: unknown response, see output above
```
|
priority
|
neb publish on qa returns error using nebuchadnezzar neb publish qa change agrarian to antiquarian unknown response status contents bad gateway the server returned an invalid or incomplete response traceback most recent call last file usr local bin neb line in sys exit cli file usr local lib site packages click core py line in call return self main args kwargs file usr local lib site packages click core py line in main rv self invoke ctx file usr local lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file usr local lib site packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib site packages click core py line in invoke return callback args kwargs file usr local lib site packages click decorators py line in new func return f get current context args kwargs file usr local lib site packages nebu cli main py line in publish has published publish base url struct publication message file usr local lib site packages nebu cli main py line in publish raise runtimeerror unknown response see output above runtimeerror unknown response see output above
| 1
|
92,962
| 3,875,956,347
|
IssuesEvent
|
2016-04-12 04:55:17
|
Kujawadl/CSC_341_OS_Lab
|
https://api.github.com/repos/Kujawadl/CSC_341_OS_Lab
|
closed
|
Dump before exiting
|
HIGH PRIORITY
|
In addition, I'd like to have an exit function that can give details about the exit. If there are processes running on exit, it would say something about that, vs a clean shutdown.
Speaking of return values, we have functions for each machine instruction returning false on error, we should start using them. If the interpreter catches an error, it should stop execution and return to the scheduler, which should then write out something about the error, and then kill the process.
|
1.0
|
Dump before exiting - In addition, I'd like to have an exit function that can give details about the exit. If there are processes running on exit, it would say something about that, vs a clean shutdown.
Speaking of return values, we have functions for each machine instruction returning false on error, we should start using them. If the interpreter catches an error, it should stop execution and return to the scheduler, which should then write out something about the error, and then kill the process.
|
priority
|
dump before exiting in addition i d like to have an exit function that can give details about the exit if there are processes running on exit it would say something about that vs a clean shutdown speaking of return values we have functions for each machine instruction returning false on error we should start using them if the interpreter catches an error it should stop execution and return to the scheduler which should then write out something about the error and then kill the process
| 1
|
310,064
| 9,485,318,455
|
IssuesEvent
|
2019-04-22 09:52:01
|
strapi/strapi
|
https://api.github.com/repos/strapi/strapi
|
closed
|
Relation tables not created when multi many-to-many relations
|
priority: high status: confirmed type: bug 🐛
|
<!-- ⚠️ If you do not respect this template your issue will be closed. -->
<!-- =============================================================================== -->
<!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. -->
<!-- Please see the wiki for guides on upgrading to the latest release. -->
<!-- =============================================================================== -->
<!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. -->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 10.x.x -->
<!-- npm 6.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 11.4.0<!-- Please ensure you are using the Node LTS version (v10) -->
- **NPM version**: 6.4.1
- **Strapi version**: 3.0.0-alpha.16<!-- Please make sure you are on the latest version -->
- **Database**: Postgres
- **Operating system**: MacOS
**What is the current behavior?**
When I create two many-to-many relations on a Content Type, only the first one is created in the Postgres database.
**Steps to reproduce the problem**
- Create two Content Types (`post` and `tags` for example)
- Create a post
- Create a many-to-many relation between posts and tags
- Create a many-to-many relation between post and users
- Edit the post
```
{ error: relation "pages_posts__posts_pages" does not exist
at Connection.parseE (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:554:11)
at Connection.parseMessage (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:379:19)
at Socket.<anonymous> (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:119:22)
at Socket.emit (events.js:189:13)
at Socket.EventEmitter.emit (domain.js:441:20)
at addChunk (_stream_readable.js:288:12)
at readableAddChunk (_stream_readable.js:269:11)
at Socket.Readable.push (_stream_readable.js:224:10)
at TCP.onStreamRead [as onread] (internal/stream_base_commons.js:145:17)
From previous event:
at Client_PG._query (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/dialects/postgres/index.js:240:12)
at Client_PG.query (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/client.js:192:17)
at Runner.<anonymous> (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:138:36)
From previous event:
at /Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:47:21
From previous event:
at Runner.run (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:33:30)
at Builder.Target.then (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/interface.js:23:43)
at processImmediate (timers.js:632:19)
at process.topLevelDomainCallback (domain.js:120:23)
name: 'error',
length: 124,
severity: 'ERROR',
code: '42P01',
detail: undefined,
hint: undefined,
position: '158',
internalPosition: undefined,
internalQuery: undefined,
where: undefined,
schema: undefined,
table: undefined,
column: undefined,
dataType: undefined,
constraint: undefined,
file: 'parse_relation.c',
line: '1180',
routine: 'parserOpenTable' }
```
**What is the expected behavior?**
The relation table should be created.
**Suggested solutions**
If you move up the in second relation in `Page.settings.json`, it will be created.
|
1.0
|
Relation tables not created when multi many-to-many relations - <!-- ⚠️ If you do not respect this template your issue will be closed. -->
<!-- =============================================================================== -->
<!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. -->
<!-- Please see the wiki for guides on upgrading to the latest release. -->
<!-- =============================================================================== -->
<!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. -->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 10.x.x -->
<!-- npm 6.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 11.4.0<!-- Please ensure you are using the Node LTS version (v10) -->
- **NPM version**: 6.4.1
- **Strapi version**: 3.0.0-alpha.16<!-- Please make sure you are on the latest version -->
- **Database**: Postgres
- **Operating system**: MacOS
**What is the current behavior?**
When I create two many-to-many relations on a Content Type, only the first one is created in the Postgres database.
**Steps to reproduce the problem**
- Create two Content Types (`post` and `tags` for example)
- Create a post
- Create a many-to-many relation between posts and tags
- Create a many-to-many relation between post and users
- Edit the post
```
{ error: relation "pages_posts__posts_pages" does not exist
at Connection.parseE (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:554:11)
at Connection.parseMessage (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:379:19)
at Socket.<anonymous> (/Users/pierreburgy/Desktop/blog/node_modules/pg/lib/connection.js:119:22)
at Socket.emit (events.js:189:13)
at Socket.EventEmitter.emit (domain.js:441:20)
at addChunk (_stream_readable.js:288:12)
at readableAddChunk (_stream_readable.js:269:11)
at Socket.Readable.push (_stream_readable.js:224:10)
at TCP.onStreamRead [as onread] (internal/stream_base_commons.js:145:17)
From previous event:
at Client_PG._query (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/dialects/postgres/index.js:240:12)
at Client_PG.query (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/client.js:192:17)
at Runner.<anonymous> (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:138:36)
From previous event:
at /Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:47:21
From previous event:
at Runner.run (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/runner.js:33:30)
at Builder.Target.then (/Users/pierreburgy/Desktop/blog/node_modules/knex/lib/interface.js:23:43)
at processImmediate (timers.js:632:19)
at process.topLevelDomainCallback (domain.js:120:23)
name: 'error',
length: 124,
severity: 'ERROR',
code: '42P01',
detail: undefined,
hint: undefined,
position: '158',
internalPosition: undefined,
internalQuery: undefined,
where: undefined,
schema: undefined,
table: undefined,
column: undefined,
dataType: undefined,
constraint: undefined,
file: 'parse_relation.c',
line: '1180',
routine: 'parserOpenTable' }
```
**What is the expected behavior?**
The relation table should be created.
**Suggested solutions**
If you move up the in second relation in `Page.settings.json`, it will be created.
|
priority
|
relation tables not created when multi many to many relations informations node js version npm version strapi version alpha database postgres operating system macos what is the current behavior when i create two many to many relations on a content type only the first one is created in the postgres database steps to reproduce the problem create two content types post and tags for example create a post create a many to many relation between posts and tags create a many to many relation between post and users edit the post error relation pages posts posts pages does not exist at connection parsee users pierreburgy desktop blog node modules pg lib connection js at connection parsemessage users pierreburgy desktop blog node modules pg lib connection js at socket users pierreburgy desktop blog node modules pg lib connection js at socket emit events js at socket eventemitter emit domain js at addchunk stream readable js at readableaddchunk stream readable js at socket readable push stream readable js at tcp onstreamread internal stream base commons js from previous event at client pg query users pierreburgy desktop blog node modules knex lib dialects postgres index js at client pg query users pierreburgy desktop blog node modules knex lib client js at runner users pierreburgy desktop blog node modules knex lib runner js from previous event at users pierreburgy desktop blog node modules knex lib runner js from previous event at runner run users pierreburgy desktop blog node modules knex lib runner js at builder target then users pierreburgy desktop blog node modules knex lib interface js at processimmediate timers js at process topleveldomaincallback domain js name error length severity error code detail undefined hint undefined position internalposition undefined internalquery undefined where undefined schema undefined table undefined column undefined datatype undefined constraint undefined file parse relation c line routine parseropentable what is the expected behavior the relation table should be created suggested solutions if you move up the in second relation in page settings json it will be created
| 1
|
767,174
| 26,913,753,974
|
IssuesEvent
|
2023-02-07 03:30:36
|
AY2223S2-CS2103T-W07-1/tp
|
https://api.github.com/repos/AY2223S2-CS2103T-W07-1/tp
|
opened
|
Add a title to each recipe function
|
type.Story priority.High
|
As a cook, I can add a title to each recipe so that I can easily search for them
|
1.0
|
Add a title to each recipe function - As a cook, I can add a title to each recipe so that I can easily search for them
|
priority
|
add a title to each recipe function as a cook i can add a title to each recipe so that i can easily search for them
| 1
|
727,904
| 25,050,020,525
|
IssuesEvent
|
2022-11-05 19:15:41
|
umgc/fall2022
|
https://api.github.com/repos/umgc/fall2022
|
closed
|
Develop: Add Actions panel below mail images
|
enhancement *** HIGH PRIORITY ***
|
As a user, I want to be able to have a single UI panel containing the most relevant and actionable content below each mail image.
See mockups: https://xd.adobe.com/view/db7cd80e-542d-495d-8909-01d4b2607dbe-8e2f/screen/7e31c20e-0f52-470c-88d2-4912f3564f3d
Acceptance Criteria
--
- Any kind of UI panel that can support other elements being added to it in the future is appended to the email body after each image.
- Document any relevant content in the Programmer guide (draft thoughts of usage and importantly the "Why")
- Updates SRS and STP (and other docs if necessary)
- Add/Update relevant unit tests
|
1.0
|
Develop: Add Actions panel below mail images - As a user, I want to be able to have a single UI panel containing the most relevant and actionable content below each mail image.
See mockups: https://xd.adobe.com/view/db7cd80e-542d-495d-8909-01d4b2607dbe-8e2f/screen/7e31c20e-0f52-470c-88d2-4912f3564f3d
Acceptance Criteria
--
- Any kind of UI panel that can support other elements being added to it in the future is appended to the email body after each image.
- Document any relevant content in the Programmer guide (draft thoughts of usage and importantly the "Why")
- Updates SRS and STP (and other docs if necessary)
- Add/Update relevant unit tests
|
priority
|
develop add actions panel below mail images as a user i want to be able to have a single ui panel containing the most relevant and actionable content below each mail image see mockups acceptance criteria any kind of ui panel that can support other elements being added to it in the future is appended to the email body after each image document any relevant content in the programmer guide draft thoughts of usage and importantly the why updates srs and stp and other docs if necessary add update relevant unit tests
| 1
|
446,110
| 12,839,739,280
|
IssuesEvent
|
2020-07-07 19:49:17
|
radical-cybertools/radical.entk
|
https://api.github.com/repos/radical-cybertools/radical.entk
|
closed
|
EnTK script fails while running on any XSEDE resource with "resource allocation failed"
|
layer:rp priority:high topic:resource type:bug
|
Unsure what else I can provide since there are no logs (or even a sandbox) on the agent side.
Trying to run on Comet and Bridges.
Terminal output is below
```
2019-04-11 15:29:44,863: radical.entk.resource_manager.0000: MainProcess : MainThread : INFO : Resource request submission successful.. waiting for pilot to go Active
2019-04-11 15:29:44,864: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: INFO : Pilot pilot.0000 state: PMGR_LAUNCHING
2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: INFO : Pilot pilot.0000 state: FAILED
2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: ERROR : Pilot has failed
2019-04-11 15:29:53,475: radical.entk.resource_manager.0000: MainProcess : MainThread : INFO : Pilot is now active
2019-04-11 15:29:53,475: radical.entk.appmanager.0000: MainProcess : MainThread : ERROR : Error in AppManager: Cannot proceed. Resource allocation ended up in FAILED
Traceback (most recent call last):
File "/home/scm177/VirtualEnvs/Env_RepEx/local/lib/python2.7/site-packages/radical/entk/appman/appmanager.py", line 311, in run
raise EnTKError(msg="Cannot proceed. Resource allocation ended up in %s"%res_alloc_state)
EnTKError: Cannot proceed. Resource allocation ended up in FAILED
2019-04-11 15:29:53,476: radical.entk.appmanager.0000: MainProcess : MainThread : INFO : Terminating WFprocessor
2019-04-11 15:29:53,476: radical.entk.wfprocessor.0000: MainProcess : MainThread : DEBUG : WFprocessor process already terminated
```
`grep ETTOR *log` on the client side gives:
```
pmgr.0000.launching.0.child.log:2019-04-11 15:29:53,402: pmgr.0000.launching.0.child: pmgr.0000.launching.0 : MainThread : ERROR : bulk launch failed
radical.entk.appmanager.0000.log:2019-04-11 15:29:53,475: radical.entk.appmanager.0000: MainProcess : MainThread : ERROR : Error in AppManager: Cannot proceed. Resource allocation ended up in FAILED
radical.entk.resource_manager.0000.log:2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: ERROR : Pilot has failed
re.session.mcewan.engr.rutgers.edu.scm177.017997.0007.log:2019-04-11 15:29:58,632: re.session.mcewan.engr.rutgers.edu.scm177.017997.0007: MainProcess : MainThread : ERROR : failed to fet profile for pilot.0000
```
This seems to be at the EnTK layer, because I _think_ I'm able to run the RP examples on Comet?
`radical-stack` is:
```
python : 2.7.14
pythonpath :
virtualenv : /home/scm177/VirtualEnvs/Env_RepEx
radical.analytics : v0.60.0@devel
radical.entk : 0.7.16-0.7.15-12-g98ea7f1@feature-suspend_resume
radical.pilot : 0.60.0-v0.60.0@devel
radical.saga : 0.60.0
radical.utils : 0.60.0
```
Happy to help debug. Thanks.
|
1.0
|
EnTK script fails while running on any XSEDE resource with "resource allocation failed" - Unsure what else I can provide since there are no logs (or even a sandbox) on the agent side.
Trying to run on Comet and Bridges.
Terminal output is below
```
2019-04-11 15:29:44,863: radical.entk.resource_manager.0000: MainProcess : MainThread : INFO : Resource request submission successful.. waiting for pilot to go Active
2019-04-11 15:29:44,864: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: INFO : Pilot pilot.0000 state: PMGR_LAUNCHING
2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: INFO : Pilot pilot.0000 state: FAILED
2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: ERROR : Pilot has failed
2019-04-11 15:29:53,475: radical.entk.resource_manager.0000: MainProcess : MainThread : INFO : Pilot is now active
2019-04-11 15:29:53,475: radical.entk.appmanager.0000: MainProcess : MainThread : ERROR : Error in AppManager: Cannot proceed. Resource allocation ended up in FAILED
Traceback (most recent call last):
File "/home/scm177/VirtualEnvs/Env_RepEx/local/lib/python2.7/site-packages/radical/entk/appman/appmanager.py", line 311, in run
raise EnTKError(msg="Cannot proceed. Resource allocation ended up in %s"%res_alloc_state)
EnTKError: Cannot proceed. Resource allocation ended up in FAILED
2019-04-11 15:29:53,476: radical.entk.appmanager.0000: MainProcess : MainThread : INFO : Terminating WFprocessor
2019-04-11 15:29:53,476: radical.entk.wfprocessor.0000: MainProcess : MainThread : DEBUG : WFprocessor process already terminated
```
`grep ETTOR *log` on the client side gives:
```
pmgr.0000.launching.0.child.log:2019-04-11 15:29:53,402: pmgr.0000.launching.0.child: pmgr.0000.launching.0 : MainThread : ERROR : bulk launch failed
radical.entk.appmanager.0000.log:2019-04-11 15:29:53,475: radical.entk.appmanager.0000: MainProcess : MainThread : ERROR : Error in AppManager: Cannot proceed. Resource allocation ended up in FAILED
radical.entk.resource_manager.0000.log:2019-04-11 15:29:53,405: radical.entk.resource_manager.0000: MainProcess : pmgr.0000.subscriber._state_sub_cb: ERROR : Pilot has failed
re.session.mcewan.engr.rutgers.edu.scm177.017997.0007.log:2019-04-11 15:29:58,632: re.session.mcewan.engr.rutgers.edu.scm177.017997.0007: MainProcess : MainThread : ERROR : failed to fet profile for pilot.0000
```
This seems to be at the EnTK layer, because I _think_ I'm able to run the RP examples on Comet?
`radical-stack` is:
```
python : 2.7.14
pythonpath :
virtualenv : /home/scm177/VirtualEnvs/Env_RepEx
radical.analytics : v0.60.0@devel
radical.entk : 0.7.16-0.7.15-12-g98ea7f1@feature-suspend_resume
radical.pilot : 0.60.0-v0.60.0@devel
radical.saga : 0.60.0
radical.utils : 0.60.0
```
Happy to help debug. Thanks.
|
priority
|
entk script fails while running on any xsede resource with resource allocation failed unsure what else i can provide since there are no logs or even a sandbox on the agent side trying to run on comet and bridges terminal output is below radical entk resource manager mainprocess mainthread info resource request submission successful waiting for pilot to go active radical entk resource manager mainprocess pmgr subscriber state sub cb info pilot pilot state pmgr launching radical entk resource manager mainprocess pmgr subscriber state sub cb info pilot pilot state failed radical entk resource manager mainprocess pmgr subscriber state sub cb error pilot has failed radical entk resource manager mainprocess mainthread info pilot is now active radical entk appmanager mainprocess mainthread error error in appmanager cannot proceed resource allocation ended up in failed traceback most recent call last file home virtualenvs env repex local lib site packages radical entk appman appmanager py line in run raise entkerror msg cannot proceed resource allocation ended up in s res alloc state entkerror cannot proceed resource allocation ended up in failed radical entk appmanager mainprocess mainthread info terminating wfprocessor radical entk wfprocessor mainprocess mainthread debug wfprocessor process already terminated grep ettor log on the client side gives pmgr launching child log pmgr launching child pmgr launching mainthread error bulk launch failed radical entk appmanager log radical entk appmanager mainprocess mainthread error error in appmanager cannot proceed resource allocation ended up in failed radical entk resource manager log radical entk resource manager mainprocess pmgr subscriber state sub cb error pilot has failed re session mcewan engr rutgers edu log re session mcewan engr rutgers edu mainprocess mainthread error failed to fet profile for pilot this seems to be at the entk layer because i think i m able to run the rp examples on comet radical stack is python pythonpath virtualenv home virtualenvs env repex radical analytics devel radical entk feature suspend resume radical pilot devel radical saga radical utils happy to help debug thanks
| 1
|
204,996
| 7,093,397,455
|
IssuesEvent
|
2018-01-12 20:19:47
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
scim-schema-extension.config file location is wrong in doc
|
Affected/5.3.0 Priority/Highest Resolution/Fixed Type/Docs
|
scim-schema-extension.config file location is wrong in doc [1] in Step 3 of Enable Extension
Path should be [IS-HOME]/repository/conf
3. Locate the scim-schema-extension.config in the path [IS-HOME]/repository/conf/identity/ and add 'somefield' attribute.
[1] https://docs.wso2.com/display/IS530/Extensible+SCIM+User+Schemas+With+WSO2+Identity+Server
|
1.0
|
scim-schema-extension.config file location is wrong in doc - scim-schema-extension.config file location is wrong in doc [1] in Step 3 of Enable Extension
Path should be [IS-HOME]/repository/conf
3. Locate the scim-schema-extension.config in the path [IS-HOME]/repository/conf/identity/ and add 'somefield' attribute.
[1] https://docs.wso2.com/display/IS530/Extensible+SCIM+User+Schemas+With+WSO2+Identity+Server
|
priority
|
scim schema extension config file location is wrong in doc scim schema extension config file location is wrong in doc in step of enable extension path should be repository conf locate the scim schema extension config in the path repository conf identity and add somefield attribute
| 1
|
198,295
| 6,972,116,953
|
IssuesEvent
|
2017-12-11 16:04:04
|
fgpv-vpgf/fgpv-vpgf
|
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
|
closed
|
Plan API / Plugin architecture
|
feedback: discussion priority: high
|
This will be done over several steps:
- [x] Review third party API's like Google Maps to see what are some common design patterns and best practices
- [ ] Solicit feedback from interested parties both in and out of ECCC for use cases we might miss
- [x] Determine if any existing map functionality can/should be branched off to a plugin
- It's clear that for some use cases, a faster loading/highly responsive map is best (ex. PCC) with a trade-off in removing some heavy weight functionality that's not needed. Especially for public facing maps with more basic needs, just there to show some visual data on a map.
- Let's map authors be in control of just how fast vs. feature rich they want their viewer
- [x] Document the basic structure of the API and how it is accessed by page/plugin authors
- [x] ~~List core plugins we will provide and their scope~~
See the [API Manifesto](https://github.com/fgpv-vpgf/fgpv-vpgf/wiki/API-Manifesto) for latest direction.
|
1.0
|
Plan API / Plugin architecture - This will be done over several steps:
- [x] Review third party API's like Google Maps to see what are some common design patterns and best practices
- [ ] Solicit feedback from interested parties both in and out of ECCC for use cases we might miss
- [x] Determine if any existing map functionality can/should be branched off to a plugin
- It's clear that for some use cases, a faster loading/highly responsive map is best (ex. PCC) with a trade-off in removing some heavy weight functionality that's not needed. Especially for public facing maps with more basic needs, just there to show some visual data on a map.
- Let's map authors be in control of just how fast vs. feature rich they want their viewer
- [x] Document the basic structure of the API and how it is accessed by page/plugin authors
- [x] ~~List core plugins we will provide and their scope~~
See the [API Manifesto](https://github.com/fgpv-vpgf/fgpv-vpgf/wiki/API-Manifesto) for latest direction.
|
priority
|
plan api plugin architecture this will be done over several steps review third party api s like google maps to see what are some common design patterns and best practices solicit feedback from interested parties both in and out of eccc for use cases we might miss determine if any existing map functionality can should be branched off to a plugin it s clear that for some use cases a faster loading highly responsive map is best ex pcc with a trade off in removing some heavy weight functionality that s not needed especially for public facing maps with more basic needs just there to show some visual data on a map let s map authors be in control of just how fast vs feature rich they want their viewer document the basic structure of the api and how it is accessed by page plugin authors list core plugins we will provide and their scope see the for latest direction
| 1
|
383,999
| 11,372,627,545
|
IssuesEvent
|
2020-01-28 02:37:23
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
opened
|
[RFC] Support per-RPC timeouts in RPC layer.
|
high priority module: rpc
|
## 🚀 Feature: Implement per-RPC timeouts.
This RFC is similar to the proposal by @xush6528 in https://github.com/pytorch/pytorch/issues/29402. We'd like to get rid of the overall global timeout that currently applies to all RPCs in PyTorch's RPC layer, and replace it with a per-RPC timeout that users can individually configure. This will allow for better reliability in our RPC framework and avoid spurious errors being reported back to users due to misconfigured timeouts, and increase user-friendliness by allowing them to tune their use of the RPC framework to the needs of their application.
## Motivation
We currently have RPC timeouts implemented in out RPC layer, implemented in this PR: https://github.com/pytorch/pytorch/pull/28392 and https://github.com/pytorch/pytorch/pull/29601. Previous issues such as https://github.com/pytorch/pytorch/issues/29018 and https://github.com/pytorch/pytorch/issues/29402 have discussed ideas for extending these timeouts to be on a per-RPC basis.
This would be useful since it gives users the option to have more granular control over their RPCs and setting timeouts for them - if a user is doing something like creating an `RRef` to a module on another node and running an expensive forward pass or data processing operation on that node, they may want to specify a different timeout, than, for example, running a simple user defined function on a different node. As users build more customized applications on top of our rpc/model parallel primitives, having the same RPC timeout for every RPC call will not work.
Supporting per-RPC timeouts will also help internal RPC messages, since they won't have to be bound to a particular timeout that can be set from user land. For example, internal messages such as those associated with the shutdown procedure are bound to the user-set timeout currently.
## Pitch
The API design is the same as the one proposed by @xush6528 in https://github.com/pytorch/pytorch/issues/29402. We will have an optional timeout parameter exposed in all RPC APIs:
```
rpc.rpc_sync(dest, func, args, kwargs, timeout=None)
rpc.rpc_async(dest, func, args, kwargs, timeout=None)
rpc.remote(dest, func, args, kwargs, timeout=None)
```
We can then propagate this timeout to the C++ layer where it can be passed into `RpcAgent::send()`, which creates the `torch::utils::Future` corresponding to the RPC. We can then use the existing scaffolding (such as what is implemented in https://github.com/pytorch/pytorch/pull/29601) to associate the future with a timeout, and mark it completed with an exception if it does time out.
By default, if a timeout is not passed in to `RpcAgent::send()`, we will assume that no timeout is intended, so the future will never be marked as timed out. By default, we can do this for internal messages, and look into whether we want timeouts for internal messages as well - this will require coordination with the retry work going on in https://github.com/pytorch/pytorch/pull/32602/files.
We can also get rid of the existing API `rpc.set_rpc_timeout(timedelta)` that sets the timeout for all RPCs.
|
1.0
|
[RFC] Support per-RPC timeouts in RPC layer. - ## 🚀 Feature: Implement per-RPC timeouts.
This RFC is similar to the proposal by @xush6528 in https://github.com/pytorch/pytorch/issues/29402. We'd like to get rid of the overall global timeout that currently applies to all RPCs in PyTorch's RPC layer, and replace it with a per-RPC timeout that users can individually configure. This will allow for better reliability in our RPC framework and avoid spurious errors being reported back to users due to misconfigured timeouts, and increase user-friendliness by allowing them to tune their use of the RPC framework to the needs of their application.
## Motivation
We currently have RPC timeouts implemented in out RPC layer, implemented in this PR: https://github.com/pytorch/pytorch/pull/28392 and https://github.com/pytorch/pytorch/pull/29601. Previous issues such as https://github.com/pytorch/pytorch/issues/29018 and https://github.com/pytorch/pytorch/issues/29402 have discussed ideas for extending these timeouts to be on a per-RPC basis.
This would be useful since it gives users the option to have more granular control over their RPCs and setting timeouts for them - if a user is doing something like creating an `RRef` to a module on another node and running an expensive forward pass or data processing operation on that node, they may want to specify a different timeout, than, for example, running a simple user defined function on a different node. As users build more customized applications on top of our rpc/model parallel primitives, having the same RPC timeout for every RPC call will not work.
Supporting per-RPC timeouts will also help internal RPC messages, since they won't have to be bound to a particular timeout that can be set from user land. For example, internal messages such as those associated with the shutdown procedure are bound to the user-set timeout currently.
## Pitch
The API design is the same as the one proposed by @xush6528 in https://github.com/pytorch/pytorch/issues/29402. We will have an optional timeout parameter exposed in all RPC APIs:
```
rpc.rpc_sync(dest, func, args, kwargs, timeout=None)
rpc.rpc_async(dest, func, args, kwargs, timeout=None)
rpc.remote(dest, func, args, kwargs, timeout=None)
```
We can then propagate this timeout to the C++ layer where it can be passed into `RpcAgent::send()`, which creates the `torch::utils::Future` corresponding to the RPC. We can then use the existing scaffolding (such as what is implemented in https://github.com/pytorch/pytorch/pull/29601) to associate the future with a timeout, and mark it completed with an exception if it does time out.
By default, if a timeout is not passed in to `RpcAgent::send()`, we will assume that no timeout is intended, so the future will never be marked as timed out. By default, we can do this for internal messages, and look into whether we want timeouts for internal messages as well - this will require coordination with the retry work going on in https://github.com/pytorch/pytorch/pull/32602/files.
We can also get rid of the existing API `rpc.set_rpc_timeout(timedelta)` that sets the timeout for all RPCs.
|
priority
|
support per rpc timeouts in rpc layer 🚀 feature implement per rpc timeouts this rfc is similar to the proposal by in we d like to get rid of the overall global timeout that currently applies to all rpcs in pytorch s rpc layer and replace it with a per rpc timeout that users can individually configure this will allow for better reliability in our rpc framework and avoid spurious errors being reported back to users due to misconfigured timeouts and increase user friendliness by allowing them to tune their use of the rpc framework to the needs of their application motivation we currently have rpc timeouts implemented in out rpc layer implemented in this pr and previous issues such as and have discussed ideas for extending these timeouts to be on a per rpc basis this would be useful since it gives users the option to have more granular control over their rpcs and setting timeouts for them if a user is doing something like creating an rref to a module on another node and running an expensive forward pass or data processing operation on that node they may want to specify a different timeout than for example running a simple user defined function on a different node as users build more customized applications on top of our rpc model parallel primitives having the same rpc timeout for every rpc call will not work supporting per rpc timeouts will also help internal rpc messages since they won t have to be bound to a particular timeout that can be set from user land for example internal messages such as those associated with the shutdown procedure are bound to the user set timeout currently pitch the api design is the same as the one proposed by in we will have an optional timeout parameter exposed in all rpc apis rpc rpc sync dest func args kwargs timeout none rpc rpc async dest func args kwargs timeout none rpc remote dest func args kwargs timeout none we can then propagate this timeout to the c layer where it can be passed into rpcagent send which creates the torch utils future corresponding to the rpc we can then use the existing scaffolding such as what is implemented in to associate the future with a timeout and mark it completed with an exception if it does time out by default if a timeout is not passed in to rpcagent send we will assume that no timeout is intended so the future will never be marked as timed out by default we can do this for internal messages and look into whether we want timeouts for internal messages as well this will require coordination with the retry work going on in we can also get rid of the existing api rpc set rpc timeout timedelta that sets the timeout for all rpcs
| 1
|
630,888
| 20,119,928,248
|
IssuesEvent
|
2022-02-08 00:28:44
|
yukiHaga/regex-hunting
|
https://api.github.com/repos/yukiHaga/regex-hunting
|
closed
|
hh:mm:ssのhhを消す
|
Priority: high Type: improvement
|
## 概要
クリアタイムの表記がhh:mm:ssである。それをmm:ssに変更する。
## やること
- [x] クリアタイムの表記をhh:mm:ssから、mm:ssに変更する。
- [x] ランキングとマイページのsplitを消す。
## 受け入れ条件
- [x] クリアタイム、ランキング、マイページのクリアタイム表記がmm:ssになっている。
## 参考記事
特になし。
|
1.0
|
hh:mm:ssのhhを消す - ## 概要
クリアタイムの表記がhh:mm:ssである。それをmm:ssに変更する。
## やること
- [x] クリアタイムの表記をhh:mm:ssから、mm:ssに変更する。
- [x] ランキングとマイページのsplitを消す。
## 受け入れ条件
- [x] クリアタイム、ランキング、マイページのクリアタイム表記がmm:ssになっている。
## 参考記事
特になし。
|
priority
|
hh mm ssのhhを消す 概要 クリアタイムの表記がhh mm ssである。それをmm ssに変更する。 やること クリアタイムの表記をhh mm ssから、mm ssに変更する。 ランキングとマイページのsplitを消す。 受け入れ条件 クリアタイム、ランキング、マイページのクリアタイム表記がmm ssになっている。 参考記事 特になし。
| 1
|
561,175
| 16,612,590,642
|
IssuesEvent
|
2021-06-02 13:19:07
|
geneontology/go-annotation
|
https://api.github.com/repos/geneontology/go-annotation
|
closed
|
PTN000497525 cactin
|
PAINT annotation PomBase high priority
|
* **PTHR ID & PTN node:**
GO:0010468 | regulation of gene expression | IBA with WBGene00012230 , PTN000497525 , Q8WUQ7
* **Sequences with problematic annotation (ID + gene/protein name):**
SPBC2F12.12c
O14342
cactin, spliceosome complex subunit
* **Type of Issue: Erroneous source or erroneous propagation, or other issue**
Erroneous source. Cactin is a splicing factor. Splicing has plieotrophic phenotypes so a lot of the annotation ios probably indirect.Although some splicing factors are regulatory, for splice variants, this could not be transferred.
It definitely seems to be a splicing factor according to this recent Science paper (uncurated)
A human postcatalytic spliceosome structure reveals essential roles of metazoan factors for exon ligation.
Fica SM, Oubridge C, Wilkinson ME, Newman AJ, Nagai K.
Science. 2019 Feb 15;363(6428):710-714. doi: 10.1126/science.aaw5569. Epub 2019 Jan 31.
PMID: 30705154
|
1.0
|
PTN000497525 cactin - * **PTHR ID & PTN node:**
GO:0010468 | regulation of gene expression | IBA with WBGene00012230 , PTN000497525 , Q8WUQ7
* **Sequences with problematic annotation (ID + gene/protein name):**
SPBC2F12.12c
O14342
cactin, spliceosome complex subunit
* **Type of Issue: Erroneous source or erroneous propagation, or other issue**
Erroneous source. Cactin is a splicing factor. Splicing has plieotrophic phenotypes so a lot of the annotation ios probably indirect.Although some splicing factors are regulatory, for splice variants, this could not be transferred.
It definitely seems to be a splicing factor according to this recent Science paper (uncurated)
A human postcatalytic spliceosome structure reveals essential roles of metazoan factors for exon ligation.
Fica SM, Oubridge C, Wilkinson ME, Newman AJ, Nagai K.
Science. 2019 Feb 15;363(6428):710-714. doi: 10.1126/science.aaw5569. Epub 2019 Jan 31.
PMID: 30705154
|
priority
|
cactin pthr id ptn node go regulation of gene expression iba with sequences with problematic annotation id gene protein name cactin spliceosome complex subunit type of issue erroneous source or erroneous propagation or other issue erroneous source cactin is a splicing factor splicing has plieotrophic phenotypes so a lot of the annotation ios probably indirect although some splicing factors are regulatory for splice variants this could not be transferred it definitely seems to be a splicing factor according to this recent science paper uncurated a human postcatalytic spliceosome structure reveals essential roles of metazoan factors for exon ligation fica sm oubridge c wilkinson me newman aj nagai k science feb doi science epub jan pmid
| 1
|
334,880
| 10,146,840,341
|
IssuesEvent
|
2019-08-05 09:08:31
|
ICPI/OCM
|
https://api.github.com/repos/ICPI/OCM
|
reopened
|
ICPI Web Page/Internet Presence
|
Priority: High Status: Help Wanted Task: Communications reminder
|
Analytic Question: How can ICPI have a greater web presence, and through web presence/transparency, be part of the greater solution process for OGAC?
Audience: ICPI staff, HQ, Field, External Audiences
<!-- probot = {"230494":{"who":"jb3436","what":"update web page in","when":"2019-08-05T09:00:00.000Z"}} -->
|
1.0
|
ICPI Web Page/Internet Presence - Analytic Question: How can ICPI have a greater web presence, and through web presence/transparency, be part of the greater solution process for OGAC?
Audience: ICPI staff, HQ, Field, External Audiences
<!-- probot = {"230494":{"who":"jb3436","what":"update web page in","when":"2019-08-05T09:00:00.000Z"}} -->
|
priority
|
icpi web page internet presence analytic question how can icpi have a greater web presence and through web presence transparency be part of the greater solution process for ogac audience icpi staff hq field external audiences
| 1
|
459,614
| 13,195,868,597
|
IssuesEvent
|
2020-08-13 19:30:21
|
gadenbuie/shrtcts
|
https://api.github.com/repos/gadenbuie/shrtcts
|
closed
|
Populate `addins.json` automatically?
|
priority: high status: in progress type: new
|
Hey Garrick!
Awesome package, as always.
I was wondering if {shrtcts} could populate a shortcut entry in `.rstudio/keybindings/addins.json` automatically if a keybinding is given in the YAML config file?
For example, setting a shortcut to a {shrtcts} addin adds the following entry to `.rstudio/keybindings/addins.json` currently:
```json
shrtcts::shortcut_01: <shortcut>
```
This entry could be automatically created by {shrtcts} due to the ID link?
This would save some clicks and everything would be at one place.
```yml
- Name: <name>
Binding: <binding>
interactive: true
id: 1
shortcut: <shortcut>
```
The tricky part might be to also remove entries if they are removed in the YAML. Hence, {shrtcts} would need to go through the YAML file on every startup :/ Too much?
|
1.0
|
Populate `addins.json` automatically? - Hey Garrick!
Awesome package, as always.
I was wondering if {shrtcts} could populate a shortcut entry in `.rstudio/keybindings/addins.json` automatically if a keybinding is given in the YAML config file?
For example, setting a shortcut to a {shrtcts} addin adds the following entry to `.rstudio/keybindings/addins.json` currently:
```json
shrtcts::shortcut_01: <shortcut>
```
This entry could be automatically created by {shrtcts} due to the ID link?
This would save some clicks and everything would be at one place.
```yml
- Name: <name>
Binding: <binding>
interactive: true
id: 1
shortcut: <shortcut>
```
The tricky part might be to also remove entries if they are removed in the YAML. Hence, {shrtcts} would need to go through the YAML file on every startup :/ Too much?
|
priority
|
populate addins json automatically hey garrick awesome package as always i was wondering if shrtcts could populate a shortcut entry in rstudio keybindings addins json automatically if a keybinding is given in the yaml config file for example setting a shortcut to a shrtcts addin adds the following entry to rstudio keybindings addins json currently json shrtcts shortcut this entry could be automatically created by shrtcts due to the id link this would save some clicks and everything would be at one place yml name binding interactive true id shortcut the tricky part might be to also remove entries if they are removed in the yaml hence shrtcts would need to go through the yaml file on every startup too much
| 1
|
419,661
| 12,226,571,812
|
IssuesEvent
|
2020-05-03 11:33:53
|
OpenApoc/OpenApoc
|
https://api.github.com/repos/OpenApoc/OpenApoc
|
closed
|
Cityscape Debug R Command Crash
|
!BUG! HIGH PRIORITY
|
When using the Repair command in Cityscape (R-key), the game crashes with the attached error. I did have damaged terrain and buildings so there should have been things to repair.

|
1.0
|
Cityscape Debug R Command Crash - When using the Repair command in Cityscape (R-key), the game crashes with the attached error. I did have damaged terrain and buildings so there should have been things to repair.

|
priority
|
cityscape debug r command crash when using the repair command in cityscape r key the game crashes with the attached error i did have damaged terrain and buildings so there should have been things to repair
| 1
|
626,091
| 19,784,832,768
|
IssuesEvent
|
2022-01-18 04:41:32
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
opened
|
[4.1.0-M2][UI] Solace - Introduce SOLACE-PUBSUB as API Type
|
Type/Improvement Priority/High
|
### Describe your problem(s)
The current Solace AsyncAPIs Type is **WEBSUB**.


But Solace APIs need to be recognised immediately as separate APIs.
### Describe your solution
- Introduce a Solace specific API type like **SOLACE PUBSUB** or just **SOLACE**
-
- Render Solace specific API type in APIs views (List and Tiles)
### How will you implement it
<!-- If you like to suggest an approach or a design -->
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
|
1.0
|
[4.1.0-M2][UI] Solace - Introduce SOLACE-PUBSUB as API Type - ### Describe your problem(s)
The current Solace AsyncAPIs Type is **WEBSUB**.


But Solace APIs need to be recognised immediately as separate APIs.
### Describe your solution
- Introduce a Solace specific API type like **SOLACE PUBSUB** or just **SOLACE**
-
- Render Solace specific API type in APIs views (List and Tiles)
### How will you implement it
<!-- If you like to suggest an approach or a design -->
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
|
priority
|
solace introduce solace pubsub as api type describe your problem s the current solace asyncapis type is websub but solace apis need to be recognised immediately as separate apis describe your solution introduce a solace specific api type like solace pubsub or just solace render solace specific api type in apis views list and tiles how will you implement it optional fields related issues suggested labels suggested assignees
| 1
|
243,343
| 7,856,577,691
|
IssuesEvent
|
2018-06-21 08:04:12
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Latest version of a package is picked when an older version is specified in Ballerina.toml - Only for the first time
|
Priority/High Type/Bug component/Compiler
|
**Steps to reproduce:**
- Use Ballerina 0.970.1 release.
- Please make sure that you don't have any wso2/twitter package versions in your home repository.
- Use the twitter example [here](https://github.com/ballerina-guides/ballerina-demo/blob/master/4_demo_transformations.bal)
- First try to build this sample. You will get a compilation error.
- Then Create the Ballerina.toml and add a dependency to "wso2/twitter" package
```
[project]
org-name = "sameera"
version = "0.0.1"
[dependencies."wso2/twitter"]
version = "0.9.10"
```
- Again build the sample
- Now you will get the same compilation failure.
-
Build again, you won't get that error. :)
**Affected Versions:**
0.970.1
|
1.0
|
Latest version of a package is picked when an older version is specified in Ballerina.toml - Only for the first time -
**Steps to reproduce:**
- Use Ballerina 0.970.1 release.
- Please make sure that you don't have any wso2/twitter package versions in your home repository.
- Use the twitter example [here](https://github.com/ballerina-guides/ballerina-demo/blob/master/4_demo_transformations.bal)
- First try to build this sample. You will get a compilation error.
- Then Create the Ballerina.toml and add a dependency to "wso2/twitter" package
```
[project]
org-name = "sameera"
version = "0.0.1"
[dependencies."wso2/twitter"]
version = "0.9.10"
```
- Again build the sample
- Now you will get the same compilation failure.
-
Build again, you won't get that error. :)
**Affected Versions:**
0.970.1
|
priority
|
latest version of a package is picked when an older version is specified in ballerina toml only for the first time steps to reproduce use ballerina release please make sure that you don t have any twitter package versions in your home repository use the twitter example first try to build this sample you will get a compilation error then create the ballerina toml and add a dependency to twitter package org name sameera version version again build the sample now you will get the same compilation failure build again you won t get that error affected versions
| 1
|
662,648
| 22,147,534,393
|
IssuesEvent
|
2022-06-03 13:34:19
|
Prathamesh-Shanbhag/Portfolio-Website
|
https://api.github.com/repos/Prathamesh-Shanbhag/Portfolio-Website
|
closed
|
Drifting Pixels
|
🕹 aspect: interface 🟧 priority: high 🛠 goal: fix no-issue-activity
|
### Description
The drifting pixels animation may warrant an epilepsy warning - games with flashing lights, neon colour shifts, drifting snow animations and the like tend to have an epilepsy warning. You never know if a user is sensitive to the same so it's best to stay on the safe side
### Screenshots

### Additional information
_No response_
|
1.0
|
Drifting Pixels - ### Description
The drifting pixels animation may warrant an epilepsy warning - games with flashing lights, neon colour shifts, drifting snow animations and the like tend to have an epilepsy warning. You never know if a user is sensitive to the same so it's best to stay on the safe side
### Screenshots

### Additional information
_No response_
|
priority
|
drifting pixels description the drifting pixels animation may warrant an epilepsy warning games with flashing lights neon colour shifts drifting snow animations and the like tend to have an epilepsy warning you never know if a user is sensitive to the same so it s best to stay on the safe side screenshots additional information no response
| 1
|
697,497
| 23,941,663,892
|
IssuesEvent
|
2022-09-12 00:25:45
|
lilyx13/lilyx-website
|
https://api.github.com/repos/lilyx13/lilyx-website
|
opened
|
Header and Footer
|
enhancement priority | high
|
### Context
These are bundled together because they're part of the general UI of the site, wrapped around everything else and they will share a lot of the same information (such as navigation links).
### Definition of Done
- [ ] Site map is accessible and easy to get around
- [ ] Orients the User
- [ ] Mobile friendly
|
1.0
|
Header and Footer - ### Context
These are bundled together because they're part of the general UI of the site, wrapped around everything else and they will share a lot of the same information (such as navigation links).
### Definition of Done
- [ ] Site map is accessible and easy to get around
- [ ] Orients the User
- [ ] Mobile friendly
|
priority
|
header and footer context these are bundled together because they re part of the general ui of the site wrapped around everything else and they will share a lot of the same information such as navigation links definition of done site map is accessible and easy to get around orients the user mobile friendly
| 1
|
535,066
| 15,681,459,054
|
IssuesEvent
|
2021-03-25 05:24:22
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
closed
|
Add entry to download exvivo template
|
installation priority:HIGH sct_download_data
|
### Context
@charleygros created [this exvivo template](https://github.com/sct-data/exvivo-template), which is notably used by our collaborators in Australia (see [user forum](https://forum.spinalcordmri.org/t/registration-to-t1-ex-vivo-template-error/644/6)).
Currently there is no instructions as to how to download this template.
### Suggestion
Add an entry to `sct_download_data` to install this template.
Note: this template should ideally be installed under `$SCT_DIR/data/`. It would be cleaner IMHO.
I've added the label "priority" because currently the installation requires users to install https://github.com/Drulex/ghsplit/, which for some users is a complicated series of steps.
|
1.0
|
Add entry to download exvivo template - ### Context
@charleygros created [this exvivo template](https://github.com/sct-data/exvivo-template), which is notably used by our collaborators in Australia (see [user forum](https://forum.spinalcordmri.org/t/registration-to-t1-ex-vivo-template-error/644/6)).
Currently there is no instructions as to how to download this template.
### Suggestion
Add an entry to `sct_download_data` to install this template.
Note: this template should ideally be installed under `$SCT_DIR/data/`. It would be cleaner IMHO.
I've added the label "priority" because currently the installation requires users to install https://github.com/Drulex/ghsplit/, which for some users is a complicated series of steps.
|
priority
|
add entry to download exvivo template context charleygros created which is notably used by our collaborators in australia see currently there is no instructions as to how to download this template suggestion add an entry to sct download data to install this template note this template should ideally be installed under sct dir data it would be cleaner imho i ve added the label priority because currently the installation requires users to install which for some users is a complicated series of steps
| 1
|
82,065
| 3,602,416,925
|
IssuesEvent
|
2016-02-03 15:36:06
|
vozdikililer/test
|
https://api.github.com/repos/vozdikililer/test
|
closed
|
issue summary
|
auto-migrated Milestone-2009 Priority-High
|
```
issue description
```
Original issue reported on code.google.com by `Ryan.Sky...@gmail.com` on 4 Jun 2012 at 3:23
|
1.0
|
issue summary - ```
issue description
```
Original issue reported on code.google.com by `Ryan.Sky...@gmail.com` on 4 Jun 2012 at 3:23
|
priority
|
issue summary issue description original issue reported on code google com by ryan sky gmail com on jun at
| 1
|
64,293
| 3,207,159,961
|
IssuesEvent
|
2015-10-05 09:00:45
|
GoldenSoftwareLtd/gedemin
|
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
|
opened
|
GetBaseClassForRelation
|
Priority-High Type-Enhancement
|
Сейчас сканирует все дерево классов. Причем в процессе определения GetDistinctRelation дергается atDatabase для поиска таблицы по ИД (тоже перебор).
|
1.0
|
GetBaseClassForRelation - Сейчас сканирует все дерево классов. Причем в процессе определения GetDistinctRelation дергается atDatabase для поиска таблицы по ИД (тоже перебор).
|
priority
|
getbaseclassforrelation сейчас сканирует все дерево классов причем в процессе определения getdistinctrelation дергается atdatabase для поиска таблицы по ид тоже перебор
| 1
|
237,186
| 7,757,134,197
|
IssuesEvent
|
2018-05-31 15:28:54
|
jonas747/yagpdb
|
https://api.github.com/repos/jonas747/yagpdb
|
closed
|
Voice connection is getting deadlocked somehow
|
Priority: High bug
|
dundundun is this a discordgo bug or is this a jonas bug
|
1.0
|
Voice connection is getting deadlocked somehow - dundundun is this a discordgo bug or is this a jonas bug
|
priority
|
voice connection is getting deadlocked somehow dundundun is this a discordgo bug or is this a jonas bug
| 1
|
184,607
| 6,714,547,957
|
IssuesEvent
|
2017-10-13 17:22:28
|
CS2103AUG2017-W11-B2/main
|
https://api.github.com/repos/CS2103AUG2017-W11-B2/main
|
opened
|
Feature Checklist
|
priority.high type.epic
|
## Things you need to do on every new feature PR:
- [ ] Upate tests
- [ ] Update **developer guide** (your feature implementation)
- [ ] Update user guide *cheatsheet* (if necessary)
- [ ] Update **user guide main function** description
1. Text preface including [optional] important notes, background information
2. Before picture (for new functions only)
3. Instructions (with optional pictures)
4. After (success & errors), state outcome (e.g. `You should see...`)
|
1.0
|
Feature Checklist - ## Things you need to do on every new feature PR:
- [ ] Upate tests
- [ ] Update **developer guide** (your feature implementation)
- [ ] Update user guide *cheatsheet* (if necessary)
- [ ] Update **user guide main function** description
1. Text preface including [optional] important notes, background information
2. Before picture (for new functions only)
3. Instructions (with optional pictures)
4. After (success & errors), state outcome (e.g. `You should see...`)
|
priority
|
feature checklist things you need to do on every new feature pr upate tests update developer guide your feature implementation update user guide cheatsheet if necessary update user guide main function description text preface including important notes background information before picture for new functions only instructions with optional pictures after success errors state outcome e g you should see
| 1
|
813,605
| 30,464,019,583
|
IssuesEvent
|
2023-07-17 09:05:14
|
Avaiga/taipy-gui
|
https://api.github.com/repos/Avaiga/taipy-gui
|
closed
|
BUG- CSS file not discovered by Taipy Enterprise
|
Gui: Back-End 💥Malfunction 🟧 Priority: High
|
**Description**
The CSS file (named main.css) is not discovered by Taipy Enterprise by default but is discovered by Taipy Community. It can have other caveats in Taipy Gui Enterprise.
**How to reproduce**
main.py
```python
from taipy.gui import Gui
data = {"x":[1, 2, 3, 4, 5, 6],
"y":[12, None, 31, 4, 8, 6],}
md = """
<|{data}|chart|x=x|y=y|width=400%|>
"""
Gui(md).run()
```
main.css
```
.taipy-chart {
overflow: auto; /* unset*/
}
```
**Expected behavior**
The CSS file should be discovered and used by Taipy Enterprise. Other caveats related to this issue should be fixed.
**Runtime environment**
Taipy Enterprise 2.3.1
|
1.0
|
BUG- CSS file not discovered by Taipy Enterprise - **Description**
The CSS file (named main.css) is not discovered by Taipy Enterprise by default but is discovered by Taipy Community. It can have other caveats in Taipy Gui Enterprise.
**How to reproduce**
main.py
```python
from taipy.gui import Gui
data = {"x":[1, 2, 3, 4, 5, 6],
"y":[12, None, 31, 4, 8, 6],}
md = """
<|{data}|chart|x=x|y=y|width=400%|>
"""
Gui(md).run()
```
main.css
```
.taipy-chart {
overflow: auto; /* unset*/
}
```
**Expected behavior**
The CSS file should be discovered and used by Taipy Enterprise. Other caveats related to this issue should be fixed.
**Runtime environment**
Taipy Enterprise 2.3.1
|
priority
|
bug css file not discovered by taipy enterprise description the css file named main css is not discovered by taipy enterprise by default but is discovered by taipy community it can have other caveats in taipy gui enterprise how to reproduce main py python from taipy gui import gui data x y md gui md run main css taipy chart overflow auto unset expected behavior the css file should be discovered and used by taipy enterprise other caveats related to this issue should be fixed runtime environment taipy enterprise
| 1
|
275,047
| 8,570,838,209
|
IssuesEvent
|
2018-11-11 23:47:36
|
SIGBlockchain/Project-Allium
|
https://api.github.com/repos/SIGBlockchain/Project-Allium
|
closed
|
New Function: generate_public_key (follow steps carefully)
|
high priority scale
|
**Objective:**
The next key in our key set is the _public key_. If you think of the private key as a password, the public key is sort of like the username. The private key is never to be given away, but there is no harm in making the public key known (hence it is public). The public key is derived from the private key via key cryptography. Because of this they are uniquely linked to each other.
Without going into too much detail about how elliptic curve cryptography works and the mathematics behind it, basically the `ecdsa` module produces two keys relevant to the public key.
The first is the _signing key_, which is what we use to "sign" transactions. Think of it like a wax seal or a pen-signature. It is meant to be interpreted as an authentic binding to someone's identity. It represents an "approval" of something in someone's name. We need this key to sign transactions, and it can be derived from the private key.
The second key is the _verifying key_, which confirms that a given signature was created from a given private key. This is what we'll be returning from the function.
**TODO:**
In `keys.py`, create a function called `generate_public_key`. It takes in a `private_key` as a parameter.
1. The first step is to get the signing key. From the `ecdsa` module, use an instance of `SigningKey` and then use that to call the function `from_string`. `from_string` takes in two parameters: the `private_key` and the result of setting the `curve` to equal `SECP256k1` from the `ecdsa` module:

2. The next step is to get the verifying key. The newly created `signing_key` has a function called `get_verifying_key()`.

3. The final step is to return the byte string version of the verifying key, which _is_ our public key. Our newly created `verifying_key` has a function called `to_string()`, so return the result of that.
**Testing:**
1. Generate a private key, then use that to generate a public key.
2. First assert that it is an instance of bytes.
3. Next, generate a signing key from the private key, just as you did in the function.
4. Hash a random string, doesn't matter what it is. Use the imported `hash_SHA` function from `block.py`.
5. Get a signature by signing the test string. You can sign a string by calling the sign function of the signing key that was generated in step 3.

6. Generate a verifying key from the public key, by created an instance of it from the `ecdsa` module:

7. Verify the signature is valid by calling the `verify` function from the verifying key. The `verify` function takes the `signature` as the first parameter and the test hash string as the second parameter. It returns `True` or `False`. Assert that it is `True`.

Further test coverage: We should also test a situation that returns False from verify, but we can do that in another issue as this one is getting pretty big.
**Disclaimer:**
If you have not pip installed the `ecdsa` module, you will be unable to work on this issue. View [issue #84](https://github.com/SIGBlockchain/Project/issues/84) for guidance on how to get this module.
|
1.0
|
New Function: generate_public_key (follow steps carefully) - **Objective:**
The next key in our key set is the _public key_. If you think of the private key as a password, the public key is sort of like the username. The private key is never to be given away, but there is no harm in making the public key known (hence it is public). The public key is derived from the private key via key cryptography. Because of this they are uniquely linked to each other.
Without going into too much detail about how elliptic curve cryptography works and the mathematics behind it, basically the `ecdsa` module produces two keys relevant to the public key.
The first is the _signing key_, which is what we use to "sign" transactions. Think of it like a wax seal or a pen-signature. It is meant to be interpreted as an authentic binding to someone's identity. It represents an "approval" of something in someone's name. We need this key to sign transactions, and it can be derived from the private key.
The second key is the _verifying key_, which confirms that a given signature was created from a given private key. This is what we'll be returning from the function.
**TODO:**
In `keys.py`, create a function called `generate_public_key`. It takes in a `private_key` as a parameter.
1. The first step is to get the signing key. From the `ecdsa` module, use an instance of `SigningKey` and then use that to call the function `from_string`. `from_string` takes in two parameters: the `private_key` and the result of setting the `curve` to equal `SECP256k1` from the `ecdsa` module:

2. The next step is to get the verifying key. The newly created `signing_key` has a function called `get_verifying_key()`.

3. The final step is to return the byte string version of the verifying key, which _is_ our public key. Our newly created `verifying_key` has a function called `to_string()`, so return the result of that.
**Testing:**
1. Generate a private key, then use that to generate a public key.
2. First assert that it is an instance of bytes.
3. Next, generate a signing key from the private key, just as you did in the function.
4. Hash a random string, doesn't matter what it is. Use the imported `hash_SHA` function from `block.py`.
5. Get a signature by signing the test string. You can sign a string by calling the sign function of the signing key that was generated in step 3.

6. Generate a verifying key from the public key, by created an instance of it from the `ecdsa` module:

7. Verify the signature is valid by calling the `verify` function from the verifying key. The `verify` function takes the `signature` as the first parameter and the test hash string as the second parameter. It returns `True` or `False`. Assert that it is `True`.

Further test coverage: We should also test a situation that returns False from verify, but we can do that in another issue as this one is getting pretty big.
**Disclaimer:**
If you have not pip installed the `ecdsa` module, you will be unable to work on this issue. View [issue #84](https://github.com/SIGBlockchain/Project/issues/84) for guidance on how to get this module.
|
priority
|
new function generate public key follow steps carefully objective the next key in our key set is the public key if you think of the private key as a password the public key is sort of like the username the private key is never to be given away but there is no harm in making the public key known hence it is public the public key is derived from the private key via key cryptography because of this they are uniquely linked to each other without going into too much detail about how elliptic curve cryptography works and the mathematics behind it basically the ecdsa module produces two keys relevant to the public key the first is the signing key which is what we use to sign transactions think of it like a wax seal or a pen signature it is meant to be interpreted as an authentic binding to someone s identity it represents an approval of something in someone s name we need this key to sign transactions and it can be derived from the private key the second key is the verifying key which confirms that a given signature was created from a given private key this is what we ll be returning from the function todo in keys py create a function called generate public key it takes in a private key as a parameter the first step is to get the signing key from the ecdsa module use an instance of signingkey and then use that to call the function from string from string takes in two parameters the private key and the result of setting the curve to equal from the ecdsa module the next step is to get the verifying key the newly created signing key has a function called get verifying key the final step is to return the byte string version of the verifying key which is our public key our newly created verifying key has a function called to string so return the result of that testing generate a private key then use that to generate a public key first assert that it is an instance of bytes next generate a signing key from the private key just as you did in the function hash a random string doesn t matter what it is use the imported hash sha function from block py get a signature by signing the test string you can sign a string by calling the sign function of the signing key that was generated in step generate a verifying key from the public key by created an instance of it from the ecdsa module verify the signature is valid by calling the verify function from the verifying key the verify function takes the signature as the first parameter and the test hash string as the second parameter it returns true or false assert that it is true further test coverage we should also test a situation that returns false from verify but we can do that in another issue as this one is getting pretty big disclaimer if you have not pip installed the ecdsa module you will be unable to work on this issue view for guidance on how to get this module
| 1
|
677,209
| 23,155,015,782
|
IssuesEvent
|
2022-07-29 12:11:19
|
MirageAegis/Sushi-Bot
|
https://api.github.com/repos/MirageAegis/Sushi-Bot
|
opened
|
Moderation commands
|
priority: high
|
**Feature description**
Moderation commands for the bot. These will make it easier for moderators to moderate Discord servers.
**Acceptance criteria**
- Sushi Bot shall have a command for kicking users.
- Sushi Bot shall have a command for banning users.
- Sushi Bot shall have a command for blacklisting users.
- Sushi Bot shall have a command for bulk deleting messages in a Discord channel.
- Sushi Bot shall have a command for changing a user's nickname.
- Sushi Bot shall have a command for displaying details of a user, such as the time at which the user joined the server the command is executed in.
**Additional context**
These commands will be ported from the current Sushi Bot and made into slash commands.
|
1.0
|
Moderation commands - **Feature description**
Moderation commands for the bot. These will make it easier for moderators to moderate Discord servers.
**Acceptance criteria**
- Sushi Bot shall have a command for kicking users.
- Sushi Bot shall have a command for banning users.
- Sushi Bot shall have a command for blacklisting users.
- Sushi Bot shall have a command for bulk deleting messages in a Discord channel.
- Sushi Bot shall have a command for changing a user's nickname.
- Sushi Bot shall have a command for displaying details of a user, such as the time at which the user joined the server the command is executed in.
**Additional context**
These commands will be ported from the current Sushi Bot and made into slash commands.
|
priority
|
moderation commands feature description moderation commands for the bot these will make it easier for moderators to moderate discord servers acceptance criteria sushi bot shall have a command for kicking users sushi bot shall have a command for banning users sushi bot shall have a command for blacklisting users sushi bot shall have a command for bulk deleting messages in a discord channel sushi bot shall have a command for changing a user s nickname sushi bot shall have a command for displaying details of a user such as the time at which the user joined the server the command is executed in additional context these commands will be ported from the current sushi bot and made into slash commands
| 1
|
193,171
| 6,882,202,392
|
IssuesEvent
|
2017-11-21 02:29:29
|
appirio-tech/direct-app
|
https://api.github.com/repos/appirio-tech/direct-app
|
closed
|
Check group permissions when adding challenge resources
|
high priority
|
When you add resources to a challenge that is restricted to a group, it doesn't currently let you do it since the legacy apps (OR, /tc) check user_group_xref instead of the new group api. We need to update the apps to check groups against the new api.
So far, I've noticed this when adding resources via OR and also on the reviewer registration pages.
|
1.0
|
Check group permissions when adding challenge resources - When you add resources to a challenge that is restricted to a group, it doesn't currently let you do it since the legacy apps (OR, /tc) check user_group_xref instead of the new group api. We need to update the apps to check groups against the new api.
So far, I've noticed this when adding resources via OR and also on the reviewer registration pages.
|
priority
|
check group permissions when adding challenge resources when you add resources to a challenge that is restricted to a group it doesn t currently let you do it since the legacy apps or tc check user group xref instead of the new group api we need to update the apps to check groups against the new api so far i ve noticed this when adding resources via or and also on the reviewer registration pages
| 1
|
722,207
| 24,854,683,098
|
IssuesEvent
|
2022-10-27 00:20:51
|
enjoythecode/scrum-wizards-cs321
|
https://api.github.com/repos/enjoythecode/scrum-wizards-cs321
|
closed
|
COACH: View team dashboard
|
high priority @Coach
|
As a coach, I want the collective team fatigue, nutrition, and sleep graphs to be the first things I see when I log on, so that on the go I can get a sense for how the team as a whole is doing and see overall trends.
|
1.0
|
COACH: View team dashboard - As a coach, I want the collective team fatigue, nutrition, and sleep graphs to be the first things I see when I log on, so that on the go I can get a sense for how the team as a whole is doing and see overall trends.
|
priority
|
coach view team dashboard as a coach i want the collective team fatigue nutrition and sleep graphs to be the first things i see when i log on so that on the go i can get a sense for how the team as a whole is doing and see overall trends
| 1
|
593,048
| 17,936,776,001
|
IssuesEvent
|
2021-09-10 16:18:48
|
cyntaria/UniPal-Backend
|
https://api.github.com/repos/cyntaria/UniPal-Backend
|
reopened
|
[PATCH] A Activity Type
|
Status: Completed Priority: High user story Type: Feature
|
### Summary
As an `admin`, I should be able to **update the details of an activity type**, so that I can **fix old or inconsistent entries**.
### Acceptance Criteria
**GIVEN** an `admin` is *editing an activity type* in the app
**WHEN** the app hits the `/activity-types/:id` endpoint with a valid PATCH request, containing:-
The path parameter:
- `:id`, the unique id of the entity of which the details are edited.
And any of the following body parameters:
- activity type
**THEN** the app should receive a status `200`
**AND** in the response, the following information should be returned:
- header message indicating update operation success
- rows matched
- rows changed
Sample Request/Sample Response
```
headers: {
error: 0,
message: "The specified item was updated successfully"
}
body: {
rows_matched: 1,
rows_changed: 1,
info: "..."
}
```
### Resources
- Development URL: {Here goes a URL to the feature on development API}
- Production URL: {Here goes a URL to the feature on production API}
### Dev Notes
{Some complementary notes if necessary}
### Testing Notes
#### Scenario 1: PATCH request is successful
1. Update an activity type with a **PATCH** request to `/activity-types/:id` endpoint
2. A subsequent **GET** request to `/activity-types/:id` endpoint should return a status code `200`
3. And the activity type details with the updated information i.e. matching the initially sent body.
4. Resubmit a **PATCH** request to `/activity-types/:id` endpoint to reverse the change and ensure status code `200` is returned.
#### Scenario 2: PATCH request is unsuccessful
1. Update a activity type with a **PATCH** request to `/activity-types/:id` endpoint containing a non-existent `activity_type_id`.
2. Ensure a `404` status code is returned.
3. And the response headers' `code` parameter should contain "**_NotFoundException_**".
#### Scenario 3: PATCH request is incorrect
1. Send a **PATCH** request to `/activity-types/:id` endpoint with an incorrect key name in the body
2. Ensure a `422` status code is returned
3. And the response headers' `code` parameter should contain "**_InvalidPropertiesException_**".
4. And the response headers' `data` parameter should contain the name of the invalid parameter.
#### Scenario 4: PATCH request is forbidden
1. Send a **PATCH** request to `/activity-types/:id` endpoint using a `student` account token.
2. Ensure the response returns a `403` forbidden status code.
3. And the response headers' `code` parameter should contain "**_ForbiddenException_**"
#### Scenario 5: PATCH request is unauthorized
1. Send a **PATCH** request to `/activity-types/:id` endpoint without an **authorization token**
2. Ensure a `401` unauthorized status code is returned.
3. And the response headers' `code` parameter should contain "**_TokenMissingException_**"
|
1.0
|
[PATCH] A Activity Type - ### Summary
As an `admin`, I should be able to **update the details of an activity type**, so that I can **fix old or inconsistent entries**.
### Acceptance Criteria
**GIVEN** an `admin` is *editing an activity type* in the app
**WHEN** the app hits the `/activity-types/:id` endpoint with a valid PATCH request, containing:-
The path parameter:
- `:id`, the unique id of the entity of which the details are edited.
And any of the following body parameters:
- activity type
**THEN** the app should receive a status `200`
**AND** in the response, the following information should be returned:
- header message indicating update operation success
- rows matched
- rows changed
Sample Request/Sample Response
```
headers: {
error: 0,
message: "The specified item was updated successfully"
}
body: {
rows_matched: 1,
rows_changed: 1,
info: "..."
}
```
### Resources
- Development URL: {Here goes a URL to the feature on development API}
- Production URL: {Here goes a URL to the feature on production API}
### Dev Notes
{Some complementary notes if necessary}
### Testing Notes
#### Scenario 1: PATCH request is successful
1. Update an activity type with a **PATCH** request to `/activity-types/:id` endpoint
2. A subsequent **GET** request to `/activity-types/:id` endpoint should return a status code `200`
3. And the activity type details with the updated information i.e. matching the initially sent body.
4. Resubmit a **PATCH** request to `/activity-types/:id` endpoint to reverse the change and ensure status code `200` is returned.
#### Scenario 2: PATCH request is unsuccessful
1. Update a activity type with a **PATCH** request to `/activity-types/:id` endpoint containing a non-existent `activity_type_id`.
2. Ensure a `404` status code is returned.
3. And the response headers' `code` parameter should contain "**_NotFoundException_**".
#### Scenario 3: PATCH request is incorrect
1. Send a **PATCH** request to `/activity-types/:id` endpoint with an incorrect key name in the body
2. Ensure a `422` status code is returned
3. And the response headers' `code` parameter should contain "**_InvalidPropertiesException_**".
4. And the response headers' `data` parameter should contain the name of the invalid parameter.
#### Scenario 4: PATCH request is forbidden
1. Send a **PATCH** request to `/activity-types/:id` endpoint using a `student` account token.
2. Ensure the response returns a `403` forbidden status code.
3. And the response headers' `code` parameter should contain "**_ForbiddenException_**"
#### Scenario 5: PATCH request is unauthorized
1. Send a **PATCH** request to `/activity-types/:id` endpoint without an **authorization token**
2. Ensure a `401` unauthorized status code is returned.
3. And the response headers' `code` parameter should contain "**_TokenMissingException_**"
|
priority
|
a activity type summary as an admin i should be able to update the details of an activity type so that i can fix old or inconsistent entries acceptance criteria given an admin is editing an activity type in the app when the app hits the activity types id endpoint with a valid patch request containing the path parameter id the unique id of the entity of which the details are edited and any of the following body parameters activity type then the app should receive a status and in the response the following information should be returned header message indicating update operation success rows matched rows changed sample request sample response headers error message the specified item was updated successfully body rows matched rows changed info resources development url here goes a url to the feature on development api production url here goes a url to the feature on production api dev notes some complementary notes if necessary testing notes scenario patch request is successful update an activity type with a patch request to activity types id endpoint a subsequent get request to activity types id endpoint should return a status code and the activity type details with the updated information i e matching the initially sent body resubmit a patch request to activity types id endpoint to reverse the change and ensure status code is returned scenario patch request is unsuccessful update a activity type with a patch request to activity types id endpoint containing a non existent activity type id ensure a status code is returned and the response headers code parameter should contain notfoundexception scenario patch request is incorrect send a patch request to activity types id endpoint with an incorrect key name in the body ensure a status code is returned and the response headers code parameter should contain invalidpropertiesexception and the response headers data parameter should contain the name of the invalid parameter scenario patch request is forbidden send a patch request to activity types id endpoint using a student account token ensure the response returns a forbidden status code and the response headers code parameter should contain forbiddenexception scenario patch request is unauthorized send a patch request to activity types id endpoint without an authorization token ensure a unauthorized status code is returned and the response headers code parameter should contain tokenmissingexception
| 1
|
21,497
| 2,641,176,541
|
IssuesEvent
|
2015-03-11 16:24:02
|
ChristopheBoucaut/cb-angular-utils
|
https://api.github.com/repos/ChristopheBoucaut/cb-angular-utils
|
closed
|
Refacto modules names and directories.
|
high priority
|
More modules (one by features) and change directories names to name feature directory
|
1.0
|
Refacto modules names and directories. - More modules (one by features) and change directories names to name feature directory
|
priority
|
refacto modules names and directories more modules one by features and change directories names to name feature directory
| 1
|
476,788
| 13,750,223,652
|
IssuesEvent
|
2020-10-06 11:43:33
|
scandipwa/base-theme
|
https://api.github.com/repos/scandipwa/base-theme
|
closed
|
Auto-scrolling to the start on the massive category pages
|
Core High Priority Type: bug Waiting Estimate
|
**Description**:
While scrolling down the category page with a lot of products, it automatically scrolls back to the start after certain break-point.
**Expected behavior**:
Controlled scrolling
**To Reproduce**:
1. Go to any category page where a lot of products are exist, i.e. [demo scandipwa cabels category](https://demo.scandipwa.com/computers-peripherals-accessories.html?customFilters=category_id:198) and choose mobile view
2. Try to scroll to the end
3. It gets you back to the start
**Versions**:
- ScandiPWA: 3.1.1
|
1.0
|
Auto-scrolling to the start on the massive category pages - **Description**:
While scrolling down the category page with a lot of products, it automatically scrolls back to the start after certain break-point.
**Expected behavior**:
Controlled scrolling
**To Reproduce**:
1. Go to any category page where a lot of products are exist, i.e. [demo scandipwa cabels category](https://demo.scandipwa.com/computers-peripherals-accessories.html?customFilters=category_id:198) and choose mobile view
2. Try to scroll to the end
3. It gets you back to the start
**Versions**:
- ScandiPWA: 3.1.1
|
priority
|
auto scrolling to the start on the massive category pages description while scrolling down the category page with a lot of products it automatically scrolls back to the start after certain break point expected behavior controlled scrolling to reproduce go to any category page where a lot of products are exist i e and choose mobile view try to scroll to the end it gets you back to the start versions scandipwa
| 1
|
148,240
| 5,661,413,038
|
IssuesEvent
|
2017-04-10 17:18:22
|
elementary/camera
|
https://api.github.com/repos/elementary/camera
|
opened
|
Generate po files for extra/pot
|
Priority: High
|
We can't upload these translations to weblate until we first generate the missing po files
|
1.0
|
Generate po files for extra/pot - We can't upload these translations to weblate until we first generate the missing po files
|
priority
|
generate po files for extra pot we can t upload these translations to weblate until we first generate the missing po files
| 1
|
769,385
| 27,003,953,303
|
IssuesEvent
|
2023-02-10 10:07:45
|
Avaiga/taipy-studio-config
|
https://api.github.com/repos/Avaiga/taipy-studio-config
|
closed
|
Bug: selecting function for Task in details
|
💥Malfunction 🟧 Priority: High
|
When selecting a module + function in the details Task, the value added to TOML contains the module twice.
+ add the filename
+ select file in config TreeView on showPerspective if no selection
|
1.0
|
Bug: selecting function for Task in details - When selecting a module + function in the details Task, the value added to TOML contains the module twice.
+ add the filename
+ select file in config TreeView on showPerspective if no selection
|
priority
|
bug selecting function for task in details when selecting a module function in the details task the value added to toml contains the module twice add the filename select file in config treeview on showperspective if no selection
| 1
|
7,074
| 2,597,070,071
|
IssuesEvent
|
2015-02-21 02:19:44
|
jasonsanjose/brackets-sass
|
https://api.github.com/repos/jasonsanjose/brackets-sass
|
closed
|
Add Ruby sass compiler option
|
high priority
|
If the user has the `sass` command line tool installed, allow switching the compiler from the libsass compiler packaged with this extension.
|
1.0
|
Add Ruby sass compiler option - If the user has the `sass` command line tool installed, allow switching the compiler from the libsass compiler packaged with this extension.
|
priority
|
add ruby sass compiler option if the user has the sass command line tool installed allow switching the compiler from the libsass compiler packaged with this extension
| 1
|
405,854
| 11,883,583,216
|
IssuesEvent
|
2020-03-27 16:11:09
|
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
|
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
|
closed
|
Weird holding names
|
:beetle: bug - localisation :scroll: :exclamation: priority high
|
**Mod Version**
37d54aac5
**Are you using any submods/mods? If so, which?**
GoA
**Please explain your issue in as much detail as possible:**
? instead of actual symbol
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details>
|
1.0
|
Weird holding names - **Mod Version**
37d54aac5
**Are you using any submods/mods? If so, which?**
GoA
**Please explain your issue in as much detail as possible:**
? instead of actual symbol
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details>
|
priority
|
weird holding names mod version are you using any submods mods if so which goa please explain your issue in as much detail as possible instead of actual symbol upload screenshots of the problem localization click to expand
| 1
|
753,462
| 26,347,626,307
|
IssuesEvent
|
2023-01-11 00:07:42
|
gamefreedomgit/Maelstrom
|
https://api.github.com/repos/gamefreedomgit/Maelstrom
|
opened
|
[NPC] [Cata][Quest] Undying Twilight
|
NPC Quest - Cataclysm (80+) Quest - Event Priority: High Status: Confirmed
|
**How to reproduce:** npc for quest Undying Twilight they just fell through ground once i tried to fight them, event related to it was not scripted at all, no friendly npcs nothing..


**How it should work:** they should not fell through the ground and event should work properly like this video
https://youtu.be/DMABtRpgkMA?t=74
|
1.0
|
[NPC] [Cata][Quest] Undying Twilight -
**How to reproduce:** npc for quest Undying Twilight they just fell through ground once i tried to fight them, event related to it was not scripted at all, no friendly npcs nothing..


**How it should work:** they should not fell through the ground and event should work properly like this video
https://youtu.be/DMABtRpgkMA?t=74
|
priority
|
undying twilight how to reproduce npc for quest undying twilight they just fell through ground once i tried to fight them event related to it was not scripted at all no friendly npcs nothing how it should work they should not fell through the ground and event should work properly like this video
| 1
|
375,665
| 11,115,047,776
|
IssuesEvent
|
2019-12-18 09:57:11
|
openmsupply/mobile
|
https://api.github.com/repos/openmsupply/mobile
|
opened
|
SimpleLabel component
|
Docs: not needed Effort: small Feature Module: dispensary Priority: high
|
## Is your feature request related to a problem? Please describe.
Various places in the prescription form need a label
## Describe the solution you'd like
A simple display component
## Implementation
Simple component taking some text and possibly style parameters for size etc.
## Describe alternatives you've considered
NA
## Additional context
N/A
|
1.0
|
SimpleLabel component - ## Is your feature request related to a problem? Please describe.
Various places in the prescription form need a label
## Describe the solution you'd like
A simple display component
## Implementation
Simple component taking some text and possibly style parameters for size etc.
## Describe alternatives you've considered
NA
## Additional context
N/A
|
priority
|
simplelabel component is your feature request related to a problem please describe various places in the prescription form need a label describe the solution you d like a simple display component implementation simple component taking some text and possibly style parameters for size etc describe alternatives you ve considered na additional context n a
| 1
|
587,206
| 17,607,134,188
|
IssuesEvent
|
2021-08-17 18:38:36
|
CO2-Abgabe/massnahmenpakete
|
https://api.github.com/repos/CO2-Abgabe/massnahmenpakete
|
closed
|
Allow filtering results by state
|
enhancement high priority
|
die neue Tabelle ist in der Produktionsumgebung angelegt.
Abfrage per API wie folgt:
$optionValues = civicrm_api4('OptionValue', 'get', [
'select' => [
'name',
'value',
],
'where' => [
['option_group_id', '=', 112],
],
'limit' => 25,
]);
Wobei: name=Wahlkreisnr., value=Name des Bundeslands
|
1.0
|
Allow filtering results by state - die neue Tabelle ist in der Produktionsumgebung angelegt.
Abfrage per API wie folgt:
$optionValues = civicrm_api4('OptionValue', 'get', [
'select' => [
'name',
'value',
],
'where' => [
['option_group_id', '=', 112],
],
'limit' => 25,
]);
Wobei: name=Wahlkreisnr., value=Name des Bundeslands
|
priority
|
allow filtering results by state die neue tabelle ist in der produktionsumgebung angelegt abfrage per api wie folgt optionvalues civicrm optionvalue get select name value where limit wobei name wahlkreisnr value name des bundeslands
| 1
|
691,165
| 23,685,627,494
|
IssuesEvent
|
2022-08-29 05:54:07
|
aulasoftwarelibre/Rx-Flappy
|
https://api.github.com/repos/aulasoftwarelibre/Rx-Flappy
|
closed
|
Houston, Houston, we have a font license problem here!
|
help wanted High priority
|
@p52safer, @mbritTech!
We need to check the app's **fonts licenses** before continue.
Can we trace them?
|
1.0
|
Houston, Houston, we have a font license problem here! - @p52safer, @mbritTech!
We need to check the app's **fonts licenses** before continue.
Can we trace them?
|
priority
|
houston houston we have a font license problem here mbrittech we need to check the app s fonts licenses before continue can we trace them
| 1
|
197,391
| 6,954,943,499
|
IssuesEvent
|
2017-12-07 04:35:44
|
TwidereProject/Twidere-Android
|
https://api.github.com/repos/TwidereProject/Twidere-Android
|
closed
|
Error logging in to Fanfou
|
category:functionality platform:fanfou priority:P1:high type:bug
|
When constructing OAuth sign URL, scheme should be `http` instead of `https`
|
1.0
|
Error logging in to Fanfou - When constructing OAuth sign URL, scheme should be `http` instead of `https`
|
priority
|
error logging in to fanfou when constructing oauth sign url scheme should be http instead of https
| 1
|
367,366
| 10,852,544,327
|
IssuesEvent
|
2019-11-13 13:01:56
|
project-koku/koku
|
https://api.github.com/repos/project-koku/koku
|
closed
|
Fix Report Download Retry logic
|
Koku'(x) bug priority - high
|
**Describe the bug**
Our retry logic for downloading and processing tasks currently isn't catching all jobs that need to retry and is possibly retrying jobs that haven't even started yet.
See https://github.com/project-koku/koku/blob/master/koku/masu/external/downloader/report_downloader_base.py#L65 for reference code.
In particular the logic here: https://github.com/project-koku/koku/blob/master/koku/masu/external/downloader/report_downloader_base.py#L92-L97 does not also include handling for the case when no completed datetime exists.
## Proposed Solution -- Needs to be checked for feasibility
1) Store the celery async task ID for download/process in the `CostUsageReportManifest` model/table. See (https://celery.readthedocs.io/en/latest/userguide/tasks.html#bound-tasks) for getting task id
2) Use `from celery.task.control import inspect` to see if the task ID is scheduled, active, or reserved. If so don't retry.
~2) Set this celery option to true http://docs.celeryproject.org/en/latest/userguide/configuration.html#task-track-started~
~3) Add logic to `check_if_manifest_should_be_downloaded` to check the status of the celery task. If it is still running don't retry. This should be the first check before any of the other logic.~
4) Add logic to `check_if_manifest_should_be_downloaded` for when the celery task is no longer running, and there are entries in `CostUsageReportStatus` that have a `last_started_datetime` but no `last_completed_datetime` and need to be retried.
|
1.0
|
Fix Report Download Retry logic - **Describe the bug**
Our retry logic for downloading and processing tasks currently isn't catching all jobs that need to retry and is possibly retrying jobs that haven't even started yet.
See https://github.com/project-koku/koku/blob/master/koku/masu/external/downloader/report_downloader_base.py#L65 for reference code.
In particular the logic here: https://github.com/project-koku/koku/blob/master/koku/masu/external/downloader/report_downloader_base.py#L92-L97 does not also include handling for the case when no completed datetime exists.
## Proposed Solution -- Needs to be checked for feasibility
1) Store the celery async task ID for download/process in the `CostUsageReportManifest` model/table. See (https://celery.readthedocs.io/en/latest/userguide/tasks.html#bound-tasks) for getting task id
2) Use `from celery.task.control import inspect` to see if the task ID is scheduled, active, or reserved. If so don't retry.
~2) Set this celery option to true http://docs.celeryproject.org/en/latest/userguide/configuration.html#task-track-started~
~3) Add logic to `check_if_manifest_should_be_downloaded` to check the status of the celery task. If it is still running don't retry. This should be the first check before any of the other logic.~
4) Add logic to `check_if_manifest_should_be_downloaded` for when the celery task is no longer running, and there are entries in `CostUsageReportStatus` that have a `last_started_datetime` but no `last_completed_datetime` and need to be retried.
|
priority
|
fix report download retry logic describe the bug our retry logic for downloading and processing tasks currently isn t catching all jobs that need to retry and is possibly retrying jobs that haven t even started yet see for reference code in particular the logic here does not also include handling for the case when no completed datetime exists proposed solution needs to be checked for feasibility store the celery async task id for download process in the costusagereportmanifest model table see for getting task id use from celery task control import inspect to see if the task id is scheduled active or reserved if so don t retry set this celery option to true add logic to check if manifest should be downloaded to check the status of the celery task if it is still running don t retry this should be the first check before any of the other logic add logic to check if manifest should be downloaded for when the celery task is no longer running and there are entries in costusagereportstatus that have a last started datetime but no last completed datetime and need to be retried
| 1
|
108,589
| 4,347,832,127
|
IssuesEvent
|
2016-07-29 21:00:24
|
mantidproject/mantid
|
https://api.github.com/repos/mantidproject/mantid
|
opened
|
Out-of-bounds memory access in CompactMDTest
|
Component: Framework Misc: Bug Priority: High
|
<!-- TEMPLATE FOR BUG REPORTS -->
### Expected behavior
run `MDAlgorithmsTest_CompactMDTest` without out-of-bounds memory access.
### Actual behavior
MDHistoWorkspaceIterator::jumpTo is being passed a very large and out of bounds index. I traced it back to MDHistoWorkspaceIterator::jumpToNearest, which in one case returns a negative value.
I also get the warning below. Not sure if it is relevant.
```
IntegrateMDHistoWorkspace-[Warning] Rounding min from: -4 to the nearest whole width at: -5
IntegrateMDHistoWorkspace-[Warning] Rounding max from: 4 to the nearest whole width at: 5
```
I believe there is a logic error in either `IntegrateMDHistoWorkspace` or `CompactMD`, but having trouble tracking it down.
### Steps to reproduce the behavior
run MDAlgorithmsTest_CompactMDTest while checking [here](https://github.com/mantidproject/mantid/blob/master/Framework/DataObjects/src/MDHistoWorkspaceIterator.cpp#L279) if `dExact` < 0.0.
### Platforms affected
All
|
1.0
|
Out-of-bounds memory access in CompactMDTest - <!-- TEMPLATE FOR BUG REPORTS -->
### Expected behavior
run `MDAlgorithmsTest_CompactMDTest` without out-of-bounds memory access.
### Actual behavior
MDHistoWorkspaceIterator::jumpTo is being passed a very large and out of bounds index. I traced it back to MDHistoWorkspaceIterator::jumpToNearest, which in one case returns a negative value.
I also get the warning below. Not sure if it is relevant.
```
IntegrateMDHistoWorkspace-[Warning] Rounding min from: -4 to the nearest whole width at: -5
IntegrateMDHistoWorkspace-[Warning] Rounding max from: 4 to the nearest whole width at: 5
```
I believe there is a logic error in either `IntegrateMDHistoWorkspace` or `CompactMD`, but having trouble tracking it down.
### Steps to reproduce the behavior
run MDAlgorithmsTest_CompactMDTest while checking [here](https://github.com/mantidproject/mantid/blob/master/Framework/DataObjects/src/MDHistoWorkspaceIterator.cpp#L279) if `dExact` < 0.0.
### Platforms affected
All
|
priority
|
out of bounds memory access in compactmdtest expected behavior run mdalgorithmstest compactmdtest without out of bounds memory access actual behavior mdhistoworkspaceiterator jumpto is being passed a very large and out of bounds index i traced it back to mdhistoworkspaceiterator jumptonearest which in one case returns a negative value i also get the warning below not sure if it is relevant integratemdhistoworkspace rounding min from to the nearest whole width at integratemdhistoworkspace rounding max from to the nearest whole width at i believe there is a logic error in either integratemdhistoworkspace or compactmd but having trouble tracking it down steps to reproduce the behavior run mdalgorithmstest compactmdtest while checking if dexact platforms affected all
| 1
|
695,026
| 23,840,765,631
|
IssuesEvent
|
2022-09-06 10:01:28
|
pika-org/pika
|
https://api.github.com/repos/pika-org/pika
|
opened
|
Reduce debug bloat
|
effort: 3 effort: 4 priority: high category: senders/receivers type: cleanup
|
This is not ok:
```
> ls -lh test/unit/eigensolver/test_tridiag_solver_full
-rwxr-xr-x 1 simbergm csstaff 1.5G Sep 6 11:49 test/unit/eigensolver/test_tridiag_solver_full
```
|
1.0
|
Reduce debug bloat - This is not ok:
```
> ls -lh test/unit/eigensolver/test_tridiag_solver_full
-rwxr-xr-x 1 simbergm csstaff 1.5G Sep 6 11:49 test/unit/eigensolver/test_tridiag_solver_full
```
|
priority
|
reduce debug bloat this is not ok ls lh test unit eigensolver test tridiag solver full rwxr xr x simbergm csstaff sep test unit eigensolver test tridiag solver full
| 1
|
262,194
| 8,256,627,347
|
IssuesEvent
|
2018-09-13 00:03:49
|
ClinGen/clincoded
|
https://api.github.com/repos/ClinGen/clincoded
|
closed
|
Apache Restarting process. Memory usage exceeds limit
|
R22 bug in progress priority: high
|
We are receiving errors from Apache logs stating:
`ERROR [clincoded.memlimit][Dummy-1] Restarting process. Memory usage exceeds limit of 524288000: 728248320`
What this is, is a cleanup script that is killing threads in apache/mod_wsgi/python that run over the memlimit (500mb), BUT it will end those threads only AFTER finishing the request. So this is typically fine, what we will be doing though is matching these errors in the log with requests from the apache log and see if there is any correlation.
|
1.0
|
Apache Restarting process. Memory usage exceeds limit - We are receiving errors from Apache logs stating:
`ERROR [clincoded.memlimit][Dummy-1] Restarting process. Memory usage exceeds limit of 524288000: 728248320`
What this is, is a cleanup script that is killing threads in apache/mod_wsgi/python that run over the memlimit (500mb), BUT it will end those threads only AFTER finishing the request. So this is typically fine, what we will be doing though is matching these errors in the log with requests from the apache log and see if there is any correlation.
|
priority
|
apache restarting process memory usage exceeds limit we are receiving errors from apache logs stating error restarting process memory usage exceeds limit of what this is is a cleanup script that is killing threads in apache mod wsgi python that run over the memlimit but it will end those threads only after finishing the request so this is typically fine what we will be doing though is matching these errors in the log with requests from the apache log and see if there is any correlation
| 1
|
333,588
| 10,128,741,846
|
IssuesEvent
|
2019-08-01 13:26:32
|
bbc/psammead
|
https://api.github.com/repos/bbc/psammead
|
closed
|
Code Coverage neither reporting nor a status check
|
high priority infra-tooling-stream
|
**Is your feature request related to a problem? Please describe.**
In [this file](https://github.com/bbc/psammead/blob/latest/Makefile) we are reporting code coverage but it's never reaching [CodeClimate](https://codeclimate.com/github/bbc/psammead), because of this we're unable to have status checks on the code coverage of a PR or on the repo overall.
**Describe the solution you'd like**
- Configure the code coverage reporting to report to CodeClimate
- Add status checks for code coverage on PRs and overall
**Describe alternatives you've considered**
None, this is a no-brainer
**Testing notes**
[Tester to complete]
Dev insight: This is to add tests, no tests on the tests would be beneficial
**Additional context**
Add any other context or screenshots about the feature request here.
- [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/psammead/labels/Refinement%20Needed)
|
1.0
|
Code Coverage neither reporting nor a status check - **Is your feature request related to a problem? Please describe.**
In [this file](https://github.com/bbc/psammead/blob/latest/Makefile) we are reporting code coverage but it's never reaching [CodeClimate](https://codeclimate.com/github/bbc/psammead), because of this we're unable to have status checks on the code coverage of a PR or on the repo overall.
**Describe the solution you'd like**
- Configure the code coverage reporting to report to CodeClimate
- Add status checks for code coverage on PRs and overall
**Describe alternatives you've considered**
None, this is a no-brainer
**Testing notes**
[Tester to complete]
Dev insight: This is to add tests, no tests on the tests would be beneficial
**Additional context**
Add any other context or screenshots about the feature request here.
- [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/psammead/labels/Refinement%20Needed)
|
priority
|
code coverage neither reporting nor a status check is your feature request related to a problem please describe in we are reporting code coverage but it s never reaching because of this we re unable to have status checks on the code coverage of a pr or on the repo overall describe the solution you d like configure the code coverage reporting to report to codeclimate add status checks for code coverage on prs and overall describe alternatives you ve considered none this is a no brainer testing notes dev insight this is to add tests no tests on the tests would be beneficial additional context add any other context or screenshots about the feature request here initially labelled with
| 1
|
167,909
| 6,348,817,268
|
IssuesEvent
|
2017-07-28 11:03:55
|
dbcollection/dbcollection
|
https://api.github.com/repos/dbcollection/dbcollection
|
opened
|
Fix 'default' task name not returning the true name
|
api bug high priority
|
When using the `default` task tag for loading a dataset, it does not return the actual task name. This causes confusion when trying to understand what task does the loaded object contain.
|
1.0
|
Fix 'default' task name not returning the true name - When using the `default` task tag for loading a dataset, it does not return the actual task name. This causes confusion when trying to understand what task does the loaded object contain.
|
priority
|
fix default task name not returning the true name when using the default task tag for loading a dataset it does not return the actual task name this causes confusion when trying to understand what task does the loaded object contain
| 1
|
400,021
| 11,765,732,597
|
IssuesEvent
|
2020-03-14 18:45:48
|
Benjamin-Bichel/GAMZ
|
https://api.github.com/repos/Benjamin-Bichel/GAMZ
|
closed
|
Validation of Travis with Git
|
high-priority
|
Changing out CI and CD from app center as that requires one to be utilizing the MS visual studio or another microsoft IDE
|
1.0
|
Validation of Travis with Git - Changing out CI and CD from app center as that requires one to be utilizing the MS visual studio or another microsoft IDE
|
priority
|
validation of travis with git changing out ci and cd from app center as that requires one to be utilizing the ms visual studio or another microsoft ide
| 1
|
540,182
| 15,802,192,342
|
IssuesEvent
|
2021-04-03 08:31:21
|
bryntum/support
|
https://api.github.com/repos/bryntum/support
|
closed
|
Selecting one record highlights two elements on specific screen size
|
Runtime Error bug high-priority premium resolved
|
[Reported here](https://www.bryntum.com/forum/viewtopic.php?p=79726#p79726)
[Open in Rootcause](https://app.therootcause.io/45eacc19548ea67060cf0a2c9849108a20341883)
http://lh/bryntum-suite/gantt/examples/advanced/
Set browser window to 1366 x 657. Window height is important
Select 1.2.1 task => see both 1.2.1 and 1.4.2.1 are highlighted
Select 1.2.2 task => see both 1.2.2 and 1.4.2.2 are highlighted
Select 1.2.3 task => see both 1.2.3 and 1.4.2.3 are highlighted
|
1.0
|
Selecting one record highlights two elements on specific screen size - [Reported here](https://www.bryntum.com/forum/viewtopic.php?p=79726#p79726)
[Open in Rootcause](https://app.therootcause.io/45eacc19548ea67060cf0a2c9849108a20341883)
http://lh/bryntum-suite/gantt/examples/advanced/
Set browser window to 1366 x 657. Window height is important
Select 1.2.1 task => see both 1.2.1 and 1.4.2.1 are highlighted
Select 1.2.2 task => see both 1.2.2 and 1.4.2.2 are highlighted
Select 1.2.3 task => see both 1.2.3 and 1.4.2.3 are highlighted
|
priority
|
selecting one record highlights two elements on specific screen size set browser window to x window height is important select task see both and are highlighted select task see both and are highlighted select task see both and are highlighted
| 1
|
674,950
| 23,071,779,626
|
IssuesEvent
|
2022-07-25 18:48:42
|
bennyboer/table-engine
|
https://api.github.com/repos/bennyboer/table-engine
|
closed
|
Dragging cell selection in scrollable page will scroll the page
|
bug high priority
|
- Drag cell selection in scrollable page
- page will scroll, but should not
|
1.0
|
Dragging cell selection in scrollable page will scroll the page - - Drag cell selection in scrollable page
- page will scroll, but should not
|
priority
|
dragging cell selection in scrollable page will scroll the page drag cell selection in scrollable page page will scroll but should not
| 1
|
486,254
| 14,006,351,187
|
IssuesEvent
|
2020-10-28 19:49:16
|
ooni/explorer
|
https://api.github.com/repos/ooni/explorer
|
closed
|
Update MK specific measurement details label
|
bug effort/S priority/high
|
As reported by @bassosimone
## Expected Behavior
The engine behind ooniprobe products for a few months has been the new `probe-engine`. For older probes, this used to be 'measurement-kit'. Measurement pages should specify which engine produced the measurement and the corresponding version number.
## Actual Behavior
Explorer currently shows a MK specific label even for `probe-engine` measurements. e.g in [this measurement](https://explorer.ooni.org/measurement/20200807T222018Z_AS33363_FTKkox83LnAxEXkK8hJQuuqzBjZu2nyTl87aLXQ7MbCPvZIEgW)

## Steps to Reproduce the Problem
1. Open any measurement page.
2. Scroll to the section right before the raw measurement data
|
1.0
|
Update MK specific measurement details label - As reported by @bassosimone
## Expected Behavior
The engine behind ooniprobe products for a few months has been the new `probe-engine`. For older probes, this used to be 'measurement-kit'. Measurement pages should specify which engine produced the measurement and the corresponding version number.
## Actual Behavior
Explorer currently shows a MK specific label even for `probe-engine` measurements. e.g in [this measurement](https://explorer.ooni.org/measurement/20200807T222018Z_AS33363_FTKkox83LnAxEXkK8hJQuuqzBjZu2nyTl87aLXQ7MbCPvZIEgW)

## Steps to Reproduce the Problem
1. Open any measurement page.
2. Scroll to the section right before the raw measurement data
|
priority
|
update mk specific measurement details label as reported by bassosimone expected behavior the engine behind ooniprobe products for a few months has been the new probe engine for older probes this used to be measurement kit measurement pages should specify which engine produced the measurement and the corresponding version number actual behavior explorer currently shows a mk specific label even for probe engine measurements e g in steps to reproduce the problem open any measurement page scroll to the section right before the raw measurement data
| 1
|
137,438
| 5,309,805,085
|
IssuesEvent
|
2017-02-12 14:31:07
|
UoS-SNe/CoCo
|
https://api.github.com/repos/UoS-SNe/CoCo
|
closed
|
Phases are not adjusted by the scale factor
|
bug high priority
|
Currently phases are saved as the difference between the MJD at any point and the MJD of the light curve peak. This needs to be adjusted by the scale factor.
|
1.0
|
Phases are not adjusted by the scale factor - Currently phases are saved as the difference between the MJD at any point and the MJD of the light curve peak. This needs to be adjusted by the scale factor.
|
priority
|
phases are not adjusted by the scale factor currently phases are saved as the difference between the mjd at any point and the mjd of the light curve peak this needs to be adjusted by the scale factor
| 1
|
581,341
| 17,291,543,408
|
IssuesEvent
|
2021-07-24 21:32:20
|
LesFruitsDefendus/saskatoon-ng
|
https://api.github.com/repos/LesFruitsDefendus/saskatoon-ng
|
closed
|
Add forget password feature to current login form
|
high priority
|
Add forget password feature to current login form
|
1.0
|
Add forget password feature to current login form - Add forget password feature to current login form
|
priority
|
add forget password feature to current login form add forget password feature to current login form
| 1
|
527,588
| 15,345,246,195
|
IssuesEvent
|
2021-02-28 05:59:21
|
parkourtheory/datapipe
|
https://api.github.com/repos/parkourtheory/datapipe
|
opened
|
ExtrapolationMask task
|
focus high priority medium
|
How does dgl map node names to ids when converting from networks?
Left notes in comments for two TODOs
|
1.0
|
ExtrapolationMask task - How does dgl map node names to ids when converting from networks?
Left notes in comments for two TODOs
|
priority
|
extrapolationmask task how does dgl map node names to ids when converting from networks left notes in comments for two todos
| 1
|
702,791
| 24,136,252,491
|
IssuesEvent
|
2022-09-21 11:33:38
|
wp-media/wp-rocket
|
https://api.github.com/repos/wp-media/wp-rocket
|
closed
|
Improve detection of missing AS tables
|
type: bug module: preload priority: high effort: [S] severity: major module: remove unused css
|
**Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
In the current approach we have a bug related to the not expected characters in the database name. Our query here is not using `` causing fail of the check and displaying false message:
https://github.com/wp-media/wp-rocket/blob/ba69c9365ac25fe9eb3c09d6910799bdfa13e225/inc/Engine/Preload/Activation/Activation.php#L117
**Expected behavior**
We should check the existence of the tables only in the specific conditions:
1. When activating the plugin, before Preload do any operation on the tables
2. When enabling RUCSS
3. When enabling Preload
4. During each update of the plugin
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
1.0
|
Improve detection of missing AS tables - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
In the current approach we have a bug related to the not expected characters in the database name. Our query here is not using `` causing fail of the check and displaying false message:
https://github.com/wp-media/wp-rocket/blob/ba69c9365ac25fe9eb3c09d6910799bdfa13e225/inc/Engine/Preload/Activation/Activation.php#L117
**Expected behavior**
We should check the existence of the tables only in the specific conditions:
1. When activating the plugin, before Preload do any operation on the tables
2. When enabling RUCSS
3. When enabling Preload
4. During each update of the plugin
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
priority
|
improve detection of missing as tables before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug in the current approach we have a bug related to the not expected characters in the database name our query here is not using causing fail of the check and displaying false message expected behavior we should check the existence of the tables only in the specific conditions when activating the plugin before preload do any operation on the tables when enabling rucss when enabling preload during each update of the plugin backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort
| 1
|
364,391
| 10,763,685,196
|
IssuesEvent
|
2019-11-01 05:15:29
|
AY1920S1-CS2103T-W11-1/main
|
https://api.github.com/repos/AY1920S1-CS2103T-W11-1/main
|
closed
|
Add more meaningful sample data
|
priority.High type.Task
|
- Add sample data for Tags, Find Rules
- Update existing sample data for Problems and Plans
|
1.0
|
Add more meaningful sample data - - Add sample data for Tags, Find Rules
- Update existing sample data for Problems and Plans
|
priority
|
add more meaningful sample data add sample data for tags find rules update existing sample data for problems and plans
| 1
|
251,912
| 8,029,318,495
|
IssuesEvent
|
2018-07-27 15:37:07
|
HBHWoolacotts/RPii
|
https://api.github.com/repos/HBHWoolacotts/RPii
|
opened
|
Delivery Questionnaire - Integrated into EPOS Sale Screen
|
High Priority Label: Future Dev Request
|
We have questionnaires that are filled in by the sales staff for large appliances (such as Range cookers and American Style fridge freezers). They're to do with access to property - door sizes, distances, etc etc. The problem is, they get filled out and sent over to the delivery team by email or fax and then lost.
Could we get something electronic implemented within the RPii sale screen that they can do, perhaps an add-on to the Delivery Note - which then prints when they print the delivery note(s)? And also stores against the sale so you can see it's been done (and add to it if it's not)?
- Create an Icon button in the Sale Screen to open the questionnaire, and perhaps a tick against it when it's been completed, to show it's been done?
- When they print the Delivery Notes, it would need to print this with them (if one has been completed).
- Would also need entries in the audit log as to who filled it out and when.

|
1.0
|
Delivery Questionnaire - Integrated into EPOS Sale Screen - We have questionnaires that are filled in by the sales staff for large appliances (such as Range cookers and American Style fridge freezers). They're to do with access to property - door sizes, distances, etc etc. The problem is, they get filled out and sent over to the delivery team by email or fax and then lost.
Could we get something electronic implemented within the RPii sale screen that they can do, perhaps an add-on to the Delivery Note - which then prints when they print the delivery note(s)? And also stores against the sale so you can see it's been done (and add to it if it's not)?
- Create an Icon button in the Sale Screen to open the questionnaire, and perhaps a tick against it when it's been completed, to show it's been done?
- When they print the Delivery Notes, it would need to print this with them (if one has been completed).
- Would also need entries in the audit log as to who filled it out and when.

|
priority
|
delivery questionnaire integrated into epos sale screen we have questionnaires that are filled in by the sales staff for large appliances such as range cookers and american style fridge freezers they re to do with access to property door sizes distances etc etc the problem is they get filled out and sent over to the delivery team by email or fax and then lost could we get something electronic implemented within the rpii sale screen that they can do perhaps an add on to the delivery note which then prints when they print the delivery note s and also stores against the sale so you can see it s been done and add to it if it s not create an icon button in the sale screen to open the questionnaire and perhaps a tick against it when it s been completed to show it s been done when they print the delivery notes it would need to print this with them if one has been completed would also need entries in the audit log as to who filled it out and when
| 1
|
781,400
| 27,436,297,392
|
IssuesEvent
|
2023-03-02 07:43:50
|
BlueCodeSystems/opensrp-client-ecap-chw
|
https://api.github.com/repos/BlueCodeSystems/opensrp-client-ecap-chw
|
closed
|
Feedback on the last build that was share - 09/05/2022
|
High Priority
|
- [x] When the feature Vulnerability assessment tool and HIV risk assessment has been filled, make sure data is returned immediately. Right now the user has to first exit the profile then reopens the form to see the information that was saved.
- [x] On the child safety plan when I save once the form is duplicated, this should be worked on.
- [x] Make sure a user can fill in information on the graduation assessment form, right now I am not able to test or do anything because the page is getting stuck.
- [x] If the household member does not qualify to be enrolled into the program, it’s best we just show the gray line only and remove the graduation icon from that member because why should we be able to graduate them when we did not enroll them in the first place. I should not even be able to see the other forms because the process has ended there.
- [x] When the VCA has been deleted from the household, this should show immediately, currently a user will first have to exit the page then reopens it to see the changes and the number of the members on the household profile is not updated.
- [ ] So on “DELETE” let us make sure the user can delete those households that were duplicated due to the system remember. Am not able to delete the duplicates having the same household IDs and also am not able to DELETE some mothers in the mother index register - The credentials I am using are username: gyabs and password 1234.If we can not DELETE the duplicates this will affect the data cleaning process.
|
1.0
|
Feedback on the last build that was share - 09/05/2022 -
- [x] When the feature Vulnerability assessment tool and HIV risk assessment has been filled, make sure data is returned immediately. Right now the user has to first exit the profile then reopens the form to see the information that was saved.
- [x] On the child safety plan when I save once the form is duplicated, this should be worked on.
- [x] Make sure a user can fill in information on the graduation assessment form, right now I am not able to test or do anything because the page is getting stuck.
- [x] If the household member does not qualify to be enrolled into the program, it’s best we just show the gray line only and remove the graduation icon from that member because why should we be able to graduate them when we did not enroll them in the first place. I should not even be able to see the other forms because the process has ended there.
- [x] When the VCA has been deleted from the household, this should show immediately, currently a user will first have to exit the page then reopens it to see the changes and the number of the members on the household profile is not updated.
- [ ] So on “DELETE” let us make sure the user can delete those households that were duplicated due to the system remember. Am not able to delete the duplicates having the same household IDs and also am not able to DELETE some mothers in the mother index register - The credentials I am using are username: gyabs and password 1234.If we can not DELETE the duplicates this will affect the data cleaning process.
|
priority
|
feedback on the last build that was share when the feature vulnerability assessment tool and hiv risk assessment has been filled make sure data is returned immediately right now the user has to first exit the profile then reopens the form to see the information that was saved on the child safety plan when i save once the form is duplicated this should be worked on make sure a user can fill in information on the graduation assessment form right now i am not able to test or do anything because the page is getting stuck if the household member does not qualify to be enrolled into the program it’s best we just show the gray line only and remove the graduation icon from that member because why should we be able to graduate them when we did not enroll them in the first place i should not even be able to see the other forms because the process has ended there when the vca has been deleted from the household this should show immediately currently a user will first have to exit the page then reopens it to see the changes and the number of the members on the household profile is not updated so on “delete” let us make sure the user can delete those households that were duplicated due to the system remember am not able to delete the duplicates having the same household ids and also am not able to delete some mothers in the mother index register the credentials i am using are username gyabs and password if we can not delete the duplicates this will affect the data cleaning process
| 1
|
536,392
| 15,708,243,959
|
IssuesEvent
|
2021-03-26 20:12:28
|
JDMCreator/LaTeXTableEditor
|
https://api.github.com/repos/JDMCreator/LaTeXTableEditor
|
closed
|
Mathematical expressions in unordered lists
|
Priority: High Status: Confirmed Type: bug
|
Hello there,
Thanks for the table editor! It is my favourite.
I am not sure if it is only me but when I use the mathematical expression using the mathematical icon as proposed ( $x^2$), it does not appear as superscript but instead, the code generated results in undefined~.
could please you give me a hand with this?
Edit: I think _the problem occurs when I use bullets combined with mathematical expression_. When I don't have the bullets the mathematical icon works perfectly.
Thanks!
_Originally posted by @fdchicom in https://github.com/JDMCreator/LaTeXTableEditor/issues/24#issuecomment-709038237_
|
1.0
|
Mathematical expressions in unordered lists - Hello there,
Thanks for the table editor! It is my favourite.
I am not sure if it is only me but when I use the mathematical expression using the mathematical icon as proposed ( $x^2$), it does not appear as superscript but instead, the code generated results in undefined~.
could please you give me a hand with this?
Edit: I think _the problem occurs when I use bullets combined with mathematical expression_. When I don't have the bullets the mathematical icon works perfectly.
Thanks!
_Originally posted by @fdchicom in https://github.com/JDMCreator/LaTeXTableEditor/issues/24#issuecomment-709038237_
|
priority
|
mathematical expressions in unordered lists hello there thanks for the table editor it is my favourite i am not sure if it is only me but when i use the mathematical expression using the mathematical icon as proposed x it does not appear as superscript but instead the code generated results in undefined could please you give me a hand with this edit i think the problem occurs when i use bullets combined with mathematical expression when i don t have the bullets the mathematical icon works perfectly thanks originally posted by fdchicom in
| 1
|
792,396
| 27,958,687,542
|
IssuesEvent
|
2023-03-24 14:11:57
|
timeseriesAI/tsai
|
https://api.github.com/repos/timeseriesAI/tsai
|
closed
|
multi-horizon forecasting
|
bug under review high-priority answered?
|
Hello, I get an error when I use many-Plus models for multi-horizon forecasting: TypeError: __init__() got an unexpected keyword argument 'custom_head'
my code as follows:
<img width="451" alt="image" src="https://user-images.githubusercontent.com/48724350/194741413-f00943cc-b252-4ec5-925f-638ef065801a.png">
those models include:
(For multi-horizon forecast, have no custom_head)
* FCN
* FCNPlus
* InceptionTime
* MLP
* RNN
* LSTM
* GRU
* RNN_FCN
* MRNN_FCN
* LSTM_FCN
* MLSTM_FCN
* GRU_FCN
* MGRU_FCN
* ResCNN
* ResNet
* ResNetPlus
* TCN
* XceptionTime
* XceptionTimePlus
* ResNetPlus - need seq_len
* RNN_FCNPlus - cannot multiply
* MRNN_FCNPlus - cannot multiply
* LSTM_FCNPlus - cannot multiply
* MLSTM_FCNPlus - cannot multiply
* GRU_FCNPlus - cannot multiply
* MGRU_FCNPlus - cannot multiply
|
1.0
|
multi-horizon forecasting - Hello, I get an error when I use many-Plus models for multi-horizon forecasting: TypeError: __init__() got an unexpected keyword argument 'custom_head'
my code as follows:
<img width="451" alt="image" src="https://user-images.githubusercontent.com/48724350/194741413-f00943cc-b252-4ec5-925f-638ef065801a.png">
those models include:
(For multi-horizon forecast, have no custom_head)
* FCN
* FCNPlus
* InceptionTime
* MLP
* RNN
* LSTM
* GRU
* RNN_FCN
* MRNN_FCN
* LSTM_FCN
* MLSTM_FCN
* GRU_FCN
* MGRU_FCN
* ResCNN
* ResNet
* ResNetPlus
* TCN
* XceptionTime
* XceptionTimePlus
* ResNetPlus - need seq_len
* RNN_FCNPlus - cannot multiply
* MRNN_FCNPlus - cannot multiply
* LSTM_FCNPlus - cannot multiply
* MLSTM_FCNPlus - cannot multiply
* GRU_FCNPlus - cannot multiply
* MGRU_FCNPlus - cannot multiply
|
priority
|
multi horizon forecasting hello i get an error when i use many plus models for multi horizon forecasting typeerror init got an unexpected keyword argument custom head my code as follows: img width alt image src those models include: for multi horizon forecast have no custom head fcn fcnplus inceptiontime mlp rnn lstm gru rnn fcn mrnn fcn lstm fcn mlstm fcn gru fcn mgru fcn rescnn resnet resnetplus tcn xceptiontime xceptiontimeplus resnetplus need seq len rnn fcnplus cannot multiply mrnn fcnplus cannot multiply lstm fcnplus cannot multiply mlstm fcnplus cannot multiply gru fcnplus cannot multiply mgru fcnplus cannot multiply
| 1
|
191,755
| 6,842,355,272
|
IssuesEvent
|
2017-11-12 00:21:29
|
theQRL/qrl-wallet
|
https://api.github.com/repos/theQRL/qrl-wallet
|
closed
|
Implement multiple gRPC backend connections
|
Priority: High Status: In Progress Type: Bug
|
Currently if a user selects a new Node to relay transactions for on the site, this will affect all other users. Each node should have it's own state managed on the Meteor server - and should be accessible to all frontend users.
For now only a single node is enabled.
|
1.0
|
Implement multiple gRPC backend connections - Currently if a user selects a new Node to relay transactions for on the site, this will affect all other users. Each node should have it's own state managed on the Meteor server - and should be accessible to all frontend users.
For now only a single node is enabled.
|
priority
|
implement multiple grpc backend connections currently if a user selects a new node to relay transactions for on the site this will affect all other users each node should have it s own state managed on the meteor server and should be accessible to all frontend users for now only a single node is enabled
| 1
|
22,078
| 2,645,269,249
|
IssuesEvent
|
2015-03-12 21:39:22
|
Connexions/webview
|
https://api.github.com/repos/Connexions/webview
|
opened
|
Incorrect user login
|
High Priority
|
In beta.cnx.org, webview will automatically sign in under the wrong user name without a password. For example, when visiting beta.cnx.org my browser will automatically sign into 'amwebb' without any prompt.

@reedstrm suggests that this is a problem of incorrectly caching with varnish, because clearing my browser cache removes the 'amwebb' login. This issue was first reported on Connexions/rhaptos.cnxmlutils/issues/120
|
1.0
|
Incorrect user login - In beta.cnx.org, webview will automatically sign in under the wrong user name without a password. For example, when visiting beta.cnx.org my browser will automatically sign into 'amwebb' without any prompt.

@reedstrm suggests that this is a problem of incorrectly caching with varnish, because clearing my browser cache removes the 'amwebb' login. This issue was first reported on Connexions/rhaptos.cnxmlutils/issues/120
|
priority
|
incorrect user login in beta cnx org webview will automatically sign in under the wrong user name without a password for example when visiting beta cnx org my browser will automatically sign into amwebb without any prompt reedstrm suggests that this is a problem of incorrectly caching with varnish because clearing my browser cache removes the amwebb login this issue was first reported on connexions rhaptos cnxmlutils issues
| 1
|
690,529
| 23,663,104,873
|
IssuesEvent
|
2022-08-26 17:39:16
|
zitadel/zitadel
|
https://api.github.com/repos/zitadel/zitadel
|
closed
|
Contributor guides for v2
|
category: docs priority: high
|
Particularly, we should fix and/or redesign paths and environment variables.
- [ ] Review contibution readmes
- [ ] How to contribute to console
- [ ] How to contibute to login
- [ ] How to contribute to backend
|
1.0
|
Contributor guides for v2 - Particularly, we should fix and/or redesign paths and environment variables.
- [ ] Review contibution readmes
- [ ] How to contribute to console
- [ ] How to contibute to login
- [ ] How to contribute to backend
|
priority
|
contributor guides for particularly we should fix and or redesign paths and environment variables review contibution readmes how to contribute to console how to contibute to login how to contribute to backend
| 1
|
113,719
| 4,567,600,789
|
IssuesEvent
|
2016-09-15 11:48:03
|
ponylang/ponyc
|
https://api.github.com/repos/ponylang/ponyc
|
closed
|
release 0.3.1
|
priority: 3 - high
|
We need to do a new release 0.3.1 in order to get the fix (a reversion in this case) for #1212 out to the public.
I'm going to be doing this release and documenting based on the work that @killerswan did and I'll be getting his help with it so it will be a little slow in happening, but we should come out of it with some good documentation that anyone can follow.
Given that we said we would be doing monthly releases as our minimum, good documentation for this is essential.
|
1.0
|
release 0.3.1 - We need to do a new release 0.3.1 in order to get the fix (a reversion in this case) for #1212 out to the public.
I'm going to be doing this release and documenting based on the work that @killerswan did and I'll be getting his help with it so it will be a little slow in happening, but we should come out of it with some good documentation that anyone can follow.
Given that we said we would be doing monthly releases as our minimum, good documentation for this is essential.
|
priority
|
release we need to do a new release in order to get the fix a reversion in this case for out to the public i m going to be doing this release and documenting based on the work that killerswan did and i ll be getting his help with it so it will be a little slow in happening but we should come out of it with some good documentation that anyone can follow given that we said we would be doing monthly releases as our minimum good documentation for this is essential
| 1
|
232,949
| 7,687,949,278
|
IssuesEvent
|
2018-05-17 07:54:22
|
HGustavs/LenaSYS
|
https://api.github.com/repos/HGustavs/LenaSYS
|
closed
|
SectionED: Hide the statistics tab on mobile
|
Grupp 3 (2018) Grupp 3 (2018) Sektion-Editor highPriority
|
Currently when on mobile mode the statistics tab dosn't behave well, so the idea was to hide it untill we figure out a method to display it.
## The issue
<img width="406" alt="skarmavbild 2018-05-17 kl 9 02 29 fm" src="https://user-images.githubusercontent.com/37795608/40161438-0cb7e85e-59b1-11e8-8c0e-2ee82dd10ac0.png">
|
1.0
|
SectionED: Hide the statistics tab on mobile - Currently when on mobile mode the statistics tab dosn't behave well, so the idea was to hide it untill we figure out a method to display it.
## The issue
<img width="406" alt="skarmavbild 2018-05-17 kl 9 02 29 fm" src="https://user-images.githubusercontent.com/37795608/40161438-0cb7e85e-59b1-11e8-8c0e-2ee82dd10ac0.png">
|
priority
|
sectioned hide the statistics tab on mobile currently when on mobile mode the statistics tab dosn t behave well so the idea was to hide it untill we figure out a method to display it the issue img width alt skarmavbild kl fm src
| 1
|
401,039
| 11,784,467,164
|
IssuesEvent
|
2020-03-17 08:25:31
|
AY1920S2-CS2103T-W12-1/main
|
https://api.github.com/repos/AY1920S2-CS2103T-W12-1/main
|
closed
|
As a courier, I want an application that allows me to view delivery orders based on a given postal sector
|
priority.High type.Story
|
So that I can easily find delivery orders in the same general location.
|
1.0
|
As a courier, I want an application that allows me to view delivery orders based on a given postal sector - So that I can easily find delivery orders in the same general location.
|
priority
|
as a courier i want an application that allows me to view delivery orders based on a given postal sector so that i can easily find delivery orders in the same general location
| 1
|
189,892
| 6,802,683,928
|
IssuesEvent
|
2017-11-02 21:05:44
|
jonathanGB/faulty-robots-simulation
|
https://api.github.com/repos/jonathanGB/faulty-robots-simulation
|
closed
|
Handle "generate"
|
high-priority
|
- [x] Only enable when there is >= 2 robots
- [x] On click, disable event listeners on "setup" robots
* keep listeners on double click
* make top inputs & textarea disabled (change css cursor on disabled too?)
- [x] display generation "001"
* pre-generate 10 generations in advance using the webworker?
* show icon on the side to generate next "wave"?
|
1.0
|
Handle "generate" - - [x] Only enable when there is >= 2 robots
- [x] On click, disable event listeners on "setup" robots
* keep listeners on double click
* make top inputs & textarea disabled (change css cursor on disabled too?)
- [x] display generation "001"
* pre-generate 10 generations in advance using the webworker?
* show icon on the side to generate next "wave"?
|
priority
|
handle generate only enable when there is robots on click disable event listeners on setup robots keep listeners on double click make top inputs textarea disabled change css cursor on disabled too display generation pre generate generations in advance using the webworker show icon on the side to generate next wave
| 1
|
103,526
| 4,174,765,735
|
IssuesEvent
|
2016-06-21 14:57:53
|
smartdevicelink/sdl_android
|
https://api.github.com/repos/smartdevicelink/sdl_android
|
closed
|
Proxy doesn't handle hybrid service packets correctly
|
bug high priority
|
Currently the proxy will throw away any packets that come through the Hybrid/bulk data service.
```Java
if (message.getSessionType().equals(SessionType.RPC)) {
...
}else{
// Handle other protocol message types here
}
```
See https://github.com/smartdevicelink/sdl_android/blob/master/sdl_android_lib/src/com/smartdevicelink/proxy/SdlProxyBase.java#L1332
Quickest fix is to just add an or statement that includes if it's type Hybrid/bulk data
```Java
if (message.getSessionType().equals(SessionType.RPC) || message.getSessionType().equals(SessionType. BULK_DATA ) {...}
````
|
1.0
|
Proxy doesn't handle hybrid service packets correctly - Currently the proxy will throw away any packets that come through the Hybrid/bulk data service.
```Java
if (message.getSessionType().equals(SessionType.RPC)) {
...
}else{
// Handle other protocol message types here
}
```
See https://github.com/smartdevicelink/sdl_android/blob/master/sdl_android_lib/src/com/smartdevicelink/proxy/SdlProxyBase.java#L1332
Quickest fix is to just add an or statement that includes if it's type Hybrid/bulk data
```Java
if (message.getSessionType().equals(SessionType.RPC) || message.getSessionType().equals(SessionType. BULK_DATA ) {...}
````
|
priority
|
proxy doesn t handle hybrid service packets correctly currently the proxy will throw away any packets that come through the hybrid bulk data service java if message getsessiontype equals sessiontype rpc else handle other protocol message types here see quickest fix is to just add an or statement that includes if it s type hybrid bulk data java if message getsessiontype equals sessiontype rpc message getsessiontype equals sessiontype bulk data
| 1
|
323,245
| 9,852,351,442
|
IssuesEvent
|
2019-06-19 12:41:00
|
eaudeweb/ozone
|
https://api.github.com/repos/eaudeweb/ozone
|
closed
|
Extend the "Transfers of production rights" form
|
Complexity: Low Component: Vue Priority: Highest
|
To show (read-only) the related structured data entered by the secretariat in the transfer forms.
|
1.0
|
Extend the "Transfers of production rights" form - To show (read-only) the related structured data entered by the secretariat in the transfer forms.
|
priority
|
extend the transfers of production rights form to show read only the related structured data entered by the secretariat in the transfer forms
| 1
|
175,455
| 6,551,211,393
|
IssuesEvent
|
2017-09-05 14:03:19
|
twosigma/beakerx
|
https://api.github.com/repos/twosigma/beakerx
|
closed
|
pandas dataframes not showing
|
Bug Priority High Runtime Python
|
first cell of tableApi_python.ipynb produces no output and no error.
|
1.0
|
pandas dataframes not showing - first cell of tableApi_python.ipynb produces no output and no error.
|
priority
|
pandas dataframes not showing first cell of tableapi python ipynb produces no output and no error
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.