Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
655,348 | 21,686,411,345 | IssuesEvent | 2022-05-09 11:44:37 | AgnostiqHQ/covalent | https://api.github.com/repos/AgnostiqHQ/covalent | closed | Decouple Dask from Covalent | feature priority / high covalent-mono | * Remove the dask dependency entirely from Covalent
* Use asyncio event loops to run workflows (single dispatch) at a time | 1.0 | Decouple Dask from Covalent - * Remove the dask dependency entirely from Covalent
* Use asyncio event loops to run workflows (single dispatch) at a time | priority | decouple dask from covalent remove the dask dependency entirely from covalent use asyncio event loops to run workflows single dispatch at a time | 1 |
351,412 | 10,518,134,108 | IssuesEvent | 2019-09-29 08:37:03 | phansch/dotfiles | https://api.github.com/repos/phansch/dotfiles | closed | Integrate ./util/dev fmt into Clippy workflow | Ansible enhancement priority:high | I don't think I want to run this on every save, but rather in a pre-commit hook. | 1.0 | Integrate ./util/dev fmt into Clippy workflow - I don't think I want to run this on every save, but rather in a pre-commit hook. | priority | integrate util dev fmt into clippy workflow i don t think i want to run this on every save but rather in a pre commit hook | 1 |
636,466 | 20,601,094,832 | IssuesEvent | 2022-03-06 09:10:17 | AY2122S2-CS2103-F09-3/tp | https://api.github.com/repos/AY2122S2-CS2103-F09-3/tp | reopened | Add appointment | type.Story priority.High | As an insurance agent, I can add new appointments so that I can note down any future meetings I have with my client | 1.0 | Add appointment - As an insurance agent, I can add new appointments so that I can note down any future meetings I have with my client | priority | add appointment as an insurance agent i can add new appointments so that i can note down any future meetings i have with my client | 1 |
506,638 | 14,669,592,529 | IssuesEvent | 2020-12-30 01:22:17 | napari/napari | https://api.github.com/repos/napari/napari | closed | Adding points on heavily-scaled data slices them on the wrong plane | bug high priority | ## 🐛 Bug
```python
import napari
from skimage import data
image = data.binary_blobs(128, n_dim=3)[:60]
with napari.gui_qt():
v = napari.view_image(image, scale=[0.29, 0.26, 0.26])
pts = v.add_points(name='interactive', size=1, ndim=3)
pts.mode = 'add'
v.dims.set_point(0, 30 * 0.29) # middle plane
```
Clicking on points adds them on the right plane, but they are not shown. One must slice several planes lower to see them. When you switch to 3D mode, they appear in the middle plane correctly.
| 1.0 | Adding points on heavily-scaled data slices them on the wrong plane - ## 🐛 Bug
```python
import napari
from skimage import data
image = data.binary_blobs(128, n_dim=3)[:60]
with napari.gui_qt():
v = napari.view_image(image, scale=[0.29, 0.26, 0.26])
pts = v.add_points(name='interactive', size=1, ndim=3)
pts.mode = 'add'
v.dims.set_point(0, 30 * 0.29) # middle plane
```
Clicking on points adds them on the right plane, but they are not shown. One must slice several planes lower to see them. When you switch to 3D mode, they appear in the middle plane correctly.
| priority | adding points on heavily scaled data slices them on the wrong plane 🐛 bug python import napari from skimage import data image data binary blobs n dim with napari gui qt v napari view image image scale pts v add points name interactive size ndim pts mode add v dims set point middle plane clicking on points adds them on the right plane but they are not shown one must slice several planes lower to see them when you switch to mode they appear in the middle plane correctly | 1 |
420,109 | 12,233,439,095 | IssuesEvent | 2020-05-04 11:38:33 | joonaspaakko/ScriptUI-Dialog-Builder-Joonas | https://api.github.com/repos/joonaspaakko/ScriptUI-Dialog-Builder-Joonas | opened | VerticalTabbedPanel + TabbedPanel → Automatic minimum size not working reliably | .BUG Priority: High ScriptUI - Disparity | I think I might've made some changes to the regular `TabbedPanel` when I added `VerticalTabbedPanel`, which caused this issue in both.
The expected behavior is for tabs to be the same size at all times.
Issue gif:

| 1.0 | VerticalTabbedPanel + TabbedPanel → Automatic minimum size not working reliably - I think I might've made some changes to the regular `TabbedPanel` when I added `VerticalTabbedPanel`, which caused this issue in both.
The expected behavior is for tabs to be the same size at all times.
Issue gif:

| priority | verticaltabbedpanel tabbedpanel → automatic minimum size not working reliably i think i might ve made some changes to the regular tabbedpanel when i added verticaltabbedpanel which caused this issue in both the expected behavior is for tabs to be the same size at all times issue gif | 1 |
540,370 | 15,806,988,403 | IssuesEvent | 2021-04-04 08:17:10 | AY2021S2-CS2113T-W09-2/tp | https://api.github.com/repos/AY2021S2-CS2113T-W09-2/tp | closed | [PE-D] Update prerequisite after marked as done | priority.High severity.Low severity.Medium type.Bug | No details provided.

After marking a module as done, it was still possible to update the mod with a prerequisite of an undone mod. Perhaps can consider checking for this and throwing an error?
<!--session: 1617437456087-b11cd5d4-595b-4323-aa93-43e1ba798752-->
-------------
Labels: `severity.Low` `type.FunctionalityBug`
original: AlexanderTanJunAn/ped#5 | 1.0 | [PE-D] Update prerequisite after marked as done - No details provided.

After marking a module as done, it was still possible to update the mod with a prerequisite of an undone mod. Perhaps can consider checking for this and throwing an error?
<!--session: 1617437456087-b11cd5d4-595b-4323-aa93-43e1ba798752-->
-------------
Labels: `severity.Low` `type.FunctionalityBug`
original: AlexanderTanJunAn/ped#5 | priority | update prerequisite after marked as done no details provided after marking a module as done it was still possible to update the mod with a prerequisite of an undone mod perhaps can consider checking for this and throwing an error labels severity low type functionalitybug original alexandertanjunan ped | 1 |
645,821 | 21,016,381,335 | IssuesEvent | 2022-03-30 11:25:06 | aave/interface | https://api.github.com/repos/aave/interface | closed | Governance page missing remaining time | feature priority:high | **Is your feature request related to a problem? Please describe.**
On the proposal page in governance it's not clear how much time is remaining:
https://app.aave.com/governance/ <- here it's visible
https://app.aave.com/governance/proposal/68/ <- here it's not
**Describe the solution you'd like**
Show the raiming time, as it's an important information.
**Additional context**


| 1.0 | Governance page missing remaining time - **Is your feature request related to a problem? Please describe.**
On the proposal page in governance it's not clear how much time is remaining:
https://app.aave.com/governance/ <- here it's visible
https://app.aave.com/governance/proposal/68/ <- here it's not
**Describe the solution you'd like**
Show the raiming time, as it's an important information.
**Additional context**


| priority | governance page missing remaining time is your feature request related to a problem please describe on the proposal page in governance it s not clear how much time is remaining here it s visible here it s not describe the solution you d like show the raiming time as it s an important information additional context | 1 |
715,554 | 24,603,846,535 | IssuesEvent | 2022-10-14 14:35:05 | epicmaxco/vuestic-ui | https://api.github.com/repos/epicmaxco/vuestic-ui | closed | Dropdown doesn't work on SSR | BUG HIGH PRIORITY | **Vuestic-ui version:** 1.5.0
Just doesn't open on click

Same with popover, very dead
 | 1.0 | Dropdown doesn't work on SSR - **Vuestic-ui version:** 1.5.0
Just doesn't open on click

Same with popover, very dead
 | priority | dropdown doesn t work on ssr vuestic ui version just doesn t open on click same with popover very dead | 1 |
598,855 | 18,257,127,919 | IssuesEvent | 2021-10-03 08:06:42 | AY2122S1-CS2113T-W12-1/tp | https://api.github.com/repos/AY2122S1-CS2113T-W12-1/tp | opened | Implement database class which contains all food data | priorityHigh | Implement database class that contains an ArrayList that stores all food data. | 1.0 | Implement database class which contains all food data - Implement database class that contains an ArrayList that stores all food data. | priority | implement database class which contains all food data implement database class that contains an arraylist that stores all food data | 1 |
455,885 | 13,133,734,399 | IssuesEvent | 2020-08-06 21:33:02 | CityOfNewYork/poletop-finder | https://api.github.com/repos/CityOfNewYork/poletop-finder | closed | Include symbology in details panel | priority: high | If equipment is installed vs not installed, appropriate icon should display next to the Poletop Reservation ID. | 1.0 | Include symbology in details panel - If equipment is installed vs not installed, appropriate icon should display next to the Poletop Reservation ID. | priority | include symbology in details panel if equipment is installed vs not installed appropriate icon should display next to the poletop reservation id | 1 |
689,836 | 23,636,003,648 | IssuesEvent | 2022-08-25 13:21:57 | kubermatic/kubermatic | https://api.github.com/repos/kubermatic/kubermatic | closed | Random failures with OIDC kubeconfig workflow | kind/bug priority/high | ### What happened?
<!-- Try to provide as much information as possible.
If you're reporting a security issue, please check the guidelines for reporting security issues:
https://github.com/kubermatic/kubermatic/blob/master/CONTRIBUTING.md#reporting-a-security-vulnerability -->
The OIDC Kubeconfig feature seems to be broken. While trying the workflow from the dashboard, we keep ending up with random errors, any one from this list:
```
- {"error":{"code":500,"message":"securecookie: base64 decode failed - caused by: illegal base64 data at input byte 3"}}
- {"error":{"code":400,"message":"incorrect value of state parameter: tmbhb7s9jj"}}
- {"error":{"code":400,"message":"incorrect value of cookie or cookie not set: http: named cookie not present"}}
- {"error":{"code":400,"message":"error while exchanging oidc code for token: oauth2: cannot fetch token: 400 Bad Request\nResponse: {\"error\":\"invalid_request\",\"error_description\":\"redirect_uri did not match URI from initial request.\"}"}}
```
Although without making any changes, sometimes it works.
### How to reproduce the issue?
<!-- Please provide as much information as possible, so we can reproduce the issue on our own. -->
Using our dashboard:
- Ensure that OIDC Kubeconfig is enabled in admin settings
- Create a Kubernetes Cluster
- Click on `Get Kubeconfig` to retrieve the OIDC Kubeconfig


### How is your environment configured?
- KKP version:
- Shared or separate master/seed clusters?:
### Provide your KKP manifest here (if applicable)
<!-- Providing an applicable manifest (KubermaticConfiguration, Seed, Cluster or other resources) will help us to reproduce the issue.
Please make sure to redact all secrets (e.g. passwords, URLs...)! -->
<details>
```yaml
# paste manifest here
```
</details>
### What cloud provider are you running on?
<!-- AWS, Azure, DigitalOcean, GCP, Hetzner Cloud, Nutanix, OpenStack, Equinix Metal (Packet), VMware vSphere, Other (e.g. baremetal or non-natively supported provider) -->
### What operating system are you running in your user cluster?
<!-- Ubuntu 20.04, CentOS 7, Rocky Linux 8, Flatcar Linux, ... (optional, bug might not be related to user cluster) -->
### Additional information
<!-- Additional information about the bug you're reporting (optional). -->
| 1.0 | Random failures with OIDC kubeconfig workflow - ### What happened?
<!-- Try to provide as much information as possible.
If you're reporting a security issue, please check the guidelines for reporting security issues:
https://github.com/kubermatic/kubermatic/blob/master/CONTRIBUTING.md#reporting-a-security-vulnerability -->
The OIDC Kubeconfig feature seems to be broken. While trying the workflow from the dashboard, we keep ending up with random errors, any one from this list:
```
- {"error":{"code":500,"message":"securecookie: base64 decode failed - caused by: illegal base64 data at input byte 3"}}
- {"error":{"code":400,"message":"incorrect value of state parameter: tmbhb7s9jj"}}
- {"error":{"code":400,"message":"incorrect value of cookie or cookie not set: http: named cookie not present"}}
- {"error":{"code":400,"message":"error while exchanging oidc code for token: oauth2: cannot fetch token: 400 Bad Request\nResponse: {\"error\":\"invalid_request\",\"error_description\":\"redirect_uri did not match URI from initial request.\"}"}}
```
Although without making any changes, sometimes it works.
### How to reproduce the issue?
<!-- Please provide as much information as possible, so we can reproduce the issue on our own. -->
Using our dashboard:
- Ensure that OIDC Kubeconfig is enabled in admin settings
- Create a Kubernetes Cluster
- Click on `Get Kubeconfig` to retrieve the OIDC Kubeconfig


### How is your environment configured?
- KKP version:
- Shared or separate master/seed clusters?:
### Provide your KKP manifest here (if applicable)
<!-- Providing an applicable manifest (KubermaticConfiguration, Seed, Cluster or other resources) will help us to reproduce the issue.
Please make sure to redact all secrets (e.g. passwords, URLs...)! -->
<details>
```yaml
# paste manifest here
```
</details>
### What cloud provider are you running on?
<!-- AWS, Azure, DigitalOcean, GCP, Hetzner Cloud, Nutanix, OpenStack, Equinix Metal (Packet), VMware vSphere, Other (e.g. baremetal or non-natively supported provider) -->
### What operating system are you running in your user cluster?
<!-- Ubuntu 20.04, CentOS 7, Rocky Linux 8, Flatcar Linux, ... (optional, bug might not be related to user cluster) -->
### Additional information
<!-- Additional information about the bug you're reporting (optional). -->
| priority | random failures with oidc kubeconfig workflow what happened try to provide as much information as possible if you re reporting a security issue please check the guidelines for reporting security issues the oidc kubeconfig feature seems to be broken while trying the workflow from the dashboard we keep ending up with random errors any one from this list error code message securecookie decode failed caused by illegal data at input byte error code message incorrect value of state parameter error code message incorrect value of cookie or cookie not set http named cookie not present error code message error while exchanging oidc code for token cannot fetch token bad request nresponse error invalid request error description redirect uri did not match uri from initial request although without making any changes sometimes it works how to reproduce the issue using our dashboard ensure that oidc kubeconfig is enabled in admin settings create a kubernetes cluster click on get kubeconfig to retrieve the oidc kubeconfig how is your environment configured kkp version shared or separate master seed clusters provide your kkp manifest here if applicable providing an applicable manifest kubermaticconfiguration seed cluster or other resources will help us to reproduce the issue please make sure to redact all secrets e g passwords urls yaml paste manifest here what cloud provider are you running on what operating system are you running in your user cluster additional information | 1 |
604,659 | 18,716,136,627 | IssuesEvent | 2021-11-03 05:10:48 | AY2122S1-CS2113T-T09-4/tp | https://api.github.com/repos/AY2122S1-CS2113T-T09-4/tp | closed | [PE-D] Inconsistent input flags and params | priority.High | Expected to be able to input with the presence of the ‘n/’ and ‘c/’ flag but required the presence of a space in front of input to be added. However, when adding in a similar format to modules it was allowed. Could make it consistent throughout the program.

-No space needed

-Space needed
Same problem for journal entries

<!--session: 1635513809369-5bee5e77-b44a-43bb-a435-5814eef45f7f--><!--Version: Web v3.4.1-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: STAung07/ped#2 | 1.0 | [PE-D] Inconsistent input flags and params - Expected to be able to input with the presence of the ‘n/’ and ‘c/’ flag but required the presence of a space in front of input to be added. However, when adding in a similar format to modules it was allowed. Could make it consistent throughout the program.

-No space needed

-Space needed
Same problem for journal entries

<!--session: 1635513809369-5bee5e77-b44a-43bb-a435-5814eef45f7f--><!--Version: Web v3.4.1-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: STAung07/ped#2 | priority | inconsistent input flags and params expected to be able to input with the presence of the ‘n ’ and ‘c ’ flag but required the presence of a space in front of input to be added however when adding in a similar format to modules it was allowed could make it consistent throughout the program no space needed space needed same problem for journal entries labels severity medium type functionalitybug original ped | 1 |
214,304 | 7,268,685,607 | IssuesEvent | 2018-02-20 10:57:49 | TCA-Team/TumCampusApp | https://api.github.com/repos/TCA-Team/TumCampusApp | closed | Crash after upgrade to v1.5.2 | Bug High Priority | ## Problem
Since upgrading to v1.5.2 the app crashes all the time when I open it. It also crashes from time to time while the app is not even used at all. Chat was never used before.
## Stacktrace

### Environment
* Phone: Nexus 7 (2013)
* OS version: 6.0.1
* TUM Campus App version: v1.5.2
* Language: de
| 1.0 | Crash after upgrade to v1.5.2 - ## Problem
Since upgrading to v1.5.2 the app crashes all the time when I open it. It also crashes from time to time while the app is not even used at all. Chat was never used before.
## Stacktrace

### Environment
* Phone: Nexus 7 (2013)
* OS version: 6.0.1
* TUM Campus App version: v1.5.2
* Language: de
| priority | crash after upgrade to problem since upgrading to the app crashes all the time when i open it it also crashes from time to time while the app is not even used at all chat was never used before stacktrace environment phone nexus os version tum campus app version language de | 1 |
338,608 | 10,232,261,592 | IssuesEvent | 2019-08-18 16:05:49 | fossasia/open-event-attendee-android | https://api.github.com/repos/fossasia/open-event-attendee-android | closed | App is crashing due to invalid timezone | Priority: High bug | **Describe the bug**
```
D/AndroidRuntime: Shutting down VM
E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.eventyay.attendee, PID: 17006
org.threeten.bp.DateTimeException: Invalid ID for region-based ZoneId, invalid format:
at org.threeten.bp.ZoneRegion.ofId(ZoneRegion.java:138)
at org.threeten.bp.ZoneId.of(ZoneId.java:358)
at org.fossasia.openevent.general.event.EventUtils.getEventDateTime(EventUtils.kt:40)
at org.fossasia.openevent.general.sessions.SessionViewHolder.bind(SessionViewHolder.kt:44)
at org.fossasia.openevent.general.sessions.SessionRecyclerAdapter.onBindViewHolder(SessionRecyclerAdapter.kt:29)
at org.fossasia.openevent.general.sessions.SessionRecyclerAdapter.onBindViewHolder(SessionRecyclerAdapter.kt:9)
```
**Would you like to work on the issue?**
<!-- Please let us know if you can work on it or the issue should be assigned to someone else. -->
yes
| 1.0 | App is crashing due to invalid timezone - **Describe the bug**
```
D/AndroidRuntime: Shutting down VM
E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.eventyay.attendee, PID: 17006
org.threeten.bp.DateTimeException: Invalid ID for region-based ZoneId, invalid format:
at org.threeten.bp.ZoneRegion.ofId(ZoneRegion.java:138)
at org.threeten.bp.ZoneId.of(ZoneId.java:358)
at org.fossasia.openevent.general.event.EventUtils.getEventDateTime(EventUtils.kt:40)
at org.fossasia.openevent.general.sessions.SessionViewHolder.bind(SessionViewHolder.kt:44)
at org.fossasia.openevent.general.sessions.SessionRecyclerAdapter.onBindViewHolder(SessionRecyclerAdapter.kt:29)
at org.fossasia.openevent.general.sessions.SessionRecyclerAdapter.onBindViewHolder(SessionRecyclerAdapter.kt:9)
```
**Would you like to work on the issue?**
<!-- Please let us know if you can work on it or the issue should be assigned to someone else. -->
yes
| priority | app is crashing due to invalid timezone describe the bug d androidruntime shutting down vm e androidruntime fatal exception main process com eventyay attendee pid org threeten bp datetimeexception invalid id for region based zoneid invalid format at org threeten bp zoneregion ofid zoneregion java at org threeten bp zoneid of zoneid java at org fossasia openevent general event eventutils geteventdatetime eventutils kt at org fossasia openevent general sessions sessionviewholder bind sessionviewholder kt at org fossasia openevent general sessions sessionrecycleradapter onbindviewholder sessionrecycleradapter kt at org fossasia openevent general sessions sessionrecycleradapter onbindviewholder sessionrecycleradapter kt would you like to work on the issue yes | 1 |
228,010 | 7,545,065,962 | IssuesEvent | 2018-04-17 20:25:52 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | BSOD: CLOCK_WATCHDOG_TIMEOUT | High Priority | On Windows 10, while running the latest version of the game, I will randomly experience a blue screen crash. The first time it happened was the last time I played on this computer 2 years ago. That crash corrupted a world, which I then reported and then stopped playing until now. This time, instead of running my own server, I'm playing in another. The first time it happened with the latest build I was actively mining stone, however there were other programs running at the same time. The second time I was away from keyboard doing nothing and came back to the blue screen in progress.
I never get this with any other software, so something must be going on with Eco to trigger it. However, it doesn't seem to be caused by any specific in game action.
System info:
OS: Windows 10 Professional 64-bit
Processor: Intel Core i7-5820k CPU @ 3.30 Ghz (water cooled)
GPU: Nvidia 980 Ti (water cooled)
RAM: 64.0 GB
| 1.0 | BSOD: CLOCK_WATCHDOG_TIMEOUT - On Windows 10, while running the latest version of the game, I will randomly experience a blue screen crash. The first time it happened was the last time I played on this computer 2 years ago. That crash corrupted a world, which I then reported and then stopped playing until now. This time, instead of running my own server, I'm playing in another. The first time it happened with the latest build I was actively mining stone, however there were other programs running at the same time. The second time I was away from keyboard doing nothing and came back to the blue screen in progress.
I never get this with any other software, so something must be going on with Eco to trigger it. However, it doesn't seem to be caused by any specific in game action.
System info:
OS: Windows 10 Professional 64-bit
Processor: Intel Core i7-5820k CPU @ 3.30 Ghz (water cooled)
GPU: Nvidia 980 Ti (water cooled)
RAM: 64.0 GB
| priority | bsod clock watchdog timeout on windows while running the latest version of the game i will randomly experience a blue screen crash the first time it happened was the last time i played on this computer years ago that crash corrupted a world which i then reported and then stopped playing until now this time instead of running my own server i m playing in another the first time it happened with the latest build i was actively mining stone however there were other programs running at the same time the second time i was away from keyboard doing nothing and came back to the blue screen in progress i never get this with any other software so something must be going on with eco to trigger it however it doesn t seem to be caused by any specific in game action system info os windows professional bit processor intel core cpu ghz water cooled gpu nvidia ti water cooled ram gb | 1 |
215,673 | 7,296,573,272 | IssuesEvent | 2018-02-26 11:14:12 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | No tooltips for any of the buttons in the Dashboard | Affected/5.5.0-Alpha2 Priority/High Severity/Minor Type/Improvement | The following were noted in the Consent Management UI


| 1.0 | No tooltips for any of the buttons in the Dashboard - The following were noted in the Consent Management UI


| priority | no tooltips for any of the buttons in the dashboard the following were noted in the consent management ui | 1 |
675,074 | 23,077,694,048 | IssuesEvent | 2022-07-26 02:24:41 | ballerina-platform/ballerina-dev-website | https://api.github.com/repos/ballerina-platform/ballerina-dev-website | closed | Remove `-ing` Form From All BBE Titles | Priority/Highest Type/Task Area/BBEs | ## Description
There are BBEs with the `-ing` form added in the titles. Need to remove them. For example, [Ignoring return values and errors](https://ballerina.io/learn/by-example/ignoring-return-values-and-errors.html), [Controlling openness](https://ballerina.io/learn/by-example/controlling-openness.html), etc.
Area/BBEs
<!--Area/BBEs-->
<!--Area/HomePageSamples-->
<!--Area/LearnPages-->
<!--Area/CommonPages-->
<!--Area/Backend-->
<!--Area/UIUX-->
<!--Area/Workflows-->
<!--Area/Blog-->
## Describe your task(s)
> A detailed description of the task.
## Related issue(s) (optional)
> Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc.
## Suggested label(s) (optional)
> Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels.
## Suggested assignee(s) (optional)
> Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
| 1.0 | Remove `-ing` Form From All BBE Titles - ## Description
There are BBEs with the `-ing` form added in the titles. Need to remove them. For example, [Ignoring return values and errors](https://ballerina.io/learn/by-example/ignoring-return-values-and-errors.html), [Controlling openness](https://ballerina.io/learn/by-example/controlling-openness.html), etc.
Area/BBEs
<!--Area/BBEs-->
<!--Area/HomePageSamples-->
<!--Area/LearnPages-->
<!--Area/CommonPages-->
<!--Area/Backend-->
<!--Area/UIUX-->
<!--Area/Workflows-->
<!--Area/Blog-->
## Describe your task(s)
> A detailed description of the task.
## Related issue(s) (optional)
> Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc.
## Suggested label(s) (optional)
> Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels.
## Suggested assignee(s) (optional)
> Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
| priority | remove ing form from all bbe titles description there are bbes with the ing form added in the titles need to remove them for example etc area bbes describe your task s a detailed description of the task related issue s optional any related issues such as sub tasks and issues reported in other repositories e g component repositories similar problems etc suggested label s optional optional comma separated list of suggested labels non committers can’t assign labels to issues and thereby this will help issue creators who are not a committer to suggest possible labels suggested assignee s optional optional comma separated list of suggested team members who should attend the issue non committers can’t assign issues to assignees and thereby this will help issue creators who are not a committer to suggest possible assignees | 1 |
253,914 | 8,067,358,629 | IssuesEvent | 2018-08-05 06:12:37 | Entrana/EntranaBugs | https://api.github.com/repos/Entrana/EntranaBugs | opened | Bank Tabs | enhancement high priority server issue | With the new bank interface being packed, we need the bank configuration server sided to accept, save, configure, and load bank tabs. Reference Impact for configurations | 1.0 | Bank Tabs - With the new bank interface being packed, we need the bank configuration server sided to accept, save, configure, and load bank tabs. Reference Impact for configurations | priority | bank tabs with the new bank interface being packed we need the bank configuration server sided to accept save configure and load bank tabs reference impact for configurations | 1 |
130,223 | 5,111,988,579 | IssuesEvent | 2017-01-06 09:20:40 | hpi-swt2/workshop-portal | https://api.github.com/repos/hpi-swt2/workshop-portal | closed | List of event drafts | High Priority ready team-helene | **As**
organizer
**I want to**
see the list of event drafts on the events overview page
**in order to**
edit them after saving
# Acceptance criteria
- When an organizer creates an events and saves it without publishing it, the entry is shown in the list of events
- Draft, that are not published yet are visually marked as drafts (status: draft/published)
- The edit page of a draft has the publish button for making it visible to applicants
- Coaches and other users do not see the drafts in the list. (only organizers & admin can see them) | 1.0 | List of event drafts - **As**
organizer
**I want to**
see the list of event drafts on the events overview page
**in order to**
edit them after saving
# Acceptance criteria
- When an organizer creates an events and saves it without publishing it, the entry is shown in the list of events
- Draft, that are not published yet are visually marked as drafts (status: draft/published)
- The edit page of a draft has the publish button for making it visible to applicants
- Coaches and other users do not see the drafts in the list. (only organizers & admin can see them) | priority | list of event drafts as organizer i want to see the list of event drafts on the events overview page in order to edit them after saving acceptance criteria when an organizer creates an events and saves it without publishing it the entry is shown in the list of events draft that are not published yet are visually marked as drafts status draft published the edit page of a draft has the publish button for making it visible to applicants coaches and other users do not see the drafts in the list only organizers admin can see them | 1 |
311,300 | 9,531,666,039 | IssuesEvent | 2019-04-29 16:33:28 | minio/minio | https://api.github.com/repos/minio/minio | closed | Bucket not resolved when using MINIO_DOMAIN | community priority: high | ## Expected Behavior
When the environment variable MINIO_DOMAIN is set, MinIO should support virtual-host-style requests as explained in the [documentation](https://github.com/minio/minio/tree/master/docs/config#domain). For a request with a host `mybucket.miniotest.com` and with `MINIO_DOMAIN=miniotest.com`, MinIO should pick `mybucket` as the bucket.
I've been using MinIO locally with docker, and tracked down that the last release where it works according to this expected behaviour in *RELEASE.2019-03-27T22-35-21Z*, while in the release after that, *RELEASE.2019-04-04T18-31-46Z*, it does not work anymore.
## Current Behavior
From *RELEASE.2019-04-04T18-31-46Z* onwards, the result is that when requesting `mybucket.miniotest.com` with `MINIO_DOMAIN=miniotest.com`, the bucket isn't resolved correctly.
## Steps to Reproduce (for bugs)
1. Use this `docker-compose.yml`:
````
version: "3"
services:
minio:
image: minio/minio:RELEASE.2019-04-04T18-31-46Z
#image: minio/minio:RELEASE.2019-03-27T22-35-21Z # <- Works in this version
container_name: "miniotest"
hostname: miniotest.com
ports:
- "8001:9000"
volumes:
- "./minio-data:/data"
environment:
- "MINIO_ACCESS_KEY=minio"
- "MINIO_SECRET_KEY=miniopass"
- "MINIO_REGION=us-east-1"
- "MINIO_DOMAIN=miniotest.com"
command: server /data
````
2. `docker-compose up -d`
3. Put into /etc/hosts:
````
127.0.0.1 miniotest.com mybucket.miniotest.com
````
4. Doing a simple GET request with curl gives the following:
````
$ curl http://mybucket.miniotest.com:8001/mykey
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>AccessDenied</Code><Message>Access Denied.</Message><BucketName>mykey</BucketName><Resource>/mykey</Resource><RequestId>15985D7FA186B0D8</RequestId><HostId>c4afcac1-d638-4352-87b2-6374ce681e73</HostId></Error>
````
Here it incorrectly says BucketName=mykey when it should get the bucket from the host. Running the same with the older version *RELEASE.2019-03-27T22-35-21Z* gives:
````
$ curl http://mybucket.miniotest.com:8001/mykey
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>AccessDenied</Code><Message>Access Denied.</Message><Key>mykey</Key><BucketName>mybucket</BucketName><Resource>/mykey</Resource><RequestId>15985DDC715A2E78</RequestId><HostId>c4afcac1-d638-4352-87b2-6374ce681e73</HostId></Error>
````
Here the bucket is resolved correctly with BucketName=mybucket.
Tried also with having the parameter `--address miniotest.com:9000` when starting the service but that did not help.
## Context
Running MinIO locally in a docker environment which involves AWS client software that only supports virtual-host-style requests.
## Regression
Seems to be regression. Last version where it works is [RELEASE.2019-03-27T22-35-21Z](https://github.com/minio/minio/releases/tag/RELEASE.2019-03-27T22-35-21Z) when the next version [RELEASE.2019-04-04T18-31-46Z](https://github.com/minio/minio/releases/tag/RELEASE.2019-04-04T18-31-46Z) breaks the functionality.
## Your Environment
* Version used (`minio version`): RELEASE.2019-04-04T18-31-46Z
* Environment name and version (e.g. nginx 1.9.1): [MinIO docker images](https://hub.docker.com/r/minio/minio), Docker Version 2.0.0.3
| 1.0 | Bucket not resolved when using MINIO_DOMAIN - ## Expected Behavior
When the environment variable MINIO_DOMAIN is set, MinIO should support virtual-host-style requests as explained in the [documentation](https://github.com/minio/minio/tree/master/docs/config#domain). For a request with a host `mybucket.miniotest.com` and with `MINIO_DOMAIN=miniotest.com`, MinIO should pick `mybucket` as the bucket.
I've been using MinIO locally with docker, and tracked down that the last release where it works according to this expected behaviour in *RELEASE.2019-03-27T22-35-21Z*, while in the release after that, *RELEASE.2019-04-04T18-31-46Z*, it does not work anymore.
## Current Behavior
From *RELEASE.2019-04-04T18-31-46Z* onwards, the result is that when requesting `mybucket.miniotest.com` with `MINIO_DOMAIN=miniotest.com`, the bucket isn't resolved correctly.
## Steps to Reproduce (for bugs)
1. Use this `docker-compose.yml`:
````
version: "3"
services:
minio:
image: minio/minio:RELEASE.2019-04-04T18-31-46Z
#image: minio/minio:RELEASE.2019-03-27T22-35-21Z # <- Works in this version
container_name: "miniotest"
hostname: miniotest.com
ports:
- "8001:9000"
volumes:
- "./minio-data:/data"
environment:
- "MINIO_ACCESS_KEY=minio"
- "MINIO_SECRET_KEY=miniopass"
- "MINIO_REGION=us-east-1"
- "MINIO_DOMAIN=miniotest.com"
command: server /data
````
2. `docker-compose up -d`
3. Put into /etc/hosts:
````
127.0.0.1 miniotest.com mybucket.miniotest.com
````
4. Doing a simple GET request with curl gives the following:
````
$ curl http://mybucket.miniotest.com:8001/mykey
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>AccessDenied</Code><Message>Access Denied.</Message><BucketName>mykey</BucketName><Resource>/mykey</Resource><RequestId>15985D7FA186B0D8</RequestId><HostId>c4afcac1-d638-4352-87b2-6374ce681e73</HostId></Error>
````
Here it incorrectly says BucketName=mykey when it should get the bucket from the host. Running the same with the older version *RELEASE.2019-03-27T22-35-21Z* gives:
````
$ curl http://mybucket.miniotest.com:8001/mykey
<?xml version="1.0" encoding="UTF-8"?>
<Error><Code>AccessDenied</Code><Message>Access Denied.</Message><Key>mykey</Key><BucketName>mybucket</BucketName><Resource>/mykey</Resource><RequestId>15985DDC715A2E78</RequestId><HostId>c4afcac1-d638-4352-87b2-6374ce681e73</HostId></Error>
````
Here the bucket is resolved correctly with BucketName=mybucket.
Tried also with having the parameter `--address miniotest.com:9000` when starting the service but that did not help.
## Context
Running MinIO locally in a docker environment which involves AWS client software that only supports virtual-host-style requests.
## Regression
Seems to be regression. Last version where it works is [RELEASE.2019-03-27T22-35-21Z](https://github.com/minio/minio/releases/tag/RELEASE.2019-03-27T22-35-21Z) when the next version [RELEASE.2019-04-04T18-31-46Z](https://github.com/minio/minio/releases/tag/RELEASE.2019-04-04T18-31-46Z) breaks the functionality.
## Your Environment
* Version used (`minio version`): RELEASE.2019-04-04T18-31-46Z
* Environment name and version (e.g. nginx 1.9.1): [MinIO docker images](https://hub.docker.com/r/minio/minio), Docker Version 2.0.0.3
| priority | bucket not resolved when using minio domain expected behavior when the environment variable minio domain is set minio should support virtual host style requests as explained in the for a request with a host mybucket miniotest com and with minio domain miniotest com minio should pick mybucket as the bucket i ve been using minio locally with docker and tracked down that the last release where it works according to this expected behaviour in release while in the release after that release it does not work anymore current behavior from release onwards the result is that when requesting mybucket miniotest com with minio domain miniotest com the bucket isn t resolved correctly steps to reproduce for bugs use this docker compose yml version services minio image minio minio release image minio minio release works in this version container name miniotest hostname miniotest com ports volumes minio data data environment minio access key minio minio secret key miniopass minio region us east minio domain miniotest com command server data docker compose up d put into etc hosts miniotest com mybucket miniotest com doing a simple get request with curl gives the following curl accessdenied access denied mykey mykey here it incorrectly says bucketname mykey when it should get the bucket from the host running the same with the older version release gives curl accessdenied access denied mykey mybucket mykey here the bucket is resolved correctly with bucketname mybucket tried also with having the parameter address miniotest com when starting the service but that did not help context running minio locally in a docker environment which involves aws client software that only supports virtual host style requests regression seems to be regression last version where it works is when the next version breaks the functionality your environment version used minio version release environment name and version e g nginx docker version | 1 |
757,610 | 26,521,116,865 | IssuesEvent | 2023-01-19 02:47:27 | EinStealth/EinStealth | https://api.github.com/repos/EinStealth/EinStealth | closed | クリア後のResult画面を別Fragmentにする | Mobile Priority: High | # 概要
これもMainFragmentでそのままやってるので、ResultFragmentに移動するようにする
# 関連するissue
#43
# 参考
<!-- 参考資料などはここに書いてね -->
# チェックリスト
- [ ] 優先度tagを設定しましたか?
| 1.0 | クリア後のResult画面を別Fragmentにする - # 概要
これもMainFragmentでそのままやってるので、ResultFragmentに移動するようにする
# 関連するissue
#43
# 参考
<!-- 参考資料などはここに書いてね -->
# チェックリスト
- [ ] 優先度tagを設定しましたか?
| priority | クリア後のresult画面を別fragmentにする 概要 これもmainfragmentでそのままやってるので、resultfragmentに移動するようにする 関連するissue 参考 チェックリスト 優先度tagを設定しましたか? | 1 |
273,907 | 8,554,718,254 | IssuesEvent | 2018-11-08 07:40:46 | BuildCraft/BuildCraft | https://api.github.com/repos/BuildCraft/BuildCraft | closed | Chunk loading checking code doesn't work quite right | complexity: simple priority: high status: fixed/implemented in dev type: bug version: 1.12.2 | `ChunkLoaderManager.loadChunksForTile` requests a ticket from forge, then checks the config. It should instead check to see if it can load *first* and exit immediately if it shouldn't.
It would also be good to have some more (flag-enabled) logging for when chunks are loaded. (Behind a new debug flag "lib.chunkloading"). | 1.0 | Chunk loading checking code doesn't work quite right - `ChunkLoaderManager.loadChunksForTile` requests a ticket from forge, then checks the config. It should instead check to see if it can load *first* and exit immediately if it shouldn't.
It would also be good to have some more (flag-enabled) logging for when chunks are loaded. (Behind a new debug flag "lib.chunkloading"). | priority | chunk loading checking code doesn t work quite right chunkloadermanager loadchunksfortile requests a ticket from forge then checks the config it should instead check to see if it can load first and exit immediately if it shouldn t it would also be good to have some more flag enabled logging for when chunks are loaded behind a new debug flag lib chunkloading | 1 |
223,307 | 7,451,919,893 | IssuesEvent | 2018-03-29 06:07:12 | ballerina-lang/ballerina | https://api.github.com/repos/ballerina-lang/ballerina | opened | Populate the endpoint attributes within, based on the endpoint type | Priority/Highest Severity/Major Type/Improvement component/language-server | **Description:**
When the endpoint template is added, currently the attribute list is empty. What if we fill the attributes with the completion template aswell? | 1.0 | Populate the endpoint attributes within, based on the endpoint type - **Description:**
When the endpoint template is added, currently the attribute list is empty. What if we fill the attributes with the completion template aswell? | priority | populate the endpoint attributes within based on the endpoint type description when the endpoint template is added currently the attribute list is empty what if we fill the attributes with the completion template aswell | 1 |
525,567 | 15,256,342,552 | IssuesEvent | 2021-02-20 19:47:03 | actually-colab/editor | https://api.github.com/repos/actually-colab/editor | closed | Create active sessions table | database difficulty: medium priority: high socket | Create a table in dynamodb to store users with active sessions in a specific notebook
### Schema
```
uid: DUser['uid']
nb_id: DNotebook['nb_id']
time_connected: Datetime
time_disconnected?: Datetime
last_event: Datetime
```
Hash key: uid + nb_id
Range key: time_connected | 1.0 | Create active sessions table - Create a table in dynamodb to store users with active sessions in a specific notebook
### Schema
```
uid: DUser['uid']
nb_id: DNotebook['nb_id']
time_connected: Datetime
time_disconnected?: Datetime
last_event: Datetime
```
Hash key: uid + nb_id
Range key: time_connected | priority | create active sessions table create a table in dynamodb to store users with active sessions in a specific notebook schema uid duser nb id dnotebook time connected datetime time disconnected datetime last event datetime hash key uid nb id range key time connected | 1 |
266,033 | 8,361,940,696 | IssuesEvent | 2018-10-03 15:31:16 | LeedsCC/Helm-PHR-Project | https://api.github.com/repos/LeedsCC/Helm-PHR-Project | opened | OpenID Connect - end users | High Priority QEWD in progress | - [ ] 2FA
- [ ] PW complexity
- [ ] force change of temporary pw
- [ ] Password reset
-####-
Aligning with CitizenID, enforce timeout due to inactivity after 60minutes | 1.0 | OpenID Connect - end users - - [ ] 2FA
- [ ] PW complexity
- [ ] force change of temporary pw
- [ ] Password reset
-####-
Aligning with CitizenID, enforce timeout due to inactivity after 60minutes | priority | openid connect end users pw complexity force change of temporary pw password reset aligning with citizenid enforce timeout due to inactivity after | 1 |
628,685 | 20,010,780,202 | IssuesEvent | 2022-02-01 06:01:33 | TeamSparker/Spark-iOS | https://api.github.com/repos/TeamSparker/Spark-iOS | closed | [Feat] 습관방에서 멤버 상태스티커 반영 | Feat 👼타락pOwEr천사현규 P1 / Priority High | ## 📌 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
- 습관방에서 다른 멤버들의 상태가 반영되지 않음
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [x] 습관방에서 다른 멤버들의 상태 스티커가 반영되지 않음
| 1.0 | [Feat] 습관방에서 멤버 상태스티커 반영 - ## 📌 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
- 습관방에서 다른 멤버들의 상태가 반영되지 않음
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [x] 습관방에서 다른 멤버들의 상태 스티커가 반영되지 않음
| priority | 습관방에서 멤버 상태스티커 반영 📌 issue 습관방에서 다른 멤버들의 상태가 반영되지 않음 📝 to do 습관방에서 다른 멤버들의 상태 스티커가 반영되지 않음 | 1 |
647,428 | 21,103,505,753 | IssuesEvent | 2022-04-04 16:26:08 | AY2122S2-CS2103T-T12-4/tp | https://api.github.com/repos/AY2122S2-CS2103T-T12-4/tp | closed | [PE-D] Github Username needs to handle hyphens | priority.High | ## Steps to reproduce
1. enter `addc n/Nicole Lee p/12345678 e/nicole@stffhub.org u/nicole-lee` into the command line
2. System reports `Github usernames should only contain alphanumeric characters.`
## Expected
Usernames on GitHub [have hyphens in them](https://docs.github.com/en/site-policy/other-site-policies/github-username-policy). If a student's username already has a hyphen in it, NUS Classes will not be able to process it.
## Actual
NUS Classes was not able to create my `nicole-lee` contact

<!--session: 1648793079029-094b23bf-c065-4806-bff4-e64ce7b438fa--><!--Version: Web v3.4.2-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: ian-from-dover/ped#5 | 1.0 | [PE-D] Github Username needs to handle hyphens - ## Steps to reproduce
1. enter `addc n/Nicole Lee p/12345678 e/nicole@stffhub.org u/nicole-lee` into the command line
2. System reports `Github usernames should only contain alphanumeric characters.`
## Expected
Usernames on GitHub [have hyphens in them](https://docs.github.com/en/site-policy/other-site-policies/github-username-policy). If a student's username already has a hyphen in it, NUS Classes will not be able to process it.
## Actual
NUS Classes was not able to create my `nicole-lee` contact

<!--session: 1648793079029-094b23bf-c065-4806-bff4-e64ce7b438fa--><!--Version: Web v3.4.2-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: ian-from-dover/ped#5 | priority | github username needs to handle hyphens steps to reproduce enter addc n nicole lee p e nicole stffhub org u nicole lee into the command line system reports github usernames should only contain alphanumeric characters expected usernames on github if a student s username already has a hyphen in it nus classes will not be able to process it actual nus classes was not able to create my nicole lee contact labels severity medium type featureflaw original ian from dover ped | 1 |
340,553 | 10,273,176,593 | IssuesEvent | 2019-08-23 18:31:42 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | Create a sample export plugin | addition: plugin priority: high | Create a simple sample export plugin which renders the map and watermark image on top of it for testing plugin loading and component positioning. | 1.0 | Create a sample export plugin - Create a simple sample export plugin which renders the map and watermark image on top of it for testing plugin loading and component positioning. | priority | create a sample export plugin create a simple sample export plugin which renders the map and watermark image on top of it for testing plugin loading and component positioning | 1 |
147,074 | 5,633,407,228 | IssuesEvent | 2017-04-05 18:48:53 | theam/haskell-do | https://api.github.com/repos/theam/haskell-do | closed | Console not working / empty | Bug High priority | No idea if I'm just doing something wrong here but the console panel is just empty grey rectangle with nothing in it. I can click a mouse at the top of it to get an input text field there and I can write text there but nothing happens, no matter how many times I press Enter. In the README instructions, there's at least `>` symbol visible but I don't have even that. Any ideas? | 1.0 | Console not working / empty - No idea if I'm just doing something wrong here but the console panel is just empty grey rectangle with nothing in it. I can click a mouse at the top of it to get an input text field there and I can write text there but nothing happens, no matter how many times I press Enter. In the README instructions, there's at least `>` symbol visible but I don't have even that. Any ideas? | priority | console not working empty no idea if i m just doing something wrong here but the console panel is just empty grey rectangle with nothing in it i can click a mouse at the top of it to get an input text field there and i can write text there but nothing happens no matter how many times i press enter in the readme instructions there s at least symbol visible but i don t have even that any ideas | 1 |
483,850 | 13,930,482,074 | IssuesEvent | 2020-10-22 02:34:24 | alibaba/sentinel-golang | https://api.github.com/repos/alibaba/sentinel-golang | closed | Purge existing panic operations | kind/enhancement priority/high | <!-- Here is for bug reports and feature requests ONLY!
If you're looking for help, please check our mail list and the Gitter room.
Please try to use English to describe your issue, or at least provide a snippet of English translation.
-->
## Issue Description
clean up panic
Type: *feature request*
### Describe what feature you want
Replace the existing panic operation in the project with log printing,any other better Suggestions?
### Additional context
Add any other context or screenshots about the feature request here.
| 1.0 | Purge existing panic operations - <!-- Here is for bug reports and feature requests ONLY!
If you're looking for help, please check our mail list and the Gitter room.
Please try to use English to describe your issue, or at least provide a snippet of English translation.
-->
## Issue Description
clean up panic
Type: *feature request*
### Describe what feature you want
Replace the existing panic operation in the project with log printing,any other better Suggestions?
### Additional context
Add any other context or screenshots about the feature request here.
| priority | purge existing panic operations here is for bug reports and feature requests only if you re looking for help please check our mail list and the gitter room please try to use english to describe your issue or at least provide a snippet of english translation issue description clean up panic type feature request describe what feature you want replace the existing panic operation in the project with log printing,any other better suggestions? additional context add any other context or screenshots about the feature request here | 1 |
207,813 | 7,134,078,192 | IssuesEvent | 2018-01-22 19:36:33 | PARINetwork/pari | https://api.github.com/repos/PARINetwork/pari | opened | Search does not work for Non-English languages | High Priority Small Work UI good first issue | https://ruralindiaonline.org/search/?q=تعلیم+کے+جنگل+میں+چاول+پر+بحث&type=article&type=video&type=audio&language=en
https://ruralindiaonline.org/search/?q=शिक्षा+के+वन+में+चावल+पर+बहस&type=article&type=video&type=audio&language=en
This is because the default search is English. Remove the default English option. | 1.0 | Search does not work for Non-English languages - https://ruralindiaonline.org/search/?q=تعلیم+کے+جنگل+میں+چاول+پر+بحث&type=article&type=video&type=audio&language=en
https://ruralindiaonline.org/search/?q=शिक्षा+के+वन+में+चावल+पर+बहस&type=article&type=video&type=audio&language=en
This is because the default search is English. Remove the default English option. | priority | search does not work for non english languages this is because the default search is english remove the default english option | 1 |
3,294 | 2,537,630,520 | IssuesEvent | 2015-01-26 21:54:04 | newca12/gapt | https://api.github.com/repos/newca12/gapt | closed | Implementation of First order logic | 1 star Component-Logic imported Milestone-Release1.0 Priority-High Type-Task | _From [shaoli...@gmail.com](https://code.google.com/u/113190107447576027220/) on October 07, 2009 10:46:22_
to implement FOL on top of higher order logic
_Original issue: http://code.google.com/p/gapt/issues/detail?id=14_ | 1.0 | Implementation of First order logic - _From [shaoli...@gmail.com](https://code.google.com/u/113190107447576027220/) on October 07, 2009 10:46:22_
to implement FOL on top of higher order logic
_Original issue: http://code.google.com/p/gapt/issues/detail?id=14_ | priority | implementation of first order logic from on october to implement fol on top of higher order logic original issue | 1 |
194,375 | 6,894,327,851 | IssuesEvent | 2017-11-23 09:35:14 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Apply MIT license to our repository | Effort: Low Priority: High Status: Sprint | Now that we have decided to license Uwazi under the MIT license (per issue #840), we need to apply the license to our repository. Here is some [info on best practices from Github](https://help.github.com/articles/licensing-a-repository/):
Most people place their license text in a file named LICENSE.txt (or LICENSE.md) in the root of the repository; [here's an example from Hubot](https://github.com/github/hubot/blob/master/LICENSE.md).
Some projects include information about their license in their README. For example, a project's README may include a note saying "This project is licensed under the terms of the MIT license."
As a best practice, we encourage you to include the license file with your project.
| 1.0 | Apply MIT license to our repository - Now that we have decided to license Uwazi under the MIT license (per issue #840), we need to apply the license to our repository. Here is some [info on best practices from Github](https://help.github.com/articles/licensing-a-repository/):
Most people place their license text in a file named LICENSE.txt (or LICENSE.md) in the root of the repository; [here's an example from Hubot](https://github.com/github/hubot/blob/master/LICENSE.md).
Some projects include information about their license in their README. For example, a project's README may include a note saying "This project is licensed under the terms of the MIT license."
As a best practice, we encourage you to include the license file with your project.
| priority | apply mit license to our repository now that we have decided to license uwazi under the mit license per issue we need to apply the license to our repository here is some most people place their license text in a file named license txt or license md in the root of the repository some projects include information about their license in their readme for example a project s readme may include a note saying this project is licensed under the terms of the mit license as a best practice we encourage you to include the license file with your project | 1 |
168,826 | 6,387,424,165 | IssuesEvent | 2017-08-03 13:39:03 | GoogleCloudPlatform/google-cloud-eclipse | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse | closed | Link in deploy dialog to create new GCP project doesn't seem to work | bug high priority | Not sure if we had a similar issue here or internally.

No matter which account is selected, I am presented with the following screen on the Developer Console:

| 1.0 | Link in deploy dialog to create new GCP project doesn't seem to work - Not sure if we had a similar issue here or internally.

No matter which account is selected, I am presented with the following screen on the Developer Console:

| priority | link in deploy dialog to create new gcp project doesn t seem to work not sure if we had a similar issue here or internally no matter which account is selected i am presented with the following screen on the developer console | 1 |
367,749 | 10,861,472,275 | IssuesEvent | 2019-11-14 11:09:31 | wso2/analytics-apim | https://api.github.com/repos/wso2/analytics-apim | closed | Dashboard DB cannot be used with PostgreSQL | APIM 3.0.0 Docs/Has Impact Priority/High | **Description:**
When a PostgeSQL database is setup for the dashboard DB, it gives the following error on startup:
`Caused by: java.sql.SQLFeatureNotSupportedException: Method org.postgresql.jdbc.PgConnection.createBlob() is not yet implemented.`
**OS, DB, other environment details and versions:**
Postgres DB version: 12.0
Postgres JDBC driver version: 42.2.8 | 1.0 | Dashboard DB cannot be used with PostgreSQL - **Description:**
When a PostgeSQL database is setup for the dashboard DB, it gives the following error on startup:
`Caused by: java.sql.SQLFeatureNotSupportedException: Method org.postgresql.jdbc.PgConnection.createBlob() is not yet implemented.`
**OS, DB, other environment details and versions:**
Postgres DB version: 12.0
Postgres JDBC driver version: 42.2.8 | priority | dashboard db cannot be used with postgresql description when a postgesql database is setup for the dashboard db it gives the following error on startup caused by java sql sqlfeaturenotsupportedexception method org postgresql jdbc pgconnection createblob is not yet implemented os db other environment details and versions postgres db version postgres jdbc driver version | 1 |
650,111 | 21,335,193,410 | IssuesEvent | 2022-04-18 13:48:11 | NEAR-Edu/near.academy | https://api.github.com/repos/NEAR-Edu/near.academy | opened | Install Google Analytics (probably via Google Tag Manager) properly | High Priority | https://github.com/NEAR-Edu/near.academy/blob/11d7421b3aeed804c92ec7800c3d40497648ee7a/src/frontend/src/app/App.analytics.tsx seems to have no options defined and seems to use https://github.com/react-ga/react-ga which seems not to be actively maintained. | 1.0 | Install Google Analytics (probably via Google Tag Manager) properly - https://github.com/NEAR-Edu/near.academy/blob/11d7421b3aeed804c92ec7800c3d40497648ee7a/src/frontend/src/app/App.analytics.tsx seems to have no options defined and seems to use https://github.com/react-ga/react-ga which seems not to be actively maintained. | priority | install google analytics probably via google tag manager properly seems to have no options defined and seems to use which seems not to be actively maintained | 1 |
403,039 | 11,834,782,391 | IssuesEvent | 2020-03-23 09:31:34 | FStarLang/FStar | https://api.github.com/repos/FStarLang/FStar | closed | Better pretty printer | area/syntax area/usability kind/enhancement priority/high | NS: better printing of terms would improve the readability of error messages immediately (and would also be useful groundwork for better visualizations in the future)
NS: About debug level low: It's not designed to be robust right now. Hence one of my tasks on the list I sent last night. It's using a very naive normalizer to print types. That thing blows up on anything but the smallest types.
| 1.0 | Better pretty printer - NS: better printing of terms would improve the readability of error messages immediately (and would also be useful groundwork for better visualizations in the future)
NS: About debug level low: It's not designed to be robust right now. Hence one of my tasks on the list I sent last night. It's using a very naive normalizer to print types. That thing blows up on anything but the smallest types.
| priority | better pretty printer ns better printing of terms would improve the readability of error messages immediately and would also be useful groundwork for better visualizations in the future ns about debug level low it s not designed to be robust right now hence one of my tasks on the list i sent last night it s using a very naive normalizer to print types that thing blows up on anything but the smallest types | 1 |
351,580 | 10,520,730,915 | IssuesEvent | 2019-09-30 02:43:18 | AY1920S1-CS2113T-W17-2/main | https://api.github.com/repos/AY1920S1-CS2113T-W17-2/main | closed | As a user, I can update my expenses record so that I can consistently keep track of my money. | priority.High type.Story | Expenses feature - Update record for tracking | 1.0 | As a user, I can update my expenses record so that I can consistently keep track of my money. - Expenses feature - Update record for tracking | priority | as a user i can update my expenses record so that i can consistently keep track of my money expenses feature update record for tracking | 1 |
714,550 | 24,566,061,201 | IssuesEvent | 2022-10-13 03:12:24 | AY2223S1-CS2103T-T09-4/tp | https://api.github.com/repos/AY2223S1-CS2103T-T09-4/tp | closed | Add the ability to edit next of kin phone numbers | priority.HIGH type.Task type.Task.Update | Update phone number that will be of 8 digits in length, starting with 8 or 9 | 1.0 | Add the ability to edit next of kin phone numbers - Update phone number that will be of 8 digits in length, starting with 8 or 9 | priority | add the ability to edit next of kin phone numbers update phone number that will be of digits in length starting with or | 1 |
152,312 | 5,844,270,139 | IssuesEvent | 2017-05-10 11:26:02 | GluuFederation/oxTrust | https://api.github.com/repos/GluuFederation/oxTrust | closed | Unable to remove Custom Script from oxTrust UI | bug High Priority | OS: CentOS 6.8 (possibly affecting all other)
Package: 2.4.4 sp2 upgraded to 3.0.1
Ticket: https://support.gluu.org/upgrade/3982/migrated-244sp2-301-unable-to-delete-unwanted-person-authentication-scripts/
Steps to reproduce: Install 2.4.4 and update to `2.4.4-sp2` be replacing the `identity.war, `oxauth.war` from maven and upgrade to 3.0.1. Log in and try to remove custom scripts it will throw error.
Report: https://github.com/GluuFederation/gluu-qa/wiki/Itemized-Reports#upgrade-script
Observation: Issue confirmed with similar stack trace given in the ticket. | 1.0 | Unable to remove Custom Script from oxTrust UI - OS: CentOS 6.8 (possibly affecting all other)
Package: 2.4.4 sp2 upgraded to 3.0.1
Ticket: https://support.gluu.org/upgrade/3982/migrated-244sp2-301-unable-to-delete-unwanted-person-authentication-scripts/
Steps to reproduce: Install 2.4.4 and update to `2.4.4-sp2` be replacing the `identity.war, `oxauth.war` from maven and upgrade to 3.0.1. Log in and try to remove custom scripts it will throw error.
Report: https://github.com/GluuFederation/gluu-qa/wiki/Itemized-Reports#upgrade-script
Observation: Issue confirmed with similar stack trace given in the ticket. | priority | unable to remove custom script from oxtrust ui os centos possibly affecting all other package upgraded to ticket steps to reproduce install and update to be replacing the identity war oxauth war from maven and upgrade to log in and try to remove custom scripts it will throw error report observation issue confirmed with similar stack trace given in the ticket | 1 |
793,678 | 28,006,913,331 | IssuesEvent | 2023-03-27 15:48:54 | ccrcomplete/CaffeStore | https://api.github.com/repos/ccrcomplete/CaffeStore | closed | Edit Product | High Priority Useful | the admin could be edit the info of whatever product, through a button in the list, between the id | 1.0 | Edit Product - the admin could be edit the info of whatever product, through a button in the list, between the id | priority | edit product the admin could be edit the info of whatever product through a button in the list between the id | 1 |
543,270 | 15,879,340,069 | IssuesEvent | 2021-04-09 12:19:46 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | BPS profile name front end validation missing which must only contain letters and numbers | Affected/5.4.0 Affected/5.4.0-Update3 Priority/High bug | When creating BPS profile, ** Profile Name** must only contain letters and numbers. But in fronted its not validated and only giving error as
> Error when adding the BPS profile.
Below error can be seen in the backend
```
[-1234] [] [2018-01-05 07:51:38,255] ERROR {org.wso2.carbon.identity.workflow.impl.WorkflowImplAdminService} - Server error when adding the profile bps_1
org.wso2.carbon.identity.workflow.impl.WorkflowImplException: Profile name should be a not null alpha numeric string, if its not the default embedded BPS.
at org.wso2.carbon.identity.workflow.impl.listener.WorkflowImplValidationListener.doPreAddBPSProfile(WorkflowImplValidationListener.java:46)
at org.wso2.carbon.identity.workflow.impl.WorkflowImplServiceImpl.addBPSProfile(WorkflowImplServiceImpl.java:85)
at org.wso2.carbon.identity.workflow.impl.WorkflowImplAdminService.addBPSProfile(WorkflowImplAdminService.java:50)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.axis2.rpc.receivers.RPCUtil.invokeServiceClass(RPCUtil.java:212)
at org.apache.axis2.rpc.receivers.RPCMessageReceiver.invokeBusinessLogic(RPCMessageReceiver.java:117)
at org.apache.axis2.receivers.AbstractInOutMessageReceiver.invokeBusinessLogic(AbstractInOutMessageReceiver.java:40)
at org.apache.axis2.receivers.AbstractMessageReceiver.receive(AbstractMessageReceiver.java:110)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:169)
at org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:82)
at org.wso2.carbon.core.transports.local.CarbonLocalTransportSender.finalizeSendWithToAddress(CarbonLocalTransportSender.java:45)
at org.apache.axis2.transport.local.LocalTransportSender.invoke(LocalTransportSender.java:77)
at org.apache.axis2.engine.AxisEngine.send(AxisEngine.java:442)
at org.apache.axis2.description.OutInAxisOperationClient.send(OutInAxisOperation.java:430)
at org.apache.axis2.description.OutInAxisOperationClient.executeImpl(OutInAxisOperation.java:225)
at org.apache.axis2.client.OperationClient.execute(OperationClient.java:149)
at org.wso2.carbon.identity.workflow.impl.stub.WorkflowImplAdminServiceStub.addBPSProfile(WorkflowImplAdminServiceStub.java:1354)
at org.wso2.carbon.identity.workflow.impl.ui.WorkflowImplAdminServiceClient.addBPSProfile(WorkflowImplAdminServiceClient.java:62)
at org.apache.jsp.workflow_002dimpl.update_002dbps_002dprofile_002dfinish_002dajaxprocessor_jsp._jspService(update_002dbps_002dprofile_002dfinish_002dajaxprocessor_jsp.java:157)
at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:70)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:439)
at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:395)
at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:339)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.wso2.carbon.ui.JspServlet.service(JspServlet.java:155)
at org.wso2.carbon.ui.TilesJspServlet.service(TilesJspServlet.java:80)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.eclipse.equinox.http.helper.ContextPathServletAdaptor.service(ContextPathServletAdaptor.java:37)
at org.eclipse.equinox.http.servlet.internal.ServletRegistration.service(ServletRegistration.java:61)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.processAlias(ProxyServlet.java:128)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.service(ProxyServlet.java:68)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.service(DelegationServlet.java:68)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.owasp.csrfguard.CsrfGuardFilter.doFilter(CsrfGuardFilter.java:88)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:124)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.wso2.carbon.tomcat.ext.filter.CharacterSetFilter.doFilter(CharacterSetFilter.java:65)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:124)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:110)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:506)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:169)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103)
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:80)
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:91)
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:60)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99)
at org.wso2.carbon.tomcat.ext.valves.CarbonTomcatValve$1.invoke(CarbonTomcatValve.java:47)
at org.wso2.carbon.webapp.mgt.TenantLazyLoaderValve.invoke(TenantLazyLoaderValve.java:57)
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:47)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62)
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:159)
at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:962)
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:445)
at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1115)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:637)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1775)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1734)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:745)
TID
```: | 1.0 | BPS profile name front end validation missing which must only contain letters and numbers - When creating BPS profile, ** Profile Name** must only contain letters and numbers. But in fronted its not validated and only giving error as
> Error when adding the BPS profile.
Below error can be seen in the backend
```
[-1234] [] [2018-01-05 07:51:38,255] ERROR {org.wso2.carbon.identity.workflow.impl.WorkflowImplAdminService} - Server error when adding the profile bps_1
org.wso2.carbon.identity.workflow.impl.WorkflowImplException: Profile name should be a not null alpha numeric string, if its not the default embedded BPS.
at org.wso2.carbon.identity.workflow.impl.listener.WorkflowImplValidationListener.doPreAddBPSProfile(WorkflowImplValidationListener.java:46)
at org.wso2.carbon.identity.workflow.impl.WorkflowImplServiceImpl.addBPSProfile(WorkflowImplServiceImpl.java:85)
at org.wso2.carbon.identity.workflow.impl.WorkflowImplAdminService.addBPSProfile(WorkflowImplAdminService.java:50)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.axis2.rpc.receivers.RPCUtil.invokeServiceClass(RPCUtil.java:212)
at org.apache.axis2.rpc.receivers.RPCMessageReceiver.invokeBusinessLogic(RPCMessageReceiver.java:117)
at org.apache.axis2.receivers.AbstractInOutMessageReceiver.invokeBusinessLogic(AbstractInOutMessageReceiver.java:40)
at org.apache.axis2.receivers.AbstractMessageReceiver.receive(AbstractMessageReceiver.java:110)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:169)
at org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:82)
at org.wso2.carbon.core.transports.local.CarbonLocalTransportSender.finalizeSendWithToAddress(CarbonLocalTransportSender.java:45)
at org.apache.axis2.transport.local.LocalTransportSender.invoke(LocalTransportSender.java:77)
at org.apache.axis2.engine.AxisEngine.send(AxisEngine.java:442)
at org.apache.axis2.description.OutInAxisOperationClient.send(OutInAxisOperation.java:430)
at org.apache.axis2.description.OutInAxisOperationClient.executeImpl(OutInAxisOperation.java:225)
at org.apache.axis2.client.OperationClient.execute(OperationClient.java:149)
at org.wso2.carbon.identity.workflow.impl.stub.WorkflowImplAdminServiceStub.addBPSProfile(WorkflowImplAdminServiceStub.java:1354)
at org.wso2.carbon.identity.workflow.impl.ui.WorkflowImplAdminServiceClient.addBPSProfile(WorkflowImplAdminServiceClient.java:62)
at org.apache.jsp.workflow_002dimpl.update_002dbps_002dprofile_002dfinish_002dajaxprocessor_jsp._jspService(update_002dbps_002dprofile_002dfinish_002dajaxprocessor_jsp.java:157)
at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:70)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:439)
at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:395)
at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:339)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.wso2.carbon.ui.JspServlet.service(JspServlet.java:155)
at org.wso2.carbon.ui.TilesJspServlet.service(TilesJspServlet.java:80)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.eclipse.equinox.http.helper.ContextPathServletAdaptor.service(ContextPathServletAdaptor.java:37)
at org.eclipse.equinox.http.servlet.internal.ServletRegistration.service(ServletRegistration.java:61)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.processAlias(ProxyServlet.java:128)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.service(ProxyServlet.java:68)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:731)
at org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.service(DelegationServlet.java:68)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.owasp.csrfguard.CsrfGuardFilter.doFilter(CsrfGuardFilter.java:88)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:124)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.wso2.carbon.tomcat.ext.filter.CharacterSetFilter.doFilter(CharacterSetFilter.java:65)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:124)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:110)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:506)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:169)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103)
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:80)
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:91)
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:60)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99)
at org.wso2.carbon.tomcat.ext.valves.CarbonTomcatValve$1.invoke(CarbonTomcatValve.java:47)
at org.wso2.carbon.webapp.mgt.TenantLazyLoaderValve.invoke(TenantLazyLoaderValve.java:57)
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:47)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62)
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:159)
at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:962)
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:445)
at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1115)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:637)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1775)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1734)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:745)
TID
```: | priority | bps profile name front end validation missing which must only contain letters and numbers when creating bps profile profile name must only contain letters and numbers but in fronted its not validated and only giving error as error when adding the bps profile below error can be seen in the backend error org carbon identity workflow impl workflowimpladminservice server error when adding the profile bps org carbon identity workflow impl workflowimplexception profile name should be a not null alpha numeric string if its not the default embedded bps at org carbon identity workflow impl listener workflowimplvalidationlistener dopreaddbpsprofile workflowimplvalidationlistener java at org carbon identity workflow impl workflowimplserviceimpl addbpsprofile workflowimplserviceimpl java at org carbon identity workflow impl workflowimpladminservice addbpsprofile workflowimpladminservice java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache rpc receivers rpcutil invokeserviceclass rpcutil java at org apache rpc receivers rpcmessagereceiver invokebusinesslogic rpcmessagereceiver java at org apache receivers abstractinoutmessagereceiver invokebusinesslogic abstractinoutmessagereceiver java at org apache receivers abstractmessagereceiver receive abstractmessagereceiver java at org apache engine axisengine receive axisengine java at org apache transport local localtransportreceiver processmessage localtransportreceiver java at org apache transport local localtransportreceiver processmessage localtransportreceiver java at org carbon core transports local carbonlocaltransportsender finalizesendwithtoaddress carbonlocaltransportsender java at org apache transport local localtransportsender invoke localtransportsender java at org apache engine axisengine send axisengine java at org apache description outinaxisoperationclient send outinaxisoperation java at org apache description outinaxisoperationclient executeimpl outinaxisoperation java at org apache client operationclient execute operationclient java at org carbon identity workflow impl stub workflowimpladminservicestub addbpsprofile workflowimpladminservicestub java at org carbon identity workflow impl ui workflowimpladminserviceclient addbpsprofile workflowimpladminserviceclient java at org apache jsp workflow update jsp jspservice update jsp java at org apache jasper runtime httpjspbase service httpjspbase java at javax servlet http httpservlet service httpservlet java at org apache jasper servlet jspservletwrapper service jspservletwrapper java at org apache jasper servlet jspservlet servicejspfile jspservlet java at org apache jasper servlet jspservlet service jspservlet java at javax servlet http httpservlet service httpservlet java at org carbon ui jspservlet service jspservlet java at org carbon ui tilesjspservlet service tilesjspservlet java at javax servlet http httpservlet service httpservlet java at org eclipse equinox http helper contextpathservletadaptor service contextpathservletadaptor java at org eclipse equinox http servlet internal servletregistration service servletregistration java at org eclipse equinox http servlet internal proxyservlet processalias proxyservlet java at org eclipse equinox http servlet internal proxyservlet service proxyservlet java at javax servlet http httpservlet service httpservlet java at org carbon tomcat ext servlet delegationservlet service delegationservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org owasp csrfguard csrfguardfilter dofilter csrfguardfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina filters httpheadersecurityfilter dofilter httpheadersecurityfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org carbon tomcat ext filter charactersetfilter dofilter charactersetfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina filters httpheadersecurityfilter dofilter httpheadersecurityfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org carbon identity context rewrite valve tenantcontextrewritevalve invoke tenantcontextrewritevalve java at org carbon identity authz valve authorizationvalve invoke authorizationvalve java at org carbon identity auth valve authenticationvalve invoke authenticationvalve java at org carbon tomcat ext valves compositevalve continueinvocation compositevalve java at org carbon tomcat ext valves carbontomcatvalve invoke carbontomcatvalve java at org carbon webapp mgt tenantlazyloadervalve invoke tenantlazyloadervalve java at org carbon tomcat ext valves tomcatvalvecontainer invokevalves tomcatvalvecontainer java at org carbon tomcat ext valves compositevalve invoke compositevalve java at org carbon tomcat ext valves carbonstuckthreaddetectionvalve invoke carbonstuckthreaddetectionvalve java at org apache catalina valves accesslogvalve invoke accesslogvalve java at org carbon tomcat ext valves carboncontextcreatorvalve invoke carboncontextcreatorvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote process java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net nioendpoint socketprocessor run nioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java tid | 1 |
372,958 | 11,031,057,669 | IssuesEvent | 2019-12-06 16:56:11 | medic/cht-core | https://api.github.com/repos/medic/cht-core | opened | Db-doc access fails when accessing a report about a deleted contact | Priority: 1 - High Type: Bug | In an attempt to make db-doc access quicker for offline users, I implemented a reverse authorization mechanism that created a "scoped" version of the authorization context, relevant just to the requested doc. However, I completely overlooked tombstones.
As such, when requesting a report that is about a deleted contact, an offline user will receive a faulty 403 error.
**To Reproduce**
1. Login as an offline user.
2. Create a report about an existent contact.
3. Delete the contact and sync.
4. Wipe data from the device / browser and login fresh.
5. Notice that you get a 403 error when trying to get the report's attachment.
**Expected behavior**
Replication should complete without error. All docs that are received via the _changes feed should be valid for download.
**Environment**
- App: api
- Version: 3.8
**Additional context**
Introduced during the development of 3.8. Related issue: https://github.com/medic/cht-core/issues/5550
| 1.0 | Db-doc access fails when accessing a report about a deleted contact - In an attempt to make db-doc access quicker for offline users, I implemented a reverse authorization mechanism that created a "scoped" version of the authorization context, relevant just to the requested doc. However, I completely overlooked tombstones.
As such, when requesting a report that is about a deleted contact, an offline user will receive a faulty 403 error.
**To Reproduce**
1. Login as an offline user.
2. Create a report about an existent contact.
3. Delete the contact and sync.
4. Wipe data from the device / browser and login fresh.
5. Notice that you get a 403 error when trying to get the report's attachment.
**Expected behavior**
Replication should complete without error. All docs that are received via the _changes feed should be valid for download.
**Environment**
- App: api
- Version: 3.8
**Additional context**
Introduced during the development of 3.8. Related issue: https://github.com/medic/cht-core/issues/5550
| priority | db doc access fails when accessing a report about a deleted contact in an attempt to make db doc access quicker for offline users i implemented a reverse authorization mechanism that created a scoped version of the authorization context relevant just to the requested doc however i completely overlooked tombstones as such when requesting a report that is about a deleted contact an offline user will receive a faulty error to reproduce login as an offline user create a report about an existent contact delete the contact and sync wipe data from the device browser and login fresh notice that you get a error when trying to get the report s attachment expected behavior replication should complete without error all docs that are received via the changes feed should be valid for download environment app api version additional context introduced during the development of related issue | 1 |
99,538 | 4,056,704,383 | IssuesEvent | 2016-05-24 19:32:41 | RestComm/mediaserver | https://api.github.com/repos/RestComm/mediaserver | opened | Expose MS external address on MGCP Endpoints | bug High-Priority | If Media Server is configured to expose a public address, then this address should be bound to the MGCP endpoints as well. Currently, the endpoints are always bound to the private address.
If Media Server is installed in a different server than RestComm and it only exposes the private address as result of MGCP operations, then this may cause interop issues with RestComm:
- RestComm will send CRCX to external address of MS
- MS replies with 200 OK but new endpoint ID will be bound to MS internal address.
- RestComm will read endpoint ID and verifies domain is different. As result, subsequent requests will be sent to MS internal address.
The attached pcap
[remoteRMS1RC.pcap](https://github.com/RestComm/mediaserver/files/280343/remoteRMS1RC.pcap.zip)
shows an example of the problem. | 1.0 | Expose MS external address on MGCP Endpoints - If Media Server is configured to expose a public address, then this address should be bound to the MGCP endpoints as well. Currently, the endpoints are always bound to the private address.
If Media Server is installed in a different server than RestComm and it only exposes the private address as result of MGCP operations, then this may cause interop issues with RestComm:
- RestComm will send CRCX to external address of MS
- MS replies with 200 OK but new endpoint ID will be bound to MS internal address.
- RestComm will read endpoint ID and verifies domain is different. As result, subsequent requests will be sent to MS internal address.
The attached pcap
[remoteRMS1RC.pcap](https://github.com/RestComm/mediaserver/files/280343/remoteRMS1RC.pcap.zip)
shows an example of the problem. | priority | expose ms external address on mgcp endpoints if media server is configured to expose a public address then this address should be bound to the mgcp endpoints as well currently the endpoints are always bound to the private address if media server is installed in a different server than restcomm and it only exposes the private address as result of mgcp operations then this may cause interop issues with restcomm restcomm will send crcx to external address of ms ms replies with ok but new endpoint id will be bound to ms internal address restcomm will read endpoint id and verifies domain is different as result subsequent requests will be sent to ms internal address the attached pcap shows an example of the problem | 1 |
539,298 | 15,786,620,455 | IssuesEvent | 2021-04-01 18:00:48 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | reopened | Parts of reconciliation batches can fail silently; and parts of reconciled column lookup batches can fail silently | bug priority: High reconciliation | This may be a duplicate of an existing bug or bugs, but I find it incredibly frustrating: when items are 'unmatched' it appears to be impossible to distinguish whether there is really no match for the item, or whether the reconciliation process simply failed to complete for some sub-batch of entries.
A couple of days ago, I was running a reconciliation of a set of authors with LoC authority IDs against Wikidata using the "reconcili.link" service. I ran it once, and about 80% reconciled. I then filtered for the unmatched authors, reconciled again, and a further 15% of the original set reconciled.
I presume that what happened was that during the first reconciliation, one or more sub-batches of contiguous records timed out, or otherwise failed to return. (The problem did appear to be affecting contiguous records). That, I guess, is something that can always happen. The problem is that OpenRefine then marked those records' reconciliation status as "unmatched" rather than "unknown", not differentiating between items for which the reconcilation had failed to complete from those for which the reconcilation had concluded but found nothing.
Similarly, when I then added a column looking up LoC IDs based on the reconciled values (which I find is a necessary thing to have to do to sanity check reconcilations), only about 80% of the IDs got added, with more added when I re-ran the lookup for those without values.
This is frustrating enough for me, aware that this can happen, knowing that I may need to re-run reconciliations or augmentations two or three or even more times to be sure. But the even worse impact is with unaware users, who as a result may then add new duplicate items to Wikidata for their unreconciled entries, without realising that the unreconciled status cannot be depended on.
| 1.0 | Parts of reconciliation batches can fail silently; and parts of reconciled column lookup batches can fail silently - This may be a duplicate of an existing bug or bugs, but I find it incredibly frustrating: when items are 'unmatched' it appears to be impossible to distinguish whether there is really no match for the item, or whether the reconciliation process simply failed to complete for some sub-batch of entries.
A couple of days ago, I was running a reconciliation of a set of authors with LoC authority IDs against Wikidata using the "reconcili.link" service. I ran it once, and about 80% reconciled. I then filtered for the unmatched authors, reconciled again, and a further 15% of the original set reconciled.
I presume that what happened was that during the first reconciliation, one or more sub-batches of contiguous records timed out, or otherwise failed to return. (The problem did appear to be affecting contiguous records). That, I guess, is something that can always happen. The problem is that OpenRefine then marked those records' reconciliation status as "unmatched" rather than "unknown", not differentiating between items for which the reconcilation had failed to complete from those for which the reconcilation had concluded but found nothing.
Similarly, when I then added a column looking up LoC IDs based on the reconciled values (which I find is a necessary thing to have to do to sanity check reconcilations), only about 80% of the IDs got added, with more added when I re-ran the lookup for those without values.
This is frustrating enough for me, aware that this can happen, knowing that I may need to re-run reconciliations or augmentations two or three or even more times to be sure. But the even worse impact is with unaware users, who as a result may then add new duplicate items to Wikidata for their unreconciled entries, without realising that the unreconciled status cannot be depended on.
| priority | parts of reconciliation batches can fail silently and parts of reconciled column lookup batches can fail silently this may be a duplicate of an existing bug or bugs but i find it incredibly frustrating when items are unmatched it appears to be impossible to distinguish whether there is really no match for the item or whether the reconciliation process simply failed to complete for some sub batch of entries a couple of days ago i was running a reconciliation of a set of authors with loc authority ids against wikidata using the reconcili link service i ran it once and about reconciled i then filtered for the unmatched authors reconciled again and a further of the original set reconciled i presume that what happened was that during the first reconciliation one or more sub batches of contiguous records timed out or otherwise failed to return the problem did appear to be affecting contiguous records that i guess is something that can always happen the problem is that openrefine then marked those records reconciliation status as unmatched rather than unknown not differentiating between items for which the reconcilation had failed to complete from those for which the reconcilation had concluded but found nothing similarly when i then added a column looking up loc ids based on the reconciled values which i find is a necessary thing to have to do to sanity check reconcilations only about of the ids got added with more added when i re ran the lookup for those without values this is frustrating enough for me aware that this can happen knowing that i may need to re run reconciliations or augmentations two or three or even more times to be sure but the even worse impact is with unaware users who as a result may then add new duplicate items to wikidata for their unreconciled entries without realising that the unreconciled status cannot be depended on | 1 |
220,186 | 7,353,942,006 | IssuesEvent | 2018-03-09 03:35:59 | openaq/openaq-fetch | https://api.github.com/repos/openaq/openaq-fetch | closed | New adapter needed for India Data Source | help wanted high priority new data | The Central Pollution Control Board has a beautiful new website to share AQ data, but since we access the data by scraping and not programmatically, this unfortunately breaks the existing adapter for this data source, and we will need a new adapter to connect to the new website.
Here is the new website with the station-level raw data:
https://app.cpcbccr.com/ccr/#/caaqm-dashboard-all/caaqm-landing
Here is the previous adapter: https://github.com/openaq/openaq-fetch/blob/develop/adapters/cpcb.js
Thanks very much in advance for any and all code contributions to get this back up and operating for the community!
| 1.0 | New adapter needed for India Data Source - The Central Pollution Control Board has a beautiful new website to share AQ data, but since we access the data by scraping and not programmatically, this unfortunately breaks the existing adapter for this data source, and we will need a new adapter to connect to the new website.
Here is the new website with the station-level raw data:
https://app.cpcbccr.com/ccr/#/caaqm-dashboard-all/caaqm-landing
Here is the previous adapter: https://github.com/openaq/openaq-fetch/blob/develop/adapters/cpcb.js
Thanks very much in advance for any and all code contributions to get this back up and operating for the community!
| priority | new adapter needed for india data source the central pollution control board has a beautiful new website to share aq data but since we access the data by scraping and not programmatically this unfortunately breaks the existing adapter for this data source and we will need a new adapter to connect to the new website here is the new website with the station level raw data here is the previous adapter thanks very much in advance for any and all code contributions to get this back up and operating for the community | 1 |
64,606 | 3,213,214,707 | IssuesEvent | 2015-10-06 18:52:07 | cs2103aug2015-w15-2j/main | https://api.github.com/repos/cs2103aug2015-w15-2j/main | closed | A user can delete tasks and events | priority.high type.story | so that the user can remove items that the user doesn't want to do anymore / remove completed task with no archive intentions / remove items keyed in wrongly. | 1.0 | A user can delete tasks and events - so that the user can remove items that the user doesn't want to do anymore / remove completed task with no archive intentions / remove items keyed in wrongly. | priority | a user can delete tasks and events so that the user can remove items that the user doesn t want to do anymore remove completed task with no archive intentions remove items keyed in wrongly | 1 |
316,619 | 9,652,521,647 | IssuesEvent | 2019-05-18 17:54:18 | openshift/odo | https://api.github.com/repos/openshift/odo | closed | .gitignore and .odoignore not working correctly under Windows | kind/bug os/windows priority/High | [kind/bug]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## What versions of software are you using?
- Operating System: Windows
- Output of `odo version`: master
## How did you run odo exactly?
Using `odo watch` after deploying a component
## Actual behavior
.odoignore and .gitignore isn't working correctly.
## Expected behavior
To ignore files correctly
## Any logs, error output, etc?
See: https://github.com/openshift/odo/pull/1601 after findings implementing the relative path PR, it looks like ignoring files hasn't worked on Windows for a while..
| 1.0 | .gitignore and .odoignore not working correctly under Windows - [kind/bug]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## What versions of software are you using?
- Operating System: Windows
- Output of `odo version`: master
## How did you run odo exactly?
Using `odo watch` after deploying a component
## Actual behavior
.odoignore and .gitignore isn't working correctly.
## Expected behavior
To ignore files correctly
## Any logs, error output, etc?
See: https://github.com/openshift/odo/pull/1601 after findings implementing the relative path PR, it looks like ignoring files hasn't worked on Windows for a while..
| priority | gitignore and odoignore not working correctly under windows welcome we kindly ask you to fill out the issue template below use the chat and talk to us if you have a question rather than a bug or feature request the chat room is at thanks for understanding and for contributing to the project what versions of software are you using operating system windows output of odo version master how did you run odo exactly using odo watch after deploying a component actual behavior odoignore and gitignore isn t working correctly expected behavior to ignore files correctly any logs error output etc see after findings implementing the relative path pr it looks like ignoring files hasn t worked on windows for a while | 1 |
605,069 | 18,724,156,821 | IssuesEvent | 2021-11-03 14:46:02 | canonical-web-and-design/snapcraft.io | https://api.github.com/repos/canonical-web-and-design/snapcraft.io | closed | Opening and closing menu on /admin crashes the view | Priority: High | ### Expected behaviour
Opening and closing the menu on small screens should not change the view in anyway.
### Steps to reproduce the problem
- Grab your phone and go to https://snapcraft.io/admin
- Login and see the view load correctly
- Hit the menu in the primary nav
- Close the primary menu and see a infinite spinner | 1.0 | Opening and closing menu on /admin crashes the view - ### Expected behaviour
Opening and closing the menu on small screens should not change the view in anyway.
### Steps to reproduce the problem
- Grab your phone and go to https://snapcraft.io/admin
- Login and see the view load correctly
- Hit the menu in the primary nav
- Close the primary menu and see a infinite spinner | priority | opening and closing menu on admin crashes the view expected behaviour opening and closing the menu on small screens should not change the view in anyway steps to reproduce the problem grab your phone and go to login and see the view load correctly hit the menu in the primary nav close the primary menu and see a infinite spinner | 1 |
658,662 | 21,899,546,577 | IssuesEvent | 2022-05-20 12:06:45 | bounswe/bounswe2022group9 | https://api.github.com/repos/bounswe/bounswe2022group9 | closed | Practice-App: Create an API for practice-app | Enhancement Priority: High Practice Application | Deadline: 19.05.2022 23.00
- [x] Design and create an API that has GET and POST methods
- [x] GET method returns a JSON that includes a list of all countries and their codes
- [x] POST method should return the information of the country that was sent with POST call | 1.0 | Practice-App: Create an API for practice-app - Deadline: 19.05.2022 23.00
- [x] Design and create an API that has GET and POST methods
- [x] GET method returns a JSON that includes a list of all countries and their codes
- [x] POST method should return the information of the country that was sent with POST call | priority | practice app create an api for practice app deadline design and create an api that has get and post methods get method returns a json that includes a list of all countries and their codes post method should return the information of the country that was sent with post call | 1 |
703,201 | 24,150,184,176 | IssuesEvent | 2022-09-21 23:16:47 | OregonDigital/OD2 | https://api.github.com/repos/OregonDigital/OD2 | closed | Search from collection landing page returns inaccurate results | Bug Metadata Priority - High Features QA | ### Descriptive summary
On the collection landing page in "Search this collection," searching using any of the free text fields (All Fields, Title, Creator, Description, Subject) returns all items in the collection no matter the search term used. This is true whether the search string actually occurs in all records, some records, or no records.
### Expected behavior
Searching in a field from the collection landing page returns works in that collection that contain the specified search string in the specified field.
### Related work
| 1.0 | Search from collection landing page returns inaccurate results - ### Descriptive summary
On the collection landing page in "Search this collection," searching using any of the free text fields (All Fields, Title, Creator, Description, Subject) returns all items in the collection no matter the search term used. This is true whether the search string actually occurs in all records, some records, or no records.
### Expected behavior
Searching in a field from the collection landing page returns works in that collection that contain the specified search string in the specified field.
### Related work
| priority | search from collection landing page returns inaccurate results descriptive summary on the collection landing page in search this collection searching using any of the free text fields all fields title creator description subject returns all items in the collection no matter the search term used this is true whether the search string actually occurs in all records some records or no records expected behavior searching in a field from the collection landing page returns works in that collection that contain the specified search string in the specified field related work | 1 |
541,231 | 15,823,623,843 | IssuesEvent | 2021-04-06 01:14:03 | Conder-Systems/tuna-lang | https://api.github.com/repos/Conder-Systems/tuna-lang | closed | Local testing framework | High Priority | Test can run in memory.
Should look like this:
```
test ValidateSomeBehavior() {
... do things
assertions...
}
```
Out of scope:
parameterization
run against database. | 1.0 | Local testing framework - Test can run in memory.
Should look like this:
```
test ValidateSomeBehavior() {
... do things
assertions...
}
```
Out of scope:
parameterization
run against database. | priority | local testing framework test can run in memory should look like this test validatesomebehavior do things assertions out of scope parameterization run against database | 1 |
183,641 | 6,690,181,401 | IssuesEvent | 2017-10-09 08:01:16 | bitshares/bitshares-ui | https://api.github.com/repos/bitshares/bitshares-ui | closed | [1] Continued issues with node stats | bug high priority | This needs to be resolved and is a high priority issue. We are giving incorrect information and it causes people to mistrust the network. I'm giving it 4 hours. I really hope someone will claim this and work through it.
Nodes that are clearly active and running are being displayed as down with no latency indicators. @xeroc if you can lend a hand here, it's my opinion that this should not be delayed.

| 1.0 | [1] Continued issues with node stats - This needs to be resolved and is a high priority issue. We are giving incorrect information and it causes people to mistrust the network. I'm giving it 4 hours. I really hope someone will claim this and work through it.
Nodes that are clearly active and running are being displayed as down with no latency indicators. @xeroc if you can lend a hand here, it's my opinion that this should not be delayed.

| priority | continued issues with node stats this needs to be resolved and is a high priority issue we are giving incorrect information and it causes people to mistrust the network i m giving it hours i really hope someone will claim this and work through it nodes that are clearly active and running are being displayed as down with no latency indicators xeroc if you can lend a hand here it s my opinion that this should not be delayed | 1 |
580,744 | 17,265,727,754 | IssuesEvent | 2021-07-22 13:35:57 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | v11: testing AD sources doesn't work anymore | Priority: High Type: Bug | **Describe the bug**
if I create an AD source using Venom and try to test it using API call, it fail with following error:
```
:[] message:Attribute (verify) does not pass the type constraint because: Validation failed for 'Str' with value undef at constructor pf::Authentication::Source::ADSource::new (defined at /usr/local/pf/lib/pf/Authentication/Source/ADSource.pm line 96) line 422
pf::Authentication::Source::ADSource::new('pf::Authentication::Source::ADSource', 'HASH(0x55b7beb496e0)') called at /usr/local/pf/lib/pf/authentication.pm line 128
pf::authentication::newAuthenticationSource('AD', 'test', 'HASH(0x55b7bf57a130)') called at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/Sources.pm line 153
eval {...} at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/Sources.pm line 152
pf::UnifiedApi::Controller::Config::Sources::test('pf::UnifiedApi::Controller::Config::Sources=HASH(0x55b7c0566c30)') called at /usr/local/pf/lib_perl/lib/perl5/Mojolicious.pm line 190
```
**To Reproduce**
Steps to reproduce the behavior:
1. Create a valid AD source and try to test it
**Screenshots**
I see a different error when I test using UI

**Additional context**
Seems related to #6363 | 1.0 | v11: testing AD sources doesn't work anymore - **Describe the bug**
if I create an AD source using Venom and try to test it using API call, it fail with following error:
```
:[] message:Attribute (verify) does not pass the type constraint because: Validation failed for 'Str' with value undef at constructor pf::Authentication::Source::ADSource::new (defined at /usr/local/pf/lib/pf/Authentication/Source/ADSource.pm line 96) line 422
pf::Authentication::Source::ADSource::new('pf::Authentication::Source::ADSource', 'HASH(0x55b7beb496e0)') called at /usr/local/pf/lib/pf/authentication.pm line 128
pf::authentication::newAuthenticationSource('AD', 'test', 'HASH(0x55b7bf57a130)') called at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/Sources.pm line 153
eval {...} at /usr/local/pf/lib/pf/UnifiedApi/Controller/Config/Sources.pm line 152
pf::UnifiedApi::Controller::Config::Sources::test('pf::UnifiedApi::Controller::Config::Sources=HASH(0x55b7c0566c30)') called at /usr/local/pf/lib_perl/lib/perl5/Mojolicious.pm line 190
```
**To Reproduce**
Steps to reproduce the behavior:
1. Create a valid AD source and try to test it
**Screenshots**
I see a different error when I test using UI

**Additional context**
Seems related to #6363 | priority | testing ad sources doesn t work anymore describe the bug if i create an ad source using venom and try to test it using api call it fail with following error message attribute verify does not pass the type constraint because validation failed for str with value undef at constructor pf authentication source adsource new defined at usr local pf lib pf authentication source adsource pm line line pf authentication source adsource new pf authentication source adsource hash called at usr local pf lib pf authentication pm line pf authentication newauthenticationsource ad test hash called at usr local pf lib pf unifiedapi controller config sources pm line eval at usr local pf lib pf unifiedapi controller config sources pm line pf unifiedapi controller config sources test pf unifiedapi controller config sources hash called at usr local pf lib perl lib mojolicious pm line to reproduce steps to reproduce the behavior create a valid ad source and try to test it screenshots i see a different error when i test using ui additional context seems related to | 1 |
330,203 | 10,036,688,748 | IssuesEvent | 2019-07-18 11:19:25 | canonical-web-and-design/ubuntu.com | https://api.github.com/repos/canonical-web-and-design/ubuntu.com | closed | /server/hyperscale - Form issue | Priority: High | There needs to be space added above the h3 title 'Here’s why Ubuntu is the best answer to the scale-out challenge:'
The contact form in the section 'Your roadmap for hyperscale success' which states 'Contact us about hyperscale ›' - when you go through and complete the form, it says "Thanks for enquiring about Ubuntu Advantage" not Hyperscale.
<img width="1122" alt="Screenshot 2019-06-07 at 11 39 50" src="https://user-images.githubusercontent.com/42238722/59098875-7fa1ab00-8919-11e9-9207-04dc6aadc4ad.png">
---
*Reported from: https://www.ubuntu.com/server/hyperscale* | 1.0 | /server/hyperscale - Form issue - There needs to be space added above the h3 title 'Here’s why Ubuntu is the best answer to the scale-out challenge:'
The contact form in the section 'Your roadmap for hyperscale success' which states 'Contact us about hyperscale ›' - when you go through and complete the form, it says "Thanks for enquiring about Ubuntu Advantage" not Hyperscale.
<img width="1122" alt="Screenshot 2019-06-07 at 11 39 50" src="https://user-images.githubusercontent.com/42238722/59098875-7fa1ab00-8919-11e9-9207-04dc6aadc4ad.png">
---
*Reported from: https://www.ubuntu.com/server/hyperscale* | priority | server hyperscale form issue there needs to be space added above the title here’s why ubuntu is the best answer to the scale out challenge the contact form in the section your roadmap for hyperscale success which states contact us about hyperscale › when you go through and complete the form it says thanks for enquiring about ubuntu advantage not hyperscale img width alt screenshot at src reported from | 1 |
154,498 | 5,919,891,379 | IssuesEvent | 2017-05-22 18:52:01 | jacobrh91/Swift-T-Variant-Calling | https://api.github.com/repos/jacobrh91/Swift-T-Variant-Calling | closed | Make workflow keep going even if one branch of loop fails for as long as possible | enhancement High Priority | Even if a given sample fails during the analysis within a loop, the rest of the loop should keep going until no further progress can be made without all of the samples being present (the "gathering" stages of the workflow, where branches are brought back together, such as the Joint Genotyping stage)
If we add this feature, it is obvious that we should then add another feature that would allow one to jump back into the workflow right before the "gathering" stage, either after the failure with the killed sample branch has been resolved, or the "gathering" can take place without the output of the failed branch.
Potential Solutions:
Write app calls that catch non-zero exits, do something, and return an exit code of 0 to swift | 1.0 | Make workflow keep going even if one branch of loop fails for as long as possible - Even if a given sample fails during the analysis within a loop, the rest of the loop should keep going until no further progress can be made without all of the samples being present (the "gathering" stages of the workflow, where branches are brought back together, such as the Joint Genotyping stage)
If we add this feature, it is obvious that we should then add another feature that would allow one to jump back into the workflow right before the "gathering" stage, either after the failure with the killed sample branch has been resolved, or the "gathering" can take place without the output of the failed branch.
Potential Solutions:
Write app calls that catch non-zero exits, do something, and return an exit code of 0 to swift | priority | make workflow keep going even if one branch of loop fails for as long as possible even if a given sample fails during the analysis within a loop the rest of the loop should keep going until no further progress can be made without all of the samples being present the gathering stages of the workflow where branches are brought back together such as the joint genotyping stage if we add this feature it is obvious that we should then add another feature that would allow one to jump back into the workflow right before the gathering stage either after the failure with the killed sample branch has been resolved or the gathering can take place without the output of the failed branch potential solutions write app calls that catch non zero exits do something and return an exit code of to swift | 1 |
120,960 | 4,803,090,059 | IssuesEvent | 2016-11-02 08:56:49 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | When adding a layer to the TOC the layer goes to bottom rather than to top | enhancement Priority: High | When adding a new layer from the catalog, the layer ends up in the TOC but at the bottom of it which I think is misleading for our users.
I would expect the layer to appear on top of the existing layers.
| 1.0 | When adding a layer to the TOC the layer goes to bottom rather than to top - When adding a new layer from the catalog, the layer ends up in the TOC but at the bottom of it which I think is misleading for our users.
I would expect the layer to appear on top of the existing layers.
| priority | when adding a layer to the toc the layer goes to bottom rather than to top when adding a new layer from the catalog the layer ends up in the toc but at the bottom of it which i think is misleading for our users i would expect the layer to appear on top of the existing layers | 1 |
225,469 | 7,481,920,128 | IssuesEvent | 2018-04-04 22:28:37 | EvictionLab/eviction-maps | https://api.github.com/repos/EvictionLab/eviction-maps | closed | General accessibility updates, review | enhancement high priority | Creating this as a catchall for any accessibility issues that we need to fix | 1.0 | General accessibility updates, review - Creating this as a catchall for any accessibility issues that we need to fix | priority | general accessibility updates review creating this as a catchall for any accessibility issues that we need to fix | 1 |
601,003 | 18,363,394,743 | IssuesEvent | 2021-10-09 16:22:50 | AY2122S1-CS2103T-W11-3/tp | https://api.github.com/repos/AY2122S1-CS2103T-W11-3/tp | closed | Update basic Functionalities for Task component | type.Epic priority.High | - Add a module : addMod
- Delete a module : deleteMod
- Update a module : updateMod | 1.0 | Update basic Functionalities for Task component - - Add a module : addMod
- Delete a module : deleteMod
- Update a module : updateMod | priority | update basic functionalities for task component add a module addmod delete a module deletemod update a module updatemod | 1 |
113,582 | 4,565,471,315 | IssuesEvent | 2016-09-15 00:36:52 | SpeedCurve-Metrics/SpeedCurve | https://api.github.com/repos/SpeedCurve-Metrics/SpeedCurve | closed | Duplication of Device and Server Selection | priority high type bug | I don't know if this issue is specific to my account but over the last few days I keep getting a huge duplication of the device and server selections on the benchmarking page.
Please find image attached:

Regards,
Ric Riley | 1.0 | Duplication of Device and Server Selection - I don't know if this issue is specific to my account but over the last few days I keep getting a huge duplication of the device and server selections on the benchmarking page.
Please find image attached:

Regards,
Ric Riley | priority | duplication of device and server selection i don t know if this issue is specific to my account but over the last few days i keep getting a huge duplication of the device and server selections on the benchmarking page please find image attached regards ric riley | 1 |
673,141 | 22,949,237,993 | IssuesEvent | 2022-07-19 05:23:31 | okTurtles/group-income | https://api.github.com/repos/okTurtles/group-income | closed | Implement payment distribution date dropdown | Note:Up-for-grabs App:Frontend Priority:High Note:UI/UX | ### Problem
The new posthumous distribution needs a distribution date picker!
### Solution
Implement this design: https://www.figma.com/file/mxGadAHfkWH6qApebQvcdN/Group-Income-2.0?node-id=22579%3A203497
Do not set a default date. Make sure that the user cannot move on until they select a date themselves. | 1.0 | Implement payment distribution date dropdown - ### Problem
The new posthumous distribution needs a distribution date picker!
### Solution
Implement this design: https://www.figma.com/file/mxGadAHfkWH6qApebQvcdN/Group-Income-2.0?node-id=22579%3A203497
Do not set a default date. Make sure that the user cannot move on until they select a date themselves. | priority | implement payment distribution date dropdown problem the new posthumous distribution needs a distribution date picker solution implement this design do not set a default date make sure that the user cannot move on until they select a date themselves | 1 |
633,468 | 20,255,559,548 | IssuesEvent | 2022-02-14 22:43:43 | bcgov/foi-flow | https://api.github.com/repos/bcgov/foi-flow | closed | UX Assurance - Date Format Display | bug UX high priority | User feedback at the last sprint demo indicated a strong preference to have all dates in the application displayed in MMM DD YYYY format (i.e. DEC 12 2021)
- [x] Update date format in Intake Team queue
- [x] Update date format on Ministry Queue
- [ ] Update date format on Request view for Ministry users
Intake Team Queue:

Ministry Queue:

Ministry Request View:
 | 1.0 | UX Assurance - Date Format Display - User feedback at the last sprint demo indicated a strong preference to have all dates in the application displayed in MMM DD YYYY format (i.e. DEC 12 2021)
- [x] Update date format in Intake Team queue
- [x] Update date format on Ministry Queue
- [ ] Update date format on Request view for Ministry users
Intake Team Queue:

Ministry Queue:

Ministry Request View:
 | priority | ux assurance date format display user feedback at the last sprint demo indicated a strong preference to have all dates in the application displayed in mmm dd yyyy format i e dec update date format in intake team queue update date format on ministry queue update date format on request view for ministry users intake team queue ministry queue ministry request view | 1 |
188,640 | 6,778,657,647 | IssuesEvent | 2017-10-28 13:48:19 | dankito/DeepThought | https://api.github.com/repos/dankito/DeepThought | closed | Read later entries: "Save" option might be misleading | Priority High question | Entries in the "Read Later" section have an option to "Save" the artice. This means that the article is moved to the "Entries" section. But isn't "Save" misleading here, because the article was already saved in order to read it later.
### Proposal
Personally, I thought the save button was a bug at first. Especially after the artice was "gone" after pressing "Save". Ok, it was **not** gone but moved to the "Entries" section, but since I was still in the "Read Later" section, the first thing that I thouht was "Hmm, why is the article gone now after saving?".
So, my proposal would be to call this command "Remember", "Keep" (dt: Merken, Behalten) or anything like that. | 1.0 | Read later entries: "Save" option might be misleading - Entries in the "Read Later" section have an option to "Save" the artice. This means that the article is moved to the "Entries" section. But isn't "Save" misleading here, because the article was already saved in order to read it later.
### Proposal
Personally, I thought the save button was a bug at first. Especially after the artice was "gone" after pressing "Save". Ok, it was **not** gone but moved to the "Entries" section, but since I was still in the "Read Later" section, the first thing that I thouht was "Hmm, why is the article gone now after saving?".
So, my proposal would be to call this command "Remember", "Keep" (dt: Merken, Behalten) or anything like that. | priority | read later entries save option might be misleading entries in the read later section have an option to save the artice this means that the article is moved to the entries section but isn t save misleading here because the article was already saved in order to read it later proposal personally i thought the save button was a bug at first especially after the artice was gone after pressing save ok it was not gone but moved to the entries section but since i was still in the read later section the first thing that i thouht was hmm why is the article gone now after saving so my proposal would be to call this command remember keep dt merken behalten or anything like that | 1 |
314,015 | 9,583,387,266 | IssuesEvent | 2019-05-08 05:13:46 | actiontech/dble | https://api.github.com/repos/actiontech/dble | closed | only part of multiple statement executed,and hang when try to get resultsets | Priority/High resolve | * **dble version:**
5.6.29-dble-9.9.9.9-2d18ed6c8dae540e4becc2ba03472b2284238f58-20190429070246
* **preconditions :**
complete dble configue refer to autotest sql_cover_mixed_bk
complete c++ code for test refer to autotest c_mysql_api/mysql_real_connect.cpp
* **configs:**
**schema.xml**
```
<schema name="schema1" sqlMaxLimit="100" dataNode="dn5">
<table name="sharding_4_t1" dataNode="dn1,dn2,dn3,dn4" rule="hash-four" />
</schema>
<dataNode name="dn1" dataHost="172.100.9.5" database="db1" />
<dataNode name="dn2" dataHost="172.100.9.6" database="db1" />
<dataNode name="dn3" dataHost="172.100.9.5" database="db2" />
<dataNode name="dn4" dataHost="172.100.9.6" database="db2" />
<dataNode name="dn5" dataHost="172.100.9.5" database="db3" />
```
**rule.xml**
```
```
**server.xml**
```
```
* **steps:**
step1. execute multiple statements with c++
```mysql
use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
CREATE TABLE sharding_4_t1(id INT);\
INSERT INTO sharding_4_t1 VALUES(10);\
INSERT INTO sharding_4_t1 VALUES(20);\
INSERT INTO sharding_4_t1 VALUES(30);\
SELECT * FROM sharding_4_t1;
```
reference c++ code:
```c++
void case_mysql_real_connect(MYSQL* conn){
// case1: default_db=null, port=0
MYSQL* test_conn = mysql_init(NULL);
mysql_options(test_conn,MYSQL_OPT_COMPRESS,"0");
mysql_options(test_conn,MYSQL_READ_DEFAULT_GROUP,"odbc");
mysql_options(test_conn,MYSQL_INIT_COMMAND,"SET autocommit=0");
mysql_real_connect(test_conn, HOST_DBLE, TEST_USER, TEST_USER_PASSWD, NULL, DBLE_PORT,NULL, CLIENT_DEPRECATE_EOF|CLIENT_MULTI_STATEMENTS);
if (test_conn == NULL) {
printf("Error connecting to database: %s\n", mysql_error(test_conn));
exit(1);
}else{
printf("connect success");
}
//case: multi query, and multi resultsets
/* execute multiple statements */
int status = mysql_query(test_conn,
"use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
CREATE TABLE sharding_4_t1(id INT);\
INSERT INTO sharding_4_t1 VALUES(10);\
INSERT INTO sharding_4_t1 VALUES(20);\
INSERT INTO sharding_4_t1 VALUES(30);\
SELECT * FROM sharding_4_t1;");
if (status)
{
printf("execute multi statement(s) Err, %s \n", mysql_error(test_conn));
mysql_close(test_conn);
exit(1);
}else{
printf(" pass! multi sqls success\n");
}
printf(" print multi resultsets:\n");
/* process each statement result */
do {
/* did current statement return data? */
MYSQL_RES *result = mysql_store_result(test_conn);
if (result)
{
/* yes; process rows and free the result set */
doPrintResult(test_conn, result);
mysql_free_result(result);
}
else /* no result set or error */
{
if (mysql_field_count(test_conn) == 0)
{
printf(" %lld rows affected\n",
mysql_affected_rows(test_conn));
}
else /* some error occurred */
{
printf("Could not retrieve result set\n");
exit(1);
}
}
if(mysql_more_results(test_conn)){
printf(" More results exist\n");
}
/* more results? -1 = no, >0 = error, 0 = yes (keep looping) */
if ((status = mysql_next_result(test_conn)) > 0){
printf("Could not get next result\n");
exit(1);
}
} while (status == 0);
mysql_autocommit(test_conn, 1);
mysql_close(test_conn);
}
```
* **expect result:**
1. multiple statements execute success
* **real result:**
1. multiple statements hang when try to fetch the 3rd resultsets, it seems dble execute only the first two quries "
```mysql
use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
```
* **supplements:**
1.dble log
[dble.log](https://github.com/actiontech/dble/files/3130079/dble.log)
/label ~BUG
| 1.0 | only part of multiple statement executed,and hang when try to get resultsets - * **dble version:**
5.6.29-dble-9.9.9.9-2d18ed6c8dae540e4becc2ba03472b2284238f58-20190429070246
* **preconditions :**
complete dble configue refer to autotest sql_cover_mixed_bk
complete c++ code for test refer to autotest c_mysql_api/mysql_real_connect.cpp
* **configs:**
**schema.xml**
```
<schema name="schema1" sqlMaxLimit="100" dataNode="dn5">
<table name="sharding_4_t1" dataNode="dn1,dn2,dn3,dn4" rule="hash-four" />
</schema>
<dataNode name="dn1" dataHost="172.100.9.5" database="db1" />
<dataNode name="dn2" dataHost="172.100.9.6" database="db1" />
<dataNode name="dn3" dataHost="172.100.9.5" database="db2" />
<dataNode name="dn4" dataHost="172.100.9.6" database="db2" />
<dataNode name="dn5" dataHost="172.100.9.5" database="db3" />
```
**rule.xml**
```
```
**server.xml**
```
```
* **steps:**
step1. execute multiple statements with c++
```mysql
use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
CREATE TABLE sharding_4_t1(id INT);\
INSERT INTO sharding_4_t1 VALUES(10);\
INSERT INTO sharding_4_t1 VALUES(20);\
INSERT INTO sharding_4_t1 VALUES(30);\
SELECT * FROM sharding_4_t1;
```
reference c++ code:
```c++
void case_mysql_real_connect(MYSQL* conn){
// case1: default_db=null, port=0
MYSQL* test_conn = mysql_init(NULL);
mysql_options(test_conn,MYSQL_OPT_COMPRESS,"0");
mysql_options(test_conn,MYSQL_READ_DEFAULT_GROUP,"odbc");
mysql_options(test_conn,MYSQL_INIT_COMMAND,"SET autocommit=0");
mysql_real_connect(test_conn, HOST_DBLE, TEST_USER, TEST_USER_PASSWD, NULL, DBLE_PORT,NULL, CLIENT_DEPRECATE_EOF|CLIENT_MULTI_STATEMENTS);
if (test_conn == NULL) {
printf("Error connecting to database: %s\n", mysql_error(test_conn));
exit(1);
}else{
printf("connect success");
}
//case: multi query, and multi resultsets
/* execute multiple statements */
int status = mysql_query(test_conn,
"use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
CREATE TABLE sharding_4_t1(id INT);\
INSERT INTO sharding_4_t1 VALUES(10);\
INSERT INTO sharding_4_t1 VALUES(20);\
INSERT INTO sharding_4_t1 VALUES(30);\
SELECT * FROM sharding_4_t1;");
if (status)
{
printf("execute multi statement(s) Err, %s \n", mysql_error(test_conn));
mysql_close(test_conn);
exit(1);
}else{
printf(" pass! multi sqls success\n");
}
printf(" print multi resultsets:\n");
/* process each statement result */
do {
/* did current statement return data? */
MYSQL_RES *result = mysql_store_result(test_conn);
if (result)
{
/* yes; process rows and free the result set */
doPrintResult(test_conn, result);
mysql_free_result(result);
}
else /* no result set or error */
{
if (mysql_field_count(test_conn) == 0)
{
printf(" %lld rows affected\n",
mysql_affected_rows(test_conn));
}
else /* some error occurred */
{
printf("Could not retrieve result set\n");
exit(1);
}
}
if(mysql_more_results(test_conn)){
printf(" More results exist\n");
}
/* more results? -1 = no, >0 = error, 0 = yes (keep looping) */
if ((status = mysql_next_result(test_conn)) > 0){
printf("Could not get next result\n");
exit(1);
}
} while (status == 0);
mysql_autocommit(test_conn, 1);
mysql_close(test_conn);
}
```
* **expect result:**
1. multiple statements execute success
* **real result:**
1. multiple statements hang when try to fetch the 3rd resultsets, it seems dble execute only the first two quries "
```mysql
use schema1; \
DROP TABLE IF EXISTS sharding_4_t1;\
```
* **supplements:**
1.dble log
[dble.log](https://github.com/actiontech/dble/files/3130079/dble.log)
/label ~BUG
| priority | only part of multiple statement executed and hang when try to get resultsets dble version: dble preconditions : complete dble configue refer to autotest sql cover mixed bk complete c code for test refer to autotest c mysql api mysql real connect cpp configs: schema xml rule xml server xml steps: execute multiple statements with c mysql use drop table if exists sharding create table sharding id int insert into sharding values insert into sharding values insert into sharding values select from sharding reference c code c void case mysql real connect mysql conn default db null port mysql test conn mysql init null mysql options test conn mysql opt compress mysql options test conn mysql read default group odbc mysql options test conn mysql init command set autocommit mysql real connect test conn host dble test user test user passwd null dble port null client deprecate eof client multi statements if test conn null printf error connecting to database s n mysql error test conn exit else printf connect success case multi query and multi resultsets execute multiple statements int status mysql query test conn use drop table if exists sharding create table sharding id int insert into sharding values insert into sharding values insert into sharding values select from sharding if status printf execute multi statement s err s n mysql error test conn mysql close test conn exit else printf pass multi sqls success n printf print multi resultsets n process each statement result do did current statement return data mysql res result mysql store result test conn if result yes process rows and free the result set doprintresult test conn result mysql free result result else no result set or error if mysql field count test conn printf lld rows affected n mysql affected rows test conn else some error occurred printf could not retrieve result set n exit if mysql more results test conn printf more results exist n more results no error yes keep looping if status mysql next result test conn printf could not get next result n exit while status mysql autocommit test conn mysql close test conn expect result: multiple statements execute success real result: multiple statements hang when try to fetch the resultsets it seems dble execute only the first two quries mysql use drop table if exists sharding supplements: dble log label bug | 1 |
130,037 | 5,108,369,326 | IssuesEvent | 2017-01-05 17:28:55 | MAPC/trailmap2 | https://api.github.com/repos/MAPC/trailmap2 | closed | Add "proposed" layer to map; default to existing bike/ped features | enhancement high priority in progress | Default map to show EXISTING bike/ped features. Create filter to show PROPOSED, and style proposed layer as a dashed line. | 1.0 | Add "proposed" layer to map; default to existing bike/ped features - Default map to show EXISTING bike/ped features. Create filter to show PROPOSED, and style proposed layer as a dashed line. | priority | add proposed layer to map default to existing bike ped features default map to show existing bike ped features create filter to show proposed and style proposed layer as a dashed line | 1 |
679,768 | 23,244,519,076 | IssuesEvent | 2022-08-03 18:43:51 | RESOStandards/reso-certification-common | https://api.github.com/repos/RESOStandards/reso-certification-common | opened | Migrate JSON -> XML Config Utils | priority-high | The Commander currently uses legacy XML configuration files. In the long run we will migrate to JSON based configs natively, but we have JSON to XML config generators that we're using in the meantime. They're currently in the certification backend repo, but should be in Common instead since other services need them. | 1.0 | Migrate JSON -> XML Config Utils - The Commander currently uses legacy XML configuration files. In the long run we will migrate to JSON based configs natively, but we have JSON to XML config generators that we're using in the meantime. They're currently in the certification backend repo, but should be in Common instead since other services need them. | priority | migrate json xml config utils the commander currently uses legacy xml configuration files in the long run we will migrate to json based configs natively but we have json to xml config generators that we re using in the meantime they re currently in the certification backend repo but should be in common instead since other services need them | 1 |
63,897 | 3,202,018,563 | IssuesEvent | 2015-10-02 11:29:51 | washingtontrails/vms | https://api.github.com/repos/washingtontrails/vms | closed | Work party registration confrimation screen shows todays date, not date of WP | Bug High Priority Pyramid Reviewing VMS BUDGET | When you regester for a work party you directed to a Work Party registraion page (see attached). The date on this screen shows the current system date (time) and not the date/start time of the work party.
When I registered for this work party: http://staging.wta.org:6543/workparty/a2eK00000000iUN/
I got this confirmation. Note the difference in the dates.

| 1.0 | Work party registration confrimation screen shows todays date, not date of WP - When you regester for a work party you directed to a Work Party registraion page (see attached). The date on this screen shows the current system date (time) and not the date/start time of the work party.
When I registered for this work party: http://staging.wta.org:6543/workparty/a2eK00000000iUN/
I got this confirmation. Note the difference in the dates.

| priority | work party registration confrimation screen shows todays date not date of wp when you regester for a work party you directed to a work party registraion page see attached the date on this screen shows the current system date time and not the date start time of the work party when i registered for this work party i got this confirmation note the difference in the dates | 1 |
400,315 | 11,772,846,999 | IssuesEvent | 2020-03-16 05:27:10 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Control API visibility in the Developer Portal not working as expected | 3.1.0 Priority/Highest Severity/Critical Type/Bug | ### Description:
1. Create tenant wso2.com
2. login to publisher from tenant admin.
3. Create an API (TestAPI) and enable Developer Portal Visibility - Restricted by Roles.
4. Add internal/publisher role to the list of roles.
5. Login to carbon console as tenant admin.
6. Create a user ( chanaka ) and add all available roles except admin and internal/publisher.
7. Login to devportal as chanaka@wso2.com.
8. Go to wso2.com tenant store.
9. The TestAPI is visible to chanaka@wso2.com.
10. The TestAPI tags are visible to chanaka@wso2.com in the tag cloud.
Expected: TestAPI should be invisible to the user chanaka@wso2.com. And TestAPI tags should not be visible to chanaka@wso2.com | 1.0 | Control API visibility in the Developer Portal not working as expected - ### Description:
1. Create tenant wso2.com
2. login to publisher from tenant admin.
3. Create an API (TestAPI) and enable Developer Portal Visibility - Restricted by Roles.
4. Add internal/publisher role to the list of roles.
5. Login to carbon console as tenant admin.
6. Create a user ( chanaka ) and add all available roles except admin and internal/publisher.
7. Login to devportal as chanaka@wso2.com.
8. Go to wso2.com tenant store.
9. The TestAPI is visible to chanaka@wso2.com.
10. The TestAPI tags are visible to chanaka@wso2.com in the tag cloud.
Expected: TestAPI should be invisible to the user chanaka@wso2.com. And TestAPI tags should not be visible to chanaka@wso2.com | priority | control api visibility in the developer portal not working as expected description create tenant com login to publisher from tenant admin create an api testapi and enable developer portal visibility restricted by roles add internal publisher role to the list of roles login to carbon console as tenant admin create a user chanaka and add all available roles except admin and internal publisher login to devportal as chanaka com go to com tenant store the testapi is visible to chanaka com the testapi tags are visible to chanaka com in the tag cloud expected testapi should be invisible to the user chanaka com and testapi tags should not be visible to chanaka com | 1 |
756,375 | 26,468,709,856 | IssuesEvent | 2023-01-17 04:14:20 | Automattic/woocommerce-payments | https://api.github.com/repos/Automattic/woocommerce-payments | closed | Some Users Unable to Save Settings Due to Console Error Related to APMs | type: bug priority: high status: has pr component: alternative payment methods impact: high component: upe category: core needs prioritisation | ### Describe the bug
Some [WooCommerce Payments](https://woocommerce.com/products/woocommerce-payments/?aff=10486&cid=1131038) merchants are unable to save their settings at **Payments > Settings**. When they attempt to, they see an `Error Saving Settings` pop-up, and the following error appears in console:
```
{
"code": "rest_invalid_param",
"message": "Invalid parameter(s): enabled_payment_method_ids",
"data": {
"status": 400,
"params": {
"enabled_payment_method_ids": "enabled_payment_method_ids[1] is not one of card, bancontact, eps, giropay, ideal, sofort, sepa_debit, and p24."
},
"details": {
"enabled_payment_method_ids": {
"code": "rest_not_in_enum",
"message": "enabled_payment_method_ids[1] is not one of card, bancontact, eps, giropay, ideal, sofort, sepa_debit, and p24.",
"data": null
}
}
}
}
```
Only merchants utilizing UPE are affected.
### To Reproduce
We've been unable to reproduce this so far.
### Workaround
Disabling UPE and re-enabling again fixes the issue.
### Additional context
Affected merchants so far:
- 5607423-zen
- 5607722-zen
- 5608298-zen
- 5608374-zen | 1.0 | Some Users Unable to Save Settings Due to Console Error Related to APMs - ### Describe the bug
Some [WooCommerce Payments](https://woocommerce.com/products/woocommerce-payments/?aff=10486&cid=1131038) merchants are unable to save their settings at **Payments > Settings**. When they attempt to, they see an `Error Saving Settings` pop-up, and the following error appears in console:
```
{
"code": "rest_invalid_param",
"message": "Invalid parameter(s): enabled_payment_method_ids",
"data": {
"status": 400,
"params": {
"enabled_payment_method_ids": "enabled_payment_method_ids[1] is not one of card, bancontact, eps, giropay, ideal, sofort, sepa_debit, and p24."
},
"details": {
"enabled_payment_method_ids": {
"code": "rest_not_in_enum",
"message": "enabled_payment_method_ids[1] is not one of card, bancontact, eps, giropay, ideal, sofort, sepa_debit, and p24.",
"data": null
}
}
}
}
```
Only merchants utilizing UPE are affected.
### To Reproduce
We've been unable to reproduce this so far.
### Workaround
Disabling UPE and re-enabling again fixes the issue.
### Additional context
Affected merchants so far:
- 5607423-zen
- 5607722-zen
- 5608298-zen
- 5608374-zen | priority | some users unable to save settings due to console error related to apms describe the bug some merchants are unable to save their settings at payments settings when they attempt to they see an error saving settings pop up and the following error appears in console code rest invalid param message invalid parameter s enabled payment method ids data status params enabled payment method ids enabled payment method ids is not one of card bancontact eps giropay ideal sofort sepa debit and details enabled payment method ids code rest not in enum message enabled payment method ids is not one of card bancontact eps giropay ideal sofort sepa debit and data null only merchants utilizing upe are affected to reproduce we ve been unable to reproduce this so far workaround disabling upe and re enabling again fixes the issue additional context affected merchants so far zen zen zen zen | 1 |
346,815 | 10,420,270,815 | IssuesEvent | 2019-09-15 23:01:00 | mgallegos/decima-purchase | https://api.github.com/repos/mgallegos/decima-purchase | closed | Pantalla "Punto de compra" | enhancement high priority | Controlador
PurchasePointManager.php
Servicio:
PurchaseOrderManager.php (ya existe)
Vista:
purchase-point.blade.php
Prefijos:
purch-pp-, purchPp, purch_pp_
Menú:
Compras -> Transacciones -> Punto de compra
URL para routes.php:
/purchases/transactions/purchase-point
Diferencias con punto de venta:
- El botón del proveedor se coloca donde esta el nombre de la organización (deja siempre botón de agregar y editar)
- Giro, dirección, teléfono colocar en encabezado, email y web.
- La primera fila será: fecha, NRC, NIT, y DUI
- Cond. y forma de pago se mantienen igual.
- Detalle se mantiene funcionalidad (revisar nombre de compras)
- En impuesto cambia la forma de calcular el total, así: $Order->purchase_total - $Order->collected_tax_amount_sum + $Order->withheld_tax_amount_sum
| 1.0 | Pantalla "Punto de compra" - Controlador
PurchasePointManager.php
Servicio:
PurchaseOrderManager.php (ya existe)
Vista:
purchase-point.blade.php
Prefijos:
purch-pp-, purchPp, purch_pp_
Menú:
Compras -> Transacciones -> Punto de compra
URL para routes.php:
/purchases/transactions/purchase-point
Diferencias con punto de venta:
- El botón del proveedor se coloca donde esta el nombre de la organización (deja siempre botón de agregar y editar)
- Giro, dirección, teléfono colocar en encabezado, email y web.
- La primera fila será: fecha, NRC, NIT, y DUI
- Cond. y forma de pago se mantienen igual.
- Detalle se mantiene funcionalidad (revisar nombre de compras)
- En impuesto cambia la forma de calcular el total, así: $Order->purchase_total - $Order->collected_tax_amount_sum + $Order->withheld_tax_amount_sum
| priority | pantalla punto de compra controlador purchasepointmanager php servicio purchaseordermanager php ya existe vista purchase point blade php prefijos purch pp purchpp purch pp menú compras transacciones punto de compra url para routes php purchases transactions purchase point diferencias con punto de venta el botón del proveedor se coloca donde esta el nombre de la organización deja siempre botón de agregar y editar giro dirección teléfono colocar en encabezado email y web la primera fila será fecha nrc nit y dui cond y forma de pago se mantienen igual detalle se mantiene funcionalidad revisar nombre de compras en impuesto cambia la forma de calcular el total así order purchase total order collected tax amount sum order withheld tax amount sum | 1 |
563,393 | 16,682,496,834 | IssuesEvent | 2021-06-08 02:47:46 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Full definition of a union type given in error messages related to union typed variables | Area/Compiler Area/DiagnosticAPI Area/Diagnostics Area/SemanticAPI Priority/High SwanLakeDump Team/CompilerFE Type/Bug | Consider the following program:
```ballerina
type ValueType int|float|string|boolean|byte;
type DataType ValueType|table|json|xml|map|();
public function main() {
map<DataType> m;
map<int> m2 = m;
}
```
Running the above program results in the following compilation error:
```
error: .::test2.bal:6:19: incompatible types: expected 'map<int>', found 'map<int|float|string|boolean|byte|table|json|xml|map?>'
```
Instead of saying `map<int|float|string|boolean|byte|table|json|xml|map?>`, it should be saying `map<DataType>`. | 1.0 | Full definition of a union type given in error messages related to union typed variables - Consider the following program:
```ballerina
type ValueType int|float|string|boolean|byte;
type DataType ValueType|table|json|xml|map|();
public function main() {
map<DataType> m;
map<int> m2 = m;
}
```
Running the above program results in the following compilation error:
```
error: .::test2.bal:6:19: incompatible types: expected 'map<int>', found 'map<int|float|string|boolean|byte|table|json|xml|map?>'
```
Instead of saying `map<int|float|string|boolean|byte|table|json|xml|map?>`, it should be saying `map<DataType>`. | priority | full definition of a union type given in error messages related to union typed variables consider the following program ballerina type valuetype int float string boolean byte type datatype valuetype table json xml map public function main map m map m running the above program results in the following compilation error error bal incompatible types expected map found map instead of saying map it should be saying map | 1 |
532,886 | 15,572,692,707 | IssuesEvent | 2021-03-17 07:29:17 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.youtube.com - video or audio doesn't play | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Firefox 87.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:87.0) Gecko/20100101 Firefox/87.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68367 -->
**URL**: https://www.youtube.com/
**Browser / Version**: Firefox 87.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
youtube videos does not playing
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/b9e535fb-4f3f-4aec-af7a-18882142180f.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210315170302</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/888aa10c-2af2-4e10-8d67-f8bf4ef7a571)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.youtube.com - video or audio doesn't play - <!-- @browser: Firefox 87.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:87.0) Gecko/20100101 Firefox/87.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68367 -->
**URL**: https://www.youtube.com/
**Browser / Version**: Firefox 87.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
youtube videos does not playing
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/b9e535fb-4f3f-4aec-af7a-18882142180f.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210315170302</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/888aa10c-2af2-4e10-8d67-f8bf4ef7a571)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce youtube videos does not playing view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
194,736 | 6,898,251,212 | IssuesEvent | 2017-11-24 08:43:00 | Mandiklopper/UBA-HR-Queries-Issues | https://api.github.com/repos/Mandiklopper/UBA-HR-Queries-Issues | closed | State Lookup Button Not Working for Address Tab of Employee Link | High Priority | The look up list for 'State' does not show values. This means the user needs to know the state code he wants to select and then enter it in manually. | 1.0 | State Lookup Button Not Working for Address Tab of Employee Link - The look up list for 'State' does not show values. This means the user needs to know the state code he wants to select and then enter it in manually. | priority | state lookup button not working for address tab of employee link the look up list for state does not show values this means the user needs to know the state code he wants to select and then enter it in manually | 1 |
144,633 | 5,543,508,728 | IssuesEvent | 2017-03-22 17:05:02 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | AMP forces sticky element to use top CSS property | in progress P1: High Priority | AMP forces my sticky element to be 'top: calc(0px)'.
To reproduce:
1. create an element with the next css:
{
position:sticky;
bottom:0;
}
2. When I render in the browser AMP has added a top property to my element:

I understant that this behaviour is related to this fix https://github.com/ampproject/amphtml/pull/7767 , are you planing on implementing the bottom property as well when it's sticky?
Thanks,
| 1.0 | AMP forces sticky element to use top CSS property - AMP forces my sticky element to be 'top: calc(0px)'.
To reproduce:
1. create an element with the next css:
{
position:sticky;
bottom:0;
}
2. When I render in the browser AMP has added a top property to my element:

I understant that this behaviour is related to this fix https://github.com/ampproject/amphtml/pull/7767 , are you planing on implementing the bottom property as well when it's sticky?
Thanks,
| priority | amp forces sticky element to use top css property amp forces my sticky element to be top calc to reproduce create an element with the next css position sticky bottom when i render in the browser amp has added a top property to my element i understant that this behaviour is related to this fix are you planing on implementing the bottom property as well when it s sticky thanks | 1 |
627,013 | 19,849,876,697 | IssuesEvent | 2022-01-21 11:03:32 | Galaxia5987/robot-2022 | https://api.github.com/repos/Galaxia5987/robot-2022 | opened | Centripetal Swerve Pose Estimator | help wanted subsystem Priority: High | We can't really know where the robot is using vision, because we can see the target from every position of the field, and we would classify it the same, so we drop the position, we just know how far are we, and the angle from the target | 1.0 | Centripetal Swerve Pose Estimator - We can't really know where the robot is using vision, because we can see the target from every position of the field, and we would classify it the same, so we drop the position, we just know how far are we, and the angle from the target | priority | centripetal swerve pose estimator we can t really know where the robot is using vision because we can see the target from every position of the field and we would classify it the same so we drop the position we just know how far are we and the angle from the target | 1 |
370,911 | 10,958,385,811 | IssuesEvent | 2019-11-27 09:18:19 | wso2/ballerina-integrator | https://api.github.com/repos/wso2/ballerina-integrator | closed | [Querying a MySQL Database]Improvement points | Priority/High Type/Docs | **Description:**
Once I was going through [1] tutorial, I found that some of the content in the tutorial needs to be improved as below.
1. The employees.sql can be found in the zip folder which is asked to download at the beginning of the tutorial. But it has been asked to run the empoyees.sql file after creating the project by executing commands. It is better if we can ask the user to download the project zip file and employees.sql file can be found in the querying-mysql-database/src/querying_mysql_database/resources folder.
2. In the downloaded zip file, inside querying-mysql-database/src/querying_mysql_database/resources folder, it does not contain the employees.sql file. Instead, it contains sampleScript.sql file. Therefore rename the script name.
3. The module name which is added to ballerina.toml file needs to be renamed as 'querying_mysql_database'.
4. In the content which needs to be coped to ballerina.conf file, the html <br/> tags need to be removed.
5. In ballerina.conf file, it needs to add the parameters as below. It is necessary to add double-quotes.
```
MYSQL_URL = "jdbc:mysql://localhost:3306/Employee"
MYSQL_USERNAME = "root"
MYSQL_PASSWORD = "root"
```
6. Specify the format of the JDBC URL which needs to be added to the ballerina.conf file.
**Suggested Labels:**
type/docs
priority/ high
**Suggested Assignees:**
@samgnaniah
[1] - https://ei.docs.wso2.com/en/latest/ballerina-integrator/learn/tutorials/database-integrations/querying-mysql-database/1/ | 1.0 | [Querying a MySQL Database]Improvement points - **Description:**
Once I was going through [1] tutorial, I found that some of the content in the tutorial needs to be improved as below.
1. The employees.sql can be found in the zip folder which is asked to download at the beginning of the tutorial. But it has been asked to run the empoyees.sql file after creating the project by executing commands. It is better if we can ask the user to download the project zip file and employees.sql file can be found in the querying-mysql-database/src/querying_mysql_database/resources folder.
2. In the downloaded zip file, inside querying-mysql-database/src/querying_mysql_database/resources folder, it does not contain the employees.sql file. Instead, it contains sampleScript.sql file. Therefore rename the script name.
3. The module name which is added to ballerina.toml file needs to be renamed as 'querying_mysql_database'.
4. In the content which needs to be coped to ballerina.conf file, the html <br/> tags need to be removed.
5. In ballerina.conf file, it needs to add the parameters as below. It is necessary to add double-quotes.
```
MYSQL_URL = "jdbc:mysql://localhost:3306/Employee"
MYSQL_USERNAME = "root"
MYSQL_PASSWORD = "root"
```
6. Specify the format of the JDBC URL which needs to be added to the ballerina.conf file.
**Suggested Labels:**
type/docs
priority/ high
**Suggested Assignees:**
@samgnaniah
[1] - https://ei.docs.wso2.com/en/latest/ballerina-integrator/learn/tutorials/database-integrations/querying-mysql-database/1/ | priority | improvement points description once i was going through tutorial i found that some of the content in the tutorial needs to be improved as below the employees sql can be found in the zip folder which is asked to download at the beginning of the tutorial but it has been asked to run the empoyees sql file after creating the project by executing commands it is better if we can ask the user to download the project zip file and employees sql file can be found in the querying mysql database src querying mysql database resources folder in the downloaded zip file inside querying mysql database src querying mysql database resources folder it does not contain the employees sql file instead it contains samplescript sql file therefore rename the script name the module name which is added to ballerina toml file needs to be renamed as querying mysql database in the content which needs to be coped to ballerina conf file the html tags need to be removed in ballerina conf file it needs to add the parameters as below it is necessary to add double quotes mysql url jdbc mysql localhost employee mysql username root mysql password root specify the format of the jdbc url which needs to be added to the ballerina conf file suggested labels type docs priority high suggested assignees samgnaniah | 1 |
808,038 | 30,031,285,272 | IssuesEvent | 2023-06-27 09:45:22 | PluginBugs/Issues-ItemsAdder | https://api.github.com/repos/PluginBugs/Issues-ItemsAdder | closed | Issue with some recipes | type: bug priority: HIGH result: breaks gameplay result: breaks configurations | Some particular recipes have issues, caused by recipes hashing collisions.
Source: https://discord.com/channels/533407895010803753/533413171403292712/1121162472594870425


Full config:
```yml
info:
namespace: recipe_bug
recipes:
crafting_table:
bamboo_wooden_hoe2:
enabled: true
pattern:
- Xww
- XsX
- XsX
ingredients:
w: BAMBOO_PLANKS
s: STICK
result:
item: item1
amount: 1
crimson_wooden_shovel:
enabled: true
pattern:
- XwX
- XsX
- XsX
ingredients:
w: CRIMSON_PLANKS
s: STICK
result:
item: item2
amount: 1
items:
item1:
enabled: true
display_name: item1
resource:
generate: false
model_path: minecraft:item/emerald
material: EMERALD
item2:
enabled: true
display_name: item2
resource:
generate: false
model_path: minecraft:item/diamond
material: DIAMOND
```
| 1.0 | Issue with some recipes - Some particular recipes have issues, caused by recipes hashing collisions.
Source: https://discord.com/channels/533407895010803753/533413171403292712/1121162472594870425


Full config:
```yml
info:
namespace: recipe_bug
recipes:
crafting_table:
bamboo_wooden_hoe2:
enabled: true
pattern:
- Xww
- XsX
- XsX
ingredients:
w: BAMBOO_PLANKS
s: STICK
result:
item: item1
amount: 1
crimson_wooden_shovel:
enabled: true
pattern:
- XwX
- XsX
- XsX
ingredients:
w: CRIMSON_PLANKS
s: STICK
result:
item: item2
amount: 1
items:
item1:
enabled: true
display_name: item1
resource:
generate: false
model_path: minecraft:item/emerald
material: EMERALD
item2:
enabled: true
display_name: item2
resource:
generate: false
model_path: minecraft:item/diamond
material: DIAMOND
```
| priority | issue with some recipes some particular recipes have issues caused by recipes hashing collisions source full config yml info namespace recipe bug recipes crafting table bamboo wooden enabled true pattern xww xsx xsx ingredients w bamboo planks s stick result item amount crimson wooden shovel enabled true pattern xwx xsx xsx ingredients w crimson planks s stick result item amount items enabled true display name resource generate false model path minecraft item emerald material emerald enabled true display name resource generate false model path minecraft item diamond material diamond | 1 |
145,249 | 5,561,286,170 | IssuesEvent | 2017-03-24 22:00:18 | sul-dlss/ld4p | https://api.github.com/repos/sul-dlss/ld4p | closed | pick ~20 BF2 instances to seed our indexing code | high priority in progress | - want a sorting title diff from full title
- want records with 700at if there are any (for addl authors (?))
- want records with diff 008 byte 6 and 7 values (for diff formats)
send some sample RDFXML to Shelley, along with corresponding marcxml | 1.0 | pick ~20 BF2 instances to seed our indexing code - - want a sorting title diff from full title
- want records with 700at if there are any (for addl authors (?))
- want records with diff 008 byte 6 and 7 values (for diff formats)
send some sample RDFXML to Shelley, along with corresponding marcxml | priority | pick instances to seed our indexing code want a sorting title diff from full title want records with if there are any for addl authors want records with diff byte and values for diff formats send some sample rdfxml to shelley along with corresponding marcxml | 1 |
458,732 | 13,180,697,647 | IssuesEvent | 2020-08-12 13:13:55 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | Remove the need for manual moderation | :rotating_light: High Priority | CDNJS requires a certain amount of manual review and moderation. This review is required both to allow new projects to be added to the project, and to handle certain types of project changes which can't be automatically imported.
This issue is a place to discuss how we might eliminate this requirement in the future.
I believe the goal of this project should be to take everything which requires a manually-reviewed PR now, and make it happen automatically. I have categorized a snapshot of the currently open Pull Requests here: https://docs.google.com/spreadsheets/d/18-HyNKxfXvzCLr6v57UrGHtchvDhNcrI8JvGn4gUwck/edit#gid=0
As you can see the majority of these issues can be divided into:
- The CI errored, the user then fixed the error, but it now requires a human to review
- A human has to look at the code to confirm the project is:
- Popular enough
- Includes the right code (can't really be verified)
- Has the right glob pattern of files to allow for auto-updating in the future
- A handful of anomalous conditions which are likely less important to automate
With that we would love to hear the communities ideas for how to eliminate the labor and danger of this manual review!
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/83667391-remove-the-need-for-manual-moderation?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Remove the need for manual moderation - CDNJS requires a certain amount of manual review and moderation. This review is required both to allow new projects to be added to the project, and to handle certain types of project changes which can't be automatically imported.
This issue is a place to discuss how we might eliminate this requirement in the future.
I believe the goal of this project should be to take everything which requires a manually-reviewed PR now, and make it happen automatically. I have categorized a snapshot of the currently open Pull Requests here: https://docs.google.com/spreadsheets/d/18-HyNKxfXvzCLr6v57UrGHtchvDhNcrI8JvGn4gUwck/edit#gid=0
As you can see the majority of these issues can be divided into:
- The CI errored, the user then fixed the error, but it now requires a human to review
- A human has to look at the code to confirm the project is:
- Popular enough
- Includes the right code (can't really be verified)
- Has the right glob pattern of files to allow for auto-updating in the future
- A handful of anomalous conditions which are likely less important to automate
With that we would love to hear the communities ideas for how to eliminate the labor and danger of this manual review!
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/83667391-remove-the-need-for-manual-moderation?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F32893&utm_medium=issues&utm_source=github).
</bountysource-plugin> | priority | remove the need for manual moderation cdnjs requires a certain amount of manual review and moderation this review is required both to allow new projects to be added to the project and to handle certain types of project changes which can t be automatically imported this issue is a place to discuss how we might eliminate this requirement in the future i believe the goal of this project should be to take everything which requires a manually reviewed pr now and make it happen automatically i have categorized a snapshot of the currently open pull requests here as you can see the majority of these issues can be divided into the ci errored the user then fixed the error but it now requires a human to review a human has to look at the code to confirm the project is popular enough includes the right code can t really be verified has the right glob pattern of files to allow for auto updating in the future a handful of anomalous conditions which are likely less important to automate with that we would love to hear the communities ideas for how to eliminate the labor and danger of this manual review want to back this issue we accept bounties via | 1 |
735,462 | 25,399,423,686 | IssuesEvent | 2022-11-22 10:57:01 | wsEM2022-24/Seminar-Projekt01 | https://api.github.com/repos/wsEM2022-24/Seminar-Projekt01 | closed | Render Priority for helpers | priority::high current-sprint | The Axishelper and Gridhelper would overlap, so theyre a little bit above each other.
to prevent the problems with inaccuracy, the render priority can be modified, so that each helper gets rendered ontop of each other. | 1.0 | Render Priority for helpers - The Axishelper and Gridhelper would overlap, so theyre a little bit above each other.
to prevent the problems with inaccuracy, the render priority can be modified, so that each helper gets rendered ontop of each other. | priority | render priority for helpers the axishelper and gridhelper would overlap so theyre a little bit above each other to prevent the problems with inaccuracy the render priority can be modified so that each helper gets rendered ontop of each other | 1 |
609,021 | 18,851,921,648 | IssuesEvent | 2021-11-11 22:10:32 | UnknownShadow200/MCGalaxy | https://api.github.com/repos/UnknownShadow200/MCGalaxy | opened | Updates dialog breaks server in some cases on Linux | bug high priority | Mostly observed with recent Gnome DE + Ubuntu

A short time after, the server will die with a message like this

Issue seems to be due to the way Updates MessageBox is spawned from a background thread (So as not to block the main UI), but this triggers a multithreading issue
https://bugzilla.xamarin.com/28/28047/bug.html
github.com/ ArduPilot/ MissionPlanner/issues/208 | 1.0 | Updates dialog breaks server in some cases on Linux - Mostly observed with recent Gnome DE + Ubuntu

A short time after, the server will die with a message like this

Issue seems to be due to the way Updates MessageBox is spawned from a background thread (So as not to block the main UI), but this triggers a multithreading issue
https://bugzilla.xamarin.com/28/28047/bug.html
github.com/ ArduPilot/ MissionPlanner/issues/208 | priority | updates dialog breaks server in some cases on linux mostly observed with recent gnome de ubuntu a short time after the server will die with a message like this issue seems to be due to the way updates messagebox is spawned from a background thread so as not to block the main ui but this triggers a multithreading issue github com ardupilot missionplanner issues | 1 |
24,102 | 2,666,081,634 | IssuesEvent | 2015-03-21 05:20:09 | cs2103jan2015-f13-2j/main | https://api.github.com/repos/cs2103jan2015-f13-2j/main | closed | bug in hotkey | priority.high type.bug | there exists a bug in hotkey.
user only able to press ctrl + space once, and escape to close it.
user then unable to press ctrl + space to bring up the main textbox again.
i believe it may be due to javafx application cannot be called more than once. it has something to do with starting & terminating threads as well. | 1.0 | bug in hotkey - there exists a bug in hotkey.
user only able to press ctrl + space once, and escape to close it.
user then unable to press ctrl + space to bring up the main textbox again.
i believe it may be due to javafx application cannot be called more than once. it has something to do with starting & terminating threads as well. | priority | bug in hotkey there exists a bug in hotkey user only able to press ctrl space once and escape to close it user then unable to press ctrl space to bring up the main textbox again i believe it may be due to javafx application cannot be called more than once it has something to do with starting terminating threads as well | 1 |
412,849 | 12,056,992,526 | IssuesEvent | 2020-04-15 15:12:42 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | SDK Server Runs Out of Memory | Needed for V2 launch Priority: Very High type/bug | Command:
```yarn flash sdk-server -c '{"ethereum": {"http":"http://127.0.0.1:8545"}, "gnosis": {"http": "http://localhost:8888/api/"}}'```
Error:
```
Running SDK server. Press ENTER to quit:
<--- Last few GCs --->
[76936:0x102642000] 4952354 ms: Mark-sweep 1310.2 (1456.0) -> 1310.0 (1456.0) MB, 1267.8 / 0.0 ms (average mu = 0.108, current mu = 0.015) allocation failure GC in old space requested
[76936:0x102642000] 4953639 ms: Mark-sweep 1310.0 (1456.0) -> 1310.0 (1456.0) MB, 1266.1 / 0.0 ms (average mu = 0.064, current mu = 0.015) allocation failure GC in old space requested
<--- JS stacktrace --->
==== JS stack trace =========================================
0: ExitFrame [pc: 0x28100745be3d]
Security context: 0x3e40bb39e6e1 <JSObject>
1: defineReadonlyProperties(aka defineReadonlyProperties) [0x3e40dcda9201] [/Users/robertdavidson/projects/augur/augur/node_modules/indexeddbshim/dist/indexeddbshim-node.js:~1116] [pc=0x281007eeb8d0](this=0x3e40f16026f1 <undefined>,/* anonymous */=0x3e4058a05129 <IDBRequest map = 0x3e40776b07d1>,/* anonymous */=0x3e40dcda9899 <JSArray[3]>,/* anonymous */...
FATAL ERROR: Ineffective mark-compacts near heap limit Allocation failed - JavaScript heap out of memory
1: 0x10003b125 node::Abort() [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
2: 0x10003b32f node::OnFatalError(char const*, char const*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
3: 0x1001a89a5 v8::internal::V8::FatalProcessOutOfMemory(v8::internal::Isolate*, char const*, bool) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
4: 0x100573dc2 v8::internal::Heap::FatalProcessOutOfMemory(char const*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
5: 0x100576895 v8::internal::Heap::CheckIneffectiveMarkCompact(unsigned long, double) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
6: 0x10057273f v8::internal::Heap::PerformGarbageCollection(v8::internal::GarbageCollector, v8::GCCallbackFlags) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
7: 0x100570914 v8::internal::Heap::CollectGarbage(v8::internal::AllocationSpace, v8::internal::GarbageCollectionReason, v8::GCCallbackFlags) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
8: 0x100570e15 v8::internal::Heap::CollectAllAvailableGarbage(v8::internal::GarbageCollectionReason) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
9: 0x10057d281 v8::internal::Heap::AllocateRawWithRetryOrFail(int, v8::internal::AllocationSpace, v8::internal::AllocationAlignment) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
10: 0x1005626a2 v8::internal::Handle<v8::internal::TransitionArray> v8::internal::Factory::NewWeakFixedArrayWithMap<v8::internal::TransitionArray>(v8::internal::Heap::RootListIndex, int, v8::internal::PretenureFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
11: 0x1005549aa v8::internal::Factory::NewTransitionArray(int, int) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
12: 0x10086b06d v8::internal::TransitionsAccessor::Insert(v8::internal::Handle<v8::internal::Name>, v8::internal::Handle<v8::internal::Map>, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
13: 0x1006991f8 v8::internal::Map::ConnectTransition(v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Name>, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
14: 0x100688151 v8::internal::Map::CopyReplaceDescriptors(v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::DescriptorArray>, v8::internal::Handle<v8::internal::LayoutDescriptor>, v8::internal::TransitionFlag, v8::internal::MaybeHandle<v8::internal::Name>, char const*, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
15: 0x10068484f v8::internal::Map::CopyAddDescriptor(v8::internal::Handle<v8::internal::Map>, v8::internal::Descriptor*, v8::internal::TransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
16: 0x10069b88c v8::internal::Map::TransitionToAccessorProperty(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Name>, int, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
17: 0x100666145 v8::internal::LookupIterator::TransitionToAccessorProperty(v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
18: 0x100694c3a v8::internal::JSObject::DefineAccessor(v8::internal::LookupIterator*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
19: 0x100698706 v8::internal::JSObject::DefineAccessor(v8::internal::Handle<v8::internal::JSObject>, v8::internal::Handle<v8::internal::Name>, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
20: 0x1007fa82c v8::internal::Runtime_DefineGetterPropertyUnchecked(int, v8::internal::Object**, v8::internal::Isolate*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
21: 0x28100745be3d
error Command failed with signal "SIGABRT".
error Command failed.
Exit code: 1
Command: /Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node
Arguments: /usr/local/Cellar/yarn/1.13.0/libexec/lib/cli.js --silent flash sdk-server -c {"ethereum": {"http":"http://127.0.0.1:8545"}, "gnosis": {"http": "http://localhost:8888/api/"}}
Directory: /Users/robertdavidson/projects/augur/augur/packages/augur-tools
Output:
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.``` | 1.0 | SDK Server Runs Out of Memory - Command:
```yarn flash sdk-server -c '{"ethereum": {"http":"http://127.0.0.1:8545"}, "gnosis": {"http": "http://localhost:8888/api/"}}'```
Error:
```
Running SDK server. Press ENTER to quit:
<--- Last few GCs --->
[76936:0x102642000] 4952354 ms: Mark-sweep 1310.2 (1456.0) -> 1310.0 (1456.0) MB, 1267.8 / 0.0 ms (average mu = 0.108, current mu = 0.015) allocation failure GC in old space requested
[76936:0x102642000] 4953639 ms: Mark-sweep 1310.0 (1456.0) -> 1310.0 (1456.0) MB, 1266.1 / 0.0 ms (average mu = 0.064, current mu = 0.015) allocation failure GC in old space requested
<--- JS stacktrace --->
==== JS stack trace =========================================
0: ExitFrame [pc: 0x28100745be3d]
Security context: 0x3e40bb39e6e1 <JSObject>
1: defineReadonlyProperties(aka defineReadonlyProperties) [0x3e40dcda9201] [/Users/robertdavidson/projects/augur/augur/node_modules/indexeddbshim/dist/indexeddbshim-node.js:~1116] [pc=0x281007eeb8d0](this=0x3e40f16026f1 <undefined>,/* anonymous */=0x3e4058a05129 <IDBRequest map = 0x3e40776b07d1>,/* anonymous */=0x3e40dcda9899 <JSArray[3]>,/* anonymous */...
FATAL ERROR: Ineffective mark-compacts near heap limit Allocation failed - JavaScript heap out of memory
1: 0x10003b125 node::Abort() [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
2: 0x10003b32f node::OnFatalError(char const*, char const*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
3: 0x1001a89a5 v8::internal::V8::FatalProcessOutOfMemory(v8::internal::Isolate*, char const*, bool) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
4: 0x100573dc2 v8::internal::Heap::FatalProcessOutOfMemory(char const*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
5: 0x100576895 v8::internal::Heap::CheckIneffectiveMarkCompact(unsigned long, double) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
6: 0x10057273f v8::internal::Heap::PerformGarbageCollection(v8::internal::GarbageCollector, v8::GCCallbackFlags) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
7: 0x100570914 v8::internal::Heap::CollectGarbage(v8::internal::AllocationSpace, v8::internal::GarbageCollectionReason, v8::GCCallbackFlags) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
8: 0x100570e15 v8::internal::Heap::CollectAllAvailableGarbage(v8::internal::GarbageCollectionReason) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
9: 0x10057d281 v8::internal::Heap::AllocateRawWithRetryOrFail(int, v8::internal::AllocationSpace, v8::internal::AllocationAlignment) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
10: 0x1005626a2 v8::internal::Handle<v8::internal::TransitionArray> v8::internal::Factory::NewWeakFixedArrayWithMap<v8::internal::TransitionArray>(v8::internal::Heap::RootListIndex, int, v8::internal::PretenureFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
11: 0x1005549aa v8::internal::Factory::NewTransitionArray(int, int) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
12: 0x10086b06d v8::internal::TransitionsAccessor::Insert(v8::internal::Handle<v8::internal::Name>, v8::internal::Handle<v8::internal::Map>, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
13: 0x1006991f8 v8::internal::Map::ConnectTransition(v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Name>, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
14: 0x100688151 v8::internal::Map::CopyReplaceDescriptors(v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::DescriptorArray>, v8::internal::Handle<v8::internal::LayoutDescriptor>, v8::internal::TransitionFlag, v8::internal::MaybeHandle<v8::internal::Name>, char const*, v8::internal::SimpleTransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
15: 0x10068484f v8::internal::Map::CopyAddDescriptor(v8::internal::Handle<v8::internal::Map>, v8::internal::Descriptor*, v8::internal::TransitionFlag) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
16: 0x10069b88c v8::internal::Map::TransitionToAccessorProperty(v8::internal::Isolate*, v8::internal::Handle<v8::internal::Map>, v8::internal::Handle<v8::internal::Name>, int, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
17: 0x100666145 v8::internal::LookupIterator::TransitionToAccessorProperty(v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
18: 0x100694c3a v8::internal::JSObject::DefineAccessor(v8::internal::LookupIterator*, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
19: 0x100698706 v8::internal::JSObject::DefineAccessor(v8::internal::Handle<v8::internal::JSObject>, v8::internal::Handle<v8::internal::Name>, v8::internal::Handle<v8::internal::Object>, v8::internal::Handle<v8::internal::Object>, v8::internal::PropertyAttributes) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
20: 0x1007fa82c v8::internal::Runtime_DefineGetterPropertyUnchecked(int, v8::internal::Object**, v8::internal::Isolate*) [/Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node]
21: 0x28100745be3d
error Command failed with signal "SIGABRT".
error Command failed.
Exit code: 1
Command: /Users/robertdavidson/.nvm/versions/node/v10.14.2/bin/node
Arguments: /usr/local/Cellar/yarn/1.13.0/libexec/lib/cli.js --silent flash sdk-server -c {"ethereum": {"http":"http://127.0.0.1:8545"}, "gnosis": {"http": "http://localhost:8888/api/"}}
Directory: /Users/robertdavidson/projects/augur/augur/packages/augur-tools
Output:
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.``` | priority | sdk server runs out of memory command yarn flash sdk server c ethereum http gnosis http error running sdk server press enter to quit ms mark sweep mb ms average mu current mu allocation failure gc in old space requested ms mark sweep mb ms average mu current mu allocation failure gc in old space requested js stack trace exitframe security context definereadonlyproperties aka definereadonlyproperties this anonymous anonymous anonymous fatal error ineffective mark compacts near heap limit allocation failed javascript heap out of memory node abort node onfatalerror char const char const internal fatalprocessoutofmemory internal isolate char const bool internal heap fatalprocessoutofmemory char const internal heap checkineffectivemarkcompact unsigned long double internal heap performgarbagecollection internal garbagecollector gccallbackflags internal heap collectgarbage internal allocationspace internal garbagecollectionreason gccallbackflags internal heap collectallavailablegarbage internal garbagecollectionreason internal heap allocaterawwithretryorfail int internal allocationspace internal allocationalignment internal handle internal factory newweakfixedarraywithmap internal heap rootlistindex int internal pretenureflag internal factory newtransitionarray int int internal transitionsaccessor insert internal handle internal handle internal simpletransitionflag internal map connecttransition internal handle internal handle internal handle internal simpletransitionflag internal map copyreplacedescriptors internal handle internal handle internal handle internal transitionflag internal maybehandle char const internal simpletransitionflag internal map copyadddescriptor internal handle internal descriptor internal transitionflag internal map transitiontoaccessorproperty internal isolate internal handle internal handle int internal handle internal handle internal propertyattributes internal lookupiterator transitiontoaccessorproperty internal handle internal handle internal propertyattributes internal jsobject defineaccessor internal lookupiterator internal handle internal handle internal propertyattributes internal jsobject defineaccessor internal handle internal handle internal handle internal handle internal propertyattributes internal runtime definegetterpropertyunchecked int internal object internal isolate error command failed with signal sigabrt error command failed exit code command users robertdavidson nvm versions node bin node arguments usr local cellar yarn libexec lib cli js silent flash sdk server c ethereum http gnosis http directory users robertdavidson projects augur augur packages augur tools output error command failed with exit code info visit for documentation about this command | 1 |
549,270 | 16,089,191,352 | IssuesEvent | 2021-04-26 14:47:01 | sopra-fs21-group-05/group-05-server | https://api.github.com/repos/sopra-fs21-group-05/group-05-server | closed | getPictureGrid endpoint and method | high priority task | - [x] create endpoint in respective controller
- [x] create method getPictureGrid()
Time estimate: 2h
| 1.0 | getPictureGrid endpoint and method - - [x] create endpoint in respective controller
- [x] create method getPictureGrid()
Time estimate: 2h
| priority | getpicturegrid endpoint and method create endpoint in respective controller create method getpicturegrid time estimate | 1 |
744,079 | 25,928,064,234 | IssuesEvent | 2022-12-16 07:16:51 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | POI to Polygon conflation of cemeteries doesn't work | Type: Bug Type: Task Category: Core Priority: High | @curranMapper found a dataset with POIs with `amenity=grave_yard` and polygons with `landuse=cemetery` that weren't conflating, fix that issue. | 1.0 | POI to Polygon conflation of cemeteries doesn't work - @curranMapper found a dataset with POIs with `amenity=grave_yard` and polygons with `landuse=cemetery` that weren't conflating, fix that issue. | priority | poi to polygon conflation of cemeteries doesn t work curranmapper found a dataset with pois with amenity grave yard and polygons with landuse cemetery that weren t conflating fix that issue | 1 |
228,386 | 7,550,433,498 | IssuesEvent | 2018-04-18 16:53:49 | mono/monodevelop | https://api.github.com/repos/mono/monodevelop | closed | New F# Blank Forms App has incorrect page name | Area: F# high-priority vs-sync | Using VS for Mac 7.5
File->New->Blank Xamarin Forms app (F#).
Give the project a name.
In 7.4 the page created would be <ProjectName>Page, inside <ProjectName>Page.xaml and <ProjectName>Page.xaml.fs.
In 7.5 it is created with the correct class name, but incorrect file names - it always uses AppPage.xaml and AppPage.xaml.fs. See the attached screenshot.

> VS bug [#601029](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/601029) | 1.0 | New F# Blank Forms App has incorrect page name - Using VS for Mac 7.5
File->New->Blank Xamarin Forms app (F#).
Give the project a name.
In 7.4 the page created would be <ProjectName>Page, inside <ProjectName>Page.xaml and <ProjectName>Page.xaml.fs.
In 7.5 it is created with the correct class name, but incorrect file names - it always uses AppPage.xaml and AppPage.xaml.fs. See the attached screenshot.

> VS bug [#601029](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/601029) | priority | new f blank forms app has incorrect page name using vs for mac file new blank xamarin forms app f give the project a name in the page created would be page inside page xaml and page xaml fs in it is created with the correct class name but incorrect file names it always uses apppage xaml and apppage xaml fs see the attached screenshot vs bug | 1 |
503,581 | 14,594,831,362 | IssuesEvent | 2020-12-20 08:16:13 | OpenMined/PySyft | https://api.github.com/repos/OpenMined/PySyft | closed | Cuda support for Encrypted Computation | 0.2.x Priority: 2 - High :cold_sweat: Status: Investigating :mag: Status: Stale :bread: Type: New Feature :heavy_plus_sign: | ## Feature Description
SPDZ has 7x more matmul ops than plaintext computations. Using cuda to optimize those would give significant speedups!
SecureNN / FSS use many intensive ops that could also be improved by cuda support. In particular, FSS uses many invocations to AES, and cuda support for AES from https://github.com/pytorch/csprng could be useful!
First investigation is being made here: https://github.com/OpenMined/PySyft/tree/ryffel/cuda + https://github.com/OpenMined/PySyft/tree/ryffel/cuda_fss
Help welcome! :)
| 1.0 | Cuda support for Encrypted Computation - ## Feature Description
SPDZ has 7x more matmul ops than plaintext computations. Using cuda to optimize those would give significant speedups!
SecureNN / FSS use many intensive ops that could also be improved by cuda support. In particular, FSS uses many invocations to AES, and cuda support for AES from https://github.com/pytorch/csprng could be useful!
First investigation is being made here: https://github.com/OpenMined/PySyft/tree/ryffel/cuda + https://github.com/OpenMined/PySyft/tree/ryffel/cuda_fss
Help welcome! :)
| priority | cuda support for encrypted computation feature description spdz has more matmul ops than plaintext computations using cuda to optimize those would give significant speedups securenn fss use many intensive ops that could also be improved by cuda support in particular fss uses many invocations to aes and cuda support for aes from could be useful first investigation is being made here help welcome | 1 |
791,484 | 27,865,156,322 | IssuesEvent | 2023-03-21 09:45:38 | Snapmaker/Luban | https://api.github.com/repos/Snapmaker/Luban | closed | [Feature request] Change Nozzle size | Category: 3DP Priority: High | Is it possible to change to nozzle size?
I would like to print with woodfill, which recommends .5 mm nozzles. It would also reduce print time with large prints without too much loss in quality.
In the Luban software I cannot find the nozzle settings, only layer height etc. | 1.0 | [Feature request] Change Nozzle size - Is it possible to change to nozzle size?
I would like to print with woodfill, which recommends .5 mm nozzles. It would also reduce print time with large prints without too much loss in quality.
In the Luban software I cannot find the nozzle settings, only layer height etc. | priority | change nozzle size is it possible to change to nozzle size i would like to print with woodfill which recommends mm nozzles it would also reduce print time with large prints without too much loss in quality in the luban software i cannot find the nozzle settings only layer height etc | 1 |
758,447 | 26,555,805,767 | IssuesEvent | 2023-01-20 11:55:11 | saudalnasser/starlux | https://api.github.com/repos/saudalnasser/starlux | opened | feat: user commands | type: feature priority: high | ## Problem
need to create user commands in an organized way and with minimal effort.
## Solution(s)
provide an easy way of:
- creating user commands
- organizing user commands
| 1.0 | feat: user commands - ## Problem
need to create user commands in an organized way and with minimal effort.
## Solution(s)
provide an easy way of:
- creating user commands
- organizing user commands
| priority | feat user commands problem need to create user commands in an organized way and with minimal effort solution s provide an easy way of creating user commands organizing user commands | 1 |
481,274 | 13,882,928,647 | IssuesEvent | 2020-10-18 09:29:47 | AY2021S1-CS2103-F10-4/tp | https://api.github.com/repos/AY2021S1-CS2103-F10-4/tp | closed | Disallow assignment of worker to another role in the same shift | class.Assignment priority.High | A worker should not be assigned to more than one role in a single shift. A new assignment of the same worker to the same shift should replace the old. | 1.0 | Disallow assignment of worker to another role in the same shift - A worker should not be assigned to more than one role in a single shift. A new assignment of the same worker to the same shift should replace the old. | priority | disallow assignment of worker to another role in the same shift a worker should not be assigned to more than one role in a single shift a new assignment of the same worker to the same shift should replace the old | 1 |
387,766 | 11,470,268,737 | IssuesEvent | 2020-02-09 01:34:19 | minetest/minetest_game | https://api.github.com/repos/minetest/minetest_game | closed | Papyrus wants default:dirt but only default:dry_dirt exists in savanna_shore biome | Bug High priority Mapgen | Unless I'm completely missing something, 5.1.0 generates no papyrus at all, since it includes #2422 (which broke this) but not #2535 (which added an alternative).
| 1.0 | Papyrus wants default:dirt but only default:dry_dirt exists in savanna_shore biome - Unless I'm completely missing something, 5.1.0 generates no papyrus at all, since it includes #2422 (which broke this) but not #2535 (which added an alternative).
| priority | papyrus wants default dirt but only default dry dirt exists in savanna shore biome unless i m completely missing something generates no papyrus at all since it includes which broke this but not which added an alternative | 1 |
308,277 | 9,437,255,438 | IssuesEvent | 2019-04-13 13:55:15 | cs2103-ay1819s2-w10-2/main | https://api.github.com/repos/cs2103-ay1819s2-w10-2/main | closed | As a software enginerring manager, I can find projects based on deadline | priority.High type.Story | so that I can check if there are any projects that will be due soon and plan my workload properly. | 1.0 | As a software enginerring manager, I can find projects based on deadline - so that I can check if there are any projects that will be due soon and plan my workload properly. | priority | as a software enginerring manager i can find projects based on deadline so that i can check if there are any projects that will be due soon and plan my workload properly | 1 |
5,703 | 2,578,332,629 | IssuesEvent | 2015-02-12 22:32:35 | nprapps/syria | https://api.github.com/repos/nprapps/syria | closed | Get final (or at least latest) text | Priority: High | The latest edit sounds pretty significant. Should get that onto the page so we can work with it. | 1.0 | Get final (or at least latest) text - The latest edit sounds pretty significant. Should get that onto the page so we can work with it. | priority | get final or at least latest text the latest edit sounds pretty significant should get that onto the page so we can work with it | 1 |
344,837 | 10,349,662,482 | IssuesEvent | 2019-09-04 23:23:31 | oslc-op/jira-migration-landfill | https://api.github.com/repos/oslc-op/jira-migration-landfill | closed | Enforce consecutive order of the trs:order numbers | Core: TRS Jira: trs Priority: High Xtra: Jira | If TRS events are to be distributed via messaging systems without strict ordering guarantees (eg partitioned Kafka topics), a Resequencer EIP pattern might need to be applied. The pattern requires the message order ids to be sequential in order to unambiguously define whether there are any out-of-order messages still missing from the internal resequencing buffer.
I think a single atomic counter is not too much to ask from the TRS server implementers.
[http://www.enterpriseintegrationpatterns.com/patterns/messaging/Resequencer.html](http://www.enterpriseintegrationpatterns.com/patterns/messaging/Resequencer.html)
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-93 (opened by @berezovskyi; previously assigned to @undefined)_
| 1.0 | Enforce consecutive order of the trs:order numbers - If TRS events are to be distributed via messaging systems without strict ordering guarantees (eg partitioned Kafka topics), a Resequencer EIP pattern might need to be applied. The pattern requires the message order ids to be sequential in order to unambiguously define whether there are any out-of-order messages still missing from the internal resequencing buffer.
I think a single atomic counter is not too much to ask from the TRS server implementers.
[http://www.enterpriseintegrationpatterns.com/patterns/messaging/Resequencer.html](http://www.enterpriseintegrationpatterns.com/patterns/messaging/Resequencer.html)
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-93 (opened by @berezovskyi; previously assigned to @undefined)_
| priority | enforce consecutive order of the trs order numbers if trs events are to be distributed via messaging systems without strict ordering guarantees eg partitioned kafka topics a resequencer eip pattern might need to be applied the pattern requires the message order ids to be sequential in order to unambiguously define whether there are any out of order messages still missing from the internal resequencing buffer i think a single atomic counter is not too much to ask from the trs server implementers migrated from opened by berezovskyi previously assigned to undefined | 1 |
168,523 | 6,377,382,966 | IssuesEvent | 2017-08-02 09:53:43 | AppertaFoundation/Diadem | https://api.github.com/repos/AppertaFoundation/Diadem | opened | Users can get context-sensitive help from user manual sourced content | high priority | - [ ] complete the 'message' GUI module
- [ ] extract context specfic text from the existing 'manual' document
- [ ] update context specific text for each screen in the app
- [ ] create a local store for this text (in a form that can be updated easily)
- [ ] create a system of extracting relevant content for each screen
- [ ] create 'calling' GUI in the app (i.e. a "I" button)
| 1.0 | Users can get context-sensitive help from user manual sourced content - - [ ] complete the 'message' GUI module
- [ ] extract context specfic text from the existing 'manual' document
- [ ] update context specific text for each screen in the app
- [ ] create a local store for this text (in a form that can be updated easily)
- [ ] create a system of extracting relevant content for each screen
- [ ] create 'calling' GUI in the app (i.e. a "I" button)
| priority | users can get context sensitive help from user manual sourced content complete the message gui module extract context specfic text from the existing manual document update context specific text for each screen in the app create a local store for this text in a form that can be updated easily create a system of extracting relevant content for each screen create calling gui in the app i e a i button | 1 |
528,455 | 15,367,388,218 | IssuesEvent | 2021-03-02 03:11:26 | Katsute/Mal4J | https://api.github.com/repos/Katsute/Mal4J | closed | Unicode characters are not being decoded correctly. | MyAnimeList API (external) critical bug high priority | For some reason Japanese characters are being read as string literals instead of unicode.
Ex: `\u3055\u304f\u3089\u8358\u306e\u30da\u30c3\u30c8\u306a\u5f7c\u5973` is presented instead of `さくら荘のペットな彼女`.
Either the API is returning unicode escape strings, or the API call is not reading the Json properly. | 1.0 | Unicode characters are not being decoded correctly. - For some reason Japanese characters are being read as string literals instead of unicode.
Ex: `\u3055\u304f\u3089\u8358\u306e\u30da\u30c3\u30c8\u306a\u5f7c\u5973` is presented instead of `さくら荘のペットな彼女`.
Either the API is returning unicode escape strings, or the API call is not reading the Json properly. | priority | unicode characters are not being decoded correctly for some reason japanese characters are being read as string literals instead of unicode ex is presented instead of さくら荘のペットな彼女 either the api is returning unicode escape strings or the api call is not reading the json properly | 1 |
383,881 | 11,363,794,585 | IssuesEvent | 2020-01-27 05:55:34 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Introduce method to check domain availability in TenantMgtAdminService | Affected/5.9.0 Complexity/Low Component/Tenant Mgt Priority/High | Currently, there is no method to check domain availability. As a workaround, we can call the get tenant by tenant domain. But this involves an extra database call. So better to have a separate method | 1.0 | Introduce method to check domain availability in TenantMgtAdminService - Currently, there is no method to check domain availability. As a workaround, we can call the get tenant by tenant domain. But this involves an extra database call. So better to have a separate method | priority | introduce method to check domain availability in tenantmgtadminservice currently there is no method to check domain availability as a workaround we can call the get tenant by tenant domain but this involves an extra database call so better to have a separate method | 1 |
400,689 | 11,779,208,486 | IssuesEvent | 2020-03-16 17:35:42 | traffic-control-fyp-aub/ns3-gym | https://api.github.com/repos/traffic-control-fyp-aub/ns3-gym | closed | Trained PPO online agent square scenario plot points | High Priority | Run online trained PPO agent (cars=10) on square scenario as test to collect plot points. | 1.0 | Trained PPO online agent square scenario plot points - Run online trained PPO agent (cars=10) on square scenario as test to collect plot points. | priority | trained ppo online agent square scenario plot points run online trained ppo agent cars on square scenario as test to collect plot points | 1 |
208,109 | 7,135,785,690 | IssuesEvent | 2018-01-23 03:00:31 | cssconf/2018.cssconf.eu | https://api.github.com/repos/cssconf/2018.cssconf.eu | closed | Add a top bar navigation, with a mobile version | has PR priority: high | @RalloTubbs we'll need this soon as we'll add more pages, so navigating is easy – right now it's inconvenient and hard to find in the footer, esp. on longer pages | 1.0 | Add a top bar navigation, with a mobile version - @RalloTubbs we'll need this soon as we'll add more pages, so navigating is easy – right now it's inconvenient and hard to find in the footer, esp. on longer pages | priority | add a top bar navigation with a mobile version rallotubbs we ll need this soon as we ll add more pages so navigating is easy – right now it s inconvenient and hard to find in the footer esp on longer pages | 1 |
3,583 | 2,538,767,974 | IssuesEvent | 2015-01-27 10:07:57 | newca12/gapt | https://api.github.com/repos/newca12/gapt | closed | Sequent Occurence to Sequent: LKsK and Resolution | 1 star Component-LogicalDataStructures enhancement imported Milestone-Release2.0 Priority-High | _From [Martin.R...@googlemail.com](https://code.google.com/u/113922991942493634983/) on September 14, 2011 13:46:29_
This is just for splitting up the work on the conversion.
_Original issue: http://code.google.com/p/gapt/issues/detail?id=174_ | 1.0 | Sequent Occurence to Sequent: LKsK and Resolution - _From [Martin.R...@googlemail.com](https://code.google.com/u/113922991942493634983/) on September 14, 2011 13:46:29_
This is just for splitting up the work on the conversion.
_Original issue: http://code.google.com/p/gapt/issues/detail?id=174_ | priority | sequent occurence to sequent lksk and resolution from on september this is just for splitting up the work on the conversion original issue | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.