Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
212,896 | 7,243,823,068 | IssuesEvent | 2018-02-14 13:12:35 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | opened | Portal redirection: Destination URL should be ignored for portal detection | Priority: Medium Type: Bug | As experienced by @lzammit, he was redirected to `http://captive.apple.com/hotspot-detect.html` at the end of the portal.
He should instead have been redirected to the redirect URL value in the profile
Not sure if its v8 related or affects v7 as well. If v7 is affected I'll cherry-pick in maintenance | 1.0 | Portal redirection: Destination URL should be ignored for portal detection - As experienced by @lzammit, he was redirected to `http://captive.apple.com/hotspot-detect.html` at the end of the portal.
He should instead have been redirected to the redirect URL value in the profile
Not sure if its v8 related or affects v7 as well. If v7 is affected I'll cherry-pick in maintenance | priority | portal redirection destination url should be ignored for portal detection as experienced by lzammit he was redirected to at the end of the portal he should instead have been redirected to the redirect url value in the profile not sure if its related or affects as well if is affected i ll cherry pick in maintenance | 1 |
11,541 | 2,610,140,426 | IssuesEvent | 2015-02-26 18:44:16 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Allow resizing window | auto-migrated Priority-Medium Type-Enhancement | ```
I have a dual-monitor setup with my primary monitor on the right hand side.
The first window (where you choose single player vs network) correct shows only
on the primary monitor.
The actual game tries to span both monitors.
Under settings, I can choose only 2304x1024.
I'm on Ubuntu 10.10 / NVidia 6600 / TwinView setup.
```
-----
Original issue reported on code.google.com by `igi...@gmail.com` on 18 Nov 2010 at 10:33 | 1.0 | Allow resizing window - ```
I have a dual-monitor setup with my primary monitor on the right hand side.
The first window (where you choose single player vs network) correct shows only
on the primary monitor.
The actual game tries to span both monitors.
Under settings, I can choose only 2304x1024.
I'm on Ubuntu 10.10 / NVidia 6600 / TwinView setup.
```
-----
Original issue reported on code.google.com by `igi...@gmail.com` on 18 Nov 2010 at 10:33 | priority | allow resizing window i have a dual monitor setup with my primary monitor on the right hand side the first window where you choose single player vs network correct shows only on the primary monitor the actual game tries to span both monitors under settings i can choose only i m on ubuntu nvidia twinview setup original issue reported on code google com by igi gmail com on nov at | 1 |
793,343 | 27,991,644,591 | IssuesEvent | 2023-03-27 04:38:47 | WordPress/openverse | https://api.github.com/repos/WordPress/openverse | opened | Consider disabling exposed ports in CI to avoid port conflict flakiness | 🟨 priority: medium 🛠 goal: fix 🤖 aspect: dx 🧱 stack: mgmt | ## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
CI jobs sometimes fail due to port conflicts.
cf #990 and #200
## Description
<!-- Describe the feature and how it solves the problem. -->
One way to avoid this is to not bind ports at all in CI. At a glance, I don't think we need to bind ports _at all_ in CI as we _never_ make requests to running containers in CI: everything happens _inside_ the containers and we just check the output.
To implement this we'd need to remove all port declarations from the base `docker-compose.yml` file and add them in a `docker-compose.development.yml` as overrides instead. See this comment: https://github.com/docker/compose/issues/3729#issuecomment-232473164
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
Do the retry approach that @dhruvkb implemented in #990. This could work but is potentially still flaky and adds complexity to the CI. | 1.0 | Consider disabling exposed ports in CI to avoid port conflict flakiness - ## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
CI jobs sometimes fail due to port conflicts.
cf #990 and #200
## Description
<!-- Describe the feature and how it solves the problem. -->
One way to avoid this is to not bind ports at all in CI. At a glance, I don't think we need to bind ports _at all_ in CI as we _never_ make requests to running containers in CI: everything happens _inside_ the containers and we just check the output.
To implement this we'd need to remove all port declarations from the base `docker-compose.yml` file and add them in a `docker-compose.development.yml` as overrides instead. See this comment: https://github.com/docker/compose/issues/3729#issuecomment-232473164
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
Do the retry approach that @dhruvkb implemented in #990. This could work but is potentially still flaky and adds complexity to the CI. | priority | consider disabling exposed ports in ci to avoid port conflict flakiness problem ci jobs sometimes fail due to port conflicts cf and description one way to avoid this is to not bind ports at all in ci at a glance i don t think we need to bind ports at all in ci as we never make requests to running containers in ci everything happens inside the containers and we just check the output to implement this we d need to remove all port declarations from the base docker compose yml file and add them in a docker compose development yml as overrides instead see this comment alternatives do the retry approach that dhruvkb implemented in this could work but is potentially still flaky and adds complexity to the ci | 1 |
690,964 | 23,679,710,931 | IssuesEvent | 2022-08-28 16:08:03 | pokt-network/pocket | https://api.github.com/repos/pokt-network/pocket | opened | [Development] Documentation for local debugging | priority:medium infra tooling | ## Objective
Document how to use the `dlv` debugger in LocalNet.
## Origin Document
The `LocalNet` CLI and logging are useful, but we need to be able to step through the code in unit tests and in a dev environment
## Goals
- [ ] Anyone who spins up a V1 LocalNet should be able to easily use the debugger, regardless of whether they have experience with it in the past
- [ ] Support `vscode` and `Goland` as a first iteration
## Deliverables
- [ ] Documentation on how to use the `dlv` debugger using `vscode`
- [ ] Documentation on how to use the `dlv` debugger using `Goland`
- [ ] A gif or video with an example of how the debugger works
- [ ] Modifications to the supporting tooling (config files, docker-compose files, etc if necessary)
## Non-goals / Non-deliverables
- Modifying the core protocol implementation
- Enabling debugging in DevNet
- Introducing any new telemetry to the codebase
## General issue deliverables
- [ ] Update the appropriate CHANGELOG
- [ ] Update any relevant READMEs (local and/or global)
- [ ] Update any relevant global documentation & references
- [ ] If applicable, update the source code tree explanation
- [ ] If applicable, add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
## [Optional] Testing Methodology
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
---
**Creator**: @Olshansk
**Co-Owners**: @andrewnguyen22
| 1.0 | [Development] Documentation for local debugging - ## Objective
Document how to use the `dlv` debugger in LocalNet.
## Origin Document
The `LocalNet` CLI and logging are useful, but we need to be able to step through the code in unit tests and in a dev environment
## Goals
- [ ] Anyone who spins up a V1 LocalNet should be able to easily use the debugger, regardless of whether they have experience with it in the past
- [ ] Support `vscode` and `Goland` as a first iteration
## Deliverables
- [ ] Documentation on how to use the `dlv` debugger using `vscode`
- [ ] Documentation on how to use the `dlv` debugger using `Goland`
- [ ] A gif or video with an example of how the debugger works
- [ ] Modifications to the supporting tooling (config files, docker-compose files, etc if necessary)
## Non-goals / Non-deliverables
- Modifying the core protocol implementation
- Enabling debugging in DevNet
- Introducing any new telemetry to the codebase
## General issue deliverables
- [ ] Update the appropriate CHANGELOG
- [ ] Update any relevant READMEs (local and/or global)
- [ ] Update any relevant global documentation & references
- [ ] If applicable, update the source code tree explanation
- [ ] If applicable, add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
## [Optional] Testing Methodology
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
---
**Creator**: @Olshansk
**Co-Owners**: @andrewnguyen22
| priority | documentation for local debugging objective document how to use the dlv debugger in localnet origin document the localnet cli and logging are useful but we need to be able to step through the code in unit tests and in a dev environment goals anyone who spins up a localnet should be able to easily use the debugger regardless of whether they have experience with it in the past support vscode and goland as a first iteration deliverables documentation on how to use the dlv debugger using vscode documentation on how to use the dlv debugger using goland a gif or video with an example of how the debugger works modifications to the supporting tooling config files docker compose files etc if necessary non goals non deliverables modifying the core protocol implementation enabling debugging in devnet introducing any new telemetry to the codebase general issue deliverables update the appropriate changelog update any relevant readmes local and or global update any relevant global documentation references if applicable update the source code tree explanation if applicable add or update a state sequence or flowchart diagram using testing methodology localnet verify a localnet is still functioning correctly by following the instructions at creator olshansk co owners | 1 |
657,834 | 21,869,490,858 | IssuesEvent | 2022-05-19 03:04:07 | pixley/TimelineBuilder | https://api.github.com/repos/pixley/TimelineBuilder | opened | Calendar of Harptos script | type: feature status: to do priority: medium | Create a Python script to implement the Calendar of Harptos, the main calendar for the Faerun of Forgotten Realms. | 1.0 | Calendar of Harptos script - Create a Python script to implement the Calendar of Harptos, the main calendar for the Faerun of Forgotten Realms. | priority | calendar of harptos script create a python script to implement the calendar of harptos the main calendar for the faerun of forgotten realms | 1 |
304,368 | 9,331,347,658 | IssuesEvent | 2019-03-28 09:33:09 | CS2103-AY1819S2-W10-1/main | https://api.github.com/repos/CS2103-AY1819S2-W10-1/main | closed | UI: Display current context | priority.Medium type.Enhancement | Right now, users only know which context they are in after executing a context-switching command, so displaying the context somewhere would help users orientate themselves. | 1.0 | UI: Display current context - Right now, users only know which context they are in after executing a context-switching command, so displaying the context somewhere would help users orientate themselves. | priority | ui display current context right now users only know which context they are in after executing a context switching command so displaying the context somewhere would help users orientate themselves | 1 |
5,324 | 2,574,225,165 | IssuesEvent | 2015-02-11 15:46:37 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | PyCondorPlugin error | Medium Priority | Message: Unhandled exception while calling update method for plugin PyCondorPlugin
Failed to fetch ads from schedd | 1.0 | PyCondorPlugin error - Message: Unhandled exception while calling update method for plugin PyCondorPlugin
Failed to fetch ads from schedd | priority | pycondorplugin error message unhandled exception while calling update method for plugin pycondorplugin failed to fetch ads from schedd | 1 |
415,543 | 12,130,453,266 | IssuesEvent | 2020-04-23 01:36:53 | minio/mc | https://api.github.com/repos/minio/mc | closed | allow fetching individual object's public url | community priority: medium stale | ## Expected behavior
It is possible to fetch an object's public url without side-effects.
## Actual behavior
The only current way (from what I can tell) to get an individual object's public url is using the following command:
`mc share download --json --expire=1s $OBJECT | jq -r .url`
This is undesirable because it creates a presigned url (even if it expires almost immediately).
The alternatives:
1. `mc policy -r links $OBJECT` -> this does not produce any output against an individual object, with or without `-r`, it appears to only enumerate whole buckets/prefixes.
2. `mc ls --json $OBJECT` -> does not have such a field
3. `mc find $OBJECT --print {url}` -> equivalent to share download, but with a long expiry
## Steps to reproduce the behavior
Go through the documentation and find no alternatives.
## mc --version
- `mcli version RELEASE.2019-12-17T23-26-28Z`
## System information
Arch Linux amd64
## Possible Solutions
a) allow `mc policy links $OBJECT` to run against individual objects
b) add public url as an information point under `mc ls --json`, if the object in question is an object and not a directory (and, potentially, has a download policy)
c) add a {puburl} property to `mc find`'s print
d) have some way to print a host's prefix (e.g https://play.min.io:9000), for both config-based hosts and environment-configured hosts
## Alternate Scenario
It's possible that this functionality already exists, but is either not documented, or documented in an unusual place, in which case this bug should be about improving the documentation to make it more obvious. | 1.0 | allow fetching individual object's public url - ## Expected behavior
It is possible to fetch an object's public url without side-effects.
## Actual behavior
The only current way (from what I can tell) to get an individual object's public url is using the following command:
`mc share download --json --expire=1s $OBJECT | jq -r .url`
This is undesirable because it creates a presigned url (even if it expires almost immediately).
The alternatives:
1. `mc policy -r links $OBJECT` -> this does not produce any output against an individual object, with or without `-r`, it appears to only enumerate whole buckets/prefixes.
2. `mc ls --json $OBJECT` -> does not have such a field
3. `mc find $OBJECT --print {url}` -> equivalent to share download, but with a long expiry
## Steps to reproduce the behavior
Go through the documentation and find no alternatives.
## mc --version
- `mcli version RELEASE.2019-12-17T23-26-28Z`
## System information
Arch Linux amd64
## Possible Solutions
a) allow `mc policy links $OBJECT` to run against individual objects
b) add public url as an information point under `mc ls --json`, if the object in question is an object and not a directory (and, potentially, has a download policy)
c) add a {puburl} property to `mc find`'s print
d) have some way to print a host's prefix (e.g https://play.min.io:9000), for both config-based hosts and environment-configured hosts
## Alternate Scenario
It's possible that this functionality already exists, but is either not documented, or documented in an unusual place, in which case this bug should be about improving the documentation to make it more obvious. | priority | allow fetching individual object s public url expected behavior it is possible to fetch an object s public url without side effects actual behavior the only current way from what i can tell to get an individual object s public url is using the following command mc share download json expire object jq r url this is undesirable because it creates a presigned url even if it expires almost immediately the alternatives mc policy r links object this does not produce any output against an individual object with or without r it appears to only enumerate whole buckets prefixes mc ls json object does not have such a field mc find object print url equivalent to share download but with a long expiry steps to reproduce the behavior go through the documentation and find no alternatives mc version mcli version release system information arch linux possible solutions a allow mc policy links object to run against individual objects b add public url as an information point under mc ls json if the object in question is an object and not a directory and potentially has a download policy c add a puburl property to mc find s print d have some way to print a host s prefix e g for both config based hosts and environment configured hosts alternate scenario it s possible that this functionality already exists but is either not documented or documented in an unusual place in which case this bug should be about improving the documentation to make it more obvious | 1 |
150,043 | 5,733,138,889 | IssuesEvent | 2017-04-21 16:27:49 | knipferrc/plate | https://api.github.com/repos/knipferrc/plate | closed | Content flash unauthorized | Priority: Medium Type: Bug | If manually changing the URL to a page which you are not authorized you get a flash of unauthorized content. | 1.0 | Content flash unauthorized - If manually changing the URL to a page which you are not authorized you get a flash of unauthorized content. | priority | content flash unauthorized if manually changing the url to a page which you are not authorized you get a flash of unauthorized content | 1 |
676,399 | 23,124,241,495 | IssuesEvent | 2022-07-28 02:47:09 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [docdb] RegistrationTest.TestTabletReports is flaky some 3/1000 times | kind/bug area/docdb priority/medium | Jira Link: [DB-1366](https://yugabyte.atlassian.net/browse/DB-1366)
Follow up to #3327. Seems like there's a genuine logical bug in the test expectation.
| 1.0 | [docdb] RegistrationTest.TestTabletReports is flaky some 3/1000 times - Jira Link: [DB-1366](https://yugabyte.atlassian.net/browse/DB-1366)
Follow up to #3327. Seems like there's a genuine logical bug in the test expectation.
| priority | registrationtest testtabletreports is flaky some times jira link follow up to seems like there s a genuine logical bug in the test expectation | 1 |
25,887 | 2,684,026,415 | IssuesEvent | 2015-03-28 15:47:10 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | bring window to top with putty in console mode | 1 star bug imported Priority-Medium | _From [mickem](https://code.google.com/u/mickem/) on October 28, 2011 22:45:21_
When using putty in console mode you cannot bring the window to top by clicking it.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=448_ | 1.0 | bring window to top with putty in console mode - _From [mickem](https://code.google.com/u/mickem/) on October 28, 2011 22:45:21_
When using putty in console mode you cannot bring the window to top by clicking it.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=448_ | priority | bring window to top with putty in console mode from on october when using putty in console mode you cannot bring the window to top by clicking it original issue | 1 |
540,047 | 15,798,817,460 | IssuesEvent | 2021-04-02 19:30:23 | itslupus/gamersnet | https://api.github.com/repos/itslupus/gamersnet | closed | Profile Game Ranking | medium priority user story | **Description**:
As a user, I want to be able to manage my game specific ranks
**Acceptance Criteria**:
* User profile displays the user's game ranks
* User can update the game ranks
**Dev Tasks**:
[Edit game ranks backend](https://github.com/itslupus/gamersnet/issues/48)
[UI to edit game ranks](https://github.com/itslupus/gamersnet/issues/47)
**Story Points (1 - 5**): 3
| 1.0 | Profile Game Ranking - **Description**:
As a user, I want to be able to manage my game specific ranks
**Acceptance Criteria**:
* User profile displays the user's game ranks
* User can update the game ranks
**Dev Tasks**:
[Edit game ranks backend](https://github.com/itslupus/gamersnet/issues/48)
[UI to edit game ranks](https://github.com/itslupus/gamersnet/issues/47)
**Story Points (1 - 5**): 3
| priority | profile game ranking description as a user i want to be able to manage my game specific ranks acceptance criteria user profile displays the user s game ranks user can update the game ranks dev tasks story points | 1 |
542,583 | 15,862,803,537 | IssuesEvent | 2021-04-08 12:04:33 | sopra-fs21-group-08/sopra-fs21-group08-client | https://api.github.com/repos/sopra-fs21-group-08/sopra-fs21-group08-client | closed | Create a pop up window with the game rules that can be accessed from the lobby of the game. | lobby medium priority task | Time estimate: 3h
This task is part of user story #25 | 1.0 | Create a pop up window with the game rules that can be accessed from the lobby of the game. - Time estimate: 3h
This task is part of user story #25 | priority | create a pop up window with the game rules that can be accessed from the lobby of the game time estimate this task is part of user story | 1 |
515,392 | 14,961,744,743 | IssuesEvent | 2021-01-27 08:16:28 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | opened | Add Missing Image Dimensions is not applied if only single attribute (width or height) is missing | module: media priority: medium severity: major type: bug | **Before submitting an issue please check that you’ve completed the following steps:**
- [ ] Made sure you’re on the latest version
- [ ] Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When image is missing width or height, Add missing image dimensions is not working (it only works only when both Height and Width are missing)
**Precondition**
Having a page with image that is missing height or width
**To Reproduce**
Steps to reproduce the behavior:
1. Endable add missing image dimensions in media page
2. Clear cache
3. Visit the page mentioned in precondition
4. In source image dimensions are still missing
**Expected behavior**
Image dimensions are added as same as that when both width and height were missing
**Screenshots**

**Additional context**
Add missing image dimensions is working fine whenever **both** height and width are missing
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | Add Missing Image Dimensions is not applied if only single attribute (width or height) is missing - **Before submitting an issue please check that you’ve completed the following steps:**
- [ ] Made sure you’re on the latest version
- [ ] Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When image is missing width or height, Add missing image dimensions is not working (it only works only when both Height and Width are missing)
**Precondition**
Having a page with image that is missing height or width
**To Reproduce**
Steps to reproduce the behavior:
1. Endable add missing image dimensions in media page
2. Clear cache
3. Visit the page mentioned in precondition
4. In source image dimensions are still missing
**Expected behavior**
Image dimensions are added as same as that when both width and height were missing
**Screenshots**

**Additional context**
Add missing image dimensions is working fine whenever **both** height and width are missing
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| priority | add missing image dimensions is not applied if only single attribute width or height is missing before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug when image is missing width or height add missing image dimensions is not working it only works only when both height and width are missing precondition having a page with image that is missing height or width to reproduce steps to reproduce the behavior endable add missing image dimensions in media page clear cache visit the page mentioned in precondition in source image dimensions are still missing expected behavior image dimensions are added as same as that when both width and height were missing screenshots additional context add missing image dimensions is working fine whenever both height and width are missing backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
267,361 | 8,388,142,295 | IssuesEvent | 2018-10-09 04:42:05 | MagiCircles/BanGDream | https://api.github.com/repos/MagiCircles/BanGDream | closed | Replace Stamp Fields With Foreign Key To Stamp Assets | feature medium priority optimization | Pretty Self-Explanatory. Title would function the same as Stamp Translation, and each image would be displayed under their proper version.
Would reduce images uploaded in the future & duplication of inputted info, so good all around. Most difficult thing would probably be configuring it to work with our special event magicollection/js code. | 1.0 | Replace Stamp Fields With Foreign Key To Stamp Assets - Pretty Self-Explanatory. Title would function the same as Stamp Translation, and each image would be displayed under their proper version.
Would reduce images uploaded in the future & duplication of inputted info, so good all around. Most difficult thing would probably be configuring it to work with our special event magicollection/js code. | priority | replace stamp fields with foreign key to stamp assets pretty self explanatory title would function the same as stamp translation and each image would be displayed under their proper version would reduce images uploaded in the future duplication of inputted info so good all around most difficult thing would probably be configuring it to work with our special event magicollection js code | 1 |
818,951 | 30,712,941,941 | IssuesEvent | 2023-07-27 11:04:56 | kubesphere/ks-devops | https://api.github.com/repos/kubesphere/ks-devops | closed | When executing a pipeline that requires input parameters, no input box pops up | kind/bug priority/medium | **Versions used**
KubeSphere: `v3.2.0`
How To Reproduce
Steps to reproduce the behavior:
1、Go to pipeline list
2、Select pipeline
3、Click on 'Run'

```
pipeline {
agent any
parameters {
string(name: 'PERSON', defaultValue: 'Mr Jenkins', description: 'Who should I say hello to?')
}
stages {
stage('Example') {
steps {
echo "Hello ${params.PERSON}"
}
}
}
}
```
**Expected behavior**
When you click to run, the following pop-up window is displayed.

/kind bug
/cc @kubesphere/sig-devops
/priority medium | 1.0 | When executing a pipeline that requires input parameters, no input box pops up - **Versions used**
KubeSphere: `v3.2.0`
How To Reproduce
Steps to reproduce the behavior:
1、Go to pipeline list
2、Select pipeline
3、Click on 'Run'

```
pipeline {
agent any
parameters {
string(name: 'PERSON', defaultValue: 'Mr Jenkins', description: 'Who should I say hello to?')
}
stages {
stage('Example') {
steps {
echo "Hello ${params.PERSON}"
}
}
}
}
```
**Expected behavior**
When you click to run, the following pop-up window is displayed.

/kind bug
/cc @kubesphere/sig-devops
/priority medium | priority | when executing a pipeline that requires input parameters no input box pops up versions used kubesphere how to reproduce steps to reproduce the behavior 、go to pipeline list 、select pipeline 、click on run pipeline agent any parameters string name person defaultvalue mr jenkins description who should i say hello to stages stage example steps echo hello params person expected behavior when you click to run the following pop up window is displayed kind bug cc kubesphere sig devops priority medium | 1 |
45,112 | 2,920,434,479 | IssuesEvent | 2015-06-24 18:55:09 | ashanbh/chrome-rest-client | https://api.github.com/repos/ashanbh/chrome-rest-client | closed | Support HAR import | auto-migrated OpSys-All Priority-Medium Type- Usability | ```
Feature suggestion: allow to open a request from a HAR/curl format.
Since Chrome Developer Tools allows to copy the request from the Network tab on
the HAR/curl format, it would be nice that the application can directly import
a request parsing that format, making the replaying of a specific request
easier.
Thanks!
```
Original issue reported on code.google.com by `Alpha...@gmail.com` on 19 Aug 2013 at 4:50 | 1.0 | Support HAR import - ```
Feature suggestion: allow to open a request from a HAR/curl format.
Since Chrome Developer Tools allows to copy the request from the Network tab on
the HAR/curl format, it would be nice that the application can directly import
a request parsing that format, making the replaying of a specific request
easier.
Thanks!
```
Original issue reported on code.google.com by `Alpha...@gmail.com` on 19 Aug 2013 at 4:50 | priority | support har import feature suggestion allow to open a request from a har curl format since chrome developer tools allows to copy the request from the network tab on the har curl format it would be nice that the application can directly import a request parsing that format making the replaying of a specific request easier thanks original issue reported on code google com by alpha gmail com on aug at | 1 |
523,406 | 15,181,149,809 | IssuesEvent | 2021-02-15 02:33:41 | gianfrancodumoulinbertucci/SOEN341 | https://api.github.com/repos/gianfrancodumoulinbertucci/SOEN341 | opened | (Feature) As a user, I want to see a timeline page | Priority: Medium feature front-end | ### **FRONT END**
To do:
# Template for the timeline page needs to be created and should have the following:
- Timeline page should have a navbar
- navbar should contain the <a href = "/timeline">LOGO<a/> on the top left and a "POST" button on the top right
- timeline page should have places for posted pictures to be displayed | 1.0 | (Feature) As a user, I want to see a timeline page - ### **FRONT END**
To do:
# Template for the timeline page needs to be created and should have the following:
- Timeline page should have a navbar
- navbar should contain the <a href = "/timeline">LOGO<a/> on the top left and a "POST" button on the top right
- timeline page should have places for posted pictures to be displayed | priority | feature as a user i want to see a timeline page front end to do template for the timeline page needs to be created and should have the following timeline page should have a navbar navbar should contain the logo on the top left and a post button on the top right timeline page should have places for posted pictures to be displayed | 1 |
55,275 | 3,072,620,009 | IssuesEvent | 2015-08-19 17:50:27 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | junit.framework.AssertionFailedError: Button with index 2131230847 is not available! | bug imported invalid Priority-Medium | _From [Surhum...@gmail.com](https://code.google.com/u/107563385414450670028/) on April 25, 2012 02:53:27_
What steps will reproduce the problem? 1.solo = new Solo(getInstrumentation(), getActivity());
2.solo.clickOnButton(dk.lector.ao.mobile.R.id.productButton);
3.solo.assertCurrentActivity("ProductSearch", ProductSearchActivity.class); What is the expected output? What do you see instead? I expected an Activity launched.
but got the title of this isssue instead. What version of the product are you using? On what operating system? using 3.2.1 on ubuntu 11.10. with eclipse Please provide any additional information below. My manifest does have the wiki manifest added, no change.
Here is my testclass:
package dk.lector.ao.mobile.test;
import android.test.ActivityInstrumentationTestCase2;
import com.jayway.android.robotium.solo.Solo;
import dk.lector.ao.mobile.AOinThePocket;
import dk.lector.ao.mobile.product.ProductSearchActivity;
public class TestAOinThePocket extends ActivityInstrumentationTestCase2<AOinThePocket> {
private Solo solo;
public TestAOinThePocket() {
super("dk.lector.ao.mobile", AOinThePocket.class);
}
@Override
protected void setUp() throws Exception {
super.setUp();
solo = new Solo(getInstrumentation(), getActivity());
synchronized (solo) {
solo.wait(2000);
}
}
public void testDisplayBlackBox() {
solo.clickOnButton(dk.lector.ao.mobile.R.id.productButton);
solo.assertCurrentActivity("ProductSearch", ProductSearchActivity.class);
}
@Override
protected void tearDown() throws Exception {
solo.finishOpenedActivities();
super.tearDown();
}
}
_Original issue: http://code.google.com/p/robotium/issues/detail?id=252_ | 1.0 | junit.framework.AssertionFailedError: Button with index 2131230847 is not available! - _From [Surhum...@gmail.com](https://code.google.com/u/107563385414450670028/) on April 25, 2012 02:53:27_
What steps will reproduce the problem? 1.solo = new Solo(getInstrumentation(), getActivity());
2.solo.clickOnButton(dk.lector.ao.mobile.R.id.productButton);
3.solo.assertCurrentActivity("ProductSearch", ProductSearchActivity.class); What is the expected output? What do you see instead? I expected an Activity launched.
but got the title of this isssue instead. What version of the product are you using? On what operating system? using 3.2.1 on ubuntu 11.10. with eclipse Please provide any additional information below. My manifest does have the wiki manifest added, no change.
Here is my testclass:
package dk.lector.ao.mobile.test;
import android.test.ActivityInstrumentationTestCase2;
import com.jayway.android.robotium.solo.Solo;
import dk.lector.ao.mobile.AOinThePocket;
import dk.lector.ao.mobile.product.ProductSearchActivity;
public class TestAOinThePocket extends ActivityInstrumentationTestCase2<AOinThePocket> {
private Solo solo;
public TestAOinThePocket() {
super("dk.lector.ao.mobile", AOinThePocket.class);
}
@Override
protected void setUp() throws Exception {
super.setUp();
solo = new Solo(getInstrumentation(), getActivity());
synchronized (solo) {
solo.wait(2000);
}
}
public void testDisplayBlackBox() {
solo.clickOnButton(dk.lector.ao.mobile.R.id.productButton);
solo.assertCurrentActivity("ProductSearch", ProductSearchActivity.class);
}
@Override
protected void tearDown() throws Exception {
solo.finishOpenedActivities();
super.tearDown();
}
}
_Original issue: http://code.google.com/p/robotium/issues/detail?id=252_ | priority | junit framework assertionfailederror button with index is not available from on april what steps will reproduce the problem solo new solo getinstrumentation getactivity solo clickonbutton dk lector ao mobile r id productbutton solo assertcurrentactivity productsearch productsearchactivity class what is the expected output what do you see instead i expected an activity launched but got the title of this isssue instead what version of the product are you using on what operating system using on ubuntu with eclipse please provide any additional information below my manifest does have the wiki manifest added no change here is my testclass package dk lector ao mobile test import android test import com jayway android robotium solo solo import dk lector ao mobile aointhepocket import dk lector ao mobile product productsearchactivity public class testaointhepocket extends private solo solo public testaointhepocket super dk lector ao mobile aointhepocket class override protected void setup throws exception super setup solo new solo getinstrumentation getactivity synchronized solo solo wait public void testdisplayblackbox solo clickonbutton dk lector ao mobile r id productbutton solo assertcurrentactivity productsearch productsearchactivity class override protected void teardown throws exception solo finishopenedactivities super teardown original issue | 1 |
282,496 | 8,707,050,086 | IssuesEvent | 2018-12-06 06:06:05 | magda-io/magda | https://api.github.com/repos/magda-io/magda | closed | Indexer load queue doesn't backpressure | bug help wanted priority: medium refined | ### Problem description
When the indexer does a full reindex of the registry it's supposed to crawl through and keep up to 1000 datasets in a buffer, but crawling and ingestion should proceed at roughly the right rate because of akka backpressure.
However recent improvements to the crawl speed in the database have revealed that this doesn't actually happen - if the database crawl goes faster than the elasticsearch ingestion (usual) then by the end of a full recrawl the indexer ends up with 10s of thousands of datasets backed up, which takes up a lot of memory.
We need to fix this so that the crawl only ever gets 1000 datasets or so in front of the ingestion.
### Screenshot / Design / File reference
Log:
```
[INFO] [08/23/2018 07:30:39.713] [indexer-akka.actor.default-dispatcher-13] [RegistryCrawler(akka://indexer)] Crawled 64000 datasets from registry
[INFO] [08/23/2018 07:30:42.079] [indexer-akka.actor.default-dispatcher-16] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:30:48.762] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:30:55.056] [indexer-akka.actor.default-dispatcher-26] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:00.296] [indexer-akka.actor.default-dispatcher-29] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[WARN] [08/23/2018 07:31:05.950] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Failure when indexing ds-dga-f07db0a8-de52-48c4-a115-9d8a4eca379f: Some(BulkError(mapper_parsing_exception,failed to parse [spatial.geoJson],null,0,null))
[INFO] [08/23/2018 07:31:05.951] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Excluded dataset ds-dga-f07db0a8-de52-48c4-a115-9d8a4eca379f due to bad geojson - trying these again with spatial.geoJson excluded
[INFO] [08/23/2018 07:31:05.951] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 999 datasets
[INFO] [08/23/2018 07:31:11.543] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:15.677] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:19.584] [indexer-akka.actor.default-dispatcher-22] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:23.696] [indexer-akka.actor.default-dispatcher-32] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:30.272] [indexer-akka.actor.default-dispatcher-31] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:35.286] [indexer-akka.actor.default-dispatcher-33] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:39.789] [indexer-akka.actor.default-dispatcher-33] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:45.148] [indexer-akka.actor.default-dispatcher-34] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:50.213] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:55.496] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:00.883] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:06.324] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:11.643] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:15.995] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:20.921] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:26.195] [indexer-akka.actor.default-dispatcher-36] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:30.956] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:35.426] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:39.786] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:43.891] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:48.866] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:55.219] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:59.162] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:04.994] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:11.094] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:15.493] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:18.343] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:20.662] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:23.396] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:26.169] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:28.932] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
```
| 1.0 | Indexer load queue doesn't backpressure - ### Problem description
When the indexer does a full reindex of the registry it's supposed to crawl through and keep up to 1000 datasets in a buffer, but crawling and ingestion should proceed at roughly the right rate because of akka backpressure.
However recent improvements to the crawl speed in the database have revealed that this doesn't actually happen - if the database crawl goes faster than the elasticsearch ingestion (usual) then by the end of a full recrawl the indexer ends up with 10s of thousands of datasets backed up, which takes up a lot of memory.
We need to fix this so that the crawl only ever gets 1000 datasets or so in front of the ingestion.
### Screenshot / Design / File reference
Log:
```
[INFO] [08/23/2018 07:30:39.713] [indexer-akka.actor.default-dispatcher-13] [RegistryCrawler(akka://indexer)] Crawled 64000 datasets from registry
[INFO] [08/23/2018 07:30:42.079] [indexer-akka.actor.default-dispatcher-16] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:30:48.762] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:30:55.056] [indexer-akka.actor.default-dispatcher-26] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:00.296] [indexer-akka.actor.default-dispatcher-29] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[WARN] [08/23/2018 07:31:05.950] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Failure when indexing ds-dga-f07db0a8-de52-48c4-a115-9d8a4eca379f: Some(BulkError(mapper_parsing_exception,failed to parse [spatial.geoJson],null,0,null))
[INFO] [08/23/2018 07:31:05.951] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Excluded dataset ds-dga-f07db0a8-de52-48c4-a115-9d8a4eca379f due to bad geojson - trying these again with spatial.geoJson excluded
[INFO] [08/23/2018 07:31:05.951] [indexer-akka.actor.default-dispatcher-30] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 999 datasets
[INFO] [08/23/2018 07:31:11.543] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:15.677] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:19.584] [indexer-akka.actor.default-dispatcher-22] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:23.696] [indexer-akka.actor.default-dispatcher-32] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:30.272] [indexer-akka.actor.default-dispatcher-31] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:35.286] [indexer-akka.actor.default-dispatcher-33] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:39.789] [indexer-akka.actor.default-dispatcher-33] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:45.148] [indexer-akka.actor.default-dispatcher-34] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:50.213] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:31:55.496] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:00.883] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:06.324] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:11.643] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:15.995] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:20.921] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:26.195] [indexer-akka.actor.default-dispatcher-36] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:30.956] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:35.426] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:39.786] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:43.891] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:48.866] [indexer-akka.actor.default-dispatcher-27] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:55.219] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:32:59.162] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:04.994] [indexer-akka.actor.default-dispatcher-23] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:11.094] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:15.493] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:18.343] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:20.662] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:23.396] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:26.169] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
[INFO] [08/23/2018 07:33:28.932] [indexer-akka.actor.default-dispatcher-35] [akka.actor.ActorSystemImpl(indexer)] Successfully indexed 1000 datasets
```
| priority | indexer load queue doesn t backpressure problem description when the indexer does a full reindex of the registry it s supposed to crawl through and keep up to datasets in a buffer but crawling and ingestion should proceed at roughly the right rate because of akka backpressure however recent improvements to the crawl speed in the database have revealed that this doesn t actually happen if the database crawl goes faster than the elasticsearch ingestion usual then by the end of a full recrawl the indexer ends up with of thousands of datasets backed up which takes up a lot of memory we need to fix this so that the crawl only ever gets datasets or so in front of the ingestion screenshot design file reference log crawled datasets from registry successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets failure when indexing ds dga some bulkerror mapper parsing exception failed to parse null null excluded dataset ds dga due to bad geojson trying these again with spatial geojson excluded successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets successfully indexed datasets | 1 |
136,811 | 5,289,019,720 | IssuesEvent | 2017-02-08 16:27:14 | jeveloper/jayrock | https://api.github.com/repos/jeveloper/jayrock | closed | Add Ext.Direct support | auto-migrated Priority-Medium Type-Enhancement | ```
- What new or enhanced feature are you proposing?
Ext.Direct requires a server-side stack that fits pretty well with what Jayrock
already provides. It has a lot of support within the Ext framework for
populating stores, and generally calling services.
See the spec here: http://www.sencha.com/products/js/direct.php
- What goal would this enhancement help you achieve?
Better integration, and simplified client-side code when using Jayrock enabled
services.
```
Original issue reported on code.google.com by `westy.at...@gmail.com` on 9 Dec 2010 at 9:45
| 1.0 | Add Ext.Direct support - ```
- What new or enhanced feature are you proposing?
Ext.Direct requires a server-side stack that fits pretty well with what Jayrock
already provides. It has a lot of support within the Ext framework for
populating stores, and generally calling services.
See the spec here: http://www.sencha.com/products/js/direct.php
- What goal would this enhancement help you achieve?
Better integration, and simplified client-side code when using Jayrock enabled
services.
```
Original issue reported on code.google.com by `westy.at...@gmail.com` on 9 Dec 2010 at 9:45
| priority | add ext direct support what new or enhanced feature are you proposing ext direct requires a server side stack that fits pretty well with what jayrock already provides it has a lot of support within the ext framework for populating stores and generally calling services see the spec here what goal would this enhancement help you achieve better integration and simplified client side code when using jayrock enabled services original issue reported on code google com by westy at gmail com on dec at | 1 |
187,230 | 6,750,243,586 | IssuesEvent | 2017-10-23 03:12:26 | cdierdorff/quiz | https://api.github.com/repos/cdierdorff/quiz | closed | User deletes quiz | good first issue medium priority | A user will be able to log on and view a list of all available quizzes. The user will have the option to navigate to a page listing his/her own quizzes and delete the quiz. A dialogue asking if the user is sure will be displayed as the action cannot be undone. | 1.0 | User deletes quiz - A user will be able to log on and view a list of all available quizzes. The user will have the option to navigate to a page listing his/her own quizzes and delete the quiz. A dialogue asking if the user is sure will be displayed as the action cannot be undone. | priority | user deletes quiz a user will be able to log on and view a list of all available quizzes the user will have the option to navigate to a page listing his her own quizzes and delete the quiz a dialogue asking if the user is sure will be displayed as the action cannot be undone | 1 |
174,051 | 6,536,038,654 | IssuesEvent | 2017-08-31 16:34:06 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | Share link title in linkedin share is "GeoSolutions" and is not overridable | bug Priority: Medium Project: C040 | Share buttons trigger use the title "GeoSolutions" for the shared resources.
The share title is used only by Linkedin, as far as I know. This causes the link you are sharing on Linkedin contains the "GeoSolutions" as title.
We should get the title of the current page or map.
| 1.0 | Share link title in linkedin share is "GeoSolutions" and is not overridable - Share buttons trigger use the title "GeoSolutions" for the shared resources.
The share title is used only by Linkedin, as far as I know. This causes the link you are sharing on Linkedin contains the "GeoSolutions" as title.
We should get the title of the current page or map.
| priority | share link title in linkedin share is geosolutions and is not overridable share buttons trigger use the title geosolutions for the shared resources the share title is used only by linkedin as far as i know this causes the link you are sharing on linkedin contains the geosolutions as title we should get the title of the current page or map | 1 |
306,034 | 9,379,806,044 | IssuesEvent | 2019-04-04 15:39:04 | CanadianClimateDataPortal/Canadian-Climate-Data-Portal | https://api.github.com/repos/CanadianClimateDataPortal/Canadian-Climate-Data-Portal | opened | IDF curves - Bulk download | 3- Portal F.E- Download Priority - Medium WG- Data | - [ ] Possibility to download multiple or all data on a single station
- [ ] Possibility to download multiple station at the same time by selecting them
Possibility to go through the download section of the portal instead of the vizual mapping representation | 1.0 | IDF curves - Bulk download - - [ ] Possibility to download multiple or all data on a single station
- [ ] Possibility to download multiple station at the same time by selecting them
Possibility to go through the download section of the portal instead of the vizual mapping representation | priority | idf curves bulk download possibility to download multiple or all data on a single station possibility to download multiple station at the same time by selecting them possibility to go through the download section of the portal instead of the vizual mapping representation | 1 |
505,789 | 14,645,638,514 | IssuesEvent | 2020-12-26 09:08:56 | schemathesis/schemathesis | https://api.github.com/repos/schemathesis/schemathesis | closed | [BUG] Unsatisfiable errors in explicit example breaks the CLI execution | Difficulty: Medium Priority: Low Type: Bug | **Describe the bug**
If the schema provides examples that lead to generating data for unsatisfiable schemas, then the execution process stops.
Reported in gitter - I'll update this issue once I have more info
| 1.0 | [BUG] Unsatisfiable errors in explicit example breaks the CLI execution - **Describe the bug**
If the schema provides examples that lead to generating data for unsatisfiable schemas, then the execution process stops.
Reported in gitter - I'll update this issue once I have more info
| priority | unsatisfiable errors in explicit example breaks the cli execution describe the bug if the schema provides examples that lead to generating data for unsatisfiable schemas then the execution process stops reported in gitter i ll update this issue once i have more info | 1 |
708,897 | 24,359,954,653 | IssuesEvent | 2022-10-03 10:47:47 | CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | closed | [Collaboration UI] Display Question Bounds | Module/Front-End Status/Medium-Priority Type/Feature | ## Description
The UI should show the bounds of the question's input variable(s)
## Parent Task
- #104 | 1.0 | [Collaboration UI] Display Question Bounds - ## Description
The UI should show the bounds of the question's input variable(s)
## Parent Task
- #104 | priority | display question bounds description the ui should show the bounds of the question s input variable s parent task | 1 |
404,928 | 11,864,652,324 | IssuesEvent | 2020-03-25 22:09:33 | cds-snc/report-a-cybercrime | https://api.github.com/repos/cds-snc/report-a-cybercrime | closed | Tweak "They asked for" lines in email | medium priority | ## Summary
From Trello https://trello.com/c/gvgdJKYX
> The best I can tell, it looks like the victim selected personal info was affected on the form > What information did they ask for? SIN [Report reads: They asked for: sin] > What information did they obtain? nil [Report reads: They asked for: undefined].
>
>Confusing to Analyst because:
>- the field looks like a duplicate on the report
>- field multiplies for however many ways they selected being "affected"
>- interrupts the flow
## Design detail
Make it more explicit what those lines are. Combine the line for the "other" field with the line for the joined list.
| 1.0 | Tweak "They asked for" lines in email - ## Summary
From Trello https://trello.com/c/gvgdJKYX
> The best I can tell, it looks like the victim selected personal info was affected on the form > What information did they ask for? SIN [Report reads: They asked for: sin] > What information did they obtain? nil [Report reads: They asked for: undefined].
>
>Confusing to Analyst because:
>- the field looks like a duplicate on the report
>- field multiplies for however many ways they selected being "affected"
>- interrupts the flow
## Design detail
Make it more explicit what those lines are. Combine the line for the "other" field with the line for the joined list.
| priority | tweak they asked for lines in email summary from trello the best i can tell it looks like the victim selected personal info was affected on the form what information did they ask for sin what information did they obtain nil confusing to analyst because the field looks like a duplicate on the report field multiplies for however many ways they selected being affected interrupts the flow design detail make it more explicit what those lines are combine the line for the other field with the line for the joined list | 1 |
795,966 | 28,094,325,935 | IssuesEvent | 2023-03-30 14:52:05 | KDT3-Final-6/final-project-BE | https://api.github.com/repos/KDT3-Final-6/final-project-BE | closed | feat: 컨트롤러에서 토큰을 받게 수정 | Type: Feature Status: In Progress Priority: Medium For: API For: Backend | ## Description
컨트롤러에서 토큰을 받게 수정
## Tasks(Process)
- [x] 컨트롤러 파라미터 수정
## References
| 1.0 | feat: 컨트롤러에서 토큰을 받게 수정 - ## Description
컨트롤러에서 토큰을 받게 수정
## Tasks(Process)
- [x] 컨트롤러 파라미터 수정
## References
| priority | feat 컨트롤러에서 토큰을 받게 수정 description 컨트롤러에서 토큰을 받게 수정 tasks process 컨트롤러 파라미터 수정 references | 1 |
134,057 | 5,219,503,501 | IssuesEvent | 2017-01-26 19:16:13 | ualbertalib/HydraNorth | https://api.github.com/repos/ualbertalib/HydraNorth | closed | Create a good four-digit date field | hydranorth2 priority:medium size:TBD team:metadata | Replace the current year_created field with a well-though-out four-digit date field for use in faceting in the public interface and for reporting publication year in DOI and ARK. It should work well with all of our content types and sources (Dataverse etc.)
| 1.0 | Create a good four-digit date field - Replace the current year_created field with a well-though-out four-digit date field for use in faceting in the public interface and for reporting publication year in DOI and ARK. It should work well with all of our content types and sources (Dataverse etc.)
| priority | create a good four digit date field replace the current year created field with a well though out four digit date field for use in faceting in the public interface and for reporting publication year in doi and ark it should work well with all of our content types and sources dataverse etc | 1 |
157,191 | 5,996,385,255 | IssuesEvent | 2017-06-03 13:50:28 | Rsl1122/Plan-PlayerAnalytics | https://api.github.com/repos/Rsl1122/Plan-PlayerAnalytics | opened | ConcurrentModificationException during Command use save | Bug Priority: MEDIUM | **Plan Version:** 3.3.0
**Server Version:** 1.8.0
**Database Type:** Unknown
**Command Causing Issue:** None
**Description:**
ConcurrentModificationException during Command use save
(Command was issued during save)
**Steps to Reproduce:**
1. Issue command while commanduse is being saved
**Proposed Solution:**
One line fix:
Copy map on save
**Stack Trace - Console/Errors.txt contents:**
```
[12:16:05] [Craft Scheduler Thread - 2372/WARN]: [Plan] Plugin Plan v3.3.0 generated an exception while executing task 1226
java.util.ConcurrentModificationException
at java.util.HashMap$HashIterator.nextNode(HashMap.java:1437) ~[?:1.8.0_131]
at java.util.HashMap$KeyIterator.next(HashMap.java:1461) ~[?:1.8.0_131]
at main.java.com.djrapitops.plan.database.tables.CommandUseTable.saveCommandUse(CommandUseTable.java:95) ~[?:?]
at main.java.com.djrapitops.plan.database.Database.saveCommandUse(Database.java:199) ~[?:?]
at main.java.com.djrapitops.plan.data.cache.DataCacheHandler.saveCommandUse(DataCacheHandler.java:341) ~[?:?]
at main.java.com.djrapitops.plan.data.cache.DataCacheHandler$1.run(DataCacheHandler.java:151) ~[?:?]
at org.bukkit.craftbukkit.v1_11_R1.scheduler.CraftTask.run(CraftTask.java:58) ~[patched_1.11.2.jar:git-Paper-1104]
at org.bukkit.craftbukkit.v1_11_R1.scheduler.CraftAsyncTask.run(CraftAsyncTask.java:52) [patched_1.11.2.jar:git-Paper-1104]
at com.destroystokyo.paper.ServerSchedulerReportingWrapper.run(ServerSchedulerReportingWrapper.java:22) [patched_1.11.2.jar:git-Paper-1104]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_131]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_131]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_131]
```
| 1.0 | ConcurrentModificationException during Command use save - **Plan Version:** 3.3.0
**Server Version:** 1.8.0
**Database Type:** Unknown
**Command Causing Issue:** None
**Description:**
ConcurrentModificationException during Command use save
(Command was issued during save)
**Steps to Reproduce:**
1. Issue command while commanduse is being saved
**Proposed Solution:**
One line fix:
Copy map on save
**Stack Trace - Console/Errors.txt contents:**
```
[12:16:05] [Craft Scheduler Thread - 2372/WARN]: [Plan] Plugin Plan v3.3.0 generated an exception while executing task 1226
java.util.ConcurrentModificationException
at java.util.HashMap$HashIterator.nextNode(HashMap.java:1437) ~[?:1.8.0_131]
at java.util.HashMap$KeyIterator.next(HashMap.java:1461) ~[?:1.8.0_131]
at main.java.com.djrapitops.plan.database.tables.CommandUseTable.saveCommandUse(CommandUseTable.java:95) ~[?:?]
at main.java.com.djrapitops.plan.database.Database.saveCommandUse(Database.java:199) ~[?:?]
at main.java.com.djrapitops.plan.data.cache.DataCacheHandler.saveCommandUse(DataCacheHandler.java:341) ~[?:?]
at main.java.com.djrapitops.plan.data.cache.DataCacheHandler$1.run(DataCacheHandler.java:151) ~[?:?]
at org.bukkit.craftbukkit.v1_11_R1.scheduler.CraftTask.run(CraftTask.java:58) ~[patched_1.11.2.jar:git-Paper-1104]
at org.bukkit.craftbukkit.v1_11_R1.scheduler.CraftAsyncTask.run(CraftAsyncTask.java:52) [patched_1.11.2.jar:git-Paper-1104]
at com.destroystokyo.paper.ServerSchedulerReportingWrapper.run(ServerSchedulerReportingWrapper.java:22) [patched_1.11.2.jar:git-Paper-1104]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_131]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_131]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_131]
```
| priority | concurrentmodificationexception during command use save plan version server version database type unknown command causing issue none description concurrentmodificationexception during command use save command was issued during save steps to reproduce issue command while commanduse is being saved proposed solution one line fix copy map on save stack trace console errors txt contents plugin plan generated an exception while executing task java util concurrentmodificationexception at java util hashmap hashiterator nextnode hashmap java at java util hashmap keyiterator next hashmap java at main java com djrapitops plan database tables commandusetable savecommanduse commandusetable java at main java com djrapitops plan database database savecommanduse database java at main java com djrapitops plan data cache datacachehandler savecommanduse datacachehandler java at main java com djrapitops plan data cache datacachehandler run datacachehandler java at org bukkit craftbukkit scheduler crafttask run crafttask java at org bukkit craftbukkit scheduler craftasynctask run craftasynctask java at com destroystokyo paper serverschedulerreportingwrapper run serverschedulerreportingwrapper java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java | 1 |
375,528 | 11,104,736,807 | IssuesEvent | 2019-12-17 08:20:59 | OpenSourceEconomics/soepy | https://api.github.com/repos/OpenSourceEconomics/soepy | opened | update numerical integration | pb package priority medium size medium | We made a lot of progress with improving the quality of numerical integration in the `construct_emax` over the last months. This needs to be incorporated in this package as well. | 1.0 | update numerical integration - We made a lot of progress with improving the quality of numerical integration in the `construct_emax` over the last months. This needs to be incorporated in this package as well. | priority | update numerical integration we made a lot of progress with improving the quality of numerical integration in the construct emax over the last months this needs to be incorporated in this package as well | 1 |
786,230 | 27,639,861,087 | IssuesEvent | 2023-03-10 17:08:01 | KinsonDigital/BranchValidator | https://api.github.com/repos/KinsonDigital/BranchValidator | closed | 🚧Setup dependabot | medium priority preview cicd | ### Complete The Item Below
- [X] I have updated the title without removing the 🚧 emoji.
### Description
Setup [dependabot](https://docs.github.com/en/code-security/dependabot) for the project.
Refer to the **Velaptor** project for the how to setup the **YAML** config tile.
The name of the file is [dependabot.yml](https://github.com/KinsonDigital/Velaptor/blob/e228cf09e27bdc45c23806ed0e80a54c94f3ec76/.github/dependabot.yml#L4)
### Acceptance Criteria
- [x] Dependabot setup for project
### ToDo Items
- [X] Change type labels added to this issue. Refer to the _**Change Type Labels**_ section below.
- [X] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below.
- [X] Issue linked to the correct project _(if applicable)_.
- [x] Issue linked to the correct milestone _(if applicable)_.
- [x] Draft pull request created and linked to this issue _(only required with code changes)_.
### Issue Dependencies
- #65
> **Note** This is because if this issue is completed first, it will start the bot and all of the NuGet package updates will be detected and a lot of PR's will be created. The initial update of the NuGet packages should be done first to help save work.
### Related Work
_No response_
### Additional Information:
**_<details closed><summary>Change Type Labels</summary>_**
| Change Type | Label |
|---------------------|----------------------|
| Bug Fixes | `🐛bug` |
| Breaking Changes | `🧨breaking changes` |
| New Feature | `✨new feature` |
| Workflow Changes | `workflow` |
| Code Doc Changes | `🗒️documentation/code` |
| Product Doc Changes | `📝documentation/product` |
</details>
**_<details closed><summary>Priority Type Labels</summary>_**
| Priority Type | Label |
|---------------------|-------------------|
| Low Priority | `low priority` |
| Medium Priority | `medium priority` |
| High Priority | `high priority` |
</details>
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct. | 1.0 | 🚧Setup dependabot - ### Complete The Item Below
- [X] I have updated the title without removing the 🚧 emoji.
### Description
Setup [dependabot](https://docs.github.com/en/code-security/dependabot) for the project.
Refer to the **Velaptor** project for the how to setup the **YAML** config tile.
The name of the file is [dependabot.yml](https://github.com/KinsonDigital/Velaptor/blob/e228cf09e27bdc45c23806ed0e80a54c94f3ec76/.github/dependabot.yml#L4)
### Acceptance Criteria
- [x] Dependabot setup for project
### ToDo Items
- [X] Change type labels added to this issue. Refer to the _**Change Type Labels**_ section below.
- [X] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below.
- [X] Issue linked to the correct project _(if applicable)_.
- [x] Issue linked to the correct milestone _(if applicable)_.
- [x] Draft pull request created and linked to this issue _(only required with code changes)_.
### Issue Dependencies
- #65
> **Note** This is because if this issue is completed first, it will start the bot and all of the NuGet package updates will be detected and a lot of PR's will be created. The initial update of the NuGet packages should be done first to help save work.
### Related Work
_No response_
### Additional Information:
**_<details closed><summary>Change Type Labels</summary>_**
| Change Type | Label |
|---------------------|----------------------|
| Bug Fixes | `🐛bug` |
| Breaking Changes | `🧨breaking changes` |
| New Feature | `✨new feature` |
| Workflow Changes | `workflow` |
| Code Doc Changes | `🗒️documentation/code` |
| Product Doc Changes | `📝documentation/product` |
</details>
**_<details closed><summary>Priority Type Labels</summary>_**
| Priority Type | Label |
|---------------------|-------------------|
| Low Priority | `low priority` |
| Medium Priority | `medium priority` |
| High Priority | `high priority` |
</details>
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct. | priority | 🚧setup dependabot complete the item below i have updated the title without removing the 🚧 emoji description setup for the project refer to the velaptor project for the how to setup the yaml config tile the name of the file is acceptance criteria dependabot setup for project todo items change type labels added to this issue refer to the change type labels section below priority label added to this issue refer to the priority type labels section below issue linked to the correct project if applicable issue linked to the correct milestone if applicable draft pull request created and linked to this issue only required with code changes issue dependencies note this is because if this issue is completed first it will start the bot and all of the nuget package updates will be detected and a lot of pr s will be created the initial update of the nuget packages should be done first to help save work related work no response additional information change type labels change type label bug fixes 🐛bug breaking changes 🧨breaking changes new feature ✨new feature workflow changes workflow code doc changes 🗒️documentation code product doc changes 📝documentation product priority type labels priority type label low priority low priority medium priority medium priority high priority high priority code of conduct i agree to follow this project s code of conduct | 1 |
787,421 | 27,717,305,407 | IssuesEvent | 2023-03-14 17:48:13 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | [TT-1955] UseCertificate field is used even if AuthToken is disable | bug Priority: Medium zendesk | **Branch/Environment/Version**
v2.9.3
**Describe the bug**
UseCertificate field is checked even if AuthToken is disabled
**Reproduction steps**
Steps to reproduce the behavior:
1. Create API with Auth token authentication mode and enable client certificate field.
2. Create Key for API
3. Update API Authentication Mode from AuthToken to MutualTLS
4. Send the request to the API using key created in Step 2.
**Actual behavior**
Request returns back correct response
**Expected behavior**
Request should fail with `Authorization field missing` error
| 1.0 | [TT-1955] UseCertificate field is used even if AuthToken is disable - **Branch/Environment/Version**
v2.9.3
**Describe the bug**
UseCertificate field is checked even if AuthToken is disabled
**Reproduction steps**
Steps to reproduce the behavior:
1. Create API with Auth token authentication mode and enable client certificate field.
2. Create Key for API
3. Update API Authentication Mode from AuthToken to MutualTLS
4. Send the request to the API using key created in Step 2.
**Actual behavior**
Request returns back correct response
**Expected behavior**
Request should fail with `Authorization field missing` error
| priority | usecertificate field is used even if authtoken is disable branch environment version describe the bug usecertificate field is checked even if authtoken is disabled reproduction steps steps to reproduce the behavior create api with auth token authentication mode and enable client certificate field create key for api update api authentication mode from authtoken to mutualtls send the request to the api using key created in step actual behavior request returns back correct response expected behavior request should fail with authorization field missing error | 1 |
257,970 | 8,149,305,705 | IssuesEvent | 2018-08-22 09:12:58 | Xceptance/neodymium-library | https://api.github.com/repos/Xceptance/neodymium-library | closed | Validate the value attribute using a regex | Medium Priority doneInDevelop feature | Add a condition that can validate the value attribute using a regex.
The Condition should be located in the SelenideAddons class. | 1.0 | Validate the value attribute using a regex - Add a condition that can validate the value attribute using a regex.
The Condition should be located in the SelenideAddons class. | priority | validate the value attribute using a regex add a condition that can validate the value attribute using a regex the condition should be located in the selenideaddons class | 1 |
462,456 | 13,247,719,078 | IssuesEvent | 2020-08-19 17:43:51 | radical-cybertools/radical.entk | https://api.github.com/repos/radical-cybertools/radical.entk | closed | unexpected keyboardinterrupt messages | layer:saga priority:medium topic:execution topic:resource topic:termination type:bug | My jobs gives some keyboardinterrupt messages after it runs as expected and finishes once walltime runs out. I didn't keyboard interrupt, it automatically quits due to walltime run out. This error is just cosmetic, doesn't affect me otherwise.
```
+ re.session.login1.eh22.018310.0003 (json)
+ pilot.0000 (profiles)
+ pilot.0000 (logfiles)
session lifetime: 718.7s ok
All components terminated
Traceback (most recent call last):
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 431, in run
self.terminate()
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 458, in terminate
self._prof.prof('term_start', uid=self._uid)
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/utils/profile.py", line 276, in prof
self._handle.write(data)
KeyboardInterrupt
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "extasy.py", line 275, in <module>
appman.run()
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 439, in run
raise KeyboardInterrupt
KeyboardInterrupt
```
```
radical-stack
python : 3.7.6
pythonpath : /sw/summit/xalt/1.2.0/site:/sw/summit/xalt/1.2.0/libexec
virtualenv : vamp11
radical.analytics : 0.90.7-v0.72.0-38-g14b9581@devel
radical.entk : 1.0.1-v1.0.1-2-g92797c9@devel
radical.pilot : 1.1.1-v1.1.1-2-g8385a7d@devel
radical.saga : 1.1.0-v1.1@devel
radical.utils : 1.1.1-v1.1.1-2-gfe6c424@devel
```
| 1.0 | unexpected keyboardinterrupt messages - My jobs gives some keyboardinterrupt messages after it runs as expected and finishes once walltime runs out. I didn't keyboard interrupt, it automatically quits due to walltime run out. This error is just cosmetic, doesn't affect me otherwise.
```
+ re.session.login1.eh22.018310.0003 (json)
+ pilot.0000 (profiles)
+ pilot.0000 (logfiles)
session lifetime: 718.7s ok
All components terminated
Traceback (most recent call last):
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 431, in run
self.terminate()
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 458, in terminate
self._prof.prof('term_start', uid=self._uid)
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/utils/profile.py", line 276, in prof
self._handle.write(data)
KeyboardInterrupt
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "extasy.py", line 275, in <module>
appman.run()
File "/ccs/home/eh22/.conda/envs/vamp11/lib/python3.7/site-packages/radical/entk/appman/appmanager.py", line 439, in run
raise KeyboardInterrupt
KeyboardInterrupt
```
```
radical-stack
python : 3.7.6
pythonpath : /sw/summit/xalt/1.2.0/site:/sw/summit/xalt/1.2.0/libexec
virtualenv : vamp11
radical.analytics : 0.90.7-v0.72.0-38-g14b9581@devel
radical.entk : 1.0.1-v1.0.1-2-g92797c9@devel
radical.pilot : 1.1.1-v1.1.1-2-g8385a7d@devel
radical.saga : 1.1.0-v1.1@devel
radical.utils : 1.1.1-v1.1.1-2-gfe6c424@devel
```
| priority | unexpected keyboardinterrupt messages my jobs gives some keyboardinterrupt messages after it runs as expected and finishes once walltime runs out i didn t keyboard interrupt it automatically quits due to walltime run out this error is just cosmetic doesn t affect me otherwise re session json pilot profiles pilot logfiles session lifetime ok all components terminated traceback most recent call last file ccs home conda envs lib site packages radical entk appman appmanager py line in run self terminate file ccs home conda envs lib site packages radical entk appman appmanager py line in terminate self prof prof term start uid self uid file ccs home conda envs lib site packages radical utils profile py line in prof self handle write data keyboardinterrupt during handling of the above exception another exception occurred traceback most recent call last file extasy py line in appman run file ccs home conda envs lib site packages radical entk appman appmanager py line in run raise keyboardinterrupt keyboardinterrupt radical stack python pythonpath sw summit xalt site sw summit xalt libexec virtualenv radical analytics devel radical entk devel radical pilot devel radical saga devel radical utils devel | 1 |
654,881 | 21,672,800,819 | IssuesEvent | 2022-05-08 08:16:00 | uriahf/rtichoke | https://api.github.com/repos/uriahf/rtichoke | opened | Fix range for decision curves | Difficulty: novice :suspect: Priority: medium | Limits for the y axis should be according to the prevalence with constant margins. | 1.0 | Fix range for decision curves - Limits for the y axis should be according to the prevalence with constant margins. | priority | fix range for decision curves limits for the y axis should be according to the prevalence with constant margins | 1 |
490,477 | 14,134,989,264 | IssuesEvent | 2020-11-10 00:35:31 | dnnsoftware/Dnn.Platform | https://api.github.com/repos/dnnsoftware/Dnn.Platform | closed | Export data tier application in Azure not possible (DNN database) | Area: Platform > Providers Effort: Medium Priority: Medium Status: Ready for Development Type: Maintenance stale | ## Current result
Try to make backup of the DNN database created as Azure SQL database.
1. In SQL Server Management Studio click on Export data tier application (generate backup, BACPAK file).
## Expected result
BACPAK generation can be done successfully.
## Error log
TITLE: Microsoft SQL Server Management Studio
------------------------------
One or more unsupported elements were found in the schema used as part of a data package.
Error SQL46010: Procedure: [dbo].[AddTab] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateTab] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[DeleteUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[DeleteUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddSiteLog] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrls] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[SaveTabUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateUrlTrackingStats] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetTabUrls] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateServer] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[AddPackage] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_UpdateLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_GetLinks] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[UpdatePackage] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_GetLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[dnnLinks_AddLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: View: [dbo].[vw_Tabs] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
(Microsoft.SqlServer.Dac)
## Additional context
The issues can be fixed in the stored procedures and views by putting "url" reference column into brackets => [url]
## Affected version
9.3.2 | 1.0 | Export data tier application in Azure not possible (DNN database) - ## Current result
Try to make backup of the DNN database created as Azure SQL database.
1. In SQL Server Management Studio click on Export data tier application (generate backup, BACPAK file).
## Expected result
BACPAK generation can be done successfully.
## Error log
TITLE: Microsoft SQL Server Management Studio
------------------------------
One or more unsupported elements were found in the schema used as part of a data package.
Error SQL46010: Procedure: [dbo].[AddTab] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateTab] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[DeleteUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[DeleteUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddSiteLog] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrls] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[SaveTabUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateUrlTrackingStats] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateUrlTracking] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[AddUrl] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[GetTabUrls] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[UpdateServer] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[AddPackage] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_UpdateLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_GetLinks] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[UpdatePackage] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: Procedure: [dbo].[dnnLinks_GetLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near URL.
Error SQL46010: Procedure: [dbo].[dnnLinks_AddLink] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
Error SQL46010: View: [dbo].[vw_Tabs] contains a statement that is not supported on Microsoft Azure SQL Database v12. The specific error is: Incorrect syntax near Url.
(Microsoft.SqlServer.Dac)
## Additional context
The issues can be fixed in the stored procedures and views by putting "url" reference column into brackets => [url]
## Affected version
9.3.2 | priority | export data tier application in azure not possible dnn database current result try to make backup of the dnn database created as azure sql database in sql server management studio click on export data tier application generate backup bacpak file expected result bacpak generation can be done successfully error log title microsoft sql server management studio one or more unsupported elements were found in the schema used as part of a data package error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error procedure contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url error view contains a statement that is not supported on microsoft azure sql database the specific error is incorrect syntax near url microsoft sqlserver dac additional context the issues can be fixed in the stored procedures and views by putting url reference column into brackets affected version | 1 |
258,925 | 8,180,901,512 | IssuesEvent | 2018-08-28 20:57:32 | VulcanForge/pvp-mode | https://api.github.com/repos/VulcanForge/pvp-mode | opened | Add the transformer exclusions annotation to the coremod | cleanup medium priority | Exclude:
* The coremod package
* The API packages | 1.0 | Add the transformer exclusions annotation to the coremod - Exclude:
* The coremod package
* The API packages | priority | add the transformer exclusions annotation to the coremod exclude the coremod package the api packages | 1 |
26,502 | 2,684,633,791 | IssuesEvent | 2015-03-29 06:08:12 | gtcasl/gpuocelot | https://api.github.com/repos/gtcasl/gpuocelot | opened | Error while installing in Arch Linux | bug imported Priority-Medium | _From [zaidho...@gmail.com](https://code.google.com/u/101209106559517423416/) on September 13, 2014 18:27:08_
What steps will reproduce the problem? After fixing the ptxgrammer issue I am getting an error with I think the boost libraries when I run the ./build.py --install command. I get Build Error. What is the expected output? What do you see instead? I am getting the error shown in the attached file. It seems to be a problem with this command:
g++ -o .release_build/ocelot/analysis/implementation/ControlTree.os -c -O2 -Wall -Werror -std=c++0x -D_GNU_SOURCE -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -fPIC -I. -I/usr/include -I/usr/include -I.release_build -I. ocelot/analysis/implementation/ControlTree.cpp What version of the product are you using? On what operating system? Latest version of each of the dependencies, and the latest version of gpuocelot. I am running Arch Linux Please provide any additional information below.
**Attachment:** [error.txt](http://code.google.com/p/gpuocelot/issues/detail?id=97)
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=97_ | 1.0 | Error while installing in Arch Linux - _From [zaidho...@gmail.com](https://code.google.com/u/101209106559517423416/) on September 13, 2014 18:27:08_
What steps will reproduce the problem? After fixing the ptxgrammer issue I am getting an error with I think the boost libraries when I run the ./build.py --install command. I get Build Error. What is the expected output? What do you see instead? I am getting the error shown in the attached file. It seems to be a problem with this command:
g++ -o .release_build/ocelot/analysis/implementation/ControlTree.os -c -O2 -Wall -Werror -std=c++0x -D_GNU_SOURCE -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -fPIC -I. -I/usr/include -I/usr/include -I.release_build -I. ocelot/analysis/implementation/ControlTree.cpp What version of the product are you using? On what operating system? Latest version of each of the dependencies, and the latest version of gpuocelot. I am running Arch Linux Please provide any additional information below.
**Attachment:** [error.txt](http://code.google.com/p/gpuocelot/issues/detail?id=97)
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=97_ | priority | error while installing in arch linux from on september what steps will reproduce the problem after fixing the ptxgrammer issue i am getting an error with i think the boost libraries when i run the build py install command i get build error what is the expected output what do you see instead i am getting the error shown in the attached file it seems to be a problem with this command g o release build ocelot analysis implementation controltree os c wall werror std c d gnu source d stdc constant macros d stdc format macros d stdc limit macros fpic i i usr include i usr include i release build i ocelot analysis implementation controltree cpp what version of the product are you using on what operating system latest version of each of the dependencies and the latest version of gpuocelot i am running arch linux please provide any additional information below attachment original issue | 1 |
328,411 | 9,994,475,692 | IssuesEvent | 2019-07-11 17:46:11 | minio/mc | https://api.github.com/repos/minio/mc | closed | do not edit user configuration files without notifying user/confirmation | community priority: medium | ## Expected behavior
don't edit user configuration files without explicit confirmation
## Actual behavior
adds
```
autoload -U +X bashcompinit && bashcompinit
complete -o nospace -C /home/atomi/go/bin/mc mc
```
to `.zshrc`
## Steps to reproduce the behavior
run mc
## mc version
- (paste output of `mc version`)
## System information
guys this is no bueno.
| 1.0 | do not edit user configuration files without notifying user/confirmation - ## Expected behavior
don't edit user configuration files without explicit confirmation
## Actual behavior
adds
```
autoload -U +X bashcompinit && bashcompinit
complete -o nospace -C /home/atomi/go/bin/mc mc
```
to `.zshrc`
## Steps to reproduce the behavior
run mc
## mc version
- (paste output of `mc version`)
## System information
guys this is no bueno.
| priority | do not edit user configuration files without notifying user confirmation expected behavior don t edit user configuration files without explicit confirmation actual behavior adds autoload u x bashcompinit bashcompinit complete o nospace c home atomi go bin mc mc to zshrc steps to reproduce the behavior run mc mc version paste output of mc version system information guys this is no bueno | 1 |
92,174 | 3,868,680,929 | IssuesEvent | 2016-04-10 03:54:59 | HubTurbo/HubTurbo | https://api.github.com/repos/HubTurbo/HubTurbo | closed | User can add issues to a 'watch list' | feature-filters feature-panels priority.medium type.story | ... so that users can maintain a 'watch list' of selected issues.
User should be able to right click on an issue card and choose `add to new watch list`. This will create a new panel with the corresponding `id:repo#id` filter. User will be prompted to give a name for the panel.
After that, user should be able to add more issues to the that panel by right-clicking on issues and choosing `add to watch list` --> `panel name`
| 1.0 | User can add issues to a 'watch list' - ... so that users can maintain a 'watch list' of selected issues.
User should be able to right click on an issue card and choose `add to new watch list`. This will create a new panel with the corresponding `id:repo#id` filter. User will be prompted to give a name for the panel.
After that, user should be able to add more issues to the that panel by right-clicking on issues and choosing `add to watch list` --> `panel name`
| priority | user can add issues to a watch list so that users can maintain a watch list of selected issues user should be able to right click on an issue card and choose add to new watch list this will create a new panel with the corresponding id repo id filter user will be prompted to give a name for the panel after that user should be able to add more issues to the that panel by right clicking on issues and choosing add to watch list panel name | 1 |
198,691 | 6,975,786,474 | IssuesEvent | 2017-12-12 08:43:55 | webpack/webpack-cli | https://api.github.com/repos/webpack/webpack-cli | opened | [Init] Add keyboard functionality for interpretor lists | CLI Good First Contribution Priority: Medium | <!-- Before creating an issue please make sure you are using the latest version of webpack. -->
**Do you want to request a *feature* or report a *bug*?**
<!-- Please ask questions on StackOverflow or the webpack Gitter (https://gitter.im/webpack/webpack). Questions will be closed. -->
Bug (Usability)
**What is the current behavior?**
Currently when you get a multiple choice list from the init generator, there is no way to key through the options with arrow keys. (Technically with interpreter this is possible [cant remember if we use that])
**If the current behavior is a bug, please provide the steps to reproduce.**
<!-- A great way to do this is to provide your configuration via a GitHub gist. -->
**What is the expected behavior?**
Be able to key up and down through lists in the cli generators with arrow keys.
**If this is a feature request, what is motivation or use case for changing the behavior?**
Usability!
webpack-cli@latest
| 1.0 | [Init] Add keyboard functionality for interpretor lists - <!-- Before creating an issue please make sure you are using the latest version of webpack. -->
**Do you want to request a *feature* or report a *bug*?**
<!-- Please ask questions on StackOverflow or the webpack Gitter (https://gitter.im/webpack/webpack). Questions will be closed. -->
Bug (Usability)
**What is the current behavior?**
Currently when you get a multiple choice list from the init generator, there is no way to key through the options with arrow keys. (Technically with interpreter this is possible [cant remember if we use that])
**If the current behavior is a bug, please provide the steps to reproduce.**
<!-- A great way to do this is to provide your configuration via a GitHub gist. -->
**What is the expected behavior?**
Be able to key up and down through lists in the cli generators with arrow keys.
**If this is a feature request, what is motivation or use case for changing the behavior?**
Usability!
webpack-cli@latest
| priority | add keyboard functionality for interpretor lists do you want to request a feature or report a bug bug usability what is the current behavior currently when you get a multiple choice list from the init generator there is no way to key through the options with arrow keys technically with interpreter this is possible if the current behavior is a bug please provide the steps to reproduce what is the expected behavior be able to key up and down through lists in the cli generators with arrow keys if this is a feature request what is motivation or use case for changing the behavior usability webpack cli latest | 1 |
350,097 | 10,478,446,362 | IssuesEvent | 2019-09-24 00:04:24 | BCcampus/edehr | https://api.github.com/repos/BCcampus/edehr | closed | Add "back to assignments" link | Effort - Low Priority - Medium ~Feature | Give users an easier way to get back to where they came from.
Please add a `< Back to assignments` link to the right of the "new assignments button | 1.0 | Add "back to assignments" link - Give users an easier way to get back to where they came from.
Please add a `< Back to assignments` link to the right of the "new assignments button | priority | add back to assignments link give users an easier way to get back to where they came from please add a back to assignments link to the right of the new assignments button | 1 |
328,901 | 10,001,235,175 | IssuesEvent | 2019-07-12 15:09:50 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | purchasing quest scrolls / bundles should have confirmation popup | good first issue priority: medium section: Quest Shop status: issue: in progress | Quest scroll purchases don't have a confirmation step.
When you buy other gem-purchasable items such as hatching potions, you see a confirmation dialog that tells you how many gems the total purchase will be and asks you to confirm:

The same thing should happen when you buy quest scrolls or quest scroll bundles. | 1.0 | purchasing quest scrolls / bundles should have confirmation popup - Quest scroll purchases don't have a confirmation step.
When you buy other gem-purchasable items such as hatching potions, you see a confirmation dialog that tells you how many gems the total purchase will be and asks you to confirm:

The same thing should happen when you buy quest scrolls or quest scroll bundles. | priority | purchasing quest scrolls bundles should have confirmation popup quest scroll purchases don t have a confirmation step when you buy other gem purchasable items such as hatching potions you see a confirmation dialog that tells you how many gems the total purchase will be and asks you to confirm the same thing should happen when you buy quest scrolls or quest scroll bundles | 1 |
184,220 | 6,710,957,318 | IssuesEvent | 2017-10-13 00:29:10 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | opened | Minor updates to Enketo UI | Enketo Priority: 2 - Medium UI/UX | This issue updates the UI for Enketo forms on mobile and desktop. These are small tweaks for now and we hope to come back and make additional changes in the future. [Design spec](https://docs.google.com/document/d/1bnbQ17I9fq1ZRbu2v9otaRpUXn_7eODe7YTYzsXQoDg/edit).
**Update form header on desktop only**:
- [ ] form title is 20px bold, in black
- [ ] title is left-aligned, 20px from the left edge of the form card
- [ ] remove the "Cancel" button and replace it with an X on the right side of the header (vertically centered)
**Update form contents**:
- [ ] Put at 20px buffer on the left and right sides of the form card
- [ ] Form group titles should be blue, 18px, bold
- [ ] If there are nested groups, the sub-title (nested group name) should be black, 18px, normal weight
- [ ] 30px spacing between the group title and the first question OR between the group title and group subtitle (if there is a subtitle, which will be there if there are nested groups)
- [ ] If there is a subtitle, 20px between the subtitle and the first question
- [ ] 30px between answer options/user inputs and the next question
- [ ] 20px spacing between the question and the answer/user input
- [ ] Increase size of radio buttons to 24px x 24px
- [ ] Increase vertical space between radio buttons and checkboxes to 16px
- [ ] Increase the minimum height of text boxes (user input) to 45px tall and add 4px rounded corners and 1px gray border (text boxes should also have the same 20px buffer on either side that the rest of the form has)
**Update previous/next button styling**:
- [ ] Increase the height of buttons to 45px and increase width of buttons to 125px
- [ ] Add 4px rounded corners to buttons
- [ ] Next button is right-aligned with a 20px buffer
- [ ] Previous button is left-aligned with a 20px buffer
- [ ] Previous button has a 2px gray border (#BCBEC0) and no fill color. Color of the text is gray (#A7A9AC)
### Screen shots for reference
**Mobile**

**Desktop**
 | 1.0 | Minor updates to Enketo UI - This issue updates the UI for Enketo forms on mobile and desktop. These are small tweaks for now and we hope to come back and make additional changes in the future. [Design spec](https://docs.google.com/document/d/1bnbQ17I9fq1ZRbu2v9otaRpUXn_7eODe7YTYzsXQoDg/edit).
**Update form header on desktop only**:
- [ ] form title is 20px bold, in black
- [ ] title is left-aligned, 20px from the left edge of the form card
- [ ] remove the "Cancel" button and replace it with an X on the right side of the header (vertically centered)
**Update form contents**:
- [ ] Put at 20px buffer on the left and right sides of the form card
- [ ] Form group titles should be blue, 18px, bold
- [ ] If there are nested groups, the sub-title (nested group name) should be black, 18px, normal weight
- [ ] 30px spacing between the group title and the first question OR between the group title and group subtitle (if there is a subtitle, which will be there if there are nested groups)
- [ ] If there is a subtitle, 20px between the subtitle and the first question
- [ ] 30px between answer options/user inputs and the next question
- [ ] 20px spacing between the question and the answer/user input
- [ ] Increase size of radio buttons to 24px x 24px
- [ ] Increase vertical space between radio buttons and checkboxes to 16px
- [ ] Increase the minimum height of text boxes (user input) to 45px tall and add 4px rounded corners and 1px gray border (text boxes should also have the same 20px buffer on either side that the rest of the form has)
**Update previous/next button styling**:
- [ ] Increase the height of buttons to 45px and increase width of buttons to 125px
- [ ] Add 4px rounded corners to buttons
- [ ] Next button is right-aligned with a 20px buffer
- [ ] Previous button is left-aligned with a 20px buffer
- [ ] Previous button has a 2px gray border (#BCBEC0) and no fill color. Color of the text is gray (#A7A9AC)
### Screen shots for reference
**Mobile**

**Desktop**
 | priority | minor updates to enketo ui this issue updates the ui for enketo forms on mobile and desktop these are small tweaks for now and we hope to come back and make additional changes in the future update form header on desktop only form title is bold in black title is left aligned from the left edge of the form card remove the cancel button and replace it with an x on the right side of the header vertically centered update form contents put at buffer on the left and right sides of the form card form group titles should be blue bold if there are nested groups the sub title nested group name should be black normal weight spacing between the group title and the first question or between the group title and group subtitle if there is a subtitle which will be there if there are nested groups if there is a subtitle between the subtitle and the first question between answer options user inputs and the next question spacing between the question and the answer user input increase size of radio buttons to x increase vertical space between radio buttons and checkboxes to increase the minimum height of text boxes user input to tall and add rounded corners and gray border text boxes should also have the same buffer on either side that the rest of the form has update previous next button styling increase the height of buttons to and increase width of buttons to add rounded corners to buttons next button is right aligned with a buffer previous button is left aligned with a buffer previous button has a gray border and no fill color color of the text is gray screen shots for reference mobile desktop | 1 |
517,282 | 14,998,668,474 | IssuesEvent | 2021-01-29 18:47:58 | JamieMason/syncpack | https://api.github.com/repos/JamieMason/syncpack | closed | Use workspace's packages version as source of truth in fix-mismatches | Priority: Medium Status: Awaiting Release Type: Feat | ## Description
When doing `fix-mismatches` it would be cool to sync dependencies version to use the local version, taken from the `package.json` version field instead of using the higher version found in the deps tree
<!--
Describe why this change is required, what problem it solves, and what
alternatives exist that you might have considered. This helps reviewers
understand the value of this change, or to highlight unnecessary changes which
can be avoided.
-->
## Suggested Solution
Use the package.json version field of workspace packages as a range present in the dependency tree
<!--
Propose some possible solutions to this issue. Try to highlight positives or
negatives of this approach compared to possible alternatives
-->
## Help Needed
<!--
What kind of help or information do you need to help you deliver this feature?
This can help Contributors understand what they can do to help you get started.
-->
| 1.0 | Use workspace's packages version as source of truth in fix-mismatches - ## Description
When doing `fix-mismatches` it would be cool to sync dependencies version to use the local version, taken from the `package.json` version field instead of using the higher version found in the deps tree
<!--
Describe why this change is required, what problem it solves, and what
alternatives exist that you might have considered. This helps reviewers
understand the value of this change, or to highlight unnecessary changes which
can be avoided.
-->
## Suggested Solution
Use the package.json version field of workspace packages as a range present in the dependency tree
<!--
Propose some possible solutions to this issue. Try to highlight positives or
negatives of this approach compared to possible alternatives
-->
## Help Needed
<!--
What kind of help or information do you need to help you deliver this feature?
This can help Contributors understand what they can do to help you get started.
-->
| priority | use workspace s packages version as source of truth in fix mismatches description when doing fix mismatches it would be cool to sync dependencies version to use the local version taken from the package json version field instead of using the higher version found in the deps tree describe why this change is required what problem it solves and what alternatives exist that you might have considered this helps reviewers understand the value of this change or to highlight unnecessary changes which can be avoided suggested solution use the package json version field of workspace packages as a range present in the dependency tree propose some possible solutions to this issue try to highlight positives or negatives of this approach compared to possible alternatives help needed what kind of help or information do you need to help you deliver this feature this can help contributors understand what they can do to help you get started | 1 |
817,399 | 30,639,684,970 | IssuesEvent | 2023-07-24 20:45:35 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Unnecessary yb_lsm.h header | kind/enhancement area/ysql priority/medium status/awaiting-triage | Jira Link: [DB-7387](https://yugabyte.atlassian.net/browse/DB-7387)
### Description
All funcs in that header could be made static.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-7387]: https://yugabyte.atlassian.net/browse/DB-7387?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YSQL] Unnecessary yb_lsm.h header - Jira Link: [DB-7387](https://yugabyte.atlassian.net/browse/DB-7387)
### Description
All funcs in that header could be made static.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-7387]: https://yugabyte.atlassian.net/browse/DB-7387?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | unnecessary yb lsm h header jira link description all funcs in that header could be made static warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information | 1 |
776,108 | 27,247,248,352 | IssuesEvent | 2023-02-22 03:47:10 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | closed | azure_rm_privatednszonelink - vNet subscription_id being overridden by value from profile | medium_priority new_featrue work in | <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
When executing module, supplying entire resource ID to the "virtual_network" parameter, seeing unexpected behavior. Ansible module seems to be reading the information provided to it correctly, however reviewing Azure logs would indicate that the wrong subscription ID is being provided to the call to ARM. The correct name and resource groups are being parsed and used.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
azure_rm_privatednszonelink
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible [core 2.13.6]
config file = None
configured module search path = ['/home/username/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/dist-packages/ansible
ansible collection location = /home/username/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/local/bin/ansible
python version = 3.9.2 (default, Feb 28 2021, 17:03:44) [GCC 10.2.1 20210110]
jinja version = 3.1.2
libyaml = True
```
##### COLLECTION VERSION
<!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes
for example: ansible-galaxy collection list community.general
-->
```paste below
# /home/username/.ansible/collections/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.14.0
# /usr/local/lib/python3.9/dist-packages/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.13.0
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
<empty>
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
username@hostname:~$ cat /etc/debian_version
11.5
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- hosts: all
connection: local
vars:
core_subscription_id: "55555555-4444-3333-2222-111111111111"
vnets:
- name: vnet-1
region: eastus
subscription: "11111111-2222-3333-4444-555555555555"
resource_group: rg-networking
dnsvnetlinks:
- zone_name: zone.example.com
registration_enabled: yes
resource_group: rg-networking
- name: Add Private DNS Zone vnet links
azure.azcollection.azure_rm_privatednszonelink:
name: "dnsvnetlink-{{ item.0['name']|regex_replace('^vnet-', '') }}"
profile: core
registration_enabled: "{{ item.1['registration_enabled'] }}"
resource_group: "{{ item.1['resource_group'] }}"
subscription_id: "{{ core_subscription_id }}"
virtual_network: "/subscriptions/{{ item.0['subscription'] }}/resourceGroups/{{ item.0['resource_group'] }}/providers/Microsoft.Network/virtualNetworks/{{ item.0['name'] }}"
zone_name: "{{ item.1['zone_name'] }}"
loop: "{{ vnets|subelements('dnsvnetlinks') }}"
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expect the vnet link to be created, or if it already exists for it to indicate so and not make any changes.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
In the command output I get an error as seen below. Also worth noting that in the verbose output (invocation section) I see the vNet with the correct ID.
In the Azure portal activity logs I get ```"message": "Virtual network resource not found for '/subscriptions/55555555-4444-3333-2222-111111111111/resourceGroups/rg-networking/providers/Microsoft.Network/virtualNetworks/vnet-1'"```. The subscription_id for the vNet is the one related to the profile being used in the task, *NOT* the one provided in the virtual_network parameter nor the one showing in the verbose ```ansible-playbook``` output.
The text error in the output from ```ansible-playbook``` would indicate an operational conflict, however that error message is less deterministic than the indication from Azure Resource Manager that an incorrect resource ID is being used.
<!--- Paste verbatim command output between quotes -->
```paste below
<azure-prod> ESTABLISH LOCAL CONNECTION FOR USER: username
<azure-prod> EXEC /bin/sh -c 'echo ~username && sleep 0'
<azure-prod> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/username/.ansible/tmp `"&& mkdir "` echo /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601 `" && echo ansible-tmp-1668013990.5085132-7259-92530896260601="` echo /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601 `" ) && sleep 0'
Using module file /home/username/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_privatednszonelink.py
<azure-prod> PUT /home/username/.ansible/tmp/ansible-local-7040joyl_665/tmp9usjihry TO /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py
<azure-prod> EXEC /bin/sh -c 'chmod u+x /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/ /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py && sleep 0'
<azure-prod> EXEC /bin/sh -c '/usr/bin/python3 /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py && sleep 0'
<azure-prod> EXEC /bin/sh -c 'rm -f -r /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/ > /dev/null 2>&1 && sleep 0'
The full traceback is:
File "/tmp/ansible_azure.azcollection.azure_rm_privatednszonelink_payload_n7i7562_/ansible_azure.azcollection.azure_rm_privatednszonelink_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_privatednszonelink.py", line 291, in create_or_update_network_link
File "/usr/local/lib/python3.9/dist-packages/azure/mgmt/privatedns/operations/_virtual_network_links_operations.py", line 163, in begin_create_or_update
raw_result = self._create_or_update_initial(
File "/usr/local/lib/python3.9/dist-packages/azure/mgmt/privatedns/operations/_virtual_network_links_operations.py", line 101, in _create_or_update_initial
map_error(status_code=response.status_code, response=response, error_map=error_map)
File "/usr/local/lib/python3.9/dist-packages/azure/core/exceptions.py", line 107, in map_error
raise error
[WARNING]: Azure API profile latest does not define an entry for PrivateDnsManagementClient
failed: [azure-prod] (item=[{'name': 'vnet-1', 'region': 'eastus', 'subscription': '11111111-2222-3333-4444-555555555555', 'resource_group': 'rg-networking', 'dnsvnetlinks': [{'zone_name': 'zone.example.com', 'registration_enabled': True, 'resource_group': 'rg-networking'}]}, {'zone_name': 'zone.example.com', 'registration_enabled': True}]) => {
"ansible_loop_var": "item",
"changed": false,
"invocation": {
"module_args": {
"ad_user": null,
"adfs_authority_url": null,
"api_profile": "latest",
"append_tags": true,
"auth_source": "auto",
"cert_validation_mode": null,
"client_id": null,
"cloud_environment": "AzureCloud",
"log_mode": null,
"log_path": null,
"name": "dnsvnetlink-1",
"password": null,
"profile": "core",
"registration_enabled": true,
"resource_group": "rg-networking",
"secret": null,
"state": "present",
"subscription_id": "55555555-4444-3333-2222-111111111111",
"tags": null,
"tenant": null,
"thumbprint": null,
"virtual_network": "/subscriptions/11111111-2222-3333-4444-555555555555/resourceGroups/rg-networking/providers/Microsoft.Network/virtualNetworks/vnet-1",
"x509_certificate_path": null,
"zone_name": "zone.example.com"
}
},
"item": [
{
"dnsvnetlinks": [
{
"registration_enabled": true,
"zone_name": "zone.example.com",
"resource_group": "rg-networking"
}
],
"name": "vnet-1",
"resource_group": "rg-networking",
"subscription": "11111111-2222-3333-4444-555555555555"
},
{
"registration_enabled": true,
"zone_name": "zone.example.com",
"resource_group": "rg-networking"
}
],
"msg": "Error creating or updating virtual network link dnsvnetlink-1 - (Conflict) Operation group '/operations/groups/id/|virtualNetworkLinks|55555555-4444-3333-2222-111111111111|rg-networking|zone.example.com|dnsvnetlink-1' already has 1 operations like '/operations/type/UpsertVirtualNetworkLink/id/5c97da98-4765-4cf8-86f1-088ce764027b_650f74e6-1373-4e94-8eac-a63e81064d9b' queued.\nCode: Conflict\nMessage: Operation group '/operations/groups/id/|virtualNetworkLinks|55555555-4444-3333-2222-111111111111|rg-networking|zone.example.com|dnsvnetlink-1' already has 1 operations like '/operations/type/UpsertVirtualNetworkLink/id/5c97da98-4765-4cf8-86f1-088ce764027b_650f74e6-1373-4e94-8eac-a63e81064d9b' queued."
```
| 1.0 | azure_rm_privatednszonelink - vNet subscription_id being overridden by value from profile - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
When executing module, supplying entire resource ID to the "virtual_network" parameter, seeing unexpected behavior. Ansible module seems to be reading the information provided to it correctly, however reviewing Azure logs would indicate that the wrong subscription ID is being provided to the call to ARM. The correct name and resource groups are being parsed and used.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
azure_rm_privatednszonelink
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible [core 2.13.6]
config file = None
configured module search path = ['/home/username/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/dist-packages/ansible
ansible collection location = /home/username/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/local/bin/ansible
python version = 3.9.2 (default, Feb 28 2021, 17:03:44) [GCC 10.2.1 20210110]
jinja version = 3.1.2
libyaml = True
```
##### COLLECTION VERSION
<!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes
for example: ansible-galaxy collection list community.general
-->
```paste below
# /home/username/.ansible/collections/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.14.0
# /usr/local/lib/python3.9/dist-packages/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.13.0
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
<empty>
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
username@hostname:~$ cat /etc/debian_version
11.5
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- hosts: all
connection: local
vars:
core_subscription_id: "55555555-4444-3333-2222-111111111111"
vnets:
- name: vnet-1
region: eastus
subscription: "11111111-2222-3333-4444-555555555555"
resource_group: rg-networking
dnsvnetlinks:
- zone_name: zone.example.com
registration_enabled: yes
resource_group: rg-networking
- name: Add Private DNS Zone vnet links
azure.azcollection.azure_rm_privatednszonelink:
name: "dnsvnetlink-{{ item.0['name']|regex_replace('^vnet-', '') }}"
profile: core
registration_enabled: "{{ item.1['registration_enabled'] }}"
resource_group: "{{ item.1['resource_group'] }}"
subscription_id: "{{ core_subscription_id }}"
virtual_network: "/subscriptions/{{ item.0['subscription'] }}/resourceGroups/{{ item.0['resource_group'] }}/providers/Microsoft.Network/virtualNetworks/{{ item.0['name'] }}"
zone_name: "{{ item.1['zone_name'] }}"
loop: "{{ vnets|subelements('dnsvnetlinks') }}"
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expect the vnet link to be created, or if it already exists for it to indicate so and not make any changes.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
In the command output I get an error as seen below. Also worth noting that in the verbose output (invocation section) I see the vNet with the correct ID.
In the Azure portal activity logs I get ```"message": "Virtual network resource not found for '/subscriptions/55555555-4444-3333-2222-111111111111/resourceGroups/rg-networking/providers/Microsoft.Network/virtualNetworks/vnet-1'"```. The subscription_id for the vNet is the one related to the profile being used in the task, *NOT* the one provided in the virtual_network parameter nor the one showing in the verbose ```ansible-playbook``` output.
The text error in the output from ```ansible-playbook``` would indicate an operational conflict, however that error message is less deterministic than the indication from Azure Resource Manager that an incorrect resource ID is being used.
<!--- Paste verbatim command output between quotes -->
```paste below
<azure-prod> ESTABLISH LOCAL CONNECTION FOR USER: username
<azure-prod> EXEC /bin/sh -c 'echo ~username && sleep 0'
<azure-prod> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/username/.ansible/tmp `"&& mkdir "` echo /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601 `" && echo ansible-tmp-1668013990.5085132-7259-92530896260601="` echo /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601 `" ) && sleep 0'
Using module file /home/username/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_privatednszonelink.py
<azure-prod> PUT /home/username/.ansible/tmp/ansible-local-7040joyl_665/tmp9usjihry TO /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py
<azure-prod> EXEC /bin/sh -c 'chmod u+x /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/ /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py && sleep 0'
<azure-prod> EXEC /bin/sh -c '/usr/bin/python3 /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/AnsiballZ_azure_rm_privatednszonelink.py && sleep 0'
<azure-prod> EXEC /bin/sh -c 'rm -f -r /home/username/.ansible/tmp/ansible-tmp-1668013990.5085132-7259-92530896260601/ > /dev/null 2>&1 && sleep 0'
The full traceback is:
File "/tmp/ansible_azure.azcollection.azure_rm_privatednszonelink_payload_n7i7562_/ansible_azure.azcollection.azure_rm_privatednszonelink_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_privatednszonelink.py", line 291, in create_or_update_network_link
File "/usr/local/lib/python3.9/dist-packages/azure/mgmt/privatedns/operations/_virtual_network_links_operations.py", line 163, in begin_create_or_update
raw_result = self._create_or_update_initial(
File "/usr/local/lib/python3.9/dist-packages/azure/mgmt/privatedns/operations/_virtual_network_links_operations.py", line 101, in _create_or_update_initial
map_error(status_code=response.status_code, response=response, error_map=error_map)
File "/usr/local/lib/python3.9/dist-packages/azure/core/exceptions.py", line 107, in map_error
raise error
[WARNING]: Azure API profile latest does not define an entry for PrivateDnsManagementClient
failed: [azure-prod] (item=[{'name': 'vnet-1', 'region': 'eastus', 'subscription': '11111111-2222-3333-4444-555555555555', 'resource_group': 'rg-networking', 'dnsvnetlinks': [{'zone_name': 'zone.example.com', 'registration_enabled': True, 'resource_group': 'rg-networking'}]}, {'zone_name': 'zone.example.com', 'registration_enabled': True}]) => {
"ansible_loop_var": "item",
"changed": false,
"invocation": {
"module_args": {
"ad_user": null,
"adfs_authority_url": null,
"api_profile": "latest",
"append_tags": true,
"auth_source": "auto",
"cert_validation_mode": null,
"client_id": null,
"cloud_environment": "AzureCloud",
"log_mode": null,
"log_path": null,
"name": "dnsvnetlink-1",
"password": null,
"profile": "core",
"registration_enabled": true,
"resource_group": "rg-networking",
"secret": null,
"state": "present",
"subscription_id": "55555555-4444-3333-2222-111111111111",
"tags": null,
"tenant": null,
"thumbprint": null,
"virtual_network": "/subscriptions/11111111-2222-3333-4444-555555555555/resourceGroups/rg-networking/providers/Microsoft.Network/virtualNetworks/vnet-1",
"x509_certificate_path": null,
"zone_name": "zone.example.com"
}
},
"item": [
{
"dnsvnetlinks": [
{
"registration_enabled": true,
"zone_name": "zone.example.com",
"resource_group": "rg-networking"
}
],
"name": "vnet-1",
"resource_group": "rg-networking",
"subscription": "11111111-2222-3333-4444-555555555555"
},
{
"registration_enabled": true,
"zone_name": "zone.example.com",
"resource_group": "rg-networking"
}
],
"msg": "Error creating or updating virtual network link dnsvnetlink-1 - (Conflict) Operation group '/operations/groups/id/|virtualNetworkLinks|55555555-4444-3333-2222-111111111111|rg-networking|zone.example.com|dnsvnetlink-1' already has 1 operations like '/operations/type/UpsertVirtualNetworkLink/id/5c97da98-4765-4cf8-86f1-088ce764027b_650f74e6-1373-4e94-8eac-a63e81064d9b' queued.\nCode: Conflict\nMessage: Operation group '/operations/groups/id/|virtualNetworkLinks|55555555-4444-3333-2222-111111111111|rg-networking|zone.example.com|dnsvnetlink-1' already has 1 operations like '/operations/type/UpsertVirtualNetworkLink/id/5c97da98-4765-4cf8-86f1-088ce764027b_650f74e6-1373-4e94-8eac-a63e81064d9b' queued."
```
| priority | azure rm privatednszonelink vnet subscription id being overridden by value from profile summary when executing module supplying entire resource id to the virtual network parameter seeing unexpected behavior ansible module seems to be reading the information provided to it correctly however reviewing azure logs would indicate that the wrong subscription id is being provided to the call to arm the correct name and resource groups are being parsed and used issue type bug report component name azure rm privatednszonelink ansible version paste below ansible config file none configured module search path ansible python module location usr local lib dist packages ansible ansible collection location home username ansible collections usr share ansible collections executable location usr local bin ansible python version default feb jinja version libyaml true collection version between the quotes for example ansible galaxy collection list community general paste below home username ansible collections ansible collections collection version azure azcollection usr local lib dist packages ansible collections collection version azure azcollection configuration paste below os environment username hostname cat etc debian version steps to reproduce yaml hosts all connection local vars core subscription id vnets name vnet region eastus subscription resource group rg networking dnsvnetlinks zone name zone example com registration enabled yes resource group rg networking name add private dns zone vnet links azure azcollection azure rm privatednszonelink name dnsvnetlink item regex replace vnet profile core registration enabled item resource group item subscription id core subscription id virtual network subscriptions item resourcegroups item providers microsoft network virtualnetworks item zone name item loop vnets subelements dnsvnetlinks expected results expect the vnet link to be created or if it already exists for it to indicate so and not make any changes actual results in the command output i get an error as seen below also worth noting that in the verbose output invocation section i see the vnet with the correct id in the azure portal activity logs i get message virtual network resource not found for subscriptions resourcegroups rg networking providers microsoft network virtualnetworks vnet the subscription id for the vnet is the one related to the profile being used in the task not the one provided in the virtual network parameter nor the one showing in the verbose ansible playbook output the text error in the output from ansible playbook would indicate an operational conflict however that error message is less deterministic than the indication from azure resource manager that an incorrect resource id is being used paste below establish local connection for user username exec bin sh c echo username sleep exec bin sh c umask mkdir p echo home username ansible tmp mkdir echo home username ansible tmp ansible tmp echo ansible tmp echo home username ansible tmp ansible tmp sleep using module file home username ansible collections ansible collections azure azcollection plugins modules azure rm privatednszonelink py put home username ansible tmp ansible local to home username ansible tmp ansible tmp ansiballz azure rm privatednszonelink py exec bin sh c chmod u x home username ansible tmp ansible tmp home username ansible tmp ansible tmp ansiballz azure rm privatednszonelink py sleep exec bin sh c usr bin home username ansible tmp ansible tmp ansiballz azure rm privatednszonelink py sleep exec bin sh c rm f r home username ansible tmp ansible tmp dev null sleep the full traceback is file tmp ansible azure azcollection azure rm privatednszonelink payload ansible azure azcollection azure rm privatednszonelink payload zip ansible collections azure azcollection plugins modules azure rm privatednszonelink py line in create or update network link file usr local lib dist packages azure mgmt privatedns operations virtual network links operations py line in begin create or update raw result self create or update initial file usr local lib dist packages azure mgmt privatedns operations virtual network links operations py line in create or update initial map error status code response status code response response error map error map file usr local lib dist packages azure core exceptions py line in map error raise error azure api profile latest does not define an entry for privatednsmanagementclient failed item zone name zone example com registration enabled true ansible loop var item changed false invocation module args ad user null adfs authority url null api profile latest append tags true auth source auto cert validation mode null client id null cloud environment azurecloud log mode null log path null name dnsvnetlink password null profile core registration enabled true resource group rg networking secret null state present subscription id tags null tenant null thumbprint null virtual network subscriptions resourcegroups rg networking providers microsoft network virtualnetworks vnet certificate path null zone name zone example com item dnsvnetlinks registration enabled true zone name zone example com resource group rg networking name vnet resource group rg networking subscription registration enabled true zone name zone example com resource group rg networking msg error creating or updating virtual network link dnsvnetlink conflict operation group operations groups id virtualnetworklinks rg networking zone example com dnsvnetlink already has operations like operations type upsertvirtualnetworklink id queued ncode conflict nmessage operation group operations groups id virtualnetworklinks rg networking zone example com dnsvnetlink already has operations like operations type upsertvirtualnetworklink id queued | 1 |
465,540 | 13,387,711,492 | IssuesEvent | 2020-09-02 16:21:43 | ansible/ansible-lint | https://api.github.com/repos/ansible/ansible-lint | opened | Add rule warning about use of run_once with strategy: free | priority/medium status/new | <!--- Verify first that your feature was not already discussed on GitHub -->
##### Summary
<!--- Describe the new feature/improvement briefly below -->
It would be useful for ansible-lint to have a rule that warns against the use of `run_once` with `strategy: free`. Ansible ignores `run_once` with the free strategy which means your tasks are run many times, once for each valid inventory host. Ansible itself does warn against this as well but doesn't fail directly. Being able to catch this early would be great.
##### Issue Type
- Feature Idea | 1.0 | Add rule warning about use of run_once with strategy: free - <!--- Verify first that your feature was not already discussed on GitHub -->
##### Summary
<!--- Describe the new feature/improvement briefly below -->
It would be useful for ansible-lint to have a rule that warns against the use of `run_once` with `strategy: free`. Ansible ignores `run_once` with the free strategy which means your tasks are run many times, once for each valid inventory host. Ansible itself does warn against this as well but doesn't fail directly. Being able to catch this early would be great.
##### Issue Type
- Feature Idea | priority | add rule warning about use of run once with strategy free summary it would be useful for ansible lint to have a rule that warns against the use of run once with strategy free ansible ignores run once with the free strategy which means your tasks are run many times once for each valid inventory host ansible itself does warn against this as well but doesn t fail directly being able to catch this early would be great issue type feature idea | 1 |
654,049 | 21,635,223,236 | IssuesEvent | 2022-05-05 13:41:19 | ita-social-projects/horondi_admin | https://api.github.com/repos/ita-social-projects/horondi_admin | closed | [Sidebar] Move links to materials, news, about us pages, etc from business pages to static pages as subfields. | priority: medium UI severity: major | Move links to materials, news, about us pages, etc from business pages to static pages as subfields.


| 1.0 | [Sidebar] Move links to materials, news, about us pages, etc from business pages to static pages as subfields. - Move links to materials, news, about us pages, etc from business pages to static pages as subfields.


| priority | move links to materials news about us pages etc from business pages to static pages as subfields move links to materials news about us pages etc from business pages to static pages as subfields | 1 |
31,397 | 2,732,898,868 | IssuesEvent | 2015-04-17 10:04:55 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | Option for selecting the size of a PNG export | auto-migrated OpSys-All Priority-Medium Type-Enhancement | ```
It would be nice to have an option, e.g. as a query string parameter of the REST
resource, to specify the dimensions of an exported PNG file.
If that is not possible. at least increasing the default resolution would be of
help,
since the current export is not even usable for presentations not to speak of
printed media.
At the moment the only workaround I found for getting higher resolution images
out of Oryx is to open the SVG in Gimp. However, that's not really user
friendly and
to some extent contravenes the principle of being able to access and share
models via links.
```
Original issue reported on code.google.com by `falko.me...@gmail.com` on 14 Nov 2008 at 4:01 | 1.0 | Option for selecting the size of a PNG export - ```
It would be nice to have an option, e.g. as a query string parameter of the REST
resource, to specify the dimensions of an exported PNG file.
If that is not possible. at least increasing the default resolution would be of
help,
since the current export is not even usable for presentations not to speak of
printed media.
At the moment the only workaround I found for getting higher resolution images
out of Oryx is to open the SVG in Gimp. However, that's not really user
friendly and
to some extent contravenes the principle of being able to access and share
models via links.
```
Original issue reported on code.google.com by `falko.me...@gmail.com` on 14 Nov 2008 at 4:01 | priority | option for selecting the size of a png export it would be nice to have an option e g as a query string parameter of the rest resource to specify the dimensions of an exported png file if that is not possible at least increasing the default resolution would be of help since the current export is not even usable for presentations not to speak of printed media at the moment the only workaround i found for getting higher resolution images out of oryx is to open the svg in gimp however that s not really user friendly and to some extent contravenes the principle of being able to access and share models via links original issue reported on code google com by falko me gmail com on nov at | 1 |
685,401 | 23,454,857,437 | IssuesEvent | 2022-08-16 07:58:09 | lukasoppermann/hourglass | https://api.github.com/repos/lukasoppermann/hourglass | closed | Task details | medium priority | As a user I want to _add details to tasks_ to _remember what to do_.
### add notes (text)
- add lists (using markdown lists in note field or pressing enter in list)
- recognise links and email addresses (using regex? Lukas will supply the regex)
### add budget
see budget task
### add files
- via drag and drop (files are only aliases, open file in default app)
- [x] design for drop indicator
- [x] design for removing file
### Access timing sessions
- click on time opens timing panel
- another click closes the timing panel (either goes back to details or to closes task, depending on previous state)
- [x] design needed
### Closing details
- whenever another detail is opened (or the list is closed) the details window is closed
- a doubleclick closes the details
### Links
Markdown Parser Library https://github.com/Cocoanetics/DTMarkdownParser (although we do not really parse markdown, only lists, but maybe we can use the list logic)
| 1.0 | Task details - As a user I want to _add details to tasks_ to _remember what to do_.
### add notes (text)
- add lists (using markdown lists in note field or pressing enter in list)
- recognise links and email addresses (using regex? Lukas will supply the regex)
### add budget
see budget task
### add files
- via drag and drop (files are only aliases, open file in default app)
- [x] design for drop indicator
- [x] design for removing file
### Access timing sessions
- click on time opens timing panel
- another click closes the timing panel (either goes back to details or to closes task, depending on previous state)
- [x] design needed
### Closing details
- whenever another detail is opened (or the list is closed) the details window is closed
- a doubleclick closes the details
### Links
Markdown Parser Library https://github.com/Cocoanetics/DTMarkdownParser (although we do not really parse markdown, only lists, but maybe we can use the list logic)
| priority | task details as a user i want to add details to tasks to remember what to do add notes text add lists using markdown lists in note field or pressing enter in list recognise links and email addresses using regex lukas will supply the regex add budget see budget task add files via drag and drop files are only aliases open file in default app design for drop indicator design for removing file access timing sessions click on time opens timing panel another click closes the timing panel either goes back to details or to closes task depending on previous state design needed closing details whenever another detail is opened or the list is closed the details window is closed a doubleclick closes the details links markdown parser library although we do not really parse markdown only lists but maybe we can use the list logic | 1 |
826,094 | 31,552,182,959 | IssuesEvent | 2023-09-02 07:17:54 | vicholp/template-laravel | https://api.github.com/repos/vicholp/template-laravel | closed | Check outdated dependencies | priority: medium | Remove dependabot
```
composer outdated -D --no-dev -m -f json
```
```
npm outdated -j -l
```
| 1.0 | Check outdated dependencies - Remove dependabot
```
composer outdated -D --no-dev -m -f json
```
```
npm outdated -j -l
```
| priority | check outdated dependencies remove dependabot composer outdated d no dev m f json npm outdated j l | 1 |
606,900 | 18,770,094,622 | IssuesEvent | 2021-11-06 17:26:17 | AY2122S1-CS2113T-T12-3/tp | https://api.github.com/repos/AY2122S1-CS2113T-T12-3/tp | closed | [PE-D] Hard-to-test features related to datetime | priority.Medium | Since date-time is set as the current date when the user adds in new items, it is very hard to test when working with date-time in the past or in the future.
<!--session: 1635497070033-2edf8de8-fb5a-49df-be08-9aeb1415bdf6-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: nvbinh15/ped#5 | 1.0 | [PE-D] Hard-to-test features related to datetime - Since date-time is set as the current date when the user adds in new items, it is very hard to test when working with date-time in the past or in the future.
<!--session: 1635497070033-2edf8de8-fb5a-49df-be08-9aeb1415bdf6-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Medium` `type.FeatureFlaw`
original: nvbinh15/ped#5 | priority | hard to test features related to datetime since date time is set as the current date when the user adds in new items it is very hard to test when working with date time in the past or in the future labels severity medium type featureflaw original ped | 1 |
561,725 | 16,622,136,739 | IssuesEvent | 2021-06-03 03:43:05 | rstudio/gt | https://api.github.com/repos/rstudio/gt | closed | Add support for side-by-side regression tables? | Difficulty: [2] Intermediate Effort: [2] Medium Priority: [2] Medium Type: ★ Enhancement | This is a phenomenal package and I'm a huge fan of the API for creating tables.
At least two other packages provide support for side-by-side regression tables ([**stargazer**](https://cran.r-project.org/web/packages/stargazer/index.html) and [**huxtable**](https://github.com/hughjonesd/huxtable)), but both have limitations: **stargazer** only supports HTML and TeX and doesn't play well with knitr, and **huxtable** supports HTML and TeX with minimal Word support, since it creates Markdown tables that don't support column spans or other fancier table features.
I wonder if it would be possible to provide support for regression tables similar to **stargazer** and **huxtable** tables, given that the format fits well in the gt API paradigm.
``` r
library(tidyverse)
library(gt)
library(huxtable)
model1 <- lm(mpg_c ~ hp, data = gtcars)
model2 <- lm(mpg_c ~ hp + trq, data = gtcars)
model3 <- lm(mpg_c ~ hp + trq + year, data = gtcars)
huxreg(model1, model2, model3)
```
```
─────────────────────────────────────────────────────────────
(1) (2) (3)
───────────────────────────────────────────────
(Intercept) 23.932 *** 22.422 *** -1275.504
(1.540) (1.777) (1174.083)
hp -0.017 *** -0.024 *** -0.021 ***
(0.003) (0.005) (0.006)
trq 0.012 0.008
(0.007) (0.008)
year 0.644
(0.582)
───────────────────────────────────────────────
N 46 46 46
R2 0.431 0.463 0.479
logLik -108.446 -107.086 -106.426
AIC 222.892 222.172 222.853
─────────────────────────────────────────────────────────────
*** p < 0.001; ** p < 0.01; * p < 0.05.
Column names: names, model1, model2, model3
```
Right now there's a way to fake it very uglyly by extracting coefficients and model details using functions from **broom** (which **huxtable** does behind the scenes too), but it'd be cool if there was a less clunky way to make side-by-side regression tables with **gt**:
```r
library(broom)
library(glue)
models_combined <- tibble(model = list(model1, model2, model3)) %>%
mutate(coefs = model %>% map(tidy),
details = model %>% map(glance),
model_number = 1:n())
model_coefs <- models_combined %>%
unnest(coefs) %>%
mutate(value = as.character(glue("{round(estimate, 2)} ({round(std.error, 2)})"))) %>%
select(model_number, term, value) %>%
spread(model_number, value)
model_details <- models_combined %>%
unnest(details) %>%
mutate(N = model %>% map_dbl(nobs),
R2 = round(adj.r.squared, 2),
AIC = round(AIC, 2)) %>%
select(model_number, N, R2, AIC) %>%
gather(term, value, -model_number) %>%
spread(model_number, value) %>%
mutate_at(vars(-term), as.character)
ugly_blank_row <- tibble(term = NA, `1` = NA, `2` = NA, `3` = NA)
bind_rows(model_coefs,ugly_blank_row, model_details) %>%
gt() %>%
tab_header(title = "Side-by-side regression table")
```
<!--html_preserve-->
<div id="gefoycaxtg">
<!--gt table start-->
<table class='gt_table'>
<thead>
<tr>
<th class="gt_heading gt_title gt_font_normal gt_center" colspan="4">
Side-by-side regression table
</th>
</tr>
<tr>
<th class="gt_heading gt_subtitle gt_font_normal gt_center gt_bottom_border" colspan="4">
</th>
</tr>
<tr>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
term
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
1
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
2
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
3
</th>
</tr>
<tbody class="gt_table_body gt_striped">
<tr>
<td class="gt_row gt_left">
(Intercept)
</td>
<td class="gt_row gt_left">
23.93 (1.54)
</td>
<td class="gt_row gt_left">
22.42 (1.78)
</td>
<td class="gt_row gt_left">
\-1275.5 (1174.08)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
hp
</td>
<td class="gt_row gt_left">
\-0.02 (0)
</td>
<td class="gt_row gt_left">
\-0.02 (0.01)
</td>
<td class="gt_row gt_left">
\-0.02 (0.01)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
trq
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
0.01 (0.01)
</td>
<td class="gt_row gt_left">
0.01 (0.01)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
year
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
0.64 (0.58)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
</tr>
<tr>
<td class="gt_row gt_left">
AIC
</td>
<td class="gt_row gt_left">
222.89
</td>
<td class="gt_row gt_left">
222.17
</td>
<td class="gt_row gt_left">
222.85
</td>
</tr>
<tr>
<td class="gt_row gt_left">
N
</td>
<td class="gt_row gt_left">
46
</td>
<td class="gt_row gt_left">
46
</td>
<td class="gt_row gt_left">
46
</td>
</tr>
<tr>
<td class="gt_row gt_left">
R2
</td>
<td class="gt_row gt_left">
0.42
</td>
<td class="gt_row gt_left">
0.44
</td>
<td class="gt_row gt_left">
0.44
</td>
</tr>
</tbody>
</table>
<!--gt table end-->
</div>
<!--/html_preserve--> | 1.0 | Add support for side-by-side regression tables? - This is a phenomenal package and I'm a huge fan of the API for creating tables.
At least two other packages provide support for side-by-side regression tables ([**stargazer**](https://cran.r-project.org/web/packages/stargazer/index.html) and [**huxtable**](https://github.com/hughjonesd/huxtable)), but both have limitations: **stargazer** only supports HTML and TeX and doesn't play well with knitr, and **huxtable** supports HTML and TeX with minimal Word support, since it creates Markdown tables that don't support column spans or other fancier table features.
I wonder if it would be possible to provide support for regression tables similar to **stargazer** and **huxtable** tables, given that the format fits well in the gt API paradigm.
``` r
library(tidyverse)
library(gt)
library(huxtable)
model1 <- lm(mpg_c ~ hp, data = gtcars)
model2 <- lm(mpg_c ~ hp + trq, data = gtcars)
model3 <- lm(mpg_c ~ hp + trq + year, data = gtcars)
huxreg(model1, model2, model3)
```
```
─────────────────────────────────────────────────────────────
(1) (2) (3)
───────────────────────────────────────────────
(Intercept) 23.932 *** 22.422 *** -1275.504
(1.540) (1.777) (1174.083)
hp -0.017 *** -0.024 *** -0.021 ***
(0.003) (0.005) (0.006)
trq 0.012 0.008
(0.007) (0.008)
year 0.644
(0.582)
───────────────────────────────────────────────
N 46 46 46
R2 0.431 0.463 0.479
logLik -108.446 -107.086 -106.426
AIC 222.892 222.172 222.853
─────────────────────────────────────────────────────────────
*** p < 0.001; ** p < 0.01; * p < 0.05.
Column names: names, model1, model2, model3
```
Right now there's a way to fake it very uglyly by extracting coefficients and model details using functions from **broom** (which **huxtable** does behind the scenes too), but it'd be cool if there was a less clunky way to make side-by-side regression tables with **gt**:
```r
library(broom)
library(glue)
models_combined <- tibble(model = list(model1, model2, model3)) %>%
mutate(coefs = model %>% map(tidy),
details = model %>% map(glance),
model_number = 1:n())
model_coefs <- models_combined %>%
unnest(coefs) %>%
mutate(value = as.character(glue("{round(estimate, 2)} ({round(std.error, 2)})"))) %>%
select(model_number, term, value) %>%
spread(model_number, value)
model_details <- models_combined %>%
unnest(details) %>%
mutate(N = model %>% map_dbl(nobs),
R2 = round(adj.r.squared, 2),
AIC = round(AIC, 2)) %>%
select(model_number, N, R2, AIC) %>%
gather(term, value, -model_number) %>%
spread(model_number, value) %>%
mutate_at(vars(-term), as.character)
ugly_blank_row <- tibble(term = NA, `1` = NA, `2` = NA, `3` = NA)
bind_rows(model_coefs,ugly_blank_row, model_details) %>%
gt() %>%
tab_header(title = "Side-by-side regression table")
```
<!--html_preserve-->
<div id="gefoycaxtg">
<!--gt table start-->
<table class='gt_table'>
<thead>
<tr>
<th class="gt_heading gt_title gt_font_normal gt_center" colspan="4">
Side-by-side regression table
</th>
</tr>
<tr>
<th class="gt_heading gt_subtitle gt_font_normal gt_center gt_bottom_border" colspan="4">
</th>
</tr>
<tr>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
term
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
1
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
2
</th>
<th class="gt_col_heading gt_left" rowspan="1" colspan="1">
3
</th>
</tr>
<tbody class="gt_table_body gt_striped">
<tr>
<td class="gt_row gt_left">
(Intercept)
</td>
<td class="gt_row gt_left">
23.93 (1.54)
</td>
<td class="gt_row gt_left">
22.42 (1.78)
</td>
<td class="gt_row gt_left">
\-1275.5 (1174.08)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
hp
</td>
<td class="gt_row gt_left">
\-0.02 (0)
</td>
<td class="gt_row gt_left">
\-0.02 (0.01)
</td>
<td class="gt_row gt_left">
\-0.02 (0.01)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
trq
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
0.01 (0.01)
</td>
<td class="gt_row gt_left">
0.01 (0.01)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
year
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
0.64 (0.58)
</td>
</tr>
<tr>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
<td class="gt_row gt_left">
NA
</td>
</tr>
<tr>
<td class="gt_row gt_left">
AIC
</td>
<td class="gt_row gt_left">
222.89
</td>
<td class="gt_row gt_left">
222.17
</td>
<td class="gt_row gt_left">
222.85
</td>
</tr>
<tr>
<td class="gt_row gt_left">
N
</td>
<td class="gt_row gt_left">
46
</td>
<td class="gt_row gt_left">
46
</td>
<td class="gt_row gt_left">
46
</td>
</tr>
<tr>
<td class="gt_row gt_left">
R2
</td>
<td class="gt_row gt_left">
0.42
</td>
<td class="gt_row gt_left">
0.44
</td>
<td class="gt_row gt_left">
0.44
</td>
</tr>
</tbody>
</table>
<!--gt table end-->
</div>
<!--/html_preserve--> | priority | add support for side by side regression tables this is a phenomenal package and i m a huge fan of the api for creating tables at least two other packages provide support for side by side regression tables and but both have limitations stargazer only supports html and tex and doesn t play well with knitr and huxtable supports html and tex with minimal word support since it creates markdown tables that don t support column spans or other fancier table features i wonder if it would be possible to provide support for regression tables similar to stargazer and huxtable tables given that the format fits well in the gt api paradigm r library tidyverse library gt library huxtable lm mpg c hp data gtcars lm mpg c hp trq data gtcars lm mpg c hp trq year data gtcars huxreg ───────────────────────────────────────────────────────────── ─────────────────────────────────────────────── intercept hp trq year ─────────────────────────────────────────────── n loglik aic ───────────────────────────────────────────────────────────── p p p column names names right now there s a way to fake it very uglyly by extracting coefficients and model details using functions from broom which huxtable does behind the scenes too but it d be cool if there was a less clunky way to make side by side regression tables with gt r library broom library glue models combined mutate coefs model map tidy details model map glance model number n model coefs unnest coefs mutate value as character glue round estimate round std error select model number term value spread model number value model details unnest details mutate n model map dbl nobs round adj r squared aic round aic select model number n aic gather term value model number spread model number value mutate at vars term as character ugly blank row tibble term na na na na bind rows model coefs ugly blank row model details gt tab header title side by side regression table side by side regression table term intercept hp trq na year na na na na na na aic n | 1 |
531,373 | 15,496,262,882 | IssuesEvent | 2021-03-11 02:21:16 | dtcenter/MET | https://api.github.com/repos/dtcenter/MET | closed | For Python embedding, support the grid attribute being defined as a grid specification string. | alert: NEED ACCOUNT KEY component: python interface priority: medium requestor: METplus Team type: enhancement | ## Describe the Enhancement ##
Python embedding currently requires a user to define all attributes about the grid that their data is on. Could we support numbered grids already known to MET, either via a new Python embedding-specific attribute (i.e. grid_number) or via the set_attr_grid command line argument like shown in #1020 ? This would end up supporting a wide variety of data on commonly-used grids already known to MET, and relieve the user from needing to re-describe their data using all of the required grid attributes MET needs in those cases.
### Time Estimate ###
1-2 days
### Relevant Deadlines ###
NONE
### Funding Source ###
NONE
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required: John HG
- [x] Select **scientist(s)** or **no scientist** required: none required but have Dan A review the PR
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones
- [x] Select **milestone**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [x] Complete the development and test your changes.
- [x] Add/update unit tests.
- [x] Add/update documentation.
- [x] Push local changes to GitHub.
- [x] Submit a pull request to merge into **develop** and link the pull request to this issue..
Pull request: `feature <Issue Number> <Description>`
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | For Python embedding, support the grid attribute being defined as a grid specification string. - ## Describe the Enhancement ##
Python embedding currently requires a user to define all attributes about the grid that their data is on. Could we support numbered grids already known to MET, either via a new Python embedding-specific attribute (i.e. grid_number) or via the set_attr_grid command line argument like shown in #1020 ? This would end up supporting a wide variety of data on commonly-used grids already known to MET, and relieve the user from needing to re-describe their data using all of the required grid attributes MET needs in those cases.
### Time Estimate ###
1-2 days
### Relevant Deadlines ###
NONE
### Funding Source ###
NONE
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required: John HG
- [x] Select **scientist(s)** or **no scientist** required: none required but have Dan A review the PR
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones
- [x] Select **milestone**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [x] Complete the development and test your changes.
- [x] Add/update unit tests.
- [x] Add/update documentation.
- [x] Push local changes to GitHub.
- [x] Submit a pull request to merge into **develop** and link the pull request to this issue..
Pull request: `feature <Issue Number> <Description>`
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | for python embedding support the grid attribute being defined as a grid specification string describe the enhancement python embedding currently requires a user to define all attributes about the grid that their data is on could we support numbered grids already known to met either via a new python embedding specific attribute i e grid number or via the set attr grid command line argument like shown in this would end up supporting a wide variety of data on commonly used grids already known to met and relieve the user from needing to re describe their data using all of the required grid attributes met needs in those cases time estimate days relevant deadlines none funding source none assignee select engineer s or no engineer required john hg select scientist s or no scientist required none required but have dan a review the pr labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones select milestone define related issue s consider the impact to the other metplus components no impacts enhancement checklist see the for details complete the issue definition above fork this repository or create a branch of develop branch name feature complete the development and test your changes add update unit tests add update documentation push local changes to github submit a pull request to merge into develop and link the pull request to this issue pull request feature iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
479,781 | 13,805,704,651 | IssuesEvent | 2020-10-11 14:46:44 | ayumi-cloud/oc2-security-module | https://api.github.com/repos/ayumi-cloud/oc2-security-module | opened | Move `Cloudflare` Crawlers into bot module | Add to Whitelist Firewall In-progress Priority: Medium | ### Enhancement idea
- [ ] Move `Cloudflare` Crawlers into bot module.
| 1.0 | Move `Cloudflare` Crawlers into bot module - ### Enhancement idea
- [ ] Move `Cloudflare` Crawlers into bot module.
| priority | move cloudflare crawlers into bot module enhancement idea move cloudflare crawlers into bot module | 1 |
518,770 | 15,034,446,064 | IssuesEvent | 2021-02-02 12:54:20 | gnosis/conditional-tokens-explorer | https://api.github.com/repos/gnosis/conditional-tokens-explorer | closed | Redeem positions: Display details on question and outcome names for Omen/Reality.eth positions. | Medium priority enhancement feature requested | Case 5 in the #546 | 1.0 | Redeem positions: Display details on question and outcome names for Omen/Reality.eth positions. - Case 5 in the #546 | priority | redeem positions display details on question and outcome names for omen reality eth positions case in the | 1 |
140,925 | 5,426,114,869 | IssuesEvent | 2017-03-03 09:05:29 | NostraliaWoW/mangoszero | https://api.github.com/repos/NostraliaWoW/mangoszero | closed | Zul'farrak event | Dungeon Priority - Medium System | Name - event after the stairs part, to engage the prisoners and have the goblin destroy the wall to allow access to final boss.
Summary - after finishing the stairs event and boss there I attempted to talk to goblin dude to get him to blow the wall, allowing access up to the final boss. He would only greet me, nothing more. Additionally the other NPC's all hearthstoned out about 5-10 seconds later without us having the chance to fight them for the quest
Steps to Reproduce - complete stairs event, wait for them to hearth, talk to goblin.
Screen Shots/Video -
Reported By paprika
Extra Information - I believe the npcs hearthing away may have been due to the group being half horde half alliance. From memory only horde fight them for a quest, alliance don't.
post,
https://nostralia.org/forums/viewtopic.php?f=24&t=274 | 1.0 | Zul'farrak event - Name - event after the stairs part, to engage the prisoners and have the goblin destroy the wall to allow access to final boss.
Summary - after finishing the stairs event and boss there I attempted to talk to goblin dude to get him to blow the wall, allowing access up to the final boss. He would only greet me, nothing more. Additionally the other NPC's all hearthstoned out about 5-10 seconds later without us having the chance to fight them for the quest
Steps to Reproduce - complete stairs event, wait for them to hearth, talk to goblin.
Screen Shots/Video -
Reported By paprika
Extra Information - I believe the npcs hearthing away may have been due to the group being half horde half alliance. From memory only horde fight them for a quest, alliance don't.
post,
https://nostralia.org/forums/viewtopic.php?f=24&t=274 | priority | zul farrak event name event after the stairs part to engage the prisoners and have the goblin destroy the wall to allow access to final boss summary after finishing the stairs event and boss there i attempted to talk to goblin dude to get him to blow the wall allowing access up to the final boss he would only greet me nothing more additionally the other npc s all hearthstoned out about seconds later without us having the chance to fight them for the quest steps to reproduce complete stairs event wait for them to hearth talk to goblin screen shots video reported by paprika extra information i believe the npcs hearthing away may have been due to the group being half horde half alliance from memory only horde fight them for a quest alliance don t post | 1 |
34,262 | 2,776,595,235 | IssuesEvent | 2015-05-04 22:43:06 | umutafacan/bounswe2015group3 | https://api.github.com/repos/umutafacan/bounswe2015group3 | reopened | adding new use cases and creating use case diagram | auto-migrated Priority-Medium Type-Task | ```
add new use cases and create use case diagram
```
Original issue reported on code.google.com by `bayrakta...@gmail.com` on 14 Mar 2015 at 5:26 | 1.0 | adding new use cases and creating use case diagram - ```
add new use cases and create use case diagram
```
Original issue reported on code.google.com by `bayrakta...@gmail.com` on 14 Mar 2015 at 5:26 | priority | adding new use cases and creating use case diagram add new use cases and create use case diagram original issue reported on code google com by bayrakta gmail com on mar at | 1 |
821,602 | 30,828,069,464 | IssuesEvent | 2023-08-01 21:57:52 | Haidoe/arc | https://api.github.com/repos/Haidoe/arc | closed | Date selector should be centered on mobile screen | priority-medium style | ## Bug Report
**Reporter: ❗️**
@ksdhir
**Describe the bug: ❗️**
The date selector is left aligned on the mobile screen. Should be centered as displayed in the mockups.
**Steps to reproduce: ❗️**
1. Go to 'dashboard' of any production
2. Resize your browser to mobile screen.
3. Scroll to the Production Progress chart.
**Screenshots or Video**

**Expected behavior: ❗️**
The date selector should be centered as displayed in mockups.
**Actual behavior: ❗️**
The date selector is left aligned.
**Follow up checklist: ❗️**
- [x] Add Assignee
- [x] Label priority ( priority-low, priority-medium, priority-high )
- [x] Label Milestone ( Alpha, Beta )
- [x] Label Issue Type ( Style, Code, API )
| 1.0 | Date selector should be centered on mobile screen - ## Bug Report
**Reporter: ❗️**
@ksdhir
**Describe the bug: ❗️**
The date selector is left aligned on the mobile screen. Should be centered as displayed in the mockups.
**Steps to reproduce: ❗️**
1. Go to 'dashboard' of any production
2. Resize your browser to mobile screen.
3. Scroll to the Production Progress chart.
**Screenshots or Video**

**Expected behavior: ❗️**
The date selector should be centered as displayed in mockups.
**Actual behavior: ❗️**
The date selector is left aligned.
**Follow up checklist: ❗️**
- [x] Add Assignee
- [x] Label priority ( priority-low, priority-medium, priority-high )
- [x] Label Milestone ( Alpha, Beta )
- [x] Label Issue Type ( Style, Code, API )
| priority | date selector should be centered on mobile screen bug report reporter ❗️ ksdhir describe the bug ❗️ the date selector is left aligned on the mobile screen should be centered as displayed in the mockups steps to reproduce ❗️ go to dashboard of any production resize your browser to mobile screen scroll to the production progress chart screenshots or video expected behavior ❗️ the date selector should be centered as displayed in mockups actual behavior ❗️ the date selector is left aligned follow up checklist ❗️ add assignee label priority priority low priority medium priority high label milestone alpha beta label issue type style code api | 1 |
420,069 | 12,232,724,143 | IssuesEvent | 2020-05-04 10:13:22 | osmontrouge/caresteouvert | https://api.github.com/repos/osmontrouge/caresteouvert | closed | Layout issues with longer translations in Missing + report dialogues | bug priority: medium | **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://www.caresteouvert.fr/@48.864702,2.332550,17.08/place/n682347923
2. Switch language to "Deutsch".
3. Click on "Fehlendes Geschäft melden"
4. See layout issues in both dialogues.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**

**Desktop:**
- OS: Windows 10
- Browser: Firefox
- Version: 77.0a1
**Additional context**
(None.)
| 1.0 | Layout issues with longer translations in Missing + report dialogues - **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://www.caresteouvert.fr/@48.864702,2.332550,17.08/place/n682347923
2. Switch language to "Deutsch".
3. Click on "Fehlendes Geschäft melden"
4. See layout issues in both dialogues.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**

**Desktop:**
- OS: Windows 10
- Browser: Firefox
- Version: 77.0a1
**Additional context**
(None.)
| priority | layout issues with longer translations in missing report dialogues describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior go to switch language to deutsch click on fehlendes geschäft melden see layout issues in both dialogues expected behavior a clear and concise description of what you expected to happen screenshots desktop os windows browser firefox version additional context none | 1 |
815,590 | 30,563,287,367 | IssuesEvent | 2023-07-20 15:54:55 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Mesh warnings from franka_description with recent Drake versions | type: bug priority: medium component: geometry perception | ### What happened?
In `drake/manipulation/models/franka_description/urdf` we provide some sample URDFs for the Franka Panda robot.
If one of those models is loaded into a scene that contains render engines (e.g., for camera simulation) then there are a log of warnings spammed to the console:
For example:
```
[console] [warning] warning: Drake currently only supports OBJs that use a single material across the whole mesh; for drake_models/franka_description/meshes/visual/hand.obj, 5 materials were used: 'Part__Feature001_008_005', 'Part__Feature002_005_005', 'Part__Feature005_001_005', 'Part__Feature005_001_005_001', 'Part__Feature_009_005'. The parsed materials will not be used.
```
It's not OK for Drake models to trigger Drake warnings. We either need to fix the model, or nerf the warning.
For now, our plan is to nerf the warning.
### Version
1.19.0
### What operating system are you using?
Ubuntu 22.04
### What installation option are you using?
compiled from source code using Bazel
### Relevant log output
```
jwnimmer@call-cps:~/jwnimmer-tri/drake$ bazel run //tools:model_visualizer -- package://drake/manipulation/models/franka_description/urdf/hand.urdf --show_rgbd_sensor
INFO: Meshcat listening for connections at http://localhost:7000
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/finger.obj, 2 materials were used: 'Part__Feature001_006', 'Part__Feature_007'. The parsed materials will not be used.
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/finger.obj, 2 materials were used: 'Part__Feature001_006', 'Part__Feature_007'. The parsed materials will not be used.
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/hand.obj, 5 materials were used: 'Part__Feature001_008_005', 'Part__Feature002_005_005', 'Part__Feature005_001_005', 'Part__Feature005_001_005_001', 'Part__Feature_009_005'. The parsed materials will not be used.
``` | 1.0 | Mesh warnings from franka_description with recent Drake versions - ### What happened?
In `drake/manipulation/models/franka_description/urdf` we provide some sample URDFs for the Franka Panda robot.
If one of those models is loaded into a scene that contains render engines (e.g., for camera simulation) then there are a log of warnings spammed to the console:
For example:
```
[console] [warning] warning: Drake currently only supports OBJs that use a single material across the whole mesh; for drake_models/franka_description/meshes/visual/hand.obj, 5 materials were used: 'Part__Feature001_008_005', 'Part__Feature002_005_005', 'Part__Feature005_001_005', 'Part__Feature005_001_005_001', 'Part__Feature_009_005'. The parsed materials will not be used.
```
It's not OK for Drake models to trigger Drake warnings. We either need to fix the model, or nerf the warning.
For now, our plan is to nerf the warning.
### Version
1.19.0
### What operating system are you using?
Ubuntu 22.04
### What installation option are you using?
compiled from source code using Bazel
### Relevant log output
```
jwnimmer@call-cps:~/jwnimmer-tri/drake$ bazel run //tools:model_visualizer -- package://drake/manipulation/models/franka_description/urdf/hand.urdf --show_rgbd_sensor
INFO: Meshcat listening for connections at http://localhost:7000
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/finger.obj, 2 materials were used: 'Part__Feature001_006', 'Part__Feature_007'. The parsed materials will not be used.
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/finger.obj, 2 materials were used: 'Part__Feature001_006', 'Part__Feature_007'. The parsed materials will not be used.
WARNING: warning: Drake currently only supports OBJs that use a single material across the whole mesh; for .../drake_models/franka_description/meshes/visual/hand.obj, 5 materials were used: 'Part__Feature001_008_005', 'Part__Feature002_005_005', 'Part__Feature005_001_005', 'Part__Feature005_001_005_001', 'Part__Feature_009_005'. The parsed materials will not be used.
``` | priority | mesh warnings from franka description with recent drake versions what happened in drake manipulation models franka description urdf we provide some sample urdfs for the franka panda robot if one of those models is loaded into a scene that contains render engines e g for camera simulation then there are a log of warnings spammed to the console for example warning drake currently only supports objs that use a single material across the whole mesh for drake models franka description meshes visual hand obj materials were used part part part part part feature the parsed materials will not be used it s not ok for drake models to trigger drake warnings we either need to fix the model or nerf the warning for now our plan is to nerf the warning version what operating system are you using ubuntu what installation option are you using compiled from source code using bazel relevant log output jwnimmer call cps jwnimmer tri drake bazel run tools model visualizer package drake manipulation models franka description urdf hand urdf show rgbd sensor info meshcat listening for connections at warning warning drake currently only supports objs that use a single material across the whole mesh for drake models franka description meshes visual finger obj materials were used part part feature the parsed materials will not be used warning warning drake currently only supports objs that use a single material across the whole mesh for drake models franka description meshes visual finger obj materials were used part part feature the parsed materials will not be used warning warning drake currently only supports objs that use a single material across the whole mesh for drake models franka description meshes visual hand obj materials were used part part part part part feature the parsed materials will not be used | 1 |
202,369 | 7,047,328,141 | IssuesEvent | 2018-01-02 12:58:37 | Casumo/BarcelonaOffice | https://api.github.com/repos/Casumo/BarcelonaOffice | closed | Accomodate Dimitri Visnadi | Medium Priority | Dimitri is joining our team. He arrives the 29th of December and will stay at least for 6 weeks until he finds an apartment for him. | 1.0 | Accomodate Dimitri Visnadi - Dimitri is joining our team. He arrives the 29th of December and will stay at least for 6 weeks until he finds an apartment for him. | priority | accomodate dimitri visnadi dimitri is joining our team he arrives the of december and will stay at least for weeks until he finds an apartment for him | 1 |
498,025 | 14,399,041,415 | IssuesEvent | 2020-12-03 10:23:20 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Add Search box for photos and album page | component: media feature: enhancement priority: medium | **Is your feature request related to a problem? Please describe.**
Although we have the same functionality on the global photos page. Users can search for photos from the global photos page but not from Member Profile and Group photos page.
**Describe the solution you'd like**
Add Search box for photos and album page
**Screenshots**
https://prnt.sc/vmn8cf
https://prnt.sc/vmn968
https://i.imgur.com/FLA21E0.jpg | 1.0 | Add Search box for photos and album page - **Is your feature request related to a problem? Please describe.**
Although we have the same functionality on the global photos page. Users can search for photos from the global photos page but not from Member Profile and Group photos page.
**Describe the solution you'd like**
Add Search box for photos and album page
**Screenshots**
https://prnt.sc/vmn8cf
https://prnt.sc/vmn968
https://i.imgur.com/FLA21E0.jpg | priority | add search box for photos and album page is your feature request related to a problem please describe although we have the same functionality on the global photos page users can search for photos from the global photos page but not from member profile and group photos page describe the solution you d like add search box for photos and album page screenshots | 1 |
91,873 | 3,863,516,300 | IssuesEvent | 2016-04-08 09:45:33 | iamxavier/elmah | https://api.github.com/repos/iamxavier/elmah | closed | elmah usage example with windows based project in framework 3.5 | auto-migrated Priority-Medium Type-Enhancement | ```
Please let me know Is it possible to use the Elmah in windows or console based
project? If yes please provide a sample code with app.config and required code
```
Original issue reported on code.google.com by `anjali.b...@gmail.com` on 10 Sep 2012 at 10:54 | 1.0 | elmah usage example with windows based project in framework 3.5 - ```
Please let me know Is it possible to use the Elmah in windows or console based
project? If yes please provide a sample code with app.config and required code
```
Original issue reported on code.google.com by `anjali.b...@gmail.com` on 10 Sep 2012 at 10:54 | priority | elmah usage example with windows based project in framework please let me know is it possible to use the elmah in windows or console based project if yes please provide a sample code with app config and required code original issue reported on code google com by anjali b gmail com on sep at | 1 |
235,025 | 7,733,878,929 | IssuesEvent | 2018-05-26 17:05:48 | vinitkumar/googlecl | https://api.github.com/repos/vinitkumar/googlecl | closed | GoogleCL is not available through Google's linux repositories | Priority-Medium bug imported | _From [jonasfa@gmail.com](https://code.google.com/u/jonasfa@gmail.com/) on June 20, 2010 06:13:14_
http://www.google.com/linuxrepositories/
_Original issue: http://code.google.com/p/googlecl/issues/detail?id=92_
| 1.0 | GoogleCL is not available through Google's linux repositories - _From [jonasfa@gmail.com](https://code.google.com/u/jonasfa@gmail.com/) on June 20, 2010 06:13:14_
http://www.google.com/linuxrepositories/
_Original issue: http://code.google.com/p/googlecl/issues/detail?id=92_
| priority | googlecl is not available through google s linux repositories from on june original issue | 1 |
85,423 | 3,690,560,953 | IssuesEvent | 2016-02-25 20:28:35 | BCGamer/website | https://api.github.com/repos/BCGamer/website | opened | New URI: /organize/ | enhancement medium priority | To be used by event organizers for reviewing/exporting critical data.
/organize/
-Requires event organizer group
-Show a list of all conventions
-Link to list view of all events for convention
/organize/<convention-slug>/
-Requires event organizer group
-Show a list of all events
-Display .count() of registrations for event
-Link to a detail view for events
/organize/<convention-slug>/<event-slug>
-Requires event organizer group
-This view will show all registrations
-Show user's email (user.email)
-Button for export to CSV
/organize/<convention-slug>/<event-slug>/export
-Requires event organizer group
-Generate CSV of the parent event | 1.0 | New URI: /organize/ - To be used by event organizers for reviewing/exporting critical data.
/organize/
-Requires event organizer group
-Show a list of all conventions
-Link to list view of all events for convention
/organize/<convention-slug>/
-Requires event organizer group
-Show a list of all events
-Display .count() of registrations for event
-Link to a detail view for events
/organize/<convention-slug>/<event-slug>
-Requires event organizer group
-This view will show all registrations
-Show user's email (user.email)
-Button for export to CSV
/organize/<convention-slug>/<event-slug>/export
-Requires event organizer group
-Generate CSV of the parent event | priority | new uri organize to be used by event organizers for reviewing exporting critical data organize requires event organizer group show a list of all conventions link to list view of all events for convention organize requires event organizer group show a list of all events display count of registrations for event link to a detail view for events organize requires event organizer group this view will show all registrations show user s email user email button for export to csv organize export requires event organizer group generate csv of the parent event | 1 |
364,187 | 10,760,246,099 | IssuesEvent | 2019-10-31 18:11:32 | Vhoyon/Vramework | https://api.github.com/repos/Vhoyon/Vramework | opened | Add default implementation to createRequest() method in AbstractCommandRouter | !Enhancement: Framework !Priority: Medium ~owner-interactions | https://github.com/Vhoyon/Vramework/blob/91877d3525fcda00bb2036b3089b69a1c9c43526/src/main/java/io/github/vhoyon/vramework/abstracts/AbstractCommandRouter.java#L66
https://github.com/Vhoyon/Discord-Bot/blob/3f21717dfbaf6456fd5bc26211ebd2ee1a664446/src/main/java/io/github/vhoyon/bot/app/CommandRouter.java#L47-L51
The implementation in our bot is generic enough to be applied by every project by default, while allowing the override if wanted.
Porting this to our framework would let a first timer do less on initial setup.
Also, this would be backward compatible, as we're not changing the core functionnality, simply providing a default implementation - of course projects who already override it won't see a change, but they could always remove it if they don't need a custom implementation anymore! | 1.0 | Add default implementation to createRequest() method in AbstractCommandRouter - https://github.com/Vhoyon/Vramework/blob/91877d3525fcda00bb2036b3089b69a1c9c43526/src/main/java/io/github/vhoyon/vramework/abstracts/AbstractCommandRouter.java#L66
https://github.com/Vhoyon/Discord-Bot/blob/3f21717dfbaf6456fd5bc26211ebd2ee1a664446/src/main/java/io/github/vhoyon/bot/app/CommandRouter.java#L47-L51
The implementation in our bot is generic enough to be applied by every project by default, while allowing the override if wanted.
Porting this to our framework would let a first timer do less on initial setup.
Also, this would be backward compatible, as we're not changing the core functionnality, simply providing a default implementation - of course projects who already override it won't see a change, but they could always remove it if they don't need a custom implementation anymore! | priority | add default implementation to createrequest method in abstractcommandrouter the implementation in our bot is generic enough to be applied by every project by default while allowing the override if wanted porting this to our framework would let a first timer do less on initial setup also this would be backward compatible as we re not changing the core functionnality simply providing a default implementation of course projects who already override it won t see a change but they could always remove it if they don t need a custom implementation anymore | 1 |
69,384 | 3,298,154,306 | IssuesEvent | 2015-11-02 13:09:40 | ox-it/ords | https://api.github.com/repos/ox-it/ords | opened | Changing a field name breaks saved queries involving that field | David Paine Priority-Medium | If a user changes a field name in the schema designer, any saved queries that mention this field break (as one would expect).
It would be nice if (dynamic) saved queries were updated when field names were changed.
I appreciate that there are plenty of other ways of breaking saved queries by editing the schema that are much harder to deal with. | 1.0 | Changing a field name breaks saved queries involving that field - If a user changes a field name in the schema designer, any saved queries that mention this field break (as one would expect).
It would be nice if (dynamic) saved queries were updated when field names were changed.
I appreciate that there are plenty of other ways of breaking saved queries by editing the schema that are much harder to deal with. | priority | changing a field name breaks saved queries involving that field if a user changes a field name in the schema designer any saved queries that mention this field break as one would expect it would be nice if dynamic saved queries were updated when field names were changed i appreciate that there are plenty of other ways of breaking saved queries by editing the schema that are much harder to deal with | 1 |
418,753 | 12,202,971,088 | IssuesEvent | 2020-04-30 09:49:08 | kenodressel/quarantine-hero | https://api.github.com/repos/kenodressel/quarantine-hero | reopened | Improve email-verification redirect logic | priority-medium | Currently, when attempting to verify your email address, you need to click the verification link in the email, which takes you to the google auth success page. It would be better UX however if the link would redirect you to the website again after successful verification, similar to the sign-in logic for the `notify-me` component.
Also, if you verify your email address and do *not* reload the page, you need to re-login for the website to notice that your email is indeed verified. This is also bad UX and could be resolved as well by improving the redirect logic to automatically redirect the user, and reload the page automatically. | 1.0 | Improve email-verification redirect logic - Currently, when attempting to verify your email address, you need to click the verification link in the email, which takes you to the google auth success page. It would be better UX however if the link would redirect you to the website again after successful verification, similar to the sign-in logic for the `notify-me` component.
Also, if you verify your email address and do *not* reload the page, you need to re-login for the website to notice that your email is indeed verified. This is also bad UX and could be resolved as well by improving the redirect logic to automatically redirect the user, and reload the page automatically. | priority | improve email verification redirect logic currently when attempting to verify your email address you need to click the verification link in the email which takes you to the google auth success page it would be better ux however if the link would redirect you to the website again after successful verification similar to the sign in logic for the notify me component also if you verify your email address and do not reload the page you need to re login for the website to notice that your email is indeed verified this is also bad ux and could be resolved as well by improving the redirect logic to automatically redirect the user and reload the page automatically | 1 |
544,577 | 15,894,729,110 | IssuesEvent | 2021-04-11 11:21:14 | marcusolsson/grafana-json-datasource | https://api.github.com/repos/marcusolsson/grafana-json-datasource | closed | Params in "Params" is not updating when variables changes. | priority/medium type/enhancement | When using variables in the "Params" option and change the variable, the query doesn't change.
https://user-images.githubusercontent.com/39485579/112290812-df547c00-8c8f-11eb-9c18-00eb63027d2d.mp4
### Param in "Path" panel


### Param in "Params" panel


### Some additional notes
There is no query string added to the datasource setup.
Docker and the API is in a docker container.

Grafana version: 7.4.3
Plugin version: 1.0.1
| 1.0 | Params in "Params" is not updating when variables changes. - When using variables in the "Params" option and change the variable, the query doesn't change.
https://user-images.githubusercontent.com/39485579/112290812-df547c00-8c8f-11eb-9c18-00eb63027d2d.mp4
### Param in "Path" panel


### Param in "Params" panel


### Some additional notes
There is no query string added to the datasource setup.
Docker and the API is in a docker container.

Grafana version: 7.4.3
Plugin version: 1.0.1
| priority | params in params is not updating when variables changes when using variables in the params option and change the variable the query doesn t change param in path panel param in params panel some additional notes there is no query string added to the datasource setup docker and the api is in a docker container grafana version plugin version | 1 |
775,593 | 27,234,906,221 | IssuesEvent | 2023-02-21 15:40:18 | cs-utulsa/Encrypted-Chat-Service | https://api.github.com/repos/cs-utulsa/Encrypted-Chat-Service | closed | Testing: Unit Testing | coding Priority 1 Medium Effort | Goal of this story is to complete our sprint specific assignment. Each of us need to complete 3 unit tests and 1 integration test. This is required as part of this sprint by the class.
Each unit test can be small or large. Just pick a small segment of code and ensure it works as intended for each unit test. The integration test also simply needs to ensure different parts work properly together. This could be between two methods or classes. Once you have completed your tests, add writeups of each in a comment on this issue.
- [x] Conner - 3 unit
- [x] Kate - 3 unit
- [x] Declan - 3 unit
- [x] Eliga - 3 unit and 1 integration test.
- [x] James - 3 unit
- [x] Joel - 3 unit
- [x] Dawson - 3 unit
- [x] Integration Test (Unassigned)
- [x] All unit tests completed and added to the sprint doc (Conner) | 1.0 | Testing: Unit Testing - Goal of this story is to complete our sprint specific assignment. Each of us need to complete 3 unit tests and 1 integration test. This is required as part of this sprint by the class.
Each unit test can be small or large. Just pick a small segment of code and ensure it works as intended for each unit test. The integration test also simply needs to ensure different parts work properly together. This could be between two methods or classes. Once you have completed your tests, add writeups of each in a comment on this issue.
- [x] Conner - 3 unit
- [x] Kate - 3 unit
- [x] Declan - 3 unit
- [x] Eliga - 3 unit and 1 integration test.
- [x] James - 3 unit
- [x] Joel - 3 unit
- [x] Dawson - 3 unit
- [x] Integration Test (Unassigned)
- [x] All unit tests completed and added to the sprint doc (Conner) | priority | testing unit testing goal of this story is to complete our sprint specific assignment each of us need to complete unit tests and integration test this is required as part of this sprint by the class each unit test can be small or large just pick a small segment of code and ensure it works as intended for each unit test the integration test also simply needs to ensure different parts work properly together this could be between two methods or classes once you have completed your tests add writeups of each in a comment on this issue conner unit kate unit declan unit eliga unit and integration test james unit joel unit dawson unit integration test unassigned all unit tests completed and added to the sprint doc conner | 1 |
547,144 | 16,038,242,669 | IssuesEvent | 2021-04-22 02:30:40 | rubyforgood/casa | https://api.github.com/repos/rubyforgood/casa | closed | CodeClimate Maintainability Badge is broken | :sparkles: :computer: Contributor Friendly / Devel Help Wanted Priority: Medium | The CodeClimate Maintainability Badge is no longer working. It should be fixed or replaced. | 1.0 | CodeClimate Maintainability Badge is broken - The CodeClimate Maintainability Badge is no longer working. It should be fixed or replaced. | priority | codeclimate maintainability badge is broken the codeclimate maintainability badge is no longer working it should be fixed or replaced | 1 |
642,572 | 20,907,572,652 | IssuesEvent | 2022-03-24 05:07:53 | Frontesque/VueTube | https://api.github.com/repos/Frontesque/VueTube | closed | Caching pages to reduce unnecessary fetching | feature request Priority: Medium | **Is your feature request related to a problem? Please describe.**
When the view switches to subscriptions and back to home, the home page so refreshed, making it feel like a slow experience
**Describe the solution you'd like**
The contents of the homepage should be temporarily cached for the app session. | 1.0 | Caching pages to reduce unnecessary fetching - **Is your feature request related to a problem? Please describe.**
When the view switches to subscriptions and back to home, the home page so refreshed, making it feel like a slow experience
**Describe the solution you'd like**
The contents of the homepage should be temporarily cached for the app session. | priority | caching pages to reduce unnecessary fetching is your feature request related to a problem please describe when the view switches to subscriptions and back to home the home page so refreshed making it feel like a slow experience describe the solution you d like the contents of the homepage should be temporarily cached for the app session | 1 |
178,202 | 6,599,875,695 | IssuesEvent | 2017-09-17 02:49:39 | ProjectSidewalk/SidewalkWebpage | https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage | closed | Clicking outside the survey modal container dismisses it | Priority: Medium pull-request-submitted ui-bug | This is a problem that most of us would have come across before with the popups. Clicking anywhere outside the actual popup closes it. This is a problem for surveys, especially when it is mandatory to answer some of the questions. The users can just dismiss these by clicking on the background and is never shown the survey again. This happens inspite of setting the backdrop to be static. Needs to be resolved!! | 1.0 | Clicking outside the survey modal container dismisses it - This is a problem that most of us would have come across before with the popups. Clicking anywhere outside the actual popup closes it. This is a problem for surveys, especially when it is mandatory to answer some of the questions. The users can just dismiss these by clicking on the background and is never shown the survey again. This happens inspite of setting the backdrop to be static. Needs to be resolved!! | priority | clicking outside the survey modal container dismisses it this is a problem that most of us would have come across before with the popups clicking anywhere outside the actual popup closes it this is a problem for surveys especially when it is mandatory to answer some of the questions the users can just dismiss these by clicking on the background and is never shown the survey again this happens inspite of setting the backdrop to be static needs to be resolved | 1 |
114,428 | 4,630,995,261 | IssuesEvent | 2016-09-28 14:23:30 | meetinghouse/cms | https://api.github.com/repos/meetinghouse/cms | opened | LIGHT THEME: Finish Drag and Drop Admin Menu Sort Feature. | Discuss First Medium Priority | @vivek-chaudhari Vivek, we have a drag and drop Admin Menu setup in the admin on **light theme**. It works to a point, but does not include the new sub-menu capability. Can you take a look at that and let us know if you can make that work and give us a time estimate? It should work something like this: https://johnny.github.io/jquery-sortable/ Thanks, John | 1.0 | LIGHT THEME: Finish Drag and Drop Admin Menu Sort Feature. - @vivek-chaudhari Vivek, we have a drag and drop Admin Menu setup in the admin on **light theme**. It works to a point, but does not include the new sub-menu capability. Can you take a look at that and let us know if you can make that work and give us a time estimate? It should work something like this: https://johnny.github.io/jquery-sortable/ Thanks, John | priority | light theme finish drag and drop admin menu sort feature vivek chaudhari vivek we have a drag and drop admin menu setup in the admin on light theme it works to a point but does not include the new sub menu capability can you take a look at that and let us know if you can make that work and give us a time estimate it should work something like this thanks john | 1 |
452,224 | 13,047,162,125 | IssuesEvent | 2020-07-29 10:12:37 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | 404 error occurred while changing the state of the warning policy | area/console kind/bug priority/medium |
**Describe the Bug**
An error occurred while changing the state of the warning policy
<img width="925" alt="alert" src="https://user-images.githubusercontent.com/36271543/88783945-066eb280-d1c2-11ea-918b-271e86b5caf2.png">
**Versions Used**
KubeSphere:3.0.0
Kubernetes:
host-v1.15.12
member1-v1.17.9
**Environment**
host: 3node /ubuntu 16.04 4cpu/16g
member1: 2 nodes /centos7 8cpu/16g
/kind bug
/area console
/assign @leoendless
/milestone 3.0.0
/priority medium | 1.0 | 404 error occurred while changing the state of the warning policy -
**Describe the Bug**
An error occurred while changing the state of the warning policy
<img width="925" alt="alert" src="https://user-images.githubusercontent.com/36271543/88783945-066eb280-d1c2-11ea-918b-271e86b5caf2.png">
**Versions Used**
KubeSphere:3.0.0
Kubernetes:
host-v1.15.12
member1-v1.17.9
**Environment**
host: 3node /ubuntu 16.04 4cpu/16g
member1: 2 nodes /centos7 8cpu/16g
/kind bug
/area console
/assign @leoendless
/milestone 3.0.0
/priority medium | priority | error occurred while changing the state of the warning policy describe the bug an error occurred while changing the state of the warning policy img width alt alert src versions used kubesphere kubernetes host environment host ubuntu nodes kind bug area console assign leoendless milestone priority medium | 1 |
138,460 | 5,341,510,576 | IssuesEvent | 2017-02-17 03:14:08 | mmisw/mmiorr | https://api.github.com/repos/mmisw/mmiorr | closed | Supporting Authority Approval "Bumper Stickers" | Addressed_in_ORR3 enhancement imported Priority-Medium xdomes | _From [mike.e.b...@gmail.com](https://code.google.com/u/104806978225310943594/) on October 11, 2010 09:30:01_
We should somehow encourage the stated approval from authoritative bodies (IOOS, Qartoc, OGC, NMMO, NOAA, etc) of full ontologies or individual terms. Then the potential user of a term can determine how strong that term is within their community.
Not being as Ontology savvy as I need to be, I don't know the best approach, whether it be through a predefined property in a class or through mappings between terms and the authority ("approved by") using VINE, etc.
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=286_
| 1.0 | Supporting Authority Approval "Bumper Stickers" - _From [mike.e.b...@gmail.com](https://code.google.com/u/104806978225310943594/) on October 11, 2010 09:30:01_
We should somehow encourage the stated approval from authoritative bodies (IOOS, Qartoc, OGC, NMMO, NOAA, etc) of full ontologies or individual terms. Then the potential user of a term can determine how strong that term is within their community.
Not being as Ontology savvy as I need to be, I don't know the best approach, whether it be through a predefined property in a class or through mappings between terms and the authority ("approved by") using VINE, etc.
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=286_
| priority | supporting authority approval bumper stickers from on october we should somehow encourage the stated approval from authoritative bodies ioos qartoc ogc nmmo noaa etc of full ontologies or individual terms then the potential user of a term can determine how strong that term is within their community not being as ontology savvy as i need to be i don t know the best approach whether it be through a predefined property in a class or through mappings between terms and the authority approved by using vine etc original issue | 1 |
314,973 | 9,605,248,244 | IssuesEvent | 2019-05-10 23:00:41 | umple/umple | https://api.github.com/repos/umple/umple | closed | Make the E G S T D A M icons in UmpleOnline context sensitive | Component-UmpleOnline Diffic-Med Priority-Medium ucosp | The 'icons' that allow the user to control the diagram type, whether text or diagrams are displayed, and whether attributes or methods are displayed are currently not sensitive to the current state.
A lighter shade should be shown for currently inactive states. So only one of the E, G or S icons should be in normal shade. If Text is hidden, or diagram is hidden, the T or D should be in lighter shade. And the A or M would also be sensitive to the current state. By default at startup, the E, T, D and A icons would appear normal.
The initial setting of these items occurs in umple.php after line 200, but these would have to be given ids to allow control. css style button2 is set around line 111. There is a button2:active style that is the same as when hovering, but it is never being used.
Initial state is known at php startup for some of these settings, e.g. $diagramType, $showText. But the setting would also have to be controlled by javascript when the various settings change. | 1.0 | Make the E G S T D A M icons in UmpleOnline context sensitive - The 'icons' that allow the user to control the diagram type, whether text or diagrams are displayed, and whether attributes or methods are displayed are currently not sensitive to the current state.
A lighter shade should be shown for currently inactive states. So only one of the E, G or S icons should be in normal shade. If Text is hidden, or diagram is hidden, the T or D should be in lighter shade. And the A or M would also be sensitive to the current state. By default at startup, the E, T, D and A icons would appear normal.
The initial setting of these items occurs in umple.php after line 200, but these would have to be given ids to allow control. css style button2 is set around line 111. There is a button2:active style that is the same as when hovering, but it is never being used.
Initial state is known at php startup for some of these settings, e.g. $diagramType, $showText. But the setting would also have to be controlled by javascript when the various settings change. | priority | make the e g s t d a m icons in umpleonline context sensitive the icons that allow the user to control the diagram type whether text or diagrams are displayed and whether attributes or methods are displayed are currently not sensitive to the current state a lighter shade should be shown for currently inactive states so only one of the e g or s icons should be in normal shade if text is hidden or diagram is hidden the t or d should be in lighter shade and the a or m would also be sensitive to the current state by default at startup the e t d and a icons would appear normal the initial setting of these items occurs in umple php after line but these would have to be given ids to allow control css style is set around line there is a active style that is the same as when hovering but it is never being used initial state is known at php startup for some of these settings e g diagramtype showtext but the setting would also have to be controlled by javascript when the various settings change | 1 |
141,984 | 5,447,890,144 | IssuesEvent | 2017-03-07 14:42:03 | DOAJ/doaj | https://api.github.com/repos/DOAJ/doaj | closed | Add the URL shortener functionality to Share button on ToCs | medium priority ready for review tnm | As per https://github.com/DOAJ/doaj/issues/983#issuecomment-231721859, we should reproduce the URL shortener on ToCs too.
| 1.0 | Add the URL shortener functionality to Share button on ToCs - As per https://github.com/DOAJ/doaj/issues/983#issuecomment-231721859, we should reproduce the URL shortener on ToCs too.
| priority | add the url shortener functionality to share button on tocs as per we should reproduce the url shortener on tocs too | 1 |
282,244 | 8,704,739,205 | IssuesEvent | 2018-12-05 20:16:26 | Shymain/Monochrome | https://api.github.com/repos/Shymain/Monochrome | opened | Image Displaying | GUI New Feature Priority: Medium | Write code to display images in the top-left Image Panel. Support a background and foreground image being simultaneously loaded and de-coupled to each other. | 1.0 | Image Displaying - Write code to display images in the top-left Image Panel. Support a background and foreground image being simultaneously loaded and de-coupled to each other. | priority | image displaying write code to display images in the top left image panel support a background and foreground image being simultaneously loaded and de coupled to each other | 1 |
24,301 | 2,667,292,238 | IssuesEvent | 2015-03-22 13:25:24 | andresriancho/w3af | https://api.github.com/repos/andresriancho/w3af | closed | Vulnerability name titles - Assert that all names are in vulns.py | core improvement priority:medium | Assert that all names are in https://github.com/andresriancho/w3af/blob/master/w3af/core/data/constants/vulns.py , when during tests write the names that are NOT in the vulns.py in a text file, assert that the file is empty at the end of the whole test run | 1.0 | Vulnerability name titles - Assert that all names are in vulns.py - Assert that all names are in https://github.com/andresriancho/w3af/blob/master/w3af/core/data/constants/vulns.py , when during tests write the names that are NOT in the vulns.py in a text file, assert that the file is empty at the end of the whole test run | priority | vulnerability name titles assert that all names are in vulns py assert that all names are in when during tests write the names that are not in the vulns py in a text file assert that the file is empty at the end of the whole test run | 1 |
733,046 | 25,285,640,305 | IssuesEvent | 2022-11-16 19:03:27 | svthalia/concrexit | https://api.github.com/repos/svthalia/concrexit | closed | Calendar wrongly shows registered for an event. | priority: medium bug | ### Describe the bug
I was registered for an event, and then deregistered, however the calendar still shows "you are registered for this event". This includes the event being magenta. When going to the event page it does not show as registered.
### How to reproduce
Steps to reproduce the behaviour:
1. Register for an event
2. Deregister
3. Go to calendar
4. See magenta event that you are not registered for
### Expected behaviour
The event should not show as registered
### Screenshots


### Additional context
I am sorry for breaking the 2^6 issues. Clearly the only solution to getting back to such a good number is to fix this issue ASAP | 1.0 | Calendar wrongly shows registered for an event. - ### Describe the bug
I was registered for an event, and then deregistered, however the calendar still shows "you are registered for this event". This includes the event being magenta. When going to the event page it does not show as registered.
### How to reproduce
Steps to reproduce the behaviour:
1. Register for an event
2. Deregister
3. Go to calendar
4. See magenta event that you are not registered for
### Expected behaviour
The event should not show as registered
### Screenshots


### Additional context
I am sorry for breaking the 2^6 issues. Clearly the only solution to getting back to such a good number is to fix this issue ASAP | priority | calendar wrongly shows registered for an event describe the bug i was registered for an event and then deregistered however the calendar still shows you are registered for this event this includes the event being magenta when going to the event page it does not show as registered how to reproduce steps to reproduce the behaviour register for an event deregister go to calendar see magenta event that you are not registered for expected behaviour the event should not show as registered screenshots additional context i am sorry for breaking the issues clearly the only solution to getting back to such a good number is to fix this issue asap | 1 |
25,137 | 2,676,643,921 | IssuesEvent | 2015-03-25 18:46:01 | dump247/udt-net | https://api.github.com/repos/dump247/udt-net | closed | Needs help on TcpClient and TcpListener | auto-migrated Priority-Medium Type-Enhancement | ```
As mentioned in ToDo list "Something similar to TcpClient and TcpListener", are
you working on this ? If yes when can we expect it ?
```
Original issue reported on code.google.com by `ppdevm...@gmail.com` on 27 Nov 2012 at 4:25 | 1.0 | Needs help on TcpClient and TcpListener - ```
As mentioned in ToDo list "Something similar to TcpClient and TcpListener", are
you working on this ? If yes when can we expect it ?
```
Original issue reported on code.google.com by `ppdevm...@gmail.com` on 27 Nov 2012 at 4:25 | priority | needs help on tcpclient and tcplistener as mentioned in todo list something similar to tcpclient and tcplistener are you working on this if yes when can we expect it original issue reported on code google com by ppdevm gmail com on nov at | 1 |
188,122 | 6,772,906,755 | IssuesEvent | 2017-10-27 02:02:22 | CS2103AUG2017-F10-B1/main | https://api.github.com/repos/CS2103AUG2017-F10-B1/main | closed | Improve UI for user input | priority.medium status.ongoing type.enhancement type.task | Currently, although user inputs and commands such as `add` allow for any ordering (e.g. you can `add n/[name] a/[address] e/[email]` or `add n/[name] e/[email] a/[address]`), it is still not very intuitive and user friendly.
Look into improving the UI for user input so that it becomes more intuitive. | 1.0 | Improve UI for user input - Currently, although user inputs and commands such as `add` allow for any ordering (e.g. you can `add n/[name] a/[address] e/[email]` or `add n/[name] e/[email] a/[address]`), it is still not very intuitive and user friendly.
Look into improving the UI for user input so that it becomes more intuitive. | priority | improve ui for user input currently although user inputs and commands such as add allow for any ordering e g you can add n a e or add n e a it is still not very intuitive and user friendly look into improving the ui for user input so that it becomes more intuitive | 1 |
216,985 | 7,313,602,363 | IssuesEvent | 2018-03-01 02:01:18 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Markers don't handle world wrapping | Medium Priority | 
The marker should be pointing at the cart (across the world border) | 1.0 | Markers don't handle world wrapping - 
The marker should be pointing at the cart (across the world border) | priority | markers don t handle world wrapping the marker should be pointing at the cart across the world border | 1 |
671,517 | 22,764,370,706 | IssuesEvent | 2022-07-08 01:45:01 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | closed | Route Table of Subnet gets cleared when using a AKS | medium_priority not a bug | <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
Recreating a Azure Subnet flushes the route_table with the entries genereated by an AKS
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_subnet.py
##### ANSIBLE VERSION
ansible 2.10.7
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible
python version = 3.9.4 (default, Apr 10 2021, 15:39:19) [GCC 8.3.0]
##### COLLECTION VERSION
1.12.0
##### STEPS TO REPRODUCE
1. create a Subnet
2. create a AKS that uses the Subnet
3. run step 1 again and it flushed the `route_table`
##### EXPECTED RESULTS
As ansible tasks should be idempotent and the AKS should still work when running the steps above again and again
##### ACTUAL RESULTS
AKS does not work anymore when running a second time (internal networking does not function) and we have to restart to whole cluster to make it work again.
The problem seems to be that the AKS-Module also automatically creates a routing_table. but when we rerun the "create subnet" step it doesn't know that there is already routing_table assigned (and therefore flushes it. this problem seems to be caused by this change in the subnet module:
###### 1.10.0
```python
if self.route_table is not None and self.route_table != results['route_table'].get('id'):
changed = True
results['route_table']['id'] = self.route_table
self.log("CHANGED: subnet {0} route_table to {1}".format(self.name, route_table.get('name')))
```
###### 1.11.0
```python
if self.route_table is not None:
if self.route_table != results['route_table'].get('id'):
changed = True
results['route_table']['id'] = self.route_table
self.log("CHANGED: subnet {0} route_table to {1}".format(self.name, route_table.get('name')))
else:
if results['route_table'].get('id') is not None:
changed = True
results['route_table']['id'] = None
self.log("CHANGED: subnet {0} will dissociate to route_table {1}".format(self.name, route_table.get('name')))
```
I also created a diff between the outputs of version 1.4.0 (left) and 1.12.0 (right)

as you can see the old version keeps the route_table and the new one resets it. after this the name resolution in the AKS does not work anymore and we have to restart it. | 1.0 | Route Table of Subnet gets cleared when using a AKS - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
Recreating a Azure Subnet flushes the route_table with the entries genereated by an AKS
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_subnet.py
##### ANSIBLE VERSION
ansible 2.10.7
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible
python version = 3.9.4 (default, Apr 10 2021, 15:39:19) [GCC 8.3.0]
##### COLLECTION VERSION
1.12.0
##### STEPS TO REPRODUCE
1. create a Subnet
2. create a AKS that uses the Subnet
3. run step 1 again and it flushed the `route_table`
##### EXPECTED RESULTS
As ansible tasks should be idempotent and the AKS should still work when running the steps above again and again
##### ACTUAL RESULTS
AKS does not work anymore when running a second time (internal networking does not function) and we have to restart to whole cluster to make it work again.
The problem seems to be that the AKS-Module also automatically creates a routing_table. but when we rerun the "create subnet" step it doesn't know that there is already routing_table assigned (and therefore flushes it. this problem seems to be caused by this change in the subnet module:
###### 1.10.0
```python
if self.route_table is not None and self.route_table != results['route_table'].get('id'):
changed = True
results['route_table']['id'] = self.route_table
self.log("CHANGED: subnet {0} route_table to {1}".format(self.name, route_table.get('name')))
```
###### 1.11.0
```python
if self.route_table is not None:
if self.route_table != results['route_table'].get('id'):
changed = True
results['route_table']['id'] = self.route_table
self.log("CHANGED: subnet {0} route_table to {1}".format(self.name, route_table.get('name')))
else:
if results['route_table'].get('id') is not None:
changed = True
results['route_table']['id'] = None
self.log("CHANGED: subnet {0} will dissociate to route_table {1}".format(self.name, route_table.get('name')))
```
I also created a diff between the outputs of version 1.4.0 (left) and 1.12.0 (right)

as you can see the old version keeps the route_table and the new one resets it. after this the name resolution in the AKS does not work anymore and we have to restart it. | priority | route table of subnet gets cleared when using a aks summary recreating a azure subnet flushes the route table with the entries genereated by an aks issue type bug report component name azure rm subnet py ansible version ansible config file none configured module search path ansible python module location usr local lib site packages ansible executable location usr local bin ansible python version default apr collection version steps to reproduce create a subnet create a aks that uses the subnet run step again and it flushed the route table expected results as ansible tasks should be idempotent and the aks should still work when running the steps above again and again actual results aks does not work anymore when running a second time internal networking does not function and we have to restart to whole cluster to make it work again the problem seems to be that the aks module also automatically creates a routing table but when we rerun the create subnet step it doesn t know that there is already routing table assigned and therefore flushes it this problem seems to be caused by this change in the subnet module python if self route table is not none and self route table results get id changed true results self route table self log changed subnet route table to format self name route table get name python if self route table is not none if self route table results get id changed true results self route table self log changed subnet route table to format self name route table get name else if results get id is not none changed true results none self log changed subnet will dissociate to route table format self name route table get name i also created a diff between the outputs of version left and right as you can see the old version keeps the route table and the new one resets it after this the name resolution in the aks does not work anymore and we have to restart it | 1 |
3,146 | 2,537,382,260 | IssuesEvent | 2015-01-26 20:11:55 | web2py/web2py | https://api.github.com/repos/web2py/web2py | opened | Update legacy table using primarykey failed which type is not string failed | 1 star bug imported Priority-Medium | _From [ochib...@gmail.com](https://code.google.com/u/107398878481677837585/) on February 10, 2012 16:29:16_
It will failed with the follwing error. It works fine if the field type is string in web2py and varchar in MSSQL.
From the syntax error, it's ignore the
'integer' from define_table and always tried with string...
TRACEBACK
\-----------------------------
SyntaxError: user is tampering with form's record_id: {'mykey':
'111'} != {'mykey': 111}
What steps will reproduce the problem? (Tested with MSSQL Server)
1. Create table
CREATE TABLE [dbo].[Test](
[mykey] [int] NOT NULL,
[myvalue] [varchar](50) NULL,
CONSTRAINT [PK_Test] PRIMARY KEY CLUSTERED
(
[mykey] ASC
)WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, IGNORE_DUP_KEY
= OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON) ON [PRIMARY]
) ON [PRIMARY]
2. Add record
mykey = 111
myvalue = 'AAA'
3. Model - define table
db.define_table('Test',
Field('mykey','integer'),
Field('myvalue'),
primarykey=['mykey'])
4. Contoller
def index():
form=crud.update(db.Test,db.Test(db.Test.mykey==111))
return dict(form=form)
5. View (default/index)
{{extend 'layout.html'}}
{{=form}} What is the expected output? What do you see instead? No error What version of the product are you using? On what operating system? Version 1.99.4 (2011-12-14 14:46:14) stable Please provide any additional information below. Forum link https://groups.google.com/group/web2py/browse_thread/thread/9df8d6b842594735/30bb082a5d5443e6#30bb082a5d5443e6
_Original issue: http://code.google.com/p/web2py/issues/detail?id=656_ | 1.0 | Update legacy table using primarykey failed which type is not string failed - _From [ochib...@gmail.com](https://code.google.com/u/107398878481677837585/) on February 10, 2012 16:29:16_
It will failed with the follwing error. It works fine if the field type is string in web2py and varchar in MSSQL.
From the syntax error, it's ignore the
'integer' from define_table and always tried with string...
TRACEBACK
\-----------------------------
SyntaxError: user is tampering with form's record_id: {'mykey':
'111'} != {'mykey': 111}
What steps will reproduce the problem? (Tested with MSSQL Server)
1. Create table
CREATE TABLE [dbo].[Test](
[mykey] [int] NOT NULL,
[myvalue] [varchar](50) NULL,
CONSTRAINT [PK_Test] PRIMARY KEY CLUSTERED
(
[mykey] ASC
)WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, IGNORE_DUP_KEY
= OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON) ON [PRIMARY]
) ON [PRIMARY]
2. Add record
mykey = 111
myvalue = 'AAA'
3. Model - define table
db.define_table('Test',
Field('mykey','integer'),
Field('myvalue'),
primarykey=['mykey'])
4. Contoller
def index():
form=crud.update(db.Test,db.Test(db.Test.mykey==111))
return dict(form=form)
5. View (default/index)
{{extend 'layout.html'}}
{{=form}} What is the expected output? What do you see instead? No error What version of the product are you using? On what operating system? Version 1.99.4 (2011-12-14 14:46:14) stable Please provide any additional information below. Forum link https://groups.google.com/group/web2py/browse_thread/thread/9df8d6b842594735/30bb082a5d5443e6#30bb082a5d5443e6
_Original issue: http://code.google.com/p/web2py/issues/detail?id=656_ | priority | update legacy table using primarykey failed which type is not string failed from on february it will failed with the follwing error it works fine if the field type is string in and varchar in mssql from the syntax error it s ignore the integer from define table and always tried with string traceback syntaxerror user is tampering with form s record id mykey mykey what steps will reproduce the problem tested with mssql server create table create table not null null constraint primary key clustered asc with pad index off statistics norecompute off ignore dup key off allow row locks on allow page locks on on on add record mykey myvalue aaa model define table db define table test field mykey integer field myvalue primarykey contoller def index form crud update db test db test db test mykey return dict form form view default index extend layout html form what is the expected output what do you see instead no error what version of the product are you using on what operating system version stable please provide any additional information below forum link original issue | 1 |
611,608 | 18,959,496,755 | IssuesEvent | 2021-11-19 01:41:05 | airshipit/airshipctl | https://api.github.com/repos/airshipit/airshipctl | closed | Generic container timeout validation & enforcement | enhancement priority/medium size l | Building on #544, provide a compliance mechanism to validate the timeout has been acknowledged & action has been taken.
| 1.0 | Generic container timeout validation & enforcement - Building on #544, provide a compliance mechanism to validate the timeout has been acknowledged & action has been taken.
| priority | generic container timeout validation enforcement building on provide a compliance mechanism to validate the timeout has been acknowledged action has been taken | 1 |
60,433 | 3,127,057,892 | IssuesEvent | 2015-09-08 13:04:36 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | Relevance sorting on search seems odd when searching for CODs | CODmigration Priority-Medium | The following search "afghanistan common operational dataset" (on stag) does return the desired dataset, but only as the 6th entry (Afghanistan Population Stats). That dataset contains a tag "common operational dataset".

Is it because the title is weighted higher? Maybe we could tweak this a bit to make it easier to find the CODs.
| 1.0 | Relevance sorting on search seems odd when searching for CODs - The following search "afghanistan common operational dataset" (on stag) does return the desired dataset, but only as the 6th entry (Afghanistan Population Stats). That dataset contains a tag "common operational dataset".

Is it because the title is weighted higher? Maybe we could tweak this a bit to make it easier to find the CODs.
| priority | relevance sorting on search seems odd when searching for cods the following search afghanistan common operational dataset on stag does return the desired dataset but only as the entry afghanistan population stats that dataset contains a tag common operational dataset is it because the title is weighted higher maybe we could tweak this a bit to make it easier to find the cods | 1 |
532,166 | 15,530,832,399 | IssuesEvent | 2021-03-13 20:44:43 | MessageKit/MessageKit | https://api.github.com/repos/MessageKit/MessageKit | closed | Scroll problem on new messages with keyboard open | confirmed bug priority: medium | **Describe the bug**
When the keyboard is open with not enough message to need scroll, if message keep arriving or you add new messages without dismissing the keyboard, the UICollectionView scroll is bugged and thinks that the UICollectionView keeps occupying the whole screen.
**To Reproduce**
Steps/code to reproduce the behavior:
I comment your loadFirstMessages function on your ChatViewController code
```
override func viewDidLoad() {
super.viewDidLoad()
configureMessageCollectionView()
configureMessageInputBar()
// loadFirstMessages()
title = "MessageKit"
}
```
and I wait until messages keep arriving with your MockSocket, once enough message has arrived to need scroll it can be seen that the scroll container it's badly calculated.
**Expected behavior**
Content scrollable in the space visible not under the keyboard.
**Screenshots**
https://user-images.githubusercontent.com/15628852/105724803-8931b980-5f28-11eb-8764-b10adf32c29b.mov
**Environment**
- What version of MessageKit are you using? 3.5.0
- What version of iOS are you running on? 14.3
- What version of Swift are you running on? Swift 5
- What device(s) are you testing on? Are these simulators? iPhone X (physical device), iPhone SE (physical device), and iPhone 12 Pro (simulator)
- Is the issue you're experiencing reproducible in the example app? Yes
**Additional context**
Add any other context about the problem here.
In the testing app, it's failing with pod and in my application with SPM
| 1.0 | Scroll problem on new messages with keyboard open - **Describe the bug**
When the keyboard is open with not enough message to need scroll, if message keep arriving or you add new messages without dismissing the keyboard, the UICollectionView scroll is bugged and thinks that the UICollectionView keeps occupying the whole screen.
**To Reproduce**
Steps/code to reproduce the behavior:
I comment your loadFirstMessages function on your ChatViewController code
```
override func viewDidLoad() {
super.viewDidLoad()
configureMessageCollectionView()
configureMessageInputBar()
// loadFirstMessages()
title = "MessageKit"
}
```
and I wait until messages keep arriving with your MockSocket, once enough message has arrived to need scroll it can be seen that the scroll container it's badly calculated.
**Expected behavior**
Content scrollable in the space visible not under the keyboard.
**Screenshots**
https://user-images.githubusercontent.com/15628852/105724803-8931b980-5f28-11eb-8764-b10adf32c29b.mov
**Environment**
- What version of MessageKit are you using? 3.5.0
- What version of iOS are you running on? 14.3
- What version of Swift are you running on? Swift 5
- What device(s) are you testing on? Are these simulators? iPhone X (physical device), iPhone SE (physical device), and iPhone 12 Pro (simulator)
- Is the issue you're experiencing reproducible in the example app? Yes
**Additional context**
Add any other context about the problem here.
In the testing app, it's failing with pod and in my application with SPM
| priority | scroll problem on new messages with keyboard open describe the bug when the keyboard is open with not enough message to need scroll if message keep arriving or you add new messages without dismissing the keyboard the uicollectionview scroll is bugged and thinks that the uicollectionview keeps occupying the whole screen to reproduce steps code to reproduce the behavior i comment your loadfirstmessages function on your chatviewcontroller code override func viewdidload super viewdidload configuremessagecollectionview configuremessageinputbar loadfirstmessages title messagekit and i wait until messages keep arriving with your mocksocket once enough message has arrived to need scroll it can be seen that the scroll container it s badly calculated expected behavior content scrollable in the space visible not under the keyboard screenshots environment what version of messagekit are you using what version of ios are you running on what version of swift are you running on swift what device s are you testing on are these simulators iphone x physical device iphone se physical device and iphone pro simulator is the issue you re experiencing reproducible in the example app yes additional context add any other context about the problem here in the testing app it s failing with pod and in my application with spm | 1 |
372,470 | 11,015,261,306 | IssuesEvent | 2019-12-05 01:01:27 | medic/cht-core | https://api.github.com/repos/medic/cht-core | closed | Deleting a report in mobile view doesn't redirect to list | Priority: 2 - Medium Type: Bug | **Describe the bug**
When deleting a report in mobile view, post deletion, instead of being redirected to the reports list, you end up on an empty "report-content" page.
**To Reproduce**
Steps to reproduce the behavior:
1. Either use a phone or use a desktop browser to simulate a mobile (narrow) view
2. Go to Reports and open a Report
3. Delete the report
4. Notice you're not redirected to the reports list, and you need to click the top "back arrow" to view the reports list again.
**Expected behavior**
You should be redirected to the reports list.
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Environment**
- Instance: local
- Browser: any
- App: webapp
- Version: 3.8.0
| 1.0 | Deleting a report in mobile view doesn't redirect to list - **Describe the bug**
When deleting a report in mobile view, post deletion, instead of being redirected to the reports list, you end up on an empty "report-content" page.
**To Reproduce**
Steps to reproduce the behavior:
1. Either use a phone or use a desktop browser to simulate a mobile (narrow) view
2. Go to Reports and open a Report
3. Delete the report
4. Notice you're not redirected to the reports list, and you need to click the top "back arrow" to view the reports list again.
**Expected behavior**
You should be redirected to the reports list.
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Environment**
- Instance: local
- Browser: any
- App: webapp
- Version: 3.8.0
| priority | deleting a report in mobile view doesn t redirect to list describe the bug when deleting a report in mobile view post deletion instead of being redirected to the reports list you end up on an empty report content page to reproduce steps to reproduce the behavior either use a phone or use a desktop browser to simulate a mobile narrow view go to reports and open a report delete the report notice you re not redirected to the reports list and you need to click the top back arrow to view the reports list again expected behavior you should be redirected to the reports list screenshots if applicable add screenshots to help explain your problem environment instance local browser any app webapp version | 1 |
204,476 | 7,088,052,862 | IssuesEvent | 2018-01-11 20:02:44 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | A4NH - Project Managing Partners - Question about who can remove them | Priority - Medium Specificity-A4NH Type - Bug Type - Question | >**Amanda W:** It looks like only the CRP Admin/PMU can remove Managing Partners once they've been assigned to a project. Is this true? I just want to make sure it's not a bug. In this case, the Managing Partner did not have any funding sources/budget or deliverables associated with them in this project. | 1.0 | A4NH - Project Managing Partners - Question about who can remove them - >**Amanda W:** It looks like only the CRP Admin/PMU can remove Managing Partners once they've been assigned to a project. Is this true? I just want to make sure it's not a bug. In this case, the Managing Partner did not have any funding sources/budget or deliverables associated with them in this project. | priority | project managing partners question about who can remove them amanda w it looks like only the crp admin pmu can remove managing partners once they ve been assigned to a project is this true i just want to make sure it s not a bug in this case the managing partner did not have any funding sources budget or deliverables associated with them in this project | 1 |
527,154 | 15,309,156,905 | IssuesEvent | 2021-02-24 23:51:01 | returntocorp/semgrep | https://api.github.com/repos/returntocorp/semgrep | closed | Ignore typescript .d.ts files | good first issue lang:typescript priority:medium stale | When semgrep scans for typescript files, it looks for `*.ts`, which includes files ending in `.d.ts` which are [typed interfaces for javascript code](https://www.typescriptlang.org/docs/handbook/declaration-files/templates/module-d-ts.html). Our parser currently doesn't support them, resulting in error messages that don't benefit the user.
Suggestion: filter out files with the `.d.ts` suffix in addition to filtering out files that don't end in `.ts`.
| 1.0 | Ignore typescript .d.ts files - When semgrep scans for typescript files, it looks for `*.ts`, which includes files ending in `.d.ts` which are [typed interfaces for javascript code](https://www.typescriptlang.org/docs/handbook/declaration-files/templates/module-d-ts.html). Our parser currently doesn't support them, resulting in error messages that don't benefit the user.
Suggestion: filter out files with the `.d.ts` suffix in addition to filtering out files that don't end in `.ts`.
| priority | ignore typescript d ts files when semgrep scans for typescript files it looks for ts which includes files ending in d ts which are our parser currently doesn t support them resulting in error messages that don t benefit the user suggestion filter out files with the d ts suffix in addition to filtering out files that don t end in ts | 1 |
77,197 | 3,506,271,004 | IssuesEvent | 2016-01-08 05:10:11 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Error compil on Ubuntu (BB #259) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 08.08.2010 19:00:17 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/259
<hr>
make[2]: *** [src/shared/vmap/CMakeFiles/vmaps.dir/VMapManager2.cpp.o] Error 1
make[1]: *** [src/shared/vmap/CMakeFiles/vmaps.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs..
its a big problem... | 1.0 | Error compil on Ubuntu (BB #259) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 08.08.2010 19:00:17 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/259
<hr>
make[2]: *** [src/shared/vmap/CMakeFiles/vmaps.dir/VMapManager2.cpp.o] Error 1
make[1]: *** [src/shared/vmap/CMakeFiles/vmaps.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs..
its a big problem... | priority | error compil on ubuntu bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state resolved direct link make error make error make waiting for unfinished jobs its a big problem | 1 |
739,310 | 25,590,985,188 | IssuesEvent | 2022-12-01 13:00:34 | eth-cscs/DLA-Future | https://api.github.com/repos/eth-cscs/DLA-Future | closed | `transform` tries to instantiate callables without a stream argument | bug Task Priority:Medium | #679 added a workaround to not use `unwrapping` with nullary callables. The workaround introduced another problem when the callable is nullary only for CPU and unary for GPU. The `is_invocable_v<F>` check (in https://github.com/eth-cscs/DLA-Future/blob/c1720b1d43f2beeb42dac4784d7c357e0e08fa3f/include/dlaf/sender/transform.h#L99) attempts to instantiate the callable with zero arguments also on the GPU path. If the callable is unconstrained this will lead to compilation errors. An example of a call that will currently fail:
```
transform(policy, [](auto&&... ts) {
if constexpr (device == Device::CPU) {
f_cpu();
} else {
f_gpu(ts...); // instantiated with ts as an empty pack by is_invocable_v<F>
}
})
```
This should be fixed at the same time as #681, but if a workaround is needed one can write the lambda as something like this:
```
if constexpr (device == Device::CPU) {
f_cpu();
} else if constexpr (sizeof...(ts) == 1) {
f_gpu(ts...);
} // ignore the other cases
``` | 1.0 | `transform` tries to instantiate callables without a stream argument - #679 added a workaround to not use `unwrapping` with nullary callables. The workaround introduced another problem when the callable is nullary only for CPU and unary for GPU. The `is_invocable_v<F>` check (in https://github.com/eth-cscs/DLA-Future/blob/c1720b1d43f2beeb42dac4784d7c357e0e08fa3f/include/dlaf/sender/transform.h#L99) attempts to instantiate the callable with zero arguments also on the GPU path. If the callable is unconstrained this will lead to compilation errors. An example of a call that will currently fail:
```
transform(policy, [](auto&&... ts) {
if constexpr (device == Device::CPU) {
f_cpu();
} else {
f_gpu(ts...); // instantiated with ts as an empty pack by is_invocable_v<F>
}
})
```
This should be fixed at the same time as #681, but if a workaround is needed one can write the lambda as something like this:
```
if constexpr (device == Device::CPU) {
f_cpu();
} else if constexpr (sizeof...(ts) == 1) {
f_gpu(ts...);
} // ignore the other cases
``` | priority | transform tries to instantiate callables without a stream argument added a workaround to not use unwrapping with nullary callables the workaround introduced another problem when the callable is nullary only for cpu and unary for gpu the is invocable v check in attempts to instantiate the callable with zero arguments also on the gpu path if the callable is unconstrained this will lead to compilation errors an example of a call that will currently fail transform policy auto ts if constexpr device device cpu f cpu else f gpu ts instantiated with ts as an empty pack by is invocable v this should be fixed at the same time as but if a workaround is needed one can write the lambda as something like this if constexpr device device cpu f cpu else if constexpr sizeof ts f gpu ts ignore the other cases | 1 |
388,282 | 11,485,850,613 | IssuesEvent | 2020-02-11 08:43:27 | DigitalCampus/moodle-block_oppia_mobile_export | https://api.github.com/repos/DigitalCampus/moodle-block_oppia_mobile_export | closed | Remove code related to export2print? | medium priority question | Not sure if this option is ever actually used. Even if the underlying code is not removed, the link/options in the block interface could be removed? | 1.0 | Remove code related to export2print? - Not sure if this option is ever actually used. Even if the underlying code is not removed, the link/options in the block interface could be removed? | priority | remove code related to not sure if this option is ever actually used even if the underlying code is not removed the link options in the block interface could be removed | 1 |
441,006 | 12,707,015,093 | IssuesEvent | 2020-06-23 08:13:07 | luna/luna | https://api.github.com/repos/luna/luna | closed | Runtime Architecture Overview | Category: Compiler Category: Runtime Category: Semantics Category: Type-System Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement | ### Summary
Before any development on the runtime can actually take place, an overview of its architecture must be constructed. This will consist of both diagrammatic overviews and textual (design) overviews.
### Value
Having a comprehensive architecture overview of the runtime design will both enable the team to check for flaws in its architecture and create an overarching plan for its implementation.
### Specification
- Creation of a diagrammatic overview of the architecture for the new runtime.
- Creation of detailed textual descriptions to describe each portion of the diagram.
### Acceptance Criteria & Test Cases
- [ ] A diagrammatic overview exists.
- [ ] It is accompanied by textual descriptions.
| 1.0 | Runtime Architecture Overview - ### Summary
Before any development on the runtime can actually take place, an overview of its architecture must be constructed. This will consist of both diagrammatic overviews and textual (design) overviews.
### Value
Having a comprehensive architecture overview of the runtime design will both enable the team to check for flaws in its architecture and create an overarching plan for its implementation.
### Specification
- Creation of a diagrammatic overview of the architecture for the new runtime.
- Creation of detailed textual descriptions to describe each portion of the diagram.
### Acceptance Criteria & Test Cases
- [ ] A diagrammatic overview exists.
- [ ] It is accompanied by textual descriptions.
| priority | runtime architecture overview summary before any development on the runtime can actually take place an overview of its architecture must be constructed this will consist of both diagrammatic overviews and textual design overviews value having a comprehensive architecture overview of the runtime design will both enable the team to check for flaws in its architecture and create an overarching plan for its implementation specification creation of a diagrammatic overview of the architecture for the new runtime creation of detailed textual descriptions to describe each portion of the diagram acceptance criteria test cases a diagrammatic overview exists it is accompanied by textual descriptions | 1 |
662,680 | 22,149,338,531 | IssuesEvent | 2022-06-03 15:10:28 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | UBSAN error in AlterTableWithConcurrentTxnTest.TServerLeaderChange | kind/bug area/docdb priority/medium | Jira Link: [DB-584](https://yugabyte.atlassian.net/browse/DB-584)
### Description
https://gist.githubusercontent.com/mbautin/f861305bf8ab284b2918c3cbc3e19c54/raw | 1.0 | UBSAN error in AlterTableWithConcurrentTxnTest.TServerLeaderChange - Jira Link: [DB-584](https://yugabyte.atlassian.net/browse/DB-584)
### Description
https://gist.githubusercontent.com/mbautin/f861305bf8ab284b2918c3cbc3e19c54/raw | priority | ubsan error in altertablewithconcurrenttxntest tserverleaderchange jira link description | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.