Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
94,445 | 3,926,005,312 | IssuesEvent | 2016-04-22 21:17:41 | CCAFS/ccafs-ap | https://api.github.com/repos/CCAFS/ccafs-ap | closed | Create the CRP indicators section | auto-migrated enhancement Priority-Medium Section-General | ```
This section will be used by the coordinating unit users
```
Original issue reported on code.google.com by `carvajal.hernandavid@gmail.com` on 15 Apr 2014 at 7:39 | 1.0 | Create the CRP indicators section - ```
This section will be used by the coordinating unit users
```
Original issue reported on code.google.com by `carvajal.hernandavid@gmail.com` on 15 Apr 2014 at 7:39 | priority | create the crp indicators section this section will be used by the coordinating unit users original issue reported on code google com by carvajal hernandavid gmail com on apr at | 1 |
746,452 | 26,030,707,075 | IssuesEvent | 2022-12-21 20:55:18 | bounswe/bounswe2022group8 | https://api.github.com/repos/bounswe/bounswe2022group8 | opened | BE-39: Backend Unit Tests | Effort: Medium Priority: Medium Status: In Progress Coding Team: Backend | ### What's up?
We have already implemented some unit tests, which can be found [here](https://github.com/bounswe/bounswe2022group8/tree/master/App/backend/api/tests), to test our helper functions and models. Within the scope of this task, my aim is to enrich the content of our unit tests to cover all the API endpoints I've implemented so far.
### To Do
- [ ] Implement unit tests using `django.test` and `faker` (if appropriate), that covers the following APIs. (`RequestFactory` can be used to generate requests):
- [ ] Signup/login/logout APIs
- [ ] Follow/unfollow APIs
- [ ] Artitem related APIs
- [ ] Profile APIs
- [ ] Exhibition related APIs (Virtual exhibitions + offline exhibitions)
- [ ] Annotation related APIs (Image annotation + Text annotation)
Keep in mind that each unit test must have a unique id to refer to in test traceability matrix.
### Deadline
27.12.2022 @23.59
### Additional Information
_No response_
### Reviewers
@BElifb | 1.0 | BE-39: Backend Unit Tests - ### What's up?
We have already implemented some unit tests, which can be found [here](https://github.com/bounswe/bounswe2022group8/tree/master/App/backend/api/tests), to test our helper functions and models. Within the scope of this task, my aim is to enrich the content of our unit tests to cover all the API endpoints I've implemented so far.
### To Do
- [ ] Implement unit tests using `django.test` and `faker` (if appropriate), that covers the following APIs. (`RequestFactory` can be used to generate requests):
- [ ] Signup/login/logout APIs
- [ ] Follow/unfollow APIs
- [ ] Artitem related APIs
- [ ] Profile APIs
- [ ] Exhibition related APIs (Virtual exhibitions + offline exhibitions)
- [ ] Annotation related APIs (Image annotation + Text annotation)
Keep in mind that each unit test must have a unique id to refer to in test traceability matrix.
### Deadline
27.12.2022 @23.59
### Additional Information
_No response_
### Reviewers
@BElifb | priority | be backend unit tests what s up we have already implemented some unit tests which can be found to test our helper functions and models within the scope of this task my aim is to enrich the content of our unit tests to cover all the api endpoints i ve implemented so far to do implement unit tests using django test and faker if appropriate that covers the following apis requestfactory can be used to generate requests signup login logout apis follow unfollow apis artitem related apis profile apis exhibition related apis virtual exhibitions offline exhibitions annotation related apis image annotation text annotation keep in mind that each unit test must have a unique id to refer to in test traceability matrix deadline additional information no response reviewers belifb | 1 |
487,656 | 14,049,933,750 | IssuesEvent | 2020-11-02 10:59:05 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | Widgets: Improve legend presentation for pie chart | Good first issue Priority: Medium Widgets enhancement investigation | # Description
The legend provided by recharts for pie charts is ugly.

It should be aligned on a side.
# Acceptance criteria AC
- The legend should be aligned on one side for pie charts
# Implementation notes
This is a sample with legend aligned to the right.
https://jsfiddle.net/offtherailz/ducnsveo/1/
WIP on this branch: https://github.com/offtherailz/MapStore2/tree/pie_legend_wip
Needings (See the branch `pie_legend_wip`):
- Update recharts to the latest version
- Fix issues due to update (I noticed now the grid of the chart is above the chart, not below, and mouse hover on bar charts hides the bar)
- Auto positioning of legend OR advanced option that defines legend position and size
- Responsive layout with every widget size | 1.0 | Widgets: Improve legend presentation for pie chart - # Description
The legend provided by recharts for pie charts is ugly.

It should be aligned on a side.
# Acceptance criteria AC
- The legend should be aligned on one side for pie charts
# Implementation notes
This is a sample with legend aligned to the right.
https://jsfiddle.net/offtherailz/ducnsveo/1/
WIP on this branch: https://github.com/offtherailz/MapStore2/tree/pie_legend_wip
Needings (See the branch `pie_legend_wip`):
- Update recharts to the latest version
- Fix issues due to update (I noticed now the grid of the chart is above the chart, not below, and mouse hover on bar charts hides the bar)
- Auto positioning of legend OR advanced option that defines legend position and size
- Responsive layout with every widget size | priority | widgets improve legend presentation for pie chart description the legend provided by recharts for pie charts is ugly it should be aligned on a side acceptance criteria ac the legend should be aligned on one side for pie charts implementation notes this is a sample with legend aligned to the right wip on this branch needings see the branch pie legend wip update recharts to the latest version fix issues due to update i noticed now the grid of the chart is above the chart not below and mouse hover on bar charts hides the bar auto positioning of legend or advanced option that defines legend position and size responsive layout with every widget size | 1 |
576,314 | 17,084,063,402 | IssuesEvent | 2021-07-08 09:29:57 | kleros/court | https://api.github.com/repos/kleros/court | closed | Maintenance on 'Where to Buy' | Priority: Medium Status: In Progress Type: Enhancement :sparkles: | 
Ninja is retired, Ethfinex no longer exists, IDEX is dried up on liquidity. Also, cards have excess space on the right side (a.k.a. not centered). | 1.0 | Maintenance on 'Where to Buy' - 
Ninja is retired, Ethfinex no longer exists, IDEX is dried up on liquidity. Also, cards have excess space on the right side (a.k.a. not centered). | priority | maintenance on where to buy ninja is retired ethfinex no longer exists idex is dried up on liquidity also cards have excess space on the right side a k a not centered | 1 |
499,970 | 14,483,486,328 | IssuesEvent | 2020-12-10 15:13:17 | xournalpp/xournalpp | https://api.github.com/repos/xournalpp/xournalpp | closed | Command line option --page not working in 1.1.0+dev | bug priority::medium regression | **Affects versions :**
- OS: Linux, Windows
- (Linux only) Desktop environment: X11
- Which version of libgtk do you use: 3.24.23
- Version of Xournal++: 1.1.0+dev (at least from git commit 8afb3841 to the latest version)
- Installation method: All of PPA, flatpak and building from source
**Describe the bug**
The command line options `-n` (in short) and `--page` (in long) for scrolling to the specified page do not work any more in version 1.1.0+dev. Simply no scroll is taking place.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a multipage document, say `foo.xopp` with >2 pages
2. Run xournalpp with the --page option, e.g. `xournalpp --page=2 foo.xopp`
3. Observe that the document loads without scrolling to the 2nd page.
**Expected behavior**
I would expect scrolling to the specified page after the document has loaded.
**Additional context**
- In version 1.0.18 and 1.0.19 the options
`--page=2` or `--page 2` or `-n 2` work, the option `-n=2` does not (and probably is not supposed to work).
In version 1.1.0+dev the page number is still correctly read with the former 3 options, but the scrolling does not take place.
- Even with working --page (or -n) option the page number is shifted. According to `xournalpp --help` the first page number should be 1, but it is 0 indeed. So `--page=2` scrolls to the 3rd page instead of the 2nd one. That should be corrected, too.
- The error seems to happen because of an invalid Metadata entry. | 1.0 | Command line option --page not working in 1.1.0+dev - **Affects versions :**
- OS: Linux, Windows
- (Linux only) Desktop environment: X11
- Which version of libgtk do you use: 3.24.23
- Version of Xournal++: 1.1.0+dev (at least from git commit 8afb3841 to the latest version)
- Installation method: All of PPA, flatpak and building from source
**Describe the bug**
The command line options `-n` (in short) and `--page` (in long) for scrolling to the specified page do not work any more in version 1.1.0+dev. Simply no scroll is taking place.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a multipage document, say `foo.xopp` with >2 pages
2. Run xournalpp with the --page option, e.g. `xournalpp --page=2 foo.xopp`
3. Observe that the document loads without scrolling to the 2nd page.
**Expected behavior**
I would expect scrolling to the specified page after the document has loaded.
**Additional context**
- In version 1.0.18 and 1.0.19 the options
`--page=2` or `--page 2` or `-n 2` work, the option `-n=2` does not (and probably is not supposed to work).
In version 1.1.0+dev the page number is still correctly read with the former 3 options, but the scrolling does not take place.
- Even with working --page (or -n) option the page number is shifted. According to `xournalpp --help` the first page number should be 1, but it is 0 indeed. So `--page=2` scrolls to the 3rd page instead of the 2nd one. That should be corrected, too.
- The error seems to happen because of an invalid Metadata entry. | priority | command line option page not working in dev affects versions os linux windows linux only desktop environment which version of libgtk do you use version of xournal dev at least from git commit to the latest version installation method all of ppa flatpak and building from source describe the bug the command line options n in short and page in long for scrolling to the specified page do not work any more in version dev simply no scroll is taking place to reproduce steps to reproduce the behavior create a multipage document say foo xopp with pages run xournalpp with the page option e g xournalpp page foo xopp observe that the document loads without scrolling to the page expected behavior i would expect scrolling to the specified page after the document has loaded additional context in version and the options page or page or n work the option n does not and probably is not supposed to work in version dev the page number is still correctly read with the former options but the scrolling does not take place even with working page or n option the page number is shifted according to xournalpp help the first page number should be but it is indeed so page scrolls to the page instead of the one that should be corrected too the error seems to happen because of an invalid metadata entry | 1 |
412,816 | 12,056,562,399 | IssuesEvent | 2020-04-15 14:37:54 | AY1920S2-CS2103-W14-1/main | https://api.github.com/repos/AY1920S2-CS2103-W14-1/main | closed | As a user, I want to make analysis based on student sign-up/ drop out rate, so that I can know the current running business makes money | feature.Student priority.Medium type.Story | We do not delete students from the database!? | 1.0 | As a user, I want to make analysis based on student sign-up/ drop out rate, so that I can know the current running business makes money - We do not delete students from the database!? | priority | as a user i want to make analysis based on student sign up drop out rate so that i can know the current running business makes money we do not delete students from the database | 1 |
754,646 | 26,397,192,566 | IssuesEvent | 2023-01-12 20:36:14 | inlang/inlang | https://api.github.com/repos/inlang/inlang | closed | auto-rebase when changes from main to inlang branch | type: improvement scope: editor priority: medium | ## Problem
#219 introduced an automatic switch to a branch called "inlang". Changes from other branches are not auto-merged into this branch. This is fine as long as a user is aware of that fact. The current flow does not make the user aware of the fact that an inlang branch is created that could be outdated.
## Proposal
Signal to the user that a dedicated "inlang" branch has been created (if not exists) that could be outdated.
### Ideas:
- create a toast notification when the inlang branch has been created.
- auto-merge changes from main into inlang on clone if possible?
| 1.0 | auto-rebase when changes from main to inlang branch - ## Problem
#219 introduced an automatic switch to a branch called "inlang". Changes from other branches are not auto-merged into this branch. This is fine as long as a user is aware of that fact. The current flow does not make the user aware of the fact that an inlang branch is created that could be outdated.
## Proposal
Signal to the user that a dedicated "inlang" branch has been created (if not exists) that could be outdated.
### Ideas:
- create a toast notification when the inlang branch has been created.
- auto-merge changes from main into inlang on clone if possible?
| priority | auto rebase when changes from main to inlang branch problem introduced an automatic switch to a branch called inlang changes from other branches are not auto merged into this branch this is fine as long as a user is aware of that fact the current flow does not make the user aware of the fact that an inlang branch is created that could be outdated proposal signal to the user that a dedicated inlang branch has been created if not exists that could be outdated ideas create a toast notification when the inlang branch has been created auto merge changes from main into inlang on clone if possible | 1 |
650,699 | 21,413,872,781 | IssuesEvent | 2022-04-22 08:59:10 | ros-controls/ros2_control | https://api.github.com/repos/ros-controls/ros2_control | closed | As a ros2_control user I would like to define transmissions for my robots | medium-priority | # Design
Port over [transmission_interface](https://github.com/ros-controls/ros_control/tree/melodic-devel/transmission_interface) | 1.0 | As a ros2_control user I would like to define transmissions for my robots - # Design
Port over [transmission_interface](https://github.com/ros-controls/ros_control/tree/melodic-devel/transmission_interface) | priority | as a control user i would like to define transmissions for my robots design port over | 1 |
138,378 | 5,332,746,618 | IssuesEvent | 2017-02-15 22:56:06 | cuappdev/tcat-ios | https://api.github.com/repos/cuappdev/tcat-ios | closed | Design Data Model | Priority: Medium | Regardless of what the design ends up looking like, we will need models of stops, routes, etc. We should plan out what models we want, what fields/functions they have, and how they're related (at least roughly).
| 1.0 | Design Data Model - Regardless of what the design ends up looking like, we will need models of stops, routes, etc. We should plan out what models we want, what fields/functions they have, and how they're related (at least roughly).
| priority | design data model regardless of what the design ends up looking like we will need models of stops routes etc we should plan out what models we want what fields functions they have and how they re related at least roughly | 1 |
718,411 | 24,716,334,697 | IssuesEvent | 2022-10-20 07:16:54 | AY2223S1-CS2103T-T14-3/tp | https://api.github.com/repos/AY2223S1-CS2103T-T14-3/tp | closed | Add birthdate field for patients | enhancement priority.Medium | - Implement `Birthdate` class to encapsulate birthdate of a patient, and modify relevant classes and tests to reflect latest change. | 1.0 | Add birthdate field for patients - - Implement `Birthdate` class to encapsulate birthdate of a patient, and modify relevant classes and tests to reflect latest change. | priority | add birthdate field for patients implement birthdate class to encapsulate birthdate of a patient and modify relevant classes and tests to reflect latest change | 1 |
523,402 | 15,181,072,903 | IssuesEvent | 2021-02-15 02:17:49 | silentium-labs/merlin-gql | https://api.github.com/repos/silentium-labs/merlin-gql | closed | new command should also allow to setup ngrok for `basic` template | Priority: Medium Status: Pending Type: Enhancement | otherwise we would be forcing everyone to start with the `example` template to get all the features and the intention is the opposite. We want people to use the `examplte` template to learn the framework and the `basic` template to start an actual project from scratch
| 1.0 | new command should also allow to setup ngrok for `basic` template - otherwise we would be forcing everyone to start with the `example` template to get all the features and the intention is the opposite. We want people to use the `examplte` template to learn the framework and the `basic` template to start an actual project from scratch
| priority | new command should also allow to setup ngrok for basic template otherwise we would be forcing everyone to start with the example template to get all the features and the intention is the opposite we want people to use the examplte template to learn the framework and the basic template to start an actual project from scratch | 1 |
57,199 | 3,081,247,695 | IssuesEvent | 2015-08-22 14:38:35 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | opened | Playback games enhancements | enhancement imported Priority-Medium | _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on April 04, 2014 23:54:58_
When playing back games we should add (at least) the following features:
1. Some way to tell the name of the current player you're following
2. Have a quick key to travel to a 'person of interest'. Example CTRL+SPACE in a CTF game might go to whoever has the flag or is closest to it
3. (Maybe bugfix) Apply team-viewing rules to mines/spybugs. Right now if you are on Blue team, you can see Red team's mines/spybugs
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=420_ | 1.0 | Playback games enhancements - _From [buckyballreaction](https://code.google.com/u/buckyballreaction/) on April 04, 2014 23:54:58_
When playing back games we should add (at least) the following features:
1. Some way to tell the name of the current player you're following
2. Have a quick key to travel to a 'person of interest'. Example CTRL+SPACE in a CTF game might go to whoever has the flag or is closest to it
3. (Maybe bugfix) Apply team-viewing rules to mines/spybugs. Right now if you are on Blue team, you can see Red team's mines/spybugs
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=420_ | priority | playback games enhancements from on april when playing back games we should add at least the following features some way to tell the name of the current player you re following have a quick key to travel to a person of interest example ctrl space in a ctf game might go to whoever has the flag or is closest to it maybe bugfix apply team viewing rules to mines spybugs right now if you are on blue team you can see red team s mines spybugs original issue | 1 |
458,634 | 13,179,023,757 | IssuesEvent | 2020-08-12 10:09:20 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Cron: scanDownloadServer get versions from both buckets | Category: Accounts Priority: Medium Status: Fixed | Get versions from both buckets with paging (re-calling url with marker of a last version) | 1.0 | Cron: scanDownloadServer get versions from both buckets - Get versions from both buckets with paging (re-calling url with marker of a last version) | priority | cron scandownloadserver get versions from both buckets get versions from both buckets with paging re calling url with marker of a last version | 1 |
430,490 | 12,453,876,746 | IssuesEvent | 2020-05-27 14:28:00 | bounswe/bounswe2020group7 | https://api.github.com/repos/bounswe/bounswe2020group7 | closed | Practice App - Search | Priority: Medium Status: Done Status: Under review Type: New Feature | Implementation of the search functionality of **practice-app**.
- [x] Implementation of the Code
- [x] Documentation
- [x] Endpoint Documentation
- [x] Unit Tests | 1.0 | Practice App - Search - Implementation of the search functionality of **practice-app**.
- [x] Implementation of the Code
- [x] Documentation
- [x] Endpoint Documentation
- [x] Unit Tests | priority | practice app search implementation of the search functionality of practice app implementation of the code documentation endpoint documentation unit tests | 1 |
54,856 | 3,071,449,147 | IssuesEvent | 2015-08-19 12:12:08 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Вывод поисковых запросов в быстром поиске | Component-UI enhancement imported Priority-Medium Usability | _From [Tirael...@gmail.com](https://code.google.com/u/108935377450235604965/) on September 26, 2010 15:38:53_
1.Кнопка быстрого поиска в окошке флая хоть и разворачивается но там всегда пусто, сделайте так чтобы там выводились те же запросы что и сохранённые в обычном поиске.
2. Сделайте нормально пункт сброс в сохранённых запросах, а то список очищается, а при открытии нового окна опять загружаются поисковые запросы из базы.
**Attachment:** [Быстрый поиск.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=183)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=183_ | 1.0 | Вывод поисковых запросов в быстром поиске - _From [Tirael...@gmail.com](https://code.google.com/u/108935377450235604965/) on September 26, 2010 15:38:53_
1.Кнопка быстрого поиска в окошке флая хоть и разворачивается но там всегда пусто, сделайте так чтобы там выводились те же запросы что и сохранённые в обычном поиске.
2. Сделайте нормально пункт сброс в сохранённых запросах, а то список очищается, а при открытии нового окна опять загружаются поисковые запросы из базы.
**Attachment:** [Быстрый поиск.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=183)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=183_ | priority | вывод поисковых запросов в быстром поиске from on september кнопка быстрого поиска в окошке флая хоть и разворачивается но там всегда пусто сделайте так чтобы там выводились те же запросы что и сохранённые в обычном поиске сделайте нормально пункт сброс в сохранённых запросах а то список очищается а при открытии нового окна опять загружаются поисковые запросы из базы attachment original issue | 1 |
109,970 | 4,417,088,805 | IssuesEvent | 2016-08-15 02:04:21 | slackhq/node-slack-sdk | https://api.github.com/repos/slackhq/node-slack-sdk | closed | Support for incoming webhooks? | Feature Request Priority—Medium | Are y'all interested in exposing an object for posting [incoming webhooks](https://api.slack.com/incoming-webhooks)?
I was thinking an interface like this:
```js
var IncomingWebhook = require('@slack/client').IncomingWebhook;
// slackUrl is the webhook url provided when creating a new incoming webhook
// options would mirror the payload, so you could set default bot name, bot channel, bot emoji, etc
var iw = new IncomingWebhook(slackUrl, options);
// can pass a string argument
iw.send('some text', cb); // sends a payload of { "text": "some text" } along with any defaults set up in the initialization options
// can pass a payload object, anything not provided uses initialization defaults
iw.send({
text: 'some text',
attachments: [
// attachment data
],
username: 'some username',
icon_emoji: ':robot_face:'
}, cb);
```
If that's desired, I'm interested in working on it. | 1.0 | Support for incoming webhooks? - Are y'all interested in exposing an object for posting [incoming webhooks](https://api.slack.com/incoming-webhooks)?
I was thinking an interface like this:
```js
var IncomingWebhook = require('@slack/client').IncomingWebhook;
// slackUrl is the webhook url provided when creating a new incoming webhook
// options would mirror the payload, so you could set default bot name, bot channel, bot emoji, etc
var iw = new IncomingWebhook(slackUrl, options);
// can pass a string argument
iw.send('some text', cb); // sends a payload of { "text": "some text" } along with any defaults set up in the initialization options
// can pass a payload object, anything not provided uses initialization defaults
iw.send({
text: 'some text',
attachments: [
// attachment data
],
username: 'some username',
icon_emoji: ':robot_face:'
}, cb);
```
If that's desired, I'm interested in working on it. | priority | support for incoming webhooks are y all interested in exposing an object for posting i was thinking an interface like this js var incomingwebhook require slack client incomingwebhook slackurl is the webhook url provided when creating a new incoming webhook options would mirror the payload so you could set default bot name bot channel bot emoji etc var iw new incomingwebhook slackurl options can pass a string argument iw send some text cb sends a payload of text some text along with any defaults set up in the initialization options can pass a payload object anything not provided uses initialization defaults iw send text some text attachments attachment data username some username icon emoji robot face cb if that s desired i m interested in working on it | 1 |
9,052 | 2,607,904,819 | IssuesEvent | 2015-02-26 00:15:10 | chrsmithdemos/zen-coding | https://api.github.com/repos/chrsmithdemos/zen-coding | opened | Textpad Support | auto-migrated Priority-Medium Type-EditorSupport | ```
Any chance for Textpad support?
```
-----
Original issue reported on code.google.com by `Alarzel...@gmail.com` on 19 May 2010 at 6:31 | 1.0 | Textpad Support - ```
Any chance for Textpad support?
```
-----
Original issue reported on code.google.com by `Alarzel...@gmail.com` on 19 May 2010 at 6:31 | priority | textpad support any chance for textpad support original issue reported on code google com by alarzel gmail com on may at | 1 |
25,631 | 2,683,870,288 | IssuesEvent | 2015-03-28 12:09:16 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | ConEmu.exe 2010.2.17 падение Conemu/Conhost при turn bufferheight ON | 2–5 stars bug imported Priority-Medium | _From [Zero...@gmail.com](https://code.google.com/u/103642962356045697092/) on February 18, 2010 00:30:19_
Версия ОС:Win7 x86 Ultimate/W2003r2sp2
Версия FAR: 1.75 build 2621 x86/v2.0 (build 1406) x86 Описание бага... 1) Падение Conemu при работе с FAR2. Из плагинов только conemu.dll (Без
плагина conemu.dll - нет и падения).
через кнопку делаем turn bufferheight ON.
подводим курсор к правому краю окна Conemu. Видим что курсор линейки
прокрутки в самом низу. Мышкой тянем берём за него и ещё тянем вниз в
итоге процесс Conemu.exe падает.
воспроизведение 100%. на W2003R2 sp2 аналогично падает (дамп лог
прилагается). С FAR1.75 в такой ситуации не падаем под w2003r2.
1) Падение Conhost при работе с FAR1.75. Запуск без плагинов вообще (даже
без conemu.dll).
через кнопку делаем turn bufferheight ON в итоге Conhost сразу валится c
кодом исключения 0xc0000005.
на W2003 не падаем в такой ситуации.
воспроизведение 100%
ps. В этих 2х случаях падения есть и при дефолтных настройках Conemu.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=192_ | 1.0 | ConEmu.exe 2010.2.17 падение Conemu/Conhost при turn bufferheight ON - _From [Zero...@gmail.com](https://code.google.com/u/103642962356045697092/) on February 18, 2010 00:30:19_
Версия ОС:Win7 x86 Ultimate/W2003r2sp2
Версия FAR: 1.75 build 2621 x86/v2.0 (build 1406) x86 Описание бага... 1) Падение Conemu при работе с FAR2. Из плагинов только conemu.dll (Без
плагина conemu.dll - нет и падения).
через кнопку делаем turn bufferheight ON.
подводим курсор к правому краю окна Conemu. Видим что курсор линейки
прокрутки в самом низу. Мышкой тянем берём за него и ещё тянем вниз в
итоге процесс Conemu.exe падает.
воспроизведение 100%. на W2003R2 sp2 аналогично падает (дамп лог
прилагается). С FAR1.75 в такой ситуации не падаем под w2003r2.
1) Падение Conhost при работе с FAR1.75. Запуск без плагинов вообще (даже
без conemu.dll).
через кнопку делаем turn bufferheight ON в итоге Conhost сразу валится c
кодом исключения 0xc0000005.
на W2003 не падаем в такой ситуации.
воспроизведение 100%
ps. В этих 2х случаях падения есть и при дефолтных настройках Conemu.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=192_ | priority | conemu exe падение conemu conhost при turn bufferheight on from on february версия ос ultimate версия far build build описание бага падение conemu при работе с из плагинов только conemu dll без плагина conemu dll нет и падения через кнопку делаем turn bufferheight on подводим курсор к правому краю окна conemu видим что курсор линейки прокрутки в самом низу мышкой тянем берём за него и ещё тянем вниз в итоге процесс conemu exe падает воспроизведение на аналогично падает дамп лог прилагается с в такой ситуации не падаем под падение conhost при работе с запуск без плагинов вообще даже без conemu dll через кнопку делаем turn bufferheight on в итоге conhost сразу валится c кодом исключения на не падаем в такой ситуации воспроизведение ps в этих случаях падения есть и при дефолтных настройках conemu original issue | 1 |
3,887 | 2,541,648,936 | IssuesEvent | 2015-01-28 10:34:37 | jackjonesfashion/tasks | https://api.github.com/repos/jackjonesfashion/tasks | closed | Content - Update brand_footer with job links | In progress Priority: Medium Task | The current footer job links need to be split into stores and office.
Original ticket: jackjonesfashion/wiz/issues/77
Danish copy:
Job & Karriere butik
Job & Karriere kontor
English copy:
Jobs & Careers Stores
Jobs & Careers Office
Links:
JACK & JONES store: http://www.aboutbestseller.com/en/JobsContent/Search.aspx?CoId=0&CaId=0&BrId=44983&JlId=0&RId=0
JACK & JONES office: http://www.aboutbestseller.com/en/JobsContent/Search.aspx?CoId=0&CaId=0&BrId=8087&JlId=0&RId=0
- [x] Phase 1: Deploy all markets in english
- [ ] Phase 2: Deploy localized versions | 1.0 | Content - Update brand_footer with job links - The current footer job links need to be split into stores and office.
Original ticket: jackjonesfashion/wiz/issues/77
Danish copy:
Job & Karriere butik
Job & Karriere kontor
English copy:
Jobs & Careers Stores
Jobs & Careers Office
Links:
JACK & JONES store: http://www.aboutbestseller.com/en/JobsContent/Search.aspx?CoId=0&CaId=0&BrId=44983&JlId=0&RId=0
JACK & JONES office: http://www.aboutbestseller.com/en/JobsContent/Search.aspx?CoId=0&CaId=0&BrId=8087&JlId=0&RId=0
- [x] Phase 1: Deploy all markets in english
- [ ] Phase 2: Deploy localized versions | priority | content update brand footer with job links the current footer job links need to be split into stores and office original ticket jackjonesfashion wiz issues danish copy job karriere butik job karriere kontor english copy jobs careers stores jobs careers office links jack jones store jack jones office phase deploy all markets in english phase deploy localized versions | 1 |
57,296 | 3,081,254,854 | IssuesEvent | 2015-08-22 14:46:44 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Levels From Pleadies | enhancement imported invalid Priority-Medium | _From [Xelor41...@gmail.com](https://code.google.com/u/116935753572643655650/) on May 16, 2015 19:39:36_
Implement a System Where players can play Levels by getting them from pleadies from Pleiades Have a Search and and filter systems as well.
This would alow for Greater Level Variety as well as reduce Bitfighters memory Size.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=512_ | 1.0 | Levels From Pleadies - _From [Xelor41...@gmail.com](https://code.google.com/u/116935753572643655650/) on May 16, 2015 19:39:36_
Implement a System Where players can play Levels by getting them from pleadies from Pleiades Have a Search and and filter systems as well.
This would alow for Greater Level Variety as well as reduce Bitfighters memory Size.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=512_ | priority | levels from pleadies from on may implement a system where players can play levels by getting them from pleadies from pleiades have a search and and filter systems as well this would alow for greater level variety as well as reduce bitfighters memory size original issue | 1 |
160,468 | 6,098,109,474 | IssuesEvent | 2017-06-20 06:27:48 | intel-analytics/BigDL | https://api.github.com/repos/intel-analytics/BigDL | opened | SpatialConvolutionMap's function for creating connection table is unimplemented in python. | medium priority python | SpatialConvolutionMap.full(nin: Int, nout: In)
SpatialConvolutionMap.oneToOne(nfeat: Int)
SpatialConvolutionMap.random(nin: Int, nout: Int, nto: Int) | 1.0 | SpatialConvolutionMap's function for creating connection table is unimplemented in python. - SpatialConvolutionMap.full(nin: Int, nout: In)
SpatialConvolutionMap.oneToOne(nfeat: Int)
SpatialConvolutionMap.random(nin: Int, nout: Int, nto: Int) | priority | spatialconvolutionmap s function for creating connection table is unimplemented in python spatialconvolutionmap full nin int nout in spatialconvolutionmap onetoone nfeat int spatialconvolutionmap random nin int nout int nto int | 1 |
30,657 | 2,724,508,145 | IssuesEvent | 2015-04-14 18:13:50 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | Rest proxies does not recognizes methods on parent interfaces | bug imported Milestone-M14-C2 Priority-Medium | _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on April 08, 2014 17:01:44_
What steps will reproduce the problem? 1. Create a rest service that implements an interface
2. On that interface, declare a method with crux rest annotations What is the expected output? What do you see instead? Crux does not found the method on the interface
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=345_ | 1.0 | Rest proxies does not recognizes methods on parent interfaces - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on April 08, 2014 17:01:44_
What steps will reproduce the problem? 1. Create a rest service that implements an interface
2. On that interface, declare a method with crux rest annotations What is the expected output? What do you see instead? Crux does not found the method on the interface
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=345_ | priority | rest proxies does not recognizes methods on parent interfaces from on april what steps will reproduce the problem create a rest service that implements an interface on that interface declare a method with crux rest annotations what is the expected output what do you see instead crux does not found the method on the interface original issue | 1 |
31,743 | 2,736,708,171 | IssuesEvent | 2015-04-19 18:04:44 | devsnd/cherrymusic | https://api.github.com/repos/devsnd/cherrymusic | closed | Cannot load playlists with single quote in the name | bug priority medium | Cannot load playlists with single quote in the name. :crying_cat_face: | 1.0 | Cannot load playlists with single quote in the name - Cannot load playlists with single quote in the name. :crying_cat_face: | priority | cannot load playlists with single quote in the name cannot load playlists with single quote in the name crying cat face | 1 |
196,036 | 6,923,547,512 | IssuesEvent | 2017-11-30 09:27:35 | datavisyn/tdp_core | https://api.github.com/repos/datavisyn/tdp_core | opened | implement option for custom security check for views | priority: medium type: feature | e.g. whether the user has a certain role
```python
.security(lambda user: user.has_role('role_name'))
``` | 1.0 | implement option for custom security check for views - e.g. whether the user has a certain role
```python
.security(lambda user: user.has_role('role_name'))
``` | priority | implement option for custom security check for views e g whether the user has a certain role python security lambda user user has role role name | 1 |
637,158 | 20,622,341,846 | IssuesEvent | 2022-03-07 18:41:12 | hapi-server/data-specification | https://api.github.com/repos/hapi-server/data-specification | closed | Need to revise old docs about Unicode | priority-medium | State that Unicode is not allowed in JSON and Unicode parameters are not allowed. This will be a minor revision. | 1.0 | Need to revise old docs about Unicode - State that Unicode is not allowed in JSON and Unicode parameters are not allowed. This will be a minor revision. | priority | need to revise old docs about unicode state that unicode is not allowed in json and unicode parameters are not allowed this will be a minor revision | 1 |
98,787 | 4,031,271,469 | IssuesEvent | 2016-05-18 16:37:17 | navacohen90/Click-a-Table | https://api.github.com/repos/navacohen90/Click-a-Table | opened | Course details page | 0 - Backlog points: 10 priority2 - MEDIUM | client side - andular + routing
<!---
@huboard:{"milestone_order":0.000244140625,"order":0.000244140625}
-->
| 1.0 | Course details page - client side - andular + routing
<!---
@huboard:{"milestone_order":0.000244140625,"order":0.000244140625}
-->
| priority | course details page client side andular routing huboard milestone order order | 1 |
807,363 | 29,997,890,886 | IssuesEvent | 2023-06-26 07:15:21 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [World Event] [Midsummer] [Quest] More Torch Tossing | Priority: Medium Quest Achievement | [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
Until today I did not have any issues with the quest but today I have failed the quest 6 times, when you pick up the quest 2 braziers are disabled and you have to work with the other 3 working, however best I could do is get to 19/20 before the time is up and the quest is marked as failed
**How to reproduce:**
https://www.wowhead.com/quest=11921/more-torch-tossing take it in Stormwind city and do the quest normally
**How it should work:**
https://www.youtube.com/watch?v=SkCVDK3IpL0
Edit:
I just took the other Ally available quest in Darnassuss, it works as intended there, so I have issues with the Stormwind one
| 1.0 | [World Event] [Midsummer] [Quest] More Torch Tossing - [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
Until today I did not have any issues with the quest but today I have failed the quest 6 times, when you pick up the quest 2 braziers are disabled and you have to work with the other 3 working, however best I could do is get to 19/20 before the time is up and the quest is marked as failed
**How to reproduce:**
https://www.wowhead.com/quest=11921/more-torch-tossing take it in Stormwind city and do the quest normally
**How it should work:**
https://www.youtube.com/watch?v=SkCVDK3IpL0
Edit:
I just took the other Ally available quest in Darnassuss, it works as intended there, so I have issues with the Stormwind one
| priority | more torch tossing rembember add links to things related to the bug using for example cata twinhead twinstar cz description until today i did not have any issues with the quest but today i have failed the quest times when you pick up the quest braziers are disabled and you have to work with the other working however best i could do is get to before the time is up and the quest is marked as failed how to reproduce take it in stormwind city and do the quest normally how it should work edit i just took the other ally available quest in darnassuss it works as intended there so i have issues with the stormwind one | 1 |
600,234 | 18,291,968,197 | IssuesEvent | 2021-10-05 16:09:06 | parzh/iso4217 | https://api.github.com/repos/parzh/iso4217 | opened | Automatically export all definitions from `*.type.ts` files | Change: patch Domain: meta Pending: blocked Priority: medium Type: improvement | Currently, when another entity definition is added to a `*.type.ts` file, to re-export it through entry-point `index.ts` file it is necessary to reference this definition again via <code>export { <I>Entity</i> } from "<I>entity</I>.type.ts";</code> statement:
```diff
export type Type1 = any;
+ export type Type2 = any;
```
```diff
export {
Type1,
+ Type2,
} from "module.type.ts";
```
To comply to D.R.Y., the export statement should automatically pick up and re-export all definitions:
```ts
export * from "module.type.ts";
```
However, the latest version of TypeScript (4.4.x) preserves this export for type entities, causing runtime error due to these errors missing.
- [ ] Await for https://github.com/microsoft/TypeScript/issues/37238 to get resolved
- [ ] Change named exports to asterisk-exports | 1.0 | Automatically export all definitions from `*.type.ts` files - Currently, when another entity definition is added to a `*.type.ts` file, to re-export it through entry-point `index.ts` file it is necessary to reference this definition again via <code>export { <I>Entity</i> } from "<I>entity</I>.type.ts";</code> statement:
```diff
export type Type1 = any;
+ export type Type2 = any;
```
```diff
export {
Type1,
+ Type2,
} from "module.type.ts";
```
To comply to D.R.Y., the export statement should automatically pick up and re-export all definitions:
```ts
export * from "module.type.ts";
```
However, the latest version of TypeScript (4.4.x) preserves this export for type entities, causing runtime error due to these errors missing.
- [ ] Await for https://github.com/microsoft/TypeScript/issues/37238 to get resolved
- [ ] Change named exports to asterisk-exports | priority | automatically export all definitions from type ts files currently when another entity definition is added to a type ts file to re export it through entry point index ts file it is necessary to reference this definition again via export entity from entity type ts statement diff export type any export type any diff export from module type ts to comply to d r y the export statement should automatically pick up and re export all definitions ts export from module type ts however the latest version of typescript x preserves this export for type entities causing runtime error due to these errors missing await for to get resolved change named exports to asterisk exports | 1 |
458,715 | 13,180,247,570 | IssuesEvent | 2020-08-12 12:28:36 | mjuenema/python-terrascript | https://api.github.com/repos/mjuenema/python-terrascript | opened | Looking for new maintainer. | Priority: Medium help wanted | As many of you will have already noticed, unfortunately I don't have sufficient time to properly maintain python-terrascript anymore.
What was intended to be a just small experiment turned into my most popular Github project with more than 250 stars. In turn this means that at least 250 other people have a certain level of expectation towards me actually mainatining the project to a certain level of quality. I am simply not able to meet this expectation anymore and don't want to pretend that I could. I am happy to pass the baton on to someone else who's willing to jump in. | 1.0 | Looking for new maintainer. - As many of you will have already noticed, unfortunately I don't have sufficient time to properly maintain python-terrascript anymore.
What was intended to be a just small experiment turned into my most popular Github project with more than 250 stars. In turn this means that at least 250 other people have a certain level of expectation towards me actually mainatining the project to a certain level of quality. I am simply not able to meet this expectation anymore and don't want to pretend that I could. I am happy to pass the baton on to someone else who's willing to jump in. | priority | looking for new maintainer as many of you will have already noticed unfortunately i don t have sufficient time to properly maintain python terrascript anymore what was intended to be a just small experiment turned into my most popular github project with more than stars in turn this means that at least other people have a certain level of expectation towards me actually mainatining the project to a certain level of quality i am simply not able to meet this expectation anymore and don t want to pretend that i could i am happy to pass the baton on to someone else who s willing to jump in | 1 |
583,585 | 17,393,141,183 | IssuesEvent | 2021-08-02 10:00:57 | stackabletech/t2 | https://api.github.com/repos/stackabletech/t2 | closed | Remove cluster-signing params in K3s installation | priority/medium status/blocked | Due to a bug in K3s, we had to add heaps of params to the installation: https://github.com/stackabletech/t2/commit/cd3047437e25fbb2fb6de0be4987c568216f76f1
The newest release (https://github.com/k3s-io/k3s/releases/tag/v1.21.1%2Bk3s1) has this problem fixed and we should be able to go without these params
- [x] remove params
- [x] Test | 1.0 | Remove cluster-signing params in K3s installation - Due to a bug in K3s, we had to add heaps of params to the installation: https://github.com/stackabletech/t2/commit/cd3047437e25fbb2fb6de0be4987c568216f76f1
The newest release (https://github.com/k3s-io/k3s/releases/tag/v1.21.1%2Bk3s1) has this problem fixed and we should be able to go without these params
- [x] remove params
- [x] Test | priority | remove cluster signing params in installation due to a bug in we had to add heaps of params to the installation the newest release has this problem fixed and we should be able to go without these params remove params test | 1 |
242,268 | 7,840,282,019 | IssuesEvent | 2018-06-18 15:54:31 | DistrictDataLabs/yellowbrick | https://api.github.com/repos/DistrictDataLabs/yellowbrick | closed | Add a new parameter to the PCADecomposition visualizer in order to plot the feature columns in the projected space (biplot) | level: novice priority: medium type: feature | ### Proposal
Add a new parameter to the PCADecomposition class in order to have the option to plot the input columns.
I would like to have something like a biplot.

I can start working on the code but I would like to know if you think its a good idea.
| 1.0 | Add a new parameter to the PCADecomposition visualizer in order to plot the feature columns in the projected space (biplot) - ### Proposal
Add a new parameter to the PCADecomposition class in order to have the option to plot the input columns.
I would like to have something like a biplot.

I can start working on the code but I would like to know if you think its a good idea.
| priority | add a new parameter to the pcadecomposition visualizer in order to plot the feature columns in the projected space biplot proposal add a new parameter to the pcadecomposition class in order to have the option to plot the input columns i would like to have something like a biplot i can start working on the code but i would like to know if you think its a good idea | 1 |
605,172 | 18,726,308,039 | IssuesEvent | 2021-11-03 16:36:20 | canonical-web-and-design/snapcraft.io | https://api.github.com/repos/canonical-web-and-design/snapcraft.io | closed | https://snapcraft.io/build: support building snaps from more than the default branch | Priority: Medium Feature 🎁 | Last I checked, https://snapcraft.io/build only supported building snaps from the default branch. Is that still the case? What about tags?
First of all, only supporting the default branch assumes a particular workflow that not all projects use. For example, gitflow doesn't use the default branch as a development trunk at all. This also doesn't support the _very_ common use-case of release branches (a 3.0 branch that has 3.0.1, 3.0.2, etc).
@Lukewh asked me to specify what I'd _like_ to see. In an ideal world, I'd like to be able to specify a relationship between the git branch and the snap track. For example, the Nextcloud snap has `11`, `12`, `13`, `14`, `15`, `16`, `17`, and `master` branches. Whenever I land a commit to `master` I get a snap on `latest/edge`. Whenever I land a commit on `17`, I get a snap on `17/edge`, and so on. The only way I can do this today is to setup a sync from github to launchpad and then use launchpad to build a snap on changes to a given branch. This type of workflow is super common for larger projects, but getting that setup in Launchpad is non-trivial. It'd be great if https://snapcraft.io/build supported it with a decent UI.
That would be a great story, but I realize there's quite a bit of work outlined there. Is there a shorter-term thing we can do to enable projects using release branches like that without making them build snaps themselves or manually setup builds in LP? For example, regarding my above question: if we build on tag pushes we have somewhat of a workaround, it's just not entirely clear if that's supported. | 1.0 | https://snapcraft.io/build: support building snaps from more than the default branch - Last I checked, https://snapcraft.io/build only supported building snaps from the default branch. Is that still the case? What about tags?
First of all, only supporting the default branch assumes a particular workflow that not all projects use. For example, gitflow doesn't use the default branch as a development trunk at all. This also doesn't support the _very_ common use-case of release branches (a 3.0 branch that has 3.0.1, 3.0.2, etc).
@Lukewh asked me to specify what I'd _like_ to see. In an ideal world, I'd like to be able to specify a relationship between the git branch and the snap track. For example, the Nextcloud snap has `11`, `12`, `13`, `14`, `15`, `16`, `17`, and `master` branches. Whenever I land a commit to `master` I get a snap on `latest/edge`. Whenever I land a commit on `17`, I get a snap on `17/edge`, and so on. The only way I can do this today is to setup a sync from github to launchpad and then use launchpad to build a snap on changes to a given branch. This type of workflow is super common for larger projects, but getting that setup in Launchpad is non-trivial. It'd be great if https://snapcraft.io/build supported it with a decent UI.
That would be a great story, but I realize there's quite a bit of work outlined there. Is there a shorter-term thing we can do to enable projects using release branches like that without making them build snaps themselves or manually setup builds in LP? For example, regarding my above question: if we build on tag pushes we have somewhat of a workaround, it's just not entirely clear if that's supported. | priority | support building snaps from more than the default branch last i checked only supported building snaps from the default branch is that still the case what about tags first of all only supporting the default branch assumes a particular workflow that not all projects use for example gitflow doesn t use the default branch as a development trunk at all this also doesn t support the very common use case of release branches a branch that has etc lukewh asked me to specify what i d like to see in an ideal world i d like to be able to specify a relationship between the git branch and the snap track for example the nextcloud snap has and master branches whenever i land a commit to master i get a snap on latest edge whenever i land a commit on i get a snap on edge and so on the only way i can do this today is to setup a sync from github to launchpad and then use launchpad to build a snap on changes to a given branch this type of workflow is super common for larger projects but getting that setup in launchpad is non trivial it d be great if supported it with a decent ui that would be a great story but i realize there s quite a bit of work outlined there is there a shorter term thing we can do to enable projects using release branches like that without making them build snaps themselves or manually setup builds in lp for example regarding my above question if we build on tag pushes we have somewhat of a workaround it s just not entirely clear if that s supported | 1 |
565,444 | 16,761,420,473 | IssuesEvent | 2021-06-13 21:33:41 | CookieJarApps/SmartCookieWeb | https://api.github.com/repos/CookieJarApps/SmartCookieWeb | closed | [Bug] Bookmark Issues | P2: Medium priority bug stale | When setting bookmarks as start page, you will find that:
New bookmarks are not beeing sorted alphabetically as before,
although they are still sorted in the bookmark menu.
When exporting bookmarks, you will find that:
New bookmarks (see above) are added at the very end of the exported file.
This is related to all versions after version 9.x of SmartCookieWeb.
| 1.0 | [Bug] Bookmark Issues - When setting bookmarks as start page, you will find that:
New bookmarks are not beeing sorted alphabetically as before,
although they are still sorted in the bookmark menu.
When exporting bookmarks, you will find that:
New bookmarks (see above) are added at the very end of the exported file.
This is related to all versions after version 9.x of SmartCookieWeb.
| priority | bookmark issues when setting bookmarks as start page you will find that new bookmarks are not beeing sorted alphabetically as before although they are still sorted in the bookmark menu when exporting bookmarks you will find that new bookmarks see above are added at the very end of the exported file this is related to all versions after version x of smartcookieweb | 1 |
538,072 | 15,761,951,595 | IssuesEvent | 2021-03-31 10:31:59 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | Review of layer Identify behavior from Search tool | Accepted Internal Priority: Medium enhancement | ## Description
<!-- A few sentences describing new feature -->
<!-- screenshot, video, or link to mockup/prototype are welcome -->
Currently if you perform an Indentify request from the Search tool configuring it as reported below

'application/json' ('PROPERTIES') is always used as GetFeatureInfo output format so that MS is able to filter GFI results and show in the Identify panel only the one selected in Seach tool results (this means the filtering is made client side).

However, the Identify response should be provided in the format choosen by the user for the layer (through Layer Settings or in General Options). It is anyway useful, at the same time, to show in Identify panel only the information related to the selected item in Search tool. We should improve the current functionality to:
- Consider to have the current behavior configurable for retro compatibility (active or not with a new true/false property). To discuss.
- If the current behavior is disabled, always use the Identify format configured for the layer and not always 'application/json'
- Include the support for [GeoServer vendor params](https://docs.geoserver.org/stable/en/user/services/wms/reference.html#getfeatureinfo) like [filter](https://docs.geoserver.org/latest/en/user/services/wms/vendor.html#cql-filter) or [featureid](https://docs.geoserver.org/latest/en/user/services/wms/vendor.html#featureid) so that it is possible to filter the Identify result directly server side instead of client side (WMS servers that don't support the GeoServer vendor param will simply ignore it)
**What kind of improvement you want to add?** (check one with "x", remove the others)
- [X] Minor changes to existing features
- [ ] Code style update (formatting, local variables)
- [ ] Refactoring (no functional changes, no api changes)
- [ ] Build related changes
- [ ] CI related changes
- [ ] Other... Please describe:
## Other useful information
Draft of the work is available in a custom branch of MS [here](https://github.com/geosolutions-it/MapStore2/commit/16b7ea04936cd967cf414c10fee3c3112133996b) where the Identify format set for the layer is always used. | 1.0 | Review of layer Identify behavior from Search tool - ## Description
<!-- A few sentences describing new feature -->
<!-- screenshot, video, or link to mockup/prototype are welcome -->
Currently if you perform an Indentify request from the Search tool configuring it as reported below

'application/json' ('PROPERTIES') is always used as GetFeatureInfo output format so that MS is able to filter GFI results and show in the Identify panel only the one selected in Seach tool results (this means the filtering is made client side).

However, the Identify response should be provided in the format choosen by the user for the layer (through Layer Settings or in General Options). It is anyway useful, at the same time, to show in Identify panel only the information related to the selected item in Search tool. We should improve the current functionality to:
- Consider to have the current behavior configurable for retro compatibility (active or not with a new true/false property). To discuss.
- If the current behavior is disabled, always use the Identify format configured for the layer and not always 'application/json'
- Include the support for [GeoServer vendor params](https://docs.geoserver.org/stable/en/user/services/wms/reference.html#getfeatureinfo) like [filter](https://docs.geoserver.org/latest/en/user/services/wms/vendor.html#cql-filter) or [featureid](https://docs.geoserver.org/latest/en/user/services/wms/vendor.html#featureid) so that it is possible to filter the Identify result directly server side instead of client side (WMS servers that don't support the GeoServer vendor param will simply ignore it)
**What kind of improvement you want to add?** (check one with "x", remove the others)
- [X] Minor changes to existing features
- [ ] Code style update (formatting, local variables)
- [ ] Refactoring (no functional changes, no api changes)
- [ ] Build related changes
- [ ] CI related changes
- [ ] Other... Please describe:
## Other useful information
Draft of the work is available in a custom branch of MS [here](https://github.com/geosolutions-it/MapStore2/commit/16b7ea04936cd967cf414c10fee3c3112133996b) where the Identify format set for the layer is always used. | priority | review of layer identify behavior from search tool description currently if you perform an indentify request from the search tool configuring it as reported below application json properties is always used as getfeatureinfo output format so that ms is able to filter gfi results and show in the identify panel only the one selected in seach tool results this means the filtering is made client side however the identify response should be provided in the format choosen by the user for the layer through layer settings or in general options it is anyway useful at the same time to show in identify panel only the information related to the selected item in search tool we should improve the current functionality to consider to have the current behavior configurable for retro compatibility active or not with a new true false property to discuss if the current behavior is disabled always use the identify format configured for the layer and not always application json include the support for like or so that it is possible to filter the identify result directly server side instead of client side wms servers that don t support the geoserver vendor param will simply ignore it what kind of improvement you want to add check one with x remove the others minor changes to existing features code style update formatting local variables refactoring no functional changes no api changes build related changes ci related changes other please describe other useful information draft of the work is available in a custom branch of ms where the identify format set for the layer is always used | 1 |
814,528 | 30,510,597,438 | IssuesEvent | 2023-07-18 20:30:10 | hpc-io/pdc | https://api.github.com/repos/hpc-io/pdc | opened | Tests should return 0 for success and -1 for failure. | enhancement priority:medium | Currently, some tests are returning 1 for failure. | 1.0 | Tests should return 0 for success and -1 for failure. - Currently, some tests are returning 1 for failure. | priority | tests should return for success and for failure currently some tests are returning for failure | 1 |
557,564 | 16,511,744,576 | IssuesEvent | 2021-05-26 05:35:55 | Praful932/Kitabe | https://api.github.com/repos/Praful932/Kitabe | closed | Better Recommendations🧙♂️ | GSSOC21 Level3 assigned enhancement priority:medium | **Is your feature request related to a problem? Please describe.**
There are cases when there is not enough books to recommend the user, In this cases currently next top books are taken up according to TF-IDF approach.
**Describe the solution you'd like**
A better approach would be to take in equally from TF-IDF also similar genre books. If the remaining books to be recommended are 4, 2 - TF-IDF & 2 - Similar genres book. Also take into account already rated & included books before adding those 2 similar genres book.
[Concerning Code Line](https://github.com/Praful932/Kitabe/blob/master/mainapp/helpers.py#L187)
**Describe alternatives you've considered**
If there is any other alternative you would like to suggest go ahead.
**Additional context**
Concerning tests should be written.
| 1.0 | Better Recommendations🧙♂️ - **Is your feature request related to a problem? Please describe.**
There are cases when there is not enough books to recommend the user, In this cases currently next top books are taken up according to TF-IDF approach.
**Describe the solution you'd like**
A better approach would be to take in equally from TF-IDF also similar genre books. If the remaining books to be recommended are 4, 2 - TF-IDF & 2 - Similar genres book. Also take into account already rated & included books before adding those 2 similar genres book.
[Concerning Code Line](https://github.com/Praful932/Kitabe/blob/master/mainapp/helpers.py#L187)
**Describe alternatives you've considered**
If there is any other alternative you would like to suggest go ahead.
**Additional context**
Concerning tests should be written.
| priority | better recommendations🧙♂️ is your feature request related to a problem please describe there are cases when there is not enough books to recommend the user in this cases currently next top books are taken up according to tf idf approach describe the solution you d like a better approach would be to take in equally from tf idf also similar genre books if the remaining books to be recommended are tf idf similar genres book also take into account already rated included books before adding those similar genres book describe alternatives you ve considered if there is any other alternative you would like to suggest go ahead additional context concerning tests should be written | 1 |
632,444 | 20,197,054,518 | IssuesEvent | 2022-02-11 11:38:54 | nens/threedi-api-qgis-client | https://api.github.com/repos/nens/threedi-api-qgis-client | closed | When new upload starts, check if loaded model (3Di Toolbox) is same as loaded schematisation (3Di API Client) | ⏰ Priority: 3. Medium | If this isn't the case, warn the user about this
Perhaps we can include this in the overview page of the upload wizard? | 1.0 | When new upload starts, check if loaded model (3Di Toolbox) is same as loaded schematisation (3Di API Client) - If this isn't the case, warn the user about this
Perhaps we can include this in the overview page of the upload wizard? | priority | when new upload starts check if loaded model toolbox is same as loaded schematisation api client if this isn t the case warn the user about this perhaps we can include this in the overview page of the upload wizard | 1 |
20,631 | 2,622,855,194 | IssuesEvent | 2015-03-04 08:07:22 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | opened | Highlighting almost never works. | auto-migrated Priority-Medium | ```
What steps will reproduce the problem? Please include a URL.
Trying to use Page Monitor on URL:
http://www.amazon.com/gp/movers-and-shakers/toys-and-games/ref=zg_bsms_toys-and-
games_home_all?pf_rd_p=1712313502&pf_rd_s=center-1&pf_rd_t=2301&pf_rd_i=home&pf_
rd_m=ATVPDKIKX0DER&pf_rd_r=1Y81NK8KQZRX8VS8Y3JW It shows alerts often but I
almost never see any highlighting which basically makes it useless if this
can't be fixed.
What is the expected output? What do you see instead? I see the normal page
without any highlighting.
What version of the Chrome are you using? On what operating system? Using the
latest Chrome on a Windows 7 64-bit system.
What are the error log messages? To view them, go to Wrench -> Tools ->
Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" ->
Console. Have two lines of this: "WARNING: Watchdog recovered a lost timeout."
Please provide any additional information below. Useless without the
highlighting working on a consistent basis.
```
Original issue reported on code.google.com by `Atow...@gmail.com` on 31 Dec 2014 at 10:13 | 1.0 | Highlighting almost never works. - ```
What steps will reproduce the problem? Please include a URL.
Trying to use Page Monitor on URL:
http://www.amazon.com/gp/movers-and-shakers/toys-and-games/ref=zg_bsms_toys-and-
games_home_all?pf_rd_p=1712313502&pf_rd_s=center-1&pf_rd_t=2301&pf_rd_i=home&pf_
rd_m=ATVPDKIKX0DER&pf_rd_r=1Y81NK8KQZRX8VS8Y3JW It shows alerts often but I
almost never see any highlighting which basically makes it useless if this
can't be fixed.
What is the expected output? What do you see instead? I see the normal page
without any highlighting.
What version of the Chrome are you using? On what operating system? Using the
latest Chrome on a Windows 7 64-bit system.
What are the error log messages? To view them, go to Wrench -> Tools ->
Extensions -> Developer Mode -> expand Page Monitor -> "background.htm" ->
Console. Have two lines of this: "WARNING: Watchdog recovered a lost timeout."
Please provide any additional information below. Useless without the
highlighting working on a consistent basis.
```
Original issue reported on code.google.com by `Atow...@gmail.com` on 31 Dec 2014 at 10:13 | priority | highlighting almost never works what steps will reproduce the problem please include a url trying to use page monitor on url games home all pf rd p pf rd s center pf rd t pf rd i home pf rd m pf rd r it shows alerts often but i almost never see any highlighting which basically makes it useless if this can t be fixed what is the expected output what do you see instead i see the normal page without any highlighting what version of the chrome are you using on what operating system using the latest chrome on a windows bit system what are the error log messages to view them go to wrench tools extensions developer mode expand page monitor background htm console have two lines of this warning watchdog recovered a lost timeout please provide any additional information below useless without the highlighting working on a consistent basis original issue reported on code google com by atow gmail com on dec at | 1 |
239,201 | 7,787,271,034 | IssuesEvent | 2018-06-06 21:47:54 | robotframework/robotframework | https://api.github.com/repos/robotframework/robotframework | closed | Change source distribution format from `tar.gz` to `zip` | alpha 1 enhancement priority: medium | Our current source distribution format is `tar.gz` but we should change it to `zip` for these reasons:
- On Windows you can open `zip` files using the file explorer but `tar.gz` requires installing a separate too.
- There have been problems using `pip` on Windows both with Jython and IronPython, requiring users to do manual installation which requires extracting the source distribution.
In RF 3.1 we are going to start providing `wheel` distributions (#1734) so the source distribution format matters even less than earlier for `pip` users. Using a format that is easier for those with problems using `pip` thus makes sense.
The change only requires updating `BUILD.rst` to specify the format when creating the source distribution and changing references to the package in `INSTALL.rst` to use `*.zip` instead of `*.tar.gz`. | 1.0 | Change source distribution format from `tar.gz` to `zip` - Our current source distribution format is `tar.gz` but we should change it to `zip` for these reasons:
- On Windows you can open `zip` files using the file explorer but `tar.gz` requires installing a separate too.
- There have been problems using `pip` on Windows both with Jython and IronPython, requiring users to do manual installation which requires extracting the source distribution.
In RF 3.1 we are going to start providing `wheel` distributions (#1734) so the source distribution format matters even less than earlier for `pip` users. Using a format that is easier for those with problems using `pip` thus makes sense.
The change only requires updating `BUILD.rst` to specify the format when creating the source distribution and changing references to the package in `INSTALL.rst` to use `*.zip` instead of `*.tar.gz`. | priority | change source distribution format from tar gz to zip our current source distribution format is tar gz but we should change it to zip for these reasons on windows you can open zip files using the file explorer but tar gz requires installing a separate too there have been problems using pip on windows both with jython and ironpython requiring users to do manual installation which requires extracting the source distribution in rf we are going to start providing wheel distributions so the source distribution format matters even less than earlier for pip users using a format that is easier for those with problems using pip thus makes sense the change only requires updating build rst to specify the format when creating the source distribution and changing references to the package in install rst to use zip instead of tar gz | 1 |
66,246 | 3,251,417,086 | IssuesEvent | 2015-10-19 09:41:21 | cs2103aug2015-w15-3j/main | https://api.github.com/repos/cs2103aug2015-w15-3j/main | closed | Different modes of working | priority.medium type.epic | ## Insert mode
Put the focus on input bar and start adding task without typing "add"
## Search mode
Put the foucs on search bar | 1.0 | Different modes of working - ## Insert mode
Put the focus on input bar and start adding task without typing "add"
## Search mode
Put the foucs on search bar | priority | different modes of working insert mode put the focus on input bar and start adding task without typing add search mode put the foucs on search bar | 1 |
66,645 | 3,256,835,452 | IssuesEvent | 2015-10-20 15:22:30 | remkos/rads | https://api.github.com/repos/remkos/rads | opened | Replace GIA model | enhancement Priority-Medium | Peltier produced a new GIA model, called ICE6G. Apparently this appears to be a significant improvement for present-day GIA ...
http://www.atmosp.physics.utoronto.ca/~peltier/data.php | 1.0 | Replace GIA model - Peltier produced a new GIA model, called ICE6G. Apparently this appears to be a significant improvement for present-day GIA ...
http://www.atmosp.physics.utoronto.ca/~peltier/data.php | priority | replace gia model peltier produced a new gia model called apparently this appears to be a significant improvement for present day gia | 1 |
154,295 | 5,917,211,826 | IssuesEvent | 2017-05-22 12:43:03 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | opened | Finish ref ID preservation mod | Category: Algorithms Priority: Medium Status: Defined Type: Feature |
* BuildingMerger
* DualWaySplitter
* DuplicateWayRemover - Do this in #1518 instead
* ~~HighwaySnapMerger~~
* ~~IntersectionSplitter~~
* MergeNearbyNodes
* PartialNetworkMerger
* PoiPolygonMerger
* RemoveDuplicateAreaVisitor
* ~~SmallWayMerger~~
* WaySplitter | 1.0 | Finish ref ID preservation mod -
* BuildingMerger
* DualWaySplitter
* DuplicateWayRemover - Do this in #1518 instead
* ~~HighwaySnapMerger~~
* ~~IntersectionSplitter~~
* MergeNearbyNodes
* PartialNetworkMerger
* PoiPolygonMerger
* RemoveDuplicateAreaVisitor
* ~~SmallWayMerger~~
* WaySplitter | priority | finish ref id preservation mod buildingmerger dualwaysplitter duplicatewayremover do this in instead highwaysnapmerger intersectionsplitter mergenearbynodes partialnetworkmerger poipolygonmerger removeduplicateareavisitor smallwaymerger waysplitter | 1 |
207,839 | 7,134,200,441 | IssuesEvent | 2018-01-22 20:01:06 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | Unable to Edit/Delete Group User after Creation | Fix Proposed Medium Priority Resolved: Next Release bug |
<!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
After creating a new Group User, none of the action menu items work. I cannot edit the group user, or delete it.
#### Expected Behavior
I should be able to edit or delete a group user.
#### Actual Behavior
<!--- Tell us what happens instead -->
<!--- Also please check relevant logs (suitecrm.log, php error.log etc.) -->
Clicking on the 'Edit' action menu item does nothing. Clicking on the 'Delete' action menu item brings up the confirmation dialog, but clicking on 'Ok' there again does nothing.
Here is a small video (zipped to make github happy) that shows the issue plainly.
[vokoscreen-2018-01-08_14-30-24.mkv.zip](https://github.com/salesagility/SuiteCRM/files/1612809/vokoscreen-2018-01-08_14-30-24.mkv.zip)
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. Admin > Users > 'Create Group User'
2. Create a user and save
3. Try to edit or delete the user
4.
#### Context
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- If you feel this should be a low/medium/high priority then please state so -->
I cannot delete a Group User. At this point I can either dig into the Database manually or restore from a backup.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.8.9
* Browser name and version: Google Chrome 63.0.3239.108 (Official Build) (64-bit), Also tested in Firefox 57.0.3 (64-bit)
* Environment name and version (e.g. MySQL, PHP 7):
DB: mariadb-server-core-10.0 vers:10.0.31-0ubuntu0.16.04.2
PHP: php7.0 vers:7.0.22-0ubuntu0.16.04.1
Using LDAP logins.
* Operating System and version: Ubuntu 16.04
| 1.0 | Unable to Edit/Delete Group User after Creation -
<!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug -->
After creating a new Group User, none of the action menu items work. I cannot edit the group user, or delete it.
#### Expected Behavior
I should be able to edit or delete a group user.
#### Actual Behavior
<!--- Tell us what happens instead -->
<!--- Also please check relevant logs (suitecrm.log, php error.log etc.) -->
Clicking on the 'Edit' action menu item does nothing. Clicking on the 'Delete' action menu item brings up the confirmation dialog, but clicking on 'Ok' there again does nothing.
Here is a small video (zipped to make github happy) that shows the issue plainly.
[vokoscreen-2018-01-08_14-30-24.mkv.zip](https://github.com/salesagility/SuiteCRM/files/1612809/vokoscreen-2018-01-08_14-30-24.mkv.zip)
#### Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug include code to reproduce, if relevant -->
1. Admin > Users > 'Create Group User'
2. Create a user and save
3. Try to edit or delete the user
4.
#### Context
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- If you feel this should be a low/medium/high priority then please state so -->
I cannot delete a Group User. At this point I can either dig into the Database manually or restore from a backup.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.8.9
* Browser name and version: Google Chrome 63.0.3239.108 (Official Build) (64-bit), Also tested in Firefox 57.0.3 (64-bit)
* Environment name and version (e.g. MySQL, PHP 7):
DB: mariadb-server-core-10.0 vers:10.0.31-0ubuntu0.16.04.2
PHP: php7.0 vers:7.0.22-0ubuntu0.16.04.1
Using LDAP logins.
* Operating System and version: Ubuntu 16.04
| priority | unable to edit delete group user after creation issue after creating a new group user none of the action menu items work i cannot edit the group user or delete it expected behavior i should be able to edit or delete a group user actual behavior clicking on the edit action menu item does nothing clicking on the delete action menu item brings up the confirmation dialog but clicking on ok there again does nothing here is a small video zipped to make github happy that shows the issue plainly steps to reproduce admin users create group user create a user and save try to edit or delete the user context i cannot delete a group user at this point i can either dig into the database manually or restore from a backup your environment suitecrm version used version browser name and version google chrome official build bit also tested in firefox bit environment name and version e g mysql php db mariadb server core vers php vers using ldap logins operating system and version ubuntu | 1 |
23,748 | 2,663,014,708 | IssuesEvent | 2015-03-20 00:03:57 | FuturePilot/Pi2buntu | https://api.github.com/repos/FuturePilot/Pi2buntu | closed | Timezone is not configured | bug Fix Released Image Medium Priority | The timezone is not configured. Either set the timezone to something (which one?) or better, offer to configure it on first login. | 1.0 | Timezone is not configured - The timezone is not configured. Either set the timezone to something (which one?) or better, offer to configure it on first login. | priority | timezone is not configured the timezone is not configured either set the timezone to something which one or better offer to configure it on first login | 1 |
61,409 | 3,145,556,348 | IssuesEvent | 2015-09-14 18:34:54 | fusioneng/reactor | https://api.github.com/repos/fusioneng/reactor | opened | Automatically launch `vagrant rsync-auto` on `vagrant up` | priority:medium | Because we're dependent on `vagrant rsync-auto` to get our editable files in the machine, it would be nice if `vagrant up` spawned this process automatically.
I did some googlin' and couldn't find an out of the box solution for this.
Here's a related request on the Vagrant repo: https://github.com/mitchellh/vagrant/issues/3823
Here's a blog post documenting a similar implementation using `syncd`: http://drunomics.com/en/blog/syncd-sync-changes-vagrant-box
We may want to use `screen` behind the scenes, so we can recover and kill the process.
| 1.0 | Automatically launch `vagrant rsync-auto` on `vagrant up` - Because we're dependent on `vagrant rsync-auto` to get our editable files in the machine, it would be nice if `vagrant up` spawned this process automatically.
I did some googlin' and couldn't find an out of the box solution for this.
Here's a related request on the Vagrant repo: https://github.com/mitchellh/vagrant/issues/3823
Here's a blog post documenting a similar implementation using `syncd`: http://drunomics.com/en/blog/syncd-sync-changes-vagrant-box
We may want to use `screen` behind the scenes, so we can recover and kill the process.
| priority | automatically launch vagrant rsync auto on vagrant up because we re dependent on vagrant rsync auto to get our editable files in the machine it would be nice if vagrant up spawned this process automatically i did some googlin and couldn t find an out of the box solution for this here s a related request on the vagrant repo here s a blog post documenting a similar implementation using syncd we may want to use screen behind the scenes so we can recover and kill the process | 1 |
30,663 | 2,724,514,750 | IssuesEvent | 2015-04-14 18:15:58 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | DeviceAdaptiveGrid Smallcolumn layout bug | bug imported Priority-Medium | _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on April 22, 2014 15:57:16_
The ActionColumn of DeviceAdaptiveGrid had its CSS changed and the size of detail button was improved to guarantee a better usability.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=351_ | 1.0 | DeviceAdaptiveGrid Smallcolumn layout bug - _From [wes...@triggolabs.com](https://code.google.com/u/114691046055037037756/) on April 22, 2014 15:57:16_
The ActionColumn of DeviceAdaptiveGrid had its CSS changed and the size of detail button was improved to guarantee a better usability.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=351_ | priority | deviceadaptivegrid smallcolumn layout bug from on april the actioncolumn of deviceadaptivegrid had its css changed and the size of detail button was improved to guarantee a better usability original issue | 1 |
101,987 | 4,149,291,307 | IssuesEvent | 2016-06-15 14:03:53 | dhis2/dhis2-gis | https://api.github.com/repos/dhis2/dhis2-gis | closed | Earth Engine: Support storing color scale and elevation value and | enhancement medium priority Needs server work | User should be able to select colour scale and specify an elevation value for improved visualisation. This should be stored with the favorite. | 1.0 | Earth Engine: Support storing color scale and elevation value and - User should be able to select colour scale and specify an elevation value for improved visualisation. This should be stored with the favorite. | priority | earth engine support storing color scale and elevation value and user should be able to select colour scale and specify an elevation value for improved visualisation this should be stored with the favorite | 1 |
76,688 | 3,491,051,722 | IssuesEvent | 2016-01-04 13:53:51 | ngageoint/hootenanny-ui | https://api.github.com/repos/ngageoint/hootenanny-ui | opened | Click on review table zooms in too much | Category: UI Priority: Medium Status: New/Undefined Type: Bug | Create zoom threshold (level 18) for clicking on review table (red/blue). | 1.0 | Click on review table zooms in too much - Create zoom threshold (level 18) for clicking on review table (red/blue). | priority | click on review table zooms in too much create zoom threshold level for clicking on review table red blue | 1 |
378,848 | 11,209,794,694 | IssuesEvent | 2020-01-06 11:25:34 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1308] Contract notification when start to join | Fixed Medium Priority | Step to reproduce:
- spawn contract board.
- another player add contract like:

- take it.
- pay off all debt:

- complete contract:

- press okay and instanlly reload. Wait 30-60 sec
- try to join:

I have notification again before entering the world.
| 1.0 | [0.9.0 staging-1308] Contract notification when start to join - Step to reproduce:
- spawn contract board.
- another player add contract like:

- take it.
- pay off all debt:

- complete contract:

- press okay and instanlly reload. Wait 30-60 sec
- try to join:

I have notification again before entering the world.
| priority | contract notification when start to join step to reproduce spawn contract board another player add contract like take it pay off all debt complete contract press okay and instanlly reload wait sec try to join i have notification again before entering the world | 1 |
622,423 | 19,634,910,599 | IssuesEvent | 2022-01-08 04:51:42 | greencoast-studios/discord.js-extended | https://api.github.com/repos/greencoast-studios/discord.js-extended | closed | Add support for Slash Commands. | Priority: Medium Status: Awaiting PR Type: Feature Request | #### :zap: Describe the New Feature
Add support for slash commands. This will be necessary for when Discord starts [requiring verified bots a especial application to access message content](https://support-dev.discord.com/hc/en-us/articles/4404772028055).
Currently, we're waiting for discord.js to release v13 which will have support for Slash Commands.
| 1.0 | Add support for Slash Commands. - #### :zap: Describe the New Feature
Add support for slash commands. This will be necessary for when Discord starts [requiring verified bots a especial application to access message content](https://support-dev.discord.com/hc/en-us/articles/4404772028055).
Currently, we're waiting for discord.js to release v13 which will have support for Slash Commands.
| priority | add support for slash commands zap describe the new feature add support for slash commands this will be necessary for when discord starts currently we re waiting for discord js to release which will have support for slash commands | 1 |
704,389 | 24,195,163,315 | IssuesEvent | 2022-09-23 22:27:11 | MetaMask/metamask-mobile | https://api.github.com/repos/MetaMask/metamask-mobile | closed | Error message while connecting Ledger, even though it did connect successfully | type-bug needs-qa Priority - Medium team-key-management Ledger | **Describe the bug**
_A clear and concise description of what the bug is_
When connecting my Ledger to my Android device an error screen appeared when it shouldn't have
**Screenshots**
_If applicable, add screenshots or links to help explain your problem_

**To Reproduce**
_Steps to reproduce the behavior_
1. Go to 'Connect Ledger hardware wallet and follow all the steps'
2. See error screen
3. Notice that the Ledger has been connected successfully in any case
**Expected behavior**
_A clear and concise description of what you expected to happen_
This error screen shouldn't appear since the Ledger was connected successfully
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6] One Plus Nord N100
- OS: [e.g. iOS8.1] Android 11
- App Version [e.g. 1.0.0] - find version number in app from Settings > About MetaMask 5.3.1 (928)
-------------------------------------------------------------
_to be added after bug submission by internal support / PM_
**Severity**
- How critical is the impact of this bug on a user?
- Add stats if available on % of customers impacted
- Is this visible to all users?
- Is this tech debt?
| 1.0 | Error message while connecting Ledger, even though it did connect successfully - **Describe the bug**
_A clear and concise description of what the bug is_
When connecting my Ledger to my Android device an error screen appeared when it shouldn't have
**Screenshots**
_If applicable, add screenshots or links to help explain your problem_

**To Reproduce**
_Steps to reproduce the behavior_
1. Go to 'Connect Ledger hardware wallet and follow all the steps'
2. See error screen
3. Notice that the Ledger has been connected successfully in any case
**Expected behavior**
_A clear and concise description of what you expected to happen_
This error screen shouldn't appear since the Ledger was connected successfully
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6] One Plus Nord N100
- OS: [e.g. iOS8.1] Android 11
- App Version [e.g. 1.0.0] - find version number in app from Settings > About MetaMask 5.3.1 (928)
-------------------------------------------------------------
_to be added after bug submission by internal support / PM_
**Severity**
- How critical is the impact of this bug on a user?
- Add stats if available on % of customers impacted
- Is this visible to all users?
- Is this tech debt?
| priority | error message while connecting ledger even though it did connect successfully describe the bug a clear and concise description of what the bug is when connecting my ledger to my android device an error screen appeared when it shouldn t have screenshots if applicable add screenshots or links to help explain your problem to reproduce steps to reproduce the behavior go to connect ledger hardware wallet and follow all the steps see error screen notice that the ledger has been connected successfully in any case expected behavior a clear and concise description of what you expected to happen this error screen shouldn t appear since the ledger was connected successfully smartphone please complete the following information device one plus nord os android app version find version number in app from settings about metamask to be added after bug submission by internal support pm severity how critical is the impact of this bug on a user add stats if available on of customers impacted is this visible to all users is this tech debt | 1 |
52,157 | 3,021,959,457 | IssuesEvent | 2015-07-31 17:34:36 | mattwood1/zf1 | https://api.github.com/repos/mattwood1/zf1 | opened | Prevent Ranking Stack Build Up | enhancement Medium Priority ToDo Task | High Stack is building up as new models are being added. It was above 55, but currently its at 49. The more models are added to the stack the lower the High Stack will get.
Need to shift a model up when there isn't a High Stack peak.
1,1,4,7,7,7,8,8,8,8,8,1,1,1 4 would be selectable by random
1,1,3,8,7,7,8,8,8,8,8,1,1,1 3 would not be selectable as there its a gap between 8 (high)
1,1,3,7,8,7,8,8,8,8,8,1,1,1 is the same as above
1,1,3,7,7,8,8,8,8,8,8,1,1,1 3 would be selectable by random. | 1.0 | Prevent Ranking Stack Build Up - High Stack is building up as new models are being added. It was above 55, but currently its at 49. The more models are added to the stack the lower the High Stack will get.
Need to shift a model up when there isn't a High Stack peak.
1,1,4,7,7,7,8,8,8,8,8,1,1,1 4 would be selectable by random
1,1,3,8,7,7,8,8,8,8,8,1,1,1 3 would not be selectable as there its a gap between 8 (high)
1,1,3,7,8,7,8,8,8,8,8,1,1,1 is the same as above
1,1,3,7,7,8,8,8,8,8,8,1,1,1 3 would be selectable by random. | priority | prevent ranking stack build up high stack is building up as new models are being added it was above but currently its at the more models are added to the stack the lower the high stack will get need to shift a model up when there isn t a high stack peak would be selectable by random would not be selectable as there its a gap between high is the same as above would be selectable by random | 1 |
220,487 | 7,360,332,313 | IssuesEvent | 2018-03-10 17:30:14 | bounswe/bounswe2018group5 | https://api.github.com/repos/bounswe/bounswe2018group5 | opened | Revise Requirements | Effort: Medium Priority: High Status: Available Type: Wiki | Per Cihat's comment:
> * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities.
> * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides. | 1.0 | Revise Requirements - Per Cihat's comment:
> * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities.
> * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides. | priority | revise requirements per cihat s comment i am just lazy to check the previous version of your requirements and my feedback however one thing just got my attention for user requirements it seems that you have grouped the requirements according to the user types it could be better if you cluster them according to the functionalities one more comment instead of the user shall be able to see you can write them as system requirements since seeing is not something like a user requirement it just sees it if the system provides | 1 |
810,297 | 30,235,302,789 | IssuesEvent | 2023-07-06 09:47:58 | horizon-efrei/HorizonBot | https://api.github.com/repos/horizon-efrei/HorizonBot | closed | Plusieurs rôles pour les conditions des Réaction-Rôles | type: enhancement priority: medium difficulty: easy status: approved | Ajouter la possibilité de définir plusieurs rôles pour les conditions des réaction role.
BONUS: pouvoir définir si on veut que ce soit un "OU" entre tous les rôles, ou un "ET". | 1.0 | Plusieurs rôles pour les conditions des Réaction-Rôles - Ajouter la possibilité de définir plusieurs rôles pour les conditions des réaction role.
BONUS: pouvoir définir si on veut que ce soit un "OU" entre tous les rôles, ou un "ET". | priority | plusieurs rôles pour les conditions des réaction rôles ajouter la possibilité de définir plusieurs rôles pour les conditions des réaction role bonus pouvoir définir si on veut que ce soit un ou entre tous les rôles ou un et | 1 |
562,795 | 16,669,967,478 | IssuesEvent | 2021-06-07 09:37:08 | AbsaOSS/enceladus | https://api.github.com/repos/AbsaOSS/enceladus | reopened | Enceladus 3.0 deployment | Conformance Menas Standardization UX devops docs priority: medium under discussion | ## Background
Enceladus is the big new version of the project with a number of breaking changes.
## Feature
Collect the requirements needed to be done when Enceladus 3.0 will be deployed.
| 1.0 | Enceladus 3.0 deployment - ## Background
Enceladus is the big new version of the project with a number of breaking changes.
## Feature
Collect the requirements needed to be done when Enceladus 3.0 will be deployed.
| priority | enceladus deployment background enceladus is the big new version of the project with a number of breaking changes feature collect the requirements needed to be done when enceladus will be deployed | 1 |
485,164 | 13,962,074,356 | IssuesEvent | 2020-10-25 07:34:21 | AY2021S1-CS2103T-W17-1/tp | https://api.github.com/repos/AY2021S1-CS2103T-W17-1/tp | closed | Refactor some typical bugs in test data | priority.Medium type.Enhancement | Some of the sample data in test folder hasn't been modified since AddressBook3. | 1.0 | Refactor some typical bugs in test data - Some of the sample data in test folder hasn't been modified since AddressBook3. | priority | refactor some typical bugs in test data some of the sample data in test folder hasn t been modified since | 1 |
453,682 | 13,087,150,806 | IssuesEvent | 2020-08-02 10:34:45 | windowsboy111/merlin-py | https://api.github.com/repos/windowsboy111/merlin-py | closed | [BUG/ISSUE] If there are no polls, /poll end all will crash. | bug cog: Utilities priority: Medium | **Describe the bug**
the `/poll end all` command will crash if there are no pending polls to end.
**To Reproduce**
Steps to reproduce the behavior:
1. make sure you don't have open polls in the channel
2. run `/poll end all`
3. go to the log channel. you will see error message `TypeError: 'NoneType' object is not iterable`.
**Expected behavior**
it should pops out error messages telling user that there are no polls to end.
**Screenshots**

**Desktop (please complete the following information):**
- discord.py version: 1.3.4
- Linux
- python version: 3.8
| 1.0 | [BUG/ISSUE] If there are no polls, /poll end all will crash. - **Describe the bug**
the `/poll end all` command will crash if there are no pending polls to end.
**To Reproduce**
Steps to reproduce the behavior:
1. make sure you don't have open polls in the channel
2. run `/poll end all`
3. go to the log channel. you will see error message `TypeError: 'NoneType' object is not iterable`.
**Expected behavior**
it should pops out error messages telling user that there are no polls to end.
**Screenshots**

**Desktop (please complete the following information):**
- discord.py version: 1.3.4
- Linux
- python version: 3.8
| priority | if there are no polls poll end all will crash describe the bug the poll end all command will crash if there are no pending polls to end to reproduce steps to reproduce the behavior make sure you don t have open polls in the channel run poll end all go to the log channel you will see error message typeerror nonetype object is not iterable expected behavior it should pops out error messages telling user that there are no polls to end screenshots desktop please complete the following information discord py version linux python version | 1 |
57,676 | 3,083,254,162 | IssuesEvent | 2015-08-24 07:38:18 | magro/memcached-session-manager | https://api.github.com/repos/magro/memcached-session-manager | closed | Multiple session id's getting created in msm + memcached | bug imported invalid Priority-Medium | _From [yoga.und...@gmail.com](https://code.google.com/u/111652832360445102669/) on March 25, 2014 00:09:07_
<b>What steps will reproduce the problem?</b>
1. msm + JavaSerilization + spring mvc
2. <Manager className="de.javakaffee.web.msm.MemcachedBackupSessionManager"
sticky="false"
memcachedNodes="n1:lxdm620m7:11211,n2:lxdm491m7:11211"
lockingMode="all"
requestUriIgnorePattern=".*\.(png|gif|jpg|css|js|ico)$"
sessionBackupAsync="false"
sessionBackupTimeout="100"
/>
3. The application has two different webapps deployed and both are (context.xml) configured with MSM
<b>What is the expected output? What do you see instead?</b>
Expected to be one session id to be created in memcached, but I see numerous session id.
<b>What version of the product are you using? On what operating system?</b>
<b>What version of the product are you using? On what operating system?</b>
spymemcached-2.10.2.jar
memcached-session-manager-tc6-1.8.1.jar
memcached-session-manager-1.8.1.jar
tomcat-6.0.23
<b>Please provide any additional information below.</b>
lxdm491m7
========
<34 get ping-n2
>34 END
<34 set bak:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 2050 3600 6700
>34 STORED
<34 set bak:validity:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 2048 1800 20
>34 STORED
<34 get ping-n2
>34 END
<34 get ping-n2
>34 END
<34 set 474CB3EB3630CC30416ADB2413EFF407-n2.tomcat1 2050 3600 6268
>34 STORED
<34 set validity:474CB3EB3630CC30416ADB2413EFF407-n2.tomcat1 2048 1800 20
>34 STORED
<34 set A69964A85A44E1E0A98604196F06FD78-n2.tomcat1 2050 3600 6342
>34 STORED
<34 set validity:A69964A85A44E1E0A98604196F06FD78-n2.tomcat1 2048 1800 20
>34 STORED
<34 set B36436C8506165AE3FE1E6B3FCE917E6-n2.tomcat1 2050 3600 6270
>34 STORED
<34 set validity:B36436C8506165AE3FE1E6B3FCE917E6-n2.tomcat1 2048 1800 20
>34 STORED
<34 set 5050AA0FB5C31BB09A152E34A3D7266E-n2.tomcat1 2050 3600 6345
>34 STORED
<34 set validity:5050AA0FB5C31BB09A152E34A3D7266E-n2.tomcat1 2048 1800 20
>34 STORED
<34 set 702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2050 3600 6308
>34 STORED
<34 set validity:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2048 1800 20
>34 STORED
<34 get ping-n2
>34 END
<34 set bak:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2050 3600 6340
>34 STORED
<34 set bak:validity:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2050 3600 6269
>34 STORED
<34 set bak:validity:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2050 3600 6347
>34 STORED
<34 set bak:validity:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2048 1800 20
>34 STORED
lxdm620m7
=========
<34 set validity:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2050 3600 6308
>34 STORED
<34 set bak:validity:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2048 1800 20
>34 STORED
<34 add lock:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 0 5 6
>34 STORED
<34 get E659131B192E1853A99DE1C28CC10E14-n1.tomcat1
>34 END
<34 delete lock:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1
>34 DELETED
<34 set 3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2050 3600 6269
>34 STORED
<34 set validity:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2048 1800 20
>34 STORED
<34 set 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2050 3600 6347
>34 STORED
<34 set validity:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2048 1800 20
>34 STORED
35: Client using the ascii protocol
<35 get ping-n1
>35 END
<35 get ping-n1
>35 END
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 STORED
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 NOT_STORED
<35 get B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 END
<35 delete lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 DELETED
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 STORED
<35 get B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 END
<35 delete lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 DELETED
<35 add lock:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 0 5 6
>35 STORED
<35 get 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1
>35 sending key 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1
>35 END
<34 set 01AD4BC6FA04FCB4921ED4B5776FA2F9-n1.tomcat1 2050 3600 6267
>34 STORED
<34 set validity:01AD4BC6FA04FCB4921ED4B5776FA2F9-n1.tomcat1 2048 1800 20
>34 STORED
<34 set DF6023986EEEA92C41C40E0439BEF0CC-n1.tomcat1 2050 3600 6341
>34 STORED
<34 set validity:DF6023986EEEA92C41C40E0439BEF0CC-n1.tomcat1 2048 1800 20
>34 STORED
<34 set E2E7A1569540C766BF1821050DD646FD-n1.tomcat1 2050 3600 6269
>34 STORED
_Original issue: http://code.google.com/p/memcached-session-manager/issues/detail?id=202_ | 1.0 | Multiple session id's getting created in msm + memcached - _From [yoga.und...@gmail.com](https://code.google.com/u/111652832360445102669/) on March 25, 2014 00:09:07_
<b>What steps will reproduce the problem?</b>
1. msm + JavaSerilization + spring mvc
2. <Manager className="de.javakaffee.web.msm.MemcachedBackupSessionManager"
sticky="false"
memcachedNodes="n1:lxdm620m7:11211,n2:lxdm491m7:11211"
lockingMode="all"
requestUriIgnorePattern=".*\.(png|gif|jpg|css|js|ico)$"
sessionBackupAsync="false"
sessionBackupTimeout="100"
/>
3. The application has two different webapps deployed and both are (context.xml) configured with MSM
<b>What is the expected output? What do you see instead?</b>
Expected to be one session id to be created in memcached, but I see numerous session id.
<b>What version of the product are you using? On what operating system?</b>
<b>What version of the product are you using? On what operating system?</b>
spymemcached-2.10.2.jar
memcached-session-manager-tc6-1.8.1.jar
memcached-session-manager-1.8.1.jar
tomcat-6.0.23
<b>Please provide any additional information below.</b>
lxdm491m7
========
<34 get ping-n2
>34 END
<34 set bak:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 2050 3600 6700
>34 STORED
<34 set bak:validity:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 2048 1800 20
>34 STORED
<34 get ping-n2
>34 END
<34 get ping-n2
>34 END
<34 set 474CB3EB3630CC30416ADB2413EFF407-n2.tomcat1 2050 3600 6268
>34 STORED
<34 set validity:474CB3EB3630CC30416ADB2413EFF407-n2.tomcat1 2048 1800 20
>34 STORED
<34 set A69964A85A44E1E0A98604196F06FD78-n2.tomcat1 2050 3600 6342
>34 STORED
<34 set validity:A69964A85A44E1E0A98604196F06FD78-n2.tomcat1 2048 1800 20
>34 STORED
<34 set B36436C8506165AE3FE1E6B3FCE917E6-n2.tomcat1 2050 3600 6270
>34 STORED
<34 set validity:B36436C8506165AE3FE1E6B3FCE917E6-n2.tomcat1 2048 1800 20
>34 STORED
<34 set 5050AA0FB5C31BB09A152E34A3D7266E-n2.tomcat1 2050 3600 6345
>34 STORED
<34 set validity:5050AA0FB5C31BB09A152E34A3D7266E-n2.tomcat1 2048 1800 20
>34 STORED
<34 set 702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2050 3600 6308
>34 STORED
<34 set validity:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2048 1800 20
>34 STORED
<34 get ping-n2
>34 END
<34 set bak:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2050 3600 6340
>34 STORED
<34 set bak:validity:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2050 3600 6269
>34 STORED
<34 set bak:validity:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2050 3600 6347
>34 STORED
<34 set bak:validity:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2048 1800 20
>34 STORED
lxdm620m7
=========
<34 set validity:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 2048 1800 20
>34 STORED
<34 set bak:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2050 3600 6308
>34 STORED
<34 set bak:validity:702BB50D3E55B387AC78A7182C860EB6-n2.tomcat1 2048 1800 20
>34 STORED
<34 add lock:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1 0 5 6
>34 STORED
<34 get E659131B192E1853A99DE1C28CC10E14-n1.tomcat1
>34 END
<34 delete lock:E659131B192E1853A99DE1C28CC10E14-n1.tomcat1
>34 DELETED
<34 set 3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2050 3600 6269
>34 STORED
<34 set validity:3A7C2E63CD1A712F403EF78370424039-n1.tomcat1 2048 1800 20
>34 STORED
<34 set 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2050 3600 6347
>34 STORED
<34 set validity:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 2048 1800 20
>34 STORED
35: Client using the ascii protocol
<35 get ping-n1
>35 END
<35 get ping-n1
>35 END
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 STORED
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 NOT_STORED
<35 get B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 END
<35 delete lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 DELETED
<35 add lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1 0 5 6
>35 STORED
<35 get B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 END
<35 delete lock:B13619494440EB26B72524D92A0BF38D-n1.tomcat1
>35 DELETED
<35 add lock:30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1 0 5 6
>35 STORED
<35 get 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1
>35 sending key 30CF04AD9139F29A7077ED6676E1233B-n1.tomcat1
>35 END
<34 set 01AD4BC6FA04FCB4921ED4B5776FA2F9-n1.tomcat1 2050 3600 6267
>34 STORED
<34 set validity:01AD4BC6FA04FCB4921ED4B5776FA2F9-n1.tomcat1 2048 1800 20
>34 STORED
<34 set DF6023986EEEA92C41C40E0439BEF0CC-n1.tomcat1 2050 3600 6341
>34 STORED
<34 set validity:DF6023986EEEA92C41C40E0439BEF0CC-n1.tomcat1 2048 1800 20
>34 STORED
<34 set E2E7A1569540C766BF1821050DD646FD-n1.tomcat1 2050 3600 6269
>34 STORED
_Original issue: http://code.google.com/p/memcached-session-manager/issues/detail?id=202_ | priority | multiple session id s getting created in msm memcached from on march what steps will reproduce the problem msm javaserilization spring mvc lt manager classname de javakaffee web msm memcachedbackupsessionmanager sticky false memcachednodes lockingmode all requesturiignorepattern png gif jpg css js ico sessionbackupasync false sessionbackuptimeout gt the application has two different webapps deployed and both are context xml configured with msm what is the expected output what do you see instead expected to be one session id to be created in memcached but i see numerous session id what version of the product are you using on what operating system what version of the product are you using on what operating system spymemcached jar memcached session manager jar memcached session manager jar tomcat please provide any additional information below lt get ping gt end lt set bak gt stored lt set bak validity gt stored lt get ping gt end lt get ping gt end lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored lt get ping gt end lt set bak gt stored lt set bak validity gt stored lt set bak gt stored lt set bak validity gt stored lt set bak gt stored lt set bak validity gt stored lt set validity gt stored lt set bak gt stored lt set bak validity gt stored lt add lock gt stored lt get gt end lt delete lock gt deleted lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored client using the ascii protocol lt get ping gt end lt get ping gt end lt add lock gt stored lt add lock gt not stored lt get gt end lt delete lock gt deleted lt add lock gt stored lt get gt end lt delete lock gt deleted lt add lock gt stored lt get gt sending key gt end lt set gt stored lt set validity gt stored lt set gt stored lt set validity gt stored lt set gt stored original issue | 1 |
577,187 | 17,104,963,964 | IssuesEvent | 2021-07-09 16:16:02 | umple/umple | https://api.github.com/repos/umple/umple | closed | Crash in SQL generation of Association Specializations first manual example | Component-SemanticsAndGen Diffic-Med Priority-Medium associations bug sql | There is a compiler error on example of Association Specialization when execute SQL generate; I tested on-line and command line; below, first lines of log errors:
Couldn't read results from the Umple compiler!
java.lang.StackOverflowError
at cruise.umple.compiler.Multiplicity.parseInt(Umple_CodeAssociation.ump:659)
at cruise.umple.compiler.Multiplicity.getUpperBound(Umple_CodeAssociation.ump:621)
at cruise.umple.compiler.AssociationVariable.isOptionalOne(Umple_CodeAssociation.ump:382)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:224)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:230)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:230)
| 1.0 | Crash in SQL generation of Association Specializations first manual example - There is a compiler error on example of Association Specialization when execute SQL generate; I tested on-line and command line; below, first lines of log errors:
Couldn't read results from the Umple compiler!
java.lang.StackOverflowError
at cruise.umple.compiler.Multiplicity.parseInt(Umple_CodeAssociation.ump:659)
at cruise.umple.compiler.Multiplicity.getUpperBound(Umple_CodeAssociation.ump:621)
at cruise.umple.compiler.AssociationVariable.isOptionalOne(Umple_CodeAssociation.ump:382)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:224)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:230)
at cruise.umple.compiler.SqlGenerator.resolvePrimaryKey(Generator_CodeSql.ump:230)
| priority | crash in sql generation of association specializations first manual example there is a compiler error on example of association specialization when execute sql generate i tested on line and command line below first lines of log errors couldn t read results from the umple compiler java lang stackoverflowerror at cruise umple compiler multiplicity parseint umple codeassociation ump at cruise umple compiler multiplicity getupperbound umple codeassociation ump at cruise umple compiler associationvariable isoptionalone umple codeassociation ump at cruise umple compiler sqlgenerator resolveprimarykey generator codesql ump at cruise umple compiler sqlgenerator resolveprimarykey generator codesql ump at cruise umple compiler sqlgenerator resolveprimarykey generator codesql ump | 1 |
540,409 | 15,811,267,185 | IssuesEvent | 2021-04-05 01:54:41 | Kedyn/fusliez-notes | https://api.github.com/repos/Kedyn/fusliez-notes | closed | The crew mate icon is now very small. | Priority: Medium Status: In Progress Type: Maintenance | First of all, thank you for continuing to support a very useful tool.
With this update, the crew mate icon is now very small.
I want the size to be selectable.
Also, the previous drag-drop placement was convenient.
I want it to be selectable.
| 1.0 | The crew mate icon is now very small. - First of all, thank you for continuing to support a very useful tool.
With this update, the crew mate icon is now very small.
I want the size to be selectable.
Also, the previous drag-drop placement was convenient.
I want it to be selectable.
| priority | the crew mate icon is now very small first of all thank you for continuing to support a very useful tool with this update the crew mate icon is now very small i want the size to be selectable also the previous drag drop placement was convenient i want it to be selectable | 1 |
781,762 | 27,448,116,493 | IssuesEvent | 2023-03-02 15:33:23 | zowe/vscode-extension-for-zowe | https://api.github.com/repos/zowe/vscode-extension-for-zowe | closed | Regression in job-search through favorites | bug priority-medium severity-medium | There might be a slight issue with the updates made to the labels.
Based on @adam-wolfe's comments during yesterday's call, we looked into things a bit closer regarding favorites and found that the updated labels break the searching capabilities from a favorited job-search.

_Originally posted by @zFernand0 in https://github.com/zowe/vscode-extension-for-zowe/issues/2064#issuecomment-1425772824_
| 1.0 | Regression in job-search through favorites - There might be a slight issue with the updates made to the labels.
Based on @adam-wolfe's comments during yesterday's call, we looked into things a bit closer regarding favorites and found that the updated labels break the searching capabilities from a favorited job-search.

_Originally posted by @zFernand0 in https://github.com/zowe/vscode-extension-for-zowe/issues/2064#issuecomment-1425772824_
| priority | regression in job search through favorites there might be a slight issue with the updates made to the labels based on adam wolfe s comments during yesterday s call we looked into things a bit closer regarding favorites and found that the updated labels break the searching capabilities from a favorited job search originally posted by in | 1 |
744,507 | 25,945,994,711 | IssuesEvent | 2022-12-17 01:08:21 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] [YSQL] org.yb.pgsql.TestPgRegressTablegroup fails in multiple build types, cannot decode value type kGroupEnd | kind/bug area/docdb priority/medium | Jira Link: [DB-4155](https://yugabyte.atlassian.net/browse/DB-4155)
### Description
x86_64 macOS release:
https://gist.githubusercontent.com/mbautin/190048c188690a44da31fa43ab8cf0c2/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/172bf6d956705c47c9a75b8676a082dc/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"`
---
x86_64 clang15 almalinux8 release LTO with linuxbrew:
https://gist.githubusercontent.com/mbautin/52cd0689c58ae5578ce2756379f1dce0/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/0da2142c6c6c23ab3c356b74a1aca19f/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"`
--
aarch64 clang15 almalinux8 debug:
https://gist.githubusercontent.com/mbautin/7593645e6fd1ef9b88c48e692f07424f/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/acdb5495c570123a79bbb04be170f092/raw
Important part: `Timed out: Timed out waiting for Create Table`
--
aarch64 clang15 almalinux8 release:
https://gist.githubusercontent.com/mbautin/e229fa0b3c2d5b33b8b67f31fc7ab54e/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/4699b5275ed738ad1d07c6e14eebe6de/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"` | 1.0 | [DocDB] [YSQL] org.yb.pgsql.TestPgRegressTablegroup fails in multiple build types, cannot decode value type kGroupEnd - Jira Link: [DB-4155](https://yugabyte.atlassian.net/browse/DB-4155)
### Description
x86_64 macOS release:
https://gist.githubusercontent.com/mbautin/190048c188690a44da31fa43ab8cf0c2/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/172bf6d956705c47c9a75b8676a082dc/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"`
---
x86_64 clang15 almalinux8 release LTO with linuxbrew:
https://gist.githubusercontent.com/mbautin/52cd0689c58ae5578ce2756379f1dce0/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/0da2142c6c6c23ab3c356b74a1aca19f/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"`
--
aarch64 clang15 almalinux8 debug:
https://gist.githubusercontent.com/mbautin/7593645e6fd1ef9b88c48e692f07424f/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/acdb5495c570123a79bbb04be170f092/raw
Important part: `Timed out: Timed out waiting for Create Table`
--
aarch64 clang15 almalinux8 release:
https://gist.githubusercontent.com/mbautin/e229fa0b3c2d5b33b8b67f31fc7ab54e/raw
pgregress part only: https://gist.githubusercontent.com/mbautin/4699b5275ed738ad1d07c6e14eebe6de/raw
Important part: `Query error: Cannot decode value type kGroupEnd from the key encoding format: "!"` | priority | org yb pgsql testpgregresstablegroup fails in multiple build types cannot decode value type kgroupend jira link description macos release pgregress part only important part query error cannot decode value type kgroupend from the key encoding format release lto with linuxbrew pgregress part only important part query error cannot decode value type kgroupend from the key encoding format debug pgregress part only important part timed out timed out waiting for create table release pgregress part only important part query error cannot decode value type kgroupend from the key encoding format | 1 |
379,911 | 11,244,049,937 | IssuesEvent | 2020-01-10 05:45:07 | buttercup/buttercup-browser-extension | https://api.github.com/repos/buttercup/buttercup-browser-extension | closed | Focus on search field | Priority: Medium Status: Pending Type: Enhancement | I would like to request a little but handy feature - focus on search field in Buttercup dialog when a Buttercup button in an input field is pressed as not allways the relevant entry is displayed. Well, the search field can be focused by pressing `Tab` key, but those are four strokes.
For instance, the search field has focus when the Buttercup extension icon is clicked in browser toolbar.
| 1.0 | Focus on search field - I would like to request a little but handy feature - focus on search field in Buttercup dialog when a Buttercup button in an input field is pressed as not allways the relevant entry is displayed. Well, the search field can be focused by pressing `Tab` key, but those are four strokes.
For instance, the search field has focus when the Buttercup extension icon is clicked in browser toolbar.
| priority | focus on search field i would like to request a little but handy feature focus on search field in buttercup dialog when a buttercup button in an input field is pressed as not allways the relevant entry is displayed well the search field can be focused by pressing tab key but those are four strokes for instance the search field has focus when the buttercup extension icon is clicked in browser toolbar | 1 |
748,616 | 26,129,250,950 | IssuesEvent | 2022-12-29 00:43:11 | michelegargiulo/MineCrashRespawn | https://api.github.com/repos/michelegargiulo/MineCrashRespawn | opened | Add Final Tier or Ore Processing, using Impetus | Enhancement Priority Medium Multiblocked | And possibly also other capabilities (Thaumcraft aspects, EMC, GP, Starlight, etc) | 1.0 | Add Final Tier or Ore Processing, using Impetus - And possibly also other capabilities (Thaumcraft aspects, EMC, GP, Starlight, etc) | priority | add final tier or ore processing using impetus and possibly also other capabilities thaumcraft aspects emc gp starlight etc | 1 |
577,093 | 17,103,336,923 | IssuesEvent | 2021-07-09 14:17:39 | jpmorganchase/modular | https://api.github.com/repos/jpmorganchase/modular | closed | Global coverage vs local package coverage. | discussion medium priority | Since adding my partial test change back to Credit, times have improved a lot, (a recent build took 12mins as oppose to 1h 30mins) but now instead of reporting an average coverage in the changed package, we now report the average coverage of the changed file(s).
First complaint from someone with a failing build, "but I didn't create this file, I just needed to make a small change, now you say I have to get the coverage above 90%?!?!"
Devs were relying on these averages by testing some files really high (above 90%) so they could get around other files (probably ones they find/are actually tricky to test).
Now, for example, if they have a branch with a change to just this uncovered file, it will fail against the global jest.config.js coverage threshold.
With this in mind and the Sonar requirements for a repo:
Sonar only supports a coverage at the repo level.
50% coverage report for a develop/master.
70% on 90 day old code and 50% on rest.
I imagine in the short-term we will need a global jest to be sure we meet the above requirements with no single package dragging the rest down.
@NMinhNguyen pinged me some threads that will help the discussion here:
https://github.com/nrwl/nx/issues/622
https://github.com/facebook/jest/issues/2418
Quote form the nx issue: **"Having one report doesn't really work for most organizations. Imagine you have two teams building two apps in the same repo. The fact that one team failed to meet a certain threshold shouldn't affect the other team.
So we can have a report per team (app+all its libs), but then the situation tricky if you have a lot of shared libs."**
In relation to above, we are stuck at the moment with one report and if we enforce a global level coverage prior to PR merge ,teams can't impact each other.
So, I was wondering if we have already discussed ideas around this? and if not, maybe we could bake something in here so projects in the repo are certain to meet the tollgate requirements.
| 1.0 | Global coverage vs local package coverage. - Since adding my partial test change back to Credit, times have improved a lot, (a recent build took 12mins as oppose to 1h 30mins) but now instead of reporting an average coverage in the changed package, we now report the average coverage of the changed file(s).
First complaint from someone with a failing build, "but I didn't create this file, I just needed to make a small change, now you say I have to get the coverage above 90%?!?!"
Devs were relying on these averages by testing some files really high (above 90%) so they could get around other files (probably ones they find/are actually tricky to test).
Now, for example, if they have a branch with a change to just this uncovered file, it will fail against the global jest.config.js coverage threshold.
With this in mind and the Sonar requirements for a repo:
Sonar only supports a coverage at the repo level.
50% coverage report for a develop/master.
70% on 90 day old code and 50% on rest.
I imagine in the short-term we will need a global jest to be sure we meet the above requirements with no single package dragging the rest down.
@NMinhNguyen pinged me some threads that will help the discussion here:
https://github.com/nrwl/nx/issues/622
https://github.com/facebook/jest/issues/2418
Quote form the nx issue: **"Having one report doesn't really work for most organizations. Imagine you have two teams building two apps in the same repo. The fact that one team failed to meet a certain threshold shouldn't affect the other team.
So we can have a report per team (app+all its libs), but then the situation tricky if you have a lot of shared libs."**
In relation to above, we are stuck at the moment with one report and if we enforce a global level coverage prior to PR merge ,teams can't impact each other.
So, I was wondering if we have already discussed ideas around this? and if not, maybe we could bake something in here so projects in the repo are certain to meet the tollgate requirements.
| priority | global coverage vs local package coverage since adding my partial test change back to credit times have improved a lot a recent build took as oppose to but now instead of reporting an average coverage in the changed package we now report the average coverage of the changed file s first complaint from someone with a failing build but i didn t create this file i just needed to make a small change now you say i have to get the coverage above devs were relying on these averages by testing some files really high above so they could get around other files probably ones they find are actually tricky to test now for example if they have a branch with a change to just this uncovered file it will fail against the global jest config js coverage threshold with this in mind and the sonar requirements for a repo sonar only supports a coverage at the repo level coverage report for a develop master on day old code and on rest i imagine in the short term we will need a global jest to be sure we meet the above requirements with no single package dragging the rest down nminhnguyen pinged me some threads that will help the discussion here quote form the nx issue having one report doesn t really work for most organizations imagine you have two teams building two apps in the same repo the fact that one team failed to meet a certain threshold shouldn t affect the other team so we can have a report per team app all its libs but then the situation tricky if you have a lot of shared libs in relation to above we are stuck at the moment with one report and if we enforce a global level coverage prior to pr merge teams can t impact each other so i was wondering if we have already discussed ideas around this and if not maybe we could bake something in here so projects in the repo are certain to meet the tollgate requirements | 1 |
673,794 | 23,031,500,644 | IssuesEvent | 2022-07-22 14:19:54 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [backup] Restore can't create tables dropped after backup creation | kind/bug area/docdb priority/medium status/awaiting-triage | Jira Link: [DB-3031](https://yugabyte.atlassian.net/browse/DB-3031)
### Description
Case:
1. Create cluster RF=3, 5 nodes, c5.2xlarge
2. Create 10 tables and load 100 MB data each (does not matter how many). YSQL, all columns is `varchar`
3. Wait 1 min and create a backup with platform UI, on AWS S3
4. Drop tables:
```
drop table backup_4f6352_5
drop table backup_fe0236_9
drop table backup_e7e07d_3
drop table backup_c8c7a1_0
drop table backup_8991d8_6
drop table backup_bc4de0_2
```
5. Restore backup on different namespace
And here we discover that not all tables were created. In this particular example - `backup_4f6352_5`
All cluster/platform logs is here - [ycsxrotzmn.zip](https://github.com/yugabyte/yugabyte-db/files/9168035/ycsxrotzmn.zip)
version: `2.15.2.0-b21`
| 1.0 | [backup] Restore can't create tables dropped after backup creation - Jira Link: [DB-3031](https://yugabyte.atlassian.net/browse/DB-3031)
### Description
Case:
1. Create cluster RF=3, 5 nodes, c5.2xlarge
2. Create 10 tables and load 100 MB data each (does not matter how many). YSQL, all columns is `varchar`
3. Wait 1 min and create a backup with platform UI, on AWS S3
4. Drop tables:
```
drop table backup_4f6352_5
drop table backup_fe0236_9
drop table backup_e7e07d_3
drop table backup_c8c7a1_0
drop table backup_8991d8_6
drop table backup_bc4de0_2
```
5. Restore backup on different namespace
And here we discover that not all tables were created. In this particular example - `backup_4f6352_5`
All cluster/platform logs is here - [ycsxrotzmn.zip](https://github.com/yugabyte/yugabyte-db/files/9168035/ycsxrotzmn.zip)
version: `2.15.2.0-b21`
| priority | restore can t create tables dropped after backup creation jira link description case create cluster rf nodes create tables and load mb data each does not matter how many ysql all columns is varchar wait min and create a backup with platform ui on aws drop tables drop table backup drop table backup drop table backup drop table backup drop table backup drop table backup restore backup on different namespace and here we discover that not all tables were created in this particular example backup all cluster platform logs is here version | 1 |
496,351 | 14,345,580,869 | IssuesEvent | 2020-11-28 19:44:54 | NikitaKit1998/datakyt | https://api.github.com/repos/NikitaKit1998/datakyt | closed | Write employee's name and equimpment id on QR code | priority: medium type: feature | **Is your feature request related to a problem? Please describe.**
It will be convenient to have employee's first/last name and equipment id written n QR code image (sticker). This will help to quickly identify who owns some equipment without looking into db.
**Describe the solution you'd like**
Method that writes the required info on QR code PNG file.
**Additional context**
See [ImageDraw module](https://pillow.readthedocs.io/en/stable/reference/ImageDraw.html) for the reference
| 1.0 | Write employee's name and equimpment id on QR code - **Is your feature request related to a problem? Please describe.**
It will be convenient to have employee's first/last name and equipment id written n QR code image (sticker). This will help to quickly identify who owns some equipment without looking into db.
**Describe the solution you'd like**
Method that writes the required info on QR code PNG file.
**Additional context**
See [ImageDraw module](https://pillow.readthedocs.io/en/stable/reference/ImageDraw.html) for the reference
| priority | write employee s name and equimpment id on qr code is your feature request related to a problem please describe it will be convenient to have employee s first last name and equipment id written n qr code image sticker this will help to quickly identify who owns some equipment without looking into db describe the solution you d like method that writes the required info on qr code png file additional context see for the reference | 1 |
577,095 | 17,103,383,873 | IssuesEvent | 2021-07-09 14:20:42 | f-lab-edu/conference-reservation | https://api.github.com/repos/f-lab-edu/conference-reservation | opened | UserLoginDto의 Id 의 코드컨벤션 수정 | Priority: Medium Type: Feature/Function | - 추가 / 개선 요소
* UserLoginDto 의 private field 가 `Id` 입니다. `id`로 변경하는 것을 제안합니다.
| 1.0 | UserLoginDto의 Id 의 코드컨벤션 수정 - - 추가 / 개선 요소
* UserLoginDto 의 private field 가 `Id` 입니다. `id`로 변경하는 것을 제안합니다.
| priority | userlogindto의 id 의 코드컨벤션 수정 추가 개선 요소 userlogindto 의 private field 가 id 입니다 id 로 변경하는 것을 제안합니다 | 1 |
671,230 | 22,749,664,145 | IssuesEvent | 2022-07-07 12:09:42 | argosp/trialdash | https://api.github.com/repos/argosp/trialdash | closed | Dump and load an experiment | enhancement Wait for QA Priority Medium done | Add a feature to dump the entire experiment to a JSON file. and the possibility to upload it.
I think the images should be saved as binary-ascii data in the file.
If there are problems, lets talk about it. | 1.0 | Dump and load an experiment - Add a feature to dump the entire experiment to a JSON file. and the possibility to upload it.
I think the images should be saved as binary-ascii data in the file.
If there are problems, lets talk about it. | priority | dump and load an experiment add a feature to dump the entire experiment to a json file and the possibility to upload it i think the images should be saved as binary ascii data in the file if there are problems lets talk about it | 1 |
487,045 | 14,018,365,301 | IssuesEvent | 2020-10-29 16:44:40 | AY2021S1-CS2103-F09-2/tp | https://api.github.com/repos/AY2021S1-CS2103-F09-2/tp | closed | Update Suspects of Investigation Case | priority.Medium type.Story | As an investigator, I can update the list of suspects that is linked to an investigation case. | 1.0 | Update Suspects of Investigation Case - As an investigator, I can update the list of suspects that is linked to an investigation case. | priority | update suspects of investigation case as an investigator i can update the list of suspects that is linked to an investigation case | 1 |
423,323 | 12,294,396,544 | IssuesEvent | 2020-05-10 23:42:40 | minio/minio | https://api.github.com/repos/minio/minio | closed | Minio Internal Blobstore for PCF - Static IP Address for Load Balancer | community priority: medium | Feature Request:
Currently the Minio Internal Blobstore for PCF tile does not have a way to assign a static IP address to the Minio load balancer. Currently you can add IPs for the Minio nodes but not the load balancer. This makes it a problem when you redeploy and the load balancer IP address may change. It would be better to be able to assign a static ip address to the load balancer.

| 1.0 | Minio Internal Blobstore for PCF - Static IP Address for Load Balancer - Feature Request:
Currently the Minio Internal Blobstore for PCF tile does not have a way to assign a static IP address to the Minio load balancer. Currently you can add IPs for the Minio nodes but not the load balancer. This makes it a problem when you redeploy and the load balancer IP address may change. It would be better to be able to assign a static ip address to the load balancer.

| priority | minio internal blobstore for pcf static ip address for load balancer feature request currently the minio internal blobstore for pcf tile does not have a way to assign a static ip address to the minio load balancer currently you can add ips for the minio nodes but not the load balancer this makes it a problem when you redeploy and the load balancer ip address may change it would be better to be able to assign a static ip address to the load balancer | 1 |
232,848 | 7,680,841,557 | IssuesEvent | 2018-05-16 04:10:05 | socialappslab/appcivist-pb-client | https://api.github.com/repos/socialappslab/appcivist-pb-client | closed | Fix translation of type of contribution in the stats of campaign page | Priority: Medium |
### What's the problem?
The text that describes the status of proposals in the little stats of the campaign is translated in the wrong gender (should be "Publicadas" not "Publicado" because it refers to Propostas.
See example at => `https://testpb.appcivist.org/#/v2/p/assembly/cc699ccf-ffb1-47e9-8b96-2a7e7012324d/campaign/894a3177-3e4b-45f4-8bf9-9e59215b7b14`

### Where to look?
1. The problem is in the campaign page. The template is in `app/v2/partials/campaign/dashboard.html`
Look for this section of code:
```
<header class="proposals__header">
<!-- Statistics/Insights summary section-->
<div class="dash_insights">
<span ng-if="!commentsSectionExpanded && enableComments"><a ng-click="toggleCommentsSection()">{{'Show comments'|translate}}</a></span>
<span ng-if="commentsSectionExpanded && enableComments"><a ng-click="toggleCommentsSection()">{{'Hide comments'|translate}}</a></span>
<span ng-if="currentComponentType!=='IDEAS' && enableProposals">
<strong>{{'Proposals' | translate}}:</strong>
{{ insights.contributions_total.contributions_per_type.PROPOSAL.PUBLISHED || 0}}
<small>({{'Published' | translate}})</small>,
{{ insights.contributions_total.contributions_per_type.PROPOSAL.DRAFT || 0}}
<small>({{'Drafts' | translate}})</small>
</span>
<span ng-if="enableIdeas && enableProposals"> – </span>
<span ng-if="enableIdeas">
<strong>{{'Ideas' | translate}}:</strong>
{{ insights.contributions_total.contributions_per_type.IDEA.PUBLISHED || 0}}
<small>({{'Published' | translate}})</small>, {{ insights.contributions_total.contributions_per_type.IDEA.DRAFT || 0}}
<small>({{'Draft' | translate}})</small>
</span>
</div>
</header>
```
Notice how there is one that use the translation key 'Drafts'. We need the same for published: one plural version that must be added to all the languages.
2. The controller is in `app/v2/controllers/campaign/dashboard.js`, but won't be neccessary to touch this in this issue
### What to do?
- [ ] Replace `{{'Published' | translate}}` for another key that would be translated to `Publicadas` in portuguese and spanish. (i.e., `{{'Publisheds' | translate}}`)
- Since the `Publisheds` key might not exist yet in our translation files, add it to all the languages. Tranlsation files are in `assets/i18n/*.json`. Each file represents a language and contains keys and values (`"Drafts": "Drafts"` in the en-US.json file and `"Drafts": "Rascunhos"` in the pt-BR.json file for portuguese).
### Observations:
Remember: to develop this, start a branch named `feature/NUMBER_OF_ISSUE` out of the branch `develop`. When you finish, open a Pull Request to merge that branch into `develop` and move the issue to `Dev complete`, labeling with the label `Status: Review Needed`.
| 1.0 | Fix translation of type of contribution in the stats of campaign page -
### What's the problem?
The text that describes the status of proposals in the little stats of the campaign is translated in the wrong gender (should be "Publicadas" not "Publicado" because it refers to Propostas.
See example at => `https://testpb.appcivist.org/#/v2/p/assembly/cc699ccf-ffb1-47e9-8b96-2a7e7012324d/campaign/894a3177-3e4b-45f4-8bf9-9e59215b7b14`

### Where to look?
1. The problem is in the campaign page. The template is in `app/v2/partials/campaign/dashboard.html`
Look for this section of code:
```
<header class="proposals__header">
<!-- Statistics/Insights summary section-->
<div class="dash_insights">
<span ng-if="!commentsSectionExpanded && enableComments"><a ng-click="toggleCommentsSection()">{{'Show comments'|translate}}</a></span>
<span ng-if="commentsSectionExpanded && enableComments"><a ng-click="toggleCommentsSection()">{{'Hide comments'|translate}}</a></span>
<span ng-if="currentComponentType!=='IDEAS' && enableProposals">
<strong>{{'Proposals' | translate}}:</strong>
{{ insights.contributions_total.contributions_per_type.PROPOSAL.PUBLISHED || 0}}
<small>({{'Published' | translate}})</small>,
{{ insights.contributions_total.contributions_per_type.PROPOSAL.DRAFT || 0}}
<small>({{'Drafts' | translate}})</small>
</span>
<span ng-if="enableIdeas && enableProposals"> – </span>
<span ng-if="enableIdeas">
<strong>{{'Ideas' | translate}}:</strong>
{{ insights.contributions_total.contributions_per_type.IDEA.PUBLISHED || 0}}
<small>({{'Published' | translate}})</small>, {{ insights.contributions_total.contributions_per_type.IDEA.DRAFT || 0}}
<small>({{'Draft' | translate}})</small>
</span>
</div>
</header>
```
Notice how there is one that use the translation key 'Drafts'. We need the same for published: one plural version that must be added to all the languages.
2. The controller is in `app/v2/controllers/campaign/dashboard.js`, but won't be neccessary to touch this in this issue
### What to do?
- [ ] Replace `{{'Published' | translate}}` for another key that would be translated to `Publicadas` in portuguese and spanish. (i.e., `{{'Publisheds' | translate}}`)
- Since the `Publisheds` key might not exist yet in our translation files, add it to all the languages. Tranlsation files are in `assets/i18n/*.json`. Each file represents a language and contains keys and values (`"Drafts": "Drafts"` in the en-US.json file and `"Drafts": "Rascunhos"` in the pt-BR.json file for portuguese).
### Observations:
Remember: to develop this, start a branch named `feature/NUMBER_OF_ISSUE` out of the branch `develop`. When you finish, open a Pull Request to merge that branch into `develop` and move the issue to `Dev complete`, labeling with the label `Status: Review Needed`.
| priority | fix translation of type of contribution in the stats of campaign page what s the problem the text that describes the status of proposals in the little stats of the campaign is translated in the wrong gender should be publicadas not publicado because it refers to propostas see example at where to look the problem is in the campaign page the template is in app partials campaign dashboard html look for this section of code show comments translate hide comments translate proposals translate insights contributions total contributions per type proposal published published translate insights contributions total contributions per type proposal draft drafts translate nbsp ndash nbsp ideas translate insights contributions total contributions per type idea published published translate insights contributions total contributions per type idea draft draft translate notice how there is one that use the translation key drafts we need the same for published one plural version that must be added to all the languages the controller is in app controllers campaign dashboard js but won t be neccessary to touch this in this issue what to do replace published translate for another key that would be translated to publicadas in portuguese and spanish i e publisheds translate since the publisheds key might not exist yet in our translation files add it to all the languages tranlsation files are in assets json each file represents a language and contains keys and values drafts drafts in the en us json file and drafts rascunhos in the pt br json file for portuguese observations remember to develop this start a branch named feature number of issue out of the branch develop when you finish open a pull request to merge that branch into develop and move the issue to dev complete labeling with the label status review needed | 1 |
142,015 | 5,448,151,960 | IssuesEvent | 2017-03-07 15:16:25 | stats4sd/wordcloud-app | https://api.github.com/repos/stats4sd/wordcloud-app | closed | Add front page with big, easy buttons | 1 - Ready Impact-High Priority-Medium Size-Small Type-feature | 2 buttons - data session and "end of day review" (or something like that!) | 1.0 | Add front page with big, easy buttons - 2 buttons - data session and "end of day review" (or something like that!) | priority | add front page with big easy buttons buttons data session and end of day review or something like that | 1 |
763,324 | 26,752,446,964 | IssuesEvent | 2023-01-30 20:47:37 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] Deleting snapshot never finishes | kind/bug area/docdb priority/medium | Jira Link: [[DB-417]](https://yugabyte.atlassian.net/browse/DB-417)
### Description
On our cluster a backup was attempted on the 25th April 2022, this created a database snapshot, around 15 minutes after creating the snapshot the snapshot was deleted (or it's deletion was started), since then it has been unsuccessfully trying to find the leader for a large number of tablets, ever 10 minutes it fires 20 RPC requests per tablet to at least one server, sometimes multiple, this ends up with 1000's of unwanted and un-needed rpc requests on an already busy system. We are on a pre-release 2.13 build, I have put full logs on Google Drive (this is a 39 node cluster so there a lot of them). This is affecting 207 tablets.
### Snapshot
```
[yugabyte@ip-172-159-39-237 ~]$ master/bin/yb-admin -master_addresses `hostname -i`:7100 list_snapshots
Snapshot UUID State Creation Time
51fbb256-1923-42bf-a5a7-4b865ade9002 DELETING 2022-04-25 16:44:10.655943
```
### Log file out from just one of the tservers
```
8e1d886684d: Started tablet snapshot creation in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:10.661648 14742 checkpoint.cc:92] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d [I]: Started the snapshot process -- creating snapshot in directory /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp.intents
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:10.662571 14742 checkpoint.cc:92] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d [R]: Started the snapshot process -- creating snapshot in directory /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp
[...]
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.428580 10363 tablet_snapshots.cc:411] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Checkpoint created in /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.429877 10363 version_set.cc:2842] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Recovered from manifest file:/mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp/MANIFEST-000144 succeeded,manifest_file_number is 144, next_file_number is 162, last_sequence is 1125899945636174, log_number is 0,prev_log_number is 0,max_column_family is 0, flushed_values is 0x0000000124b31180 -> { op_id: 19.53572278 hybrid_time: { physical: 1650904793405367 } history_cutoff: { physical: 1650895988300693 } hybrid_time_filter: <invalid> max_value_level_ttl_expiration_time: <initial> }
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.430670 10363 tablet_snapshots.cc:175] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot creation in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002, snapshot hybrid time: { physical: 1650905050655943 }
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.373879 24136 tablet_snapshots.cc:369] T 47d7f101463b4b879b635edf21757e71 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-00004001000030008000000000007f31/tablet-47d7f101463b4b879b635edf21757e71.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.374763 14245 tablet_snapshots.cc:369] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.374953 29469 tablet_snapshots.cc:369] T 529052c83d404685a99316a121081e98 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000b861/tablet-529052c83d404685a99316a121081e98.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.375635 25922 tablet_snapshots.cc:369] T c225ba07cc994356b3114acc5636ff0b P 1bfe87e507ee42949f2e9
[...]
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:11:25.659945 16321 tablet_snapshots.cc:369] T 2968a5ed73d9489a9fcc7c479ec73580 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-2968a5ed73d9489a9fcc7c479ec73580.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:11:45.670030 2004 tablet_snapshots.cc:369] T 640077cf65e248e997ff93e6b4bc4da4 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-640077cf65e248e997ff93e6b4bc4da4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:12:05.677551 5766 tablet_snapshots.cc:369] T cdc24b57bd3e4ff2bcba9cdda141d255 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d1/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-cdc24b57bd3e4ff2bcba9cdda141d255.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:52:11.608462 2039 tablet_snapshots.cc:369] T 75af014231c843aab48023aae46d818c P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd14/tablet-75af014231c843aab48023aae46d818c.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:52:21.614275 2008 tablet_snapshots.cc:369] T 93a6d6ca32e74d2aac11a4ba361ca6b9 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd19/tablet-93a6d6ca32e74d2aac11a4ba361ca6b9.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:53:46.652258 8023 tablet_snapshots.cc:369] T 5f16be40a8ac422fb40c329c2c4141cb P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d1/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd37/tablet-5f16be40a8ac422fb40c329c2c4141cb.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
Number of rpc requests sent upto the 8th May (it is still going now, so this is nearly twice this now)
```
### Number of times that the snapshots of the tablets have been attempted to be deleted in a 15 day period
```
[yugabyte@ip-172-159-39-237 ~]$ cat /tmp/loggs|sed 's/^.*(table //'|sed 's/]).*$/])/'|grep "id="|wc -l
1976706
```
[DB-417]: https://yugabyte.atlassian.net/browse/DB-417?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [DocDB] Deleting snapshot never finishes - Jira Link: [[DB-417]](https://yugabyte.atlassian.net/browse/DB-417)
### Description
On our cluster a backup was attempted on the 25th April 2022, this created a database snapshot, around 15 minutes after creating the snapshot the snapshot was deleted (or it's deletion was started), since then it has been unsuccessfully trying to find the leader for a large number of tablets, ever 10 minutes it fires 20 RPC requests per tablet to at least one server, sometimes multiple, this ends up with 1000's of unwanted and un-needed rpc requests on an already busy system. We are on a pre-release 2.13 build, I have put full logs on Google Drive (this is a 39 node cluster so there a lot of them). This is affecting 207 tablets.
### Snapshot
```
[yugabyte@ip-172-159-39-237 ~]$ master/bin/yb-admin -master_addresses `hostname -i`:7100 list_snapshots
Snapshot UUID State Creation Time
51fbb256-1923-42bf-a5a7-4b865ade9002 DELETING 2022-04-25 16:44:10.655943
```
### Log file out from just one of the tservers
```
8e1d886684d: Started tablet snapshot creation in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:10.661648 14742 checkpoint.cc:92] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d [I]: Started the snapshot process -- creating snapshot in directory /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp.intents
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:10.662571 14742 checkpoint.cc:92] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d [R]: Started the snapshot process -- creating snapshot in directory /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp
[...]
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.428580 10363 tablet_snapshots.cc:411] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Checkpoint created in /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.429877 10363 version_set.cc:2842] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Recovered from manifest file:/mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002.tmp/MANIFEST-000144 succeeded,manifest_file_number is 144, next_file_number is 162, last_sequence is 1125899945636174, log_number is 0,prev_log_number is 0,max_column_family is 0, flushed_values is 0x0000000124b31180 -> { op_id: 19.53572278 hybrid_time: { physical: 1650904793405367 } history_cutoff: { physical: 1650895988300693 } hybrid_time_filter: <invalid> max_value_level_ttl_expiration_time: <initial> }
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 16:44:11.430670 10363 tablet_snapshots.cc:175] T 05584779522d4ab08b83ea0a2dea0791 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot creation in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fa0c/tablet-05584779522d4ab08b83ea0a2dea0791.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002, snapshot hybrid time: { physical: 1650905050655943 }
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.373879 24136 tablet_snapshots.cc:369] T 47d7f101463b4b879b635edf21757e71 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-00004001000030008000000000007f31/tablet-47d7f101463b4b879b635edf21757e71.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.374763 14245 tablet_snapshots.cc:369] T c691fcb974c54da89e758489f85a13d4 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-000040010000300080000000000061c7/tablet-c691fcb974c54da89e758489f85a13d4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.374953 29469 tablet_snapshots.cc:369] T 529052c83d404685a99316a121081e98 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000b861/tablet-529052c83d404685a99316a121081e98.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220424-192601.22298.gz:I0425 17:02:31.375635 25922 tablet_snapshots.cc:369] T c225ba07cc994356b3114acc5636ff0b P 1bfe87e507ee42949f2e9
[...]
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:11:25.659945 16321 tablet_snapshots.cc:369] T 2968a5ed73d9489a9fcc7c479ec73580 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-2968a5ed73d9489a9fcc7c479ec73580.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:11:45.670030 2004 tablet_snapshots.cc:369] T 640077cf65e248e997ff93e6b4bc4da4 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-640077cf65e248e997ff93e6b4bc4da4.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:12:05.677551 5766 tablet_snapshots.cc:369] T cdc24b57bd3e4ff2bcba9cdda141d255 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d1/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fb20/tablet-cdc24b57bd3e4ff2bcba9cdda141d255.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:52:11.608462 2039 tablet_snapshots.cc:369] T 75af014231c843aab48023aae46d818c P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd14/tablet-75af014231c843aab48023aae46d818c.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:52:21.614275 2008 tablet_snapshots.cc:369] T 93a6d6ca32e74d2aac11a4ba361ca6b9 P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d0/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd19/tablet-93a6d6ca32e74d2aac11a4ba361ca6b9.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
172.159.51.178/tserver/yb-tserver.ip-172-159-51-178.eu-west-1.compute.internal.yugabyte.log.INFO.20220508-223844.1948.gz:I0509 22:53:46.652258 8023 tablet_snapshots.cc:369] T 5f16be40a8ac422fb40c329c2c4141cb P 1bfe87e507ee42949f2e98e1d886684d: Complete snapshot deletion on tablet in folder: /mnt/d1/yb-data/tserver/data/rocksdb/table-0000400100003000800000000000fd37/tablet-5f16be40a8ac422fb40c329c2c4141cb.snapshots/51fbb256-1923-42bf-a5a7-4b865ade9002
Number of rpc requests sent upto the 8th May (it is still going now, so this is nearly twice this now)
```
### Number of times that the snapshots of the tablets have been attempted to be deleted in a 15 day period
```
[yugabyte@ip-172-159-39-237 ~]$ cat /tmp/loggs|sed 's/^.*(table //'|sed 's/]).*$/])/'|grep "id="|wc -l
1976706
```
[DB-417]: https://yugabyte.atlassian.net/browse/DB-417?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | deleting snapshot never finishes jira link description on our cluster a backup was attempted on the april this created a database snapshot around minutes after creating the snapshot the snapshot was deleted or it s deletion was started since then it has been unsuccessfully trying to find the leader for a large number of tablets ever minutes it fires rpc requests per tablet to at least one server sometimes multiple this ends up with s of unwanted and un needed rpc requests on an already busy system we are on a pre release build i have put full logs on google drive this is a node cluster so there a lot of them this is affecting tablets snapshot master bin yb admin master addresses hostname i list snapshots snapshot uuid state creation time deleting log file out from just one of the tservers started tablet snapshot creation in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz checkpoint cc t p started the snapshot process creating snapshot in directory mnt yb data tserver data rocksdb table tablet snapshots tmp intents tserver yb tserver ip eu west compute internal yugabyte log info gz checkpoint cc t p started the snapshot process creating snapshot in directory mnt yb data tserver data rocksdb table tablet snapshots tmp tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p checkpoint created in mnt yb data tserver data rocksdb table tablet snapshots tmp tserver yb tserver ip eu west compute internal yugabyte log info gz version set cc t p recovered from manifest file mnt yb data tserver data rocksdb table tablet snapshots tmp manifest succeeded manifest file number is next file number is last sequence is log number is prev log number is max column family is flushed values is op id hybrid time physical history cutoff physical hybrid time filter max value level ttl expiration time tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot creation in folder mnt yb data tserver data rocksdb table tablet snapshots snapshot hybrid time physical tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots tserver yb tserver ip eu west compute internal yugabyte log info gz tablet snapshots cc t p complete snapshot deletion on tablet in folder mnt yb data tserver data rocksdb table tablet snapshots number of rpc requests sent upto the may it is still going now so this is nearly twice this now number of times that the snapshots of the tablets have been attempted to be deleted in a day period cat tmp loggs sed s table sed s grep id wc l | 1 |
41,498 | 2,869,010,362 | IssuesEvent | 2015-06-05 22:33:38 | dart-lang/intl | https://api.github.com/repos/dart-lang/intl | closed | Escaping for plural and gender expressions using the JSON format seems to be broken | AssumedStale bug Priority-Medium | <a href="https://github.com/alan-knight"><img src="https://avatars.githubusercontent.com/u/3476088?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [alan-knight](https://github.com/alan-knight)**
_Originally opened as dart-lang/sdk#19403_
----
From the dart mailing list:
...
I translated it, and produced the following file (for generate_from_json.dart ):
{
"_locale":"EN",
"playerTextStatsIConcededGoals" : "${Intl.plural(value, one: 'conceded goal', few: 'conceded ${value} goals', other: 'conceded ${value} goals')}"
}
Key here is exactly the "name" and value has the same structure as "message", but the translated strings.
After using generate_from_json.dart as specified, I get 2 files, where one is for EN locale, containing following function:
static playerTextStatsIConcededGoals(value) => "\${Intl.plural(value, one: \'conceded goal\', few: \'conceded ${value} goals\', other: \'conceded ${value} goals\')}";
As the special characters are escaped, it produces raw, ugly string. However, when I remove the backslashes, it appears to work fine. Why does it escape them? While deleting the backslashes manually, I got a feeling this might not be the right way of doing it.
=======================
The test cases actually avoid the escaping issues by serializing the plural and gender structures. So I'm guessing that the de-serializing improperly escapes things it shouldn't.
| 1.0 | Escaping for plural and gender expressions using the JSON format seems to be broken - <a href="https://github.com/alan-knight"><img src="https://avatars.githubusercontent.com/u/3476088?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [alan-knight](https://github.com/alan-knight)**
_Originally opened as dart-lang/sdk#19403_
----
From the dart mailing list:
...
I translated it, and produced the following file (for generate_from_json.dart ):
{
"_locale":"EN",
"playerTextStatsIConcededGoals" : "${Intl.plural(value, one: 'conceded goal', few: 'conceded ${value} goals', other: 'conceded ${value} goals')}"
}
Key here is exactly the "name" and value has the same structure as "message", but the translated strings.
After using generate_from_json.dart as specified, I get 2 files, where one is for EN locale, containing following function:
static playerTextStatsIConcededGoals(value) => "\${Intl.plural(value, one: \'conceded goal\', few: \'conceded ${value} goals\', other: \'conceded ${value} goals\')}";
As the special characters are escaped, it produces raw, ugly string. However, when I remove the backslashes, it appears to work fine. Why does it escape them? While deleting the backslashes manually, I got a feeling this might not be the right way of doing it.
=======================
The test cases actually avoid the escaping issues by serializing the plural and gender structures. So I'm guessing that the de-serializing improperly escapes things it shouldn't.
| priority | escaping for plural and gender expressions using the json format seems to be broken issue by originally opened as dart lang sdk from the dart mailing list i translated it and produced the following file for generate from json dart quot locale quot quot en quot quot playertextstatsiconcededgoals quot quot intl plural value one conceded goal few conceded value goals other conceded value goals quot key here is exactly the quot name quot and value has the same structure as quot message quot but the translated strings after using generate from json dart as specified i get files where one is for en locale containing following function nbsp nbsp static playertextstatsiconcededgoals value gt quot intl plural value one conceded goal few conceded value goals other conceded value goals quot as the special characters are escaped it produces raw ugly string however when i remove the backslashes it appears to work fine why does it escape them while deleting the backslashes manually i got a feeling this might not be the right way of doing it equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals equals the test cases actually avoid the escaping issues by serializing the plural and gender structures so i m guessing that the de serializing improperly escapes things it shouldn t | 1 |
51,817 | 3,014,304,322 | IssuesEvent | 2015-07-29 14:15:33 | scalan/scalan | https://api.github.com/repos/scalan/scalan | opened | Implement Shapeless-based tuples/case classes access | enhancement medium priority | This should help with #1 and make for simpler graphs and more efficient output. | 1.0 | Implement Shapeless-based tuples/case classes access - This should help with #1 and make for simpler graphs and more efficient output. | priority | implement shapeless based tuples case classes access this should help with and make for simpler graphs and more efficient output | 1 |
15,944 | 2,611,533,928 | IssuesEvent | 2015-02-27 06:04:36 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | keyboard problems in engine chat | auto-migrated Component-Engine Milestone-NextRelease Priority-Medium Type-Regression | ```
1. chat history navigation is not working
2. nick completion is not working
3. commands like '/fullscreen' do not work in multiplayer mode
Tested under OS X
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 23 Oct 2013 at 2:00 | 1.0 | keyboard problems in engine chat - ```
1. chat history navigation is not working
2. nick completion is not working
3. commands like '/fullscreen' do not work in multiplayer mode
Tested under OS X
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 23 Oct 2013 at 2:00 | priority | keyboard problems in engine chat chat history navigation is not working nick completion is not working commands like fullscreen do not work in multiplayer mode tested under os x original issue reported on code google com by vittorio gmail com on oct at | 1 |
376,949 | 11,160,574,978 | IssuesEvent | 2019-12-26 10:12:38 | qlcchain/go-qlc | https://api.github.com/repos/qlcchain/go-qlc | closed | sign the released binaries with gpg | Priority: Medium Type: Maintenance | ### Description of the issue
import gpg private key to CI
### Issue-Type
- [ ] bug report
- [x] feature request
- [ ] Documentation improvement | 1.0 | sign the released binaries with gpg - ### Description of the issue
import gpg private key to CI
### Issue-Type
- [ ] bug report
- [x] feature request
- [ ] Documentation improvement | priority | sign the released binaries with gpg description of the issue import gpg private key to ci issue type bug report feature request documentation improvement | 1 |
734,081 | 25,337,710,092 | IssuesEvent | 2022-11-18 18:20:51 | zowe/zowe-explorer-intellij | https://api.github.com/repos/zowe/zowe-explorer-intellij | closed | 'Zowe Explorer 0.2.1' installation does remove 'For Mainframe 0.6.1' from IntellJ 2021.3.2 and vice versa | bug priority-medium severity-medium | Hi,
I have upgraded 'For Mainframe' plugin to v 0.6.1. Then I installed 'Zowe Explorer' v0.2.1.
I expected to see both plugins available in my IntellJ 2021.3.2 Community Edition.
But it seems 'Zowe Explorer' replaced 'For Mainframe'.
I tried to install 'For Mainframe' plugin once again.
This time 'Zowe Explorer' disappeared from the list of installed plugins.
The restart of IDE did not help. | 1.0 | 'Zowe Explorer 0.2.1' installation does remove 'For Mainframe 0.6.1' from IntellJ 2021.3.2 and vice versa - Hi,
I have upgraded 'For Mainframe' plugin to v 0.6.1. Then I installed 'Zowe Explorer' v0.2.1.
I expected to see both plugins available in my IntellJ 2021.3.2 Community Edition.
But it seems 'Zowe Explorer' replaced 'For Mainframe'.
I tried to install 'For Mainframe' plugin once again.
This time 'Zowe Explorer' disappeared from the list of installed plugins.
The restart of IDE did not help. | priority | zowe explorer installation does remove for mainframe from intellj and vice versa hi i have upgraded for mainframe plugin to v then i installed zowe explorer i expected to see both plugins available in my intellj community edition but it seems zowe explorer replaced for mainframe i tried to install for mainframe plugin once again this time zowe explorer disappeared from the list of installed plugins the restart of ide did not help | 1 |
71,739 | 3,367,617,954 | IssuesEvent | 2015-11-22 10:19:05 | music-encoding/music-encoding | https://api.github.com/repos/music-encoding/music-encoding | closed | <instrumentation> and subelements should allow att.edit | Priority: Medium | _From [klaus.re...@gmail.com](https://code.google.com/u/105834722588578911678/) on January 29, 2015 08:11:10_
At some point it may not be possible to identify the original instrumation of a work (completely). So it should be possible to mark the coding with @-cert.
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=227_ | 1.0 | <instrumentation> and subelements should allow att.edit - _From [klaus.re...@gmail.com](https://code.google.com/u/105834722588578911678/) on January 29, 2015 08:11:10_
At some point it may not be possible to identify the original instrumation of a work (completely). So it should be possible to mark the coding with @-cert.
_Original issue: http://code.google.com/p/music-encoding/issues/detail?id=227_ | priority | and subelements should allow att edit from on january at some point it may not be possible to identify the original instrumation of a work completely so it should be possible to mark the coding with cert original issue | 1 |
174,119 | 6,536,793,083 | IssuesEvent | 2017-08-31 19:35:21 | k0shk0sh/FastHub | https://api.github.com/repos/k0shk0sh/FastHub | closed | Enabling the Android Move To SD Card Feature? | Priority: Medium Status: Completed Type: Feature Request | I didn't find an issue about this topic so I wanted to ask if it is possible to add the optional "Move to SD Card" feature.
I myself haven't yet tried to add it or even developed Android Apps yet but after looking at the following two websites it doesn't seem very difficult (from my point of view - I am just interested in adding this feature because I have only limited internal system space so enabling this feature would be very cool):
* https://code.tutsplus.com/tutorials/quick-tip-enabling-the-android-move-to-sd-card-feature--mobile-862
* https://developer.android.com/guide/topics/data/install-location.html
It seems like you only need to modify the android manifest by adding `android:installLocation="auto"` like in the following example from [the first website](https://code.tutsplus.com/tutorials/quick-tip-enabling-the-android-move-to-sd-card-feature--mobile-862):
```
<manifest xmlns:android="http://schemas.android.com/apk/res/android" package="com.mamlambo.article.phrasebook" android:versionCode="2" android:versionName="1.0.1"
android:installLocation="auto">
...
</manifest>
```
I really like the app and hope that I didn't wasted your time with something that obviously has reasons to not be activated or can't be activated because of some services or something other.
Also sorry for my probably not so good English.
😄
**FastHub Version:** 4.1.0
***Android Version:** 6.1.0*
***Device Information:** ONE E1003* | 1.0 | Enabling the Android Move To SD Card Feature? - I didn't find an issue about this topic so I wanted to ask if it is possible to add the optional "Move to SD Card" feature.
I myself haven't yet tried to add it or even developed Android Apps yet but after looking at the following two websites it doesn't seem very difficult (from my point of view - I am just interested in adding this feature because I have only limited internal system space so enabling this feature would be very cool):
* https://code.tutsplus.com/tutorials/quick-tip-enabling-the-android-move-to-sd-card-feature--mobile-862
* https://developer.android.com/guide/topics/data/install-location.html
It seems like you only need to modify the android manifest by adding `android:installLocation="auto"` like in the following example from [the first website](https://code.tutsplus.com/tutorials/quick-tip-enabling-the-android-move-to-sd-card-feature--mobile-862):
```
<manifest xmlns:android="http://schemas.android.com/apk/res/android" package="com.mamlambo.article.phrasebook" android:versionCode="2" android:versionName="1.0.1"
android:installLocation="auto">
...
</manifest>
```
I really like the app and hope that I didn't wasted your time with something that obviously has reasons to not be activated or can't be activated because of some services or something other.
Also sorry for my probably not so good English.
😄
**FastHub Version:** 4.1.0
***Android Version:** 6.1.0*
***Device Information:** ONE E1003* | priority | enabling the android move to sd card feature i didn t find an issue about this topic so i wanted to ask if it is possible to add the optional move to sd card feature i myself haven t yet tried to add it or even developed android apps yet but after looking at the following two websites it doesn t seem very difficult from my point of view i am just interested in adding this feature because i have only limited internal system space so enabling this feature would be very cool it seems like you only need to modify the android manifest by adding android installlocation auto like in the following example from manifest xmlns android package com mamlambo article phrasebook android versioncode android versionname android installlocation auto i really like the app and hope that i didn t wasted your time with something that obviously has reasons to not be activated or can t be activated because of some services or something other also sorry for my probably not so good english 😄 fasthub version android version device information one | 1 |
727,784 | 25,046,314,732 | IssuesEvent | 2022-11-05 09:43:33 | Chatterino/chatterino2 | https://api.github.com/repos/Chatterino/chatterino2 | closed | Migrate /commercial command to Helix API | Platform: Twitch Priority: Medium Deprecation: Twitch IRC Commands hacktoberfest | As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/commercial` command needs to be migrated to use the relevant Helix API endpoint.
Helix API reference: https://dev.twitch.tv/docs/api/reference#start-commercial
Split from #3956 | 1.0 | Migrate /commercial command to Helix API - As part of Twitch's announced deprecation of IRC-based commands ([see here for more info](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486), the `/commercial` command needs to be migrated to use the relevant Helix API endpoint.
Helix API reference: https://dev.twitch.tv/docs/api/reference#start-commercial
Split from #3956 | priority | migrate commercial command to helix api as part of twitch s announced deprecation of irc based commands the commercial command needs to be migrated to use the relevant helix api endpoint helix api reference split from | 1 |
527,397 | 15,342,070,838 | IssuesEvent | 2021-02-27 14:46:05 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Show the Photo Albums in the main photo directory | feature: enhancement priority: medium | **Is your feature request related to a problem? Please describe.**
There is a link to create an album on the main photos directory but there is no way to see them on the main photos directory. Currently, you will need to go to each of the profile pages to see the available photo albums.
**Support ticket links**
https://secure.helpscout.net/conversation/1432811307/126496
| 1.0 | Show the Photo Albums in the main photo directory - **Is your feature request related to a problem? Please describe.**
There is a link to create an album on the main photos directory but there is no way to see them on the main photos directory. Currently, you will need to go to each of the profile pages to see the available photo albums.
**Support ticket links**
https://secure.helpscout.net/conversation/1432811307/126496
| priority | show the photo albums in the main photo directory is your feature request related to a problem please describe there is a link to create an album on the main photos directory but there is no way to see them on the main photos directory currently you will need to go to each of the profile pages to see the available photo albums support ticket links | 1 |
789,358 | 27,787,790,320 | IssuesEvent | 2023-03-17 05:51:08 | ainc/ainc-gatsby-sanity | https://api.github.com/repos/ainc/ainc-gatsby-sanity | opened | Create /team-alpha page | Priority: High Difficulty: Medium | https://www.awesomeinc.org/team-alpha
- Sanity Doc
- Name
- Picture
- Team/Role
- Favorite Rule
- Favorite Song
- Favorite Person
- Random Fact | 1.0 | Create /team-alpha page - https://www.awesomeinc.org/team-alpha
- Sanity Doc
- Name
- Picture
- Team/Role
- Favorite Rule
- Favorite Song
- Favorite Person
- Random Fact | priority | create team alpha page sanity doc name picture team role favorite rule favorite song favorite person random fact | 1 |
242,832 | 7,849,149,416 | IssuesEvent | 2018-06-20 01:45:58 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: eEvveerryytthhiinngg iiss ddoouubblleedd tteexxtt.. | Medium Priority | **Version:** 0.7.3.2 beta
**Steps to Reproduce:**
lLaauunncchh tthhee llaatteesstt eeccoo oonn lliinnuuxx..
**Expected behavior:**
wweellccoommee ssccrreeeenn wwiitthh uusseerrnnaammee//ppaasssswwoorrdd
**Actual behavior:**
aAss yyoouu ccaann sseeee eevveerryytthhiinngg iinn ggaammee iiss ddoouubbllee tteexxtt.. | 1.0 | USER ISSUE: eEvveerryytthhiinngg iiss ddoouubblleedd tteexxtt.. - **Version:** 0.7.3.2 beta
**Steps to Reproduce:**
lLaauunncchh tthhee llaatteesstt eeccoo oonn lliinnuuxx..
**Expected behavior:**
wweellccoommee ssccrreeeenn wwiitthh uusseerrnnaammee//ppaasssswwoorrdd
**Actual behavior:**
aAss yyoouu ccaann sseeee eevveerryytthhiinngg iinn ggaammee iiss ddoouubbllee tteexxtt.. | priority | user issue eevveerryytthhiinngg iiss ddoouubblleedd tteexxtt version beta steps to reproduce llaauunncchh tthhee llaatteesstt eeccoo oonn lliinnuuxx expected behavior wweellccoommee ssccrreeeenn wwiitthh uusseerrnnaammee ppaasssswwoorrdd actual behavior aass yyoouu ccaann sseeee eevveerryytthhiinngg iinn ggaammee iiss ddoouubbllee tteexxtt | 1 |
583,364 | 17,383,366,056 | IssuesEvent | 2021-08-01 06:13:44 | eatmyvenom/hyarcade | https://api.github.com/repos/eatmyvenom/hyarcade | closed | [Refractor] improve leaderboard generation speed. | Medium priority disc:overall refractor t:discord | utilizing some advanced JS features this can be sped up a bit. This should be done to improve quality of life overall.
| 1.0 | [Refractor] improve leaderboard generation speed. - utilizing some advanced JS features this can be sped up a bit. This should be done to improve quality of life overall.
| priority | improve leaderboard generation speed utilizing some advanced js features this can be sped up a bit this should be done to improve quality of life overall | 1 |
804,449 | 29,488,573,818 | IssuesEvent | 2023-06-02 11:45:55 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID: 316635] Macro compares unsigned to 0 in subsys/net/l2/ethernet/gptp/gptp_md.c | bug priority: medium area: Networking Coverity |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/dae79cefaabf63086946a48ccca4094f26f146c8/subsys/net/l2/ethernet/gptp/gptp_md.c#L478
Category: Integer handling issues
Function: `gptp_md_follow_up_receipt_timeout`
Component: Networking
CID: [316635](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=316635)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/dae79cefaabf63086946a48ccca4094f26f146c8/subsys/net/l2/ethernet/gptp/gptp_md.c#L478
```
472 struct gptp_sync_rcv_state *state;
473 int port;
474
475 for (port = GPTP_PORT_START; port < GPTP_PORT_END; port++) {
476 state = &GPTP_PORT_STATE(port)->sync_rcv;
477 if (timer == &state->follow_up_discard_timer) {
>>> CID 316635: Integer handling issues (NO_EFFECT)
>>> This less-than-zero comparison of an unsigned value is never true. "0UL < 0UL".
478 NET_WARN("No %s received after %s message",
479 "FOLLOWUP", "SYNC");
480 state->follow_up_timeout_expired = true;
481 }
482 }
483 }
```
For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_NO_EFFECT). ([CWE-570](http://cwe.mitre.org/data/definitions/570.html))
Please fix or provide comments in coverity using the link:
https://scan9.scan.coverity.com/reports.htm#v29271/p12996
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID: 316635] Macro compares unsigned to 0 in subsys/net/l2/ethernet/gptp/gptp_md.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/dae79cefaabf63086946a48ccca4094f26f146c8/subsys/net/l2/ethernet/gptp/gptp_md.c#L478
Category: Integer handling issues
Function: `gptp_md_follow_up_receipt_timeout`
Component: Networking
CID: [316635](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=316635)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/dae79cefaabf63086946a48ccca4094f26f146c8/subsys/net/l2/ethernet/gptp/gptp_md.c#L478
```
472 struct gptp_sync_rcv_state *state;
473 int port;
474
475 for (port = GPTP_PORT_START; port < GPTP_PORT_END; port++) {
476 state = &GPTP_PORT_STATE(port)->sync_rcv;
477 if (timer == &state->follow_up_discard_timer) {
>>> CID 316635: Integer handling issues (NO_EFFECT)
>>> This less-than-zero comparison of an unsigned value is never true. "0UL < 0UL".
478 NET_WARN("No %s received after %s message",
479 "FOLLOWUP", "SYNC");
480 state->follow_up_timeout_expired = true;
481 }
482 }
483 }
```
For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_NO_EFFECT). ([CWE-570](http://cwe.mitre.org/data/definitions/570.html))
Please fix or provide comments in coverity using the link:
https://scan9.scan.coverity.com/reports.htm#v29271/p12996
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | macro compares unsigned to in subsys net ethernet gptp gptp md c static code scan issues found in file category integer handling issues function gptp md follow up receipt timeout component networking cid details struct gptp sync rcv state state int port for port gptp port start port gptp port end port state gptp port state port sync rcv if timer state follow up discard timer cid integer handling issues no effect this less than zero comparison of an unsigned value is never true net warn no s received after s message followup sync state follow up timeout expired true for more information about the violation check the please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
286,243 | 8,785,665,487 | IssuesEvent | 2018-12-20 13:41:56 | pravega/pravega-operator | https://api.github.com/repos/pravega/pravega-operator | opened | Investigate Bookkeeper disruption cases | kind/enhancement priority/medium status/needs-investigation | We need to investigate and implement an action plan to cover all BookKeeper disruption cases. Some disruption cases that come to my mind are:
- Graceful termination. Pod will receive an TERM signal to gracefully shutdown. Some situations that cause this are:
- `kubectl drain` to remove a node from the K8 cluster.
- `kubectl delete pod` to delete a particular Bookkeeper pod, probably accidentally.
- Scale down event.
- Unexpected termination. This kind of disruptions will give pods no chance to gracefully terminate. Examples are hardware errors, VM deletion, [hard evictions](https://kubernetes.io/docs/tasks/administer-cluster/out-of-resource/#hard-eviction-thresholds), etc.
For graceful terminations, we may want to run a pre-delete hook, [`preStop handler`](https://kubernetes.io/docs/tasks/configure-pod-container/attach-handler-lifecycle-event/) in K8 terminology, to make sure that ledgers stored in that Bookie are rereplicated before it is shut down. Probably by running a [Bookkeeper manual recovery](https://bookkeeper.apache.org/docs/4.7.0/admin/autorecovery/) process.
For unexpected terminations, we may want to rely on Bookkeeper's [autorecovery](https://bookkeeper.apache.org/docs/4.7.0/admin/autorecovery/) feature and the pod disruption budget to prevent a second pod graceful termination until the terminated pod is rescheduled and recovered.
| 1.0 | Investigate Bookkeeper disruption cases - We need to investigate and implement an action plan to cover all BookKeeper disruption cases. Some disruption cases that come to my mind are:
- Graceful termination. Pod will receive an TERM signal to gracefully shutdown. Some situations that cause this are:
- `kubectl drain` to remove a node from the K8 cluster.
- `kubectl delete pod` to delete a particular Bookkeeper pod, probably accidentally.
- Scale down event.
- Unexpected termination. This kind of disruptions will give pods no chance to gracefully terminate. Examples are hardware errors, VM deletion, [hard evictions](https://kubernetes.io/docs/tasks/administer-cluster/out-of-resource/#hard-eviction-thresholds), etc.
For graceful terminations, we may want to run a pre-delete hook, [`preStop handler`](https://kubernetes.io/docs/tasks/configure-pod-container/attach-handler-lifecycle-event/) in K8 terminology, to make sure that ledgers stored in that Bookie are rereplicated before it is shut down. Probably by running a [Bookkeeper manual recovery](https://bookkeeper.apache.org/docs/4.7.0/admin/autorecovery/) process.
For unexpected terminations, we may want to rely on Bookkeeper's [autorecovery](https://bookkeeper.apache.org/docs/4.7.0/admin/autorecovery/) feature and the pod disruption budget to prevent a second pod graceful termination until the terminated pod is rescheduled and recovered.
| priority | investigate bookkeeper disruption cases we need to investigate and implement an action plan to cover all bookkeeper disruption cases some disruption cases that come to my mind are graceful termination pod will receive an term signal to gracefully shutdown some situations that cause this are kubectl drain to remove a node from the cluster kubectl delete pod to delete a particular bookkeeper pod probably accidentally scale down event unexpected termination this kind of disruptions will give pods no chance to gracefully terminate examples are hardware errors vm deletion etc for graceful terminations we may want to run a pre delete hook in terminology to make sure that ledgers stored in that bookie are rereplicated before it is shut down probably by running a process for unexpected terminations we may want to rely on bookkeeper s feature and the pod disruption budget to prevent a second pod graceful termination until the terminated pod is rescheduled and recovered | 1 |
46,677 | 2,964,265,421 | IssuesEvent | 2015-07-10 15:41:17 | Sonarr/Sonarr | https://api.github.com/repos/Sonarr/Sonarr | opened | Remove Failing Pending Items | priority:medium proposal suboptimal | If a pending download fails a number of times we should remove it from the pending queue instead of trying over and over. Recently I've seen cases where the torrent has ben removed or there is no torrent client to be sent to. | 1.0 | Remove Failing Pending Items - If a pending download fails a number of times we should remove it from the pending queue instead of trying over and over. Recently I've seen cases where the torrent has ben removed or there is no torrent client to be sent to. | priority | remove failing pending items if a pending download fails a number of times we should remove it from the pending queue instead of trying over and over recently i ve seen cases where the torrent has ben removed or there is no torrent client to be sent to | 1 |
438,325 | 12,626,399,331 | IssuesEvent | 2020-06-14 16:24:13 | rikatz/kubepug | https://api.github.com/repos/rikatz/kubepug | closed | Feature: Use YAML instead of Kubernetes as validation source | Priority/Medium enhancement | The tool currently checks objects in Kubernetes, which is great. I would also like to add it to our CI/CD to check manifests before they make it to Kubernetes. In that regards, I'd like to either pipe YAML/JSON into the tool and it provides the same output including error codes for the pipeline to fail.
Let me know your thoughts. | 1.0 | Feature: Use YAML instead of Kubernetes as validation source - The tool currently checks objects in Kubernetes, which is great. I would also like to add it to our CI/CD to check manifests before they make it to Kubernetes. In that regards, I'd like to either pipe YAML/JSON into the tool and it provides the same output including error codes for the pipeline to fail.
Let me know your thoughts. | priority | feature use yaml instead of kubernetes as validation source the tool currently checks objects in kubernetes which is great i would also like to add it to our ci cd to check manifests before they make it to kubernetes in that regards i d like to either pipe yaml json into the tool and it provides the same output including error codes for the pipeline to fail let me know your thoughts | 1 |
696,284 | 23,895,173,198 | IssuesEvent | 2022-09-08 14:18:51 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] Snapshots - Push down the filters to master instead of filtering at yb-admin client | kind/bug area/docdb priority/medium | Jira Link: [DB-3199](https://yugabyte.atlassian.net/browse/DB-3199)
### Description
Currently, when we issue `list_snapshots` command via `yb-admin`, we don't pass down certain filters to the master in the request pb and have the master filter them out accordingly. Instead, we get everything from the master and filter at the client. This creates bottlenecks on the network wire if there are a large number of snapshots of lots of tables/tablets. In particular:
1. Irrespective of whether the user passes in the `show_details` option or not, the `master` always includes all the object entries (tables/tablets/namespace) in the response PB to `yb-admin`. `yb-admin` then filters and only displays them if the `show_details` option was passed in.
2. Even if `show_details` option is passed by the user, we only display the TABLES, NAMESPACES, UDTYPES and don't display TABLETS but the response that the master returns has all 4 types.
We should push down these options in the request protobuf to the master and have the master filter them accordingly. | 1.0 | [DocDB] Snapshots - Push down the filters to master instead of filtering at yb-admin client - Jira Link: [DB-3199](https://yugabyte.atlassian.net/browse/DB-3199)
### Description
Currently, when we issue `list_snapshots` command via `yb-admin`, we don't pass down certain filters to the master in the request pb and have the master filter them out accordingly. Instead, we get everything from the master and filter at the client. This creates bottlenecks on the network wire if there are a large number of snapshots of lots of tables/tablets. In particular:
1. Irrespective of whether the user passes in the `show_details` option or not, the `master` always includes all the object entries (tables/tablets/namespace) in the response PB to `yb-admin`. `yb-admin` then filters and only displays them if the `show_details` option was passed in.
2. Even if `show_details` option is passed by the user, we only display the TABLES, NAMESPACES, UDTYPES and don't display TABLETS but the response that the master returns has all 4 types.
We should push down these options in the request protobuf to the master and have the master filter them accordingly. | priority | snapshots push down the filters to master instead of filtering at yb admin client jira link description currently when we issue list snapshots command via yb admin we don t pass down certain filters to the master in the request pb and have the master filter them out accordingly instead we get everything from the master and filter at the client this creates bottlenecks on the network wire if there are a large number of snapshots of lots of tables tablets in particular irrespective of whether the user passes in the show details option or not the master always includes all the object entries tables tablets namespace in the response pb to yb admin yb admin then filters and only displays them if the show details option was passed in even if show details option is passed by the user we only display the tables namespaces udtypes and don t display tablets but the response that the master returns has all types we should push down these options in the request protobuf to the master and have the master filter them accordingly | 1 |
636,067 | 20,574,056,260 | IssuesEvent | 2022-03-04 01:15:33 | giancarlopernudisegura/cmput404 | https://api.github.com/repos/giancarlopernudisegura/cmput404 | closed | [User Story]: Post Posts to Stream | question Priority Level: High Implementation Difficulty: Medium User Story | ### User Story
As an author, I want to post posts to my “stream”.
### Objects
Inbox | 1.0 | [User Story]: Post Posts to Stream - ### User Story
As an author, I want to post posts to my “stream”.
### Objects
Inbox | priority | post posts to stream user story as an author i want to post posts to my “stream” objects inbox | 1 |
172,291 | 6,501,397,692 | IssuesEvent | 2017-08-23 09:28:07 | Polpetta/SecurityAndRiskManagementNotes | https://api.github.com/repos/Polpetta/SecurityAndRiskManagementNotes | closed | Controlli delle frodi | bug enhancement in progress priority:medium question type:content | **Description**
I tipi di controlli delle frodi 1.3 secondo le slide sono 3 (preso dal libro essentials of corporate fraud).
I controlli compensativi (presi dal CISA Manual) che abbiamo spostato nella sezione 1.3 vengono messi in atto per contrastare l'assenza di segregation of duties (soprattutto in realtà con pochi dipendenti) e quindi **NON DOVREBBERO FARE PARTE** della sezione 1.3 . Ma dovrebbero essere sotto segregation of duty come nelle slide.


Cosa ne pensate @mzanella @Polpetta
**Other issues**
| 1.0 | Controlli delle frodi - **Description**
I tipi di controlli delle frodi 1.3 secondo le slide sono 3 (preso dal libro essentials of corporate fraud).
I controlli compensativi (presi dal CISA Manual) che abbiamo spostato nella sezione 1.3 vengono messi in atto per contrastare l'assenza di segregation of duties (soprattutto in realtà con pochi dipendenti) e quindi **NON DOVREBBERO FARE PARTE** della sezione 1.3 . Ma dovrebbero essere sotto segregation of duty come nelle slide.


Cosa ne pensate @mzanella @Polpetta
**Other issues**
| priority | controlli delle frodi description i tipi di controlli delle frodi secondo le slide sono preso dal libro essentials of corporate fraud i controlli compensativi presi dal cisa manual che abbiamo spostato nella sezione vengono messi in atto per contrastare l assenza di segregation of duties soprattutto in realtà con pochi dipendenti e quindi non dovrebbero fare parte della sezione ma dovrebbero essere sotto segregation of duty come nelle slide cosa ne pensate mzanella polpetta other issues | 1 |
19,608 | 2,622,155,504 | IssuesEvent | 2015-03-04 00:07:56 | byzhang/terrastore | https://api.github.com/repos/byzhang/terrastore | closed | Improve range queries performance. | auto-migrated Milestone-0.6.0 Priority-Medium Project-Terrastore Type-Enhancement | ```
Improve range queries performance by:
1) Merging sorted key sets rather than sorting everything again.
2) Making the merge process parallel.
3) Composing sorted keys with unsorted values through a zero-copy map with
sorted keys indirectly referring to mapped values.
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 4 Jun 2010 at 9:08 | 1.0 | Improve range queries performance. - ```
Improve range queries performance by:
1) Merging sorted key sets rather than sorting everything again.
2) Making the merge process parallel.
3) Composing sorted keys with unsorted values through a zero-copy map with
sorted keys indirectly referring to mapped values.
```
Original issue reported on code.google.com by `sergio.b...@gmail.com` on 4 Jun 2010 at 9:08 | priority | improve range queries performance improve range queries performance by merging sorted key sets rather than sorting everything again making the merge process parallel composing sorted keys with unsorted values through a zero copy map with sorted keys indirectly referring to mapped values original issue reported on code google com by sergio b gmail com on jun at | 1 |
594,515 | 18,047,338,720 | IssuesEvent | 2021-09-19 05:34:25 | AY2122S1-CS2103T-T13-4/tp | https://api.github.com/repos/AY2122S1-CS2103T-T13-4/tp | opened | Update Ui to include file import | priority.Medium priority.Low | Include a file import button/command to import from an excel file | 2.0 | Update Ui to include file import - Include a file import button/command to import from an excel file | priority | update ui to include file import include a file import button command to import from an excel file | 1 |
661,143 | 22,041,688,469 | IssuesEvent | 2022-05-29 13:00:51 | CheeseInthe-Life/Daru-iOS | https://api.github.com/repos/CheeseInthe-Life/Daru-iOS | closed | Delegate private 접근제어자 이슈 | Type: Bugfix/Function Priority: Medium Status: To Do Type: Bugfix/UI | # 발생 일시
22.05.29 21:49
# 발생 위치
homeViewController
# 증상
HeaderView안에 있는 버튼클릭액션을 처리하기 위해 Delegate Protocol을 정의하고 HomeViewController에 해당 프로토콜을 채택하려고 했지만 private 접근제어자 때문에 에러가 났다.
# 재현 방법
```
protocol MoreButtonDelegate: AnyObject {
func moreButtonDidTap(_ section: Int)
}
// MARK: - MoreButton Delegate
private extension HomeViewController: MoreButtonDelegate {
func moreButtonDidTap(_ section: Int) {
<#code#>
}
}
```
# 스크린샷

| 1.0 | Delegate private 접근제어자 이슈 - # 발생 일시
22.05.29 21:49
# 발생 위치
homeViewController
# 증상
HeaderView안에 있는 버튼클릭액션을 처리하기 위해 Delegate Protocol을 정의하고 HomeViewController에 해당 프로토콜을 채택하려고 했지만 private 접근제어자 때문에 에러가 났다.
# 재현 방법
```
protocol MoreButtonDelegate: AnyObject {
func moreButtonDidTap(_ section: Int)
}
// MARK: - MoreButton Delegate
private extension HomeViewController: MoreButtonDelegate {
func moreButtonDidTap(_ section: Int) {
<#code#>
}
}
```
# 스크린샷

| priority | delegate private 접근제어자 이슈 발생 일시 발생 위치 homeviewcontroller 증상 headerview안에 있는 버튼클릭액션을 처리하기 위해 delegate protocol을 정의하고 homeviewcontroller에 해당 프로토콜을 채택하려고 했지만 private 접근제어자 때문에 에러가 났다 재현 방법 protocol morebuttondelegate anyobject func morebuttondidtap section int mark morebutton delegate private extension homeviewcontroller morebuttondelegate func morebuttondidtap section int 스크린샷 | 1 |
257,900 | 8,148,392,973 | IssuesEvent | 2018-08-22 05:33:26 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2). | Bug C: Grid F: PDF Export Kendo1 Needs QA Priority 2 S: Wrappers (ASP.NET Core) SEV: Medium | ### Bug report
The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2).

### Reproduction of the problem
The issue can be reproduced in our demo when the filtering is on and the paperSize is set to auto in order to show a bigger part of the Grid.
### Environment
* **Browser:** [all]
| 1.0 | The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2). - ### Bug report
The Grid PDF export is not showing the icons and it is rendering the wrong header in MVC Core(Bootstrap v4 and Default2).

### Reproduction of the problem
The issue can be reproduced in our demo when the filtering is on and the paperSize is set to auto in order to show a bigger part of the Grid.
### Environment
* **Browser:** [all]
| priority | the grid pdf export is not showing the icons and it is rendering the wrong header in mvc core bootstrap and bug report the grid pdf export is not showing the icons and it is rendering the wrong header in mvc core bootstrap and reproduction of the problem the issue can be reproduced in our demo when the filtering is on and the papersize is set to auto in order to show a bigger part of the grid environment browser | 1 |
701,275 | 24,093,731,304 | IssuesEvent | 2022-09-19 16:48:31 | CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | closed | [Collaboration Service] Automatic cleanup | Module/Back-End Status/Medium-Priority Type/Feature | ## Description
The collaboration service should automatically clean up rooms with no users, and release all related resources.
## Parent Task
- #63 | 1.0 | [Collaboration Service] Automatic cleanup - ## Description
The collaboration service should automatically clean up rooms with no users, and release all related resources.
## Parent Task
- #63 | priority | automatic cleanup description the collaboration service should automatically clean up rooms with no users and release all related resources parent task | 1 |
448,997 | 12,961,677,478 | IssuesEvent | 2020-07-20 16:00:17 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | BVA Dispatch: Alphabetize Special Issues | Priority: Medium Product: caseflow-dispatch Stakeholder: BVA Team: Echo 🐬 Type: UI Change | <!-- The goal of this template is to be a tool to communicate the requirements for a story related task. It is not intended as a mandate, adapt as needed. -->
## User or job story
User story: As a BVA Dispatch user, I need to see the special issues on the special issues page during checkout alphabetized, so that I can read and select the options quickly.
## Acceptance criteria
- [ ] Please put this work behind the feature toggle: :special_issues_revamp
- [ ] This feature should be accessible to the following user groups: BVA Dispatch
- [ ] Include screenshot(s) in the Github issue if there are front-end changes
- [ ] Special Issues listed on the page are alphabetized including the new contentions added recently (MST, Burn Pit, CAVC, Blue Water)
### Technical notes
The special issues page only appears for Legacy cases at this time.
### Resources/other links
<!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
| 1.0 | BVA Dispatch: Alphabetize Special Issues - <!-- The goal of this template is to be a tool to communicate the requirements for a story related task. It is not intended as a mandate, adapt as needed. -->
## User or job story
User story: As a BVA Dispatch user, I need to see the special issues on the special issues page during checkout alphabetized, so that I can read and select the options quickly.
## Acceptance criteria
- [ ] Please put this work behind the feature toggle: :special_issues_revamp
- [ ] This feature should be accessible to the following user groups: BVA Dispatch
- [ ] Include screenshot(s) in the Github issue if there are front-end changes
- [ ] Special Issues listed on the page are alphabetized including the new contentions added recently (MST, Burn Pit, CAVC, Blue Water)
### Technical notes
The special issues page only appears for Legacy cases at this time.
### Resources/other links
<!-- E.g. links to other issues, PRs, Sentry alerts, or Slack threads, or external service requests. -->
| priority | bva dispatch alphabetize special issues user or job story user story as a bva dispatch user i need to see the special issues on the special issues page during checkout alphabetized so that i can read and select the options quickly acceptance criteria please put this work behind the feature toggle special issues revamp this feature should be accessible to the following user groups bva dispatch include screenshot s in the github issue if there are front end changes special issues listed on the page are alphabetized including the new contentions added recently mst burn pit cavc blue water technical notes the special issues page only appears for legacy cases at this time resources other links | 1 |
47,256 | 2,974,603,286 | IssuesEvent | 2015-07-15 02:16:59 | Reimashi/jotai | https://api.github.com/repos/Reimashi/jotai | closed | Asus Crosshair formula IV motherboard support | auto-migrated Priority-Medium Type-Enhancement | ```
What is the expected output? What do you see instead? 116.0
What version of the product are you using? On what operating system? 1.37 I
think
Please provide any additional information below. asus brosshair IV formula
amd am3 1090T Black
Please attach a Report created with "File / Save Report...".
Open Hardware Monitor Report
--------------------------------------------------------------------------------
Version: 0.1.37.0
--------------------------------------------------------------------------------
Common Language Runtime: 2.0.50727.4952
Operating System: Microsoft Windows NT 6.1.7600.0
Process Type: 64-Bit
--------------------------------------------------------------------------------
Sensors
|
+-+ ASUS Crosshair IV Formula (/mainboard)
|
+-+ AMD Phenom II X6 1090T Processor (/amdcpu/0)
| +- Temperature[0] : 0 : 0 : 0 : Core #1 - #6
| +- Load[0] : 7.696998 : 0 : 18.72208 : CPU Total
| +- Load[1] : 6.158614 : 0 : 24.61922 : Core #1
| +- Load[2] : 12.31214 : 0 : 100 : Core #2
| +- Load[3] : 1.543462 : 0 : 20.00406 : Core #3
| +- Load[4] : 4.62023 : 0 : 29.23436 : Core #4
| +- Load[5] : 7.696998 : 0 : 29.30409 : Core #5
| +- Load[6] : 13.85053 : 0 : 46.15658 : Core #6
|
+-+ NVIDIA GeForce GTX 480 (/nvidiagpu/0)
| +- Clock[0] : 50.625 : 0 : 405 : GPU Core
| +- Clock[1] : 135 : 135 : 324 : GPU Memory
| +- Clock[2] : 101.25 : 0 : 810.0001 : GPU Shader
| +- Temperature[0] : 52 : 52 : 55 : GPU Core
| +- Temperature[1] : 45 : 45 : 47 : GPU Board
| +- Load[0] : 1 : 0 : 49 : GPU Core
| +- Load[1] : 7 : 2 : 32 : GPU Memory Controller
| +- Load[2] : 0 : 0 : 0 : GPU Video Engine
| +- Load[3] : 11.77394 : 10.85408 : 12.31944 : GPU Memory
| +- Fan[0] : 1770 : 1740 : 1800 : GPU
| +- Control[0] : 44 : 44 : 44 : GPU Fan
|
+-+ WDC WD1002FAEX-00Y9A0 (/hdd/0)
| +- Temperature[0] : 42 : 42 : 44 : HDD
|
+-+ WDC WD6400AAKS-08A7B2 (/hdd/1)
| +- Temperature[0] : 42 : 42 : 42 : HDD
--------------------------------------------------------------------------------
Parameters
|
+-+ ASUS Crosshair IV Formula (/mainboard)
|
+-+ AMD Phenom II X6 1090T Processor (/amdcpu/0)
| +- Temperature[0] : Core #1 - #6
| +- Offset [°C] : 0 : 0
|
+-+ NVIDIA GeForce GTX 480 (/nvidiagpu/0)
|
+-+ WDC WD1002FAEX-00Y9A0 (/hdd/0)
|
+-+ WDC WD6400AAKS-08A7B2 (/hdd/1)
--------------------------------------------------------------------------------
Mainboard
BIOS Vendor: American Megatrends Inc.
BIOS Version: 0801
Mainboard Manufacturer: ASUSTeK Computer INC.
Mainboard Name: Crosshair IV Formula
SMBIOS Table
ABgAAAECAPADH5Dai38BAAAAMwUID///QW1lcmljYW4gTWVnYXRyZW5kcyBJbmMu
ADA4MDEgICAAMDUvMDMvMjAxMAAAARsBAAECAwRAmAAejAAA2/nIIM8wHwEZBgUG
```
Original issue reported on code.google.com by `MrPB...@gmail.com` on 10 Sep 2010 at 10:50
Attachments:
* [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-113/comment-0/OpenHardwareMonitor.Report.txt)
| 1.0 | Asus Crosshair formula IV motherboard support - ```
What is the expected output? What do you see instead? 116.0
What version of the product are you using? On what operating system? 1.37 I
think
Please provide any additional information below. asus brosshair IV formula
amd am3 1090T Black
Please attach a Report created with "File / Save Report...".
Open Hardware Monitor Report
--------------------------------------------------------------------------------
Version: 0.1.37.0
--------------------------------------------------------------------------------
Common Language Runtime: 2.0.50727.4952
Operating System: Microsoft Windows NT 6.1.7600.0
Process Type: 64-Bit
--------------------------------------------------------------------------------
Sensors
|
+-+ ASUS Crosshair IV Formula (/mainboard)
|
+-+ AMD Phenom II X6 1090T Processor (/amdcpu/0)
| +- Temperature[0] : 0 : 0 : 0 : Core #1 - #6
| +- Load[0] : 7.696998 : 0 : 18.72208 : CPU Total
| +- Load[1] : 6.158614 : 0 : 24.61922 : Core #1
| +- Load[2] : 12.31214 : 0 : 100 : Core #2
| +- Load[3] : 1.543462 : 0 : 20.00406 : Core #3
| +- Load[4] : 4.62023 : 0 : 29.23436 : Core #4
| +- Load[5] : 7.696998 : 0 : 29.30409 : Core #5
| +- Load[6] : 13.85053 : 0 : 46.15658 : Core #6
|
+-+ NVIDIA GeForce GTX 480 (/nvidiagpu/0)
| +- Clock[0] : 50.625 : 0 : 405 : GPU Core
| +- Clock[1] : 135 : 135 : 324 : GPU Memory
| +- Clock[2] : 101.25 : 0 : 810.0001 : GPU Shader
| +- Temperature[0] : 52 : 52 : 55 : GPU Core
| +- Temperature[1] : 45 : 45 : 47 : GPU Board
| +- Load[0] : 1 : 0 : 49 : GPU Core
| +- Load[1] : 7 : 2 : 32 : GPU Memory Controller
| +- Load[2] : 0 : 0 : 0 : GPU Video Engine
| +- Load[3] : 11.77394 : 10.85408 : 12.31944 : GPU Memory
| +- Fan[0] : 1770 : 1740 : 1800 : GPU
| +- Control[0] : 44 : 44 : 44 : GPU Fan
|
+-+ WDC WD1002FAEX-00Y9A0 (/hdd/0)
| +- Temperature[0] : 42 : 42 : 44 : HDD
|
+-+ WDC WD6400AAKS-08A7B2 (/hdd/1)
| +- Temperature[0] : 42 : 42 : 42 : HDD
--------------------------------------------------------------------------------
Parameters
|
+-+ ASUS Crosshair IV Formula (/mainboard)
|
+-+ AMD Phenom II X6 1090T Processor (/amdcpu/0)
| +- Temperature[0] : Core #1 - #6
| +- Offset [°C] : 0 : 0
|
+-+ NVIDIA GeForce GTX 480 (/nvidiagpu/0)
|
+-+ WDC WD1002FAEX-00Y9A0 (/hdd/0)
|
+-+ WDC WD6400AAKS-08A7B2 (/hdd/1)
--------------------------------------------------------------------------------
Mainboard
BIOS Vendor: American Megatrends Inc.
BIOS Version: 0801
Mainboard Manufacturer: ASUSTeK Computer INC.
Mainboard Name: Crosshair IV Formula
SMBIOS Table
ABgAAAECAPADH5Dai38BAAAAMwUID///QW1lcmljYW4gTWVnYXRyZW5kcyBJbmMu
ADA4MDEgICAAMDUvMDMvMjAxMAAAARsBAAECAwRAmAAejAAA2/nIIM8wHwEZBgUG
```
Original issue reported on code.google.com by `MrPB...@gmail.com` on 10 Sep 2010 at 10:50
Attachments:
* [OpenHardwareMonitor.Report.txt](https://storage.googleapis.com/google-code-attachments/open-hardware-monitor/issue-113/comment-0/OpenHardwareMonitor.Report.txt)
| priority | asus crosshair formula iv motherboard support what is the expected output what do you see instead what version of the product are you using on what operating system i think please provide any additional information below asus brosshair iv formula amd black please attach a report created with file save report open hardware monitor report version common language runtime operating system microsoft windows nt process type bit sensors asus crosshair iv formula mainboard amd phenom ii processor amdcpu temperature core load cpu total load core load core load core load core load core load core nvidia geforce gtx nvidiagpu clock gpu core clock gpu memory clock gpu shader temperature gpu core temperature gpu board load gpu core load gpu memory controller load gpu video engine load gpu memory fan gpu control gpu fan wdc hdd temperature hdd wdc hdd temperature hdd parameters asus crosshair iv formula mainboard amd phenom ii processor amdcpu temperature core offset nvidia geforce gtx nvidiagpu wdc hdd wdc hdd mainboard bios vendor american megatrends inc bios version mainboard manufacturer asustek computer inc mainboard name crosshair iv formula smbios table original issue reported on code google com by mrpb gmail com on sep at attachments | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.