Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
75,684 | 3,470,980,360 | IssuesEvent | 2015-12-23 12:25:09 | USGS-WiM/SiGLDMS | https://api.github.com/repos/USGS-WiM/SiGLDMS | closed | editable-error styling (CSS) | Priority: Medium | Online data location should check for “http://” when in editing mode (looks like it only checks when it’s originally being entered). -Jen
I fixed this, but not sure how to handle styling. Don't want to add any other libraries on top of what you are doing.
AngularJS Xeditable has a css that handles how the class='editable-error' looks. (http://vitalets.github.io/angular-xeditable/#onbeforesave) I'll leave this to you so I don't break anything. | 1.0 | editable-error styling (CSS) - Online data location should check for “http://” when in editing mode (looks like it only checks when it’s originally being entered). -Jen
I fixed this, but not sure how to handle styling. Don't want to add any other libraries on top of what you are doing.
AngularJS Xeditable has a css that handles how the class='editable-error' looks. (http://vitalets.github.io/angular-xeditable/#onbeforesave) I'll leave this to you so I don't break anything. | priority | editable error styling css online data location should check for “ when in editing mode looks like it only checks when it’s originally being entered jen i fixed this but not sure how to handle styling don t want to add any other libraries on top of what you are doing angularjs xeditable has a css that handles how the class editable error looks i ll leave this to you so i don t break anything | 1 |
253,445 | 8,056,208,888 | IssuesEvent | 2018-08-02 11:57:25 | canonical-websites/www.ubuntu.com | https://api.github.com/repos/canonical-websites/www.ubuntu.com | closed | Robotics Partners logos are not displayed on accessing Ubuntu robotics page | Priority: Medium Type: Bug | *Note - filling in this form will file an issue against the www.ubuntu.com website, not against Ubuntu itself, for Ubuntu bugs, please [go here](https://help.ubuntu.com/community/ReportingBugs)*
## Summary
[please describe the issue]
## Process
1. From the page https://www.ubuntu.com/download/alternative-downloads#alternate-ubuntu-server-installer
2. click on Robotics link, at the bottom of the page
## Current Result
The robotics page is displayed
but seems like the Partner logos are not displayed
Instead , in that area a text displaying that user should creating a bug
## Expected result
Logos of the related Robotics Partners need to be displayed;
If these Logos or icons have a link to the Partner site, it will be better
## Screenshot
[if relevant, include a screenshot]
| 1.0 | Robotics Partners logos are not displayed on accessing Ubuntu robotics page - *Note - filling in this form will file an issue against the www.ubuntu.com website, not against Ubuntu itself, for Ubuntu bugs, please [go here](https://help.ubuntu.com/community/ReportingBugs)*
## Summary
[please describe the issue]
## Process
1. From the page https://www.ubuntu.com/download/alternative-downloads#alternate-ubuntu-server-installer
2. click on Robotics link, at the bottom of the page
## Current Result
The robotics page is displayed
but seems like the Partner logos are not displayed
Instead , in that area a text displaying that user should creating a bug
## Expected result
Logos of the related Robotics Partners need to be displayed;
If these Logos or icons have a link to the Partner site, it will be better
## Screenshot
[if relevant, include a screenshot]
| priority | robotics partners logos are not displayed on accessing ubuntu robotics page note filling in this form will file an issue against the website not against ubuntu itself for ubuntu bugs please summary process from the page click on robotics link at the bottom of the page current result the robotics page is displayed but seems like the partner logos are not displayed instead in that area a text displaying that user should creating a bug expected result logos of the related robotics partners need to be displayed if these logos or icons have a link to the partner site it will be better screenshot | 1 |
634,381 | 20,360,108,072 | IssuesEvent | 2022-02-20 15:07:07 | dnd-side-project/dnd-6th-1-iOS | https://api.github.com/repos/dnd-side-project/dnd-6th-1-iOS | opened | 게시글 상세 화면 서버 연동 | 윤경 커뮤니티 View Priority: Medium Feature Network | ## 기능 설명
게시글 상세 화면 서버 연동
+ 커뮤니티 홈 Post Category 추가
+ 댓글, 게시글 데이터 없을때 Cell 추가
## 완료 조건
- [ ] 게시글 선택 시 데이터 GET
- [ ] 게시글 댓글 데이터 GET
- [ ] PostButtonView 댓글 버튼 연결
- [ ] 커뮤니티 홈 Post Category 추가
- [ ] 게시글 데이터 없을때 Cell 구현
- [ ] 댓글 데이터 없을때 Cell 구현
| 1.0 | 게시글 상세 화면 서버 연동 - ## 기능 설명
게시글 상세 화면 서버 연동
+ 커뮤니티 홈 Post Category 추가
+ 댓글, 게시글 데이터 없을때 Cell 추가
## 완료 조건
- [ ] 게시글 선택 시 데이터 GET
- [ ] 게시글 댓글 데이터 GET
- [ ] PostButtonView 댓글 버튼 연결
- [ ] 커뮤니티 홈 Post Category 추가
- [ ] 게시글 데이터 없을때 Cell 구현
- [ ] 댓글 데이터 없을때 Cell 구현
| priority | 게시글 상세 화면 서버 연동 기능 설명 게시글 상세 화면 서버 연동 커뮤니티 홈 post category 추가 댓글 게시글 데이터 없을때 cell 추가 완료 조건 게시글 선택 시 데이터 get 게시글 댓글 데이터 get postbuttonview 댓글 버튼 연결 커뮤니티 홈 post category 추가 게시글 데이터 없을때 cell 구현 댓글 데이터 없을때 cell 구현 | 1 |
96,404 | 3,967,991,635 | IssuesEvent | 2016-05-03 18:07:44 | soulweaver91/project-carrot | https://api.github.com/repos/soulweaver91/project-carrot | opened | Bridge event somehow despawned permanently | Priority: Medium Scope: Previously implemented feature Status: Investigating Type: Bug | It didn't come back even after going far enough to deactivate and reactivate it. I assume something somewhere cleared its tile in the tile map, but what and why is unknown. | 1.0 | Bridge event somehow despawned permanently - It didn't come back even after going far enough to deactivate and reactivate it. I assume something somewhere cleared its tile in the tile map, but what and why is unknown. | priority | bridge event somehow despawned permanently it didn t come back even after going far enough to deactivate and reactivate it i assume something somewhere cleared its tile in the tile map but what and why is unknown | 1 |
105,883 | 4,243,162,916 | IssuesEvent | 2016-07-06 21:55:51 | OperationCode/operationcode | https://api.github.com/repos/OperationCode/operationcode | opened | Add 'Code Schools Nearby' to veteran profile | Priority: Medium Status: Available Type: Enhancement | We can probably easily compare the users zip code with our list of code schools to see how many schools are in the area. | 1.0 | Add 'Code Schools Nearby' to veteran profile - We can probably easily compare the users zip code with our list of code schools to see how many schools are in the area. | priority | add code schools nearby to veteran profile we can probably easily compare the users zip code with our list of code schools to see how many schools are in the area | 1 |
406,101 | 11,886,868,906 | IssuesEvent | 2020-03-27 23:16:11 | gohornet/hornet | https://api.github.com/repos/gohornet/hornet | closed | Use the workerpool from the hive.go async package which uses ant | feature hive.go priority-medium | Ant is probably a better performing implementation of a workerpool than our own. This also decreases the technical debt of maintaining our own workerpool implementation to a certain degree. | 1.0 | Use the workerpool from the hive.go async package which uses ant - Ant is probably a better performing implementation of a workerpool than our own. This also decreases the technical debt of maintaining our own workerpool implementation to a certain degree. | priority | use the workerpool from the hive go async package which uses ant ant is probably a better performing implementation of a workerpool than our own this also decreases the technical debt of maintaining our own workerpool implementation to a certain degree | 1 |
54,835 | 3,071,427,633 | IssuesEvent | 2015-08-19 12:03:29 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Предложение: Добавить поиск kinopoisk.ru | Component-UI enhancement imported Priority-Medium Usability | _From [cane...@gmail.com](https://code.google.com/u/118030959308332740020/) on September 12, 2010 05:25:43_
Добавте пожалуйста поиск kinopoisk.ru или замените им Яндекс. Этим вы облегчите мне жизнь. Спасибо
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=163_ | 1.0 | Предложение: Добавить поиск kinopoisk.ru - _From [cane...@gmail.com](https://code.google.com/u/118030959308332740020/) on September 12, 2010 05:25:43_
Добавте пожалуйста поиск kinopoisk.ru или замените им Яндекс. Этим вы облегчите мне жизнь. Спасибо
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=163_ | priority | предложение добавить поиск kinopoisk ru from on september добавте пожалуйста поиск kinopoisk ru или замените им яндекс этим вы облегчите мне жизнь спасибо original issue | 1 |
617,368 | 19,348,748,231 | IssuesEvent | 2021-12-15 13:40:34 | Esri/solutions-widget-utility-network-trace | https://api.github.com/repos/Esri/solutions-widget-utility-network-trace | closed | Only display terminal selection option if there are 2+ terminals available | enhancement JSAPI widget Priority - Medium | Only display terminal selection option if there are 2+ terminals available, and filter out the system terminals from the list.

Also, if there are more than 1 terminal, we need to add an error to select at least 1 terminal.
Also, if there's only 1 terminal I believe we nay need to keep the expandable section with a string that says which terminal is set as an FYI to the user ? | 1.0 | Only display terminal selection option if there are 2+ terminals available - Only display terminal selection option if there are 2+ terminals available, and filter out the system terminals from the list.

Also, if there are more than 1 terminal, we need to add an error to select at least 1 terminal.
Also, if there's only 1 terminal I believe we nay need to keep the expandable section with a string that says which terminal is set as an FYI to the user ? | priority | only display terminal selection option if there are terminals available only display terminal selection option if there are terminals available and filter out the system terminals from the list also if there are more than terminal we need to add an error to select at least terminal also if there s only terminal i believe we nay need to keep the expandable section with a string that says which terminal is set as an fyi to the user | 1 |
554,498 | 16,430,872,908 | IssuesEvent | 2021-05-20 01:16:01 | KShewengger/ngx-simple-crud-recipe | https://api.github.com/repos/KShewengger/ngx-simple-crud-recipe | opened | Setup Special Ingredients Functionality | Priority: Medium Type: Feature Type: Functionality | Ingredients with a matching `ingredientId` listed in the specials response should also show the special title, type and text under the ingredient name | 1.0 | Setup Special Ingredients Functionality - Ingredients with a matching `ingredientId` listed in the specials response should also show the special title, type and text under the ingredient name | priority | setup special ingredients functionality ingredients with a matching ingredientid listed in the specials response should also show the special title type and text under the ingredient name | 1 |
721,027 | 24,815,234,446 | IssuesEvent | 2022-10-25 12:41:13 | Zenika/skillZ | https://api.github.com/repos/Zenika/skillZ | closed | Profile page empty after adding a certification | bug front Priority: Medium | Steps to reproduce :
1. Go to profile page
2. Add a certification
3. Confirm adding certification
4. The profile page are reloading and is empty

| 1.0 | Profile page empty after adding a certification - Steps to reproduce :
1. Go to profile page
2. Add a certification
3. Confirm adding certification
4. The profile page are reloading and is empty

| priority | profile page empty after adding a certification steps to reproduce go to profile page add a certification confirm adding certification the profile page are reloading and is empty | 1 |
660,463 | 21,968,729,619 | IssuesEvent | 2022-05-25 00:14:03 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [Memory leak in unit test] BackupTxnTest.Consistency | kind/bug area/docdb priority/medium status/awaiting-triage | Jira Link: [[DB-273]](https://yugabyte.atlassian.net/browse/DB-273)
### Description
Memory leak observed in unit test BackupTxnTest.Consistency
Stacktrace: https://gist.github.com/kripasreenivasan/1bfbb413e3ddb20052765ec04854fdce
[DB-273]: https://yugabyte.atlassian.net/browse/DB-273?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [Memory leak in unit test] BackupTxnTest.Consistency - Jira Link: [[DB-273]](https://yugabyte.atlassian.net/browse/DB-273)
### Description
Memory leak observed in unit test BackupTxnTest.Consistency
Stacktrace: https://gist.github.com/kripasreenivasan/1bfbb413e3ddb20052765ec04854fdce
[DB-273]: https://yugabyte.atlassian.net/browse/DB-273?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | backuptxntest consistency jira link description memory leak observed in unit test backuptxntest consistency stacktrace | 1 |
417,260 | 12,157,814,236 | IssuesEvent | 2020-04-26 00:09:44 | BGWE/BoardgameWE | https://api.github.com/repos/BGWE/BoardgameWE | closed | Suggestions for players | component/backend component/ui priority/medium sev/improvement | When creating a game, a suggestion of users could be provided as likely players (users with whom current user has played recently and/or plays often) | 1.0 | Suggestions for players - When creating a game, a suggestion of users could be provided as likely players (users with whom current user has played recently and/or plays often) | priority | suggestions for players when creating a game a suggestion of users could be provided as likely players users with whom current user has played recently and or plays often | 1 |
237,219 | 7,757,546,415 | IssuesEvent | 2018-05-31 16:38:26 | VoxelIndustry/RiseOfSteam | https://api.github.com/repos/VoxelIndustry/RiseOfSteam | closed | Loss of fluids in pipes | MEDIUM PRIORITY bug | On world reload connected fluidpipes loss a fraction of their cold stored fluid.
This is not a major issue, but need investigation in the future. | 1.0 | Loss of fluids in pipes - On world reload connected fluidpipes loss a fraction of their cold stored fluid.
This is not a major issue, but need investigation in the future. | priority | loss of fluids in pipes on world reload connected fluidpipes loss a fraction of their cold stored fluid this is not a major issue but need investigation in the future | 1 |
21,358 | 2,639,550,088 | IssuesEvent | 2015-03-11 03:00:55 | firejake308/code-versus-bugs | https://api.github.com/repos/firejake308/code-versus-bugs | closed | Minigame Completion | enhancement Medium Priority | The minigame needs some more slides. Anyone up to do it? Also, the first segment should have an output, but it doesn't show up. | 1.0 | Minigame Completion - The minigame needs some more slides. Anyone up to do it? Also, the first segment should have an output, but it doesn't show up. | priority | minigame completion the minigame needs some more slides anyone up to do it also the first segment should have an output but it doesn t show up | 1 |
40,824 | 2,868,944,397 | IssuesEvent | 2015-06-05 22:06:48 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | closed | Document dev dependencies | enhancement Fixed Priority-Medium | <a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)**
_Originally opened as dart-lang/sdk#9120_
----
Need some user docs on pub.dartlang.org. | 1.0 | Document dev dependencies - <a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)**
_Originally opened as dart-lang/sdk#9120_
----
Need some user docs on pub.dartlang.org. | priority | document dev dependencies issue by originally opened as dart lang sdk need some user docs on pub dartlang org | 1 |
114,329 | 4,628,762,255 | IssuesEvent | 2016-09-28 06:34:34 | bwapi/bwapi | https://api.github.com/repos/bwapi/bwapi | closed | Request: Function to get the unit which a given unit is currently training | New Feature Priority-Medium | Example: A function which returns the SCV that a command center is currently training.
Notes: It seems that the frame after you give a build command to a command center getRemainingTrainTime() > 0 but there is no SCV unit yet on the map. For this frame, its training queue is of size 1 and the value of it is the SCV. | 1.0 | Request: Function to get the unit which a given unit is currently training - Example: A function which returns the SCV that a command center is currently training.
Notes: It seems that the frame after you give a build command to a command center getRemainingTrainTime() > 0 but there is no SCV unit yet on the map. For this frame, its training queue is of size 1 and the value of it is the SCV. | priority | request function to get the unit which a given unit is currently training example a function which returns the scv that a command center is currently training notes it seems that the frame after you give a build command to a command center getremainingtraintime but there is no scv unit yet on the map for this frame its training queue is of size and the value of it is the scv | 1 |
487,590 | 14,048,856,613 | IssuesEvent | 2020-11-02 09:28:09 | PazerOP/tf2_bot_detector | https://api.github.com/repos/PazerOP/tf2_bot_detector | closed | [Enhancement/Tweak] Increase delay between failed votes OR assume re-vote chance | Priority: Medium Type: Enhancement | There is a cooldown between failed and successful votes.
Since we don't know if a vote was successful, I'd suggest to either increase the delay between votes a good bit (30-45 sec?, 60 sec?) or 'assume' if the vote is failing or not.

If a vote keeps failing (like 3 times in a row), delay the following votes until a sane result can be exprected (after a minute, on vote target switch, etc). | 1.0 | [Enhancement/Tweak] Increase delay between failed votes OR assume re-vote chance - There is a cooldown between failed and successful votes.
Since we don't know if a vote was successful, I'd suggest to either increase the delay between votes a good bit (30-45 sec?, 60 sec?) or 'assume' if the vote is failing or not.

If a vote keeps failing (like 3 times in a row), delay the following votes until a sane result can be exprected (after a minute, on vote target switch, etc). | priority | increase delay between failed votes or assume re vote chance there is a cooldown between failed and successful votes since we don t know if a vote was successful i d suggest to either increase the delay between votes a good bit sec sec or assume if the vote is failing or not if a vote keeps failing like times in a row delay the following votes until a sane result can be exprected after a minute on vote target switch etc | 1 |
527,233 | 15,326,568,273 | IssuesEvent | 2021-02-26 03:56:33 | stride3d/stride | https://api.github.com/repos/stride3d/stride | closed | Archetype cannot be removed on the derived asset | area-Asset bug priority-Medium work-estimate-S | **Release Type**: Official Release
**Version**: 1265 (and earlier)
**Platform(s)**: Editor
**Describe the bug**
The derived asset has an Archetype property, which allows to select another asset. There's also a 'Clear reference' button which when clicked does nothing.
**To Reproduce**
Steps to reproduce the behavior:
1. Create new project
2. Create a derived asset from the Sphere Material
3. Open the new asset and try clicking 'Clear reference' button on its Archetype property.
**Expected behavior**
It should be either possible to remove an archetype reference causing the asset to become standalone, or the clear reference button should not be there.
**Workaround**
It's possible to remove the archetype in the asset's YAML.
**Additional context**
Removing an archetype could be beneficial until #280 is fixed. | 1.0 | Archetype cannot be removed on the derived asset - **Release Type**: Official Release
**Version**: 1265 (and earlier)
**Platform(s)**: Editor
**Describe the bug**
The derived asset has an Archetype property, which allows to select another asset. There's also a 'Clear reference' button which when clicked does nothing.
**To Reproduce**
Steps to reproduce the behavior:
1. Create new project
2. Create a derived asset from the Sphere Material
3. Open the new asset and try clicking 'Clear reference' button on its Archetype property.
**Expected behavior**
It should be either possible to remove an archetype reference causing the asset to become standalone, or the clear reference button should not be there.
**Workaround**
It's possible to remove the archetype in the asset's YAML.
**Additional context**
Removing an archetype could be beneficial until #280 is fixed. | priority | archetype cannot be removed on the derived asset release type official release version and earlier platform s editor describe the bug the derived asset has an archetype property which allows to select another asset there s also a clear reference button which when clicked does nothing to reproduce steps to reproduce the behavior create new project create a derived asset from the sphere material open the new asset and try clicking clear reference button on its archetype property expected behavior it should be either possible to remove an archetype reference causing the asset to become standalone or the clear reference button should not be there workaround it s possible to remove the archetype in the asset s yaml additional context removing an archetype could be beneficial until is fixed | 1 |
331,277 | 10,063,582,755 | IssuesEvent | 2019-07-23 06:23:45 | all-contributors/all-contributors | https://api.github.com/repos/all-contributors/all-contributors | closed | HTML generated respects MD linter rule: MD013 Line length | enhancement priority: medium status: waiting for feedback | **Is your feature request related to a problem? Please describe.**
The HTML inserted into Readme.md breaks Md linter.
**Describe the solution you'd like**
Inserted HTML code have width < 80 chars.
**Describe alternatives you've considered**
Inserted code warns not to altered generated code.
**Additional context**
This is the only thing breaking the linting check in my Circle-CI orb
```
#!/bin/bash -eo pipefail
mdl ./
.//README.md:191: MD013 Line length
A detailed description of the rules is available at https://github.com/markdownlint/markdownlint/blob/master/docs/RULES.md
Exited with code 1
```
| 1.0 | HTML generated respects MD linter rule: MD013 Line length - **Is your feature request related to a problem? Please describe.**
The HTML inserted into Readme.md breaks Md linter.
**Describe the solution you'd like**
Inserted HTML code have width < 80 chars.
**Describe alternatives you've considered**
Inserted code warns not to altered generated code.
**Additional context**
This is the only thing breaking the linting check in my Circle-CI orb
```
#!/bin/bash -eo pipefail
mdl ./
.//README.md:191: MD013 Line length
A detailed description of the rules is available at https://github.com/markdownlint/markdownlint/blob/master/docs/RULES.md
Exited with code 1
```
| priority | html generated respects md linter rule line length is your feature request related to a problem please describe the html inserted into readme md breaks md linter describe the solution you d like inserted html code have width chars describe alternatives you ve considered inserted code warns not to altered generated code additional context this is the only thing breaking the linting check in my circle ci orb bin bash eo pipefail mdl readme md line length a detailed description of the rules is available at exited with code | 1 |
703,070 | 24,145,909,922 | IssuesEvent | 2022-09-21 18:43:27 | kytos-ng/flow_stats | https://api.github.com/repos/kytos-ng/flow_stats | closed | Drop OF1.0 support | priority_medium 2022.3 epic_drop_openflow_0x01 | Since `epic_drop_openflow_0x01` is progressing faster than `epic_flow_stats_v1_stable`, we'll need to drop OF1.0 before dropping `of_core` OF1.0 support https://github.com/kytos-ng/of_core/pull/82, otherwise this NApp won't load. Initially the idea was to drop on issue #35, but we'll drop OF1.0 support first since this is progressing faster to avoid getting blocked
@Alopalao, could you help out with this? Thanks
```
❯ rg 0x01 -g "*.py"
main.py
13:import pyof.v0x01.controller2switch.common as common01
19:from napps.kytos.of_core.v0x01.flow import Action as Action10
22:from pyof.v0x01.common.flow_match import FlowWildCards
30: def __init__(self, version='0x01', match=None, idle_timeout=0,
62: if self.version == '0x01':
79: if self.version == '0x01':
128: def from_flow_stats(cls, flow_stats, version='0x01'):
138: if version == '0x01':
202: if self.version == '0x01':
418: if switch.ofp_version == '0x01':
541: @listen_to('kytos/of_core.v0x01.messages.in.ofpt_stats_reply')
542: def on_stats_reply_0x01(self, event):
543: """Capture flow stats messages for v0x01 switches."""
544: self.handle_stats_reply_0x01(event)
546: def handle_stats_reply_0x01(self, event):
547: """Handle stats replies for v0x01 switches."""
``` | 1.0 | Drop OF1.0 support - Since `epic_drop_openflow_0x01` is progressing faster than `epic_flow_stats_v1_stable`, we'll need to drop OF1.0 before dropping `of_core` OF1.0 support https://github.com/kytos-ng/of_core/pull/82, otherwise this NApp won't load. Initially the idea was to drop on issue #35, but we'll drop OF1.0 support first since this is progressing faster to avoid getting blocked
@Alopalao, could you help out with this? Thanks
```
❯ rg 0x01 -g "*.py"
main.py
13:import pyof.v0x01.controller2switch.common as common01
19:from napps.kytos.of_core.v0x01.flow import Action as Action10
22:from pyof.v0x01.common.flow_match import FlowWildCards
30: def __init__(self, version='0x01', match=None, idle_timeout=0,
62: if self.version == '0x01':
79: if self.version == '0x01':
128: def from_flow_stats(cls, flow_stats, version='0x01'):
138: if version == '0x01':
202: if self.version == '0x01':
418: if switch.ofp_version == '0x01':
541: @listen_to('kytos/of_core.v0x01.messages.in.ofpt_stats_reply')
542: def on_stats_reply_0x01(self, event):
543: """Capture flow stats messages for v0x01 switches."""
544: self.handle_stats_reply_0x01(event)
546: def handle_stats_reply_0x01(self, event):
547: """Handle stats replies for v0x01 switches."""
``` | priority | drop support since epic drop openflow is progressing faster than epic flow stats stable we ll need to drop before dropping of core support otherwise this napp won t load initially the idea was to drop on issue but we ll drop support first since this is progressing faster to avoid getting blocked alopalao could you help out with this thanks ❯ rg g py main py import pyof common as from napps kytos of core flow import action as from pyof common flow match import flowwildcards def init self version match none idle timeout if self version if self version def from flow stats cls flow stats version if version if self version if switch ofp version listen to kytos of core messages in ofpt stats reply def on stats reply self event capture flow stats messages for switches self handle stats reply event def handle stats reply self event handle stats replies for switches | 1 |
87,789 | 3,757,950,082 | IssuesEvent | 2016-03-14 05:21:06 | nus-fboa2016-PL/PowerPointLabs | https://api.github.com/repos/nus-fboa2016-PL/PowerPointLabs | opened | Distribute with respect to extreme shapes | Feature.PositionsLab Priority.Medium Type.FeatureRequest | Should have 4 settings instead of 3
1. Need to rename first reference object (container or something like that)
2. Distribute between first and second objects
3. Distribute between extreme shapes (left and right most, for example)
4. Distribute within slide | 1.0 | Distribute with respect to extreme shapes - Should have 4 settings instead of 3
1. Need to rename first reference object (container or something like that)
2. Distribute between first and second objects
3. Distribute between extreme shapes (left and right most, for example)
4. Distribute within slide | priority | distribute with respect to extreme shapes should have settings instead of need to rename first reference object container or something like that distribute between first and second objects distribute between extreme shapes left and right most for example distribute within slide | 1 |
826,257 | 31,562,531,452 | IssuesEvent | 2023-09-03 12:32:53 | ASCENDynamics-NFP/AscendCoopPlatform | https://api.github.com/repos/ASCENDynamics-NFP/AscendCoopPlatform | closed | Need Unit Tests for AuthService File | good first issue priority - medium type - refactoring status - on hold effort - medium role - backend developer | ## Description
We need to create unit tests for our service file using Karma and Jasmine to ensure all functions behave as expected.
## Service File Details
- File Name: `auth.service.ts`
- Location: `src/app/services/auth.service.ts`
## Test Requirements
- Write tests to cover all functions in the service file.
- Each test should cover both the expected outcome and potential edge cases.
## Steps to Start
1. Navigate to the service file located at `src/app/services/auth.service.ts`
2. Examine the functions that are currently without tests
3. Begin writing tests in a separate `auth.service.spec.ts` file
## Expected Outcome
All functions within the service file will have associated unit tests that cover both expected behavior and potential edge cases. All tests should pass when running the `ng test` command. | 1.0 | Need Unit Tests for AuthService File - ## Description
We need to create unit tests for our service file using Karma and Jasmine to ensure all functions behave as expected.
## Service File Details
- File Name: `auth.service.ts`
- Location: `src/app/services/auth.service.ts`
## Test Requirements
- Write tests to cover all functions in the service file.
- Each test should cover both the expected outcome and potential edge cases.
## Steps to Start
1. Navigate to the service file located at `src/app/services/auth.service.ts`
2. Examine the functions that are currently without tests
3. Begin writing tests in a separate `auth.service.spec.ts` file
## Expected Outcome
All functions within the service file will have associated unit tests that cover both expected behavior and potential edge cases. All tests should pass when running the `ng test` command. | priority | need unit tests for authservice file description we need to create unit tests for our service file using karma and jasmine to ensure all functions behave as expected service file details file name auth service ts location src app services auth service ts test requirements write tests to cover all functions in the service file each test should cover both the expected outcome and potential edge cases steps to start navigate to the service file located at src app services auth service ts examine the functions that are currently without tests begin writing tests in a separate auth service spec ts file expected outcome all functions within the service file will have associated unit tests that cover both expected behavior and potential edge cases all tests should pass when running the ng test command | 1 |
196,287 | 6,926,439,595 | IssuesEvent | 2017-11-30 19:10:21 | information-artifact-ontology/IAO | https://api.github.com/repos/information-artifact-ontology/IAO | closed | broad synonym | imported Priority-Medium Type-Term | _From [dosu...@gmail.com](https://code.google.com/u/102674886352087815907/) on March 10, 2011 09:56:42_
_label for the new term_
Broad synonym
_textual definition_
def: "Information artifact that links an ontology term to a term in language that, within some specified publication or discipline, has a broader set of referents then the ontology term." Please provide any additional information below. (e.g., proposed position in the IAO hierarchy) is_a: synonym
_Original issue: http://code.google.com/p/information-artifact-ontology/issues/detail?id=104_
| 1.0 | broad synonym - _From [dosu...@gmail.com](https://code.google.com/u/102674886352087815907/) on March 10, 2011 09:56:42_
_label for the new term_
Broad synonym
_textual definition_
def: "Information artifact that links an ontology term to a term in language that, within some specified publication or discipline, has a broader set of referents then the ontology term." Please provide any additional information below. (e.g., proposed position in the IAO hierarchy) is_a: synonym
_Original issue: http://code.google.com/p/information-artifact-ontology/issues/detail?id=104_
| priority | broad synonym from on march label for the new term broad synonym textual definition def information artifact that links an ontology term to a term in language that within some specified publication or discipline has a broader set of referents then the ontology term please provide any additional information below e g proposed position in the iao hierarchy is a synonym original issue | 1 |
765,327 | 26,841,537,535 | IssuesEvent | 2023-02-03 01:10:29 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Fix migrations to properly update pg_proc | kind/bug area/ysql priority/medium | Jira Link: [DB-5369](https://yugabyte.atlassian.net/browse/DB-5369)
### Description
Commit 00e886216e63118f24e15b4d16527354f23f5359 changes UPDATE to DELETE and INSERT as a workaround to issue #13500. It doesn't properly translate the UPDATE to DELETE and INSERT because the conditions of the UPDATE don't match the condition of the DELETE. Actually, in the first place, it didn't have enough conditions.
`\d pg_proc` shows
```
Indexes:
"pg_proc_oid_index" PRIMARY KEY, lsm (oid ASC)
"pg_proc_proname_args_nsp_index" UNIQUE, lsm (proname ASC, proargtypes ASC, pronamespace ASC)
```
If using the unique index, it should condition on all three columns. Before, it had only two columns, and after, it had only one column. One possibility right now is to have user's public.pg_stat_get_progress_info get deleted.
[DB-5369]: https://yugabyte.atlassian.net/browse/DB-5369?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YSQL] Fix migrations to properly update pg_proc - Jira Link: [DB-5369](https://yugabyte.atlassian.net/browse/DB-5369)
### Description
Commit 00e886216e63118f24e15b4d16527354f23f5359 changes UPDATE to DELETE and INSERT as a workaround to issue #13500. It doesn't properly translate the UPDATE to DELETE and INSERT because the conditions of the UPDATE don't match the condition of the DELETE. Actually, in the first place, it didn't have enough conditions.
`\d pg_proc` shows
```
Indexes:
"pg_proc_oid_index" PRIMARY KEY, lsm (oid ASC)
"pg_proc_proname_args_nsp_index" UNIQUE, lsm (proname ASC, proargtypes ASC, pronamespace ASC)
```
If using the unique index, it should condition on all three columns. Before, it had only two columns, and after, it had only one column. One possibility right now is to have user's public.pg_stat_get_progress_info get deleted.
[DB-5369]: https://yugabyte.atlassian.net/browse/DB-5369?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | fix migrations to properly update pg proc jira link description commit changes update to delete and insert as a workaround to issue it doesn t properly translate the update to delete and insert because the conditions of the update don t match the condition of the delete actually in the first place it didn t have enough conditions d pg proc shows indexes pg proc oid index primary key lsm oid asc pg proc proname args nsp index unique lsm proname asc proargtypes asc pronamespace asc if using the unique index it should condition on all three columns before it had only two columns and after it had only one column one possibility right now is to have user s public pg stat get progress info get deleted | 1 |
198,888 | 6,978,687,304 | IssuesEvent | 2017-12-12 18:22:11 | statgen/locuszoom-api | https://api.github.com/repos/statgen/locuszoom-api | closed | LD data unavailable for some regions | priority: medium type: bug | # Summary
Some of the regions for the API do not return LD data. This means that (some) sample LocusZoom js plots are rendered in solid grey instead of in color. This includes the default plot that is the first demo on the landing page.
# Description/ steps to reproduce
See this page: default plot, and "top hits" links on the left-hand sidebar of the page.
http://statgen.github.io/locuszoom/
This issue causes the sample plots to appear grey instead of in color. (eg the base example and some of the top hits like TCFL72)
<img width="899" alt="screen shot 2017-12-04 at 11 46 35 am" src="https://user-images.githubusercontent.com/2957073/33564580-d86c1406-d8e8-11e7-8a65-4863d8831940.png">
Other plot regions have LD data and show up in color just fine, eg FTO.
<img width="910" alt="screen shot 2017-12-04 at 11 46 28 am" src="https://user-images.githubusercontent.com/2957073/33564584-dcebf5fa-d8e8-11e7-970a-33d92f23dc78.png">
# Scope/ impacts
For LZ demo purposes, a work around is available: we can deal with this by changing the default region on the example plot in our next release. Therefore a backend fix would be nice, but is not urgent.
# Notes
Ryan suggests that the issue is with the cache mechanism for the LD endpoint. | 1.0 | LD data unavailable for some regions - # Summary
Some of the regions for the API do not return LD data. This means that (some) sample LocusZoom js plots are rendered in solid grey instead of in color. This includes the default plot that is the first demo on the landing page.
# Description/ steps to reproduce
See this page: default plot, and "top hits" links on the left-hand sidebar of the page.
http://statgen.github.io/locuszoom/
This issue causes the sample plots to appear grey instead of in color. (eg the base example and some of the top hits like TCFL72)
<img width="899" alt="screen shot 2017-12-04 at 11 46 35 am" src="https://user-images.githubusercontent.com/2957073/33564580-d86c1406-d8e8-11e7-8a65-4863d8831940.png">
Other plot regions have LD data and show up in color just fine, eg FTO.
<img width="910" alt="screen shot 2017-12-04 at 11 46 28 am" src="https://user-images.githubusercontent.com/2957073/33564584-dcebf5fa-d8e8-11e7-970a-33d92f23dc78.png">
# Scope/ impacts
For LZ demo purposes, a work around is available: we can deal with this by changing the default region on the example plot in our next release. Therefore a backend fix would be nice, but is not urgent.
# Notes
Ryan suggests that the issue is with the cache mechanism for the LD endpoint. | priority | ld data unavailable for some regions summary some of the regions for the api do not return ld data this means that some sample locuszoom js plots are rendered in solid grey instead of in color this includes the default plot that is the first demo on the landing page description steps to reproduce see this page default plot and top hits links on the left hand sidebar of the page this issue causes the sample plots to appear grey instead of in color eg the base example and some of the top hits like img width alt screen shot at am src other plot regions have ld data and show up in color just fine eg fto img width alt screen shot at am src scope impacts for lz demo purposes a work around is available we can deal with this by changing the default region on the example plot in our next release therefore a backend fix would be nice but is not urgent notes ryan suggests that the issue is with the cache mechanism for the ld endpoint | 1 |
244,670 | 7,878,462,739 | IssuesEvent | 2018-06-26 10:18:07 | uracreative/task-management | https://api.github.com/repos/uracreative/task-management | opened | Prepare and submit appplication for Unicef Innovation Fund | Internal: Identihub Priority: Medium | Financial and technological support from UNICEF’s Venture Fund is available for companies that can show a strong founding team and a clear path to improving the lives of children. We need to have a working, open source prototype showing promising results. With this grant we aim to implement features that will help children do more with 3D printing. More specifically kids can hosting 3D Print assets and print them easily. In addition, through Identihub kids can convert SVG to a 3rd dimention and make these files ready for print.
Website for application: https://unicefinnovationfund.org/#/submit
Deadline for draft: 15.08.2018
Deadline for application: 20.08.2018 | 1.0 | Prepare and submit appplication for Unicef Innovation Fund - Financial and technological support from UNICEF’s Venture Fund is available for companies that can show a strong founding team and a clear path to improving the lives of children. We need to have a working, open source prototype showing promising results. With this grant we aim to implement features that will help children do more with 3D printing. More specifically kids can hosting 3D Print assets and print them easily. In addition, through Identihub kids can convert SVG to a 3rd dimention and make these files ready for print.
Website for application: https://unicefinnovationfund.org/#/submit
Deadline for draft: 15.08.2018
Deadline for application: 20.08.2018 | priority | prepare and submit appplication for unicef innovation fund financial and technological support from unicef’s venture fund is available for companies that can show a strong founding team and a clear path to improving the lives of children we need to have a working open source prototype showing promising results with this grant we aim to implement features that will help children do more with printing more specifically kids can hosting print assets and print them easily in addition through identihub kids can convert svg to a dimention and make these files ready for print website for application deadline for draft deadline for application | 1 |
717,075 | 24,660,449,960 | IssuesEvent | 2022-10-18 06:04:56 | OpenMined/PySyft | https://api.github.com/repos/OpenMined/PySyft | closed | Missing progress status when we clone the repo on installing hagrid for first time | Type: Improvement :chart_with_upwards_trend: Priority: 3 - Medium :unamused: | ## Description
We are missing the progress/process status when we clone the repo on installing hagrid the first time.
Refer to:
- `get_git_repo` -> Clone repo for the first time
- `update_repo` -> Update repo when launching domain using a different branch
We can either print progress info or use rich.Console to produce a process status animation (Refer to `console.status` in `quickstart_setup` in hagrid/cli.py. [Rich Documentation](https://rich.readthedocs.io/en/stable/console.html#status)) | 1.0 | Missing progress status when we clone the repo on installing hagrid for first time - ## Description
We are missing the progress/process status when we clone the repo on installing hagrid the first time.
Refer to:
- `get_git_repo` -> Clone repo for the first time
- `update_repo` -> Update repo when launching domain using a different branch
We can either print progress info or use rich.Console to produce a process status animation (Refer to `console.status` in `quickstart_setup` in hagrid/cli.py. [Rich Documentation](https://rich.readthedocs.io/en/stable/console.html#status)) | priority | missing progress status when we clone the repo on installing hagrid for first time description we are missing the progress process status when we clone the repo on installing hagrid the first time refer to get git repo clone repo for the first time update repo update repo when launching domain using a different branch we can either print progress info or use rich console to produce a process status animation refer to console status in quickstart setup in hagrid cli py | 1 |
655,418 | 21,689,815,257 | IssuesEvent | 2022-05-09 14:27:59 | zeyneplervesarp/swe574-javagang | https://api.github.com/repos/zeyneplervesarp/swe574-javagang | closed | Add FOLLOW to Activity Streams | enhancement backend medium priority difficulty-medium | Follow events had been skipped in the activity streams, it should be added.
In the end, the activity should look like:
` {
"summary": "<username> started following <username>",
"verb": "follow",
"actor": {
"objectType": "user",
"displayName": "<username>",
"id": "<id of the following user>",
"url": "/user/<id of the following user>"
},` | 1.0 | Add FOLLOW to Activity Streams - Follow events had been skipped in the activity streams, it should be added.
In the end, the activity should look like:
` {
"summary": "<username> started following <username>",
"verb": "follow",
"actor": {
"objectType": "user",
"displayName": "<username>",
"id": "<id of the following user>",
"url": "/user/<id of the following user>"
},` | priority | add follow to activity streams follow events had been skipped in the activity streams it should be added in the end the activity should look like summary started following verb follow actor objecttype user displayname id url user | 1 |
496,203 | 14,342,525,243 | IssuesEvent | 2020-11-28 03:42:00 | CMPUT301F20T41/boromi | https://api.github.com/repos/CMPUT301F20T41/boromi | closed | CI Unit Tests and Android Tests | Length: 5-10 Hours Priority: Medium Status: Abandoned Type: Maintenance | ## Description
Now that we've begun creating suites of unit tests and android tests, it may be appropriate to start adding CI to run these test suites to ensure that new functionalities aren't causing breaking changes
## Deliverables
* A CI workflow which will run unit tests
* A CI workflow which will run android tests
## Demo
1. Add a test case which always fails
2. Push to your branch
3. Make sure the build fails
4. Remove the test case
5. Push to your branch make sure the build passes | 1.0 | CI Unit Tests and Android Tests - ## Description
Now that we've begun creating suites of unit tests and android tests, it may be appropriate to start adding CI to run these test suites to ensure that new functionalities aren't causing breaking changes
## Deliverables
* A CI workflow which will run unit tests
* A CI workflow which will run android tests
## Demo
1. Add a test case which always fails
2. Push to your branch
3. Make sure the build fails
4. Remove the test case
5. Push to your branch make sure the build passes | priority | ci unit tests and android tests description now that we ve begun creating suites of unit tests and android tests it may be appropriate to start adding ci to run these test suites to ensure that new functionalities aren t causing breaking changes deliverables a ci workflow which will run unit tests a ci workflow which will run android tests demo add a test case which always fails push to your branch make sure the build fails remove the test case push to your branch make sure the build passes | 1 |
363,419 | 10,741,091,403 | IssuesEvent | 2019-10-29 19:30:31 | eriq-augustine/test-issue-copy | https://api.github.com/repos/eriq-augustine/test-issue-copy | opened | Reuse Terms in Weight Learning? | Components - Optimization Components - Weight Learning Difficulty - Medium Performance Priority - Low Type - Investigation | <a href="https://github.com/eriq-augustine"><img src="https://avatars0.githubusercontent.com/u/337857?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [eriq-augustine](https://github.com/eriq-augustine)**
_Wednesday Jan 02, 2019 at 19:41 GMT_
_Originally opened as https://github.com/eriq-augustine/psl/issues/189_
----
Right now, we throw out and re-generate terms for each run of weight learning.
It's a bit wasteful, but terms can be complex and simply updating them in-place may be difficult.
We need to see if there is a simple way to generalize this and see if it is worth the time save.
| 1.0 | Reuse Terms in Weight Learning? - <a href="https://github.com/eriq-augustine"><img src="https://avatars0.githubusercontent.com/u/337857?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [eriq-augustine](https://github.com/eriq-augustine)**
_Wednesday Jan 02, 2019 at 19:41 GMT_
_Originally opened as https://github.com/eriq-augustine/psl/issues/189_
----
Right now, we throw out and re-generate terms for each run of weight learning.
It's a bit wasteful, but terms can be complex and simply updating them in-place may be difficult.
We need to see if there is a simple way to generalize this and see if it is worth the time save.
| priority | reuse terms in weight learning issue by wednesday jan at gmt originally opened as right now we throw out and re generate terms for each run of weight learning it s a bit wasteful but terms can be complex and simply updating them in place may be difficult we need to see if there is a simple way to generalize this and see if it is worth the time save | 1 |
506,997 | 14,677,856,400 | IssuesEvent | 2020-12-31 01:01:14 | hamaluik/timecop | https://api.github.com/repos/hamaluik/timecop | closed | Allow changing default time span for the search (Feature request) | Priority: Medium Status: Accepted Type: Enhancement | The search on top seems to be limited to a month back by default (to make sure most users won't run into performance issues, I figure?).
I'm glad that I can override that to a longer span if I need to. But that always reverts back after I've done other operations in the app.
Usually I'd want the default span to be longer. So if there were an option for manually defining this default, it would be very useful! | 1.0 | Allow changing default time span for the search (Feature request) - The search on top seems to be limited to a month back by default (to make sure most users won't run into performance issues, I figure?).
I'm glad that I can override that to a longer span if I need to. But that always reverts back after I've done other operations in the app.
Usually I'd want the default span to be longer. So if there were an option for manually defining this default, it would be very useful! | priority | allow changing default time span for the search feature request the search on top seems to be limited to a month back by default to make sure most users won t run into performance issues i figure i m glad that i can override that to a longer span if i need to but that always reverts back after i ve done other operations in the app usually i d want the default span to be longer so if there were an option for manually defining this default it would be very useful | 1 |
174,912 | 6,544,343,234 | IssuesEvent | 2017-09-03 15:02:52 | therajanmaurya/android-client-2.0 | https://api.github.com/repos/therajanmaurya/android-client-2.0 | closed | Show Error UI, If API request fails in Deposit Account | enhancement GSoC 2017 Medium Priority Small Issue | **Summary:**
Currently user can't see any UI that give him clue that request is failed to fetch from server and he can retry. | 1.0 | Show Error UI, If API request fails in Deposit Account - **Summary:**
Currently user can't see any UI that give him clue that request is failed to fetch from server and he can retry. | priority | show error ui if api request fails in deposit account summary currently user can t see any ui that give him clue that request is failed to fetch from server and he can retry | 1 |
257,659 | 8,140,220,011 | IssuesEvent | 2018-08-20 20:21:12 | marklogic/marklogic-data-hub | https://api.github.com/repos/marklogic/marklogic-data-hub | opened | main.sjs not properly updated during upgrade | bug priority:medium | When I upgraded my project from 2.0.6 to 3.1 (tip of develop) using QS, the upgrade failed to update the invocation of the writer in my main.sjs, even though my main.sjs contains no customizations. The following is the change that should get made, but is not:
```
38c38
< dhf.runWriter(xdmp.function(null, './writer.sjs'), id, envelope, options);
---
> dhf.runWriter(writerPlugin, id, envelope, options);
```
Without this change, the harmonization flow fails. As a bonus, DHF (or at least QS) tells you absolutely nothing about the error that caused it to fail. I had to hack an xdmp.log statement into flow-impl.xqy to tease the error out, and then diff my upgraded project against a clean 3.x project to figure out what was wrong.
IDK if this problem will persist with upgrade changes currently in the works, but since they're not about detecting changes, I would guess the answer is "yes".
ML & DHF on RHEL 7, QS running in Chrome on Windows. | 1.0 | main.sjs not properly updated during upgrade - When I upgraded my project from 2.0.6 to 3.1 (tip of develop) using QS, the upgrade failed to update the invocation of the writer in my main.sjs, even though my main.sjs contains no customizations. The following is the change that should get made, but is not:
```
38c38
< dhf.runWriter(xdmp.function(null, './writer.sjs'), id, envelope, options);
---
> dhf.runWriter(writerPlugin, id, envelope, options);
```
Without this change, the harmonization flow fails. As a bonus, DHF (or at least QS) tells you absolutely nothing about the error that caused it to fail. I had to hack an xdmp.log statement into flow-impl.xqy to tease the error out, and then diff my upgraded project against a clean 3.x project to figure out what was wrong.
IDK if this problem will persist with upgrade changes currently in the works, but since they're not about detecting changes, I would guess the answer is "yes".
ML & DHF on RHEL 7, QS running in Chrome on Windows. | priority | main sjs not properly updated during upgrade when i upgraded my project from to tip of develop using qs the upgrade failed to update the invocation of the writer in my main sjs even though my main sjs contains no customizations the following is the change that should get made but is not dhf runwriter xdmp function null writer sjs id envelope options dhf runwriter writerplugin id envelope options without this change the harmonization flow fails as a bonus dhf or at least qs tells you absolutely nothing about the error that caused it to fail i had to hack an xdmp log statement into flow impl xqy to tease the error out and then diff my upgraded project against a clean x project to figure out what was wrong idk if this problem will persist with upgrade changes currently in the works but since they re not about detecting changes i would guess the answer is yes ml dhf on rhel qs running in chrome on windows | 1 |
240,818 | 7,806,127,871 | IssuesEvent | 2018-06-11 13:13:12 | Nukkit-coders/MobPlugin | https://api.github.com/repos/Nukkit-coders/MobPlugin | closed | Not responding to food | bug confirmed fixed medium priority | ### Issue Description
(Running Nukkit from NukkitX (version info bellow), plugin latest version)
None of the animals will respond to food.
F.ex cow, no respond to wheat (or bukket). Hold the wheat and the cow walks around totally ignoring as it does not see the wheat.
Same happens to the other animals. I have testet chicken, sheep, cow and pig.
### Steps to Reproduce the Issue
<!--- Help us to find the problem by adding steps to reproduce the issue -->
1. I just joins the server and try to breed some animals
2. ...
3. ...
### Nukkit Version and Client Version
From the 'version' command in Nukkit:
This server is running Nukkit 1.0dev 「蘋果(Apple)派(Pie)」 implementing API version 1.0.6 for Minecraft: PE v1.2.13 (protocol version 223)
* Client: iOS
#### Config.yml
```
entities:
auto-spawn-tick: 300
worlds-spawn-disabled: "exampleworld"
max-spawns:
bat: 1
blaze: 0
chicken: 1
cow: 1
creeper: 0
enderman: 1
ghast: 0
husk: 0
horse: 1
mooshroom: 1
ocelot: 1
pig: 1
polarbear: 0
rabbit: 1
stray: 0
sheep: 1
skeleton: 0
spider: 1
wolf: 1
zombie: 0
```
### Crashdump, Backtrace or Other Files
<!--- Please use gist or anything else and add links here -->
```
<!--- Copy and paste your backtrace here... -->
```
| 1.0 | Not responding to food - ### Issue Description
(Running Nukkit from NukkitX (version info bellow), plugin latest version)
None of the animals will respond to food.
F.ex cow, no respond to wheat (or bukket). Hold the wheat and the cow walks around totally ignoring as it does not see the wheat.
Same happens to the other animals. I have testet chicken, sheep, cow and pig.
### Steps to Reproduce the Issue
<!--- Help us to find the problem by adding steps to reproduce the issue -->
1. I just joins the server and try to breed some animals
2. ...
3. ...
### Nukkit Version and Client Version
From the 'version' command in Nukkit:
This server is running Nukkit 1.0dev 「蘋果(Apple)派(Pie)」 implementing API version 1.0.6 for Minecraft: PE v1.2.13 (protocol version 223)
* Client: iOS
#### Config.yml
```
entities:
auto-spawn-tick: 300
worlds-spawn-disabled: "exampleworld"
max-spawns:
bat: 1
blaze: 0
chicken: 1
cow: 1
creeper: 0
enderman: 1
ghast: 0
husk: 0
horse: 1
mooshroom: 1
ocelot: 1
pig: 1
polarbear: 0
rabbit: 1
stray: 0
sheep: 1
skeleton: 0
spider: 1
wolf: 1
zombie: 0
```
### Crashdump, Backtrace or Other Files
<!--- Please use gist or anything else and add links here -->
```
<!--- Copy and paste your backtrace here... -->
```
| priority | not responding to food issue description running nukkit from nukkitx version info bellow plugin latest version none of the animals will respond to food f ex cow no respond to wheat or bukket hold the wheat and the cow walks around totally ignoring as it does not see the wheat same happens to the other animals i have testet chicken sheep cow and pig steps to reproduce the issue i just joins the server and try to breed some animals nukkit version and client version from the version command in nukkit this server is running nukkit 「蘋果 apple 派 pie 」 implementing api version for minecraft pe protocol version client ios config yml entities auto spawn tick worlds spawn disabled exampleworld max spawns bat blaze chicken cow creeper enderman ghast husk horse mooshroom ocelot pig polarbear rabbit stray sheep skeleton spider wolf zombie crashdump backtrace or other files | 1 |
68,933 | 3,293,856,696 | IssuesEvent | 2015-10-30 21:03:44 | biocodellc/biocode-fims | https://api.github.com/repos/biocodellc/biocode-fims | closed | Create functional tests based | auto-migrated Priority-Medium Type-Enhancement | ```
Main method accepts
-spreadsheet
-configuration file
-util.List group level messages
Create a listing of Constants of group level messages in the renderers.Message
class, like:
public static final String MESSAGE_REQUIRED_COLUMN = "Yo, dude, you're missing
a required column";
public static final int MESSAGE _PRIMARY_KEY = "Primary key is missing";
This issue replaces the unit testing item, which we will NOT be doing at this
time.
```
Original issue reported on code.google.com by `jdec...@gmail.com` on 29 Dec 2014 at 11:16 | 1.0 | Create functional tests based - ```
Main method accepts
-spreadsheet
-configuration file
-util.List group level messages
Create a listing of Constants of group level messages in the renderers.Message
class, like:
public static final String MESSAGE_REQUIRED_COLUMN = "Yo, dude, you're missing
a required column";
public static final int MESSAGE _PRIMARY_KEY = "Primary key is missing";
This issue replaces the unit testing item, which we will NOT be doing at this
time.
```
Original issue reported on code.google.com by `jdec...@gmail.com` on 29 Dec 2014 at 11:16 | priority | create functional tests based main method accepts spreadsheet configuration file util list group level messages create a listing of constants of group level messages in the renderers message class like public static final string message required column yo dude you re missing a required column public static final int message primary key primary key is missing this issue replaces the unit testing item which we will not be doing at this time original issue reported on code google com by jdec gmail com on dec at | 1 |
658,714 | 21,900,897,411 | IssuesEvent | 2022-05-20 13:19:49 | bounswe/bounswe2022group9 | https://api.github.com/repos/bounswe/bounswe2022group9 | closed | Review Pull Request for "Berkant's API" | Priority: Medium Practice Application | Deadline: 20.05.2022 17.00
- [x] Review #199 . If you don't see any problem, merge it. | 1.0 | Review Pull Request for "Berkant's API" - Deadline: 20.05.2022 17.00
- [x] Review #199 . If you don't see any problem, merge it. | priority | review pull request for berkant s api deadline review if you don t see any problem merge it | 1 |
322,317 | 9,816,372,649 | IssuesEvent | 2019-06-13 14:32:28 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Allow JWT auth method to verify the signature but not enforce the policy | Priority: Medium customer request enhancement needs documenting | **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
In order to use this method, you must have permission to change the JWT and **inject** the policy id to it.
The problem is that JWTs are created and **signed** with a private key by the IdP, so injecting more info could be a problem to the API owner.
**What is the expected behavior?**
To be able to add a public key or shared key to the api def without choosing a specific policy.
Please note that this feature is not relevant to individual keys since in this case, you would need to update the JWT anyway.
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.7 | 1.0 | Allow JWT auth method to verify the signature but not enforce the policy - **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
In order to use this method, you must have permission to change the JWT and **inject** the policy id to it.
The problem is that JWTs are created and **signed** with a private key by the IdP, so injecting more info could be a problem to the API owner.
**What is the expected behavior?**
To be able to add a public key or shared key to the api def without choosing a specific policy.
Please note that this feature is not relevant to individual keys since in this case, you would need to update the JWT anyway.
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.7 | priority | allow jwt auth method to verify the signature but not enforce the policy do you want to request a feature or report a bug feature what is the current behavior in order to use this method you must have permission to change the jwt and inject the policy id to it the problem is that jwts are created and signed with a private key by the idp so injecting more info could be a problem to the api owner what is the expected behavior to be able to add a public key or shared key to the api def without choosing a specific policy please note that this feature is not relevant to individual keys since in this case you would need to update the jwt anyway if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem which versions of tyk affected by this issue did this work in previous versions of tyk | 1 |
365,280 | 10,780,370,689 | IssuesEvent | 2019-11-04 12:50:56 | emsec/hal | https://api.github.com/repos/emsec/hal | closed | Python allows setting invalid init strings | Priority: Medium Type: Feature Request discussion needed | **Describe the bug**
On a LUT6 you should set an init string this way:
```
lut.set_data("generic", "INIT", "bit_vector", "0"*16)
```
However, doing this also works, meaning you can set init strings of a wrong length:
```
lut.set_data("generic", "INIT", "bit_vector", "0"*4)
```
This causes all sorts of issues, including plugins that rely on correct config strings not working.
**To Reproduce**
Steps to reproduce the behavior:
1. Take any netlist with a LUT6.
2. Run the second code snippet.
3. See that HAL does not prevent this.
**Expected behavior**
HAL should not let the user do this if the init string clearly does not match the gate type.
**Additional context**
This works for other gates as well. | 1.0 | Python allows setting invalid init strings - **Describe the bug**
On a LUT6 you should set an init string this way:
```
lut.set_data("generic", "INIT", "bit_vector", "0"*16)
```
However, doing this also works, meaning you can set init strings of a wrong length:
```
lut.set_data("generic", "INIT", "bit_vector", "0"*4)
```
This causes all sorts of issues, including plugins that rely on correct config strings not working.
**To Reproduce**
Steps to reproduce the behavior:
1. Take any netlist with a LUT6.
2. Run the second code snippet.
3. See that HAL does not prevent this.
**Expected behavior**
HAL should not let the user do this if the init string clearly does not match the gate type.
**Additional context**
This works for other gates as well. | priority | python allows setting invalid init strings describe the bug on a you should set an init string this way lut set data generic init bit vector however doing this also works meaning you can set init strings of a wrong length lut set data generic init bit vector this causes all sorts of issues including plugins that rely on correct config strings not working to reproduce steps to reproduce the behavior take any netlist with a run the second code snippet see that hal does not prevent this expected behavior hal should not let the user do this if the init string clearly does not match the gate type additional context this works for other gates as well | 1 |
428,072 | 12,402,420,517 | IssuesEvent | 2020-05-21 11:57:29 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Have the ability to enable/disable a network behavior policy | Priority: Medium Type: Feature / Enhancement | **Is your feature request related to a problem? Please describe.**
Since we ship all of the default ones enabled by default, it may cause excessive emails from the default security events that go with them.
**Describe the solution you'd like**
We should have a flag to enable/disable them and ship the examples disabled in the future
**Describe alternatives you've considered**
Watching pictures of @lzammit to relax, didn't have much impact on what is above
**Additional context**
One of the pictures I've looked at:

| 1.0 | Have the ability to enable/disable a network behavior policy - **Is your feature request related to a problem? Please describe.**
Since we ship all of the default ones enabled by default, it may cause excessive emails from the default security events that go with them.
**Describe the solution you'd like**
We should have a flag to enable/disable them and ship the examples disabled in the future
**Describe alternatives you've considered**
Watching pictures of @lzammit to relax, didn't have much impact on what is above
**Additional context**
One of the pictures I've looked at:

| priority | have the ability to enable disable a network behavior policy is your feature request related to a problem please describe since we ship all of the default ones enabled by default it may cause excessive emails from the default security events that go with them describe the solution you d like we should have a flag to enable disable them and ship the examples disabled in the future describe alternatives you ve considered watching pictures of lzammit to relax didn t have much impact on what is above additional context one of the pictures i ve looked at | 1 |
327,565 | 9,977,465,264 | IssuesEvent | 2019-07-09 17:20:50 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | [BUG] | Nethergarde can get titles that don't belong to dark portal owner | :beetle: bug :beetle: :grey_exclamation: priority medium | **Mod Version**
4297099d
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Ironforge is a dark portal owner, but upon creation, Nethergarde get duchy that belongs to the reformed Horde
**Steps to reproduce the issue:**
_
**Upload an attachment below: .zip of your save, or screenshots:**
<details>

</details> | 1.0 | [BUG] | Nethergarde can get titles that don't belong to dark portal owner - **Mod Version**
4297099d
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Ironforge is a dark portal owner, but upon creation, Nethergarde get duchy that belongs to the reformed Horde
**Steps to reproduce the issue:**
_
**Upload an attachment below: .zip of your save, or screenshots:**
<details>

</details> | priority | nethergarde can get titles that don t belong to dark portal owner mod version what expansions do you have installed all please explain your issue in as much detail as possible ironforge is a dark portal owner but upon creation nethergarde get duchy that belongs to the reformed horde steps to reproduce the issue upload an attachment below zip of your save or screenshots | 1 |
20,859 | 2,631,827,224 | IssuesEvent | 2015-03-07 13:56:02 | trizen/sidef | https://api.github.com/repos/trizen/sidef | opened | Add non-logical boolean operators ("|" and "&") | enhancement medium priority TODO | ```ruby
true || say "hello"; # prints nothing
true | say "hello"; # prints "hello"
``` | 1.0 | Add non-logical boolean operators ("|" and "&") - ```ruby
true || say "hello"; # prints nothing
true | say "hello"; # prints "hello"
``` | priority | add non logical boolean operators and ruby true say hello prints nothing true say hello prints hello | 1 |
402,842 | 11,825,498,284 | IssuesEvent | 2020-03-21 13:13:10 | autonomio/trauma-team-international | https://api.github.com/repos/autonomio/trauma-team-international | opened | What is the patients journey to ventilation? | help-wanted priority: MEDIUM | - [ ] What is the playbook for ventilation in ICU setting? Is there an actual document we can read to learn what the decision criteria are?
- [ ] What procedures always are done before ventilation is started or can be started? | 1.0 | What is the patients journey to ventilation? - - [ ] What is the playbook for ventilation in ICU setting? Is there an actual document we can read to learn what the decision criteria are?
- [ ] What procedures always are done before ventilation is started or can be started? | priority | what is the patients journey to ventilation what is the playbook for ventilation in icu setting is there an actual document we can read to learn what the decision criteria are what procedures always are done before ventilation is started or can be started | 1 |
291,928 | 8,951,519,100 | IssuesEvent | 2019-01-25 14:11:44 | tigerfahd/TermProjectDS3 | https://api.github.com/repos/tigerfahd/TermProjectDS3 | opened | Team Member Roles and Bios | Medium Priority Writing | A listing of the team members and their assigned roles (and secondary roles), as well as short
bios to validate role choice. Role crossover is encouraged where possible, but it should be
described | 1.0 | Team Member Roles and Bios - A listing of the team members and their assigned roles (and secondary roles), as well as short
bios to validate role choice. Role crossover is encouraged where possible, but it should be
described | priority | team member roles and bios a listing of the team members and their assigned roles and secondary roles as well as short bios to validate role choice role crossover is encouraged where possible but it should be described | 1 |
437,191 | 12,564,797,859 | IssuesEvent | 2020-06-08 08:39:31 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Allow members to tag friends/contacts in media uploads | feature: enhancement priority: medium | Allow members to tag friends/contacts in media uploads.

**Describe alternatives you've considered**
This solution is working in your [BuddyBoss Media](https://www.buddyboss.com/product/buddyboss-media/) plugin.
- buddyboss-platform 1.4.1
| 1.0 | Allow members to tag friends/contacts in media uploads - Allow members to tag friends/contacts in media uploads.

**Describe alternatives you've considered**
This solution is working in your [BuddyBoss Media](https://www.buddyboss.com/product/buddyboss-media/) plugin.
- buddyboss-platform 1.4.1
| priority | allow members to tag friends contacts in media uploads allow members to tag friends contacts in media uploads describe alternatives you ve considered this solution is working in your plugin buddyboss platform | 1 |
559,590 | 16,565,797,251 | IssuesEvent | 2021-05-29 11:31:36 | sopra-fs21-group-25/sopra-fs21-jass-client | https://api.github.com/repos/sopra-fs21-group-25/sopra-fs21-jass-client | closed | ingame text chat: group text chat inside the game for all 4 game members | medium priority task | Time estimation: 24h
Part of user story #6 | 1.0 | ingame text chat: group text chat inside the game for all 4 game members - Time estimation: 24h
Part of user story #6 | priority | ingame text chat group text chat inside the game for all game members time estimation part of user story | 1 |
182,733 | 6,673,232,095 | IssuesEvent | 2017-10-04 14:28:33 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | Paging in WS /api/drugs doesn't work correctly | epic/medication kind/bug priority/medium project/reimbursement | GET {{host}}/api/drugs?innm_name=Будесонід
In response I receive 4 arrays of data, but total_entries = 1.
### Response
```
{
"paging": {
"total_pages": 1,
"total_entries": 1,
"page_size": 50,
"page_number": 1
},
"meta": {
"url": "http://dev.ehealth.world/api/drugs",
"type": "list",
"request_id": "15p46pqb0htpk3q7a7mppd3gj3cit7pp",
"code": 200
},
"data": [
{
"packages": [
{
"package_qty": 100,
"package_min_qty": 100,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
},
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 200 MKG Порошок для інгаляцій",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "116fb004-6f54-4c6c-a106-4f21cfd97e55",
"form": "INHALATION_POWDER",
"dosage": {
"numerator_value": 200,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
},
{
"packages": [
{
"package_qty": 20,
"package_min_qty": 5,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "ML",
"denumerator_value": 1,
"denumerator_unit": "CONTAINER"
}
}
],
"name": "Будесонід 0.5 MG Суспензія для розпилення",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "13cdb91d-8f4e-41bf-bb9f-9411b3ead549",
"form": "NEBULISER_SUSPENSION",
"dosage": {
"numerator_value": 0.5,
"numerator_unit": "MG",
"denumerator_value": 1,
"denumerator_unit": "ML"
}
},
{
"packages": [
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 200 MKG інгаляція під тиском",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "52a7dab5-31ac-4465-b782-03e5b9006e2a",
"form": "PRESSURISED_INHALATION",
"dosage": {
"numerator_value": 200,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
},
{
"packages": [
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 100 MKG Порошок для інгаляцій",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "f2ac4807-71f0-4bdd-b54a-2474dab03a8c",
"form": "INHALATION_POWDER",
"dosage": {
"numerator_value": 100,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
}
]
}
```
For example, in division list paging works correctly
```
{
"paging": {
"total_pages": 1,
"total_entries": 2,
"page_size": 15,
"page_number": 1
},
"meta": {
"url": "http://dev.ehealth.world/api/divisions",
"type": "list",
"request_id": "hl4b3044gv8al7nu2rve8da7nhs0bkmf",
"code": 200
},
"data": [
{
"type": "CLINIC",
"status": "ACTIVE",
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
],
"name": "Бориспільське відділення Клініки Борис",
"mountain_group": false,
"location": null,
"legal_entity_id": "443e901e-b365-45d2-be14-083ee0aba657",
"id": "6525d607-04e7-4fc8-8c6c-39849624ddbc",
"external_id": "3213213",
"email": "sp.virny+51@gmail.com",
"addresses": [
{
"zip": "02090",
"type": "REGISTRATION",
"street_type": "STREET",
"street": "вул. Ніжинська",
"settlement_type": "CITY",
"settlement_id": "707dbc55-cb6b-4aaa-97c1-2a1e03476100",
"settlement": "СОРОКИ-ЛЬВІВСЬКІ",
"region": "ПУСТОМИТІВСЬКИЙ",
"country": "UA",
"building": "15",
"area": "ЛЬВІВСЬКА",
"apartment": "23"
}
]
},
{
"type": "CLINIC",
"status": "ACTIVE",
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
],
"name": "Бориспільське відділення Клініки Борис",
"mountain_group": false,
"location": null,
"legal_entity_id": "443e901e-b365-45d2-be14-083ee0aba657",
"id": "b5fe1184-cee2-4948-a9b9-d0389dece95a",
"external_id": "3213213",
"email": "sp.virny+51@gmail.com",
"addresses": [
{
"zip": "02090",
"type": "REGISTRATION",
"street_type": "STREET",
"street": "вул. Ніжинська",
"settlement_type": "CITY",
"settlement_id": "707dbc55-cb6b-4aaa-97c1-2a1e03476100",
"settlement": "СОРОКИ-ЛЬВІВСЬКІ",
"region": "ПУСТОМИТІВСЬКИЙ",
"country": "UA",
"building": "15",
"area": "ЛЬВІВСЬКА",
"apartment": "23"
}
]
}
]
}
``` | 1.0 | Paging in WS /api/drugs doesn't work correctly - GET {{host}}/api/drugs?innm_name=Будесонід
In response I receive 4 arrays of data, but total_entries = 1.
### Response
```
{
"paging": {
"total_pages": 1,
"total_entries": 1,
"page_size": 50,
"page_number": 1
},
"meta": {
"url": "http://dev.ehealth.world/api/drugs",
"type": "list",
"request_id": "15p46pqb0htpk3q7a7mppd3gj3cit7pp",
"code": 200
},
"data": [
{
"packages": [
{
"package_qty": 100,
"package_min_qty": 100,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
},
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 200 MKG Порошок для інгаляцій",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "116fb004-6f54-4c6c-a106-4f21cfd97e55",
"form": "INHALATION_POWDER",
"dosage": {
"numerator_value": 200,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
},
{
"packages": [
{
"package_qty": 20,
"package_min_qty": 5,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "ML",
"denumerator_value": 1,
"denumerator_unit": "CONTAINER"
}
}
],
"name": "Будесонід 0.5 MG Суспензія для розпилення",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "13cdb91d-8f4e-41bf-bb9f-9411b3ead549",
"form": "NEBULISER_SUSPENSION",
"dosage": {
"numerator_value": 0.5,
"numerator_unit": "MG",
"denumerator_value": 1,
"denumerator_unit": "ML"
}
},
{
"packages": [
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 200 MKG інгаляція під тиском",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "52a7dab5-31ac-4465-b782-03e5b9006e2a",
"form": "PRESSURISED_INHALATION",
"dosage": {
"numerator_value": 200,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
},
{
"packages": [
{
"package_qty": 200,
"package_min_qty": 200,
"container_dosage": {
"numerator_value": 1,
"numerator_unit": "DOSE",
"denumerator_value": 1,
"denumerator_unit": "AEROSOL"
}
}
],
"name": "Будесонід 100 MKG Порошок для інгаляцій",
"innm": {
"sctid": null,
"name_original": "Budesonide",
"name": "Будесонід",
"id": "fcadd9e5-cf91-45aa-b7fb-ffd706b59d01"
},
"id": "f2ac4807-71f0-4bdd-b54a-2474dab03a8c",
"form": "INHALATION_POWDER",
"dosage": {
"numerator_value": 100,
"numerator_unit": "MKG",
"denumerator_value": 1,
"denumerator_unit": "DOSE"
}
}
]
}
```
For example, in division list paging works correctly
```
{
"paging": {
"total_pages": 1,
"total_entries": 2,
"page_size": 15,
"page_number": 1
},
"meta": {
"url": "http://dev.ehealth.world/api/divisions",
"type": "list",
"request_id": "hl4b3044gv8al7nu2rve8da7nhs0bkmf",
"code": 200
},
"data": [
{
"type": "CLINIC",
"status": "ACTIVE",
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
],
"name": "Бориспільське відділення Клініки Борис",
"mountain_group": false,
"location": null,
"legal_entity_id": "443e901e-b365-45d2-be14-083ee0aba657",
"id": "6525d607-04e7-4fc8-8c6c-39849624ddbc",
"external_id": "3213213",
"email": "sp.virny+51@gmail.com",
"addresses": [
{
"zip": "02090",
"type": "REGISTRATION",
"street_type": "STREET",
"street": "вул. Ніжинська",
"settlement_type": "CITY",
"settlement_id": "707dbc55-cb6b-4aaa-97c1-2a1e03476100",
"settlement": "СОРОКИ-ЛЬВІВСЬКІ",
"region": "ПУСТОМИТІВСЬКИЙ",
"country": "UA",
"building": "15",
"area": "ЛЬВІВСЬКА",
"apartment": "23"
}
]
},
{
"type": "CLINIC",
"status": "ACTIVE",
"phones": [
{
"type": "MOBILE",
"number": "+380503410870"
}
],
"name": "Бориспільське відділення Клініки Борис",
"mountain_group": false,
"location": null,
"legal_entity_id": "443e901e-b365-45d2-be14-083ee0aba657",
"id": "b5fe1184-cee2-4948-a9b9-d0389dece95a",
"external_id": "3213213",
"email": "sp.virny+51@gmail.com",
"addresses": [
{
"zip": "02090",
"type": "REGISTRATION",
"street_type": "STREET",
"street": "вул. Ніжинська",
"settlement_type": "CITY",
"settlement_id": "707dbc55-cb6b-4aaa-97c1-2a1e03476100",
"settlement": "СОРОКИ-ЛЬВІВСЬКІ",
"region": "ПУСТОМИТІВСЬКИЙ",
"country": "UA",
"building": "15",
"area": "ЛЬВІВСЬКА",
"apartment": "23"
}
]
}
]
}
``` | priority | paging in ws api drugs doesn t work correctly get host api drugs innm name будесонід in response i receive arrays of data but total entries response paging total pages total entries page size page number meta url type list request id code data packages package qty package min qty container dosage numerator value numerator unit dose denumerator value denumerator unit aerosol package qty package min qty container dosage numerator value numerator unit dose denumerator value denumerator unit aerosol name будесонід mkg порошок для інгаляцій innm sctid null name original budesonide name будесонід id id form inhalation powder dosage numerator value numerator unit mkg denumerator value denumerator unit dose packages package qty package min qty container dosage numerator value numerator unit ml denumerator value denumerator unit container name будесонід mg суспензія для розпилення innm sctid null name original budesonide name будесонід id id form nebuliser suspension dosage numerator value numerator unit mg denumerator value denumerator unit ml packages package qty package min qty container dosage numerator value numerator unit dose denumerator value denumerator unit aerosol name будесонід mkg інгаляція під тиском innm sctid null name original budesonide name будесонід id id form pressurised inhalation dosage numerator value numerator unit mkg denumerator value denumerator unit dose packages package qty package min qty container dosage numerator value numerator unit dose denumerator value denumerator unit aerosol name будесонід mkg порошок для інгаляцій innm sctid null name original budesonide name будесонід id id form inhalation powder dosage numerator value numerator unit mkg denumerator value denumerator unit dose for example in division list paging works correctly paging total pages total entries page size page number meta url type list request id code data type clinic status active phones type mobile number name бориспільське відділення клініки борис mountain group false location null legal entity id id external id email sp virny gmail com addresses zip type registration street type street street вул ніжинська settlement type city settlement id settlement сороки львівські region пустомитівський country ua building area львівська apartment type clinic status active phones type mobile number name бориспільське відділення клініки борис mountain group false location null legal entity id id external id email sp virny gmail com addresses zip type registration street type street street вул ніжинська settlement type city settlement id settlement сороки львівські region пустомитівський country ua building area львівська apartment | 1 |
656,314 | 21,726,326,425 | IssuesEvent | 2022-05-11 08:00:51 | SimplyVC/panic | https://api.github.com/repos/SimplyVC/panic | closed | Installation Wizard - Add Repos | 3 SP UI iteration 2 Priority: Medium | As a node operator, I want to be able to add repositories that I want to be alerted on.
**Requirements**
- A "home page" for Repositories setup must be created and be displayed at _/installation/repositories_.
- The page must have two central elements:
- A button with the label "Add Repository", allowing the node operator to click and open the form. The form will be shown in a modal (no redirection to a different page).
- A **svc-data-table** which is only rendered when the node operator has added at least one node. The data table will be set to **crud** mode, therefore rendering two buttons, for editing/deleting respectively:
- When the node operator clicks in the "edit" button, a modal is shown with a form populated with the repository data.
- When the node operator clicks in the "delete" button, a dialog is shown asking the node operator to confirm the action.
- The page must have other 3 secondary elements:
- A "go back". It leads the node operator to the previous step (Add Node).
- Next button, allowing the node operator to proceed to the next step (Alerts setup)
- When the "next" button is clicked, the current configuration is saved in the local storage (browser cache)
- A "help" button so the node operator can click and get to know more information about what he's configuring at the moment. This in-depth info must be displayed in a modal. | 1.0 | Installation Wizard - Add Repos - As a node operator, I want to be able to add repositories that I want to be alerted on.
**Requirements**
- A "home page" for Repositories setup must be created and be displayed at _/installation/repositories_.
- The page must have two central elements:
- A button with the label "Add Repository", allowing the node operator to click and open the form. The form will be shown in a modal (no redirection to a different page).
- A **svc-data-table** which is only rendered when the node operator has added at least one node. The data table will be set to **crud** mode, therefore rendering two buttons, for editing/deleting respectively:
- When the node operator clicks in the "edit" button, a modal is shown with a form populated with the repository data.
- When the node operator clicks in the "delete" button, a dialog is shown asking the node operator to confirm the action.
- The page must have other 3 secondary elements:
- A "go back". It leads the node operator to the previous step (Add Node).
- Next button, allowing the node operator to proceed to the next step (Alerts setup)
- When the "next" button is clicked, the current configuration is saved in the local storage (browser cache)
- A "help" button so the node operator can click and get to know more information about what he's configuring at the moment. This in-depth info must be displayed in a modal. | priority | installation wizard add repos as a node operator i want to be able to add repositories that i want to be alerted on requirements a home page for repositories setup must be created and be displayed at installation repositories the page must have two central elements a button with the label add repository allowing the node operator to click and open the form the form will be shown in a modal no redirection to a different page a svc data table which is only rendered when the node operator has added at least one node the data table will be set to crud mode therefore rendering two buttons for editing deleting respectively when the node operator clicks in the edit button a modal is shown with a form populated with the repository data when the node operator clicks in the delete button a dialog is shown asking the node operator to confirm the action the page must have other secondary elements a go back it leads the node operator to the previous step add node next button allowing the node operator to proceed to the next step alerts setup when the next button is clicked the current configuration is saved in the local storage browser cache a help button so the node operator can click and get to know more information about what he s configuring at the moment this in depth info must be displayed in a modal | 1 |
596,158 | 18,098,860,731 | IssuesEvent | 2021-09-22 12:13:51 | EddieHubCommunity/LinkFree | https://api.github.com/repos/EddieHubCommunity/LinkFree | closed | Add issue forms and pull request template | ⭐ goal: addition 🏁 status: ready for dev 🟨 priority: medium good first issue | We need issue forms, bug forms and other templates to facilitate smoother workflow | 1.0 | Add issue forms and pull request template - We need issue forms, bug forms and other templates to facilitate smoother workflow | priority | add issue forms and pull request template we need issue forms bug forms and other templates to facilitate smoother workflow | 1 |
751,824 | 26,259,958,697 | IssuesEvent | 2023-01-06 06:32:46 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [CDCSDK] DROP TABLE - ysqlsh crashes/connection reset happens when a table having an active CDC connection is dropped | kind/bug priority/medium area/cdcsdk | Jira Link: [DB-3517](https://yugabyte.atlassian.net/browse/DB-3517)
If we initiate a CDC connection to a table test, and then without closing the connection, we drop the table, the following set of events happen-
```
FATAL: terminating connection due to unexpected postmaster exit
server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
The connection to the server was lost. Attempting reset:
```
Steps to reproduce:
- Create a table test and run the java client to attach CDC to this table
- Drop the table using `drop table test;` | 1.0 | [CDCSDK] DROP TABLE - ysqlsh crashes/connection reset happens when a table having an active CDC connection is dropped - Jira Link: [DB-3517](https://yugabyte.atlassian.net/browse/DB-3517)
If we initiate a CDC connection to a table test, and then without closing the connection, we drop the table, the following set of events happen-
```
FATAL: terminating connection due to unexpected postmaster exit
server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
The connection to the server was lost. Attempting reset:
```
Steps to reproduce:
- Create a table test and run the java client to attach CDC to this table
- Drop the table using `drop table test;` | priority | drop table ysqlsh crashes connection reset happens when a table having an active cdc connection is dropped jira link if we initiate a cdc connection to a table test and then without closing the connection we drop the table the following set of events happen fatal terminating connection due to unexpected postmaster exit server closed the connection unexpectedly this probably means the server terminated abnormally before or while processing the request the connection to the server was lost attempting reset steps to reproduce create a table test and run the java client to attach cdc to this table drop the table using drop table test | 1 |
266,405 | 8,367,145,354 | IssuesEvent | 2018-10-04 11:17:56 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | Regression - pageSize "all" does not work correctly with the grid's Excel export | Bug C: Grid C: Pager Kendo1 Priority 3 SEV: Medium | ### Bug report
Ticket ID 1155619
### Reproduction of the problem
[Instructions in the Dojo](https://dojo.telerik.com/@bubblemaster/EvopOPaL)
### Current behavior
Since 2018 SP1 when pageSize is "all" or equal to the total of the dataSource, the excel export contains the original pageSize, instead of the total which results in an unexpected Excel export. This is a regression from the fix of the pager [here.](https://github.com/telerik/kendo-ui-core/issues/3424).
### Observations
The issue arises because we are setting the pageSize to undefined. During the Excel export, the grid's data source is extended with the current view and passed to the excel exporter. Since the pageSize is no longer defined(it used to be `dataSource.pageSize(dataSource.total() ? dataSource.total() : MAX_VALUE);`, $.extend takes the initial grid dataSource pageSize.
### Workaround
Add a handler to the data source [`change`](https://docs.telerik.com/kendo-ui/api/javascript/data/datasource/events/change) event and if the internal property _pageSize is `undefined`, set it as the total of the data source:
```
change: function(e){
if(!e.sender._pageSize){
e.sender._pageSize = e.sender.total();
}
}
```
### Expected/desired behavior
When the grid pager is changed to "All", all records should be exported as it is in 2018 R1 and before.
### Environment
* **Kendo UI version:** 2018.1.221
| 1.0 | Regression - pageSize "all" does not work correctly with the grid's Excel export - ### Bug report
Ticket ID 1155619
### Reproduction of the problem
[Instructions in the Dojo](https://dojo.telerik.com/@bubblemaster/EvopOPaL)
### Current behavior
Since 2018 SP1 when pageSize is "all" or equal to the total of the dataSource, the excel export contains the original pageSize, instead of the total which results in an unexpected Excel export. This is a regression from the fix of the pager [here.](https://github.com/telerik/kendo-ui-core/issues/3424).
### Observations
The issue arises because we are setting the pageSize to undefined. During the Excel export, the grid's data source is extended with the current view and passed to the excel exporter. Since the pageSize is no longer defined(it used to be `dataSource.pageSize(dataSource.total() ? dataSource.total() : MAX_VALUE);`, $.extend takes the initial grid dataSource pageSize.
### Workaround
Add a handler to the data source [`change`](https://docs.telerik.com/kendo-ui/api/javascript/data/datasource/events/change) event and if the internal property _pageSize is `undefined`, set it as the total of the data source:
```
change: function(e){
if(!e.sender._pageSize){
e.sender._pageSize = e.sender.total();
}
}
```
### Expected/desired behavior
When the grid pager is changed to "All", all records should be exported as it is in 2018 R1 and before.
### Environment
* **Kendo UI version:** 2018.1.221
| priority | regression pagesize all does not work correctly with the grid s excel export bug report ticket id reproduction of the problem current behavior since when pagesize is all or equal to the total of the datasource the excel export contains the original pagesize instead of the total which results in an unexpected excel export this is a regression from the fix of the pager observations the issue arises because we are setting the pagesize to undefined during the excel export the grid s data source is extended with the current view and passed to the excel exporter since the pagesize is no longer defined it used to be datasource pagesize datasource total datasource total max value extend takes the initial grid datasource pagesize workaround add a handler to the data source event and if the internal property pagesize is undefined set it as the total of the data source change function e if e sender pagesize e sender pagesize e sender total expected desired behavior when the grid pager is changed to all all records should be exported as it is in and before environment kendo ui version | 1 |
395,211 | 11,672,637,730 | IssuesEvent | 2020-03-04 07:11:14 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | disputing and reporting right hand panels should only switch to mobile components at mobile break. | Add post v2 launch Priority: Medium | Looks like it's just the My available rep balance and current dispute window components that are switching to mobile components too early. My available REP balance shouldn't be collapsable until mobile.
Build:

Design:
https://www.figma.com/file/aAzKHh4cA6OT2t7WFv2BQ7fB/Reporting-and-Disputing?node-id=2192%3A0
 | 1.0 | disputing and reporting right hand panels should only switch to mobile components at mobile break. - Looks like it's just the My available rep balance and current dispute window components that are switching to mobile components too early. My available REP balance shouldn't be collapsable until mobile.
Build:

Design:
https://www.figma.com/file/aAzKHh4cA6OT2t7WFv2BQ7fB/Reporting-and-Disputing?node-id=2192%3A0
 | priority | disputing and reporting right hand panels should only switch to mobile components at mobile break looks like it s just the my available rep balance and current dispute window components that are switching to mobile components too early my available rep balance shouldn t be collapsable until mobile build design | 1 |
4,665 | 2,562,579,139 | IssuesEvent | 2015-02-06 03:20:58 | cs2103jan2015-t09-3j/main | https://api.github.com/repos/cs2103jan2015-t09-3j/main | opened | As a user, I want to cancel my done tasks instead of deleting them immediately when I completed that tasks | priority.medium type.enhancement type.task | canceling instead of deleting the tasks can let user review their tasks done (if there is a need to) | 1.0 | As a user, I want to cancel my done tasks instead of deleting them immediately when I completed that tasks - canceling instead of deleting the tasks can let user review their tasks done (if there is a need to) | priority | as a user i want to cancel my done tasks instead of deleting them immediately when i completed that tasks canceling instead of deleting the tasks can let user review their tasks done if there is a need to | 1 |
392,709 | 11,594,792,938 | IssuesEvent | 2020-02-24 15:54:48 | aol/moloch | https://api.github.com/repos/aol/moloch | closed | Improve Db.update for tags,hunts,... | enhancement medium priority v2.0 viewer | Once we require ES 6 change add tags/hunts/... updates to do a retry and to actually do appending instead of replacement. So something like
<pre>"script": "if (ctx._source.containsKey(\"tags\")) {ctx._source.tag += tags;ctx._source.tagsCnt++} else {ctx._source.tags = [tag]; ctx._source.tagsCnt = 1}"
</pre>
Even better if it checks if the element is in there or not before adding. | 1.0 | Improve Db.update for tags,hunts,... - Once we require ES 6 change add tags/hunts/... updates to do a retry and to actually do appending instead of replacement. So something like
<pre>"script": "if (ctx._source.containsKey(\"tags\")) {ctx._source.tag += tags;ctx._source.tagsCnt++} else {ctx._source.tags = [tag]; ctx._source.tagsCnt = 1}"
</pre>
Even better if it checks if the element is in there or not before adding. | priority | improve db update for tags hunts once we require es change add tags hunts updates to do a retry and to actually do appending instead of replacement so something like script if ctx source containskey tags ctx source tag tags ctx source tagscnt else ctx source tags ctx source tagscnt even better if it checks if the element is in there or not before adding | 1 |
731,641 | 25,225,361,947 | IssuesEvent | 2022-11-14 15:36:28 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | closed | Frontend: .env file and using environment variables in the code | priority-medium status-new front-end | ### Issue Description
During the implementation of the first Milestone outputs, we also thought that it would be better if we use .env file and environment variables in the code instead of statically declaring the API URLs or other kinds of things similar to this. So, we decided to implement this feature in the [meeting](https://github.com/bounswe/bounswe2022group2/wiki/Meeting-%232-of-the-Frontend-Team). So, I will do this implementation
### Step Details
Steps that will be performed:
- [x] research on how to use environment variables
- [x] declare environment variables
- [x] share the .env file with the frontend team
### Final Actions
Reviewed by the team members and merged
### Deadline of the Issue
08.11.2022 - Tuesday - 23.59
### Reviewer
Altay Acar
### Deadline for the Review
09.11.2022 - Wednesday - 23.59 | 1.0 | Frontend: .env file and using environment variables in the code - ### Issue Description
During the implementation of the first Milestone outputs, we also thought that it would be better if we use .env file and environment variables in the code instead of statically declaring the API URLs or other kinds of things similar to this. So, we decided to implement this feature in the [meeting](https://github.com/bounswe/bounswe2022group2/wiki/Meeting-%232-of-the-Frontend-Team). So, I will do this implementation
### Step Details
Steps that will be performed:
- [x] research on how to use environment variables
- [x] declare environment variables
- [x] share the .env file with the frontend team
### Final Actions
Reviewed by the team members and merged
### Deadline of the Issue
08.11.2022 - Tuesday - 23.59
### Reviewer
Altay Acar
### Deadline for the Review
09.11.2022 - Wednesday - 23.59 | priority | frontend env file and using environment variables in the code issue description during the implementation of the first milestone outputs we also thought that it would be better if we use env file and environment variables in the code instead of statically declaring the api urls or other kinds of things similar to this so we decided to implement this feature in the so i will do this implementation step details steps that will be performed research on how to use environment variables declare environment variables share the env file with the frontend team final actions reviewed by the team members and merged deadline of the issue tuesday reviewer altay acar deadline for the review wednesday | 1 |
4,632 | 2,559,814,380 | IssuesEvent | 2015-02-05 05:29:11 | cs2103jan2015-t15-4j/main | https://api.github.com/repos/cs2103jan2015-t15-4j/main | opened | A user can update an existing task by selecting and then editing it | priority.medium type.story | ...so that there is no need to remove and then add a task with almost identical parameters. | 1.0 | A user can update an existing task by selecting and then editing it - ...so that there is no need to remove and then add a task with almost identical parameters. | priority | a user can update an existing task by selecting and then editing it so that there is no need to remove and then add a task with almost identical parameters | 1 |
782,777 | 27,506,784,201 | IssuesEvent | 2023-03-06 04:40:38 | priyanshujain/leben | https://api.github.com/repos/priyanshujain/leben | opened | Physical exercise tool to record variable progress | type:feature priority:medium state:research | For ex. count the goal of daily 50 pushups.
For ex. count the goal of daily 15 mins meditation.
| 1.0 | Physical exercise tool to record variable progress - For ex. count the goal of daily 50 pushups.
For ex. count the goal of daily 15 mins meditation.
| priority | physical exercise tool to record variable progress for ex count the goal of daily pushups for ex count the goal of daily mins meditation | 1 |
658,243 | 21,882,644,025 | IssuesEvent | 2022-05-19 15:32:11 | pvs-hd-tea/LapsPython | https://api.github.com/repos/pvs-hd-tea/LapsPython | opened | Implement R code generation from ASTs | Feature Request Medium Priority | Although the ASTs are based on Python, it will be possible to generate corresponding code from them written in R. This requires an additional rule DB. | 1.0 | Implement R code generation from ASTs - Although the ASTs are based on Python, it will be possible to generate corresponding code from them written in R. This requires an additional rule DB. | priority | implement r code generation from asts although the asts are based on python it will be possible to generate corresponding code from them written in r this requires an additional rule db | 1 |
281,909 | 8,700,725,820 | IssuesEvent | 2018-12-05 09:34:22 | bitshares/bitshares-ui | https://api.github.com/repos/bitshares/bitshares-ui | closed | [2] Enable Withdraw EOS.* Gateway Asset | [1c] Task [3] Refactoring [4d] Critical Priority [5c] Medium [7] Finalized | ## To enable OpenLedger EOS assets
Open Ledger has a asset type name of EOS.*
For example,
- EOS is named EOS.EOS
- IQ is named EOS.IQ
While other gateways does not have the prepended EOS name. This requires some refactoring of the code to enable EOS.* assets pairing.
This issue is to enable withdraw of said asset through the Gateway modal. | 1.0 | [2] Enable Withdraw EOS.* Gateway Asset - ## To enable OpenLedger EOS assets
Open Ledger has a asset type name of EOS.*
For example,
- EOS is named EOS.EOS
- IQ is named EOS.IQ
While other gateways does not have the prepended EOS name. This requires some refactoring of the code to enable EOS.* assets pairing.
This issue is to enable withdraw of said asset through the Gateway modal. | priority | enable withdraw eos gateway asset to enable openledger eos assets open ledger has a asset type name of eos for example eos is named eos eos iq is named eos iq while other gateways does not have the prepended eos name this requires some refactoring of the code to enable eos assets pairing this issue is to enable withdraw of said asset through the gateway modal | 1 |
341,076 | 10,282,480,591 | IssuesEvent | 2019-08-26 11:16:15 | input-output-hk/cardano-wallet | https://api.github.com/repos/input-output-hk/cardano-wallet | closed | Implement `KnownAddresses` for random scheme | PRIORITY[MEDIUM] | # Context
[Cardano.Wallet.Primitive.AddressDiscovery](https://github.com/input-output-hk/cardano-wallet/blob/e752723b35764d57fe50aa2998a0eece7167709c/lib/core/src/Cardano/Wallet/Primitive/AddressDiscovery.hs#L186-L194)
```hs
-- | Extract the list of all known addresses.
--
-- NOTE: Change addresses aren't considered "known" until they've been used. The
-- rationale is that, we don't want users or consumers of the wallet to be using
-- change addresses prematurely.
class KnownAddresses s where
knownAddresses
:: s
-> [Address]
```
# Decision
Unlike sequential derivation, we can't derive this information from the index only. We will likely need to store this information in the state too. This overlaps to some extent with the 'blacklist' approach described in #560 regarding generation of change addresses.
# Acceptance Criteria
1. `KnownAddresses` _must_ be implemented for the random scheme
---
# Development Plan
1. Use the discovered addresses list from RndState as, implemented in #560.
2. Pending addresses are not part of the list.
# PR
| Number | Base |
| --- | --- |
| #625 | `master` |
| #657 | `master` |
# QA
The properties of `KnownAddresses` are covered in `AddressDerivation.RandomSpec`.
| 1.0 | Implement `KnownAddresses` for random scheme - # Context
[Cardano.Wallet.Primitive.AddressDiscovery](https://github.com/input-output-hk/cardano-wallet/blob/e752723b35764d57fe50aa2998a0eece7167709c/lib/core/src/Cardano/Wallet/Primitive/AddressDiscovery.hs#L186-L194)
```hs
-- | Extract the list of all known addresses.
--
-- NOTE: Change addresses aren't considered "known" until they've been used. The
-- rationale is that, we don't want users or consumers of the wallet to be using
-- change addresses prematurely.
class KnownAddresses s where
knownAddresses
:: s
-> [Address]
```
# Decision
Unlike sequential derivation, we can't derive this information from the index only. We will likely need to store this information in the state too. This overlaps to some extent with the 'blacklist' approach described in #560 regarding generation of change addresses.
# Acceptance Criteria
1. `KnownAddresses` _must_ be implemented for the random scheme
---
# Development Plan
1. Use the discovered addresses list from RndState as, implemented in #560.
2. Pending addresses are not part of the list.
# PR
| Number | Base |
| --- | --- |
| #625 | `master` |
| #657 | `master` |
# QA
The properties of `KnownAddresses` are covered in `AddressDerivation.RandomSpec`.
| priority | implement knownaddresses for random scheme context hs extract the list of all known addresses note change addresses aren t considered known until they ve been used the rationale is that we don t want users or consumers of the wallet to be using change addresses prematurely class knownaddresses s where knownaddresses s decision unlike sequential derivation we can t derive this information from the index only we will likely need to store this information in the state too this overlaps to some extent with the blacklist approach described in regarding generation of change addresses acceptance criteria knownaddresses must be implemented for the random scheme development plan use the discovered addresses list from rndstate as implemented in pending addresses are not part of the list pr number base master master qa the properties of knownaddresses are covered in addressderivation randomspec | 1 |
794,065 | 28,021,132,157 | IssuesEvent | 2023-03-28 05:31:51 | rpm-software-management/dnf5 | https://api.github.com/repos/rpm-software-management/dnf5 | closed | Move a content of CHANGES-DNF-5.md to DNF5 documentation | Priority: MEDIUM | It will increase the visibility of the changes between DNF4 and DNF5 | 1.0 | Move a content of CHANGES-DNF-5.md to DNF5 documentation - It will increase the visibility of the changes between DNF4 and DNF5 | priority | move a content of changes dnf md to documentation it will increase the visibility of the changes between and | 1 |
541,685 | 15,831,495,116 | IssuesEvent | 2021-04-06 13:41:38 | inf112-v21/Trivial-Solution | https://api.github.com/repos/inf112-v21/Trivial-Solution | opened | SanityCheck | Medium priority | Vi har satt opp multiplayer slik at alle maskinene simulerer det samme hver for seg. Men hvordan kan vi være sikker på at alt blir likt hos alle? Jo, vi implementerer en klasse SanityCheck, som holder på all informasjonen til et Board på en kompakt måte, og en metode getSanityCheck i Board som lager et sånt objekt. Så trenger bare serveren og klientene å sende sånne objekter frem og tilbake en gang per runde eller noe sånt. Om alt går bra trenger vi ikke disse objektene, men om noe blir simulert ulikt vil vi nå ha en måte å oppdage det på.
Oppgave: Implementer klassen SanityCheck og metoden board.getSanityCheck(), slik at vi kan sjekke at alt er likt. | 1.0 | SanityCheck - Vi har satt opp multiplayer slik at alle maskinene simulerer det samme hver for seg. Men hvordan kan vi være sikker på at alt blir likt hos alle? Jo, vi implementerer en klasse SanityCheck, som holder på all informasjonen til et Board på en kompakt måte, og en metode getSanityCheck i Board som lager et sånt objekt. Så trenger bare serveren og klientene å sende sånne objekter frem og tilbake en gang per runde eller noe sånt. Om alt går bra trenger vi ikke disse objektene, men om noe blir simulert ulikt vil vi nå ha en måte å oppdage det på.
Oppgave: Implementer klassen SanityCheck og metoden board.getSanityCheck(), slik at vi kan sjekke at alt er likt. | priority | sanitycheck vi har satt opp multiplayer slik at alle maskinene simulerer det samme hver for seg men hvordan kan vi være sikker på at alt blir likt hos alle jo vi implementerer en klasse sanitycheck som holder på all informasjonen til et board på en kompakt måte og en metode getsanitycheck i board som lager et sånt objekt så trenger bare serveren og klientene å sende sånne objekter frem og tilbake en gang per runde eller noe sånt om alt går bra trenger vi ikke disse objektene men om noe blir simulert ulikt vil vi nå ha en måte å oppdage det på oppgave implementer klassen sanitycheck og metoden board getsanitycheck slik at vi kan sjekke at alt er likt | 1 |
792,485 | 27,962,721,316 | IssuesEvent | 2023-03-24 16:50:39 | camsaul/toucan2 | https://api.github.com/repos/camsaul/toucan2 | opened | Consider implementing a version of `before-update` that doesn't fetch the originals | enhancement under consideration medium priority performance | Sometimes all you want to do is add a key like `:updated_at (t/offset-date-time)`, there is really no need to fetch the original instance in this case. Seems like a significant performance win.
But really the last thing we need to do is make Toucan 2 even more complicated, so maybe we can noodle and think of a way that we can do this automagically. Can `before-update` determine whether any columns are accessed from the original object somehow, and only fetch it if they are? TBD. | 1.0 | Consider implementing a version of `before-update` that doesn't fetch the originals - Sometimes all you want to do is add a key like `:updated_at (t/offset-date-time)`, there is really no need to fetch the original instance in this case. Seems like a significant performance win.
But really the last thing we need to do is make Toucan 2 even more complicated, so maybe we can noodle and think of a way that we can do this automagically. Can `before-update` determine whether any columns are accessed from the original object somehow, and only fetch it if they are? TBD. | priority | consider implementing a version of before update that doesn t fetch the originals sometimes all you want to do is add a key like updated at t offset date time there is really no need to fetch the original instance in this case seems like a significant performance win but really the last thing we need to do is make toucan even more complicated so maybe we can noodle and think of a way that we can do this automagically can before update determine whether any columns are accessed from the original object somehow and only fetch it if they are tbd | 1 |
1,863 | 2,520,915,142 | IssuesEvent | 2015-01-19 09:57:52 | clementine-player/Clementine | https://api.github.com/repos/clementine-player/Clementine | closed | Soundcloud.com integration | Component-Radio enhancement imported Priority-Medium | _From [n...@leverton.org](https://code.google.com/u/114311011906226196511/) on December 31, 2011 20:48:26_
It would be really great to add integration with soundcloud.com as a music source. Play/download from there, rss feed, follow, rate, add timed comments etc :)
Also mixcrate, mixcloud, etc etc
Thanks for considering this suggestion !
_Original issue: http://code.google.com/p/clementine-player/issues/detail?id=2540_ | 1.0 | Soundcloud.com integration - _From [n...@leverton.org](https://code.google.com/u/114311011906226196511/) on December 31, 2011 20:48:26_
It would be really great to add integration with soundcloud.com as a music source. Play/download from there, rss feed, follow, rate, add timed comments etc :)
Also mixcrate, mixcloud, etc etc
Thanks for considering this suggestion !
_Original issue: http://code.google.com/p/clementine-player/issues/detail?id=2540_ | priority | soundcloud com integration from on december it would be really great to add integration with soundcloud com as a music source play download from there rss feed follow rate add timed comments etc also mixcrate mixcloud etc etc thanks for considering this suggestion original issue | 1 |
534,179 | 15,611,487,846 | IssuesEvent | 2021-03-19 14:22:36 | AY2021S2-CS2113T-W09-1/tp | https://api.github.com/repos/AY2021S2-CS2113T-W09-1/tp | closed | Modify AddCommand to accept a person's name for a loan type | priority.Medium | A `-p` option for `add -l ...` only.
Example:
`add -l 2nd loan to gerard -a 200.00 -d 18/03/2021 -p gerard` | 1.0 | Modify AddCommand to accept a person's name for a loan type - A `-p` option for `add -l ...` only.
Example:
`add -l 2nd loan to gerard -a 200.00 -d 18/03/2021 -p gerard` | priority | modify addcommand to accept a person s name for a loan type a p option for add l only example add l loan to gerard a d p gerard | 1 |
795,231 | 28,066,530,193 | IssuesEvent | 2023-03-29 15:48:09 | AY2223S2-CS2113-T12-3/tp | https://api.github.com/repos/AY2223S2-CS2113-T12-3/tp | closed | [Task] Development Guide for UI class and List commands | priority.Medium | 1. Describe the Task
Starting on the Development Guide for UI class and List commands
Draw Sequence Diagrams for the list commands
| 1.0 | [Task] Development Guide for UI class and List commands - 1. Describe the Task
Starting on the Development Guide for UI class and List commands
Draw Sequence Diagrams for the list commands
| priority | development guide for ui class and list commands describe the task starting on the development guide for ui class and list commands draw sequence diagrams for the list commands | 1 |
101,258 | 4,111,436,096 | IssuesEvent | 2016-06-07 06:07:39 | MinetestForFun/server-minetestforfun | https://api.github.com/repos/MinetestForFun/server-minetestforfun | closed | we can't use the shift-click in darkage:box | Modding Priority: Medium |

to put and get all items from and to the box | 1.0 | we can't use the shift-click in darkage:box -

to put and get all items from and to the box | priority | we can t use the shift click in darkage box to put and get all items from and to the box | 1 |
340,012 | 10,265,629,617 | IssuesEvent | 2019-08-22 19:20:16 | ualbertalib/avalon | https://api.github.com/repos/ualbertalib/avalon | closed | Item level begin/ends dates and auto-deletion | Post-launch enhancement priority:medium | ### Descriptive summary
ERA A+V is being used to host licensed videos where we agree to only host the videos for a specified period of time (e.g. 5 years) and then cut off access and delete the video once the license expires. It is currently not possible to make an item expire/auto-delete, so we're just managing this with calendar events/reminders at this point.
### Expected behaviour
Begin/end dates can be applied at the item level. Upon expiry, files are deleted automatically and access is cut off. It would also be good if there were some automated reporting on these actions (e.g. email warnings in advance of when an item is due for deletion, email confirmation once an item has been deleted).
### Actual behaviour
Expiry dates can be applied to specific users, groups and IP ranges, which is very useful, but doesn't apply to this situation:
 | 1.0 | Item level begin/ends dates and auto-deletion - ### Descriptive summary
ERA A+V is being used to host licensed videos where we agree to only host the videos for a specified period of time (e.g. 5 years) and then cut off access and delete the video once the license expires. It is currently not possible to make an item expire/auto-delete, so we're just managing this with calendar events/reminders at this point.
### Expected behaviour
Begin/end dates can be applied at the item level. Upon expiry, files are deleted automatically and access is cut off. It would also be good if there were some automated reporting on these actions (e.g. email warnings in advance of when an item is due for deletion, email confirmation once an item has been deleted).
### Actual behaviour
Expiry dates can be applied to specific users, groups and IP ranges, which is very useful, but doesn't apply to this situation:
 | priority | item level begin ends dates and auto deletion descriptive summary era a v is being used to host licensed videos where we agree to only host the videos for a specified period of time e g years and then cut off access and delete the video once the license expires it is currently not possible to make an item expire auto delete so we re just managing this with calendar events reminders at this point expected behaviour begin end dates can be applied at the item level upon expiry files are deleted automatically and access is cut off it would also be good if there were some automated reporting on these actions e g email warnings in advance of when an item is due for deletion email confirmation once an item has been deleted actual behaviour expiry dates can be applied to specific users groups and ip ranges which is very useful but doesn t apply to this situation | 1 |
712,799 | 24,507,636,226 | IssuesEvent | 2022-10-10 17:56:32 | RobertCraigie/prisma-client-py | https://api.github.com/repos/RobertCraigie/prisma-client-py | closed | Add official support for the windows platform | kind/feature level/intermediate priority/medium | ## Context
I haven't tested windows support in a while but apart from some internal changes required to make tests pass, the biggest blocker is https://github.com/prisma/prisma/issues/10270, however we can fix this ourselves.
| 1.0 | Add official support for the windows platform - ## Context
I haven't tested windows support in a while but apart from some internal changes required to make tests pass, the biggest blocker is https://github.com/prisma/prisma/issues/10270, however we can fix this ourselves.
| priority | add official support for the windows platform context i haven t tested windows support in a while but apart from some internal changes required to make tests pass the biggest blocker is however we can fix this ourselves | 1 |
58,681 | 3,090,481,382 | IssuesEvent | 2015-08-26 07:01:51 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Пропало назавание ADC хаба на вкладке | bug imported Priority-Medium | _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 28, 2014 09:56:27_
Есть один экзотический ADC-хаб на ассемблере.
С недавних пор пропало его название с вкладки и она стала размером чуть более кнопки закрывания вкладки.
Очень неудобно.
Также во флайлинке есть странный юзер без ника с CID AAAAA...
Наверное, это бот хаба, но вопрос с его отображением был недавно решен.
Видимо, не до конца решили.
**Attachment:** [Fly_x64_r17900_adc_hubname.png](http://code.google.com/p/flylinkdc/issues/detail?id=1543)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1543_ | 1.0 | Пропало назавание ADC хаба на вкладке - _From [mike.kor...@gmail.com](https://code.google.com/u/101495626515388303633/) on November 28, 2014 09:56:27_
Есть один экзотический ADC-хаб на ассемблере.
С недавних пор пропало его название с вкладки и она стала размером чуть более кнопки закрывания вкладки.
Очень неудобно.
Также во флайлинке есть странный юзер без ника с CID AAAAA...
Наверное, это бот хаба, но вопрос с его отображением был недавно решен.
Видимо, не до конца решили.
**Attachment:** [Fly_x64_r17900_adc_hubname.png](http://code.google.com/p/flylinkdc/issues/detail?id=1543)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1543_ | priority | пропало назавание adc хаба на вкладке from on november есть один экзотический adc хаб на ассемблере с недавних пор пропало его название с вкладки и она стала размером чуть более кнопки закрывания вкладки очень неудобно также во флайлинке есть странный юзер без ника с cid aaaaa наверное это бот хаба но вопрос с его отображением был недавно решен видимо не до конца решили attachment original issue | 1 |
282,246 | 8,704,757,085 | IssuesEvent | 2018-12-05 20:19:44 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | blocks behind minor details | Bug Priority: Medium | <img width="360" alt="screen shot 2018-12-05 at 9 19 23 am" src="https://user-images.githubusercontent.com/6775839/49531325-ed87ef80-f86e-11e8-9ab4-7c7604b344b5.png">
- need to round to 2 decimal places
- fix styling for mobile
- blocks behind number needs commas | 1.0 | blocks behind minor details - <img width="360" alt="screen shot 2018-12-05 at 9 19 23 am" src="https://user-images.githubusercontent.com/6775839/49531325-ed87ef80-f86e-11e8-9ab4-7c7604b344b5.png">
- need to round to 2 decimal places
- fix styling for mobile
- blocks behind number needs commas | priority | blocks behind minor details img width alt screen shot at am src need to round to decimal places fix styling for mobile blocks behind number needs commas | 1 |
470,110 | 13,531,191,027 | IssuesEvent | 2020-09-15 21:10:54 | argoproj/argo-cd | https://api.github.com/repos/argoproj/argo-cd | closed | Log output stops if message is too long | bug bug/priority:medium bug/severity:major component:ui type:scalability | Checklist:
* [x] I've searched in the docs and FAQ for my answer: http://bit.ly/argocd-faq.
* [x] I've included steps to reproduce the bug.
* [x] I've pasted the output of `argocd version`.
**Describe the bug**
When viewing logs of a container, if the output is really long (ie: 161,240 characters), output for logs stops indefinitely.
**To Reproduce**
1. Have a pod output a large string of text to stdout. Tailing logs should stop and should see the argocd-server start to output:
```
time="2020-03-20T10:01:01Z" level=warning msg="k8s pod logs scanner failed with error: bufio.Scanner: token too long" application=0xc000dc0740
```
**Expected behavior**
For the log message to be truncated so logs can continue to be streamed.
**Version**
```shell
v1.2.5
```
**Logs**
```
time="2020-03-20T10:01:01Z" level=warning msg="k8s pod logs scanner failed with error: bufio.Scanner: token too long" application=0xc000dc0740
```
| 1.0 | Log output stops if message is too long - Checklist:
* [x] I've searched in the docs and FAQ for my answer: http://bit.ly/argocd-faq.
* [x] I've included steps to reproduce the bug.
* [x] I've pasted the output of `argocd version`.
**Describe the bug**
When viewing logs of a container, if the output is really long (ie: 161,240 characters), output for logs stops indefinitely.
**To Reproduce**
1. Have a pod output a large string of text to stdout. Tailing logs should stop and should see the argocd-server start to output:
```
time="2020-03-20T10:01:01Z" level=warning msg="k8s pod logs scanner failed with error: bufio.Scanner: token too long" application=0xc000dc0740
```
**Expected behavior**
For the log message to be truncated so logs can continue to be streamed.
**Version**
```shell
v1.2.5
```
**Logs**
```
time="2020-03-20T10:01:01Z" level=warning msg="k8s pod logs scanner failed with error: bufio.Scanner: token too long" application=0xc000dc0740
```
| priority | log output stops if message is too long checklist i ve searched in the docs and faq for my answer i ve included steps to reproduce the bug i ve pasted the output of argocd version describe the bug when viewing logs of a container if the output is really long ie characters output for logs stops indefinitely to reproduce have a pod output a large string of text to stdout tailing logs should stop and should see the argocd server start to output time level warning msg pod logs scanner failed with error bufio scanner token too long application expected behavior for the log message to be truncated so logs can continue to be streamed version shell logs time level warning msg pod logs scanner failed with error bufio scanner token too long application | 1 |
681,209 | 23,300,935,999 | IssuesEvent | 2022-08-07 10:06:15 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | opened | When using GCR and mount the config.json - got: No credentials for gcr.io found | Type: Bug Priority: Medium Status: Available | <!--
Before submitting your issue, please make sure you're using the containrrr/watchtower:latest image.
If not, switch to this image prior to posting your report. Other forks, or the old `v2tec` image are **not** supported.
-->
**watchtower with GCP mount to config.json not recognize the file **
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
Steps to reproduce the behavior:
1. Setup GCP instance with Service account permissions to pull from gcr.io
2. run ```gcloud auth configure-docker```
3. Create docker-compose with
```watchtower:
image: containrrr/watchtower
restart: always
environment:
WATCHTOWER_NOTIFICATIONS: ${WATCHTOWER_NOTIFICATIONS}
WATCHTOWER_NOTIFICATION_SLACK_HOOK_URL: ${WATCHTOWER_NOTIFICATION_SLACK_HOOK_URL}
WATCHTOWER_NOTIFICATION_SLACK_IDENTIFIER: ${WATCHTOWER_NOTIFICATION_SLACK_IDENTIFIER}
WATCHTOWER_NOTIFICATION_SLACK_CHANNEL: ${WATCHTOWER_NOTIFICATION_SLACK_CHANNEL}
HOME: /root
GCR: "yes"
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- $HOME/.docker/config.json:/root/.docker/config.json
- $HOME/.config/gcloud:/root/.config/gcloud
command: --cleanup --debug --interval 15
labels:
- com.centurylinklabs.watchtower.enable=false
```
4. watch docker logs
```
time="2022-08-07T09:58:50Z" level=debug
time="2022-08-07T09:58:50Z" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
time="2022-08-07T09:58:51Z" level=debug msg="Making sure everything is sane before starting"
time="2022-08-07T09:58:51Z" level=debug msg="Retrieving running containers"
time="2022-08-07T09:58:51Z" level=debug msg="There are no additional watchtower containers"
time="2022-08-07T09:58:51Z" level=debug msg="Watchtower HTTP API skipped."
time="2022-08-07T09:58:51Z" level=info msg="Watchtower 1.4.0"
time="2022-08-07T09:58:51Z" level=info msg="Using notifications: slack"
time="2022-08-07T09:58:51Z" level=info msg="Checking all containers (except explicitly disabled with label)"
time="2022-08-07T09:58:51Z" level=info msg="Scheduling first run: 2022-08-07 09:59:06 +0000 UTC"
time="2022-08-07T09:58:51Z" level=info msg="Note that the first check will be performed in 14 seconds"
time="2022-08-07T09:59:06Z" level=debug msg="Checking containers for updated images"
time="2022-08-07T09:59:06Z" level=debug msg="Retrieving running containers"
time="2022-08-07T09:59:06Z" level=debug msg="Trying to load authentication credentials." container=/antidote-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="No credentials for gcr.io found" config_file=/config.json
time="2022-08-07T09:59:06Z" level=debug msg="Got image name: gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Checking if pull is needed" container=/image-queries-1 image="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Building challenge URL" URL="https://gcr.io/v2/"
time="2022-08-07T09:59:06Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://gcr.io/v2/token\",service=\"gcr.io\"" status="401 Unauthorized"
time="2022-08-07T09:59:06Z" level=debug msg="Checking challenge header content" realm="https://gcr.io/v2/token" service=gcr.io
time="2022-08-07T09:59:06Z" level=debug msg="Setting scope for auth token" image=gcr.io/xxx-xxx-xxx/beyond-queries scope="repository:gcr.io/xxx-xxx-xxx/beyond-queries:pull"
time="2022-08-07T09:59:06Z" level=debug msg="No credentials found."
time="2022-08-07T09:59:06Z" level=debug msg="Parsing image ref" host=gcr.io image=xxx-xxx-xxx/beyond-queries normalized="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest" tag=staging-latest
time="2022-08-07T09:59:06Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://gcr.io/v2/xxx-xxx-xxx/beyond-queries/manifests/staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Could not do a head request for \"gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest\", falling back to regular pull." container=/image-queries-1 image="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Reason: registry responded to head request with \"401 Unauthorized\", auth: \"Bearer realm=\\\"https://gcr.io/v2/token\\\",service=\\\"gcr.io\\\",scope=\\\"repository:xxx-xxx-xxx/beyond-queries:pull\\\"\"" container=/image-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Pulling image" container=/antidote-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Error pulling image gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest, Error response from daemon: unauthorized: You don't have the needed permissions to perform this operation, and you may have invalid credentials. To authenticate your request, follow the steps in: https://cloud.google.com/container-registry/docs/advanced-authentication"
time="2022-08-07T09:59:06Z" level=info msg="Unable to update container \"/image-queries-1\": Error response from daemon: unauthorized: You don't have the needed permissions to perform this operation, and you may have invalid credentials. To authenticate your request, follow the steps in: https://cloud.google.com/container-registry/docs/advanced-authentication. Proceeding to next."
```
**Expected behavior**
should be pull and restart as expected in watchtower documentation
**Screenshots**
Do not have special screenshots
**Environment**
We want to know:
- Google Cloud Platform
- Goolge instance virtual machine, ubuntu 20.04
- Docker version:
```Client: Docker Engine - Community
Version: 20.10.17
API version: 1.41
Go version: go1.17.11
Git commit: 100c701
Built: Mon Jun 6 23:02:57 2022
OS/Arch: linux/amd64
Context: default
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 20.10.17
API version: 1.41 (minimum version 1.12)
Go version: go1.17.11
Git commit: a89b842
Built: Mon Jun 6 23:01:03 2022
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: 1.6.6
GitCommit: 10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1
runc:
Version: 1.1.2
GitCommit: v1.1.2-0-ga916309
docker-init:
Version: 0.19.0
GitCommit: de40ad0```
What am i missing here? | 1.0 | When using GCR and mount the config.json - got: No credentials for gcr.io found - <!--
Before submitting your issue, please make sure you're using the containrrr/watchtower:latest image.
If not, switch to this image prior to posting your report. Other forks, or the old `v2tec` image are **not** supported.
-->
**watchtower with GCP mount to config.json not recognize the file **
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
Steps to reproduce the behavior:
1. Setup GCP instance with Service account permissions to pull from gcr.io
2. run ```gcloud auth configure-docker```
3. Create docker-compose with
```watchtower:
image: containrrr/watchtower
restart: always
environment:
WATCHTOWER_NOTIFICATIONS: ${WATCHTOWER_NOTIFICATIONS}
WATCHTOWER_NOTIFICATION_SLACK_HOOK_URL: ${WATCHTOWER_NOTIFICATION_SLACK_HOOK_URL}
WATCHTOWER_NOTIFICATION_SLACK_IDENTIFIER: ${WATCHTOWER_NOTIFICATION_SLACK_IDENTIFIER}
WATCHTOWER_NOTIFICATION_SLACK_CHANNEL: ${WATCHTOWER_NOTIFICATION_SLACK_CHANNEL}
HOME: /root
GCR: "yes"
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- $HOME/.docker/config.json:/root/.docker/config.json
- $HOME/.config/gcloud:/root/.config/gcloud
command: --cleanup --debug --interval 15
labels:
- com.centurylinklabs.watchtower.enable=false
```
4. watch docker logs
```
time="2022-08-07T09:58:50Z" level=debug
time="2022-08-07T09:58:50Z" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
time="2022-08-07T09:58:51Z" level=debug msg="Making sure everything is sane before starting"
time="2022-08-07T09:58:51Z" level=debug msg="Retrieving running containers"
time="2022-08-07T09:58:51Z" level=debug msg="There are no additional watchtower containers"
time="2022-08-07T09:58:51Z" level=debug msg="Watchtower HTTP API skipped."
time="2022-08-07T09:58:51Z" level=info msg="Watchtower 1.4.0"
time="2022-08-07T09:58:51Z" level=info msg="Using notifications: slack"
time="2022-08-07T09:58:51Z" level=info msg="Checking all containers (except explicitly disabled with label)"
time="2022-08-07T09:58:51Z" level=info msg="Scheduling first run: 2022-08-07 09:59:06 +0000 UTC"
time="2022-08-07T09:58:51Z" level=info msg="Note that the first check will be performed in 14 seconds"
time="2022-08-07T09:59:06Z" level=debug msg="Checking containers for updated images"
time="2022-08-07T09:59:06Z" level=debug msg="Retrieving running containers"
time="2022-08-07T09:59:06Z" level=debug msg="Trying to load authentication credentials." container=/antidote-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="No credentials for gcr.io found" config_file=/config.json
time="2022-08-07T09:59:06Z" level=debug msg="Got image name: gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Checking if pull is needed" container=/image-queries-1 image="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Building challenge URL" URL="https://gcr.io/v2/"
time="2022-08-07T09:59:06Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://gcr.io/v2/token\",service=\"gcr.io\"" status="401 Unauthorized"
time="2022-08-07T09:59:06Z" level=debug msg="Checking challenge header content" realm="https://gcr.io/v2/token" service=gcr.io
time="2022-08-07T09:59:06Z" level=debug msg="Setting scope for auth token" image=gcr.io/xxx-xxx-xxx/beyond-queries scope="repository:gcr.io/xxx-xxx-xxx/beyond-queries:pull"
time="2022-08-07T09:59:06Z" level=debug msg="No credentials found."
time="2022-08-07T09:59:06Z" level=debug msg="Parsing image ref" host=gcr.io image=xxx-xxx-xxx/beyond-queries normalized="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest" tag=staging-latest
time="2022-08-07T09:59:06Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://gcr.io/v2/xxx-xxx-xxx/beyond-queries/manifests/staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Could not do a head request for \"gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest\", falling back to regular pull." container=/image-queries-1 image="gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Reason: registry responded to head request with \"401 Unauthorized\", auth: \"Bearer realm=\\\"https://gcr.io/v2/token\\\",service=\\\"gcr.io\\\",scope=\\\"repository:xxx-xxx-xxx/beyond-queries:pull\\\"\"" container=/image-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Pulling image" container=/antidote-queries-1 image="gcr.io/med-and-beyond/beyond-queries:staging-latest"
time="2022-08-07T09:59:06Z" level=debug msg="Error pulling image gcr.io/xxx-xxx-xxx/beyond-queries:staging-latest, Error response from daemon: unauthorized: You don't have the needed permissions to perform this operation, and you may have invalid credentials. To authenticate your request, follow the steps in: https://cloud.google.com/container-registry/docs/advanced-authentication"
time="2022-08-07T09:59:06Z" level=info msg="Unable to update container \"/image-queries-1\": Error response from daemon: unauthorized: You don't have the needed permissions to perform this operation, and you may have invalid credentials. To authenticate your request, follow the steps in: https://cloud.google.com/container-registry/docs/advanced-authentication. Proceeding to next."
```
**Expected behavior**
should be pull and restart as expected in watchtower documentation
**Screenshots**
Do not have special screenshots
**Environment**
We want to know:
- Google Cloud Platform
- Goolge instance virtual machine, ubuntu 20.04
- Docker version:
```Client: Docker Engine - Community
Version: 20.10.17
API version: 1.41
Go version: go1.17.11
Git commit: 100c701
Built: Mon Jun 6 23:02:57 2022
OS/Arch: linux/amd64
Context: default
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 20.10.17
API version: 1.41 (minimum version 1.12)
Go version: go1.17.11
Git commit: a89b842
Built: Mon Jun 6 23:01:03 2022
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: 1.6.6
GitCommit: 10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1
runc:
Version: 1.1.2
GitCommit: v1.1.2-0-ga916309
docker-init:
Version: 0.19.0
GitCommit: de40ad0```
What am i missing here? | priority | when using gcr and mount the config json got no credentials for gcr io found before submitting your issue please make sure you re using the containrrr watchtower latest image if not switch to this image prior to posting your report other forks or the old image are not supported watchtower with gcp mount to config json not recognize the file to reproduce steps to reproduce the behavior setup gcp instance with service account permissions to pull from gcr io run gcloud auth configure docker create docker compose with watchtower image containrrr watchtower restart always environment watchtower notifications watchtower notifications watchtower notification slack hook url watchtower notification slack hook url watchtower notification slack identifier watchtower notification slack identifier watchtower notification slack channel watchtower notification slack channel home root gcr yes volumes var run docker sock var run docker sock home docker config json root docker config json home config gcloud root config gcloud command cleanup debug interval labels com centurylinklabs watchtower enable false watch docker logs time level debug time level debug msg sleeping for a second to ensure the docker api client has been properly initialized time level debug msg making sure everything is sane before starting time level debug msg retrieving running containers time level debug msg there are no additional watchtower containers time level debug msg watchtower http api skipped time level info msg watchtower time level info msg using notifications slack time level info msg checking all containers except explicitly disabled with label time level info msg scheduling first run utc time level info msg note that the first check will be performed in seconds time level debug msg checking containers for updated images time level debug msg retrieving running containers time level debug msg trying to load authentication credentials container antidote queries image gcr io med and beyond beyond queries staging latest time level debug msg no credentials for gcr io found config file config json time level debug msg got image name gcr io xxx xxx xxx beyond queries staging latest time level debug msg checking if pull is needed container image queries image gcr io xxx xxx xxx beyond queries staging latest time level debug msg building challenge url url time level debug msg got response to challenge request header bearer realm status unauthorized time level debug msg checking challenge header content realm service gcr io time level debug msg setting scope for auth token image gcr io xxx xxx xxx beyond queries scope repository gcr io xxx xxx xxx beyond queries pull time level debug msg no credentials found time level debug msg parsing image ref host gcr io image xxx xxx xxx beyond queries normalized gcr io xxx xxx xxx beyond queries staging latest tag staging latest time level debug msg doing a head request to fetch a digest url time level debug msg could not do a head request for gcr io xxx xxx xxx beyond queries staging latest falling back to regular pull container image queries image gcr io xxx xxx xxx beyond queries staging latest time level debug msg reason registry responded to head request with unauthorized auth bearer realm container image queries image gcr io med and beyond beyond queries staging latest time level debug msg pulling image container antidote queries image gcr io med and beyond beyond queries staging latest time level debug msg error pulling image gcr io xxx xxx xxx beyond queries staging latest error response from daemon unauthorized you don t have the needed permissions to perform this operation and you may have invalid credentials to authenticate your request follow the steps in time level info msg unable to update container image queries error response from daemon unauthorized you don t have the needed permissions to perform this operation and you may have invalid credentials to authenticate your request follow the steps in proceeding to next expected behavior should be pull and restart as expected in watchtower documentation screenshots do not have special screenshots environment we want to know google cloud platform goolge instance virtual machine ubuntu docker version client docker engine community version api version go version git commit built mon jun os arch linux context default experimental true server docker engine community engine version api version minimum version go version git commit built mon jun os arch linux experimental false containerd version gitcommit runc version gitcommit docker init version gitcommit what am i missing here | 1 |
57,036 | 3,081,233,146 | IssuesEvent | 2015-08-22 14:21:22 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Add build time to F7 screen | 019a bug imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 16:42:52_
Testing with other users would be much easier if we could tell which version players were running. An easy way would be to have the build time baked into the binary.
Make it so.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=267_ | 1.0 | Add build time to F7 screen - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 16:42:52_
Testing with other users would be much easier if we could tell which version players were running. An easy way would be to have the build time baked into the binary.
Make it so.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=267_ | priority | add build time to screen from on october testing with other users would be much easier if we could tell which version players were running an easy way would be to have the build time baked into the binary make it so original issue | 1 |
427,971 | 12,401,425,189 | IssuesEvent | 2020-05-21 09:50:53 | vdaas/vald | https://api.github.com/repos/vdaas/vald | opened | The SearchByVectorID result may not accurate | priority/medium team/core type/bug | ### Describe the bug:
After Vald receive the SearchByVectorID request, Vald LB Gateway will forward the UUID to the Vald Agent to perform the search.
For example we have 4 Vald Agents:
| Agent | Memory |
| 1 | V1, V2 |
| 2 | V1, V3 |
| 3 | V2, V4 |
| 4 | V4, V3 |
When user perform the searchByVectorID where the vector ID matches the vector `V2`, only Agent 1 and 3 can return the result, but the most accurate result maybe stored in Agent 2 or Agent 4.
<!-- A clear and concise description of what the bug is. -->
### To Reproduce:
<!-- Please describe the steps to reproduce the behavior: -->
### Expected behavior:
<!-- A clear and concise description of what you expected to happen. -->
All the Vald Agent should be able to perform the search.
### Environment:
<!--- Please change the versions below along with your environment -->
- Golang Version: 1.13.5
- Docker Version: 19.03.5
- Kubernetes Version: 1.16.3
- NGT Version: 1.8.4
| 1.0 | The SearchByVectorID result may not accurate - ### Describe the bug:
After Vald receive the SearchByVectorID request, Vald LB Gateway will forward the UUID to the Vald Agent to perform the search.
For example we have 4 Vald Agents:
| Agent | Memory |
| 1 | V1, V2 |
| 2 | V1, V3 |
| 3 | V2, V4 |
| 4 | V4, V3 |
When user perform the searchByVectorID where the vector ID matches the vector `V2`, only Agent 1 and 3 can return the result, but the most accurate result maybe stored in Agent 2 or Agent 4.
<!-- A clear and concise description of what the bug is. -->
### To Reproduce:
<!-- Please describe the steps to reproduce the behavior: -->
### Expected behavior:
<!-- A clear and concise description of what you expected to happen. -->
All the Vald Agent should be able to perform the search.
### Environment:
<!--- Please change the versions below along with your environment -->
- Golang Version: 1.13.5
- Docker Version: 19.03.5
- Kubernetes Version: 1.16.3
- NGT Version: 1.8.4
| priority | the searchbyvectorid result may not accurate describe the bug after vald receive the searchbyvectorid request vald lb gateway will forward the uuid to the vald agent to perform the search for example we have vald agents agent memory when user perform the searchbyvectorid where the vector id matches the vector only agent and can return the result but the most accurate result maybe stored in agent or agent to reproduce expected behavior all the vald agent should be able to perform the search environment golang version docker version kubernetes version ngt version | 1 |
194,002 | 6,890,468,027 | IssuesEvent | 2017-11-22 14:07:45 | kedgeproject/kedge | https://api.github.com/repos/kedgeproject/kedge | closed | portMappings errors with `80`, success with `80:80` | kind/bug priority/medium size/M | Looks to be an int / string parsing issue.
To replicate:
```yaml
name: wordpress
containers:
- image: wordpress:4
health:
httpGet:
path: /
port: 80
initialDelaySeconds: 20
timeoutSeconds: 5
env:
- name: WORDPRESS_DB_NAME
valueFrom:
configMapKeyRef:
name: database
key: MYSQL_DATABASE
- name: WORDPRESS_DB_USER
valueFrom:
configMapKeyRef:
name: database
key: MYSQL_USER
- name: WORDPRESS_DB_PASSWORD
valueFrom:
secretKeyRef:
name: database-user-password
key: MYSQL_PASSWORD
- name: WORDPRESS_DB_HOST
value: database:3306
services:
- name: wordpress
type: LoadBalancer
portMappings:
- 80
```
```sh
▶ ./kedge generate -f examples/wordpress/mariadb.yaml
unable to perform controller operations: unable to unmarshal data: could not unmarshal into internal struct: error unmarshaling JSON: json: cannot unmarshal number into Go struct field ServiceSpecMod.portMappings of type string
```
Work-around is to use:
```yaml
services:
- name: wordpress
type: LoadBalancer
portMappings:
- 80:80
``` | 1.0 | portMappings errors with `80`, success with `80:80` - Looks to be an int / string parsing issue.
To replicate:
```yaml
name: wordpress
containers:
- image: wordpress:4
health:
httpGet:
path: /
port: 80
initialDelaySeconds: 20
timeoutSeconds: 5
env:
- name: WORDPRESS_DB_NAME
valueFrom:
configMapKeyRef:
name: database
key: MYSQL_DATABASE
- name: WORDPRESS_DB_USER
valueFrom:
configMapKeyRef:
name: database
key: MYSQL_USER
- name: WORDPRESS_DB_PASSWORD
valueFrom:
secretKeyRef:
name: database-user-password
key: MYSQL_PASSWORD
- name: WORDPRESS_DB_HOST
value: database:3306
services:
- name: wordpress
type: LoadBalancer
portMappings:
- 80
```
```sh
▶ ./kedge generate -f examples/wordpress/mariadb.yaml
unable to perform controller operations: unable to unmarshal data: could not unmarshal into internal struct: error unmarshaling JSON: json: cannot unmarshal number into Go struct field ServiceSpecMod.portMappings of type string
```
Work-around is to use:
```yaml
services:
- name: wordpress
type: LoadBalancer
portMappings:
- 80:80
``` | priority | portmappings errors with success with looks to be an int string parsing issue to replicate yaml name wordpress containers image wordpress health httpget path port initialdelayseconds timeoutseconds env name wordpress db name valuefrom configmapkeyref name database key mysql database name wordpress db user valuefrom configmapkeyref name database key mysql user name wordpress db password valuefrom secretkeyref name database user password key mysql password name wordpress db host value database services name wordpress type loadbalancer portmappings sh ▶ kedge generate f examples wordpress mariadb yaml unable to perform controller operations unable to unmarshal data could not unmarshal into internal struct error unmarshaling json json cannot unmarshal number into go struct field servicespecmod portmappings of type string work around is to use yaml services name wordpress type loadbalancer portmappings | 1 |
155,135 | 5,949,590,693 | IssuesEvent | 2017-05-26 14:39:21 | mkdo/kapow-theme | https://api.github.com/repos/mkdo/kapow-theme | closed | retina image issue | Priority: Medium Status: Completed Type: Bug | On this site
https://dbhweb.wpengine.com/
The retina version of the image is not loading because
```
Mixed Content:
The page at 'https://dbhweb.wpengine.com/' was loaded over HTTPS,
but requested an insecure image 'http://dbhweb.wpengine.com/wp-content/uploads/2017/05/dbh-logox2.png'.
This request has been blocked; the content must be served over HTTPS.
``` | 1.0 | retina image issue - On this site
https://dbhweb.wpengine.com/
The retina version of the image is not loading because
```
Mixed Content:
The page at 'https://dbhweb.wpengine.com/' was loaded over HTTPS,
but requested an insecure image 'http://dbhweb.wpengine.com/wp-content/uploads/2017/05/dbh-logox2.png'.
This request has been blocked; the content must be served over HTTPS.
``` | priority | retina image issue on this site the retina version of the image is not loading because mixed content the page at was loaded over https but requested an insecure image this request has been blocked the content must be served over https | 1 |
639,822 | 20,766,855,286 | IssuesEvent | 2022-03-15 21:36:28 | magma/magma | https://api.github.com/repos/magma/magma | closed | [AMF]: Authentication reject on getting authentication failure from UE with different causes | type: bug wontfix component: agw priority: medium product: 5g sa | This is reference to testcase TC43d and TC43e. As part of Authentication Procedure UE is sending Authentication failure with Error as UE Security Capability Mismatch and Security Mode failures.
Although this is related to security mode but as its Authentication Failure so AMF has to respond with Authentication Reject (with the logs saying its not un-related error).
Reference : 3GPP TS 24.501 version 15.6.0 Release 15
5.4.1.2 EAP based primary authentication and key agreement procedure
General section : It states :
If the authentication of the UE completes unsuccessfully, the EAP-failure message is transported from the network to
the UE using the AUTHENTICATION RESULT message or the AUTHENTICATION REJECT message of the EAP
result message transport procedure or in a response of the initial 5GMM procedure as part of which the EAP based
primary authentication and key agreement procedure is performed | 1.0 | [AMF]: Authentication reject on getting authentication failure from UE with different causes - This is reference to testcase TC43d and TC43e. As part of Authentication Procedure UE is sending Authentication failure with Error as UE Security Capability Mismatch and Security Mode failures.
Although this is related to security mode but as its Authentication Failure so AMF has to respond with Authentication Reject (with the logs saying its not un-related error).
Reference : 3GPP TS 24.501 version 15.6.0 Release 15
5.4.1.2 EAP based primary authentication and key agreement procedure
General section : It states :
If the authentication of the UE completes unsuccessfully, the EAP-failure message is transported from the network to
the UE using the AUTHENTICATION RESULT message or the AUTHENTICATION REJECT message of the EAP
result message transport procedure or in a response of the initial 5GMM procedure as part of which the EAP based
primary authentication and key agreement procedure is performed | priority | authentication reject on getting authentication failure from ue with different causes this is reference to testcase and as part of authentication procedure ue is sending authentication failure with error as ue security capability mismatch and security mode failures although this is related to security mode but as its authentication failure so amf has to respond with authentication reject with the logs saying its not un related error reference ts version release eap based primary authentication and key agreement procedure general section it states if the authentication of the ue completes unsuccessfully the eap failure message is transported from the network to the ue using the authentication result message or the authentication reject message of the eap result message transport procedure or in a response of the initial procedure as part of which the eap based primary authentication and key agreement procedure is performed | 1 |
25,944 | 2,684,054,444 | IssuesEvent | 2015-03-28 16:19:49 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | Иногда при смене режима реальная консоль создается с неправильной высотой | 1 star bug imported Priority-Medium | _From [thecybershadow](https://code.google.com/u/thecybershadow/) on March 15, 2012 16:16:35_
OS version: Windows Server 2008 x64 ConEmu version: 120315e x86
Far version: 3.0 (build 2546) x86
Похоже на Issue 82 , но:
1) это происходит не при открытии, а при смене экранного режима
2) Alt+F9 в реальной консоли не помогает
При смене разрешения не воспроизводится, только при смене ориентации. *Steps to reproduction* 1. http://noeld.com/programs.asp#display 2. https://gist.github.com/2047520 3. Press Enter until reproduced
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=509_ | 1.0 | Иногда при смене режима реальная консоль создается с неправильной высотой - _From [thecybershadow](https://code.google.com/u/thecybershadow/) on March 15, 2012 16:16:35_
OS version: Windows Server 2008 x64 ConEmu version: 120315e x86
Far version: 3.0 (build 2546) x86
Похоже на Issue 82 , но:
1) это происходит не при открытии, а при смене экранного режима
2) Alt+F9 в реальной консоли не помогает
При смене разрешения не воспроизводится, только при смене ориентации. *Steps to reproduction* 1. http://noeld.com/programs.asp#display 2. https://gist.github.com/2047520 3. Press Enter until reproduced
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=509_ | priority | иногда при смене режима реальная консоль создается с неправильной высотой from on march os version windows server conemu version far version build похоже на issue но это происходит не при открытии а при смене экранного режима alt в реальной консоли не помогает при смене разрешения не воспроизводится только при смене ориентации steps to reproduction press enter until reproduced original issue | 1 |
593,626 | 18,012,461,178 | IssuesEvent | 2021-09-16 10:09:49 | roq-trading/roq-issues | https://api.github.com/repos/roq-trading/roq-issues | opened | [roq-fix-bridge] How to deal with the TIMEOUT OrderAck event in the FIX state change matrix | bug medium priority support | This is unclear because TIMEOUT is not a final request state.
The immediate thought is to simply suppress this update and wait for the gateway to send subsequent final request states (ACCEPTED or REJECTED).
This is a scenario that can happen for many gateways where order actions are sent on one connection and order updates may arrive on another connection. A timeout (or disconnect) could occur on the order action channel and the order update could arrive immediately after. There's also the scenario where the order request is rejected and an order update may therefore never arrive on the other channel. For these scenarios a download must be requested (another request with potential for timeout/disconnect) before one can confirm it's rejected.
This is related to
* #38
* #39
| 1.0 | [roq-fix-bridge] How to deal with the TIMEOUT OrderAck event in the FIX state change matrix - This is unclear because TIMEOUT is not a final request state.
The immediate thought is to simply suppress this update and wait for the gateway to send subsequent final request states (ACCEPTED or REJECTED).
This is a scenario that can happen for many gateways where order actions are sent on one connection and order updates may arrive on another connection. A timeout (or disconnect) could occur on the order action channel and the order update could arrive immediately after. There's also the scenario where the order request is rejected and an order update may therefore never arrive on the other channel. For these scenarios a download must be requested (another request with potential for timeout/disconnect) before one can confirm it's rejected.
This is related to
* #38
* #39
| priority | how to deal with the timeout orderack event in the fix state change matrix this is unclear because timeout is not a final request state the immediate thought is to simply suppress this update and wait for the gateway to send subsequent final request states accepted or rejected this is a scenario that can happen for many gateways where order actions are sent on one connection and order updates may arrive on another connection a timeout or disconnect could occur on the order action channel and the order update could arrive immediately after there s also the scenario where the order request is rejected and an order update may therefore never arrive on the other channel for these scenarios a download must be requested another request with potential for timeout disconnect before one can confirm it s rejected this is related to | 1 |
148,839 | 5,697,155,243 | IssuesEvent | 2017-04-16 19:03:13 | ESAPI/esapi-java-legacy | https://api.github.com/repos/ESAPI/esapi-java-legacy | closed | Update ESAPI's pom.xml to address vulnerable 3rd party components | Configuration Priority-Medium | Reported by Mateusz Krzeszowiec via private email.
Hello Kevin, Chris,
While reviewing dependencies of ESAPI I've noticed that it uses number of vulnerable components, both through "direct" dependencies (via pom.xml) and transitive ones.
Printout from dependency checker goes as follows:
commons-beanutils-core-1.8.3.jar
(commons-beanutils:commons-beanutils-core:1.8.3,
cpe:/a:apache:commons_beanutils:1.8.3) : CVE-2014-0114
commons-fileupload-1.3.1.jar
(commons-fileupload:commons-fileupload:1.3.1,
cpe:/a:apache:commons_fileupload:1.3.1) : CVE-2016-3092
commons-httpclient-3.1.jar (commons-httpclient:commons-httpclient:3.1,
cpe:/a:apache:commons-httpclient:3.1, cpe:/a:apache:httpclient:3.1) :
CVE-2015-5262, CVE-2014-3577, CVE-2012-6153
xalan-2.7.0.jar (cpe:/a:apache:xalan-java:2.7.0, xalan:xalan:2.7.0) :
CVE-2014-0107
Is it something that you might consider addressing, even though ESAPI 2 is considered legacy? I'd be more then happy to help with fixing those issues.
I'd appreciate if you could review the findings, for convenience I'm attaching dependency tree for ESAPI as well as some initial review of the issues reported through dependencies.
xalan: I did some digging for xalan one already, it's included via xom which doesn't seem to be actively maintained, still in that case I'd say it might be doable to fix the xom dependency. One option could be to move away from using xom as it seem to be ancient thing. antisamy: For antisamy (that brings in httpclient) it'd be slightly more difficult I believe as it's not being maintained for 3 years if I'm not mistaken.
commons-beanutils-core - at this moment I'm not yet sure if the CVE reported actually applies to common-beanutils-core (it does according to dependency checker but I haven't verified that) - in this case it should be as easy as changing ESAPI pom.xml commons-fileupload: it's fairly new vulnerability, again should be as
easy as changing pom.xml
---
Note: In addition to addressing these issues, we should also update to latest version of OWASP Dependency Check, which as of this writing, is 1.4.0. --kevin
| 1.0 | Update ESAPI's pom.xml to address vulnerable 3rd party components - Reported by Mateusz Krzeszowiec via private email.
Hello Kevin, Chris,
While reviewing dependencies of ESAPI I've noticed that it uses number of vulnerable components, both through "direct" dependencies (via pom.xml) and transitive ones.
Printout from dependency checker goes as follows:
commons-beanutils-core-1.8.3.jar
(commons-beanutils:commons-beanutils-core:1.8.3,
cpe:/a:apache:commons_beanutils:1.8.3) : CVE-2014-0114
commons-fileupload-1.3.1.jar
(commons-fileupload:commons-fileupload:1.3.1,
cpe:/a:apache:commons_fileupload:1.3.1) : CVE-2016-3092
commons-httpclient-3.1.jar (commons-httpclient:commons-httpclient:3.1,
cpe:/a:apache:commons-httpclient:3.1, cpe:/a:apache:httpclient:3.1) :
CVE-2015-5262, CVE-2014-3577, CVE-2012-6153
xalan-2.7.0.jar (cpe:/a:apache:xalan-java:2.7.0, xalan:xalan:2.7.0) :
CVE-2014-0107
Is it something that you might consider addressing, even though ESAPI 2 is considered legacy? I'd be more then happy to help with fixing those issues.
I'd appreciate if you could review the findings, for convenience I'm attaching dependency tree for ESAPI as well as some initial review of the issues reported through dependencies.
xalan: I did some digging for xalan one already, it's included via xom which doesn't seem to be actively maintained, still in that case I'd say it might be doable to fix the xom dependency. One option could be to move away from using xom as it seem to be ancient thing. antisamy: For antisamy (that brings in httpclient) it'd be slightly more difficult I believe as it's not being maintained for 3 years if I'm not mistaken.
commons-beanutils-core - at this moment I'm not yet sure if the CVE reported actually applies to common-beanutils-core (it does according to dependency checker but I haven't verified that) - in this case it should be as easy as changing ESAPI pom.xml commons-fileupload: it's fairly new vulnerability, again should be as
easy as changing pom.xml
---
Note: In addition to addressing these issues, we should also update to latest version of OWASP Dependency Check, which as of this writing, is 1.4.0. --kevin
| priority | update esapi s pom xml to address vulnerable party components reported by mateusz krzeszowiec via private email hello kevin chris while reviewing dependencies of esapi i ve noticed that it uses number of vulnerable components both through direct dependencies via pom xml and transitive ones printout from dependency checker goes as follows commons beanutils core jar commons beanutils commons beanutils core cpe a apache commons beanutils cve commons fileupload jar commons fileupload commons fileupload cpe a apache commons fileupload cve commons httpclient jar commons httpclient commons httpclient cpe a apache commons httpclient cpe a apache httpclient cve cve cve xalan jar cpe a apache xalan java xalan xalan cve is it something that you might consider addressing even though esapi is considered legacy i d be more then happy to help with fixing those issues i d appreciate if you could review the findings for convenience i m attaching dependency tree for esapi as well as some initial review of the issues reported through dependencies xalan i did some digging for xalan one already it s included via xom which doesn t seem to be actively maintained still in that case i d say it might be doable to fix the xom dependency one option could be to move away from using xom as it seem to be ancient thing antisamy for antisamy that brings in httpclient it d be slightly more difficult i believe as it s not being maintained for years if i m not mistaken commons beanutils core at this moment i m not yet sure if the cve reported actually applies to common beanutils core it does according to dependency checker but i haven t verified that in this case it should be as easy as changing esapi pom xml commons fileupload it s fairly new vulnerability again should be as easy as changing pom xml note in addition to addressing these issues we should also update to latest version of owasp dependency check which as of this writing is kevin | 1 |
52,287 | 3,022,480,560 | IssuesEvent | 2015-07-31 20:37:03 | information-artifact-ontology/IAO | https://api.github.com/repos/information-artifact-ontology/IAO | opened | subject agrees they understand informed consent document | imported Priority-Medium Type-Term | _From [mcour...@gmail.com](https://code.google.com/u/116795168307825520406/) on November 04, 2009 18:17:41_
term currently in OBI with note "09/28/2009 Alan Ruttenberg. There's a need
for a general process like this in IAO - document and person in, signed
document (and associated obligations, rights, out"
_Original issue: http://code.google.com/p/information-artifact-ontology/issues/detail?id=69_ | 1.0 | subject agrees they understand informed consent document - _From [mcour...@gmail.com](https://code.google.com/u/116795168307825520406/) on November 04, 2009 18:17:41_
term currently in OBI with note "09/28/2009 Alan Ruttenberg. There's a need
for a general process like this in IAO - document and person in, signed
document (and associated obligations, rights, out"
_Original issue: http://code.google.com/p/information-artifact-ontology/issues/detail?id=69_ | priority | subject agrees they understand informed consent document from on november term currently in obi with note alan ruttenberg there s a need for a general process like this in iao document and person in signed document and associated obligations rights out original issue | 1 |
317,354 | 9,663,941,600 | IssuesEvent | 2019-05-21 03:04:10 | uvicgse/project-virtual-team-3 | https://api.github.com/repos/uvicgse/project-virtual-team-3 | opened | Add tagging support | High Priority Medium Sprint 1 enhancement | **Feature request:**
Add support for tagging commits in the past and/or the current commit that has yet to be pushed. This will allow users to mark important points in their repository for example "v1.0" without needing to search for a certain commit message/checksum. Users should also be able to view the tags on the main page | 1.0 | Add tagging support - **Feature request:**
Add support for tagging commits in the past and/or the current commit that has yet to be pushed. This will allow users to mark important points in their repository for example "v1.0" without needing to search for a certain commit message/checksum. Users should also be able to view the tags on the main page | priority | add tagging support feature request add support for tagging commits in the past and or the current commit that has yet to be pushed this will allow users to mark important points in their repository for example without needing to search for a certain commit message checksum users should also be able to view the tags on the main page | 1 |
340,787 | 10,278,516,937 | IssuesEvent | 2019-08-25 15:03:35 | x13pixels/remedybg-issues | https://api.github.com/repos/x13pixels/remedybg-issues | closed | Add support for /ZI | Component: Engine Priority: 6 (Medium) Status: Completed Type: Enhancement | Related to #29. Reported: "Global symbols seem to work but not local/stack symbols". | 1.0 | Add support for /ZI - Related to #29. Reported: "Global symbols seem to work but not local/stack symbols". | priority | add support for zi related to reported global symbols seem to work but not local stack symbols | 1 |
328,858 | 10,000,857,404 | IssuesEvent | 2019-07-12 14:20:34 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | opened | [studio][studio-ui] Implement type-ahead services | enhancement priority: medium | Backend:
Implement type-ahead APIs for:
- Users
- Groups
- Organizations (placeholder)
- Sites
Submit the API definition first for review, and then we start implementation. Assign to UI once the spec is approved.
Front-end:
Use the above services to implement type-ahead across the UI for the above use-cases. | 1.0 | [studio][studio-ui] Implement type-ahead services - Backend:
Implement type-ahead APIs for:
- Users
- Groups
- Organizations (placeholder)
- Sites
Submit the API definition first for review, and then we start implementation. Assign to UI once the spec is approved.
Front-end:
Use the above services to implement type-ahead across the UI for the above use-cases. | priority | implement type ahead services backend implement type ahead apis for users groups organizations placeholder sites submit the api definition first for review and then we start implementation assign to ui once the spec is approved front end use the above services to implement type ahead across the ui for the above use cases | 1 |
346,562 | 10,416,644,345 | IssuesEvent | 2019-09-14 15:16:13 | EasyX-Community/EasyNOMP-old | https://api.github.com/repos/EasyX-Community/EasyNOMP-old | closed | Show blocks / Immature blocks / balances | enhancement gui in progress medium priority v1.1.4 website | Show blocks / immature blocks in block explorer and immature balances on worker pages | 1.0 | Show blocks / Immature blocks / balances - Show blocks / immature blocks in block explorer and immature balances on worker pages | priority | show blocks immature blocks balances show blocks immature blocks in block explorer and immature balances on worker pages | 1 |
420,669 | 12,241,045,274 | IssuesEvent | 2020-05-05 02:34:29 | projectacrn/acrn-hypervisor | https://api.github.com/repos/projectacrn/acrn-hypervisor | closed | doc: document how to launch 2 Linux UOSes using SDC2 scenario | priority: P3-Medium type: bug | ACRN introduces new SDC2 scenario recently, a tutorial is nice to have, which provide step by step instructions on how to set up the SDC2 scenario and run 2 Linux UOSes in post-launched VMs. | 1.0 | doc: document how to launch 2 Linux UOSes using SDC2 scenario - ACRN introduces new SDC2 scenario recently, a tutorial is nice to have, which provide step by step instructions on how to set up the SDC2 scenario and run 2 Linux UOSes in post-launched VMs. | priority | doc document how to launch linux uoses using scenario acrn introduces new scenario recently a tutorial is nice to have which provide step by step instructions on how to set up the scenario and run linux uoses in post launched vms | 1 |
698,998 | 23,999,785,318 | IssuesEvent | 2022-09-14 10:27:12 | CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | opened | [Collaboration UI] Question Bounds | Module/Front-End Status/Medium-Priority Type/Feature | ## Description
The UI should show the bounds of the question's input variable(s)
## Parent Task
- #72 | 1.0 | [Collaboration UI] Question Bounds - ## Description
The UI should show the bounds of the question's input variable(s)
## Parent Task
- #72 | priority | question bounds description the ui should show the bounds of the question s input variable s parent task | 1 |
727,295 | 25,030,319,866 | IssuesEvent | 2022-11-04 11:46:56 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Open/Close toolbar and action buttons on mobile | Sprint Priority: Medium Feature | **Is your feature request related to a problem? Please describe.**
On mobile, we would like to more easily use the search, sort, filter functions in the toolbar, when needed. Similarly, the new action buttons on mobile are currently always present, but can be hidden when they are not needed.
**Describe the solution you'd like**
We agree to top and bottom navigation drawers. Something like:
Link to the [Figma design](https://www.figma.com/file/kpGAA005yHdL6Ch1rncaB2/SaaS-Improvements?node-id=740%3A6977).

| 1.0 | Open/Close toolbar and action buttons on mobile - **Is your feature request related to a problem? Please describe.**
On mobile, we would like to more easily use the search, sort, filter functions in the toolbar, when needed. Similarly, the new action buttons on mobile are currently always present, but can be hidden when they are not needed.
**Describe the solution you'd like**
We agree to top and bottom navigation drawers. Something like:
Link to the [Figma design](https://www.figma.com/file/kpGAA005yHdL6Ch1rncaB2/SaaS-Improvements?node-id=740%3A6977).

| priority | open close toolbar and action buttons on mobile is your feature request related to a problem please describe on mobile we would like to more easily use the search sort filter functions in the toolbar when needed similarly the new action buttons on mobile are currently always present but can be hidden when they are not needed describe the solution you d like we agree to top and bottom navigation drawers something like link to the | 1 |
576,860 | 17,096,807,700 | IssuesEvent | 2021-07-09 04:49:22 | LibreTexts/metalc | https://api.github.com/repos/LibreTexts/metalc | closed | ZFS scrub not sending emails | good first issue medium priority | We have a cronjob running on gravity that is supposed to ssh into blackhole, our zfs machine, and run a zfs scrub script then email us back the report. For one reason or another, we have not received any emails about this in a while, so it is not working. | 1.0 | ZFS scrub not sending emails - We have a cronjob running on gravity that is supposed to ssh into blackhole, our zfs machine, and run a zfs scrub script then email us back the report. For one reason or another, we have not received any emails about this in a while, so it is not working. | priority | zfs scrub not sending emails we have a cronjob running on gravity that is supposed to ssh into blackhole our zfs machine and run a zfs scrub script then email us back the report for one reason or another we have not received any emails about this in a while so it is not working | 1 |
462,365 | 13,245,840,106 | IssuesEvent | 2020-08-19 14:54:33 | phetsims/tandem | https://api.github.com/repos/phetsims/tandem | closed | PhetioGroup is not scalable | dev:phet-io phet-io:collaboration priority:3-medium | Related to https://github.com/phetsims/studio/issues/122 (Studio handling of dynamic elements is not scalable)
In Natural Selection, bunnies are dynamically created via a PhetioGroup subclass - see BunnyGroup.js. The group can contain at most ~3000 living bunnies, and an open-ended number of dead bunnies (bunnies dies after 2 generations, or if killed by environmental factors).
Scalability problems that I see in Natural Selection:
1. PhetioGroup creates elements 1 at a time. On each cycle of the clock, bunnies are paired up and have a litter of 4 bunnies. If I have 800 bunnies, that's 400 pairs x 4 per litter = 1600 bunnies that will be created, and 1600 calls to `PhetioGroup.createNextElement`. This results in a very noticeable pause in animation (seconds).
2. PhetioGroup disposes of elements 1 at a time. If 50 bunnies die at the same time, that's 50 calls to `PhetioGroup.disposeElement`.
3. PhetioGroup `clear` calls `disposeElement` for each element in the group. Bunnies take over the world when there are > 1000 living bunnies (and an open-ended number of dead bunnies). Pressing the _Reset All_ button or _Start Over_ button in Natural Selection results in a significant lag (seconds), because a large number of calls to `PhetioGroup.disposeElement` is made.
4. Performance for all of the above is even worse in Studio, making it effectively unusable. See https://github.com/phetsims/studio/issues/122. | 1.0 | PhetioGroup is not scalable - Related to https://github.com/phetsims/studio/issues/122 (Studio handling of dynamic elements is not scalable)
In Natural Selection, bunnies are dynamically created via a PhetioGroup subclass - see BunnyGroup.js. The group can contain at most ~3000 living bunnies, and an open-ended number of dead bunnies (bunnies dies after 2 generations, or if killed by environmental factors).
Scalability problems that I see in Natural Selection:
1. PhetioGroup creates elements 1 at a time. On each cycle of the clock, bunnies are paired up and have a litter of 4 bunnies. If I have 800 bunnies, that's 400 pairs x 4 per litter = 1600 bunnies that will be created, and 1600 calls to `PhetioGroup.createNextElement`. This results in a very noticeable pause in animation (seconds).
2. PhetioGroup disposes of elements 1 at a time. If 50 bunnies die at the same time, that's 50 calls to `PhetioGroup.disposeElement`.
3. PhetioGroup `clear` calls `disposeElement` for each element in the group. Bunnies take over the world when there are > 1000 living bunnies (and an open-ended number of dead bunnies). Pressing the _Reset All_ button or _Start Over_ button in Natural Selection results in a significant lag (seconds), because a large number of calls to `PhetioGroup.disposeElement` is made.
4. Performance for all of the above is even worse in Studio, making it effectively unusable. See https://github.com/phetsims/studio/issues/122. | priority | phetiogroup is not scalable related to studio handling of dynamic elements is not scalable in natural selection bunnies are dynamically created via a phetiogroup subclass see bunnygroup js the group can contain at most living bunnies and an open ended number of dead bunnies bunnies dies after generations or if killed by environmental factors scalability problems that i see in natural selection phetiogroup creates elements at a time on each cycle of the clock bunnies are paired up and have a litter of bunnies if i have bunnies that s pairs x per litter bunnies that will be created and calls to phetiogroup createnextelement this results in a very noticeable pause in animation seconds phetiogroup disposes of elements at a time if bunnies die at the same time that s calls to phetiogroup disposeelement phetiogroup clear calls disposeelement for each element in the group bunnies take over the world when there are living bunnies and an open ended number of dead bunnies pressing the reset all button or start over button in natural selection results in a significant lag seconds because a large number of calls to phetiogroup disposeelement is made performance for all of the above is even worse in studio making it effectively unusable see | 1 |
361,560 | 10,710,670,646 | IssuesEvent | 2019-10-25 03:10:05 | AY1920S1-CS2103T-F12-3/main | https://api.github.com/repos/AY1920S1-CS2103T-F12-3/main | closed | Fix note filtering proposed feature in Developer Guide | priority.Medium type.Task | - Fixed diagram's uneven lines and split them into less complex diagrams
- Add design consideration section possibly by adding a method `ListNoteCommand#getPredicate` which will replace `ListNoteCommand#filterAndListByTag` and `ListNoteCommand#filterAndListByKeyword` method | 1.0 | Fix note filtering proposed feature in Developer Guide - - Fixed diagram's uneven lines and split them into less complex diagrams
- Add design consideration section possibly by adding a method `ListNoteCommand#getPredicate` which will replace `ListNoteCommand#filterAndListByTag` and `ListNoteCommand#filterAndListByKeyword` method | priority | fix note filtering proposed feature in developer guide fixed diagram s uneven lines and split them into less complex diagrams add design consideration section possibly by adding a method listnotecommand getpredicate which will replace listnotecommand filterandlistbytag and listnotecommand filterandlistbykeyword method | 1 |
638,966 | 20,743,591,480 | IssuesEvent | 2022-03-14 20:13:59 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | Item 22540 doesn't work after update | component:database status:confirmed component:skill priority:medium mode:renewal mode:prerenewal type:bug | <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->3be73773f
* **Client Date**: 20211103
<!-- Please specify the client date you used. -->
* **Server Mode**: Re
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Runstone_Lux doesn't work after update
* Result: <!-- Describe the issue that you experienced in detail. -->
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**: None
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| 1.0 | Item 22540 doesn't work after update - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->3be73773f
* **Client Date**: 20211103
<!-- Please specify the client date you used. -->
* **Server Mode**: Re
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Runstone_Lux doesn't work after update
* Result: <!-- Describe the issue that you experienced in detail. -->
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**: None
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| priority | item doesn t work after update rathena hash please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode re description of issue runstone lux doesn t work after update result expected result how to reproduce official information modifications that may affect results none | 1 |
294,546 | 9,036,734,406 | IssuesEvent | 2019-02-09 02:32:01 | brandon1024/find | https://api.github.com/repos/brandon1024/find | closed | Investigate Options Disabled by Default in Firefox | medium priority | ## Issue Description
After installing the extension in Firefox, some of the features/options are disabled by default when they shouldn't be. This causes some confusion, and should be addressed.
| 1.0 | Investigate Options Disabled by Default in Firefox - ## Issue Description
After installing the extension in Firefox, some of the features/options are disabled by default when they shouldn't be. This causes some confusion, and should be addressed.
| priority | investigate options disabled by default in firefox issue description after installing the extension in firefox some of the features options are disabled by default when they shouldn t be this causes some confusion and should be addressed | 1 |
185,241 | 6,720,235,887 | IssuesEvent | 2017-10-16 06:51:19 | CS2103AUG2017-T15-B2/main | https://api.github.com/repos/CS2103AUG2017-T15-B2/main | opened | As a user I want to filter my contacts based on any search field | priority.medium type.story | ... so that I can avoid going through the whole list in the address book to find a contact | 1.0 | As a user I want to filter my contacts based on any search field - ... so that I can avoid going through the whole list in the address book to find a contact | priority | as a user i want to filter my contacts based on any search field so that i can avoid going through the whole list in the address book to find a contact | 1 |
417,118 | 12,155,912,700 | IssuesEvent | 2020-04-25 15:09:50 | Scifabric/pybossa | https://api.github.com/repos/Scifabric/pybossa | closed | Import data from FTP and HTTP listings | priority.medium | I'd like to create importer for data stored on FTP and HTTP listings. Should I do it as a pull-request to the core pybossa with files under [pybossa/importers](https://github.com/Scifabric/pybossa/tree/master/pybossa/importers) or should it be a separate plugin? | 1.0 | Import data from FTP and HTTP listings - I'd like to create importer for data stored on FTP and HTTP listings. Should I do it as a pull-request to the core pybossa with files under [pybossa/importers](https://github.com/Scifabric/pybossa/tree/master/pybossa/importers) or should it be a separate plugin? | priority | import data from ftp and http listings i d like to create importer for data stored on ftp and http listings should i do it as a pull request to the core pybossa with files under or should it be a separate plugin | 1 |
829,335 | 31,864,824,433 | IssuesEvent | 2023-09-15 13:28:31 | oceanbase/odc | https://api.github.com/repos/oceanbase/odc | opened | [Bug]: The primary key expression is not recognized when it contains a newline character, causing the dlm to fail | type-bug priority-medium module-Data generation and migration | ### ODC version
ODC 4.2.0
### OB version
OB 3.2.4
### What happened?
The primary key expression is not recognized when it contains a newline character, causing the dlm to fail
### What did you expect to happen?
The primary key expression for dlm tasks recognizes newlines
### How can we reproduce it (as minimally and precisely as possible)?
CREATE TABLE `user3` (
`id
` int(11) NOT NULL,
`name
` char(120) DEFAULT NULL,
`create_time` timestamp NULL DEFAULT NULL,
PRIMARY KEY (`id
`)
)
### Anything else we need to know?
_No response_
### Cloud
_No response_ | 1.0 | [Bug]: The primary key expression is not recognized when it contains a newline character, causing the dlm to fail - ### ODC version
ODC 4.2.0
### OB version
OB 3.2.4
### What happened?
The primary key expression is not recognized when it contains a newline character, causing the dlm to fail
### What did you expect to happen?
The primary key expression for dlm tasks recognizes newlines
### How can we reproduce it (as minimally and precisely as possible)?
CREATE TABLE `user3` (
`id
` int(11) NOT NULL,
`name
` char(120) DEFAULT NULL,
`create_time` timestamp NULL DEFAULT NULL,
PRIMARY KEY (`id
`)
)
### Anything else we need to know?
_No response_
### Cloud
_No response_ | priority | the primary key expression is not recognized when it contains a newline character causing the dlm to fail odc version odc ob version ob what happened the primary key expression is not recognized when it contains a newline character causing the dlm to fail what did you expect to happen the primary key expression for dlm tasks recognizes newlines how can we reproduce it as minimally and precisely as possible create table id int not null name char default null create time timestamp null default null primary key id anything else we need to know no response cloud no response | 1 |
24,651 | 2,671,324,067 | IssuesEvent | 2015-03-24 05:00:50 | cs2103jan2015-w11-1c/main | https://api.github.com/repos/cs2103jan2015-w11-1c/main | closed | A user can set different priorities | priority.medium status.ongoing type.enhancement | so that the user can give more attention to the more pressing tasks | 1.0 | A user can set different priorities - so that the user can give more attention to the more pressing tasks | priority | a user can set different priorities so that the user can give more attention to the more pressing tasks | 1 |
652,548 | 21,555,628,761 | IssuesEvent | 2022-04-30 11:14:02 | DavidZamanian/Bachelor-thesis-blockchain-for-medical-records | https://api.github.com/repos/DavidZamanian/Bachelor-thesis-blockchain-for-medical-records | opened | Placeholders for EHR-content when empty | enhancement FRONTEND Medium Priority | Add labels like: "There are no prescriptions", "There are no diagnoses" to the EHROverview-page in cases when patients do not have them.
This can also be useful for testing. | 1.0 | Placeholders for EHR-content when empty - Add labels like: "There are no prescriptions", "There are no diagnoses" to the EHROverview-page in cases when patients do not have them.
This can also be useful for testing. | priority | placeholders for ehr content when empty add labels like there are no prescriptions there are no diagnoses to the ehroverview page in cases when patients do not have them this can also be useful for testing | 1 |
59,272 | 3,104,736,917 | IssuesEvent | 2015-08-31 17:24:01 | clouder-community/clouder | https://api.github.com/repos/clouder-community/clouder | opened | Add another abstract layer between Clouder and Docker | clouder core enhancement medium priority Need conception discussions before development | Hello everyone,
Like some of you suggested, we today have many great tools to manage Docker instances. The goal of Clouder is not to replace them, it's to be the place where the informations about application/containers/base are stored.
Since I want to keep the minimum command in Clouder, I believe we have a missing software on the process between Clouder and Docker, I mean Clouder shall give order to this software and this software will then manage the containers. As we do for the registry, this software shall also be in a container created by Clouder.
Some of you suggested docker-compose, other rancher, any other suggestions ? We will probably have to make a choice, we need to make the good one. | 1.0 | Add another abstract layer between Clouder and Docker - Hello everyone,
Like some of you suggested, we today have many great tools to manage Docker instances. The goal of Clouder is not to replace them, it's to be the place where the informations about application/containers/base are stored.
Since I want to keep the minimum command in Clouder, I believe we have a missing software on the process between Clouder and Docker, I mean Clouder shall give order to this software and this software will then manage the containers. As we do for the registry, this software shall also be in a container created by Clouder.
Some of you suggested docker-compose, other rancher, any other suggestions ? We will probably have to make a choice, we need to make the good one. | priority | add another abstract layer between clouder and docker hello everyone like some of you suggested we today have many great tools to manage docker instances the goal of clouder is not to replace them it s to be the place where the informations about application containers base are stored since i want to keep the minimum command in clouder i believe we have a missing software on the process between clouder and docker i mean clouder shall give order to this software and this software will then manage the containers as we do for the registry this software shall also be in a container created by clouder some of you suggested docker compose other rancher any other suggestions we will probably have to make a choice we need to make the good one | 1 |
30,255 | 2,723,310,526 | IssuesEvent | 2015-04-14 11:40:59 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | The Tab is not closed if opened with closeable = "false" | bug CruxWidgetsLibrary imported Milestone-2.1.0 Priority-Medium | _From [lucianor...@gmail.com](https://code.google.com/u/114491495264375709919/) on March 08, 2010 14:13:35_
The Tab is not closed if it will be opened by DynaTabs.openTab with the
attribute closeable equal false
steps to reproduce the problem:
1.Open tab with closeable attribute equal false
"DynaTabs.openTab(tabId, tabLabel, appendMainUrlParameters(tabUrl), false,
false);"
2. Close tab with method DynaTabs.closeTab(tabId)
3.The Tab is not closed
Please it verifies the behavior...
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=15_ | 1.0 | The Tab is not closed if opened with closeable = "false" - _From [lucianor...@gmail.com](https://code.google.com/u/114491495264375709919/) on March 08, 2010 14:13:35_
The Tab is not closed if it will be opened by DynaTabs.openTab with the
attribute closeable equal false
steps to reproduce the problem:
1.Open tab with closeable attribute equal false
"DynaTabs.openTab(tabId, tabLabel, appendMainUrlParameters(tabUrl), false,
false);"
2. Close tab with method DynaTabs.closeTab(tabId)
3.The Tab is not closed
Please it verifies the behavior...
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=15_ | priority | the tab is not closed if opened with closeable false from on march the tab is not closed if it will be opened by dynatabs opentab with the attribute closeable equal false steps to reproduce the problem open tab with closeable attribute equal false dynatabs opentab tabid tablabel appendmainurlparameters taburl false false close tab with method dynatabs closetab tabid the tab is not closed please it verifies the behavior original issue | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.