Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
427,105 | 12,392,975,797 | IssuesEvent | 2020-05-20 14:46:03 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Anki custom study session changes original card state | Bug Priority-Medium Reproduced | ###### Reproduction Steps
1. I started custom study session on Synonyms and antonyms

2. Reschedule turned off

3. [Apkg file of the session](https://drive.google.com/file/d/1e4s_lQewsdVVjzqYu1jzUZxKI5hBafY6/view?usp=drivesdk)
###### Expected Result
###### Actual Result
The original card session state changed

###### Debug info
AnkiDroid Version = 2.10.2
Android Version = 9
ACRA UUID = 6e2a8e53-f25c-4cc6-a183-7dab54dad6a0
###### Research
*Enter an [x] character to confirm the points below:*
- [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
- [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
- [x] I have searched for similar existing issues here and on the user forum
- [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
| 1.0 | Anki custom study session changes original card state - ###### Reproduction Steps
1. I started custom study session on Synonyms and antonyms

2. Reschedule turned off

3. [Apkg file of the session](https://drive.google.com/file/d/1e4s_lQewsdVVjzqYu1jzUZxKI5hBafY6/view?usp=drivesdk)
###### Expected Result
###### Actual Result
The original card session state changed

###### Debug info
AnkiDroid Version = 2.10.2
Android Version = 9
ACRA UUID = 6e2a8e53-f25c-4cc6-a183-7dab54dad6a0
###### Research
*Enter an [x] character to confirm the points below:*
- [x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
- [x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
- [x] I have searched for similar existing issues here and on the user forum
- [x] (Optional) I have confirmed the issue is not resolved in the latest alpha release ([instructions](https://docs.ankidroid.org/manual.html#betaTesting))
| priority | anki custom study session changes original card state reproduction steps i started custom study session on synonyms and antonyms reschedule turned off expected result actual result the original card session state changed debug info ankidroid version android version acra uuid research enter an character to confirm the points below i have read the and am reporting a bug or enhancement request specific to ankidroid i have checked the and the and could not find a solution to my issue i have searched for similar existing issues here and on the user forum optional i have confirmed the issue is not resolved in the latest alpha release | 1 |
280,616 | 8,684,068,899 | IssuesEvent | 2018-12-02 23:32:22 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | suggest recently/frequently used keys/values | Enhancement Medium Priority Work in progress | During my editing I use small set of tags. JOSM behaviour where recently used tags are suggested is very useful in such situation.
Currently Vespucci is showing possible keys/values in alphabetical order. Maybe at least part of that list may be reserved for popular/recently used keys/values? | 1.0 | suggest recently/frequently used keys/values - During my editing I use small set of tags. JOSM behaviour where recently used tags are suggested is very useful in such situation.
Currently Vespucci is showing possible keys/values in alphabetical order. Maybe at least part of that list may be reserved for popular/recently used keys/values? | priority | suggest recently frequently used keys values during my editing i use small set of tags josm behaviour where recently used tags are suggested is very useful in such situation currently vespucci is showing possible keys values in alphabetical order maybe at least part of that list may be reserved for popular recently used keys values | 1 |
684,489 | 23,419,980,840 | IssuesEvent | 2022-08-13 14:45:02 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | [Bug] Kotlin: can no longer commit hunks | Priority-Medium Keep Open Dev | It seems that due to the pre-commit hook, the full file is committed
https://github.com/ankidroid/Anki-Android/blob/master/pre-commit | 1.0 | [Bug] Kotlin: can no longer commit hunks - It seems that due to the pre-commit hook, the full file is committed
https://github.com/ankidroid/Anki-Android/blob/master/pre-commit | priority | kotlin can no longer commit hunks it seems that due to the pre commit hook the full file is committed | 1 |
183,491 | 6,689,098,860 | IssuesEvent | 2017-10-08 22:02:50 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] Error when generating docs (Parse error on line 33, Expecting 'CLOSE', got 'ID') | Priority: Medium Time: ~1 hour Type: Bug | ##### **Overview of the issue**
When trying to generate the documentation, the following error occurs:
``` bash
[11:17:26] Process page : typealiases
[11:17:26] Process page : enumerations
[11:17:26] Process page : coverage
[11:17:26] Error: Parse error on line 33:
... {{> menu data menu='normal' }}
-----------------------^
Expecting 'CLOSE', got 'ID'
```
##### **Operating System, Node.js, npm, compodoc version(s)**
- Issue occuring with `compodoc 1.0.0-beta.14` and `compodoc 1.0.0-beta.15`
(`compodoc 1.0.0-beta.13` worked fine although the generation took forever)
- Windows 7
- NodeJS 8.2.1
- Angular 4.3.2
##### **Angular configuration, a `package.json` file in the root folder**
The command we use:
``` bash
compodoc src/lib --tsconfig src/tsconfig.docs.json
```
The referenced tsconfig file:
``` json
{
"compilerOptions": {
"baseUrl": "",
"declaration": false,
"emitDecoratorMetadata": true,
"experimentalDecorators": true,
"lib": [
"es2016",
"dom"
],
"module": "es2015",
"moduleResolution": "node",
"outDir": "../out-tsc/app",
"sourceMap": true,
"target": "es6",
"types": []
},
"exclude": [
"test.ts",
"polyfills.ts",
"main.ts",
"index.ts",
"**/*.spec.ts",
"app/**/*"
]
}
```
##### **Compodoc installed globally or locally ?**
locally
##### **Motivation for or Use Case**
The documentation won't build at all. | 1.0 | [BUG] Error when generating docs (Parse error on line 33, Expecting 'CLOSE', got 'ID') - ##### **Overview of the issue**
When trying to generate the documentation, the following error occurs:
``` bash
[11:17:26] Process page : typealiases
[11:17:26] Process page : enumerations
[11:17:26] Process page : coverage
[11:17:26] Error: Parse error on line 33:
... {{> menu data menu='normal' }}
-----------------------^
Expecting 'CLOSE', got 'ID'
```
##### **Operating System, Node.js, npm, compodoc version(s)**
- Issue occuring with `compodoc 1.0.0-beta.14` and `compodoc 1.0.0-beta.15`
(`compodoc 1.0.0-beta.13` worked fine although the generation took forever)
- Windows 7
- NodeJS 8.2.1
- Angular 4.3.2
##### **Angular configuration, a `package.json` file in the root folder**
The command we use:
``` bash
compodoc src/lib --tsconfig src/tsconfig.docs.json
```
The referenced tsconfig file:
``` json
{
"compilerOptions": {
"baseUrl": "",
"declaration": false,
"emitDecoratorMetadata": true,
"experimentalDecorators": true,
"lib": [
"es2016",
"dom"
],
"module": "es2015",
"moduleResolution": "node",
"outDir": "../out-tsc/app",
"sourceMap": true,
"target": "es6",
"types": []
},
"exclude": [
"test.ts",
"polyfills.ts",
"main.ts",
"index.ts",
"**/*.spec.ts",
"app/**/*"
]
}
```
##### **Compodoc installed globally or locally ?**
locally
##### **Motivation for or Use Case**
The documentation won't build at all. | priority | error when generating docs parse error on line expecting close got id overview of the issue when trying to generate the documentation the following error occurs bash process page typealiases process page enumerations process page coverage error parse error on line menu data menu normal expecting close got id operating system node js npm compodoc version s issue occuring with compodoc beta and compodoc beta compodoc beta worked fine although the generation took forever windows nodejs angular angular configuration a package json file in the root folder the command we use bash compodoc src lib tsconfig src tsconfig docs json the referenced tsconfig file json compileroptions baseurl declaration false emitdecoratormetadata true experimentaldecorators true lib dom module moduleresolution node outdir out tsc app sourcemap true target types exclude test ts polyfills ts main ts index ts spec ts app compodoc installed globally or locally locally motivation for or use case the documentation won t build at all | 1 |
411,655 | 12,027,202,138 | IssuesEvent | 2020-04-12 17:21:27 | AY1920S2-CS2103-W15-2/main | https://api.github.com/repos/AY1920S2-CS2103-W15-2/main | closed | Generate Report of Interviews in the form of PDF | priority.Medium severity.Medium type.Enhancement | The interview report might be needed in the form of a PDF, this feature is very beneficial for the interviewee. | 1.0 | Generate Report of Interviews in the form of PDF - The interview report might be needed in the form of a PDF, this feature is very beneficial for the interviewee. | priority | generate report of interviews in the form of pdf the interview report might be needed in the form of a pdf this feature is very beneficial for the interviewee | 1 |
432,932 | 12,500,228,908 | IssuesEvent | 2020-06-01 21:47:00 | OrangeJuice7/SDL-OpenGL-Game-Framework | https://api.github.com/repos/OrangeJuice7/SDL-OpenGL-Game-Framework | opened | Better animated sprites | assets.graphics priority.medium work.medium | Also create sprites that face different directions, and track that direction in the Entity. | 1.0 | Better animated sprites - Also create sprites that face different directions, and track that direction in the Entity. | priority | better animated sprites also create sprites that face different directions and track that direction in the entity | 1 |
348,469 | 10,442,792,736 | IssuesEvent | 2019-09-18 13:44:56 | CLOSER-Cohorts/archivist | https://api.github.com/repos/CLOSER-Cohorts/archivist | closed | Unable to change topic linked to variable | Database Mappings Medium priority | In the dataset view changing the topic linked to a variable to blank, none or selecting a new topic, it processes for a while and then eventually times out. | 1.0 | Unable to change topic linked to variable - In the dataset view changing the topic linked to a variable to blank, none or selecting a new topic, it processes for a while and then eventually times out. | priority | unable to change topic linked to variable in the dataset view changing the topic linked to a variable to blank none or selecting a new topic it processes for a while and then eventually times out | 1 |
517,855 | 15,020,469,114 | IssuesEvent | 2021-02-01 14:45:03 | ansible/galaxy_ng | https://api.github.com/repos/ansible/galaxy_ng | closed | UI: Disable groups and repo management in navigation for users with no permissions | area/frontend priority/medium status/in-progress type/enhancement | Disable groups and and repo management in the navigation menu for users that don't have permissions to view these pages.
Additionally, controls for creating, editing and deleting groups as well as the button for syncing and editing remotes should be disabled based on the users permissions.

| 1.0 | UI: Disable groups and repo management in navigation for users with no permissions - Disable groups and and repo management in the navigation menu for users that don't have permissions to view these pages.
Additionally, controls for creating, editing and deleting groups as well as the button for syncing and editing remotes should be disabled based on the users permissions.

| priority | ui disable groups and repo management in navigation for users with no permissions disable groups and and repo management in the navigation menu for users that don t have permissions to view these pages additionally controls for creating editing and deleting groups as well as the button for syncing and editing remotes should be disabled based on the users permissions | 1 |
121,461 | 4,816,836,900 | IssuesEvent | 2016-11-04 11:31:17 | SuperTux/supertux | https://api.github.com/repos/SuperTux/supertux | closed | Resizable Sprites | involves:functionality priority:medium status:needs-work type:idea | Sprites and Surfaces should be resizable. This is needed for the storage of powerups and items, because the user should be allowed to add his own collectible items and it's easier for the user to be able to use a Sprite in any resolution.
It might be a good idea to resize the images, and store them in a specail directory, because else the images would have to be resized too often, and it will ensure that the game is fast. For avoiding storing too many images they should be kept and maybe deleted if they weren't used in a long time.
TODO
- [ ] Implement a scaling algorithm for PNGs
- [ ] Implement the resized image container
- [ ] Extend this to Sprites
| 1.0 | Resizable Sprites - Sprites and Surfaces should be resizable. This is needed for the storage of powerups and items, because the user should be allowed to add his own collectible items and it's easier for the user to be able to use a Sprite in any resolution.
It might be a good idea to resize the images, and store them in a specail directory, because else the images would have to be resized too often, and it will ensure that the game is fast. For avoiding storing too many images they should be kept and maybe deleted if they weren't used in a long time.
TODO
- [ ] Implement a scaling algorithm for PNGs
- [ ] Implement the resized image container
- [ ] Extend this to Sprites
| priority | resizable sprites sprites and surfaces should be resizable this is needed for the storage of powerups and items because the user should be allowed to add his own collectible items and it s easier for the user to be able to use a sprite in any resolution it might be a good idea to resize the images and store them in a specail directory because else the images would have to be resized too often and it will ensure that the game is fast for avoiding storing too many images they should be kept and maybe deleted if they weren t used in a long time todo implement a scaling algorithm for pngs implement the resized image container extend this to sprites | 1 |
104,949 | 4,227,015,495 | IssuesEvent | 2016-07-02 21:44:32 | octobercms/october | https://api.github.com/repos/octobercms/october | closed | Crop and insert can produce "Invalid selection data" when a fixed ratio is used | Priority: Medium Status: Completed Type: Bug | Trying to crop an image using a fixed ratio can produce a _Invalid selection data_ message. Since [this commit](https://github.com/tapmodo/Jcrop/commit/3a51e67f8dbced3b4f763cb832f75c0a0750ec72#diff-f93c6ae36034c63a1c3e1d298a9af0d0L656) selection data can contain decimal places, which [prevents](https://github.com/octobercms/october/blob/ab4abb9bc5c42f09028e25206cd6d575fa3b1f31/modules/cms/widgets/MediaManager.php#L1124) images from being created.
##### Expected behavior
A cropped image should be created.
##### Actual behavior
An error for _Invalid selection data_ is shown.
##### Reproduce steps
1. Select crop and insert
2. Set selection mode to fixed ratio
3. Set width and height to 4/3
4. Make a selection with width and/or height that contain decimal places
5. Clip Crop and insert
##### October build
318 | 1.0 | Crop and insert can produce "Invalid selection data" when a fixed ratio is used - Trying to crop an image using a fixed ratio can produce a _Invalid selection data_ message. Since [this commit](https://github.com/tapmodo/Jcrop/commit/3a51e67f8dbced3b4f763cb832f75c0a0750ec72#diff-f93c6ae36034c63a1c3e1d298a9af0d0L656) selection data can contain decimal places, which [prevents](https://github.com/octobercms/october/blob/ab4abb9bc5c42f09028e25206cd6d575fa3b1f31/modules/cms/widgets/MediaManager.php#L1124) images from being created.
##### Expected behavior
A cropped image should be created.
##### Actual behavior
An error for _Invalid selection data_ is shown.
##### Reproduce steps
1. Select crop and insert
2. Set selection mode to fixed ratio
3. Set width and height to 4/3
4. Make a selection with width and/or height that contain decimal places
5. Clip Crop and insert
##### October build
318 | priority | crop and insert can produce invalid selection data when a fixed ratio is used trying to crop an image using a fixed ratio can produce a invalid selection data message since selection data can contain decimal places which images from being created expected behavior a cropped image should be created actual behavior an error for invalid selection data is shown reproduce steps select crop and insert set selection mode to fixed ratio set width and height to make a selection with width and or height that contain decimal places clip crop and insert october build | 1 |
412,254 | 12,037,249,657 | IssuesEvent | 2020-04-13 21:25:37 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Eco Tree dev features | Category: Web Priority: Medium Status: Fixed | 
I'd like to expand the role of the dev icons here and make it part of our daily updates. The idea is that we make this a center for public development and personal connection between the team and community.
- [x] Jeff: add HTML code and styles for all the improvements of Tooltip, so Denys would just filled them with logic afterwards. Like `status`, `title`, `bio` and `email` links.
- [x] Denys: Add the ability to set a 'status' for each developer of what they're working on, with the date that the status was applied. Display this status in the tooltip.
- [x] Denys: Display a title for each dev (mentioned this before but would like to do it). Programmer, designer, artist, QA.
- [x] Jeff: Have two links in the tooltip that can be clicked, bio and email (not all devs have slg email, some signed up to wordpress with their private email. we dont want to publish private emails, so create new user field for email. should be optional).
- [x] Denys: Have an easy way for the team to set their status. They will be doing this everyday, so make sure it's easy. Ideally they can just open a page, click the node they're working on, and type a short status.
Bonus:
- [ ] Have 'recent updates' for each user, that links to the last few posts they made in the eco tree somewhere. | 1.0 | Eco Tree dev features - 
I'd like to expand the role of the dev icons here and make it part of our daily updates. The idea is that we make this a center for public development and personal connection between the team and community.
- [x] Jeff: add HTML code and styles for all the improvements of Tooltip, so Denys would just filled them with logic afterwards. Like `status`, `title`, `bio` and `email` links.
- [x] Denys: Add the ability to set a 'status' for each developer of what they're working on, with the date that the status was applied. Display this status in the tooltip.
- [x] Denys: Display a title for each dev (mentioned this before but would like to do it). Programmer, designer, artist, QA.
- [x] Jeff: Have two links in the tooltip that can be clicked, bio and email (not all devs have slg email, some signed up to wordpress with their private email. we dont want to publish private emails, so create new user field for email. should be optional).
- [x] Denys: Have an easy way for the team to set their status. They will be doing this everyday, so make sure it's easy. Ideally they can just open a page, click the node they're working on, and type a short status.
Bonus:
- [ ] Have 'recent updates' for each user, that links to the last few posts they made in the eco tree somewhere. | priority | eco tree dev features i d like to expand the role of the dev icons here and make it part of our daily updates the idea is that we make this a center for public development and personal connection between the team and community jeff add html code and styles for all the improvements of tooltip so denys would just filled them with logic afterwards like status title bio and email links denys add the ability to set a status for each developer of what they re working on with the date that the status was applied display this status in the tooltip denys display a title for each dev mentioned this before but would like to do it programmer designer artist qa jeff have two links in the tooltip that can be clicked bio and email not all devs have slg email some signed up to wordpress with their private email we dont want to publish private emails so create new user field for email should be optional denys have an easy way for the team to set their status they will be doing this everyday so make sure it s easy ideally they can just open a page click the node they re working on and type a short status bonus have recent updates for each user that links to the last few posts they made in the eco tree somewhere | 1 |
392,208 | 11,584,825,490 | IssuesEvent | 2020-02-22 19:38:41 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | Has a class in Artifact Conditions | :beetle: bug - localisation :scroll: :grey_exclamation: priority medium | <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Mod Version**
0317ee1deb744cfa454aa5f410a1fb475d84b255
**Are you using any submods/mods? If so, which?**
No
**Please explain your issue in as much detail as possible:**
Localization is wrong. I have a class, but it returns false.
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details>
| 1.0 | Has a class in Artifact Conditions - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Mod Version**
0317ee1deb744cfa454aa5f410a1fb475d84b255
**Are you using any submods/mods? If so, which?**
No
**Please explain your issue in as much detail as possible:**
Localization is wrong. I have a class, but it returns false.
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details>
| priority | has a class in artifact conditions do not remove pre existing lines mod version are you using any submods mods if so which no please explain your issue in as much detail as possible localization is wrong i have a class but it returns false upload screenshots of the problem localization click to expand | 1 |
68,387 | 3,286,940,858 | IssuesEvent | 2015-10-29 07:23:03 | cs2103aug2015-t16-4j/main | https://api.github.com/repos/cs2103aug2015-t16-4j/main | closed | As a user, Jim wants to be able to have the program warn him if he places 2 events at the same time | priority.medium type.story | In case he mistakenly plans his timetable with clashes | 1.0 | As a user, Jim wants to be able to have the program warn him if he places 2 events at the same time - In case he mistakenly plans his timetable with clashes | priority | as a user jim wants to be able to have the program warn him if he places events at the same time in case he mistakenly plans his timetable with clashes | 1 |
2,281 | 2,525,001,611 | IssuesEvent | 2015-01-20 21:32:07 | graybeal/ont | https://api.github.com/repos/graybeal/ont | closed | Create device ontology instantiation | 1 star enhancement imported OntDev Priority-Medium | _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on November 23, 2009 12:47:36_
Description of the action item Create device ontology instantiation and present lessons learned to the team
(thanks Bob!) Please provide any relevant information/links below http://marinemetadata.org/community/teams/ontdevices/agendas/am20091027 http://marinemetadata.org/pipermail/ontdev/2009/000344.html http://marinemetadata.org/pipermail/ontdev/2009/000356.html
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=224_ | 1.0 | Create device ontology instantiation - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on November 23, 2009 12:47:36_
Description of the action item Create device ontology instantiation and present lessons learned to the team
(thanks Bob!) Please provide any relevant information/links below http://marinemetadata.org/community/teams/ontdevices/agendas/am20091027 http://marinemetadata.org/pipermail/ontdev/2009/000344.html http://marinemetadata.org/pipermail/ontdev/2009/000356.html
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=224_ | priority | create device ontology instantiation from on november description of the action item create device ontology instantiation and present lessons learned to the team thanks bob please provide any relevant information links below original issue | 1 |
217,828 | 7,328,432,224 | IssuesEvent | 2018-03-04 20:40:19 | PaulL48/SOEN341-SC4 | https://api.github.com/repos/PaulL48/SOEN341-SC4 | opened | Improvement: Login after registering | enhancement priority: low risk: medium | It would make sense if once the user registers, they would be logged in. | 1.0 | Improvement: Login after registering - It would make sense if once the user registers, they would be logged in. | priority | improvement login after registering it would make sense if once the user registers they would be logged in | 1 |
641,984 | 20,864,112,261 | IssuesEvent | 2022-03-22 04:11:32 | AY2122S2-CS2103T-T12-4/tp | https://api.github.com/repos/AY2122S2-CS2103T-T12-4/tp | closed | Allow users to add links to tasks | priority.Medium | As a user, I can add links to a task so that I can access the link. | 1.0 | Allow users to add links to tasks - As a user, I can add links to a task so that I can access the link. | priority | allow users to add links to tasks as a user i can add links to a task so that i can access the link | 1 |
413,359 | 12,065,992,831 | IssuesEvent | 2020-04-16 10:57:14 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Dashboard should be able to provide a list of all my current keys | Priority: Medium customer request enhancement | **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
- ~~Key listing on the gateway is possible only if the keys are not hashed.~~ Possible directly with the gateway by setting `"enable_hashed_keys_listing": true,` in tyk.conf.
- Even if you enabled key listing on both dashboard and gateway we list keys under apis, users need to fetch all the apis and then call /keys per api.
- We provide keys listing in the analytics. As such, admins need to wait until the keys are being used before they can see the keys on the system.
**What is the expected behavior?**
We should add the ability to see the keys as soon as they get created since we have this info already. It will be very powerful from the dashboard users perspective.
Since we have this info why not create an appropriate endpoint?
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.8 | 1.0 | Dashboard should be able to provide a list of all my current keys - **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
- ~~Key listing on the gateway is possible only if the keys are not hashed.~~ Possible directly with the gateway by setting `"enable_hashed_keys_listing": true,` in tyk.conf.
- Even if you enabled key listing on both dashboard and gateway we list keys under apis, users need to fetch all the apis and then call /keys per api.
- We provide keys listing in the analytics. As such, admins need to wait until the keys are being used before they can see the keys on the system.
**What is the expected behavior?**
We should add the ability to see the keys as soon as they get created since we have this info already. It will be very powerful from the dashboard users perspective.
Since we have this info why not create an appropriate endpoint?
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.8 | priority | dashboard should be able to provide a list of all my current keys do you want to request a feature or report a bug feature what is the current behavior key listing on the gateway is possible only if the keys are not hashed possible directly with the gateway by setting enable hashed keys listing true in tyk conf even if you enabled key listing on both dashboard and gateway we list keys under apis users need to fetch all the apis and then call keys per api we provide keys listing in the analytics as such admins need to wait until the keys are being used before they can see the keys on the system what is the expected behavior we should add the ability to see the keys as soon as they get created since we have this info already it will be very powerful from the dashboard users perspective since we have this info why not create an appropriate endpoint if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem which versions of tyk affected by this issue did this work in previous versions of tyk | 1 |
726,442 | 24,999,504,975 | IssuesEvent | 2022-11-03 06:07:38 | factly/validly | https://api.github.com/repos/factly/validly | opened | Add new validations to check the PSU name upto proper standardisation | enhancement good first issue priority:medium | Description
Similar to States we must add PSU names from the data dictionary as a part of validly-server.
Task
Column mapping should include a new column that will incorporate PSU name columns
Then the validation rule will be added similar to how state validation is present | 1.0 | Add new validations to check the PSU name upto proper standardisation - Description
Similar to States we must add PSU names from the data dictionary as a part of validly-server.
Task
Column mapping should include a new column that will incorporate PSU name columns
Then the validation rule will be added similar to how state validation is present | priority | add new validations to check the psu name upto proper standardisation description similar to states we must add psu names from the data dictionary as a part of validly server task column mapping should include a new column that will incorporate psu name columns then the validation rule will be added similar to how state validation is present | 1 |
552,647 | 16,246,279,808 | IssuesEvent | 2021-05-07 14:57:14 | curiouslearning/followthelearners | https://api.github.com/repos/curiouslearning/followthelearners | closed | Shell scripts for Firebase emulators on dev.followthelearners.org | medium priority | - Initalization script for Firebase emulator.
- Shutdown script for Firebase emulator.
- Export script to export the data from the emulator to a file before shutdown. | 1.0 | Shell scripts for Firebase emulators on dev.followthelearners.org - - Initalization script for Firebase emulator.
- Shutdown script for Firebase emulator.
- Export script to export the data from the emulator to a file before shutdown. | priority | shell scripts for firebase emulators on dev followthelearners org initalization script for firebase emulator shutdown script for firebase emulator export script to export the data from the emulator to a file before shutdown | 1 |
257,745 | 8,141,070,923 | IssuesEvent | 2018-08-20 23:58:48 | minio/minio | https://api.github.com/repos/minio/minio | closed | Crash: Unsupported bitrot algorithm | community duplicate priority: medium | unable to heal minio
## Expected Behavior
I should be able to heal minio bucket
## Current Behavior
executing "mc admin heal --recursive --json --debug live-minio-4" ends with minio crashing.
## Logs
<details>
<summary>Logs</summary>
<pre><code>
Aug 16 22:45:51 live-minio-4 minio[11564]: API: SYSTEM()
Aug 16 22:45:51 live-minio-4 minio[11564]: Time: 22:45:51 EEST 08/16/2018
Aug 16 22:45:51 live-minio-4 minio[11564]: Error: Unsupported bitrot algorithm
Aug 16 22:45:51 live-minio-4 minio[11564]: 1: cmd/logger/logger.go:326:logger.CriticalIf()
Aug 16 22:45:51 live-minio-4 minio[11564]: 2: cmd/xl-v1-metadata.go:77:cmd.BitrotAlgorithm.New()
Aug 16 22:45:51 live-minio-4 minio[11564]: 3: cmd/erasure.go:102:cmd.NewBitrotVerifier()
Aug 16 22:45:51 live-minio-4 minio[11564]: 4: cmd/xl-v1-healing-common.go:182:cmd.disksWithAllParts()
Aug 16 22:45:51 live-minio-4 minio[11564]: 5: cmd/xl-v1-healing.go:308:cmd.healObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 6: cmd/xl-v1-healing.go:627:cmd.xlObjects.HealObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 7: cmd/xl-sets.go:1230:cmd.(*xlSets).HealObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 8: cmd/admin-heal-ops.go:658:cmd.(*healSequence).healObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 9: cmd/admin-heal-ops.go:635:cmd.(*healSequence).healBucket()
Aug 16 22:45:51 live-minio-4 minio[11564]: 10: cmd/admin-heal-ops.go:577:cmd.(*healSequence).healBuckets()
Aug 16 22:45:51 live-minio-4 minio[11564]: 11: cmd/admin-heal-ops.go:523:cmd.healBuckets)-fm()
Aug 16 22:45:51 live-minio-4 minio[11564]: 12: cmd/admin-heal-ops.go:516:cmd.(*healSequence).traverseAndHeal.func1()
Aug 16 22:45:51 live-minio-4 minio[11564]: 13: cmd/admin-heal-ops.go:523:cmd.(*healSequence).traverseAndHeal()
Aug 16 22:45:51 live-minio-4 minio[11564]: panic: (struct {}) (0x1158e60,0x1eac830)
Aug 16 22:45:51 live-minio-4 minio[11564]: goroutine 2130102 [running]:
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd/logger.CriticalIf(0x146cee0, 0xc42003c960, 0x145c4c0, 0xc4233c3860)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/logger/logger.go:327 +0x74
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.BitrotAlgorithm.New(0x0, 0x2, 0x2)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-metadata.go:77 +0xac
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.NewBitrotVerifier(0x0, 0x0, 0x0, 0x0, 0xa00000)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/erasure.go:102 +0x2f
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.disksWithAllParts(0x146cf60, 0xc424872360, 0xc422774900, 0x10, 0x10, 0xc420e18000, 0x10, 0x10, 0xc422774300, 0x10, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing-common.go:182 +0x306
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.healObject(0x146cf60, 0xc424872360, 0xc422774200, 0x10, 0x10, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0xb, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing.go:308 +0x28e
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.xlObjects.HealObject(0xc42045f340, 0xc42045f380, 0xc42045f360, 0x0, 0x0, 0x0, 0x0, 0x146cf60, 0xc424872360, 0xc42010a3eb, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing.go:627 +0x38d
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*xlSets).HealObject(0xc420432b40, 0x146cf60, 0xc424872360, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0x0, 0x0, 0x0, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-sets.go:1230 +0x128
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healObject(0xc420738300, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0x0, 0x0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:658 +0xf4
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healBucket(0xc420738300, 0xc42010a3eb, 0x7, 0xc422bb05a0, 0x2)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:635 +0x2ec
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healBuckets(0xc420738300, 0xc4200a41e0, 0xc42523df00)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:577 +0x157
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).(github.com/minio/minio/cmd.healBuckets)-fm(0xc420738300, 0x0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:523 +0x2a
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).traverseAndHeal.func1(0xc420566bd0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:516 +0x8f
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).traverseAndHeal(0xc420738300)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:523 +0xd9
Aug 16 22:45:51 live-minio-4 minio[11564]: created by github.com/minio/minio/cmd.(*healSequence).healSequenceStart
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:466 +0xf0
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Main process exited, code=exited, status=2/INVALIDARGUMENT
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Unit entered failed state.
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Failed with result 'exit-code'.
</code></pre>
</details>
## Your Environment
* Version used (`minio version`): 2018-08-02T23:11:36Z
* Operating System and version (`uname -a`): 4.9.0-6-amd64 #1 SMP Debian 4.9.88-1+deb9u1 (2018-05-07) x86_64 GNU/Linux
| 1.0 | Crash: Unsupported bitrot algorithm - unable to heal minio
## Expected Behavior
I should be able to heal minio bucket
## Current Behavior
executing "mc admin heal --recursive --json --debug live-minio-4" ends with minio crashing.
## Logs
<details>
<summary>Logs</summary>
<pre><code>
Aug 16 22:45:51 live-minio-4 minio[11564]: API: SYSTEM()
Aug 16 22:45:51 live-minio-4 minio[11564]: Time: 22:45:51 EEST 08/16/2018
Aug 16 22:45:51 live-minio-4 minio[11564]: Error: Unsupported bitrot algorithm
Aug 16 22:45:51 live-minio-4 minio[11564]: 1: cmd/logger/logger.go:326:logger.CriticalIf()
Aug 16 22:45:51 live-minio-4 minio[11564]: 2: cmd/xl-v1-metadata.go:77:cmd.BitrotAlgorithm.New()
Aug 16 22:45:51 live-minio-4 minio[11564]: 3: cmd/erasure.go:102:cmd.NewBitrotVerifier()
Aug 16 22:45:51 live-minio-4 minio[11564]: 4: cmd/xl-v1-healing-common.go:182:cmd.disksWithAllParts()
Aug 16 22:45:51 live-minio-4 minio[11564]: 5: cmd/xl-v1-healing.go:308:cmd.healObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 6: cmd/xl-v1-healing.go:627:cmd.xlObjects.HealObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 7: cmd/xl-sets.go:1230:cmd.(*xlSets).HealObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 8: cmd/admin-heal-ops.go:658:cmd.(*healSequence).healObject()
Aug 16 22:45:51 live-minio-4 minio[11564]: 9: cmd/admin-heal-ops.go:635:cmd.(*healSequence).healBucket()
Aug 16 22:45:51 live-minio-4 minio[11564]: 10: cmd/admin-heal-ops.go:577:cmd.(*healSequence).healBuckets()
Aug 16 22:45:51 live-minio-4 minio[11564]: 11: cmd/admin-heal-ops.go:523:cmd.healBuckets)-fm()
Aug 16 22:45:51 live-minio-4 minio[11564]: 12: cmd/admin-heal-ops.go:516:cmd.(*healSequence).traverseAndHeal.func1()
Aug 16 22:45:51 live-minio-4 minio[11564]: 13: cmd/admin-heal-ops.go:523:cmd.(*healSequence).traverseAndHeal()
Aug 16 22:45:51 live-minio-4 minio[11564]: panic: (struct {}) (0x1158e60,0x1eac830)
Aug 16 22:45:51 live-minio-4 minio[11564]: goroutine 2130102 [running]:
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd/logger.CriticalIf(0x146cee0, 0xc42003c960, 0x145c4c0, 0xc4233c3860)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/logger/logger.go:327 +0x74
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.BitrotAlgorithm.New(0x0, 0x2, 0x2)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-metadata.go:77 +0xac
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.NewBitrotVerifier(0x0, 0x0, 0x0, 0x0, 0xa00000)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/erasure.go:102 +0x2f
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.disksWithAllParts(0x146cf60, 0xc424872360, 0xc422774900, 0x10, 0x10, 0xc420e18000, 0x10, 0x10, 0xc422774300, 0x10, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing-common.go:182 +0x306
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.healObject(0x146cf60, 0xc424872360, 0xc422774200, 0x10, 0x10, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0xb, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing.go:308 +0x28e
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.xlObjects.HealObject(0xc42045f340, 0xc42045f380, 0xc42045f360, 0x0, 0x0, 0x0, 0x0, 0x146cf60, 0xc424872360, 0xc42010a3eb, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-v1-healing.go:627 +0x38d
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*xlSets).HealObject(0xc420432b40, 0x146cf60, 0xc424872360, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0x0, 0x0, 0x0, ...)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/xl-sets.go:1230 +0x128
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healObject(0xc420738300, 0xc42010a3eb, 0x7, 0xc42984b800, 0x34, 0x0, 0x0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:658 +0xf4
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healBucket(0xc420738300, 0xc42010a3eb, 0x7, 0xc422bb05a0, 0x2)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:635 +0x2ec
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).healBuckets(0xc420738300, 0xc4200a41e0, 0xc42523df00)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:577 +0x157
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).(github.com/minio/minio/cmd.healBuckets)-fm(0xc420738300, 0x0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:523 +0x2a
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).traverseAndHeal.func1(0xc420566bd0)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:516 +0x8f
Aug 16 22:45:51 live-minio-4 minio[11564]: github.com/minio/minio/cmd.(*healSequence).traverseAndHeal(0xc420738300)
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:523 +0xd9
Aug 16 22:45:51 live-minio-4 minio[11564]: created by github.com/minio/minio/cmd.(*healSequence).healSequenceStart
Aug 16 22:45:51 live-minio-4 minio[11564]: #011/q/.q/sources/gopath/src/github.com/minio/minio/cmd/admin-heal-ops.go:466 +0xf0
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Main process exited, code=exited, status=2/INVALIDARGUMENT
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Unit entered failed state.
Aug 16 22:45:51 live-minio-4 systemd[1]: minio.service: Failed with result 'exit-code'.
</code></pre>
</details>
## Your Environment
* Version used (`minio version`): 2018-08-02T23:11:36Z
* Operating System and version (`uname -a`): 4.9.0-6-amd64 #1 SMP Debian 4.9.88-1+deb9u1 (2018-05-07) x86_64 GNU/Linux
| priority | crash unsupported bitrot algorithm unable to heal minio expected behavior i should be able to heal minio bucket current behavior executing mc admin heal recursive json debug live minio ends with minio crashing logs logs aug live minio minio api system aug live minio minio time eest aug live minio minio error unsupported bitrot algorithm aug live minio minio cmd logger logger go logger criticalif aug live minio minio cmd xl metadata go cmd bitrotalgorithm new aug live minio minio cmd erasure go cmd newbitrotverifier aug live minio minio cmd xl healing common go cmd diskswithallparts aug live minio minio cmd xl healing go cmd healobject aug live minio minio cmd xl healing go cmd xlobjects healobject aug live minio minio cmd xl sets go cmd xlsets healobject aug live minio minio cmd admin heal ops go cmd healsequence healobject aug live minio minio cmd admin heal ops go cmd healsequence healbucket aug live minio minio cmd admin heal ops go cmd healsequence healbuckets aug live minio minio cmd admin heal ops go cmd healbuckets fm aug live minio minio cmd admin heal ops go cmd healsequence traverseandheal aug live minio minio cmd admin heal ops go cmd healsequence traverseandheal aug live minio minio panic struct aug live minio minio goroutine aug live minio minio github com minio minio cmd logger criticalif aug live minio minio q q sources gopath src github com minio minio cmd logger logger go aug live minio minio github com minio minio cmd bitrotalgorithm new aug live minio minio q q sources gopath src github com minio minio cmd xl metadata go aug live minio minio github com minio minio cmd newbitrotverifier aug live minio minio q q sources gopath src github com minio minio cmd erasure go aug live minio minio github com minio minio cmd diskswithallparts aug live minio minio q q sources gopath src github com minio minio cmd xl healing common go aug live minio minio github com minio minio cmd healobject aug live minio minio q q sources gopath src github com minio minio cmd xl healing go aug live minio minio github com minio minio cmd xlobjects healobject aug live minio minio q q sources gopath src github com minio minio cmd xl healing go aug live minio minio github com minio minio cmd xlsets healobject aug live minio minio q q sources gopath src github com minio minio cmd xl sets go aug live minio minio github com minio minio cmd healsequence healobject aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio github com minio minio cmd healsequence healbucket aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio github com minio minio cmd healsequence healbuckets aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio github com minio minio cmd healsequence github com minio minio cmd healbuckets fm aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio github com minio minio cmd healsequence traverseandheal aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio github com minio minio cmd healsequence traverseandheal aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio minio created by github com minio minio cmd healsequence healsequencestart aug live minio minio q q sources gopath src github com minio minio cmd admin heal ops go aug live minio systemd minio service main process exited code exited status invalidargument aug live minio systemd minio service unit entered failed state aug live minio systemd minio service failed with result exit code your environment version used minio version operating system and version uname a smp debian gnu linux | 1 |
816,504 | 30,601,258,518 | IssuesEvent | 2023-07-22 12:15:47 | anuthapaliy/Coursework-Planner | https://api.github.com/repos/anuthapaliy/Coursework-Planner | opened | [TECH ED] 🏝️ Stretch challenges | Week 1 🏝️ Priority Stretch 🐂 Size Medium 📅 Node | From Module-Node created by [Dedekind561](https://github.com/Dedekind561): CodeYourFuture/Module-Node#11
### Link to the coursework
https://github.com/CodeYourFuture/Module-Node/edit/main/quote-server/README.md
### Why are we doing this?
These tasks will get you to further develop your skills by implementing more functionality for your server projects.
You can find the stretch section in the README under the 🏝 **Stretch challenge** heading.
### Maximum time in hours
2
### How to get help
Share your blockers in your class channel
https://syllabus.codeyourfuture.io/guides/asking-questions
### How to submit
Follow the instructions on the linked repo | 1.0 | [TECH ED] 🏝️ Stretch challenges - From Module-Node created by [Dedekind561](https://github.com/Dedekind561): CodeYourFuture/Module-Node#11
### Link to the coursework
https://github.com/CodeYourFuture/Module-Node/edit/main/quote-server/README.md
### Why are we doing this?
These tasks will get you to further develop your skills by implementing more functionality for your server projects.
You can find the stretch section in the README under the 🏝 **Stretch challenge** heading.
### Maximum time in hours
2
### How to get help
Share your blockers in your class channel
https://syllabus.codeyourfuture.io/guides/asking-questions
### How to submit
Follow the instructions on the linked repo | priority | 🏝️ stretch challenges from module node created by codeyourfuture module node link to the coursework why are we doing this these tasks will get you to further develop your skills by implementing more functionality for your server projects you can find the stretch section in the readme under the 🏝 stretch challenge heading maximum time in hours how to get help share your blockers in your class channel how to submit follow the instructions on the linked repo | 1 |
120,536 | 4,791,031,282 | IssuesEvent | 2016-10-31 10:59:37 | JKGDevs/JediKnightGalaxies | https://api.github.com/repos/JKGDevs/JediKnightGalaxies | closed | Switching teams makes the client think they have a second starting weapon...that they don't have. | bug priority:medium | Self-explanatory. Switch teams when you have only the starter weapon, and there will be two in your inventory. The second one cannot be equipped.
| 1.0 | Switching teams makes the client think they have a second starting weapon...that they don't have. - Self-explanatory. Switch teams when you have only the starter weapon, and there will be two in your inventory. The second one cannot be equipped.
| priority | switching teams makes the client think they have a second starting weapon that they don t have self explanatory switch teams when you have only the starter weapon and there will be two in your inventory the second one cannot be equipped | 1 |
208,348 | 7,153,261,278 | IssuesEvent | 2018-01-26 00:41:57 | vmware/vic-product | https://api.github.com/repos/vmware/vic-product | closed | Support bundle of integrated OVA should collect Admiral logs | component/ova kind/enhancement priority/medium team/lifecycle triage/proposed-1.4 | Support bundle should collect Admiral container(vic-admiral) logs.
Also it might be useful if logs of the docker which is running in the integrated OVA, to be collected too for more debugging information. | 1.0 | Support bundle of integrated OVA should collect Admiral logs - Support bundle should collect Admiral container(vic-admiral) logs.
Also it might be useful if logs of the docker which is running in the integrated OVA, to be collected too for more debugging information. | priority | support bundle of integrated ova should collect admiral logs support bundle should collect admiral container vic admiral logs also it might be useful if logs of the docker which is running in the integrated ova to be collected too for more debugging information | 1 |
333,710 | 10,130,497,028 | IssuesEvent | 2019-08-01 17:06:04 | SELinuxProject/selinux-kernel | https://api.github.com/repos/SELinuxProject/selinux-kernel | closed | BUG: selinux-testsuite failes on binder tests in v5.1-rc1 | bug priority/medium | When running the selinux-testsuite, the binder tests cause a kernel panic/BUG which causes the test to block.
The test output:
```
Running as user root with context unconfined_u:unconfined_r:unconfined_t
domain_trans/test ........... ok
...
netlink_socket/test ......... ok
prlimit/test ................ ok
binder/test ................. 1/6
<test hang>
```
The relevant console output:
```
[ 823.210062] binder: release 3645:3645 transaction 2 out, still active
[ 823.214047] binder: 3644:3644 transaction failed 29189/0, size 24-8 line 2926
[ 823.218009] binder: send failed reply for transaction 2, target dead
[ 823.221329] binder: 3646:3646 transaction failed 29201/-1, size 24-8 line 3002
[ 823.232432] ------------[ cut here ]------------
[ 823.234746] kernel BUG at drivers/android/binder_alloc.c:1141!
[ 823.237447] invalid opcode: 0000 [#1] SMP PTI
[ 823.239421] CPU: 1 PID: 3644 Comm: test_binder Not tainted 5.1.0-0.rc1.git0.1.2.secnext.fc31.x86_64 #1
[ 823.243538] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[ 823.246079] RIP: 0010:binder_alloc_do_buffer_copy+0x34/0x210
[ 823.248613] Code: 0a 41 55 49 89 fb 41 54 41 89 f4 48 8d 77 38 48 8b 42 58 55 53 48 39 f1 0f 84 17 01 00 00 48 8b 49 58 48 29 c1 49 39 c9 76 02 <0f> 0b 4c 29 c9 49 39 ca 77 f6 41 f6 c2 03 75 f0 0f b6 4a 28 f6 c1
[ 823.256404] RSP: 0018:ffffb04e41093b68 EFLAGS: 00010202
[ 823.258513] RAX: 00007fb600c52000 RBX: a0d48e24a0213e28 RCX: 0000000000000020
[ 823.261375] RDX: ffff9c09b058a9c0 RSI: ffff9c09189165b0 RDI: ffff9c0918916578
[ 823.264225] RBP: ffff9c09b058a9c0 R08: ffffb04e41093c80 R09: 0000000000000028
[ 823.267044] R10: a0d48e24a0213e28 R11: ffff9c0918916578 R12: 0000000000000000
[ 823.269758] R13: ffff9c09b67c9660 R14: ffff9c09b116fb40 R15: ffffffff8acd4d08
[ 823.272482] FS: 00007fbeb3438800(0000) GS:ffff9c09b7a80000(0000) knlGS:0000000000000000
[ 823.275595] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 823.277676] CR2: 000055b102d31cc9 CR3: 0000000234648000 CR4: 00000000001406e0
[ 823.280347] Call Trace:
[ 823.281287] binder_get_object+0x60/0xf0
[ 823.282728] binder_transaction+0xc2e/0x2370
[ 823.284268] ? __check_object_size+0x41/0x15d
[ 823.285849] ? binder_thread_read+0x9e2/0x1460
[ 823.287342] ? binder_update_ref_for_handle+0x83/0x1a0
[ 823.289066] binder_thread_write+0x2ae/0xfc0
[ 823.290513] ? finish_wait+0x80/0x80
[ 823.291729] binder_ioctl+0x659/0x836
[ 823.292980] do_vfs_ioctl+0x40a/0x670
[ 823.294234] ksys_ioctl+0x5e/0x90
[ 823.295364] __x64_sys_ioctl+0x16/0x20
[ 823.296609] do_syscall_64+0x5b/0x150
[ 823.297796] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 823.299423] RIP: 0033:0x7fbeb35e782b
[ 823.300580] Code: 0f 1e fa 48 8b 05 5d 96 0c 00 64 c7 00 26 00 00 00 48 c7 c0 ff ff ff ff c3 66 0f 1f 44 00 00 f3 0f 1e fa b8 10 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 2d 96 0c 00 f7 d8 64 89 01 48
[ 823.306473] RSP: 002b:00007ffdfae2f198 EFLAGS: 00000287 ORIG_RAX: 0000000000000010
[ 823.308868] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fbeb35e782b
[ 823.311029] RDX: 00007ffdfae2f1b0 RSI: 00000000c0306201 RDI: 0000000000000003
[ 823.313206] RBP: 00007ffdfae30210 R08: 00000000010fa330 R09: 0000000000000000
[ 823.315379] R10: 0000000000400644 R11: 0000000000000287 R12: 0000000000401190
[ 823.317459] R13: 00007ffdfae304c0 R14: 0000000000000000 R15: 0000000000000000
[ 823.319510] Modules linked in: crypto_user nfnetlink xt_multiport bluetooth ecdh_generic rfkill sctp overlay ip6table_security xt_CONNSECMARK xt_SECMARK xt_state xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 libcrc32c iptable_security ah6 xfrm6_mode_transport ah4 xfrm4_mode_transport ip6table_mangle ip6table_filter ip6_tables iptable_mangle xt_mark xt_AUDIT ib_isert iscsi_target_mod ib_srpt target_core_mod ib_srp scsi_transport_srp rpcrdma rdma_ucm ib_iser ib_umad ib_ipoib rdma_cm iw_cm libiscsi scsi_transport_iscsi ib_cm mlx5_ib ib_uverbs ib_core sunrpc crct10dif_pclmul crc32_pclmul ghash_clmulni_intel joydev virtio_balloon i2c_piix4 drm_kms_helper virtio_net net_failover failover ttm drm mlx5_core crc32c_intel virtio_blk ata_generic virtio_console mlxfw serio_raw pata_acpi qemu_fw_cfg [last unloaded: arp_tables]
[ 823.339786] ---[ end trace 6f761f654b297775 ]---
```
The related code in Linus' tree (it's the BUG_ON(...) at the top):
```
static void binder_alloc_do_buffer_copy(struct binder_alloc *alloc,
bool to_buffer,
struct binder_buffer *buffer,
binder_size_t buffer_offset,
void *ptr,
size_t bytes)
{
/* All copies must be 32-bit aligned and 32-bit size */
BUG_ON(!check_buffer(alloc, buffer, buffer_offset, bytes));
while (bytes) {
unsigned long size;
struct page *page;
pgoff_t pgoff;
void *tmpptr;
void *base_ptr;
page = binder_alloc_get_page(alloc, buffer,
buffer_offset, &pgoff);
size = min_t(size_t, bytes, PAGE_SIZE - pgoff);
base_ptr = kmap_atomic(page);
tmpptr = base_ptr + pgoff;
if (to_buffer)
memcpy(tmpptr, ptr, size);
else
memcpy(ptr, tmpptr, size);
/*
* kunmap_atomic() takes care of flushing the cache
* if this device has VIVT cache arch
*/
kunmap_atomic(base_ptr);
bytes -= size;
pgoff = 0;
ptr = ptr + size;
buffer_offset += size;
}
}
```
| 1.0 | BUG: selinux-testsuite failes on binder tests in v5.1-rc1 - When running the selinux-testsuite, the binder tests cause a kernel panic/BUG which causes the test to block.
The test output:
```
Running as user root with context unconfined_u:unconfined_r:unconfined_t
domain_trans/test ........... ok
...
netlink_socket/test ......... ok
prlimit/test ................ ok
binder/test ................. 1/6
<test hang>
```
The relevant console output:
```
[ 823.210062] binder: release 3645:3645 transaction 2 out, still active
[ 823.214047] binder: 3644:3644 transaction failed 29189/0, size 24-8 line 2926
[ 823.218009] binder: send failed reply for transaction 2, target dead
[ 823.221329] binder: 3646:3646 transaction failed 29201/-1, size 24-8 line 3002
[ 823.232432] ------------[ cut here ]------------
[ 823.234746] kernel BUG at drivers/android/binder_alloc.c:1141!
[ 823.237447] invalid opcode: 0000 [#1] SMP PTI
[ 823.239421] CPU: 1 PID: 3644 Comm: test_binder Not tainted 5.1.0-0.rc1.git0.1.2.secnext.fc31.x86_64 #1
[ 823.243538] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[ 823.246079] RIP: 0010:binder_alloc_do_buffer_copy+0x34/0x210
[ 823.248613] Code: 0a 41 55 49 89 fb 41 54 41 89 f4 48 8d 77 38 48 8b 42 58 55 53 48 39 f1 0f 84 17 01 00 00 48 8b 49 58 48 29 c1 49 39 c9 76 02 <0f> 0b 4c 29 c9 49 39 ca 77 f6 41 f6 c2 03 75 f0 0f b6 4a 28 f6 c1
[ 823.256404] RSP: 0018:ffffb04e41093b68 EFLAGS: 00010202
[ 823.258513] RAX: 00007fb600c52000 RBX: a0d48e24a0213e28 RCX: 0000000000000020
[ 823.261375] RDX: ffff9c09b058a9c0 RSI: ffff9c09189165b0 RDI: ffff9c0918916578
[ 823.264225] RBP: ffff9c09b058a9c0 R08: ffffb04e41093c80 R09: 0000000000000028
[ 823.267044] R10: a0d48e24a0213e28 R11: ffff9c0918916578 R12: 0000000000000000
[ 823.269758] R13: ffff9c09b67c9660 R14: ffff9c09b116fb40 R15: ffffffff8acd4d08
[ 823.272482] FS: 00007fbeb3438800(0000) GS:ffff9c09b7a80000(0000) knlGS:0000000000000000
[ 823.275595] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 823.277676] CR2: 000055b102d31cc9 CR3: 0000000234648000 CR4: 00000000001406e0
[ 823.280347] Call Trace:
[ 823.281287] binder_get_object+0x60/0xf0
[ 823.282728] binder_transaction+0xc2e/0x2370
[ 823.284268] ? __check_object_size+0x41/0x15d
[ 823.285849] ? binder_thread_read+0x9e2/0x1460
[ 823.287342] ? binder_update_ref_for_handle+0x83/0x1a0
[ 823.289066] binder_thread_write+0x2ae/0xfc0
[ 823.290513] ? finish_wait+0x80/0x80
[ 823.291729] binder_ioctl+0x659/0x836
[ 823.292980] do_vfs_ioctl+0x40a/0x670
[ 823.294234] ksys_ioctl+0x5e/0x90
[ 823.295364] __x64_sys_ioctl+0x16/0x20
[ 823.296609] do_syscall_64+0x5b/0x150
[ 823.297796] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 823.299423] RIP: 0033:0x7fbeb35e782b
[ 823.300580] Code: 0f 1e fa 48 8b 05 5d 96 0c 00 64 c7 00 26 00 00 00 48 c7 c0 ff ff ff ff c3 66 0f 1f 44 00 00 f3 0f 1e fa b8 10 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 2d 96 0c 00 f7 d8 64 89 01 48
[ 823.306473] RSP: 002b:00007ffdfae2f198 EFLAGS: 00000287 ORIG_RAX: 0000000000000010
[ 823.308868] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fbeb35e782b
[ 823.311029] RDX: 00007ffdfae2f1b0 RSI: 00000000c0306201 RDI: 0000000000000003
[ 823.313206] RBP: 00007ffdfae30210 R08: 00000000010fa330 R09: 0000000000000000
[ 823.315379] R10: 0000000000400644 R11: 0000000000000287 R12: 0000000000401190
[ 823.317459] R13: 00007ffdfae304c0 R14: 0000000000000000 R15: 0000000000000000
[ 823.319510] Modules linked in: crypto_user nfnetlink xt_multiport bluetooth ecdh_generic rfkill sctp overlay ip6table_security xt_CONNSECMARK xt_SECMARK xt_state xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 libcrc32c iptable_security ah6 xfrm6_mode_transport ah4 xfrm4_mode_transport ip6table_mangle ip6table_filter ip6_tables iptable_mangle xt_mark xt_AUDIT ib_isert iscsi_target_mod ib_srpt target_core_mod ib_srp scsi_transport_srp rpcrdma rdma_ucm ib_iser ib_umad ib_ipoib rdma_cm iw_cm libiscsi scsi_transport_iscsi ib_cm mlx5_ib ib_uverbs ib_core sunrpc crct10dif_pclmul crc32_pclmul ghash_clmulni_intel joydev virtio_balloon i2c_piix4 drm_kms_helper virtio_net net_failover failover ttm drm mlx5_core crc32c_intel virtio_blk ata_generic virtio_console mlxfw serio_raw pata_acpi qemu_fw_cfg [last unloaded: arp_tables]
[ 823.339786] ---[ end trace 6f761f654b297775 ]---
```
The related code in Linus' tree (it's the BUG_ON(...) at the top):
```
static void binder_alloc_do_buffer_copy(struct binder_alloc *alloc,
bool to_buffer,
struct binder_buffer *buffer,
binder_size_t buffer_offset,
void *ptr,
size_t bytes)
{
/* All copies must be 32-bit aligned and 32-bit size */
BUG_ON(!check_buffer(alloc, buffer, buffer_offset, bytes));
while (bytes) {
unsigned long size;
struct page *page;
pgoff_t pgoff;
void *tmpptr;
void *base_ptr;
page = binder_alloc_get_page(alloc, buffer,
buffer_offset, &pgoff);
size = min_t(size_t, bytes, PAGE_SIZE - pgoff);
base_ptr = kmap_atomic(page);
tmpptr = base_ptr + pgoff;
if (to_buffer)
memcpy(tmpptr, ptr, size);
else
memcpy(ptr, tmpptr, size);
/*
* kunmap_atomic() takes care of flushing the cache
* if this device has VIVT cache arch
*/
kunmap_atomic(base_ptr);
bytes -= size;
pgoff = 0;
ptr = ptr + size;
buffer_offset += size;
}
}
```
| priority | bug selinux testsuite failes on binder tests in when running the selinux testsuite the binder tests cause a kernel panic bug which causes the test to block the test output running as user root with context unconfined u unconfined r unconfined t domain trans test ok netlink socket test ok prlimit test ok binder test the relevant console output binder release transaction out still active binder transaction failed size line binder send failed reply for transaction target dead binder transaction failed size line kernel bug at drivers android binder alloc c invalid opcode smp pti cpu pid comm test binder not tainted secnext hardware name red hat kvm bios rip binder alloc do buffer copy code fb ca rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es call trace binder get object binder transaction check object size binder thread read binder update ref for handle binder thread write finish wait binder ioctl do vfs ioctl ksys ioctl sys ioctl do syscall entry syscall after hwframe rip code fa ff ff ff ff fa ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp modules linked in crypto user nfnetlink xt multiport bluetooth ecdh generic rfkill sctp overlay security xt connsecmark xt secmark xt state xt conntrack nf conntrack nf defrag nf defrag iptable security mode transport mode transport mangle filter tables iptable mangle xt mark xt audit ib isert iscsi target mod ib srpt target core mod ib srp scsi transport srp rpcrdma rdma ucm ib iser ib umad ib ipoib rdma cm iw cm libiscsi scsi transport iscsi ib cm ib ib uverbs ib core sunrpc pclmul pclmul ghash clmulni intel joydev virtio balloon drm kms helper virtio net net failover failover ttm drm core intel virtio blk ata generic virtio console mlxfw serio raw pata acpi qemu fw cfg the related code in linus tree it s the bug on at the top static void binder alloc do buffer copy struct binder alloc alloc bool to buffer struct binder buffer buffer binder size t buffer offset void ptr size t bytes all copies must be bit aligned and bit size bug on check buffer alloc buffer buffer offset bytes while bytes unsigned long size struct page page pgoff t pgoff void tmpptr void base ptr page binder alloc get page alloc buffer buffer offset pgoff size min t size t bytes page size pgoff base ptr kmap atomic page tmpptr base ptr pgoff if to buffer memcpy tmpptr ptr size else memcpy ptr tmpptr size kunmap atomic takes care of flushing the cache if this device has vivt cache arch kunmap atomic base ptr bytes size pgoff ptr ptr size buffer offset size | 1 |
271,937 | 8,494,026,977 | IssuesEvent | 2018-10-28 17:44:14 | nimona/go-nimona | https://api.github.com/repos/nimona/go-nimona | closed | Allow specifying own hostname | Good first issue Priority: Medium Status: In Review Type: Enhancement | Allow for either a `--announce-address` arg or a `NIMONA_ANNOUNCE_ADDRESS` env var to force the local address to be set to its value. This should allow bootstrap nodes to always announce their addresses as their hostnames. | 1.0 | Allow specifying own hostname - Allow for either a `--announce-address` arg or a `NIMONA_ANNOUNCE_ADDRESS` env var to force the local address to be set to its value. This should allow bootstrap nodes to always announce their addresses as their hostnames. | priority | allow specifying own hostname allow for either a announce address arg or a nimona announce address env var to force the local address to be set to its value this should allow bootstrap nodes to always announce their addresses as their hostnames | 1 |
100,425 | 4,087,516,626 | IssuesEvent | 2016-06-01 10:21:19 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | Allow distinct genotypes that differ only by mating type | enhancement genotype_enhancements medium priority user interface | At present, Canto ignores the text in the background box for the purpose of determining whether one genotype differs from another in a session. That's usually OK, because if an allele is important we'll put in the genotype proper. But for mating type, we don't have a convenient single gene to use to add an allele as usual.
What we'd like is for Canto to let us use the most common mating type designations -- i.e. h-, h+, and h90 -- as alleles somehow.
Antonia suggested tickboxes; I wouldn't mind entering them in an interface more like the usual allele-addition setup, maybe with a "mating type" button analogous to the gene buttons. Whatever's easy to code ... my goal is simply to be able to represent mating type in genotypes so we can capture the phenotypes where it makes a difference.
Let me know if I should also open a Chado ticket for anything connected with this.
| 1.0 | Allow distinct genotypes that differ only by mating type - At present, Canto ignores the text in the background box for the purpose of determining whether one genotype differs from another in a session. That's usually OK, because if an allele is important we'll put in the genotype proper. But for mating type, we don't have a convenient single gene to use to add an allele as usual.
What we'd like is for Canto to let us use the most common mating type designations -- i.e. h-, h+, and h90 -- as alleles somehow.
Antonia suggested tickboxes; I wouldn't mind entering them in an interface more like the usual allele-addition setup, maybe with a "mating type" button analogous to the gene buttons. Whatever's easy to code ... my goal is simply to be able to represent mating type in genotypes so we can capture the phenotypes where it makes a difference.
Let me know if I should also open a Chado ticket for anything connected with this.
| priority | allow distinct genotypes that differ only by mating type at present canto ignores the text in the background box for the purpose of determining whether one genotype differs from another in a session that s usually ok because if an allele is important we ll put in the genotype proper but for mating type we don t have a convenient single gene to use to add an allele as usual what we d like is for canto to let us use the most common mating type designations i e h h and as alleles somehow antonia suggested tickboxes i wouldn t mind entering them in an interface more like the usual allele addition setup maybe with a mating type button analogous to the gene buttons whatever s easy to code my goal is simply to be able to represent mating type in genotypes so we can capture the phenotypes where it makes a difference let me know if i should also open a chado ticket for anything connected with this | 1 |
40,423 | 2,868,918,771 | IssuesEvent | 2015-06-05 21:57:35 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | closed | Pub should respect a DART_HOME environment variable | enhancement Priority-Medium wontfix | <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)**
_Originally opened as dart-lang/sdk#3713_
----
Setting a DART_HOME to the binary SDK would allow us to avoid passing in --sdkdir to the pub command | 1.0 | Pub should respect a DART_HOME environment variable - <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)**
_Originally opened as dart-lang/sdk#3713_
----
Setting a DART_HOME to the binary SDK would allow us to avoid passing in --sdkdir to the pub command | priority | pub should respect a dart home environment variable issue by originally opened as dart lang sdk setting a dart home to the binary sdk would allow us to avoid passing in sdkdir to the pub command | 1 |
800,047 | 28,323,833,389 | IssuesEvent | 2023-04-11 05:05:14 | WordPress/openverse | https://api.github.com/repos/WordPress/openverse | opened | Flaky playwright test `visual-regression/pages/pages-single-result.spec.ts:30:13` | 🟨 priority: medium 🛠 goal: fix 🤖 aspect: dx 🧱 stack: frontend | ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
`visual-regression/pages/pages-single-result.spec.ts:30:13 › image ltr single-result page snapshots › screen at breakpoint lg with width 1024 › from search results` is flaky: https://github.com/WordPress/openverse/actions/runs/4663921765/jobs/8255702950?pr=907
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
See the linked failure. | 1.0 | Flaky playwright test `visual-regression/pages/pages-single-result.spec.ts:30:13` - ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
`visual-regression/pages/pages-single-result.spec.ts:30:13 › image ltr single-result page snapshots › screen at breakpoint lg with width 1024 › from search results` is flaky: https://github.com/WordPress/openverse/actions/runs/4663921765/jobs/8255702950?pr=907
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
See the linked failure. | priority | flaky playwright test visual regression pages pages single result spec ts description visual regression pages pages single result spec ts › image ltr single result page snapshots › screen at breakpoint lg with width › from search results is flaky reproduction see the linked failure | 1 |
584,827 | 17,465,276,028 | IssuesEvent | 2021-08-06 15:54:52 | CookieJarApps/SmartCookieWeb | https://api.github.com/repos/CookieJarApps/SmartCookieWeb | closed | ad blocker issue | enhancement stale P2: Medium priority | some website use Ad blocker detection script (notification visible ) , please add specific website ad block enable/disable option
example
https://disableadblock.com/developers/ | 1.0 | ad blocker issue - some website use Ad blocker detection script (notification visible ) , please add specific website ad block enable/disable option
example
https://disableadblock.com/developers/ | priority | ad blocker issue some website use ad blocker detection script notification visible please add specific website ad block enable disable option example | 1 |
607,068 | 18,772,751,721 | IssuesEvent | 2021-11-07 05:21:10 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Windows can exist without low walls beneath them | Type: Bug Priority: 3-Not Required Difficulty: 2-Medium | <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
windows shouldnt be able to float like that, eris' solution is to just have the window fall and break after a delay if the low wall is destroyed first
related to #3730
**Screenshots**

**Additional context**
<!-- Add any other context about the problem here. -->
| 1.0 | Windows can exist without low walls beneath them - <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
windows shouldnt be able to float like that, eris' solution is to just have the window fall and break after a delay if the low wall is destroyed first
related to #3730
**Screenshots**

**Additional context**
<!-- Add any other context about the problem here. -->
| priority | windows can exist without low walls beneath them description windows shouldnt be able to float like that eris solution is to just have the window fall and break after a delay if the low wall is destroyed first related to screenshots additional context | 1 |
631,986 | 20,167,277,157 | IssuesEvent | 2022-02-10 06:37:15 | way-of-elendil/3.3.5 | https://api.github.com/repos/way-of-elendil/3.3.5 | closed | Changeliche selfheal dk | bug type-class priority-medium | **Description**
Quand le dk est sous changeliche il ne peut plus se soigner via le voile mortel
**Étapes de reproduction**
Prenez un dk, utilisez changeliche (avec de la vie en moins), et utilisez voile mortel sur vous même et constater.
**Comportement attendu**
Le voile mortel soigne le dk sous changeliche (c'est un mort vivant)
**Ressources additionnelles**
https://www.wowhead.com/forums/topic/lichborne-deathcoil-macro-147445
| 1.0 | Changeliche selfheal dk - **Description**
Quand le dk est sous changeliche il ne peut plus se soigner via le voile mortel
**Étapes de reproduction**
Prenez un dk, utilisez changeliche (avec de la vie en moins), et utilisez voile mortel sur vous même et constater.
**Comportement attendu**
Le voile mortel soigne le dk sous changeliche (c'est un mort vivant)
**Ressources additionnelles**
https://www.wowhead.com/forums/topic/lichborne-deathcoil-macro-147445
| priority | changeliche selfheal dk description quand le dk est sous changeliche il ne peut plus se soigner via le voile mortel étapes de reproduction prenez un dk utilisez changeliche avec de la vie en moins et utilisez voile mortel sur vous même et constater comportement attendu le voile mortel soigne le dk sous changeliche c est un mort vivant ressources additionnelles | 1 |
204,903 | 7,092,176,109 | IssuesEvent | 2018-01-12 15:42:22 | SmartlyDressedGames/Unturned-4.x-Community | https://api.github.com/repos/SmartlyDressedGames/Unturned-4.x-Community | closed | Preserve Aim Target Between First and Third Person | Priority: Medium Status: Complete Type: Bug | When looking at something and then swapping perspectives you should still be looking at the same thing. | 1.0 | Preserve Aim Target Between First and Third Person - When looking at something and then swapping perspectives you should still be looking at the same thing. | priority | preserve aim target between first and third person when looking at something and then swapping perspectives you should still be looking at the same thing | 1 |
16,942 | 2,615,127,023 | IssuesEvent | 2015-03-01 05:55:33 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Accessing Google Analytics Data (migrating from gdata-java-client to gooogle-api-java-client) | auto-migrated Priority-Medium Type-Sample | ```
Which Google API and version (e.g. Google Calendar Data API version 2)?
Google Analytics API (latest version)
What format (e.g. JSON, Atom)?
JSON
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
ClientLogin or using an API key for programatic access without having to
authorise by users during runtime.
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
External references, such as API reference guide?
Please provide any additional information below.
I am planning to migrate to google data v3 apis. At present using v2.4.
I could not work out a way to access Google Analytics (core reporting data)
using google analytics account login/password or/and API key.
```
Original issue reported on code.google.com by `goss...@gmail.com` on 18 Jun 2012 at 10:17 | 1.0 | Accessing Google Analytics Data (migrating from gdata-java-client to gooogle-api-java-client) - ```
Which Google API and version (e.g. Google Calendar Data API version 2)?
Google Analytics API (latest version)
What format (e.g. JSON, Atom)?
JSON
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?
ClientLogin or using an API key for programatic access without having to
authorise by users during runtime.
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 6
External references, such as API reference guide?
Please provide any additional information below.
I am planning to migrate to google data v3 apis. At present using v2.4.
I could not work out a way to access Google Analytics (core reporting data)
using google analytics account login/password or/and API key.
```
Original issue reported on code.google.com by `goss...@gmail.com` on 18 Jun 2012 at 10:17 | priority | accessing google analytics data migrating from gdata java client to gooogle api java client which google api and version e g google calendar data api version google analytics api latest version what format e g json atom json what authentation e g oauth oauth clientlogin clientlogin or using an api key for programatic access without having to authorise by users during runtime java environment e g java android app engine java external references such as api reference guide please provide any additional information below i am planning to migrate to google data apis at present using i could not work out a way to access google analytics core reporting data using google analytics account login password or and api key original issue reported on code google com by goss gmail com on jun at | 1 |
54,865 | 3,071,456,788 | IssuesEvent | 2015-08-19 12:15:22 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Поддержка грейлинковского формата хранения даты добавления файла в шару "Shared" | enhancement imported Priority-Medium | _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on October 05, 2010 12:31:26_
GreyLink отдает свой файллист с таким полем у файлов:
Shared="129267921732211804"
С точки зрения авторов Грея 'Shared' - это дата добавления файла в шару.
StrongDC++ такого авангардизма не понимает. Хотя ГрейЛинк в листах,
созданных Стронгом (или Флаем), дату добавления в шару показывает.
Можно попросить подружить Стронга с Греевским форматом?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=194_ | 1.0 | Поддержка грейлинковского формата хранения даты добавления файла в шару "Shared" - _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on October 05, 2010 12:31:26_
GreyLink отдает свой файллист с таким полем у файлов:
Shared="129267921732211804"
С точки зрения авторов Грея 'Shared' - это дата добавления файла в шару.
StrongDC++ такого авангардизма не понимает. Хотя ГрейЛинк в листах,
созданных Стронгом (или Флаем), дату добавления в шару показывает.
Можно попросить подружить Стронга с Греевским форматом?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=194_ | priority | поддержка грейлинковского формата хранения даты добавления файла в шару shared from on october greylink отдает свой файллист с таким полем у файлов shared с точки зрения авторов грея shared это дата добавления файла в шару strongdc такого авангардизма не понимает хотя грейлинк в листах созданных стронгом или флаем дату добавления в шару показывает можно попросить подружить стронга с греевским форматом original issue | 1 |
170,700 | 6,469,609,615 | IssuesEvent | 2017-08-17 06:37:45 | powermock/powermock | https://api.github.com/repos/powermock/powermock | closed | Typesafe reflection | enhancement Priority:Medium | _From [jan.kron...@gmail.com](https://code.google.com/u/111289048178412145391/) on November 14, 2008 18:40:19_
Refactor WhiteBox to support the following. Maybe even create separate project?
createPartialMock(SomeClass.class, methodCalled("qwe"), methodCalled("asd"));
createPartialMock(SomeClass.class, methodReturning(Qwe.class),
methodCalled("asd"));
createPartialMock(SomeClass.class, methodReturning(Qwe.class));
createPartialMock(SomeClass.class, withArguments(Qwe.class, Arg.class));
invoke(methodReturning(Qwe.class))
invoke(privateMethodReturning(Qwe.class))
invoke(methodAnnotated(Qwe.class))
_Original issue: http://code.google.com/p/powermock/issues/detail?id=76_
| 1.0 | Typesafe reflection - _From [jan.kron...@gmail.com](https://code.google.com/u/111289048178412145391/) on November 14, 2008 18:40:19_
Refactor WhiteBox to support the following. Maybe even create separate project?
createPartialMock(SomeClass.class, methodCalled("qwe"), methodCalled("asd"));
createPartialMock(SomeClass.class, methodReturning(Qwe.class),
methodCalled("asd"));
createPartialMock(SomeClass.class, methodReturning(Qwe.class));
createPartialMock(SomeClass.class, withArguments(Qwe.class, Arg.class));
invoke(methodReturning(Qwe.class))
invoke(privateMethodReturning(Qwe.class))
invoke(methodAnnotated(Qwe.class))
_Original issue: http://code.google.com/p/powermock/issues/detail?id=76_
| priority | typesafe reflection from on november refactor whitebox to support the following maybe even create separate project createpartialmock someclass class methodcalled qwe methodcalled asd createpartialmock someclass class methodreturning qwe class methodcalled asd createpartialmock someclass class methodreturning qwe class createpartialmock someclass class witharguments qwe class arg class invoke methodreturning qwe class invoke privatemethodreturning qwe class invoke methodannotated qwe class original issue | 1 |
778,107 | 27,304,143,266 | IssuesEvent | 2023-02-24 06:21:57 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [CDCSDK] Clean up unused header_schema and header_schema_version fields from ReadOpsResult | kind/enhancement priority/medium area/cdcsdk | Jira Link: [DB-5215](https://yugabyte.atlassian.net/browse/DB-5215)
### Description
These fields were added in https://github.com/yugabyte/yugabyte-db/commit/13719443684b29bd7ddb779d7ba63e42c81f1e82 but are not used anywhere. The task is to clean up these fields.
[DB-5215]: https://yugabyte.atlassian.net/browse/DB-5215?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [CDCSDK] Clean up unused header_schema and header_schema_version fields from ReadOpsResult - Jira Link: [DB-5215](https://yugabyte.atlassian.net/browse/DB-5215)
### Description
These fields were added in https://github.com/yugabyte/yugabyte-db/commit/13719443684b29bd7ddb779d7ba63e42c81f1e82 but are not used anywhere. The task is to clean up these fields.
[DB-5215]: https://yugabyte.atlassian.net/browse/DB-5215?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | clean up unused header schema and header schema version fields from readopsresult jira link description these fields were added in but are not used anywhere the task is to clean up these fields | 1 |
314,703 | 9,601,758,268 | IssuesEvent | 2019-05-10 13:01:42 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :198389]Memory - illegal accesses in /subsys/settings/src/settings_runtime.c | Coverity area: Settings bug priority: medium | Static code scan issues seen in File: /subsys/settings/src/settings_runtime.c
Category: Memory - illegal accesses
Function: settings_runtime_commit
Component: Other
CID: 198389
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996 | 1.0 | [Coverity CID :198389]Memory - illegal accesses in /subsys/settings/src/settings_runtime.c - Static code scan issues seen in File: /subsys/settings/src/settings_runtime.c
Category: Memory - illegal accesses
Function: settings_runtime_commit
Component: Other
CID: 198389
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996 | priority | memory illegal accesses in subsys settings src settings runtime c static code scan issues seen in file subsys settings src settings runtime c category memory illegal accesses function settings runtime commit component other cid please fix or provide comments to square it off in coverity in the link | 1 |
769,271 | 26,999,147,814 | IssuesEvent | 2023-02-10 05:40:18 | Fiserv/Support | https://api.github.com/repos/Fiserv/Support | closed | Changes not published | bug Priority - Medium Severity - Low MerchantAcquiringLATAM | # Reporting new issue for Merchant Acquiring - LATAM
**Region** (if applicable)
LATAM
**Page**
https://dev-developerstudio.fiserv.com/product/MerchantAcquiringLATAM/docs/?path=docs/espa%C3%B1ol/pagos/3-5-3d-secure.md&branch=develop
**Describe the bug**
Made some modifications but not published since yesterday morning. I checked yaml and markdown validator and everything it's correct.
**To Reproduce**
Steps to reproduce the behavior:
**Expected behavior**
Visualize my changes
**Screenshots**

**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Tablet (please complete the following information):**
- Device: [e.g. iPad]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
<!-- ** Recommended **
- Assign a Bug
On you right hand side, you can set Assignees by Clicking setting button next to it.
- Subscribe to a Team
At the end of your Bug template, please mention team you would like to subscribe for this Issue for future updates. You can mention teams in Github
Example: @Fiserv/fiserv-developer)
-->
| 1.0 | Changes not published - # Reporting new issue for Merchant Acquiring - LATAM
**Region** (if applicable)
LATAM
**Page**
https://dev-developerstudio.fiserv.com/product/MerchantAcquiringLATAM/docs/?path=docs/espa%C3%B1ol/pagos/3-5-3d-secure.md&branch=develop
**Describe the bug**
Made some modifications but not published since yesterday morning. I checked yaml and markdown validator and everything it's correct.
**To Reproduce**
Steps to reproduce the behavior:
**Expected behavior**
Visualize my changes
**Screenshots**

**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Tablet (please complete the following information):**
- Device: [e.g. iPad]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
<!-- ** Recommended **
- Assign a Bug
On you right hand side, you can set Assignees by Clicking setting button next to it.
- Subscribe to a Team
At the end of your Bug template, please mention team you would like to subscribe for this Issue for future updates. You can mention teams in Github
Example: @Fiserv/fiserv-developer)
-->
| priority | changes not published reporting new issue for merchant acquiring latam region if applicable latam page describe the bug made some modifications but not published since yesterday morning i checked yaml and markdown validator and everything it s correct to reproduce steps to reproduce the behavior expected behavior visualize my changes screenshots desktop please complete the following information os browser version tablet please complete the following information device os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here recommended assign a bug on you right hand side you can set assignees by clicking setting button next to it subscribe to a team at the end of your bug template please mention team you would like to subscribe for this issue for future updates you can mention teams in github example fiserv fiserv developer | 1 |
801,486 | 28,490,535,416 | IssuesEvent | 2023-04-18 10:54:36 | OpenBioML/chemnlp | https://api.github.com/repos/OpenBioML/chemnlp | opened | Investigate learning rate schedulers for training | work package: model training investigation priority: medium | Look into [default learning rate scheduler](https://huggingface.co/docs/transformers/v4.27.2/en/main_classes/optimizer_schedules#transformers.get_linear_schedule_with_warmup) provided by the Hugging Face `Trainer` and decide if we want to use it.
* Suggestions of alternatives from the literature would be welcome | 1.0 | Investigate learning rate schedulers for training - Look into [default learning rate scheduler](https://huggingface.co/docs/transformers/v4.27.2/en/main_classes/optimizer_schedules#transformers.get_linear_schedule_with_warmup) provided by the Hugging Face `Trainer` and decide if we want to use it.
* Suggestions of alternatives from the literature would be welcome | priority | investigate learning rate schedulers for training look into provided by the hugging face trainer and decide if we want to use it suggestions of alternatives from the literature would be welcome | 1 |
711,451 | 24,464,634,400 | IssuesEvent | 2022-10-07 14:03:45 | AgnostiqHQ/covalent | https://api.github.com/repos/AgnostiqHQ/covalent | closed | Remove _get_node_error from Result | priority / medium | `_get_node_error` is an internal getter in `Result` which connects to a DB to return a value. It is called once in execution.py. This must be removed from `Result` since client side object can't be aware of the DB. execution.py needs it own logic to get the value from the DB.
*Acceptance Criteria*
- [ ] `_get_node_error` is removed from `Result`
- [ ] logic added in execution.py to get error value from DB | 1.0 | Remove _get_node_error from Result - `_get_node_error` is an internal getter in `Result` which connects to a DB to return a value. It is called once in execution.py. This must be removed from `Result` since client side object can't be aware of the DB. execution.py needs it own logic to get the value from the DB.
*Acceptance Criteria*
- [ ] `_get_node_error` is removed from `Result`
- [ ] logic added in execution.py to get error value from DB | priority | remove get node error from result get node error is an internal getter in result which connects to a db to return a value it is called once in execution py this must be removed from result since client side object can t be aware of the db execution py needs it own logic to get the value from the db acceptance criteria get node error is removed from result logic added in execution py to get error value from db | 1 |
87,310 | 3,744,812,201 | IssuesEvent | 2016-03-10 04:21:27 | cs2103jan2016-t11-4c/main | https://api.github.com/repos/cs2103jan2016-t11-4c/main | opened | Create higher level Memory class from which Storage and Settings classes will inherit | priority.medium type.enhancement | because storage and settings are found to share some data information e.g. file destinations and methods e.g. validity of .txt files | 1.0 | Create higher level Memory class from which Storage and Settings classes will inherit - because storage and settings are found to share some data information e.g. file destinations and methods e.g. validity of .txt files | priority | create higher level memory class from which storage and settings classes will inherit because storage and settings are found to share some data information e g file destinations and methods e g validity of txt files | 1 |
513,112 | 14,916,215,422 | IssuesEvent | 2021-01-22 17:51:47 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Make global menu screens UI consistent. | enhancement priority: medium | There are several small differences in the design of the global menu screens.
Ping me to discuss them and solution.
| 1.0 | [studio-ui] Make global menu screens UI consistent. - There are several small differences in the design of the global menu screens.
Ping me to discuss them and solution.
| priority | make global menu screens ui consistent there are several small differences in the design of the global menu screens ping me to discuss them and solution | 1 |
688,695 | 23,593,022,019 | IssuesEvent | 2022-08-23 16:44:04 | wasmerio/wasmer | https://api.github.com/repos/wasmerio/wasmer | closed | Improve Wasmer 3.0 migration example | 🎉 enhancement priority-medium | We can improve our migration guide by:
* [x] Having an example from WasmerEnv to FunctionEnv
* [ ] Adding the new memory view API (and show how people can update from previous usage)
| 1.0 | Improve Wasmer 3.0 migration example - We can improve our migration guide by:
* [x] Having an example from WasmerEnv to FunctionEnv
* [ ] Adding the new memory view API (and show how people can update from previous usage)
| priority | improve wasmer migration example we can improve our migration guide by having an example from wasmerenv to functionenv adding the new memory view api and show how people can update from previous usage | 1 |
800,819 | 28,433,998,728 | IssuesEvent | 2023-04-15 04:44:05 | KDT3-Final-6/final-project-BE | https://api.github.com/repos/KDT3-Final-6/final-project-BE | closed | refactor: DTO 수정 | Type: Refactor Status: Completed Priority: Medium For: API For: Backend | ## Description
겹치는 DTO들을 수정함
## Tasks(Process)
- [x] Order DTO 수정
- [x] QnA DTO 수정
- [x] Review DTO 수정
## References
| 1.0 | refactor: DTO 수정 - ## Description
겹치는 DTO들을 수정함
## Tasks(Process)
- [x] Order DTO 수정
- [x] QnA DTO 수정
- [x] Review DTO 수정
## References
| priority | refactor dto 수정 description 겹치는 dto들을 수정함 tasks process order dto 수정 qna dto 수정 review dto 수정 references | 1 |
724,640 | 24,937,365,725 | IssuesEvent | 2022-10-31 16:06:00 | grafana/k6-docs | https://api.github.com/repos/grafana/k6-docs | closed | Option 'throw' is broken or badly documented | Area: Community Content Priority: Medium Status: Triage Type: Missing Content | ## Expected behaviour
Enabling [throw](https://docs.k6.io/docs/options) exits the specific test iteration as soon as an request error occurs: connection issues, status 4xx, status 5xxx.
I hoped that with this option I wouldn't have to write `check`s for every request.
## Actual behaviour
Status errors don't exit the iteration, or show anything in the log.
## Documented behaviour
A boolean, true or false, specifying whether to throw errors on failed HTTP requests or not. | 1.0 | Option 'throw' is broken or badly documented - ## Expected behaviour
Enabling [throw](https://docs.k6.io/docs/options) exits the specific test iteration as soon as an request error occurs: connection issues, status 4xx, status 5xxx.
I hoped that with this option I wouldn't have to write `check`s for every request.
## Actual behaviour
Status errors don't exit the iteration, or show anything in the log.
## Documented behaviour
A boolean, true or false, specifying whether to throw errors on failed HTTP requests or not. | priority | option throw is broken or badly documented expected behaviour enabling exits the specific test iteration as soon as an request error occurs connection issues status status i hoped that with this option i wouldn t have to write check s for every request actual behaviour status errors don t exit the iteration or show anything in the log documented behaviour a boolean true or false specifying whether to throw errors on failed http requests or not | 1 |
7,570 | 2,603,620,877 | IssuesEvent | 2015-02-24 17:01:08 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | Create a way of logging requests to the ckan API for monitoring purposes | enhancement Priority-Medium | @luiscape has sent a query to the ckan dev mailing list to see if there are any existing solutions. Once we have that input, we will decide exactly what to implement. | 1.0 | Create a way of logging requests to the ckan API for monitoring purposes - @luiscape has sent a query to the ckan dev mailing list to see if there are any existing solutions. Once we have that input, we will decide exactly what to implement. | priority | create a way of logging requests to the ckan api for monitoring purposes luiscape has sent a query to the ckan dev mailing list to see if there are any existing solutions once we have that input we will decide exactly what to implement | 1 |
270,268 | 8,453,811,218 | IssuesEvent | 2018-10-20 19:16:12 | orange-alliance/the-orange-alliance | https://api.github.com/repos/orange-alliance/the-orange-alliance | opened | API Returns teams in wrong order | API Bug Medium Priority bug | The API returns the teams ordered wrong, it will show every number starting with 1, then 2, the three. Ex:
1
11
100
1000
2
22
20000
etc.
This kinda looks unprofessional. | 1.0 | API Returns teams in wrong order - The API returns the teams ordered wrong, it will show every number starting with 1, then 2, the three. Ex:
1
11
100
1000
2
22
20000
etc.
This kinda looks unprofessional. | priority | api returns teams in wrong order the api returns the teams ordered wrong it will show every number starting with then the three ex etc this kinda looks unprofessional | 1 |
579,847 | 17,199,189,772 | IssuesEvent | 2021-07-16 23:34:39 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | reopened | Busting directory is not cleared during the cache clearing | module: minify CSS module: minify JS needs: grooming priority: medium severity: moderate type: bug | **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
We're not removing the `/wp-content/cache/busting/1/wp-content/` content when clearing the cache from the top menu neither during deactivation.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the page when the dynamic CSS file is added (dynamic_css.php, etc.)
2. Confirm that the associated file is created in the busting directory
3. Clear the cache, deactivate WP Rocket
4. Confirm that the cache is not cleared
**Expected behavior**
We should be able to clear this cache at the correct moment, including the `.gz` files which are not listed in the `$extensions` array here:
https://github.com/wp-media/wp-rocket/blob/8c02750b89412eb7e395551c150ab06e4a4b8e2b/inc/functions/files.php#L401
**Screenshots**

**Additional context**
We should think when the busting directory should be cleared exactly.
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | Busting directory is not cleared during the cache clearing - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
We're not removing the `/wp-content/cache/busting/1/wp-content/` content when clearing the cache from the top menu neither during deactivation.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the page when the dynamic CSS file is added (dynamic_css.php, etc.)
2. Confirm that the associated file is created in the busting directory
3. Clear the cache, deactivate WP Rocket
4. Confirm that the cache is not cleared
**Expected behavior**
We should be able to clear this cache at the correct moment, including the `.gz` files which are not listed in the `$extensions` array here:
https://github.com/wp-media/wp-rocket/blob/8c02750b89412eb7e395551c150ab06e4a4b8e2b/inc/functions/files.php#L401
**Screenshots**

**Additional context**
We should think when the busting directory should be cleared exactly.
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| priority | busting directory is not cleared during the cache clearing before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug we re not removing the wp content cache busting wp content content when clearing the cache from the top menu neither during deactivation to reproduce steps to reproduce the behavior visit the page when the dynamic css file is added dynamic css php etc confirm that the associated file is created in the busting directory clear the cache deactivate wp rocket confirm that the cache is not cleared expected behavior we should be able to clear this cache at the correct moment including the gz files which are not listed in the extensions array here screenshots additional context we should think when the busting directory should be cleared exactly backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
76,404 | 3,487,798,418 | IssuesEvent | 2016-01-02 09:08:52 | bounswe/bounswe2015group8 | https://api.github.com/repos/bounswe/bounswe2015group8 | closed | Edit Post Back-end Logic | Priority-Medium Type-Task | Implement the back-end logic of editing a post.
Only the original owner of the post should be able to edit it. | 1.0 | Edit Post Back-end Logic - Implement the back-end logic of editing a post.
Only the original owner of the post should be able to edit it. | priority | edit post back end logic implement the back end logic of editing a post only the original owner of the post should be able to edit it | 1 |
455,421 | 13,126,538,154 | IssuesEvent | 2020-08-06 08:45:44 | ayumi-cloud/oc2-security-module | https://api.github.com/repos/ayumi-cloud/oc2-security-module | opened | Add Linkpad bot to firewall | Add to Blacklist Add to Whitelist Firewall Priority: Medium enhancement in-progress | ### Enhancement idea
- [ ] Add Linkpad bot to firewall.
- [ ] Add to whitelists.
- [ ] Add to blacklists.
- [ ] Add option in settings.
- [ ] Add filters.
- { ] Add labelling system.
User agent example:
```
Mozilla/5.0 (compatible; LinkpadBot/2.3; +http://linkpad.org/robot/)
```
| 1.0 | Add Linkpad bot to firewall - ### Enhancement idea
- [ ] Add Linkpad bot to firewall.
- [ ] Add to whitelists.
- [ ] Add to blacklists.
- [ ] Add option in settings.
- [ ] Add filters.
- { ] Add labelling system.
User agent example:
```
Mozilla/5.0 (compatible; LinkpadBot/2.3; +http://linkpad.org/robot/)
```
| priority | add linkpad bot to firewall enhancement idea add linkpad bot to firewall add to whitelists add to blacklists add option in settings add filters add labelling system user agent example mozilla compatible linkpadbot | 1 |
202,948 | 7,056,676,367 | IssuesEvent | 2018-01-04 13:46:22 | getcanoe/canoe | https://api.github.com/repos/getcanoe/canoe | opened | Implement i18n | good first issue help wanted Priority: Medium Type: Question | Currently the Copay fork uses Crowdin as a translation management system. Unfortunately Crowdin is too expensive for our project. We need to find a cheaper solution (preferably free.)
Suggestions welcome! | 1.0 | Implement i18n - Currently the Copay fork uses Crowdin as a translation management system. Unfortunately Crowdin is too expensive for our project. We need to find a cheaper solution (preferably free.)
Suggestions welcome! | priority | implement currently the copay fork uses crowdin as a translation management system unfortunately crowdin is too expensive for our project we need to find a cheaper solution preferably free suggestions welcome | 1 |
33,837 | 2,772,886,825 | IssuesEvent | 2015-05-03 03:45:48 | QuiteRSS/quiterss | https://api.github.com/repos/QuiteRSS/quiterss | closed | Замораживание приложения (проблема sqlite) | auto-migrated bug Priority-Medium | ```
При добавлении нескольких лент с длинными
ссылками приложение перестаёт отвечать.
Проблема с sqlite - решения пока нет.
Пример ссылок:
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%A4%D0%B
5%D0%B4%D0%B5%D1%80%D0%B0%D0%BB%D1%8C%D0%BD%D0%BE%D0%B5+%D0%B3%D0%BE%D1%81%D1%83
%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+%D0%B1%D1%8E%
D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%D0%B6%D0%B4%D
0%B5%D0%BD%D0%B8%D0%B5+%22%D0%A1%D0%B5%D0%B2%D0%B5%D1%80%D0%BE-%D0%9A%D0%B0%D0%B
2%D0%BA%D0%B0%D0%B7%D1%81%D0%BA%D0%B8%D0%B9+%D0%BC%D0%BD%D0%BE%D0%B3%D0%BE%D0%BF
%D1%80%D0%BE%D1%84%D0%B8%D0%BB%D1%8C%D0%BD%D1%8B%D0%B9+%D0%BC%D0%B5%D0%B4%D0%B8%
D1%86%D0%B8%D0%BD%D1%81%D0%BA%D0%B8%D0%B9+%D1%86%D0%B5%D0%BD%D1%82%D1%80%22+%D0%
9C%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82%D0%B2%D0%B0+%D0%B7%D0%B
4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+
%D0%B8+%D1%81%D0%BE%D1%86%D0%B8%D0%B0%D0%BB%D1%8C%D0%BD%D0%BE%D0%B3%D0%BE+%D1%80
%D0%B0%D0%B7%D0%B2%D0%B8%D1%82%D0%B8%D1%8F+%D0%A0%D0%BE%D1%81%D1%81%D0%B8%D0%B9%
D1%81%D0%BA%D0%BE%D0%B9+%D0%A4%D0%B5%D0%B4%D0%B5%D1%80%D0%B0%D1%86%D0%B8%D0%B8+(
%D0%B3.%D0%91%D0%B5%D1%81%D0%BB%D0%B0%D0%BD
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%93%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%A0%D0%B5%D1%81%D0%BF%D1%83
%D0%B1%D0%BB%D0%B8%D0%BA%D0%B0%D0%BD%D1%81%D0%BA%D0%B0%D1%8F+%D0%BA%D0%BB%D0%B8%
D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+%D0%B1%D0%BE%D0%BB%D1%8C%D0%BD%D
0%B8%D1%86%D0%B0%22+%D0%9C%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82
%D0%B2%D0%B0+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%
D0%B5%D0%BD%D0%B8%D1%8F+%D0%9A%D0%91%D0%A0&customer.code=03042000029&customer.fz
94id=639454&customer.Inn=%3Cb%3E0711035586%3C%2Fb%3E&_custLev=on&_custLev=on&_cu
stLev=on&_custLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=
on&_orderStages=on&_orderStages=on&morphology=false&strictEqual=false
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%93%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%A1%D0%B0%D0%BC%D0%B0%D1%80
%D1%81%D0%BA%D0%B0%D1%8F+%D0%BE%D0%B1%D0%BB%D0%B0%D1%81%D1%82%D0%BD%D0%B0%D1%8F+
%D0%BA%D0%BB%D0%B8%D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+%D0%B1%D0%BE%
D0%BB%D1%8C%D0%BD%D0%B8%D1%86%D0%B0+%D0%B8%D0%BC.%D0%9C.%D0%98.%D0%9A%D0%B0%D0%B
B%D0%B8%D0%BD%D0%B8%D0%BD%D0%B0%22&customer.code=03422000065&customer.fz94id=675
529&customer.Inn=%3Cb%3E6312023131%3C%2Fb%3E&_custLev=on&_custLev=on&_custLev=on
&_custLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=on&_orde
rStages=on&_orderStages=on&morphology=false&strictEqual=false
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%B3%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%9A%D1%80%D0%B0%D0%B5%D0%B2
%D0%B0%D1%8F+%D0%BA%D0%BB%D0%B8%D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+
%D0%B1%D0%BE%D0%BB%D1%8C%D0%BD%D0%B8%D1%86%D0%B0+%E2%84%96+1+%D0%B8%D0%BC%D0%B5%
D0%BD%D0%B8+%D0%BF%D1%80%D0%BE%D1%84%D0%B5%D1%81%D1%81%D0%BE%D1%80%D0%B0+%D0%A1.
%D0%92.%D0%9E%D1%87%D0%B0%D0%BF%D0%BE%D0%B2%D1%81%D0%BA%D0%BE%D0%B3%D0%BE%22+%D0
%BC%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82%D0%B2%D0%B0+%D0%B7%D0%
B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F
+%D0%9A%D1%80%D0%B0%D1%81%D0%BD%D0%BE%D0%B4%D0%B0%D1%80%D1%81%D0%BA%D0%BE%D0%B3%
D0%BE+%D0%BA%D1%80%D0%B0%D1%8F&customer.code=03182000656&customer.fz94id=689156&
customer.Inn=%3Cb%3E2311040088%3C%2Fb%3E&_custLev=on&_custLev=on&_custLev=on&_cu
stLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=on&_orderSta
ges=on&_orderStages=on&morphology=false&strictEqual=false
```
Original issue reported on code.google.com by `egor.shi...@gmail.com` on 30 Jan 2015 at 7:35 | 1.0 | Замораживание приложения (проблема sqlite) - ```
При добавлении нескольких лент с длинными
ссылками приложение перестаёт отвечать.
Проблема с sqlite - решения пока нет.
Пример ссылок:
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%A4%D0%B
5%D0%B4%D0%B5%D1%80%D0%B0%D0%BB%D1%8C%D0%BD%D0%BE%D0%B5+%D0%B3%D0%BE%D1%81%D1%83
%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+%D0%B1%D1%8E%
D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%D0%B6%D0%B4%D
0%B5%D0%BD%D0%B8%D0%B5+%22%D0%A1%D0%B5%D0%B2%D0%B5%D1%80%D0%BE-%D0%9A%D0%B0%D0%B
2%D0%BA%D0%B0%D0%B7%D1%81%D0%BA%D0%B8%D0%B9+%D0%BC%D0%BD%D0%BE%D0%B3%D0%BE%D0%BF
%D1%80%D0%BE%D1%84%D0%B8%D0%BB%D1%8C%D0%BD%D1%8B%D0%B9+%D0%BC%D0%B5%D0%B4%D0%B8%
D1%86%D0%B8%D0%BD%D1%81%D0%BA%D0%B8%D0%B9+%D1%86%D0%B5%D0%BD%D1%82%D1%80%22+%D0%
9C%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82%D0%B2%D0%B0+%D0%B7%D0%B
4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+
%D0%B8+%D1%81%D0%BE%D1%86%D0%B8%D0%B0%D0%BB%D1%8C%D0%BD%D0%BE%D0%B3%D0%BE+%D1%80
%D0%B0%D0%B7%D0%B2%D0%B8%D1%82%D0%B8%D1%8F+%D0%A0%D0%BE%D1%81%D1%81%D0%B8%D0%B9%
D1%81%D0%BA%D0%BE%D0%B9+%D0%A4%D0%B5%D0%B4%D0%B5%D1%80%D0%B0%D1%86%D0%B8%D0%B8+(
%D0%B3.%D0%91%D0%B5%D1%81%D0%BB%D0%B0%D0%BD
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%93%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%A0%D0%B5%D1%81%D0%BF%D1%83
%D0%B1%D0%BB%D0%B8%D0%BA%D0%B0%D0%BD%D1%81%D0%BA%D0%B0%D1%8F+%D0%BA%D0%BB%D0%B8%
D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+%D0%B1%D0%BE%D0%BB%D1%8C%D0%BD%D
0%B8%D1%86%D0%B0%22+%D0%9C%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82
%D0%B2%D0%B0+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%
D0%B5%D0%BD%D0%B8%D1%8F+%D0%9A%D0%91%D0%A0&customer.code=03042000029&customer.fz
94id=639454&customer.Inn=%3Cb%3E0711035586%3C%2Fb%3E&_custLev=on&_custLev=on&_cu
stLev=on&_custLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=
on&_orderStages=on&_orderStages=on&morphology=false&strictEqual=false
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%93%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%A1%D0%B0%D0%BC%D0%B0%D1%80
%D1%81%D0%BA%D0%B0%D1%8F+%D0%BE%D0%B1%D0%BB%D0%B0%D1%81%D1%82%D0%BD%D0%B0%D1%8F+
%D0%BA%D0%BB%D0%B8%D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+%D0%B1%D0%BE%
D0%BB%D1%8C%D0%BD%D0%B8%D1%86%D0%B0+%D0%B8%D0%BC.%D0%9C.%D0%98.%D0%9A%D0%B0%D0%B
B%D0%B8%D0%BD%D0%B8%D0%BD%D0%B0%22&customer.code=03422000065&customer.fz94id=675
529&customer.Inn=%3Cb%3E6312023131%3C%2Fb%3E&_custLev=on&_custLev=on&_custLev=on
&_custLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=on&_orde
rStages=on&_orderStages=on&morphology=false&strictEqual=false
http://zakupki.gov.ru/epz/order/extendedsearch/rss?_aTypes=on&_aTypes=on&_aTypes
=on&placeOfSearch=FZ_44&_placeOfSearch=on&placeOfSearch=FZ_223&_placeOfSearch=on
&placeOfSearch=FZ_94&_placeOfSearch=on&_purchaseMethodList=on&_purchaseMethodLis
t=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purch
aseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodLi
st=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purc
haseMethodList=on&_purchaseMethodList=on&_purchaseMethodList=on&_purchaseMethodL
ist=on&_purchaseMethodList=on&orderPriceCurrencyId=-1&customer.title=%D0%B3%D0%B
E%D1%81%D1%83%D0%B4%D0%B0%D1%80%D1%81%D1%82%D0%B2%D0%B5%D0%BD%D0%BD%D0%BE%D0%B5+
%D0%B1%D1%8E%D0%B4%D0%B6%D0%B5%D1%82%D0%BD%D0%BE%D0%B5+%D1%83%D1%87%D1%80%D0%B5%
D0%B6%D0%B4%D0%B5%D0%BD%D0%B8%D0%B5+%D0%B7%D0%B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D
1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F+%22%D0%9A%D1%80%D0%B0%D0%B5%D0%B2
%D0%B0%D1%8F+%D0%BA%D0%BB%D0%B8%D0%BD%D0%B8%D1%87%D0%B5%D1%81%D0%BA%D0%B0%D1%8F+
%D0%B1%D0%BE%D0%BB%D1%8C%D0%BD%D0%B8%D1%86%D0%B0+%E2%84%96+1+%D0%B8%D0%BC%D0%B5%
D0%BD%D0%B8+%D0%BF%D1%80%D0%BE%D1%84%D0%B5%D1%81%D1%81%D0%BE%D1%80%D0%B0+%D0%A1.
%D0%92.%D0%9E%D1%87%D0%B0%D0%BF%D0%BE%D0%B2%D1%81%D0%BA%D0%BE%D0%B3%D0%BE%22+%D0
%BC%D0%B8%D0%BD%D0%B8%D1%81%D1%82%D0%B5%D1%80%D1%81%D1%82%D0%B2%D0%B0+%D0%B7%D0%
B4%D1%80%D0%B0%D0%B2%D0%BE%D0%BE%D1%85%D1%80%D0%B0%D0%BD%D0%B5%D0%BD%D0%B8%D1%8F
+%D0%9A%D1%80%D0%B0%D1%81%D0%BD%D0%BE%D0%B4%D0%B0%D1%80%D1%81%D0%BA%D0%BE%D0%B3%
D0%BE+%D0%BA%D1%80%D0%B0%D1%8F&customer.code=03182000656&customer.fz94id=689156&
customer.Inn=%3Cb%3E2311040088%3C%2Fb%3E&_custLev=on&_custLev=on&_custLev=on&_cu
stLev=on&orderStages=AF&_orderStages=on&orderStages=CA&_orderStages=on&_orderSta
ges=on&_orderStages=on&morphology=false&strictEqual=false
```
Original issue reported on code.google.com by `egor.shi...@gmail.com` on 30 Jan 2015 at 7:35 | priority | замораживание приложения проблема sqlite при добавлении нескольких лент с длинными ссылками приложение перестаёт отвечать проблема с sqlite решения пока нет пример ссылок on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on purchasemethodlist on purchasemethodlis t on purchasemethodlist on purchasemethodlist on purchasemethodlist on purch asemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodli st on purchasemethodlist on purchasemethodlist on purchasemethodlist on purc hasemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodl ist on purchasemethodlist on orderpricecurrencyid customer title b bb bd be be bd bd be bd be d bd be b ba ba bc bd be be bf be bb bd bc bd ba bd bd b be be bd bd be bb bd be be be ba be bb bd on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on purchasemethodlist on purchasemethodlis t on purchasemethodlist on purchasemethodlist on purchasemethodlist on purch asemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodli st on purchasemethodlist on purchasemethodlist on purchasemethodlist on purc hasemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodl ist on purchasemethodlist on orderpricecurrencyid customer title b e bd bd be bd be bd be be d bd bd bf bb ba bd ba ba bb bd ba be bb bd d bd be be bd bd customer code customer fz customer inn custlev on custlev on cu stlev on custlev on orderstages af orderstages on orderstages ca orderstages on orderstages on orderstages on morphology false strictequal false on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on purchasemethodlist on purchasemethodlis t on purchasemethodlist on purchasemethodlist on purchasemethodlist on purch asemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodli st on purchasemethodlist on purchasemethodlist on purchasemethodlist on purc hasemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodl ist on purchasemethodlist on orderpricecurrencyid customer title b e bd bd be bd be bd be be d bd bd bc ba be bb bd ba bb bd ba be bb bd bc b b bd bd customer code customer customer inn custlev on custlev on custlev on custlev on orderstages af orderstages on orderstages ca orderstages on orde rstages on orderstages on morphology false strictequal false on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on placeofsearch fz placeofsearch on purchasemethodlist on purchasemethodlis t on purchasemethodlist on purchasemethodlist on purchasemethodlist on purch asemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodli st on purchasemethodlist on purchasemethodlist on purchasemethodlist on purc hasemethodlist on purchasemethodlist on purchasemethodlist on purchasemethodl ist on purchasemethodlist on orderpricecurrencyid customer title b e bd bd be bd be bd be be d bd bd ba bb bd ba be bb bd bc bd bf be be bf be ba be be bc bd be be bd bd bd be ba be be ba customer code customer customer inn custlev on custlev on custlev on cu stlev on orderstages af orderstages on orderstages ca orderstages on ordersta ges on orderstages on morphology false strictequal false original issue reported on code google com by egor shi gmail com on jan at | 1 |
807,359 | 29,997,433,485 | IssuesEvent | 2023-06-26 06:56:38 | Riksrevisjonen/pioneeR | https://api.github.com/repos/Riksrevisjonen/pioneeR | opened | Ability to add weights to variables | Type: 2-Enhancement Priority: 3-Medium | Add UI to add weights to variables included in the analysis. | 1.0 | Ability to add weights to variables - Add UI to add weights to variables included in the analysis. | priority | ability to add weights to variables add ui to add weights to variables included in the analysis | 1 |
530,010 | 15,414,920,331 | IssuesEvent | 2021-03-05 01:22:39 | visit-dav/visit | https://api.github.com/repos/visit-dav/visit | closed | VisIt no longer displays output to the visualization window with XWin-32. | asc bug grizit impact high likelihood medium priority | I ran the development version of VisIt on quartz displaying back to my Windows 10 system running XWin-32 2018 and the visualization window was never refreshed and appeared black the whole time.
I got the following messages in the terminal window.
bin/visit
Running: gui -dv
Qt: Warning: X server does not support XInput 2
Qt: Warning: failed to get the current screen resources
Qt: Warning: QXcbConnection: XCB error: 160 (Unknown), sequence: 177, resource id: 101, major code: 148 (Unknown), minor code: 20
Running: viewer -dv -geometry 3020x1372+420+28 -borders 31,8,8,8 -shift 8,31 -preshift 0,0 -defer -host 127.0.0.1 -port 5600
Running: mdserver -dv -host 127.0.0.1 -port 5601
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Running: engine_ser -dv -host 127.0.0.1 -port 5600
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
The key messages are about "MIT_SHM" missing. This was a problem when VisIt 3.0 was first released and if you run 3.0.2 on quartz you will get those messages. I added a patch to Mesa that fixed this problem and if you run 3.1.4 on quartz those messages will not be present and graphics will show up in the visualization window.
That patch needs to be added back to bv_mesagl.sh. | 1.0 | VisIt no longer displays output to the visualization window with XWin-32. - I ran the development version of VisIt on quartz displaying back to my Windows 10 system running XWin-32 2018 and the visualization window was never refreshed and appeared black the whole time.
I got the following messages in the terminal window.
bin/visit
Running: gui -dv
Qt: Warning: X server does not support XInput 2
Qt: Warning: failed to get the current screen resources
Qt: Warning: QXcbConnection: XCB error: 160 (Unknown), sequence: 177, resource id: 101, major code: 148 (Unknown), minor code: 20
Running: viewer -dv -geometry 3020x1372+420+28 -borders 31,8,8,8 -shift 8,31 -preshift 0,0 -defer -host 127.0.0.1 -port 5600
Running: mdserver -dv -host 127.0.0.1 -port 5601
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Running: engine_ser -dv -host 127.0.0.1 -port 5600
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
Xlib: extension "MIT-SHM" missing on display "192.168.132.132:17.0".
The key messages are about "MIT_SHM" missing. This was a problem when VisIt 3.0 was first released and if you run 3.0.2 on quartz you will get those messages. I added a patch to Mesa that fixed this problem and if you run 3.1.4 on quartz those messages will not be present and graphics will show up in the visualization window.
That patch needs to be added back to bv_mesagl.sh. | priority | visit no longer displays output to the visualization window with xwin i ran the development version of visit on quartz displaying back to my windows system running xwin and the visualization window was never refreshed and appeared black the whole time i got the following messages in the terminal window bin visit running gui dv qt warning x server does not support xinput qt warning failed to get the current screen resources qt warning qxcbconnection xcb error unknown sequence resource id major code unknown minor code running viewer dv geometry borders shift preshift defer host port running mdserver dv host port xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display running engine ser dv host port xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display xlib extension mit shm missing on display the key messages are about mit shm missing this was a problem when visit was first released and if you run on quartz you will get those messages i added a patch to mesa that fixed this problem and if you run on quartz those messages will not be present and graphics will show up in the visualization window that patch needs to be added back to bv mesagl sh | 1 |
26,577 | 2,684,877,879 | IssuesEvent | 2015-03-29 13:28:36 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | Missing link dependency in Far.ExtendedConsole in Release mode | 2–5 stars bug imported Priority-Medium | _From [peter.su...@gmail.com](https://code.google.com/u/111208667693142380164/) on June 14, 2013 05:05:45_
Required information! OS version: Win7 SP1 x86/x64 ConEmu version: Latest preview Far version (if you are using Far Manager): ? *Bug description* When building ConEmu solution in Release mode a link dependency in Far. ExtendedConsole is missing. If you link with Version.lib it works. *Steps to reproduction* 1. 2. 3.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1098_ | 1.0 | Missing link dependency in Far.ExtendedConsole in Release mode - _From [peter.su...@gmail.com](https://code.google.com/u/111208667693142380164/) on June 14, 2013 05:05:45_
Required information! OS version: Win7 SP1 x86/x64 ConEmu version: Latest preview Far version (if you are using Far Manager): ? *Bug description* When building ConEmu solution in Release mode a link dependency in Far. ExtendedConsole is missing. If you link with Version.lib it works. *Steps to reproduction* 1. 2. 3.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1098_ | priority | missing link dependency in far extendedconsole in release mode from on june required information os version conemu version latest preview far version if you are using far manager bug description when building conemu solution in release mode a link dependency in far extendedconsole is missing if you link with version lib it works steps to reproduction original issue | 1 |
327,309 | 9,973,905,306 | IssuesEvent | 2019-07-09 09:23:05 | canonical-web-and-design/canonicalwebteam.blog | https://api.github.com/repos/canonical-web-and-design/canonicalwebteam.blog | closed | Add test for tag endpoint in common view logic | Priority: Medium | The `tag` endpoint should be tested in the `test_common_view_logic.py` file. | 1.0 | Add test for tag endpoint in common view logic - The `tag` endpoint should be tested in the `test_common_view_logic.py` file. | priority | add test for tag endpoint in common view logic the tag endpoint should be tested in the test common view logic py file | 1 |
5,426 | 2,575,630,504 | IssuesEvent | 2015-02-12 00:42:31 | couchbase/couchbase-lite-net | https://api.github.com/repos/couchbase/couchbase-lite-net | closed | ToLiveQuery() doesn't preserve query's attributes | bug P4: minor priority-medium size-small | Things like start key and end key are reset to null when calling ToLiveQuery() on a query | 1.0 | ToLiveQuery() doesn't preserve query's attributes - Things like start key and end key are reset to null when calling ToLiveQuery() on a query | priority | tolivequery doesn t preserve query s attributes things like start key and end key are reset to null when calling tolivequery on a query | 1 |
676,651 | 23,132,932,410 | IssuesEvent | 2022-07-28 12:05:30 | CodeYourFuture/ldn8-horizon-chatterbox-talent | https://api.github.com/repos/CodeYourFuture/ldn8-horizon-chatterbox-talent | closed | Sort #1 Level 100 sort by most recent posted | m medium priority | Sorting was a non-client requirement, however we deemed it to provide very useful UX.
The following were agreed to be potentially useful sorting categories:
By scheme start-date (note - this field may not be easily retrievable given the existing fields)
By date added to the site
By duration of the program
Add sort drop down
Identify data to sort
on select opportunities are sorted | 1.0 | Sort #1 Level 100 sort by most recent posted - Sorting was a non-client requirement, however we deemed it to provide very useful UX.
The following were agreed to be potentially useful sorting categories:
By scheme start-date (note - this field may not be easily retrievable given the existing fields)
By date added to the site
By duration of the program
Add sort drop down
Identify data to sort
on select opportunities are sorted | priority | sort level sort by most recent posted sorting was a non client requirement however we deemed it to provide very useful ux the following were agreed to be potentially useful sorting categories by scheme start date note this field may not be easily retrievable given the existing fields by date added to the site by duration of the program add sort drop down identify data to sort on select opportunities are sorted | 1 |
212,627 | 7,239,265,398 | IssuesEvent | 2018-02-13 16:57:21 | cilium/cilium | https://api.github.com/repos/cilium/cilium | closed | Cilium log entry needs appropriate log level (info/warn vs ERROR) | priority/medium | While debugging source endpoint labels observed as null in the cilium access log, the following ERROR was observed using `kubectl logs cilium-xxxxx':
```
level=info msg="Endpoint policy recalculated" containerID=16ca9971fc endpointID=9765 policyRevision=56
ERROR: logging before flag.Parse: W0213 01:24:16.672084 1 reflector.go:334] github.com/cilium/cilium/pkg/k8s/client/informers/externalversions/factory.go:71: watch of *v2.CiliumNetworkPolicy ended with: too old resource version: 132262 (132768)
```
This `ERROR` was diagnosed as harmless, thus should be logged at a less severe level, like info or warn, to avoid alarming the operator and saving the operator time for investigating the error.
More environment details:
`cilium/cilium:v0.13.1`
```
$kubectl exec -ti cilium-7bx9l -n kube-system cilium version
Client: 1.0.0-rc4 3523143 2018-01-30T17:03:18+01:00 go version go1.9 linux/amd64
Daemon: 1.0.0-rc4 3523143 2018-01-30T17:03:18+01:00 go version go1.9 linux/amd64
$kubectl version
Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:52:23Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.0", GitCommit:"0b9efaeb34a2fc51ff8e4d34ad9bc6375459c4a4", GitTreeState:"clean", BuildDate:"2017-11-29T22:43:34Z", GoVersion:"go1.9.1", Compiler:"gc", Platform:"linux/amd64"}
```
| 1.0 | Cilium log entry needs appropriate log level (info/warn vs ERROR) - While debugging source endpoint labels observed as null in the cilium access log, the following ERROR was observed using `kubectl logs cilium-xxxxx':
```
level=info msg="Endpoint policy recalculated" containerID=16ca9971fc endpointID=9765 policyRevision=56
ERROR: logging before flag.Parse: W0213 01:24:16.672084 1 reflector.go:334] github.com/cilium/cilium/pkg/k8s/client/informers/externalversions/factory.go:71: watch of *v2.CiliumNetworkPolicy ended with: too old resource version: 132262 (132768)
```
This `ERROR` was diagnosed as harmless, thus should be logged at a less severe level, like info or warn, to avoid alarming the operator and saving the operator time for investigating the error.
More environment details:
`cilium/cilium:v0.13.1`
```
$kubectl exec -ti cilium-7bx9l -n kube-system cilium version
Client: 1.0.0-rc4 3523143 2018-01-30T17:03:18+01:00 go version go1.9 linux/amd64
Daemon: 1.0.0-rc4 3523143 2018-01-30T17:03:18+01:00 go version go1.9 linux/amd64
$kubectl version
Client Version: version.Info{Major:"1", Minor:"9", GitVersion:"v1.9.1", GitCommit:"3a1c9449a956b6026f075fa3134ff92f7d55f812", GitTreeState:"clean", BuildDate:"2018-01-04T11:52:23Z", GoVersion:"go1.9.2", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"8", GitVersion:"v1.8.0", GitCommit:"0b9efaeb34a2fc51ff8e4d34ad9bc6375459c4a4", GitTreeState:"clean", BuildDate:"2017-11-29T22:43:34Z", GoVersion:"go1.9.1", Compiler:"gc", Platform:"linux/amd64"}
```
| priority | cilium log entry needs appropriate log level info warn vs error while debugging source endpoint labels observed as null in the cilium access log the following error was observed using kubectl logs cilium xxxxx level info msg endpoint policy recalculated containerid endpointid policyrevision error logging before flag parse reflector go github com cilium cilium pkg client informers externalversions factory go watch of ciliumnetworkpolicy ended with too old resource version this error was diagnosed as harmless thus should be logged at a less severe level like info or warn to avoid alarming the operator and saving the operator time for investigating the error more environment details cilium cilium kubectl exec ti cilium n kube system cilium version client go version linux daemon go version linux kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform darwin server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux | 1 |
283,576 | 8,719,962,747 | IssuesEvent | 2018-12-08 06:45:15 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Crash with arbitrary polygon mesh from kull | asc bug likelihood medium priority reviewed severity low | Filled Boundary Plot crash on later timesteps.
The mesh begins as quads but becomes arbitrary polygons as the simulation progresses.
Cyrus has test files on rz.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1444
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Crash with arbitrary polygon mesh from kull
Assigned to: Jeremy Meredith
Category:
Target version: 2.6.3
Author: Cyrus Harrison
Start: 05/03/2013
Due date:
% Done: 0
Estimated time:
Created: 05/03/2013 12:04 pm
Updated: 06/18/2013 02:52 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.2
Impact:
Expected Use:
OS: All
Support Group: DOE/ASC
Description:
Filled Boundary Plot crash on later timesteps.
The mesh begins as quads but becomes arbitrary polygons as the simulation progresses.
Cyrus has test files on rz.
Comments:
Certainty seems to be an issue with polygons.It is crashing in ZooMIR.nodeVFs comes from avt/MIR/Zoo/CellReconstructor it is defined to be size MAX_NODES_PER_ZONE, which seems to be 7.It crashes in a case where we have 9 nodes for a zone.Not sure if something slipped by the transform manager, or if we just have a corner case?
Brad said there should be so logic to split higher edge polygons, Cyrus is following up with Jeremy for guidance.
I added logic to tessellate polygons with more than 8 sides. Turns out we don't really sacrifice accuracy as long as we tessellate them on the fly during reconstruction. And since MIR is already outputting more than one output cell per input cell, it has no new consequences for operations later in the pipeline. (It would be messy and inaccurate if we tessellated any earlier than the last-possible-second; e.g. doing so in the transform mgr would be Bad).
| 1.0 | Crash with arbitrary polygon mesh from kull - Filled Boundary Plot crash on later timesteps.
The mesh begins as quads but becomes arbitrary polygons as the simulation progresses.
Cyrus has test files on rz.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1444
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Crash with arbitrary polygon mesh from kull
Assigned to: Jeremy Meredith
Category:
Target version: 2.6.3
Author: Cyrus Harrison
Start: 05/03/2013
Due date:
% Done: 0
Estimated time:
Created: 05/03/2013 12:04 pm
Updated: 06/18/2013 02:52 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.2
Impact:
Expected Use:
OS: All
Support Group: DOE/ASC
Description:
Filled Boundary Plot crash on later timesteps.
The mesh begins as quads but becomes arbitrary polygons as the simulation progresses.
Cyrus has test files on rz.
Comments:
Certainty seems to be an issue with polygons.It is crashing in ZooMIR.nodeVFs comes from avt/MIR/Zoo/CellReconstructor it is defined to be size MAX_NODES_PER_ZONE, which seems to be 7.It crashes in a case where we have 9 nodes for a zone.Not sure if something slipped by the transform manager, or if we just have a corner case?
Brad said there should be so logic to split higher edge polygons, Cyrus is following up with Jeremy for guidance.
I added logic to tessellate polygons with more than 8 sides. Turns out we don't really sacrifice accuracy as long as we tessellate them on the fly during reconstruction. And since MIR is already outputting more than one output cell per input cell, it has no new consequences for operations later in the pipeline. (It would be messy and inaccurate if we tessellated any earlier than the last-possible-second; e.g. doing so in the transform mgr would be Bad).
| priority | crash with arbitrary polygon mesh from kull filled boundary plot crash on later timesteps the mesh begins as quads but becomes arbitrary polygons as the simulation progresses cyrus has test files on rz redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject crash with arbitrary polygon mesh from kull assigned to jeremy meredith category target version author cyrus harrison start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group doe asc description filled boundary plot crash on later timesteps the mesh begins as quads but becomes arbitrary polygons as the simulation progresses cyrus has test files on rz comments certainty seems to be an issue with polygons it is crashing in zoomir nodevfs comes from avt mir zoo cellreconstructor it is defined to be size max nodes per zone which seems to be it crashes in a case where we have nodes for a zone not sure if something slipped by the transform manager or if we just have a corner case brad said there should be so logic to split higher edge polygons cyrus is following up with jeremy for guidance i added logic to tessellate polygons with more than sides turns out we don t really sacrifice accuracy as long as we tessellate them on the fly during reconstruction and since mir is already outputting more than one output cell per input cell it has no new consequences for operations later in the pipeline it would be messy and inaccurate if we tessellated any earlier than the last possible second e g doing so in the transform mgr would be bad | 1 |
751,079 | 26,229,778,820 | IssuesEvent | 2023-01-04 22:31:59 | hoffstadt/DearPyGui | https://api.github.com/repos/hoffstadt/DearPyGui | closed | Dotted line theme for drawing and plotting. | type: feature category: plots priority: medium | DPG 1.8.0
Dashed lines are often used in drawing and plotting, could this style be supported in the future. | 1.0 | Dotted line theme for drawing and plotting. - DPG 1.8.0
Dashed lines are often used in drawing and plotting, could this style be supported in the future. | priority | dotted line theme for drawing and plotting dpg dashed lines are often used in drawing and plotting could this style be supported in the future | 1 |
89,427 | 3,793,811,979 | IssuesEvent | 2016-03-22 15:05:43 | pmem/issues | https://api.github.com/repos/pmem/issues | closed | pmempool create: multiplicative suffixes KB, MB, GB etc. for size argument are not working | Exposure: Medium Priority: 3 medium Type: Bug | From pmempool-create man page:
`The size argument may be passed with multiplicative suffixes K=1024, M=1024*1024 and so on for G, T and P or the suffixes KB=1000, MB=1000*1000 and so on for GB, TB and PB`
In fact KB, MB etc. suffixes are not working at all. Pool is always created with 1024^x multiplicative suffix | 1.0 | pmempool create: multiplicative suffixes KB, MB, GB etc. for size argument are not working - From pmempool-create man page:
`The size argument may be passed with multiplicative suffixes K=1024, M=1024*1024 and so on for G, T and P or the suffixes KB=1000, MB=1000*1000 and so on for GB, TB and PB`
In fact KB, MB etc. suffixes are not working at all. Pool is always created with 1024^x multiplicative suffix | priority | pmempool create multiplicative suffixes kb mb gb etc for size argument are not working from pmempool create man page the size argument may be passed with multiplicative suffixes k m and so on for g t and p or the suffixes kb mb and so on for gb tb and pb in fact kb mb etc suffixes are not working at all pool is always created with x multiplicative suffix | 1 |
616,842 | 19,322,420,011 | IssuesEvent | 2021-12-14 07:44:01 | andrewtougas/21mafia | https://api.github.com/repos/andrewtougas/21mafia | closed | Setup CI/CD Gitlab Script and Process | enhancement component: project priority: medium | Create the `.gitlab-ci.yml` file in the root directory and setup the Gitlab CI/CD process on master push. Follow the same file and setup as in [AMSH](https://gitlab.com/andrewtougas/amsh/).
EDIT (7/29/2019 2:23pm): The script should be updated to deploy the `dist/` directory to the remote server:
- https://docs.gitlab.com/ee/ci/ssh_keys/
- https://www.linode.com/docs/security/authentication/use-public-key-authentication-with-ssh
- https://codeburst.io/gitlab-build-and-push-to-a-server-via-ssh-6d27ca1bf7b4
> Migrated from Gitlab | 1.0 | Setup CI/CD Gitlab Script and Process - Create the `.gitlab-ci.yml` file in the root directory and setup the Gitlab CI/CD process on master push. Follow the same file and setup as in [AMSH](https://gitlab.com/andrewtougas/amsh/).
EDIT (7/29/2019 2:23pm): The script should be updated to deploy the `dist/` directory to the remote server:
- https://docs.gitlab.com/ee/ci/ssh_keys/
- https://www.linode.com/docs/security/authentication/use-public-key-authentication-with-ssh
- https://codeburst.io/gitlab-build-and-push-to-a-server-via-ssh-6d27ca1bf7b4
> Migrated from Gitlab | priority | setup ci cd gitlab script and process create the gitlab ci yml file in the root directory and setup the gitlab ci cd process on master push follow the same file and setup as in edit the script should be updated to deploy the dist directory to the remote server migrated from gitlab | 1 |
316,715 | 9,653,922,274 | IssuesEvent | 2019-05-19 09:38:56 | chrisly-bear/PEBRApp | https://api.github.com/repos/chrisly-bear/PEBRApp | opened | Convert `enums` to `beans` | medium priority | Move the `enum` types in `database/models/..` to their own bean class in `database/beans/`. Take `database/beans/NoConsentReason.dart` as a blueprint for how to design a bean class. Remove the `xxxToString()` methods from `utils/Utils.dart`.
The bean class summarizes all relevant code in one file (i.e., it improves code quality). Furthermore, it allows us to customize the encoding of each bean option using the `_encoding` map (instead of relying on the enum sequence). | 1.0 | Convert `enums` to `beans` - Move the `enum` types in `database/models/..` to their own bean class in `database/beans/`. Take `database/beans/NoConsentReason.dart` as a blueprint for how to design a bean class. Remove the `xxxToString()` methods from `utils/Utils.dart`.
The bean class summarizes all relevant code in one file (i.e., it improves code quality). Furthermore, it allows us to customize the encoding of each bean option using the `_encoding` map (instead of relying on the enum sequence). | priority | convert enums to beans move the enum types in database models to their own bean class in database beans take database beans noconsentreason dart as a blueprint for how to design a bean class remove the xxxtostring methods from utils utils dart the bean class summarizes all relevant code in one file i e it improves code quality furthermore it allows us to customize the encoding of each bean option using the encoding map instead of relying on the enum sequence | 1 |
380,447 | 11,261,099,045 | IssuesEvent | 2020-01-13 11:59:03 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.0 staging-1334] Contracts: need a hint with items number for finishing a contract | Priority: Medium | When you are trying to finish a contract with some items (distribution, adding, harvesting) you have this useful popup with wanted items. But we had number of items early too (how much items you still need for finishing a contract)

| 1.0 | [0.9.0 staging-1334] Contracts: need a hint with items number for finishing a contract - When you are trying to finish a contract with some items (distribution, adding, harvesting) you have this useful popup with wanted items. But we had number of items early too (how much items you still need for finishing a contract)

| priority | contracts need a hint with items number for finishing a contract when you are trying to finish a contract with some items distribution adding harvesting you have this useful popup with wanted items but we had number of items early too how much items you still need for finishing a contract | 1 |
189,160 | 6,794,683,481 | IssuesEvent | 2017-11-01 13:12:48 | ActivityWatch/activitywatch | https://api.github.com/repos/ActivityWatch/activitywatch | closed | Move docs into a separate repository | area: ci priority: medium size: small | It feels unnecessary to make a push to the bundle repo everytime I want to add something to the docs. I think it would make sense to move it out.
It should probably not be a submodule either. | 1.0 | Move docs into a separate repository - It feels unnecessary to make a push to the bundle repo everytime I want to add something to the docs. I think it would make sense to move it out.
It should probably not be a submodule either. | priority | move docs into a separate repository it feels unnecessary to make a push to the bundle repo everytime i want to add something to the docs i think it would make sense to move it out it should probably not be a submodule either | 1 |
268,866 | 8,415,299,762 | IssuesEvent | 2018-10-13 13:18:24 | SkriptLang/Skript | https://api.github.com/repos/SkriptLang/Skript | closed | Problem with negative skript calculations | bug completed priority: medium | **Describe your problem:**
Skript dev37c no longer does calculations correctly. Indeed, when we remove a number from a variable that is not set, this value is added to the variable instead.
**Code**
```vb
command test123:
trigger:
remove 7 from {_x}
broadcast "x= %{_x}%" #Result will be 7 and not -7
```
**Debug**
__Skript:__ dev37c
__Addons:__ Tested without addons
__Server:__ CraftBukkit version git-Spigot-3d850ec-809c399 (MC: 1.12.2) **and** Paper version git-TacoSpigot-"f8ba67d6" (MC: 1.12.2) | 1.0 | Problem with negative skript calculations - **Describe your problem:**
Skript dev37c no longer does calculations correctly. Indeed, when we remove a number from a variable that is not set, this value is added to the variable instead.
**Code**
```vb
command test123:
trigger:
remove 7 from {_x}
broadcast "x= %{_x}%" #Result will be 7 and not -7
```
**Debug**
__Skript:__ dev37c
__Addons:__ Tested without addons
__Server:__ CraftBukkit version git-Spigot-3d850ec-809c399 (MC: 1.12.2) **and** Paper version git-TacoSpigot-"f8ba67d6" (MC: 1.12.2) | priority | problem with negative skript calculations describe your problem skript no longer does calculations correctly indeed when we remove a number from a variable that is not set this value is added to the variable instead code vb command trigger remove from x broadcast x x result will be and not debug skript addons tested without addons server craftbukkit version git spigot mc and paper version git tacospigot mc | 1 |
69,893 | 3,316,293,586 | IssuesEvent | 2015-11-06 16:18:39 | TeselaGen/Peony-Issue-Tracking | https://api.github.com/repos/TeselaGen/Peony-Issue-Tracking | opened | User Manager - Restrict usernames to valid Dow usernames | Customer: DAS Phase I Priority: Medium Type: Enhancement | _From @thldev on October 27, 2015 20:37_
When creating a new user using the user manager tab in the admin console, it would be nice to have it query LDAP and confirm that the username exists.
_Copied from original issue: TeselaGen/ve#1503_ | 1.0 | User Manager - Restrict usernames to valid Dow usernames - _From @thldev on October 27, 2015 20:37_
When creating a new user using the user manager tab in the admin console, it would be nice to have it query LDAP and confirm that the username exists.
_Copied from original issue: TeselaGen/ve#1503_ | priority | user manager restrict usernames to valid dow usernames from thldev on october when creating a new user using the user manager tab in the admin console it would be nice to have it query ldap and confirm that the username exists copied from original issue teselagen ve | 1 |
405,185 | 11,869,346,970 | IssuesEvent | 2020-03-26 10:47:39 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Content Types ending with "s" can't be saved | priority: medium source: plugin:content-manager status: confirmed type: bug | **Describe the bug**
When creating a content type that ends in the letter "s", Strapi automatically creates the table in the database using "es" for the plural. The problem is that if you have a component within that content type, the column with the ID added to the component table is correct without the "es", but Strapi looks for it just without the "s".
For example:
Content Type: "about_us"
Table name: "about_uses" (sounds funny, but whatever)
A component is added to the "about_us" content type. In the component table, the "about_us_id" column is added, but Strapi looks for the column "about_us**e**_id".
**Steps to reproduce the behavior**
1. Create type: "about_us"
2. Create a component and add it to the content type.
3. Start Strapi and create an item of content type "about_us".
4. Try saving. An error is returned.
**Expected behavior**
The item should be saved and Strapi should look for the column without the "e" that it added to the name.
**System**
- Node.js version: 10.18.1
- Strapi version: 3.0.0-beta.18.6
- Database: Postgres
- Operating system: Linux
| 1.0 | Content Types ending with "s" can't be saved - **Describe the bug**
When creating a content type that ends in the letter "s", Strapi automatically creates the table in the database using "es" for the plural. The problem is that if you have a component within that content type, the column with the ID added to the component table is correct without the "es", but Strapi looks for it just without the "s".
For example:
Content Type: "about_us"
Table name: "about_uses" (sounds funny, but whatever)
A component is added to the "about_us" content type. In the component table, the "about_us_id" column is added, but Strapi looks for the column "about_us**e**_id".
**Steps to reproduce the behavior**
1. Create type: "about_us"
2. Create a component and add it to the content type.
3. Start Strapi and create an item of content type "about_us".
4. Try saving. An error is returned.
**Expected behavior**
The item should be saved and Strapi should look for the column without the "e" that it added to the name.
**System**
- Node.js version: 10.18.1
- Strapi version: 3.0.0-beta.18.6
- Database: Postgres
- Operating system: Linux
| priority | content types ending with s can t be saved describe the bug when creating a content type that ends in the letter s strapi automatically creates the table in the database using es for the plural the problem is that if you have a component within that content type the column with the id added to the component table is correct without the es but strapi looks for it just without the s for example content type about us table name about uses sounds funny but whatever a component is added to the about us content type in the component table the about us id column is added but strapi looks for the column about us e id steps to reproduce the behavior create type about us create a component and add it to the content type start strapi and create an item of content type about us try saving an error is returned expected behavior the item should be saved and strapi should look for the column without the e that it added to the name system node js version strapi version beta database postgres operating system linux | 1 |
543,818 | 15,885,756,484 | IssuesEvent | 2021-04-09 21:08:55 | zeoflow/flow-kit | https://api.github.com/repos/zeoflow/flow-kit | closed | StringCreator | @feature @priority-medium | **Is your feature request related to a problem? Please describe.**
A class to create easy strings - StringCreator
###### We also happily accept [pull requests](https://github.com/zeoflow/flow-kit/pulls).
| 1.0 | StringCreator - **Is your feature request related to a problem? Please describe.**
A class to create easy strings - StringCreator
###### We also happily accept [pull requests](https://github.com/zeoflow/flow-kit/pulls).
| priority | stringcreator is your feature request related to a problem please describe a class to create easy strings stringcreator we also happily accept | 1 |
501,385 | 14,527,076,864 | IssuesEvent | 2020-12-14 14:57:05 | carbon-design-system/carbon-for-ibm-dotcom | https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom | closed | Web component: Content block - cards Prod QA testing | QA dev complete package: web components priority: medium | <!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
developer using the ibm.com Library `Content block - cards`
> I need to:
have a version of the component that has been tested for accessibility compliance as well as on multiple browsers and platforms
> so that I can:
be confident that my ibm.com web site users will have a good experience
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- [Browser Stack link](https://ibm.ent.box.com/notes/578734426612)
- [Browser Standard](https://w3.ibm.com/standards/web/browser/)
- Browser versions to be tested: Tier 1 browsers will be tested with defects created as Sev 1 or Sev 2. Tier 2 browser defects will be created as Sev 3 defects.
- Platforms to be tested, by priority: 1) Desktop 2) Mobile 3) Tablet
- Mobile & Tablet iOS versions: 13.1, 13.3 and 14
- Mobile & Tablet Android versions: 9.0 Pie and 8.1 Oreo
- Browsers to be tested: Desktop: Chrome, Firefox, Safari, Edge, Mobile: Chrome, Safari, Samsung Internet, UC Browser, Tablet: Safari, Chrome, Android
- [Accessibility Checklist](https://www.ibm.com/able/guidelines/ci162/accessibility_checklist.html)
- [Creating a QA bug](https://ibm.ent.box.com/notes/603242247385)
- **See the Epic for the Design and Functional specs information**
- Dev issue (#3782)
- Once development is finished the updated code is available in the [**Web Components Canary Environment**](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/overview-getting-started--page) for testing.
- [**React canary environment**](https://ibmdotcom-react-canary.mybluemix.net/?path=/story/overview-getting-started--page)
#### Acceptance criteria
- [ ] Accessibility testing is complete. Component is compliant.
- [ ] All browser versions are tested
- [ ] All operating systems are tested
- [ ] All devices are tested
- [ ] Defects are recorded and retested when fixed | 1.0 | Web component: Content block - cards Prod QA testing - <!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
developer using the ibm.com Library `Content block - cards`
> I need to:
have a version of the component that has been tested for accessibility compliance as well as on multiple browsers and platforms
> so that I can:
be confident that my ibm.com web site users will have a good experience
#### Additional information
<!-- {{Please provide any additional information or resources for reference}} -->
- [Browser Stack link](https://ibm.ent.box.com/notes/578734426612)
- [Browser Standard](https://w3.ibm.com/standards/web/browser/)
- Browser versions to be tested: Tier 1 browsers will be tested with defects created as Sev 1 or Sev 2. Tier 2 browser defects will be created as Sev 3 defects.
- Platforms to be tested, by priority: 1) Desktop 2) Mobile 3) Tablet
- Mobile & Tablet iOS versions: 13.1, 13.3 and 14
- Mobile & Tablet Android versions: 9.0 Pie and 8.1 Oreo
- Browsers to be tested: Desktop: Chrome, Firefox, Safari, Edge, Mobile: Chrome, Safari, Samsung Internet, UC Browser, Tablet: Safari, Chrome, Android
- [Accessibility Checklist](https://www.ibm.com/able/guidelines/ci162/accessibility_checklist.html)
- [Creating a QA bug](https://ibm.ent.box.com/notes/603242247385)
- **See the Epic for the Design and Functional specs information**
- Dev issue (#3782)
- Once development is finished the updated code is available in the [**Web Components Canary Environment**](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/overview-getting-started--page) for testing.
- [**React canary environment**](https://ibmdotcom-react-canary.mybluemix.net/?path=/story/overview-getting-started--page)
#### Acceptance criteria
- [ ] Accessibility testing is complete. Component is compliant.
- [ ] All browser versions are tested
- [ ] All operating systems are tested
- [ ] All devices are tested
- [ ] Defects are recorded and retested when fixed | priority | web component content block cards prod qa testing user story as a developer using the ibm com library content block cards i need to have a version of the component that has been tested for accessibility compliance as well as on multiple browsers and platforms so that i can be confident that my ibm com web site users will have a good experience additional information browser versions to be tested tier browsers will be tested with defects created as sev or sev tier browser defects will be created as sev defects platforms to be tested by priority desktop mobile tablet mobile tablet ios versions and mobile tablet android versions pie and oreo browsers to be tested desktop chrome firefox safari edge mobile chrome safari samsung internet uc browser tablet safari chrome android see the epic for the design and functional specs information dev issue once development is finished the updated code is available in the for testing acceptance criteria accessibility testing is complete component is compliant all browser versions are tested all operating systems are tested all devices are tested defects are recorded and retested when fixed | 1 |
656,818 | 21,776,928,198 | IssuesEvent | 2022-05-13 14:38:30 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | Useability improvement: Duplicate node names lead to inability to manage a room | internal submit priority/medium | We use the node name as the only identifier shown on multiple screens. Of especial importance is the room nodes list, which we would normally use to remove nodes we don't want in the room.
If we continue to allow duplicate node names, we need to provide an additional identifier that allows users to distinguish between multiple entries with the same name, so they can at least remove from the room the ones that are problematic. | 1.0 | Useability improvement: Duplicate node names lead to inability to manage a room - We use the node name as the only identifier shown on multiple screens. Of especial importance is the room nodes list, which we would normally use to remove nodes we don't want in the room.
If we continue to allow duplicate node names, we need to provide an additional identifier that allows users to distinguish between multiple entries with the same name, so they can at least remove from the room the ones that are problematic. | priority | useability improvement duplicate node names lead to inability to manage a room we use the node name as the only identifier shown on multiple screens of especial importance is the room nodes list which we would normally use to remove nodes we don t want in the room if we continue to allow duplicate node names we need to provide an additional identifier that allows users to distinguish between multiple entries with the same name so they can at least remove from the room the ones that are problematic | 1 |
535,423 | 15,687,921,244 | IssuesEvent | 2021-03-25 14:10:30 | CS506-Oversight/autorack-front | https://api.github.com/repos/CS506-Oversight/autorack-front | opened | Test for error snackbar pop up | Priority-Medium Type-Enhancements | Error snack bar pop-up is introduced in #11. However, the related tests are not.
### Test cases
- Enter some malformed account information for **sign up** and expect an alert to pop up
- Enter non-existent account information for **sign in** and expect an alert to pop up | 1.0 | Test for error snackbar pop up - Error snack bar pop-up is introduced in #11. However, the related tests are not.
### Test cases
- Enter some malformed account information for **sign up** and expect an alert to pop up
- Enter non-existent account information for **sign in** and expect an alert to pop up | priority | test for error snackbar pop up error snack bar pop up is introduced in however the related tests are not test cases enter some malformed account information for sign up and expect an alert to pop up enter non existent account information for sign in and expect an alert to pop up | 1 |
41,590 | 2,869,064,144 | IssuesEvent | 2015-06-05 23:02:31 | dart-lang/polymer-dart | https://api.github.com/repos/dart-lang/polymer-dart | opened | Dart Polymer element + dart2js: let me import into a Polymer JS project | enhancement NeedsInfo PolymerMilestone-Later Priority-Medium | <a href="https://github.com/kevmoo"><img src="https://avatars.githubusercontent.com/u/17034?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [kevmoo](https://github.com/kevmoo)**
_Originally opened as dart-lang/sdk#15227_
----
No clue if this is practical, but it'd be great.
1) Create polymer element with foo.dart + foo.html
2) dart2js to foo.js
3) Deploy foo.html + foo.js ... somewhere
4) Polymer JS app can <link rel="import" href="foo.html">
5) It works! | 1.0 | Dart Polymer element + dart2js: let me import into a Polymer JS project - <a href="https://github.com/kevmoo"><img src="https://avatars.githubusercontent.com/u/17034?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [kevmoo](https://github.com/kevmoo)**
_Originally opened as dart-lang/sdk#15227_
----
No clue if this is practical, but it'd be great.
1) Create polymer element with foo.dart + foo.html
2) dart2js to foo.js
3) Deploy foo.html + foo.js ... somewhere
4) Polymer JS app can <link rel="import" href="foo.html">
5) It works! | priority | dart polymer element let me import into a polymer js project issue by originally opened as dart lang sdk no clue if this is practical but it d be great create polymer element with foo dart foo html to foo js deploy foo html foo js somewhere polymer js app can lt link rel quot import quot href quot foo html quot gt it works | 1 |
220,151 | 7,353,490,514 | IssuesEvent | 2018-03-09 00:59:52 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] error type generated for argument of method, type mismatch | 1. Type: Bug Priority: Medium Status: Completed Time: ~1 hour | <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
<!-- explain the issue or feature request, if an error is being thrown a stack trace helps -->
Got a method with a few arguments defined as below:
```
public filter(term: string, fields?: (string | number)[]): void;
```
notice that the `fields` argument got a combined type `(string | number)[]`
And, I use compodoc to generate its meta info to a json file, I got the following data:
```
"jsdoctags": [
{
"name": "term",
"type": "",
"tagName": {
"text": "param"
}
},
{
"name": "fields",
"type": "string | number[]",
"optional": true,
"tagName": {
"text": "param"
}
}
]
```
Notice that the type info `string | number[]` in the generated `jsdoctags` property is not correct. The generated type got a totally different type!
##### **Operating System, Node.js, npm, compodoc version(s)**
<!-- provide all information we need -->
win7, node 6.1
##### **Angular configuration, a `package.json` file in the root folder**
<!-- This is mandatory for bug reports. This will help us to replicate the scenario. -->
yes
##### **Compodoc installed globally or locally ?**
<!-- provide all information we need -->
globally
##### **Motivation for or Use Case**
<!-- explain why this is a bug for you -->
##### **Reproduce the error**
<!-- an unambiguous set of steps to reproduce the error. -->
run the following scripts in win7:
```
git clone https://github.com/rdkmaster/jigsaw.git
cd jigsaw
compodoc src/jigsaw -p tsconfig.json --disableSourceCode --disableGraph --disableCoverage --disablePrivate --disableInternal --disableLifeCycleHooks --disableRoutesGraph --exportFormat json
```
After compodoc successfully generates the json file, open file, you can find this info near line 11185.
##### **Related issues**
<!-- has a similar issue been reported before? -->
##### **Suggest a Fix**
```
"jsdoctags": [
...
{
...
"type": "(string | number[])",
....
}
]
```
<!-- if you can't fix the bug yourself, perhaps you can point to what might be
causing the problem (line of code or commit) -->
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | 1.0 | [BUG] error type generated for argument of method, type mismatch - <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
<!-- explain the issue or feature request, if an error is being thrown a stack trace helps -->
Got a method with a few arguments defined as below:
```
public filter(term: string, fields?: (string | number)[]): void;
```
notice that the `fields` argument got a combined type `(string | number)[]`
And, I use compodoc to generate its meta info to a json file, I got the following data:
```
"jsdoctags": [
{
"name": "term",
"type": "",
"tagName": {
"text": "param"
}
},
{
"name": "fields",
"type": "string | number[]",
"optional": true,
"tagName": {
"text": "param"
}
}
]
```
Notice that the type info `string | number[]` in the generated `jsdoctags` property is not correct. The generated type got a totally different type!
##### **Operating System, Node.js, npm, compodoc version(s)**
<!-- provide all information we need -->
win7, node 6.1
##### **Angular configuration, a `package.json` file in the root folder**
<!-- This is mandatory for bug reports. This will help us to replicate the scenario. -->
yes
##### **Compodoc installed globally or locally ?**
<!-- provide all information we need -->
globally
##### **Motivation for or Use Case**
<!-- explain why this is a bug for you -->
##### **Reproduce the error**
<!-- an unambiguous set of steps to reproduce the error. -->
run the following scripts in win7:
```
git clone https://github.com/rdkmaster/jigsaw.git
cd jigsaw
compodoc src/jigsaw -p tsconfig.json --disableSourceCode --disableGraph --disableCoverage --disablePrivate --disableInternal --disableLifeCycleHooks --disableRoutesGraph --exportFormat json
```
After compodoc successfully generates the json file, open file, you can find this info near line 11185.
##### **Related issues**
<!-- has a similar issue been reported before? -->
##### **Suggest a Fix**
```
"jsdoctags": [
...
{
...
"type": "(string | number[])",
....
}
]
```
<!-- if you can't fix the bug yourself, perhaps you can point to what might be
causing the problem (line of code or commit) -->
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | priority | error type generated for argument of method type mismatch please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue got a method with a few arguments defined as below public filter term string fields string number void notice that the fields argument got a combined type string number and i use compodoc to generate its meta info to a json file i got the following data jsdoctags name term type tagname text param name fields type string number optional true tagname text param notice that the type info string number in the generated jsdoctags property is not correct the generated type got a totally different type operating system node js npm compodoc version s node angular configuration a package json file in the root folder yes compodoc installed globally or locally globally motivation for or use case reproduce the error run the following scripts in git clone cd jigsaw compodoc src jigsaw p tsconfig json disablesourcecode disablegraph disablecoverage disableprivate disableinternal disablelifecyclehooks disableroutesgraph exportformat json after compodoc successfully generates the json file open file you can find this info near line related issues suggest a fix jsdoctags type string number if you can t fix the bug yourself perhaps you can point to what might be causing the problem line of code or commit love compodoc please consider supporting our collective 👉 | 1 |
538,961 | 15,781,346,899 | IssuesEvent | 2021-04-01 11:14:22 | AbsaOSS/hyperdrive-trigger | https://api.github.com/repos/AbsaOSS/hyperdrive-trigger | closed | Proper config implementation | bug priority: medium | Proper config implementation.
Update PostgresDB ( profile.api.Database.forConfig("db") ) implementation. Implement same logic as is in Configs conf
| 1.0 | Proper config implementation - Proper config implementation.
Update PostgresDB ( profile.api.Database.forConfig("db") ) implementation. Implement same logic as is in Configs conf
| priority | proper config implementation proper config implementation update postgresdb profile api database forconfig db implementation implement same logic as is in configs conf | 1 |
91,223 | 3,840,902,266 | IssuesEvent | 2016-04-04 01:03:38 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | vm crash | Blocked Feedback required Medium Priority VM | calling a macro `defMsg` from inside a template (`x`) crashes the compiler.
```nimrod
import entoody # https://bitbucket.org/fowlmouth/entoody
proc msg* (arg:int) {.unicast.}
template x (ty;body:stmt): stmt {.immediate.} =
defMsg(ty, msg) do (arg:int):
body
type Foo = object
x(Foo):
echo "FOO!"
``` | 1.0 | vm crash - calling a macro `defMsg` from inside a template (`x`) crashes the compiler.
```nimrod
import entoody # https://bitbucket.org/fowlmouth/entoody
proc msg* (arg:int) {.unicast.}
template x (ty;body:stmt): stmt {.immediate.} =
defMsg(ty, msg) do (arg:int):
body
type Foo = object
x(Foo):
echo "FOO!"
``` | priority | vm crash calling a macro defmsg from inside a template x crashes the compiler nimrod import entoody proc msg arg int unicast template x ty body stmt stmt immediate defmsg ty msg do arg int body type foo object x foo echo foo | 1 |
174,317 | 6,539,024,329 | IssuesEvent | 2017-09-01 09:18:58 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | opened | EP: Get Drugs | epic/medication kind/task priority/medium project/reimbursement | Get Drugs WS according to requirements
Requirements: [вимоги](https://docs.google.com/document/d/15qqew84E9PdDb0LS6mhCagOLqcnNa6wmxenR5kHrUPg/edit?usp=sharing)
[BP](https://edenlab.atlassian.net/wiki/spaces/EH/pages/3380784/Create+Medication+Request+BP)
- [x] Create WS [spec](https://edenlab.atlassian.net/wiki/spaces/EH/pages/3380878/Get+Drugs+Dictionary)
- [ ] Implement
- [ ] Deploy
- [ ] Improve test scenarios | 1.0 | EP: Get Drugs - Get Drugs WS according to requirements
Requirements: [вимоги](https://docs.google.com/document/d/15qqew84E9PdDb0LS6mhCagOLqcnNa6wmxenR5kHrUPg/edit?usp=sharing)
[BP](https://edenlab.atlassian.net/wiki/spaces/EH/pages/3380784/Create+Medication+Request+BP)
- [x] Create WS [spec](https://edenlab.atlassian.net/wiki/spaces/EH/pages/3380878/Get+Drugs+Dictionary)
- [ ] Implement
- [ ] Deploy
- [ ] Improve test scenarios | priority | ep get drugs get drugs ws according to requirements requirements create ws implement deploy improve test scenarios | 1 |
622,483 | 19,636,387,653 | IssuesEvent | 2022-01-08 11:06:20 | google/mozc | https://api.github.com/repos/google/mozc | closed | 誤変換: 1爪の話は(1つ目の、1つめの) | Priority-Medium auto-migrated Type-Conversion | ```
tweetで出ていた誤変換:
1爪の話は(1つ目の、1つめの)
9爪の話は(9つ目の、9つめの)
その他:
えいえい辞典 (「英英」が収録されていない)
%に換算すると(「ぱーせんと」の第一候補が「パーセン��
�」にならない)
「%」が欲しいときは Shift+5 で一発入力すると思います
Mozc-1.4.1033.102
```
Original issue reported on code.google.com by `heathros...@gmail.com` on 18 Apr 2012 at 2:18
| 1.0 | 誤変換: 1爪の話は(1つ目の、1つめの) - ```
tweetで出ていた誤変換:
1爪の話は(1つ目の、1つめの)
9爪の話は(9つ目の、9つめの)
その他:
えいえい辞典 (「英英」が収録されていない)
%に換算すると(「ぱーせんと」の第一候補が「パーセン��
�」にならない)
「%」が欲しいときは Shift+5 で一発入力すると思います
Mozc-1.4.1033.102
```
Original issue reported on code.google.com by `heathros...@gmail.com` on 18 Apr 2012 at 2:18
| priority | 誤変換 ( 、 ) tweetで出ていた誤変換 ( 、 ) ( 、 ) その他 えいえい辞典 (「英英」が収録されていない) %に換算すると(「ぱーせんと」の第一候補が「パーセン�� �」にならない) 「%」が欲しいときは shift で一発入力すると思います mozc original issue reported on code google com by heathros gmail com on apr at | 1 |
21,412 | 2,640,188,047 | IssuesEvent | 2015-03-11 10:16:04 | handsontable/hot-table | https://api.github.com/repos/handsontable/hot-table | closed | Declarative cell renderer | Enhancement Priority: medium Released | It should be possible to use a `<template rel="hot-renderer">` as a cell renderer, so one can declare renderers with HTML. Any HTML tag, Custom Element or Polymer variable can be used in the template.
```html
<hot-column>
<template rel="hot-renderer">
<strong>{{name.last}}</strong> {{name.first}}
</template>
</hot-column>
```
NOTE: `rel` is probably not the right attribute to add semantic meaning for a `<template>`, because it should be used only with predefined values and a href (http://demosthenes.info/blog/411/Using-rel-in-HTML5). Please investigate what is the desired replacement (`class`? `data-role`?) | 1.0 | Declarative cell renderer - It should be possible to use a `<template rel="hot-renderer">` as a cell renderer, so one can declare renderers with HTML. Any HTML tag, Custom Element or Polymer variable can be used in the template.
```html
<hot-column>
<template rel="hot-renderer">
<strong>{{name.last}}</strong> {{name.first}}
</template>
</hot-column>
```
NOTE: `rel` is probably not the right attribute to add semantic meaning for a `<template>`, because it should be used only with predefined values and a href (http://demosthenes.info/blog/411/Using-rel-in-HTML5). Please investigate what is the desired replacement (`class`? `data-role`?) | priority | declarative cell renderer it should be possible to use a as a cell renderer so one can declare renderers with html any html tag custom element or polymer variable can be used in the template html name last name first note rel is probably not the right attribute to add semantic meaning for a because it should be used only with predefined values and a href please investigate what is the desired replacement class data role | 1 |
685,125 | 23,444,564,991 | IssuesEvent | 2022-08-15 18:14:14 | OpenPrinting/cups | https://api.github.com/repos/OpenPrinting/cups | closed | lpoptions -d as root | bug priority-medium | Yair Yarom submitted [Debian bug #1008053](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1008053) and observed that running lpoptions as root does not update /etc/cups/lpoptions but /root/.cups/lpoptions. He was using 2.3.3op2-3+deb11u1 and included a patch in the report.
I observe the same behaviour with 2.4.2 but not with 2.2.10-6+deb10u4. | 1.0 | lpoptions -d as root - Yair Yarom submitted [Debian bug #1008053](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1008053) and observed that running lpoptions as root does not update /etc/cups/lpoptions but /root/.cups/lpoptions. He was using 2.3.3op2-3+deb11u1 and included a patch in the report.
I observe the same behaviour with 2.4.2 but not with 2.2.10-6+deb10u4. | priority | lpoptions d as root yair yarom submitted and observed that running lpoptions as root does not update etc cups lpoptions but root cups lpoptions he was using and included a patch in the report i observe the same behaviour with but not with | 1 |
487,170 | 14,020,069,336 | IssuesEvent | 2020-10-29 19:06:35 | AY2021S1-CS2103T-T09-1/tp | https://api.github.com/repos/AY2021S1-CS2103T-T09-1/tp | closed | Update help command to include new commands | priority.Medium type.Task | Add `required`, `gem`, `science`, `search`, and `tag` commands into `help`.
Update `edit`, `delete` and `find` command formats. | 1.0 | Update help command to include new commands - Add `required`, `gem`, `science`, `search`, and `tag` commands into `help`.
Update `edit`, `delete` and `find` command formats. | priority | update help command to include new commands add required gem science search and tag commands into help update edit delete and find command formats | 1 |
326,927 | 9,962,608,749 | IssuesEvent | 2019-07-07 15:56:11 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] : --disableCoverage flag causes parse error | Priority: Medium Status: Accepted Time: ~1 hour Type: Bug | <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
Generating documentation with the --disableCoverage flag causes a parse error. This is the exact command I'm running: "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage". When I remove the --disableCoverage flag and modify nothing else, the documentation is generated just fine.
Stack trace:
Unhandled Rejection at: Promise {
<rejected> TypeError: Cannot read property 'depth' of undefined
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:8:93)
at main (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:175:32)
at Object.ret (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:178:12)
at Object.ret [as routes] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/compiler.js:526:21)
at Object.invokePartialWrapper [as invokePartial] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:72:46)
at eval (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:5:31)
at Object.prog [as fn] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:221:12)
at CompareHelper.helperFunc (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:1759:24)
at Object.<anonymous> (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:3348:38)
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:52:65) } reason: TypeError: Cannot read property 'depth' of undefined
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:8:93)
at main (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:175:32)
at Object.ret (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:178:12)
at Object.ret [as routes] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/compiler.js:526:21)
at Object.invokePartialWrapper [as invokePartial] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:72:46)
at eval (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:5:31)
at Object.prog [as fn] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:221:12)
at CompareHelper.helperFunc (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:1759:24)
at Object.<anonymous> (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:3348:38)
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:52:65)
Rest of error:
Sorry, but there was a problem during parsing or generation of the documentation. Please fill an issue on github. (https://github.com/compodoc/compodoc/issues/new)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the spirit-gui@0.0.0 compodoc script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/vmuser/.npm/_logs/2019-05-23T15_13_52_667Z-debug.log
##### **Operating System, Node.js, npm, compodoc version(s)**
Operating System: Red Hat Enterprise Linux Server release 6.9 (Santiago)
Node.js: v8.11.3
npm: 6.9.0
compodoc version: ^1.1.9
##### **Angular configuration, a `package.json` file in the root folder**
package.json:
{
"name": "spirit-gui",
"version": "0.0.0",
"scripts": {
"ng": "ng",
"start": "ng serve --proxy-config proxy.config.json",
"build": "ng build --dev --deploy-url=/",
"prod": "ng build --prod --aot --deploy-url=/",
"test": "ng test --json",
"lint": "ng lint",
"pree2e": "webdriver-manager update --standalone false --gecko false --ignore_ssl",
"e2e": "ng e2e -uw fal",
"compodoc": "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage",
"serveWithDocs": "npm run compodoc && npm start"
},
"private": true,
"dependencies": {
"@angular/animations": "^6.1.0",
"@angular/common": "^6.1.0",
"@angular/compiler": "^6.1.0",
"@angular/core": "^6.1.0",
"@angular/forms": "^6.1.0",
"@angular/http": "^6.1.0",
"@angular/platform-browser": "^6.1.0",
"@angular/platform-browser-dynamic": "^6.1.0",
"@angular/router": "^6.1.0",
"@ngrx/core": "^1.2.0",
"@ngrx/effects": "^6.1.0",
"@ngrx/schematics": "^6.1.0",
"@ngrx/store": "^6.1.0",
"chart.js": "^2.7.2",
"classlist.js": "^1.1.20150312",
"core-js": "^2.5.4",
"eventsource": "^1.0.7",
"jasmine-marbles": "^0.5.0",
"ng-event-source": "^1.0.13",
"rxjs": "^6.2.2",
"zone.js": "~0.8.26"
},
"devDependencies": {
"@angular-builders/jest": "^1.2.2",
"@angular-devkit/build-angular": "^0.13.8",
"@angular/cli": "~6.1.3",
"@angular/compiler-cli": "^6.1.0",
"@angular/language-service": "^6.1.0",
"@compodoc/compodoc": "^1.1.9",
"@ngrx/store-devtools": "^6.1.0",
"@types/chart.js": "^2.7.31",
"@types/jasmine": "~2.8.6",
"@types/jasminewd2": "~2.0.3",
"@types/node": "~10.9.1",
"codelyzer": "~4.4.4",
"jasmine-marbles": "^0.5.0",
"jasmine-core": "~3.2.1",
"jasmine-spec-reporter": "~4.2.1",
"jest": "^23.6.0",
"jest-canvas-mock": "^1.1.0",
"protractor": "^5.4.0",
"ts-node": "~7.0.1",
"tslint": "~5.11.0",
"typescript": "^2.9.2"
}
}
##### **Compodoc installed globally or locally ?**
Globally
##### **If possible sourcecode of the file where it breaks**
No source code
##### **If possible your terminal logs before the error**
0 info it worked if it ends with ok
1 verbose cli [ '/opt/node-v8.11.3-linux-x64/bin/node',
1 verbose cli '/opt/node-v8.11.3-linux-x64/bin/npm',
1 verbose cli 'run',
1 verbose cli 'compodoc' ]
2 info using npm@6.9.0
3 info using node@v8.11.3
4 verbose run-script [ 'precompodoc', 'compodoc', 'postcompodoc' ]
5 info lifecycle spirit-gui@0.0.0~precompodoc: spirit-gui@0.0.0
6 info lifecycle spirit-gui@0.0.0~compodoc: spirit-gui@0.0.0
7 verbose lifecycle spirit-gui@0.0.0~compodoc: unsafe-perm in lifecycle true
8 verbose lifecycle spirit-gui@0.0.0~compodoc: PATH: /opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/node-gyp-bin:/home/vmuser/dev/SpiritGui/node_modules/.bin:/opt/apache-maven-3.3.9/bin:/opt/node-v8.11.3-linux-x64/bin:/opt/apache-activemq-5.13.3/bin:/opt/PostgreSQL/9.6/bin:/usr/java/jdk1.8.0_181-amd64/bin:/usr/local/bin:/usr/bin:/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/vmuser/bin
9 verbose lifecycle spirit-gui@0.0.0~compodoc: CWD: /home/vmuser/dev/SpiritGui
10 silly lifecycle spirit-gui@0.0.0~compodoc: Args: [ '-c',
10 silly lifecycle 'npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName \'How to Write Documentation\' --disableCoverage' ]
11 silly lifecycle spirit-gui@0.0.0~compodoc: Returned: code: 1 signal: null
12 info lifecycle spirit-gui@0.0.0~compodoc: Failed to exec compodoc script
13 verbose stack Error: spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
13 verbose stack Exit status 1
13 verbose stack at EventEmitter.<anonymous> (/opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/index.js:301:16)
13 verbose stack at emitTwo (events.js:126:13)
13 verbose stack at EventEmitter.emit (events.js:214:7)
13 verbose stack at ChildProcess.<anonymous> (/opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/lib/spawn.js:55:14)
13 verbose stack at emitTwo (events.js:126:13)
13 verbose stack at ChildProcess.emit (events.js:214:7)
13 verbose stack at maybeClose (internal/child_process.js:925:16)
13 verbose stack at Process.ChildProcess._handle.onexit (internal/child_process.js:209:5)
14 verbose pkgid spirit-gui@0.0.0
15 verbose cwd /home/vmuser/dev/SpiritGui
16 verbose Linux 2.6.32-754.el6.x86_64
17 verbose argv "/opt/node-v8.11.3-linux-x64/bin/node" "/opt/node-v8.11.3-linux-x64/bin/npm" "run" "compodoc"
18 verbose node v8.11.3
19 verbose npm v6.9.0
20 error code ELIFECYCLE
21 error errno 1
22 error spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
22 error Exit status 1
23 error Failed at the spirit-gui@0.0.0 compodoc script.
23 error This is probably not a problem with npm. There is likely additional logging output above.
24 verbose exit [ 1, true ]
##### **Motivation for or Use Case**
We are trying to disable coverage because there are a lot of files that we don't want coverage for becasue they are dumb components/ NgRX, which we feel are unnecessary to document. Leaving them in the coverage drags down our code coverage percentage.
##### **Reproduce the error**
pakage.json:
{
"scripts": {
"compodoc": "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage"
}
}
- command used: "npm run compodoc"
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | 1.0 | [BUG] : --disableCoverage flag causes parse error - <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
Generating documentation with the --disableCoverage flag causes a parse error. This is the exact command I'm running: "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage". When I remove the --disableCoverage flag and modify nothing else, the documentation is generated just fine.
Stack trace:
Unhandled Rejection at: Promise {
<rejected> TypeError: Cannot read property 'depth' of undefined
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:8:93)
at main (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:175:32)
at Object.ret (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:178:12)
at Object.ret [as routes] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/compiler.js:526:21)
at Object.invokePartialWrapper [as invokePartial] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:72:46)
at eval (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:5:31)
at Object.prog [as fn] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:221:12)
at CompareHelper.helperFunc (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:1759:24)
at Object.<anonymous> (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:3348:38)
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:52:65) } reason: TypeError: Cannot read property 'depth' of undefined
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:8:93)
at main (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:175:32)
at Object.ret (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:178:12)
at Object.ret [as routes] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/compiler.js:526:21)
at Object.invokePartialWrapper [as invokePartial] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:72:46)
at eval (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:5:31)
at Object.prog [as fn] (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/runtime.js:221:12)
at CompareHelper.helperFunc (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:1759:24)
at Object.<anonymous> (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/dist/index-cli.js:3348:38)
at Object.eval [as main] (eval at createFunctionContext (/home/vmuser/dev/SpiritGui/node_modules/@compodoc/compodoc/node_modules/handlebars/dist/cjs/handlebars/compiler/javascript-compiler.js:257:23), <anonymous>:52:65)
Rest of error:
Sorry, but there was a problem during parsing or generation of the documentation. Please fill an issue on github. (https://github.com/compodoc/compodoc/issues/new)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the spirit-gui@0.0.0 compodoc script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/vmuser/.npm/_logs/2019-05-23T15_13_52_667Z-debug.log
##### **Operating System, Node.js, npm, compodoc version(s)**
Operating System: Red Hat Enterprise Linux Server release 6.9 (Santiago)
Node.js: v8.11.3
npm: 6.9.0
compodoc version: ^1.1.9
##### **Angular configuration, a `package.json` file in the root folder**
package.json:
{
"name": "spirit-gui",
"version": "0.0.0",
"scripts": {
"ng": "ng",
"start": "ng serve --proxy-config proxy.config.json",
"build": "ng build --dev --deploy-url=/",
"prod": "ng build --prod --aot --deploy-url=/",
"test": "ng test --json",
"lint": "ng lint",
"pree2e": "webdriver-manager update --standalone false --gecko false --ignore_ssl",
"e2e": "ng e2e -uw fal",
"compodoc": "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage",
"serveWithDocs": "npm run compodoc && npm start"
},
"private": true,
"dependencies": {
"@angular/animations": "^6.1.0",
"@angular/common": "^6.1.0",
"@angular/compiler": "^6.1.0",
"@angular/core": "^6.1.0",
"@angular/forms": "^6.1.0",
"@angular/http": "^6.1.0",
"@angular/platform-browser": "^6.1.0",
"@angular/platform-browser-dynamic": "^6.1.0",
"@angular/router": "^6.1.0",
"@ngrx/core": "^1.2.0",
"@ngrx/effects": "^6.1.0",
"@ngrx/schematics": "^6.1.0",
"@ngrx/store": "^6.1.0",
"chart.js": "^2.7.2",
"classlist.js": "^1.1.20150312",
"core-js": "^2.5.4",
"eventsource": "^1.0.7",
"jasmine-marbles": "^0.5.0",
"ng-event-source": "^1.0.13",
"rxjs": "^6.2.2",
"zone.js": "~0.8.26"
},
"devDependencies": {
"@angular-builders/jest": "^1.2.2",
"@angular-devkit/build-angular": "^0.13.8",
"@angular/cli": "~6.1.3",
"@angular/compiler-cli": "^6.1.0",
"@angular/language-service": "^6.1.0",
"@compodoc/compodoc": "^1.1.9",
"@ngrx/store-devtools": "^6.1.0",
"@types/chart.js": "^2.7.31",
"@types/jasmine": "~2.8.6",
"@types/jasminewd2": "~2.0.3",
"@types/node": "~10.9.1",
"codelyzer": "~4.4.4",
"jasmine-marbles": "^0.5.0",
"jasmine-core": "~3.2.1",
"jasmine-spec-reporter": "~4.2.1",
"jest": "^23.6.0",
"jest-canvas-mock": "^1.1.0",
"protractor": "^5.4.0",
"ts-node": "~7.0.1",
"tslint": "~5.11.0",
"typescript": "^2.9.2"
}
}
##### **Compodoc installed globally or locally ?**
Globally
##### **If possible sourcecode of the file where it breaks**
No source code
##### **If possible your terminal logs before the error**
0 info it worked if it ends with ok
1 verbose cli [ '/opt/node-v8.11.3-linux-x64/bin/node',
1 verbose cli '/opt/node-v8.11.3-linux-x64/bin/npm',
1 verbose cli 'run',
1 verbose cli 'compodoc' ]
2 info using npm@6.9.0
3 info using node@v8.11.3
4 verbose run-script [ 'precompodoc', 'compodoc', 'postcompodoc' ]
5 info lifecycle spirit-gui@0.0.0~precompodoc: spirit-gui@0.0.0
6 info lifecycle spirit-gui@0.0.0~compodoc: spirit-gui@0.0.0
7 verbose lifecycle spirit-gui@0.0.0~compodoc: unsafe-perm in lifecycle true
8 verbose lifecycle spirit-gui@0.0.0~compodoc: PATH: /opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/node-gyp-bin:/home/vmuser/dev/SpiritGui/node_modules/.bin:/opt/apache-maven-3.3.9/bin:/opt/node-v8.11.3-linux-x64/bin:/opt/apache-activemq-5.13.3/bin:/opt/PostgreSQL/9.6/bin:/usr/java/jdk1.8.0_181-amd64/bin:/usr/local/bin:/usr/bin:/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/vmuser/bin
9 verbose lifecycle spirit-gui@0.0.0~compodoc: CWD: /home/vmuser/dev/SpiritGui
10 silly lifecycle spirit-gui@0.0.0~compodoc: Args: [ '-c',
10 silly lifecycle 'npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName \'How to Write Documentation\' --disableCoverage' ]
11 silly lifecycle spirit-gui@0.0.0~compodoc: Returned: code: 1 signal: null
12 info lifecycle spirit-gui@0.0.0~compodoc: Failed to exec compodoc script
13 verbose stack Error: spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
13 verbose stack Exit status 1
13 verbose stack at EventEmitter.<anonymous> (/opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/index.js:301:16)
13 verbose stack at emitTwo (events.js:126:13)
13 verbose stack at EventEmitter.emit (events.js:214:7)
13 verbose stack at ChildProcess.<anonymous> (/opt/node-v8.11.3-linux-x64/lib/node_modules/npm/node_modules/npm-lifecycle/lib/spawn.js:55:14)
13 verbose stack at emitTwo (events.js:126:13)
13 verbose stack at ChildProcess.emit (events.js:214:7)
13 verbose stack at maybeClose (internal/child_process.js:925:16)
13 verbose stack at Process.ChildProcess._handle.onexit (internal/child_process.js:209:5)
14 verbose pkgid spirit-gui@0.0.0
15 verbose cwd /home/vmuser/dev/SpiritGui
16 verbose Linux 2.6.32-754.el6.x86_64
17 verbose argv "/opt/node-v8.11.3-linux-x64/bin/node" "/opt/node-v8.11.3-linux-x64/bin/npm" "run" "compodoc"
18 verbose node v8.11.3
19 verbose npm v6.9.0
20 error code ELIFECYCLE
21 error errno 1
22 error spirit-gui@0.0.0 compodoc: `npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage`
22 error Exit status 1
23 error Failed at the spirit-gui@0.0.0 compodoc script.
23 error This is probably not a problem with npm. There is likely additional logging output above.
24 verbose exit [ 1, true ]
##### **Motivation for or Use Case**
We are trying to disable coverage because there are a lot of files that we don't want coverage for becasue they are dumb components/ NgRX, which we feel are unnecessary to document. Leaving them in the coverage drags down our code coverage percentage.
##### **Reproduce the error**
pakage.json:
{
"scripts": {
"compodoc": "npx compodoc -p src/tsconfig.app.json --theme readthedocs --includes additional_documentation --includesName 'How to Write Documentation' --disableCoverage"
}
}
- command used: "npm run compodoc"
<!-- Love compodoc? Please consider supporting our collective:
👉 https://opencollective.com/compodoc/donate --> | priority | disablecoverage flag causes parse error please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue generating documentation with the disablecoverage flag causes a parse error this is the exact command i m running npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage when i remove the disablecoverage flag and modify nothing else the documentation is generated just fine stack trace unhandled rejection at promise typeerror cannot read property depth of undefined at object eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js at main home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at object ret home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at object ret home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler compiler js at object invokepartialwrapper home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js at object prog home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at comparehelper helperfunc home vmuser dev spiritgui node modules compodoc compodoc dist index cli js at object home vmuser dev spiritgui node modules compodoc compodoc dist index cli js at object eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js reason typeerror cannot read property depth of undefined at object eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js at main home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at object ret home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at object ret home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler compiler js at object invokepartialwrapper home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js at object prog home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars runtime js at comparehelper helperfunc home vmuser dev spiritgui node modules compodoc compodoc dist index cli js at object home vmuser dev spiritgui node modules compodoc compodoc dist index cli js at object eval eval at createfunctioncontext home vmuser dev spiritgui node modules compodoc compodoc node modules handlebars dist cjs handlebars compiler javascript compiler js rest of error sorry but there was a problem during parsing or generation of the documentation please fill an issue on github npm err code elifecycle npm err errno npm err spirit gui compodoc npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage npm err exit status npm err npm err failed at the spirit gui compodoc script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home vmuser npm logs debug log operating system node js npm compodoc version s operating system red hat enterprise linux server release santiago node js npm compodoc version angular configuration a package json file in the root folder package json name spirit gui version scripts ng ng start ng serve proxy config proxy config json build ng build dev deploy url prod ng build prod aot deploy url test ng test json lint ng lint webdriver manager update standalone false gecko false ignore ssl ng uw fal compodoc npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage servewithdocs npm run compodoc npm start private true dependencies angular animations angular common angular compiler angular core angular forms angular http angular platform browser angular platform browser dynamic angular router ngrx core ngrx effects ngrx schematics ngrx store chart js classlist js core js eventsource jasmine marbles ng event source rxjs zone js devdependencies angular builders jest angular devkit build angular angular cli angular compiler cli angular language service compodoc compodoc ngrx store devtools types chart js types jasmine types types node codelyzer jasmine marbles jasmine core jasmine spec reporter jest jest canvas mock protractor ts node tslint typescript compodoc installed globally or locally globally if possible sourcecode of the file where it breaks no source code if possible your terminal logs before the error info it worked if it ends with ok verbose cli opt node linux bin node verbose cli opt node linux bin npm verbose cli run verbose cli compodoc info using npm info using node verbose run script info lifecycle spirit gui precompodoc spirit gui info lifecycle spirit gui compodoc spirit gui verbose lifecycle spirit gui compodoc unsafe perm in lifecycle true verbose lifecycle spirit gui compodoc path opt node linux lib node modules npm node modules npm lifecycle node gyp bin home vmuser dev spiritgui node modules bin opt apache maven bin opt node linux bin opt apache activemq bin opt postgresql bin usr java bin usr local bin usr bin bin usr local sbin usr sbin sbin home vmuser bin verbose lifecycle spirit gui compodoc cwd home vmuser dev spiritgui silly lifecycle spirit gui compodoc args c silly lifecycle npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage silly lifecycle spirit gui compodoc returned code signal null info lifecycle spirit gui compodoc failed to exec compodoc script verbose stack error spirit gui compodoc npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage verbose stack exit status verbose stack at eventemitter opt node linux lib node modules npm node modules npm lifecycle index js verbose stack at emittwo events js verbose stack at eventemitter emit events js verbose stack at childprocess opt node linux lib node modules npm node modules npm lifecycle lib spawn js verbose stack at emittwo events js verbose stack at childprocess emit events js verbose stack at maybeclose internal child process js verbose stack at process childprocess handle onexit internal child process js verbose pkgid spirit gui verbose cwd home vmuser dev spiritgui verbose linux verbose argv opt node linux bin node opt node linux bin npm run compodoc verbose node verbose npm error code elifecycle error errno error spirit gui compodoc npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage error exit status error failed at the spirit gui compodoc script error this is probably not a problem with npm there is likely additional logging output above verbose exit motivation for or use case we are trying to disable coverage because there are a lot of files that we don t want coverage for becasue they are dumb components ngrx which we feel are unnecessary to document leaving them in the coverage drags down our code coverage percentage reproduce the error pakage json scripts compodoc npx compodoc p src tsconfig app json theme readthedocs includes additional documentation includesname how to write documentation disablecoverage command used npm run compodoc love compodoc please consider supporting our collective 👉 | 1 |
621,115 | 19,578,054,794 | IssuesEvent | 2022-01-04 17:28:31 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | msgq: NULL handler assertion with data cache enabled | bug priority: medium area: Kernel | **Describe the bug**
Trying to receive a message from a message queue inside a work queue causes a NULL handler assertion.
The call to 'k_msgq_get' from inside the work queue causes this to happen.
**What have you tried to diagnose or workaround this issue?**
I found that this only occurs when the data cache is enabled, I am using a Cortex-M7 chip (Atmel Sam e70).
When I comment out the following lines in the soc.c file the problem goes away:
```
if (!(SCB->CCR & SCB_CCR_DC_Msk)) {
SCB_EnableDCache();
}
```
**To Reproduce**
Here is the relevant code:
```
/* globals */
static struct k_work_q serial_link_work_q;
struct serial_link_rx_packet {
uint8_t *buff;
uint16_t length;
uint16_t index;
};
/* In init function */
BUILD_ASSERT(IS_POWER_OF_2(sizeof(struct serial_link_rx_packet)),
"error: 'struct serial_link_rx_packet' size must be a power of 2");
k_msgq_init(&data->rx_thread.packet_mq,
data->rx_thread.packet_buffer,
sizeof(struct serial_link_rx_packet),
RX_PACKET_Q_MSG_SIZE);
k_work_init_delayable(&data->rx_thread.work, _rx_thread);
k_work_init_delayable(&data->tx_thread.work, _tx_thread);
k_work_queue_start(&serial_link_work_q,
serial_link_work_q_stack,
K_THREAD_STACK_SIZEOF(serial_link_work_q_stack),
WORK_Q_PRIORITY,
NULL);
k_thread_name_set(&serial_link_work_q.thread, "serial_link_workq");
/* In work queue */
k_msgq_get(&thread->packet_mq, &packet, K_FOREVER);
/* In ISR */
k_msgq_put(&data->rx_thread.packet_mq, &packet, K_NO_WAIT);
```
I will see about making a small reproducible example, might be a while because of time constraints.
**Expected behavior**
There should be no assertion failure
**Impact**
I can work around this, maybe add some cache invalidation code or disable cache if I cannot get it working. However, would be nice if this was part of the library.
**Logs and console output**
```
ASSERTION FAIL [handler != ((void *)0)] @ WEST_TOPDIR/zephyr/kernel/work.c:650
[00:00:00.215,000] <err> os: r0/a1: 0x00000004 r1/a2: 0x0000028a r2/a3: 0x40024000
[00:00:00.215,000] <err> os: r3/a4: 0x00000000 r12/ip: 0x00000000 r14/lr: 0x004186d1
[00:00:00.215,000] <err> os: xpsr: 0x61000000
[00:00:00.215,000] <err> os: EXC_RETURN: 0x0
[00:00:00.215,000] <err> os: Faulting instruction address (r15/pc): 0x004186ca
[00:00:00.215,000] <err> os: >>> ZEPHYR FATAL ERROR 4: Kernel panic on CPU 0
[00:00:00.215,000] <err> os: Current thread: 0x20400af8 (serial_link_workq)
[00:00:00.266,000] <err> os: Halting system
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.13.1
- v2.7.0 branch
| 1.0 | msgq: NULL handler assertion with data cache enabled - **Describe the bug**
Trying to receive a message from a message queue inside a work queue causes a NULL handler assertion.
The call to 'k_msgq_get' from inside the work queue causes this to happen.
**What have you tried to diagnose or workaround this issue?**
I found that this only occurs when the data cache is enabled, I am using a Cortex-M7 chip (Atmel Sam e70).
When I comment out the following lines in the soc.c file the problem goes away:
```
if (!(SCB->CCR & SCB_CCR_DC_Msk)) {
SCB_EnableDCache();
}
```
**To Reproduce**
Here is the relevant code:
```
/* globals */
static struct k_work_q serial_link_work_q;
struct serial_link_rx_packet {
uint8_t *buff;
uint16_t length;
uint16_t index;
};
/* In init function */
BUILD_ASSERT(IS_POWER_OF_2(sizeof(struct serial_link_rx_packet)),
"error: 'struct serial_link_rx_packet' size must be a power of 2");
k_msgq_init(&data->rx_thread.packet_mq,
data->rx_thread.packet_buffer,
sizeof(struct serial_link_rx_packet),
RX_PACKET_Q_MSG_SIZE);
k_work_init_delayable(&data->rx_thread.work, _rx_thread);
k_work_init_delayable(&data->tx_thread.work, _tx_thread);
k_work_queue_start(&serial_link_work_q,
serial_link_work_q_stack,
K_THREAD_STACK_SIZEOF(serial_link_work_q_stack),
WORK_Q_PRIORITY,
NULL);
k_thread_name_set(&serial_link_work_q.thread, "serial_link_workq");
/* In work queue */
k_msgq_get(&thread->packet_mq, &packet, K_FOREVER);
/* In ISR */
k_msgq_put(&data->rx_thread.packet_mq, &packet, K_NO_WAIT);
```
I will see about making a small reproducible example, might be a while because of time constraints.
**Expected behavior**
There should be no assertion failure
**Impact**
I can work around this, maybe add some cache invalidation code or disable cache if I cannot get it working. However, would be nice if this was part of the library.
**Logs and console output**
```
ASSERTION FAIL [handler != ((void *)0)] @ WEST_TOPDIR/zephyr/kernel/work.c:650
[00:00:00.215,000] <err> os: r0/a1: 0x00000004 r1/a2: 0x0000028a r2/a3: 0x40024000
[00:00:00.215,000] <err> os: r3/a4: 0x00000000 r12/ip: 0x00000000 r14/lr: 0x004186d1
[00:00:00.215,000] <err> os: xpsr: 0x61000000
[00:00:00.215,000] <err> os: EXC_RETURN: 0x0
[00:00:00.215,000] <err> os: Faulting instruction address (r15/pc): 0x004186ca
[00:00:00.215,000] <err> os: >>> ZEPHYR FATAL ERROR 4: Kernel panic on CPU 0
[00:00:00.215,000] <err> os: Current thread: 0x20400af8 (serial_link_workq)
[00:00:00.266,000] <err> os: Halting system
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.13.1
- v2.7.0 branch
| priority | msgq null handler assertion with data cache enabled describe the bug trying to receive a message from a message queue inside a work queue causes a null handler assertion the call to k msgq get from inside the work queue causes this to happen what have you tried to diagnose or workaround this issue i found that this only occurs when the data cache is enabled i am using a cortex chip atmel sam when i comment out the following lines in the soc c file the problem goes away if scb ccr scb ccr dc msk scb enabledcache to reproduce here is the relevant code globals static struct k work q serial link work q struct serial link rx packet t buff t length t index in init function build assert is power of sizeof struct serial link rx packet error struct serial link rx packet size must be a power of k msgq init data rx thread packet mq data rx thread packet buffer sizeof struct serial link rx packet rx packet q msg size k work init delayable data rx thread work rx thread k work init delayable data tx thread work tx thread k work queue start serial link work q serial link work q stack k thread stack sizeof serial link work q stack work q priority null k thread name set serial link work q thread serial link workq in work queue k msgq get thread packet mq packet k forever in isr k msgq put data rx thread packet mq packet k no wait i will see about making a small reproducible example might be a while because of time constraints expected behavior there should be no assertion failure impact i can work around this maybe add some cache invalidation code or disable cache if i cannot get it working however would be nice if this was part of the library logs and console output assertion fail west topdir zephyr kernel work c os os ip lr os xpsr os exc return os faulting instruction address pc os zephyr fatal error kernel panic on cpu os current thread serial link workq os halting system environment please complete the following information os linux toolchain zephyr sdk branch | 1 |
585,161 | 17,481,205,727 | IssuesEvent | 2021-08-09 02:52:58 | oam-dev/kubevela | https://api.github.com/repos/oam-dev/kubevela | closed | [Feature] Allow users to specify component revision name | type/enhancement good first issue priority/important-soon effort/medium | **Is your feature request related to a problem? Please describe.**
Now our component revision is automatically generated with incremental number. It has some drawbacks:
1. Real users has their own version standard, some version number is business related. They hope the version in kubvela can match with their own revision. So they can have end to end consistency.
2. Users need to apply application first to know the revision, and list/get the created revision, and then specify revision in workflow. If they can specify a revision, the steps can merge to one.
**Describe the solution you'd like**
Add an externalRevision field in component.
```
kind: Application
spec:
components:
- name: mycomp
type: webservice
externalRevision: my-revision-v1
properties:
...
```
1. If externalRevision specified, generate the controller revision with the specified name.
2. Add webhook validation, if the `externalRevision` already exist, check their spec hash, if not equal, report an error.
3. Don't create or update component revision when specified externalRevision already exist.
4. If externalRevision not specified,just work as it is now.
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
| 1.0 | [Feature] Allow users to specify component revision name - **Is your feature request related to a problem? Please describe.**
Now our component revision is automatically generated with incremental number. It has some drawbacks:
1. Real users has their own version standard, some version number is business related. They hope the version in kubvela can match with their own revision. So they can have end to end consistency.
2. Users need to apply application first to know the revision, and list/get the created revision, and then specify revision in workflow. If they can specify a revision, the steps can merge to one.
**Describe the solution you'd like**
Add an externalRevision field in component.
```
kind: Application
spec:
components:
- name: mycomp
type: webservice
externalRevision: my-revision-v1
properties:
...
```
1. If externalRevision specified, generate the controller revision with the specified name.
2. Add webhook validation, if the `externalRevision` already exist, check their spec hash, if not equal, report an error.
3. Don't create or update component revision when specified externalRevision already exist.
4. If externalRevision not specified,just work as it is now.
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
| priority | allow users to specify component revision name is your feature request related to a problem please describe now our component revision is automatically generated with incremental number it has some drawbacks real users has their own version standard some version number is business related they hope the version in kubvela can match with their own revision so they can have end to end consistency users need to apply application first to know the revision and list get the created revision and then specify revision in workflow if they can specify a revision the steps can merge to one describe the solution you d like add an externalrevision field in component kind application spec components name mycomp type webservice externalrevision my revision properties if externalrevision specified generate the controller revision with the specified name add webhook validation if the externalrevision already exist check their spec hash if not equal report an error don t create or update component revision when specified externalrevision already exist if externalrevision not specified,just work as it is now describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 1 |
57,557 | 3,082,893,732 | IssuesEvent | 2015-08-24 03:15:36 | ROGUE-JCTD/MapLoom | https://api.github.com/repos/ROGUE-JCTD/MapLoom | closed | Summary Statistics not working | Priority: Medium Type: Bug | When you are in table view, select a column header and then click on summary statistics. The summary statistics window comes up but it's blank - no charts. | 1.0 | Summary Statistics not working - When you are in table view, select a column header and then click on summary statistics. The summary statistics window comes up but it's blank - no charts. | priority | summary statistics not working when you are in table view select a column header and then click on summary statistics the summary statistics window comes up but it s blank no charts | 1 |
596,933 | 18,150,910,508 | IssuesEvent | 2021-09-26 08:45:09 | sugkiy/ManjarotoArch | https://api.github.com/repos/sugkiy/ManjarotoArch | closed | Prompt for removal of Manjaro wallpapers | priority_medium | Feature for a future version: Prompt the user if they wish to remove the Manjaro wallpapers
I believe they are in the following package:
illyria-wallpaper | 1.0 | Prompt for removal of Manjaro wallpapers - Feature for a future version: Prompt the user if they wish to remove the Manjaro wallpapers
I believe they are in the following package:
illyria-wallpaper | priority | prompt for removal of manjaro wallpapers feature for a future version prompt the user if they wish to remove the manjaro wallpapers i believe they are in the following package illyria wallpaper | 1 |
613,917 | 19,101,538,789 | IssuesEvent | 2021-11-29 23:22:01 | CMPUT301F21T21/detes | https://api.github.com/repos/CMPUT301F21T21/detes | closed | US 05.02.01 - Habit Following and Sharing | Final checkpoint Clarification needed Medium Risk Medium Priority Updated | As a doer, I want to grant or deny another doer permission to follow all my **public** habits.
**Clarification:** The user would like to be notified if another user wants to follow their progress on their public habits, as well as be able to accept or deny their request.
Story Points: 4 | 1.0 | US 05.02.01 - Habit Following and Sharing - As a doer, I want to grant or deny another doer permission to follow all my **public** habits.
**Clarification:** The user would like to be notified if another user wants to follow their progress on their public habits, as well as be able to accept or deny their request.
Story Points: 4 | priority | us habit following and sharing as a doer i want to grant or deny another doer permission to follow all my public habits clarification the user would like to be notified if another user wants to follow their progress on their public habits as well as be able to accept or deny their request story points | 1 |
55,867 | 3,075,077,284 | IssuesEvent | 2015-08-20 11:29:55 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | Расшаривание определенных файлов из каталога | bug imported Priority-Medium | _From [Exciter1...@gmail.com](https://code.google.com/u/111711663099331309075/) on February 08, 2011 09:29:20_
Сделать возможным в меню расшаривания выбор с помощью галки нужного/ненужного файла из каталога(чтоб ненужные не шарить)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=347_ | 1.0 | Расшаривание определенных файлов из каталога - _From [Exciter1...@gmail.com](https://code.google.com/u/111711663099331309075/) on February 08, 2011 09:29:20_
Сделать возможным в меню расшаривания выбор с помощью галки нужного/ненужного файла из каталога(чтоб ненужные не шарить)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=347_ | priority | расшаривание определенных файлов из каталога from on february сделать возможным в меню расшаривания выбор с помощью галки нужного ненужного файла из каталога чтоб ненужные не шарить original issue | 1 |
752,676 | 26,295,115,408 | IssuesEvent | 2023-01-08 21:55:12 | KingSupernova31/RulesGuru | https://api.github.com/repos/KingSupernova31/RulesGuru | opened | Fix the submitted by field | enhancement medium priority | Shouldn't be editable by editors or gurus.
If question doesn't have a submitted by name, display the editor's name. | 1.0 | Fix the submitted by field - Shouldn't be editable by editors or gurus.
If question doesn't have a submitted by name, display the editor's name. | priority | fix the submitted by field shouldn t be editable by editors or gurus if question doesn t have a submitted by name display the editor s name | 1 |
312,931 | 9,555,025,872 | IssuesEvent | 2019-05-03 00:40:54 | Marri/glowfic | https://api.github.com/repos/Marri/glowfic | closed | Move post visibility checking to more complicated scoping | 3. medium priority 8. medium dev | This way, we put the overhead onto the database, but so long as we're consistent about it in the rest of the code it should be fine.
e.g.
```rb
posts = Post.where(privacy: Privacy::PUBLIC)
if logged_in?
posts = posts.or(Post.where(privacy: Privacy::CONSTELLATION)) # or whatever the privacy is
posts = posts.or(Post.where(privacy: Privacy::ACCESS_LIST, id: PostViewer.where(viewer_id: user.id).select(:id))) # or whatever viewer_id and PostViewer are
posts = posts.or(Post.where(privacy: Privacy::PRIVATE, user_id: user.id))
end
```
(Not sure if this is correct, though.) | 1.0 | Move post visibility checking to more complicated scoping - This way, we put the overhead onto the database, but so long as we're consistent about it in the rest of the code it should be fine.
e.g.
```rb
posts = Post.where(privacy: Privacy::PUBLIC)
if logged_in?
posts = posts.or(Post.where(privacy: Privacy::CONSTELLATION)) # or whatever the privacy is
posts = posts.or(Post.where(privacy: Privacy::ACCESS_LIST, id: PostViewer.where(viewer_id: user.id).select(:id))) # or whatever viewer_id and PostViewer are
posts = posts.or(Post.where(privacy: Privacy::PRIVATE, user_id: user.id))
end
```
(Not sure if this is correct, though.) | priority | move post visibility checking to more complicated scoping this way we put the overhead onto the database but so long as we re consistent about it in the rest of the code it should be fine e g rb posts post where privacy privacy public if logged in posts posts or post where privacy privacy constellation or whatever the privacy is posts posts or post where privacy privacy access list id postviewer where viewer id user id select id or whatever viewer id and postviewer are posts posts or post where privacy privacy private user id user id end not sure if this is correct though | 1 |
553,287 | 16,361,932,691 | IssuesEvent | 2021-05-14 10:50:52 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Zoom translation using TranslatePress plugin | bug priority: medium status: needs review | **Describe the bug**
Translating the zoom page using translatepress plugin
**To Reproduce**
Steps to reproduce the behavior:
1. Create new zoom meeting.
2. Go to zoom tab.
3. See error
**Expected behavior**
Either the time schedule should also be translated or not.
**Screenshots**
In english:
https://prnt.sc/tif3ib
In mandarin:
https://prnt.sc/tievn9
**Support ticket links**
https://secure.helpscout.net/conversation/1214571461/81568?folderId=3943005
Related ticket:
https://secure.helpscout.net/conversation/1217911270/82160?folderId=3709080
https://secure.helpscout.net/conversation/1220209934/82541?folderId=3709081
https://secure.helpscout.net/conversation/1374813046/116093 | 1.0 | Zoom translation using TranslatePress plugin - **Describe the bug**
Translating the zoom page using translatepress plugin
**To Reproduce**
Steps to reproduce the behavior:
1. Create new zoom meeting.
2. Go to zoom tab.
3. See error
**Expected behavior**
Either the time schedule should also be translated or not.
**Screenshots**
In english:
https://prnt.sc/tif3ib
In mandarin:
https://prnt.sc/tievn9
**Support ticket links**
https://secure.helpscout.net/conversation/1214571461/81568?folderId=3943005
Related ticket:
https://secure.helpscout.net/conversation/1217911270/82160?folderId=3709080
https://secure.helpscout.net/conversation/1220209934/82541?folderId=3709081
https://secure.helpscout.net/conversation/1374813046/116093 | priority | zoom translation using translatepress plugin describe the bug translating the zoom page using translatepress plugin to reproduce steps to reproduce the behavior create new zoom meeting go to zoom tab see error expected behavior either the time schedule should also be translated or not screenshots in english in mandarin support ticket links related ticket | 1 |
482,554 | 13,909,363,468 | IssuesEvent | 2020-10-20 14:50:18 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Validate JWT using PEM rather than x509 certificate | Priority: Medium SLA customer request enhancement | **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
Receive the following error when setting up a Tyk JWT authorizer with Okta as an authorization server:
```
curl 'https://$tyk-api/$path-with-jwt-authorizer' -H 'User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:67.0) Gecko/20100101 Firefox/67.0' -H 'Accept: */*' -H 'Accept-Language: en-US,en;q=0.5' --compressed -H 'Referer: http://localhost:3000/protected' -H 'Authorization: Bearer $okta-access-token' -H 'Origin: http://localhost:3000' -H 'DNT: 1' -H 'Connection: keep-alive'
```
```json
{
"error": "Key not authorized:no certificates in JWK"
}
```
This is due to Tyk expecting an `x5c` parameter: https://github.com/TykTechnologies/tyk/blob/f69055a931dc50abf3ef884fa9cb7277ae667e4c/gateway/mw_jwt.go#L94-L104 and not using `n` and `e` to generate a PEM if they are present.
**What is the expected behavior?**
Tyk verifies the JWT using the PEM modulo and exponent in Okta's JWT keyset.
Okta do not expose their X509 certificate in their JWKS endpoint anymore (https://stackoverflow.com/questions/38727528/is-okta-oidc-documentation-outdated/38947218#38947218)).
Example keyset response parameters (values omitted):
```
{
"keys": [
{
"alg": "RS256",
"e": ...,
"kid": ...,
"kty": "RSA",
"n": ...,
"use": "sig"
}
]
}
```
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
All versions (`<=v2.8.0` at the time of writing). | 1.0 | Validate JWT using PEM rather than x509 certificate - **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
Receive the following error when setting up a Tyk JWT authorizer with Okta as an authorization server:
```
curl 'https://$tyk-api/$path-with-jwt-authorizer' -H 'User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:67.0) Gecko/20100101 Firefox/67.0' -H 'Accept: */*' -H 'Accept-Language: en-US,en;q=0.5' --compressed -H 'Referer: http://localhost:3000/protected' -H 'Authorization: Bearer $okta-access-token' -H 'Origin: http://localhost:3000' -H 'DNT: 1' -H 'Connection: keep-alive'
```
```json
{
"error": "Key not authorized:no certificates in JWK"
}
```
This is due to Tyk expecting an `x5c` parameter: https://github.com/TykTechnologies/tyk/blob/f69055a931dc50abf3ef884fa9cb7277ae667e4c/gateway/mw_jwt.go#L94-L104 and not using `n` and `e` to generate a PEM if they are present.
**What is the expected behavior?**
Tyk verifies the JWT using the PEM modulo and exponent in Okta's JWT keyset.
Okta do not expose their X509 certificate in their JWKS endpoint anymore (https://stackoverflow.com/questions/38727528/is-okta-oidc-documentation-outdated/38947218#38947218)).
Example keyset response parameters (values omitted):
```
{
"keys": [
{
"alg": "RS256",
"e": ...,
"kid": ...,
"kty": "RSA",
"n": ...,
"use": "sig"
}
]
}
```
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
All versions (`<=v2.8.0` at the time of writing). | priority | validate jwt using pem rather than certificate do you want to request a feature or report a bug feature what is the current behavior receive the following error when setting up a tyk jwt authorizer with okta as an authorization server curl h user agent mozilla macintosh intel mac os x rv gecko firefox h accept h accept language en us en q compressed h referer h authorization bearer okta access token h origin h dnt h connection keep alive json error key not authorized no certificates in jwk this is due to tyk expecting an parameter and not using n and e to generate a pem if they are present what is the expected behavior tyk verifies the jwt using the pem modulo and exponent in okta s jwt keyset okta do not expose their certificate in their jwks endpoint anymore example keyset response parameters values omitted keys alg e kid kty rsa n use sig which versions of tyk affected by this issue did this work in previous versions of tyk all versions at the time of writing | 1 |
246,578 | 7,895,403,635 | IssuesEvent | 2018-06-29 03:03:22 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Update VisIt to use VTK 8.1 | Expected Use: 3 - Occasional Feature Impact: 3 - Medium OS: All Priority: High Support Group: Any version: 2.12.3 | We should update VisIt to use VTK 8.1
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 09/17/2015 03:04 pm
Original update: 04/04/2018 09:59 pm
Ticket number: 2379 | 1.0 | Update VisIt to use VTK 8.1 - We should update VisIt to use VTK 8.1
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 09/17/2015 03:04 pm
Original update: 04/04/2018 09:59 pm
Ticket number: 2379 | priority | update visit to use vtk we should update visit to use vtk redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation pm original update pm ticket number | 1 |
126,427 | 4,995,444,199 | IssuesEvent | 2016-12-09 10:10:20 | emoncms/MyHomeEnergyPlanner | https://api.github.com/repos/emoncms/MyHomeEnergyPlanner | opened | empty libraries | For release Medium priority | When you apply a fabric measure,if the library is empty, it tries to display walls but the list of leasure is empty. The reason is that the library is empty no that there is a bug. Anyway adding a message saying _This library is empty_ would look nicer | 1.0 | empty libraries - When you apply a fabric measure,if the library is empty, it tries to display walls but the list of leasure is empty. The reason is that the library is empty no that there is a bug. Anyway adding a message saying _This library is empty_ would look nicer | priority | empty libraries when you apply a fabric measure if the library is empty it tries to display walls but the list of leasure is empty the reason is that the library is empty no that there is a bug anyway adding a message saying this library is empty would look nicer | 1 |
253,424 | 8,056,076,599 | IssuesEvent | 2018-08-02 11:27:54 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | opened | Get XTENSA boards compliant with default configuration guidelines | area: Boards area: Xtensa enhancement good first issue priority: medium | Created from #7151, dedicated to XTENSA boards
Following introduction of "default board configuration guidelines" (#6858), one part of the work is to move boards already ported to zephyr to conform with guidelines available here:
http://docs.zephyrproject.org/porting/board_porting.html#default-board-configuration
https://github.com/zephyrproject-rtos/zephyr/tree/master/boards/arm/nucleo_f429zi could be used as reference for this update. | 1.0 | Get XTENSA boards compliant with default configuration guidelines - Created from #7151, dedicated to XTENSA boards
Following introduction of "default board configuration guidelines" (#6858), one part of the work is to move boards already ported to zephyr to conform with guidelines available here:
http://docs.zephyrproject.org/porting/board_porting.html#default-board-configuration
https://github.com/zephyrproject-rtos/zephyr/tree/master/boards/arm/nucleo_f429zi could be used as reference for this update. | priority | get xtensa boards compliant with default configuration guidelines created from dedicated to xtensa boards following introduction of default board configuration guidelines one part of the work is to move boards already ported to zephyr to conform with guidelines available here could be used as reference for this update | 1 |
540,083 | 15,800,575,127 | IssuesEvent | 2021-04-03 00:20:07 | JensenJ/EmbargoMC-IssueTracker | https://api.github.com/repos/JensenJ/EmbargoMC-IssueTracker | closed | [BUG] Chunk Errors in NE Island | bug medium-priority | **Describe the bug**
Chunk errors in NE island
**Expected behaviour**
No chunk errors, will need to regen island.
| 1.0 | [BUG] Chunk Errors in NE Island - **Describe the bug**
Chunk errors in NE island
**Expected behaviour**
No chunk errors, will need to regen island.
| priority | chunk errors in ne island describe the bug chunk errors in ne island expected behaviour no chunk errors will need to regen island | 1 |
804,408 | 29,486,323,991 | IssuesEvent | 2023-06-02 10:00:28 | ClockGU/clock-frontend | https://api.github.com/repos/ClockGU/clock-frontend | closed | Wiederholung für Wochenende herausnehmen | medium priority refactor | Beim Wiederholen von Schichten gibt es die Option "an Wochenenden (Sa-So)" - das beißt sich mit der Regel, dass am Sonntag keine Schichten angelegt werden dürfen. | 1.0 | Wiederholung für Wochenende herausnehmen - Beim Wiederholen von Schichten gibt es die Option "an Wochenenden (Sa-So)" - das beißt sich mit der Regel, dass am Sonntag keine Schichten angelegt werden dürfen. | priority | wiederholung für wochenende herausnehmen beim wiederholen von schichten gibt es die option an wochenenden sa so das beißt sich mit der regel dass am sonntag keine schichten angelegt werden dürfen | 1 |
403,936 | 11,849,325,784 | IssuesEvent | 2020-03-24 15:05:09 | cloudfoundry-incubator/kubecf | https://api.github.com/repos/cloudfoundry-incubator/kubecf | closed | Extract CI pipeline configuration to a separate file | Priority: Medium Status: Accepted Type: CI | We need a way for developers to be able to deploy a copy of [the Concourse pipeline](https://github.com/cloudfoundry-incubator/kubecf/tree/master/.concourse) in order to add steps or debug without changing the current deployed one (see also: https://github.com/cloudfoundry-incubator/kubecf/issues/430). This means, we don't want the copy to publish to any of the final S3 buckets or in general publish to public locations (e.g. GitHub commit statuses on this repo).
The preferred approach would be to extract all the relevant configuration (buckets, forks etc) to a separate file that will be interpolated in the pipeline.yaml with some templating mechanism.
A switch to completely disable publish steps (any kind) would also be useful.
| 1.0 | Extract CI pipeline configuration to a separate file - We need a way for developers to be able to deploy a copy of [the Concourse pipeline](https://github.com/cloudfoundry-incubator/kubecf/tree/master/.concourse) in order to add steps or debug without changing the current deployed one (see also: https://github.com/cloudfoundry-incubator/kubecf/issues/430). This means, we don't want the copy to publish to any of the final S3 buckets or in general publish to public locations (e.g. GitHub commit statuses on this repo).
The preferred approach would be to extract all the relevant configuration (buckets, forks etc) to a separate file that will be interpolated in the pipeline.yaml with some templating mechanism.
A switch to completely disable publish steps (any kind) would also be useful.
| priority | extract ci pipeline configuration to a separate file we need a way for developers to be able to deploy a copy of in order to add steps or debug without changing the current deployed one see also this means we don t want the copy to publish to any of the final buckets or in general publish to public locations e g github commit statuses on this repo the preferred approach would be to extract all the relevant configuration buckets forks etc to a separate file that will be interpolated in the pipeline yaml with some templating mechanism a switch to completely disable publish steps any kind would also be useful | 1 |
588,664 | 17,664,208,023 | IssuesEvent | 2021-08-22 05:44:46 | hackforla/lucky-parking | https://api.github.com/repos/hackforla/lucky-parking | closed | bar chart option for sidebar data visualization | type: enhancement role: dev priority: high size: medium | ### Overview
current app has pie chart, explore other options
### Action Items
- [ ] bar chart
- [ ] add raw data, rather percentage
### Resource
[Figma wireframe](https://www.figma.com/file/R3N7mvDtlFbFgQm4OXcMih/Lucky-Parking?node-id=135%3A7) | 1.0 | bar chart option for sidebar data visualization - ### Overview
current app has pie chart, explore other options
### Action Items
- [ ] bar chart
- [ ] add raw data, rather percentage
### Resource
[Figma wireframe](https://www.figma.com/file/R3N7mvDtlFbFgQm4OXcMih/Lucky-Parking?node-id=135%3A7) | priority | bar chart option for sidebar data visualization overview current app has pie chart explore other options action items bar chart add raw data rather percentage resource | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.