Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
125,326 | 4,955,587,017 | IssuesEvent | 2016-12-01 20:51:27 | caitlynmayers/dukes | https://api.github.com/repos/caitlynmayers/dukes | opened | Shopping Cart: Create Account Alignment | Low Priority Style | Upon checking out, "create an account" is on a separate line than the check box
<img width="189" alt="screen shot 2016-12-01 at 3 47 42 pm" src="https://cloud.githubusercontent.com/assets/24302252/20812010/fac027ea-b7dd-11e6-9e92-bb4a97466f82.png">
| 1.0 | Shopping Cart: Create Account Alignment - Upon checking out, "create an account" is on a separate line than the check box
<img width="189" alt="screen shot 2016-12-01 at 3 47 42 pm" src="https://cloud.githubusercontent.com/assets/24302252/20812010/fac027ea-b7dd-11e6-9e92-bb4a97466f82.png">
| priority | shopping cart create account alignment upon checking out create an account is on a separate line than the check box img width alt screen shot at pm src | 1 |
17,492 | 2,615,145,540 | IssuesEvent | 2015-03-01 06:21:26 | chrsmith/html5rocks | https://api.github.com/repos/chrsmith/html5rocks | closed | Need hook script for site update | auto-migrated Maintenance Milestone-2 Priority-Low Type-Enhancement | ```
- one for updating/building the slides cache manifest
- one for .zip'ing the studio samples
First is done, but should be put into a .sh script on in place of running
appcfg.py update.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 2 Aug 2010 at 4:52 | 1.0 | Need hook script for site update - ```
- one for updating/building the slides cache manifest
- one for .zip'ing the studio samples
First is done, but should be put into a .sh script on in place of running
appcfg.py update.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 2 Aug 2010 at 4:52 | priority | need hook script for site update one for updating building the slides cache manifest one for zip ing the studio samples first is done but should be put into a sh script on in place of running appcfg py update original issue reported on code google com by ericbide com on aug at | 1 |
683,499 | 23,384,600,029 | IssuesEvent | 2022-08-11 12:45:19 | TheYellowArchitect/doubledamnation | https://api.github.com/repos/TheYellowArchitect/doubledamnation | opened | Remake Enemy Behaviour | enemy ai low priority | It is no secret that the code for the enemies is **not** well-designed. After all, it was made back in 2018, I didn't even know a design pattern back then.
I didn't read on AI, it was quite literally handmade without any guidance, I still remember the "breakthroughs" and how making the AI felt like exploring, happy times.
There are some good things, like the split of enemy onto pathfinder and behaviour, and the attacks being split in classes, hence monsters are "modular" (e.g. hollow with some clicks, could get the jump attack of satyr, or the rush attack of centaur and minotaur)
However! There are 3 horrible things, which linger to this very day, as I didn't know anything about AI, and I was coding whatever came into my head:
1. **Attack Detection Hitbox and Attack Damage Hitbox are MERGED**
Yes, this is as horrible as it sounds. The same hitbox which says "if in X range, use your first attack" is also the hitbox which checks for applying damage within X range.
It isn't really noticeable in-game, because with the above, enemies attack when they get in range of their attack. But it is noticeable for Harpy's spin attack.
Harpy's spinattack is clunky/janky, because of this flawed design.
2. **Sidewall Bug**
There was a weird bug I cannot recall, when the AI was adjacent to a wall. Related to satyr, probably (this was back in late 2018, dont blame me)
To "fix" this, I did a sidewall check. Which did "work" but I didn't see at the time that it influenced other behaviour.
While I am certain the fix is simple, just refreshing myself on that codebase and what the original bug was, will take a week, and this is a chore.
It is a serious bug, however, because enemies at wall sides have bugged AI.
3. Flying Type
Instead of splitting EnemyBehaviour into Grounded and Flying, I slapped Flying code onto default EnemyBehaviour. And if (flying) then bloated code overrides default.
Not a bug, not wrong, but it makes the code ugly and bloated. | 1.0 | Remake Enemy Behaviour - It is no secret that the code for the enemies is **not** well-designed. After all, it was made back in 2018, I didn't even know a design pattern back then.
I didn't read on AI, it was quite literally handmade without any guidance, I still remember the "breakthroughs" and how making the AI felt like exploring, happy times.
There are some good things, like the split of enemy onto pathfinder and behaviour, and the attacks being split in classes, hence monsters are "modular" (e.g. hollow with some clicks, could get the jump attack of satyr, or the rush attack of centaur and minotaur)
However! There are 3 horrible things, which linger to this very day, as I didn't know anything about AI, and I was coding whatever came into my head:
1. **Attack Detection Hitbox and Attack Damage Hitbox are MERGED**
Yes, this is as horrible as it sounds. The same hitbox which says "if in X range, use your first attack" is also the hitbox which checks for applying damage within X range.
It isn't really noticeable in-game, because with the above, enemies attack when they get in range of their attack. But it is noticeable for Harpy's spin attack.
Harpy's spinattack is clunky/janky, because of this flawed design.
2. **Sidewall Bug**
There was a weird bug I cannot recall, when the AI was adjacent to a wall. Related to satyr, probably (this was back in late 2018, dont blame me)
To "fix" this, I did a sidewall check. Which did "work" but I didn't see at the time that it influenced other behaviour.
While I am certain the fix is simple, just refreshing myself on that codebase and what the original bug was, will take a week, and this is a chore.
It is a serious bug, however, because enemies at wall sides have bugged AI.
3. Flying Type
Instead of splitting EnemyBehaviour into Grounded and Flying, I slapped Flying code onto default EnemyBehaviour. And if (flying) then bloated code overrides default.
Not a bug, not wrong, but it makes the code ugly and bloated. | priority | remake enemy behaviour it is no secret that the code for the enemies is not well designed after all it was made back in i didn t even know a design pattern back then i didn t read on ai it was quite literally handmade without any guidance i still remember the breakthroughs and how making the ai felt like exploring happy times there are some good things like the split of enemy onto pathfinder and behaviour and the attacks being split in classes hence monsters are modular e g hollow with some clicks could get the jump attack of satyr or the rush attack of centaur and minotaur however there are horrible things which linger to this very day as i didn t know anything about ai and i was coding whatever came into my head attack detection hitbox and attack damage hitbox are merged yes this is as horrible as it sounds the same hitbox which says if in x range use your first attack is also the hitbox which checks for applying damage within x range it isn t really noticeable in game because with the above enemies attack when they get in range of their attack but it is noticeable for harpy s spin attack harpy s spinattack is clunky janky because of this flawed design sidewall bug there was a weird bug i cannot recall when the ai was adjacent to a wall related to satyr probably this was back in late dont blame me to fix this i did a sidewall check which did work but i didn t see at the time that it influenced other behaviour while i am certain the fix is simple just refreshing myself on that codebase and what the original bug was will take a week and this is a chore it is a serious bug however because enemies at wall sides have bugged ai flying type instead of splitting enemybehaviour into grounded and flying i slapped flying code onto default enemybehaviour and if flying then bloated code overrides default not a bug not wrong but it makes the code ugly and bloated | 1 |
780,854 | 27,411,010,206 | IssuesEvent | 2023-03-01 10:31:25 | horizon-efrei/HorizonBot | https://api.github.com/repos/horizon-efrei/HorizonBot | closed | Feature pour délégués: TODO-list de devoirs | type: feature difficulty: complex status: awaiting approval scope: class groups priority: lowest | Ensemble de commandes pour gérer de façon semi-automatique une liste de devoirs à faire pour les délégués dans les classes
Commandes à prévoir: !todo create <liste des matières>, !todo add <matière> <intitulé> <date due>, !todo remove <matière/intitulé>, !todo edit <intitulé>, !todo archive
Exemple de layout (devoirs à faire):
--- DEVOIRS A FAIRE (S8) ----
**Natural Language Processing (NLP): (projet terminé ❌ DE Passé ❌)**
- Amazon TP Analysis (01/02)
- TP1: Analysis of Presidential Speeches (05/02)
- Envoyer un HTML/PDF du Notebook (avant le 06/02)
Exemple de layout archive:
--- DEVOIRS FAITS (S7) ----
**Big Data Frameworks: (projet terminé :ballot_box_with_check: DE Passé :ballot_box_with_check: )**
- TP4 Lab1 avec Markdown (15 Oct. 23:59 sur Moodle)
- TP4 Lab2 (Yarn/JAVA) (12 Nov. 23:59 sur Moodle)
- TP5 (Hive) (15 Nov. 23:59 sur Moodle)
- TP6 (HBase) (23 Nov, pas de rendu) | 1.0 | Feature pour délégués: TODO-list de devoirs - Ensemble de commandes pour gérer de façon semi-automatique une liste de devoirs à faire pour les délégués dans les classes
Commandes à prévoir: !todo create <liste des matières>, !todo add <matière> <intitulé> <date due>, !todo remove <matière/intitulé>, !todo edit <intitulé>, !todo archive
Exemple de layout (devoirs à faire):
--- DEVOIRS A FAIRE (S8) ----
**Natural Language Processing (NLP): (projet terminé ❌ DE Passé ❌)**
- Amazon TP Analysis (01/02)
- TP1: Analysis of Presidential Speeches (05/02)
- Envoyer un HTML/PDF du Notebook (avant le 06/02)
Exemple de layout archive:
--- DEVOIRS FAITS (S7) ----
**Big Data Frameworks: (projet terminé :ballot_box_with_check: DE Passé :ballot_box_with_check: )**
- TP4 Lab1 avec Markdown (15 Oct. 23:59 sur Moodle)
- TP4 Lab2 (Yarn/JAVA) (12 Nov. 23:59 sur Moodle)
- TP5 (Hive) (15 Nov. 23:59 sur Moodle)
- TP6 (HBase) (23 Nov, pas de rendu) | priority | feature pour délégués todo list de devoirs ensemble de commandes pour gérer de façon semi automatique une liste de devoirs à faire pour les délégués dans les classes commandes à prévoir todo create todo add todo remove todo edit todo archive exemple de layout devoirs à faire devoirs a faire natural language processing nlp projet terminé ❌ de passé ❌ amazon tp analysis analysis of presidential speeches envoyer un html pdf du notebook avant le exemple de layout archive devoirs faits big data frameworks projet terminé ballot box with check de passé ballot box with check avec markdown oct sur moodle yarn java nov sur moodle hive nov sur moodle hbase nov pas de rendu | 1 |
250,614 | 7,979,146,271 | IssuesEvent | 2018-07-17 20:41:58 | conveyal/analysis-ui | https://api.github.com/repos/conveyal/analysis-ui | closed | Support email should not be hard-coded | low priority small task | We may want to specify the support email in the same place we specify API keys, instead of https://github.com/conveyal/analysis-ui/blob/c9a53e9f74386de40450d5d9e4d3b0f852b14b0a/lib/components/application.js#L245
This would make sure that our support email address only shows up for our supported deployments. | 1.0 | Support email should not be hard-coded - We may want to specify the support email in the same place we specify API keys, instead of https://github.com/conveyal/analysis-ui/blob/c9a53e9f74386de40450d5d9e4d3b0f852b14b0a/lib/components/application.js#L245
This would make sure that our support email address only shows up for our supported deployments. | priority | support email should not be hard coded we may want to specify the support email in the same place we specify api keys instead of this would make sure that our support email address only shows up for our supported deployments | 1 |
470,732 | 13,543,433,083 | IssuesEvent | 2020-09-16 18:57:16 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :212426] Unrecoverable parse warning in drivers/wifi/eswifi/eswifi_socket_offload.c | Coverity bug priority: low |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/66bd06a7d1f9e4682faafbc551046af695fa1060/drivers/wifi/eswifi/eswifi_socket_offload.c#L502
Category: Parse warnings
Function: ``
Component: Drivers
CID: [212426](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=212426)
Details:
```
496 (const struct fd_op_vtable *)
497 &eswifi_socket_fd_op_vtable);
498
499 return fd;
500 }
501
>>> CID 212426: Parse warnings (PARSE_ERROR)
>>> function "eswifi_socket_close" has already been defined
502 static int eswifi_socket_close(void *obj)
503 {
504 return eswifi_socket_close(sd);
505 }
506
507 static int eswifi_socket_ioctl(void *obj, unsigned int request, va_list args)
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID :212426] Unrecoverable parse warning in drivers/wifi/eswifi/eswifi_socket_offload.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/66bd06a7d1f9e4682faafbc551046af695fa1060/drivers/wifi/eswifi/eswifi_socket_offload.c#L502
Category: Parse warnings
Function: ``
Component: Drivers
CID: [212426](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=212426)
Details:
```
496 (const struct fd_op_vtable *)
497 &eswifi_socket_fd_op_vtable);
498
499 return fd;
500 }
501
>>> CID 212426: Parse warnings (PARSE_ERROR)
>>> function "eswifi_socket_close" has already been defined
502 static int eswifi_socket_close(void *obj)
503 {
504 return eswifi_socket_close(sd);
505 }
506
507 static int eswifi_socket_ioctl(void *obj, unsigned int request, va_list args)
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | unrecoverable parse warning in drivers wifi eswifi eswifi socket offload c static code scan issues found in file category parse warnings function component drivers cid details const struct fd op vtable eswifi socket fd op vtable return fd cid parse warnings parse error function eswifi socket close has already been defined static int eswifi socket close void obj return eswifi socket close sd static int eswifi socket ioctl void obj unsigned int request va list args please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
806,219 | 29,806,536,787 | IssuesEvent | 2023-06-16 12:08:16 | penrose/penrose | https://api.github.com/repos/penrose/penrose | opened | feat: Make it possible to "bake" diagrams | kind:enhancement system:optimization system:style kind:usability priority:feature request layout priority:low | ### Issue
Penrose uses random sampling to initialize variables (e.g., those marked `?` in a Style program, or shape properties that are not explicitly specified). If a user is happy with a particular instance of a diagram, they would like to be able to reproduce this diagram later down the road. For this reason, we introduced at some point the notion of a _variation_, i.e., a string that is used to determine the random seed prior to diagram generation. However, variations do not completely solve the problem of reproducibility, since changes to the system itself (e.g,. the compiler or optimizer) can also change the appearance of the final diagram. Hence, it would be nice to have a more "iron clad" solution for finalizing diagrams.
One answer is to simply say: the final diagram can be stored as an SVG. However, saving as SVG does not preserve editability of non-optimized features of the diagrams (e.g., colors, line widths, text labels, etc.).
### Proposal
Instead, this issue raises the following proposal:
> Add functionality to "bake" diagrams, by writing a Style file where all optimized attributes are explicitly overriden with the optimized constants.
For instance, suppose we have the programs
```
-- Domain
type Circle
```
```
-- Substance
Circle c1
```
```
-- Style
canvas {
width = 100
height = 100
}
forall Circle C {
shape C.icon = Circle {
r: 25
center: (?,?)
}
}
```
and the optimizer yields a center `(-20.21, 15.45)` for the circle. Then to bake this diagram, we would append the following lines to the Style program:
```
forall Circle `c1` {
override `c1`.icon.center.x = -20.21
override `c1`.icon.center.y = 15.45
}
```
This feature could be exposed by an additional IDE button like `Bake` that downloads the augmented Style, and a command line option `--bake-style` that likewise dumps the augmented Style.
### Implementation
There's probably some grungy stuff that would have to happen to implement this feature, but the core thing is just getting a list of (i) optimized variables, (ii) their paths in the Style program, and (iii) their final, optimized values, which shouldn't be too hard to get our hands on.
### Field-specific baking
An elaboration on this design would be to allow the Style file to specify properties that should be baked, using the usual selector paradigm. E.g., one could write
```
forall Circle `c1` {
bake `c1`.icon.center.x
bake `c1`.icon.center.y
}
```
to bake the center of one particular named `Circle`, or
```
forall Circle C {
bake C.icon.center.x
bake C.icon.center.y
}
```
to bake the centers of all circles.
| 2.0 | feat: Make it possible to "bake" diagrams - ### Issue
Penrose uses random sampling to initialize variables (e.g., those marked `?` in a Style program, or shape properties that are not explicitly specified). If a user is happy with a particular instance of a diagram, they would like to be able to reproduce this diagram later down the road. For this reason, we introduced at some point the notion of a _variation_, i.e., a string that is used to determine the random seed prior to diagram generation. However, variations do not completely solve the problem of reproducibility, since changes to the system itself (e.g,. the compiler or optimizer) can also change the appearance of the final diagram. Hence, it would be nice to have a more "iron clad" solution for finalizing diagrams.
One answer is to simply say: the final diagram can be stored as an SVG. However, saving as SVG does not preserve editability of non-optimized features of the diagrams (e.g., colors, line widths, text labels, etc.).
### Proposal
Instead, this issue raises the following proposal:
> Add functionality to "bake" diagrams, by writing a Style file where all optimized attributes are explicitly overriden with the optimized constants.
For instance, suppose we have the programs
```
-- Domain
type Circle
```
```
-- Substance
Circle c1
```
```
-- Style
canvas {
width = 100
height = 100
}
forall Circle C {
shape C.icon = Circle {
r: 25
center: (?,?)
}
}
```
and the optimizer yields a center `(-20.21, 15.45)` for the circle. Then to bake this diagram, we would append the following lines to the Style program:
```
forall Circle `c1` {
override `c1`.icon.center.x = -20.21
override `c1`.icon.center.y = 15.45
}
```
This feature could be exposed by an additional IDE button like `Bake` that downloads the augmented Style, and a command line option `--bake-style` that likewise dumps the augmented Style.
### Implementation
There's probably some grungy stuff that would have to happen to implement this feature, but the core thing is just getting a list of (i) optimized variables, (ii) their paths in the Style program, and (iii) their final, optimized values, which shouldn't be too hard to get our hands on.
### Field-specific baking
An elaboration on this design would be to allow the Style file to specify properties that should be baked, using the usual selector paradigm. E.g., one could write
```
forall Circle `c1` {
bake `c1`.icon.center.x
bake `c1`.icon.center.y
}
```
to bake the center of one particular named `Circle`, or
```
forall Circle C {
bake C.icon.center.x
bake C.icon.center.y
}
```
to bake the centers of all circles.
| priority | feat make it possible to bake diagrams issue penrose uses random sampling to initialize variables e g those marked in a style program or shape properties that are not explicitly specified if a user is happy with a particular instance of a diagram they would like to be able to reproduce this diagram later down the road for this reason we introduced at some point the notion of a variation i e a string that is used to determine the random seed prior to diagram generation however variations do not completely solve the problem of reproducibility since changes to the system itself e g the compiler or optimizer can also change the appearance of the final diagram hence it would be nice to have a more iron clad solution for finalizing diagrams one answer is to simply say the final diagram can be stored as an svg however saving as svg does not preserve editability of non optimized features of the diagrams e g colors line widths text labels etc proposal instead this issue raises the following proposal add functionality to bake diagrams by writing a style file where all optimized attributes are explicitly overriden with the optimized constants for instance suppose we have the programs domain type circle substance circle style canvas width height forall circle c shape c icon circle r center and the optimizer yields a center for the circle then to bake this diagram we would append the following lines to the style program forall circle override icon center x override icon center y this feature could be exposed by an additional ide button like bake that downloads the augmented style and a command line option bake style that likewise dumps the augmented style implementation there s probably some grungy stuff that would have to happen to implement this feature but the core thing is just getting a list of i optimized variables ii their paths in the style program and iii their final optimized values which shouldn t be too hard to get our hands on field specific baking an elaboration on this design would be to allow the style file to specify properties that should be baked using the usual selector paradigm e g one could write forall circle bake icon center x bake icon center y to bake the center of one particular named circle or forall circle c bake c icon center x bake c icon center y to bake the centers of all circles | 1 |
78,014 | 3,508,687,744 | IssuesEvent | 2016-01-08 19:02:54 | LunaNode/lobster | https://api.github.com/repos/LunaNode/lobster | closed | SolusVM with OpenVZ: burst memory on creation | low-priority | Context: for SolusVM we only require setup of a single Lobster plan. Then we use customdisk, customcpu, etc. to configure the plan so that it matches the one in the Lobster database.
For OpenVZ, the vserver-create returns an invalid custom burst memory error if we set custommemory. Currently we are using a work-around where we change the memory after provisioning instead of initially. We should improve on this if possible. | 1.0 | SolusVM with OpenVZ: burst memory on creation - Context: for SolusVM we only require setup of a single Lobster plan. Then we use customdisk, customcpu, etc. to configure the plan so that it matches the one in the Lobster database.
For OpenVZ, the vserver-create returns an invalid custom burst memory error if we set custommemory. Currently we are using a work-around where we change the memory after provisioning instead of initially. We should improve on this if possible. | priority | solusvm with openvz burst memory on creation context for solusvm we only require setup of a single lobster plan then we use customdisk customcpu etc to configure the plan so that it matches the one in the lobster database for openvz the vserver create returns an invalid custom burst memory error if we set custommemory currently we are using a work around where we change the memory after provisioning instead of initially we should improve on this if possible | 1 |
159,825 | 6,062,201,755 | IssuesEvent | 2017-06-14 08:50:56 | python/mypy | https://api.github.com/repos/python/mypy | opened | Use plugin to support SQLAlchemy table definitions | feature priority-2-low topic-plugins | We'll also need stubs for SQLAlchemy as they were removed from typeshed some time ago. Current work on stubs is happening at https://github.com/JelleZijlstra/sqlalchemy-stubs. | 1.0 | Use plugin to support SQLAlchemy table definitions - We'll also need stubs for SQLAlchemy as they were removed from typeshed some time ago. Current work on stubs is happening at https://github.com/JelleZijlstra/sqlalchemy-stubs. | priority | use plugin to support sqlalchemy table definitions we ll also need stubs for sqlalchemy as they were removed from typeshed some time ago current work on stubs is happening at | 1 |
777,495 | 27,281,970,566 | IssuesEvent | 2023-02-23 10:43:19 | uhh-cms/columnflow | https://api.github.com/repos/uhh-cms/columnflow | opened | Add tasks and helpers to write pyhf workspaces | enhancement low-priority | We define our statistical models in an experiment agnostic way, allowing for various formats for exporting actual fit models (stat. model + data). So far, we have a combine datacard writer and an accompanying task that creates cards.
We should provide the same mechanism for Pyhf workspaces, which should be fairly easy to achieve, but could help us leveraging bunch of benefits. | 1.0 | Add tasks and helpers to write pyhf workspaces - We define our statistical models in an experiment agnostic way, allowing for various formats for exporting actual fit models (stat. model + data). So far, we have a combine datacard writer and an accompanying task that creates cards.
We should provide the same mechanism for Pyhf workspaces, which should be fairly easy to achieve, but could help us leveraging bunch of benefits. | priority | add tasks and helpers to write pyhf workspaces we define our statistical models in an experiment agnostic way allowing for various formats for exporting actual fit models stat model data so far we have a combine datacard writer and an accompanying task that creates cards we should provide the same mechanism for pyhf workspaces which should be fairly easy to achieve but could help us leveraging bunch of benefits | 1 |
445,948 | 12,838,084,003 | IssuesEvent | 2020-07-07 16:49:55 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | closed | Duplicate ID caused by Optimize CSS Delivery option | community effort: [S] module: file optimization needs: acceptance criteria priority: low type: bug | **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version ✓
- Used the search feature to ensure that the bug hasn’t been reported before ✓
**Describe the bug**
Have developed a theme for a client and checked my theme with w3 validator. There are some Errors caused by a duplicate id for all Stylesheet and JavaScript Files.
Its came from the noscript at the bottom of the page, where i think you pushed it.
And the other warnings is coused by type attributes on css/js files.
So there are unnecessary, maybe you can delete them.
**To Reproduce**
Steps to reproduce the behavior:
1. Activate minimize files
2. logout
3. check page with w3 validator https://validator.w3.org
4. See error
**Expected behavior**
No w3c errors caused by wp-rocket? xD
**Screenshots**

**Backlog Grooming**
- [x] Reproduce the problem
- [x] Identify the root cause
- [x] Scope a solution
- [x] Estimate the effort | 1.0 | Duplicate ID caused by Optimize CSS Delivery option - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version ✓
- Used the search feature to ensure that the bug hasn’t been reported before ✓
**Describe the bug**
Have developed a theme for a client and checked my theme with w3 validator. There are some Errors caused by a duplicate id for all Stylesheet and JavaScript Files.
Its came from the noscript at the bottom of the page, where i think you pushed it.
And the other warnings is coused by type attributes on css/js files.
So there are unnecessary, maybe you can delete them.
**To Reproduce**
Steps to reproduce the behavior:
1. Activate minimize files
2. logout
3. check page with w3 validator https://validator.w3.org
4. See error
**Expected behavior**
No w3c errors caused by wp-rocket? xD
**Screenshots**

**Backlog Grooming**
- [x] Reproduce the problem
- [x] Identify the root cause
- [x] Scope a solution
- [x] Estimate the effort | priority | duplicate id caused by optimize css delivery option before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version ✓ used the search feature to ensure that the bug hasn’t been reported before ✓ describe the bug have developed a theme for a client and checked my theme with validator there are some errors caused by a duplicate id for all stylesheet and javascript files its came from the noscript at the bottom of the page where i think you pushed it and the other warnings is coused by type attributes on css js files so there are unnecessary maybe you can delete them to reproduce steps to reproduce the behavior activate minimize files logout check page with validator see error expected behavior no errors caused by wp rocket xd screenshots backlog grooming reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
824,789 | 31,199,304,361 | IssuesEvent | 2023-08-18 00:36:00 | awslabs/aws-ec2rescue-linux | https://api.github.com/repos/awslabs/aws-ec2rescue-linux | closed | Parse metrics from metrics collect modules | enhancement low priority request for comment | Similar reasoning behind this as #72
We could potentially use the postdiag modules here for this - basically, the ask would be to take the relevant metrics from collect modules and parse them into json for use with a monitoring system.
Since the output from tools is nonstandard we'd need to handle parsing on a per module basis, so this issue is probably a combination of work on specific postdiag modules for parsing if we feel that's the best route here, and determining what it would look like for making them available.
Simples solution would require no real work on the framework - we just run the postdiagnostic modules and the output appears in the default location. However, with #72 in mind we might want to output them in stdout for programmatic consumption there. | 1.0 | Parse metrics from metrics collect modules - Similar reasoning behind this as #72
We could potentially use the postdiag modules here for this - basically, the ask would be to take the relevant metrics from collect modules and parse them into json for use with a monitoring system.
Since the output from tools is nonstandard we'd need to handle parsing on a per module basis, so this issue is probably a combination of work on specific postdiag modules for parsing if we feel that's the best route here, and determining what it would look like for making them available.
Simples solution would require no real work on the framework - we just run the postdiagnostic modules and the output appears in the default location. However, with #72 in mind we might want to output them in stdout for programmatic consumption there. | priority | parse metrics from metrics collect modules similar reasoning behind this as we could potentially use the postdiag modules here for this basically the ask would be to take the relevant metrics from collect modules and parse them into json for use with a monitoring system since the output from tools is nonstandard we d need to handle parsing on a per module basis so this issue is probably a combination of work on specific postdiag modules for parsing if we feel that s the best route here and determining what it would look like for making them available simples solution would require no real work on the framework we just run the postdiagnostic modules and the output appears in the default location however with in mind we might want to output them in stdout for programmatic consumption there | 1 |
604,790 | 18,718,973,372 | IssuesEvent | 2021-11-03 09:34:37 | aau-giraf/weekplanner | https://api.github.com/repos/aau-giraf/weekplanner | closed | As a user I would like messages that are displayed to me to be consistent and grammatically correct | Type: chore Priority: low Good First Issue Trustee | **Is your feature request related to a problem? Please describe.**
It seems a bit lazy to not use the correct plural or singular form when managing items in the application, such as deleting and copying. There are also many instances where a question mark is missing.
**Describe the solution you'd like**
A ternary expression should be able to detect whether one or more items are being managed, and the parentheses with "(er)" and titles on warnings should therefore be of the correct form.
Generally, just go through the messages and warnings and make sure they use correct punctuation and grammar.
**Additional context**
Some cases that were found:
Should be "Aflys aktivitet" and "Vil du markere 1 aktivitet som aflyst?"

Should be "Slet aktivitet" and "Vil du slette 1 aktivitet?"

Should be "Slet ugeplan" and "Vil du slette 1 ugeplan?"

| 1.0 | As a user I would like messages that are displayed to me to be consistent and grammatically correct - **Is your feature request related to a problem? Please describe.**
It seems a bit lazy to not use the correct plural or singular form when managing items in the application, such as deleting and copying. There are also many instances where a question mark is missing.
**Describe the solution you'd like**
A ternary expression should be able to detect whether one or more items are being managed, and the parentheses with "(er)" and titles on warnings should therefore be of the correct form.
Generally, just go through the messages and warnings and make sure they use correct punctuation and grammar.
**Additional context**
Some cases that were found:
Should be "Aflys aktivitet" and "Vil du markere 1 aktivitet som aflyst?"

Should be "Slet aktivitet" and "Vil du slette 1 aktivitet?"

Should be "Slet ugeplan" and "Vil du slette 1 ugeplan?"

| priority | as a user i would like messages that are displayed to me to be consistent and grammatically correct is your feature request related to a problem please describe it seems a bit lazy to not use the correct plural or singular form when managing items in the application such as deleting and copying there are also many instances where a question mark is missing describe the solution you d like a ternary expression should be able to detect whether one or more items are being managed and the parentheses with er and titles on warnings should therefore be of the correct form generally just go through the messages and warnings and make sure they use correct punctuation and grammar additional context some cases that were found should be aflys aktivitet and vil du markere aktivitet som aflyst should be slet aktivitet and vil du slette aktivitet should be slet ugeplan and vil du slette ugeplan | 1 |
306,274 | 9,383,218,664 | IssuesEvent | 2019-04-05 02:15:21 | squizlabs/PHP_CodeSniffer | https://api.github.com/repos/squizlabs/PHP_CodeSniffer | closed | --config-set installed_paths overwrites itself | Enhancement Low Priority | phpcs 2.8.1
attempted
```php
phpcs --config-set installed_paths /path/to/one
phpcs --config-set installed_paths /path/to/two
```
### result
only `/path/to/two` was added to the installed_paths
### expected
both `/path/to/one` and `/path/to/two` should be added to the installed_paths.
### note
`phpcs --config-set installed_paths /path/to/one,/path/to/two` will add both to the installed_paths. but in the other case it's unexpected to lose `/path/to/one` by using a second call to `--config-set` | 1.0 | --config-set installed_paths overwrites itself - phpcs 2.8.1
attempted
```php
phpcs --config-set installed_paths /path/to/one
phpcs --config-set installed_paths /path/to/two
```
### result
only `/path/to/two` was added to the installed_paths
### expected
both `/path/to/one` and `/path/to/two` should be added to the installed_paths.
### note
`phpcs --config-set installed_paths /path/to/one,/path/to/two` will add both to the installed_paths. but in the other case it's unexpected to lose `/path/to/one` by using a second call to `--config-set` | priority | config set installed paths overwrites itself phpcs attempted php phpcs config set installed paths path to one phpcs config set installed paths path to two result only path to two was added to the installed paths expected both path to one and path to two should be added to the installed paths note phpcs config set installed paths path to one path to two will add both to the installed paths but in the other case it s unexpected to lose path to one by using a second call to config set | 1 |
416,129 | 12,140,046,616 | IssuesEvent | 2020-04-23 19:52:20 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Watchtower HTTP API proposal | Do not close Priority: Low Status: Available Type: Enhancement | Since Watchtower basically watches for docker registries and actively pulls images in order to check for outdated containers, it keeps continuously incrementing the registries' pull counter, making them useless. The counts can no longer be considered as user downloads/pulls.
This issue has been impacting some of my co-workers' personal projects and I have managed to use it as a working case in my bachelor's thesis. My proposal is to implement an HTTP API to enable DockerHub's Webook service to trigger Watchtower instances and make them check for updates only when a new image is pulled. Another use case would be a request made by a CI platform in order to trigger the update process.
I have already started some prototypes and I hope to be implementing a working version soon. | 1.0 | Watchtower HTTP API proposal - Since Watchtower basically watches for docker registries and actively pulls images in order to check for outdated containers, it keeps continuously incrementing the registries' pull counter, making them useless. The counts can no longer be considered as user downloads/pulls.
This issue has been impacting some of my co-workers' personal projects and I have managed to use it as a working case in my bachelor's thesis. My proposal is to implement an HTTP API to enable DockerHub's Webook service to trigger Watchtower instances and make them check for updates only when a new image is pulled. Another use case would be a request made by a CI platform in order to trigger the update process.
I have already started some prototypes and I hope to be implementing a working version soon. | priority | watchtower http api proposal since watchtower basically watches for docker registries and actively pulls images in order to check for outdated containers it keeps continuously incrementing the registries pull counter making them useless the counts can no longer be considered as user downloads pulls this issue has been impacting some of my co workers personal projects and i have managed to use it as a working case in my bachelor s thesis my proposal is to implement an http api to enable dockerhub s webook service to trigger watchtower instances and make them check for updates only when a new image is pulled another use case would be a request made by a ci platform in order to trigger the update process i have already started some prototypes and i hope to be implementing a working version soon | 1 |
831,251 | 32,042,808,542 | IssuesEvent | 2023-09-22 21:01:45 | cucapra/filament | https://api.github.com/repos/cucapra/filament | opened | Testing floating point/iterative divider implementations for accuracy instead of conformity | low priority C: tools | Currently, the `floating-point` directory of implementations tests for conformity among a couple other implementations, most of which are directly derived from an implementation in verilog. The adder was taken externally from [here](https://github.com/suhasr1991/5-Stage-Pipelined-IEEE-Single-Precision-Floating-Point-Adder-Design).
The problem with this is that if the initial verilog designs are not IEEE compliant, this can lead to a problem where correct implementations can cause failing tests. For example, #199 turns out to be because the filament library is correct, whereas the other implementations cut some corners in renormalization leading to rare problems.
Therefore, we can instead change testing to use something like [berkeley's hardfloat](https://github.com/ucb-bar/berkeley-hardfloat) instead, which also has the benefit of testing edge cases and doing better than just totally randomized testing, to check whether implementations are accurate. #174 adds edge testing capability, so we are a step closer to doing this. What's left is only to convert the output of `hardfloat` to a json file.
This would allow us to re-enable floating point testing with the filament library in the main branch as well. This is not a very high priority change, however, so can probably be deferred until later. | 1.0 | Testing floating point/iterative divider implementations for accuracy instead of conformity - Currently, the `floating-point` directory of implementations tests for conformity among a couple other implementations, most of which are directly derived from an implementation in verilog. The adder was taken externally from [here](https://github.com/suhasr1991/5-Stage-Pipelined-IEEE-Single-Precision-Floating-Point-Adder-Design).
The problem with this is that if the initial verilog designs are not IEEE compliant, this can lead to a problem where correct implementations can cause failing tests. For example, #199 turns out to be because the filament library is correct, whereas the other implementations cut some corners in renormalization leading to rare problems.
Therefore, we can instead change testing to use something like [berkeley's hardfloat](https://github.com/ucb-bar/berkeley-hardfloat) instead, which also has the benefit of testing edge cases and doing better than just totally randomized testing, to check whether implementations are accurate. #174 adds edge testing capability, so we are a step closer to doing this. What's left is only to convert the output of `hardfloat` to a json file.
This would allow us to re-enable floating point testing with the filament library in the main branch as well. This is not a very high priority change, however, so can probably be deferred until later. | priority | testing floating point iterative divider implementations for accuracy instead of conformity currently the floating point directory of implementations tests for conformity among a couple other implementations most of which are directly derived from an implementation in verilog the adder was taken externally from the problem with this is that if the initial verilog designs are not ieee compliant this can lead to a problem where correct implementations can cause failing tests for example turns out to be because the filament library is correct whereas the other implementations cut some corners in renormalization leading to rare problems therefore we can instead change testing to use something like instead which also has the benefit of testing edge cases and doing better than just totally randomized testing to check whether implementations are accurate adds edge testing capability so we are a step closer to doing this what s left is only to convert the output of hardfloat to a json file this would allow us to re enable floating point testing with the filament library in the main branch as well this is not a very high priority change however so can probably be deferred until later | 1 |
405,392 | 11,872,516,730 | IssuesEvent | 2020-03-26 15:55:48 | ComPWA/pycompwa | https://api.github.com/repos/ComPWA/pycompwa | closed | Improve urls of subpages of the pycompwa website | Priority: Low Status: In Progress Type: Enhancement Type: Maintenance | Currently, some urls of our website, contain an underscore, like so:
* https://compwa.github.io/_api/pycompwa.html
* https://compwa.github.io/_examples/Quickstart.html
This is only case for pages that have been included through a [`toctree`](https://www.sphinx-doc.org/en/master/usage/restructuredtext/directives.html#directive-toctree).
The underscores were added to render the pages correctly through the [GitHub Pages repo](https://github.com/ComPWA/ComPWA.github.io). It could be, however, that this is no longer necessary, [now that we make use of the `githubpages` Sphinx extension](https://github.com/ComPWA/pycompwa/commit/2a4a58ca69753bfde99f3409ab23e971b1252fa0#diff-555c472a89566734e0d8567fb205d5dfR70). | 1.0 | Improve urls of subpages of the pycompwa website - Currently, some urls of our website, contain an underscore, like so:
* https://compwa.github.io/_api/pycompwa.html
* https://compwa.github.io/_examples/Quickstart.html
This is only case for pages that have been included through a [`toctree`](https://www.sphinx-doc.org/en/master/usage/restructuredtext/directives.html#directive-toctree).
The underscores were added to render the pages correctly through the [GitHub Pages repo](https://github.com/ComPWA/ComPWA.github.io). It could be, however, that this is no longer necessary, [now that we make use of the `githubpages` Sphinx extension](https://github.com/ComPWA/pycompwa/commit/2a4a58ca69753bfde99f3409ab23e971b1252fa0#diff-555c472a89566734e0d8567fb205d5dfR70). | priority | improve urls of subpages of the pycompwa website currently some urls of our website contain an underscore like so this is only case for pages that have been included through a the underscores were added to render the pages correctly through the it could be however that this is no longer necessary | 1 |
544,578 | 15,894,731,289 | IssuesEvent | 2021-04-11 11:21:47 | marcusolsson/grafana-hourly-heatmap-panel | https://api.github.com/repos/marcusolsson/grafana-hourly-heatmap-panel | closed | Color option for null values | priority/low type/enhancement | If there is a null value in the data, the heatmap prints black color. It would be nice to have a color palette to select a color for null values. | 1.0 | Color option for null values - If there is a null value in the data, the heatmap prints black color. It would be nice to have a color palette to select a color for null values. | priority | color option for null values if there is a null value in the data the heatmap prints black color it would be nice to have a color palette to select a color for null values | 1 |
92,501 | 3,871,643,981 | IssuesEvent | 2016-04-11 10:40:52 | osm2vectortiles/osm2vectortiles | https://api.github.com/repos/osm2vectortiles/osm2vectortiles | opened | Implement network and shield field | low-priority | - [ ] Implement field shield of layer road_label
- [ ] Implement field network of layer rail_station_label
- The OSM key [network](http://wiki.openstreetmap.org/wiki/Key:network) can be used to determine in which country/region (network) the road or railway is. | 1.0 | Implement network and shield field - - [ ] Implement field shield of layer road_label
- [ ] Implement field network of layer rail_station_label
- The OSM key [network](http://wiki.openstreetmap.org/wiki/Key:network) can be used to determine in which country/region (network) the road or railway is. | priority | implement network and shield field implement field shield of layer road label implement field network of layer rail station label the osm key can be used to determine in which country region network the road or railway is | 1 |
438,923 | 12,663,474,754 | IssuesEvent | 2020-06-18 01:29:11 | vmware/clarity | https://api.github.com/repos/vmware/clarity | closed | Datagrid memory leak and DOM elements leak. | component: datagrid flag: has workaround priority: 1 low status: needs investigation type: bug | ```
[x] bug
[ ] feature request
[ ] enhancement
```
### Expected behavior
Datagrid not to leak memory and DOM elements.
### Actual behavior
After constantly updating the table using a socket, memory leaks and element leaks are created.
### Reproduction of behavior
- https://stackblitz.com/edit/angular-4-5-clarity-datagrid-memoy-leak
- Turn on the performance tab in chrome for one minute and you'll be surprised by the error
- https://monosnap.com/file/nz5FJKsxRrx7f5SVdH5xr2UaY6PCD8
### Environment details
* **Angular version:** 4.4.5
* **Clarity version:** 0.10.23
* **OS and version:**
* **Browser:** [ Chrome 63.0.3239.132 ]
| 1.0 | Datagrid memory leak and DOM elements leak. - ```
[x] bug
[ ] feature request
[ ] enhancement
```
### Expected behavior
Datagrid not to leak memory and DOM elements.
### Actual behavior
After constantly updating the table using a socket, memory leaks and element leaks are created.
### Reproduction of behavior
- https://stackblitz.com/edit/angular-4-5-clarity-datagrid-memoy-leak
- Turn on the performance tab in chrome for one minute and you'll be surprised by the error
- https://monosnap.com/file/nz5FJKsxRrx7f5SVdH5xr2UaY6PCD8
### Environment details
* **Angular version:** 4.4.5
* **Clarity version:** 0.10.23
* **OS and version:**
* **Browser:** [ Chrome 63.0.3239.132 ]
| priority | datagrid memory leak and dom elements leak bug feature request enhancement expected behavior datagrid not to leak memory and dom elements actual behavior after constantly updating the table using a socket memory leaks and element leaks are created reproduction of behavior turn on the performance tab in chrome for one minute and you ll be surprised by the error environment details angular version clarity version os and version browser | 1 |
438,120 | 12,619,567,073 | IssuesEvent | 2020-06-13 01:16:00 | skylight-hq/skylight.digital | https://api.github.com/repos/skylight-hq/skylight.digital | closed | Auto generate pages for blog post authors, blog post tags, and project team members | priority:low | Right now each page has to be set up manually. | 1.0 | Auto generate pages for blog post authors, blog post tags, and project team members - Right now each page has to be set up manually. | priority | auto generate pages for blog post authors blog post tags and project team members right now each page has to be set up manually | 1 |
256,216 | 8,127,038,622 | IssuesEvent | 2018-08-17 06:17:52 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Append version number onto build_visit and visit-install that we put on the Web. | Expected Use: 3 - Occasional Feature Impact: 2 - Low Priority: Normal Support Group: DOE/ASC | cq-id: VisIt00008818
cq-submitter: Brad Whitlock
cq-submit-date: 12/01/08
A couple of external users have become confused with build_visit and visit-install and have ended up using them with wrong versions of the binary distributions and source code. The users both suggested that we append the version number to the scripts that we make available for download on the Web.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 126
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Append version number onto build_visit and visit-install that we put on the Web.
Assigned to: Eric Brugger
Category:
Target version: 2.0.2
Author: Cyrus Harrison
Start:
Due date:
% Done: 100
Estimated time:
Created: 06/21/2010 07:16 pm
Updated: 07/20/2010 02:22 pm
Likelihood:
Severity:
Found in version:
Impact: 2 - Low
Expected Use: 3 - Occasional
OS: All
Support Group: DOE/ASC
Description:
cq-id: VisIt00008818
cq-submitter: Brad Whitlock
cq-submit-date: 12/01/08
A couple of external users have become confused with build_visit and visit-install and have ended up using them with wrong versions of the binary distributions and source code. The users both suggested that we append the version number to the scripts that we make available for download on the Web.
Comments:
I added the version number to the build_visit and visit-install scripts when the scripts are added to the web site.
| 1.0 | Append version number onto build_visit and visit-install that we put on the Web. - cq-id: VisIt00008818
cq-submitter: Brad Whitlock
cq-submit-date: 12/01/08
A couple of external users have become confused with build_visit and visit-install and have ended up using them with wrong versions of the binary distributions and source code. The users both suggested that we append the version number to the scripts that we make available for download on the Web.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 126
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Append version number onto build_visit and visit-install that we put on the Web.
Assigned to: Eric Brugger
Category:
Target version: 2.0.2
Author: Cyrus Harrison
Start:
Due date:
% Done: 100
Estimated time:
Created: 06/21/2010 07:16 pm
Updated: 07/20/2010 02:22 pm
Likelihood:
Severity:
Found in version:
Impact: 2 - Low
Expected Use: 3 - Occasional
OS: All
Support Group: DOE/ASC
Description:
cq-id: VisIt00008818
cq-submitter: Brad Whitlock
cq-submit-date: 12/01/08
A couple of external users have become confused with build_visit and visit-install and have ended up using them with wrong versions of the binary distributions and source code. The users both suggested that we append the version number to the scripts that we make available for download on the Web.
Comments:
I added the version number to the build_visit and visit-install scripts when the scripts are added to the web site.
| priority | append version number onto build visit and visit install that we put on the web cq id cq submitter brad whitlock cq submit date a couple of external users have become confused with build visit and visit install and have ended up using them with wrong versions of the binary distributions and source code the users both suggested that we append the version number to the scripts that we make available for download on the web redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject append version number onto build visit and visit install that we put on the web assigned to eric brugger category target version author cyrus harrison start due date done estimated time created pm updated pm likelihood severity found in version impact low expected use occasional os all support group doe asc description cq id cq submitter brad whitlock cq submit date a couple of external users have become confused with build visit and visit install and have ended up using them with wrong versions of the binary distributions and source code the users both suggested that we append the version number to the scripts that we make available for download on the web comments i added the version number to the build visit and visit install scripts when the scripts are added to the web site | 1 |
671,447 | 22,761,544,955 | IssuesEvent | 2022-07-07 21:47:20 | markmac99/UKmon-shared | https://api.github.com/repos/markmac99/UKmon-shared | closed | Add container support for cartopy | enhancement Low priority | This would allow OS maps to be used as the backdrop for the ground-track maps.
However cartopy is poorly supported and has to be built from source, along with GEOS, PROJ4 and SQLITE3, which makes it quite a task. | 1.0 | Add container support for cartopy - This would allow OS maps to be used as the backdrop for the ground-track maps.
However cartopy is poorly supported and has to be built from source, along with GEOS, PROJ4 and SQLITE3, which makes it quite a task. | priority | add container support for cartopy this would allow os maps to be used as the backdrop for the ground track maps however cartopy is poorly supported and has to be built from source along with geos and which makes it quite a task | 1 |
29,254 | 2,714,206,549 | IssuesEvent | 2015-04-10 00:43:08 | hamiltont/clasp | https://api.github.com/repos/hamiltont/clasp | opened | Option for custom Android system, boot, and kernel images. | Low priority | _From @bamos on September 24, 2014 15:33_
Specifically to run nbd -- http://bamos.github.io/2014/09/08/nbd-android/
_Copied from original issue: hamiltont/attack#53_ | 1.0 | Option for custom Android system, boot, and kernel images. - _From @bamos on September 24, 2014 15:33_
Specifically to run nbd -- http://bamos.github.io/2014/09/08/nbd-android/
_Copied from original issue: hamiltont/attack#53_ | priority | option for custom android system boot and kernel images from bamos on september specifically to run nbd copied from original issue hamiltont attack | 1 |
228,531 | 7,552,579,968 | IssuesEvent | 2018-04-19 01:09:32 | OperationCode/operationcode_backend | https://api.github.com/repos/OperationCode/operationcode_backend | closed | Rake task to tag community leaders | Priority: Low Status: In Progress Type: Feature | <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
Now that https://github.com/OperationCode/operationcode_backend/pull/292 is merged, we need a rake task to initially tag the appropriate users in prod with the community leader tag.
## What should your feature do?
- [ ] Collaborate with @hollomancer (or whomever he deems appropriate) to determine what prod users should be tagged
- [ ] Creating the rake task to [tag community leaders](https://github.com/OperationCode/operationcode_backend/blob/master/app/models/user.rb#L2) | 1.0 | Rake task to tag community leaders - <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
Now that https://github.com/OperationCode/operationcode_backend/pull/292 is merged, we need a rake task to initially tag the appropriate users in prod with the community leader tag.
## What should your feature do?
- [ ] Collaborate with @hollomancer (or whomever he deems appropriate) to determine what prod users should be tagged
- [ ] Creating the rake task to [tag community leaders](https://github.com/OperationCode/operationcode_backend/blob/master/app/models/user.rb#L2) | priority | rake task to tag community leaders feature why is this feature being added now that is merged we need a rake task to initially tag the appropriate users in prod with the community leader tag what should your feature do collaborate with hollomancer or whomever he deems appropriate to determine what prod users should be tagged creating the rake task to | 1 |
35,022 | 2,789,753,415 | IssuesEvent | 2015-05-08 21:16:41 | google/google-visualization-api-issues | https://api.github.com/repos/google/google-visualization-api-issues | opened | Visualization to understand XML | Priority-Low Type-Enhancement | Original [issue 55](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=55) created by google-admin on 2009-09-16T14:22:48.000Z:
<b>What would you like to see us add to this API?</b>
It would have been nice if all visualization also understands XML.
In that case we have to only fetch the XML from the server and pass it to
Visualization. This will remove the coding need to fill the Visualization
data.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | Visualization to understand XML - Original [issue 55](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=55) created by google-admin on 2009-09-16T14:22:48.000Z:
<b>What would you like to see us add to this API?</b>
It would have been nice if all visualization also understands XML.
In that case we have to only fetch the XML from the server and pass it to
Visualization. This will remove the coding need to fill the Visualization
data.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| priority | visualization to understand xml original created by google admin on what would you like to see us add to this api it would have been nice if all visualization also understands xml in that case we have to only fetch the xml from the server and pass it to visualization this will remove the coding need to fill the visualization data what component is this issue related to piechart linechart datatable query etc for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 1 |
614,576 | 19,185,841,565 | IssuesEvent | 2021-12-05 06:59:01 | eshoku/frontend | https://api.github.com/repos/eshoku/frontend | opened | 主催者でないのにurl経由で編集ページにアクセスできてしまう | bug Priority: low | ## 概要
主催者でないのにurl経由で編集ページにアクセスできてしまうバグがある。
バックエンド側で権限を確認しているので実際にルーム情報を編集できるわけではないが、利用者が混乱するため、修正するべき。 | 1.0 | 主催者でないのにurl経由で編集ページにアクセスできてしまう - ## 概要
主催者でないのにurl経由で編集ページにアクセスできてしまうバグがある。
バックエンド側で権限を確認しているので実際にルーム情報を編集できるわけではないが、利用者が混乱するため、修正するべき。 | priority | 主催者でないのにurl経由で編集ページにアクセスできてしまう 概要 主催者でないのにurl経由で編集ページにアクセスできてしまうバグがある。 バックエンド側で権限を確認しているので実際にルーム情報を編集できるわけではないが、利用者が混乱するため、修正するべき。 | 1 |
121,182 | 4,805,925,332 | IssuesEvent | 2016-11-02 17:13:31 | FrozenSand/UrbanTerror4 | https://api.github.com/repos/FrozenSand/UrbanTerror4 | closed | Auto-completion is not working for several commands | confirmed low priority | Auto completion in 4.3 isnt't working anymore for all auth commands, and potentially others.
| 1.0 | Auto-completion is not working for several commands - Auto completion in 4.3 isnt't working anymore for all auth commands, and potentially others.
| priority | auto completion is not working for several commands auto completion in isnt t working anymore for all auth commands and potentially others | 1 |
545,698 | 15,955,375,664 | IssuesEvent | 2021-04-15 14:32:38 | Gird-the-Grid/Grid-the-Grid | https://api.github.com/repos/Gird-the-Grid/Grid-the-Grid | opened | Frontend: Register | enhancement low priority | * Show loading button while waiting for register response
* Display error message if `Success: false` (Email user / other error) | 1.0 | Frontend: Register - * Show loading button while waiting for register response
* Display error message if `Success: false` (Email user / other error) | priority | frontend register show loading button while waiting for register response display error message if success false email user other error | 1 |
35,101 | 2,789,798,526 | IssuesEvent | 2015-05-08 21:33:42 | google/google-visualization-api-issues | https://api.github.com/repos/google/google-visualization-api-issues | opened | Comparing of different time-periods | Priority-Low Type-Enhancement | Original [issue 140](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=140) created by orwant on 2009-11-30T21:45:51.000Z:
<b>What would you like to see us add to this API?</b>
A feature where you can compare values to different time-periods would be
very helpful.
E.G. How developed my sales in the last month to the same month in the
previous year ? or how was the quarter 3/09 to 3/08. Perhaps with green.
orange and red arrows, pointing up, side or down...
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
I Guess it would be a new component which appears like a data-Table.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | Comparing of different time-periods - Original [issue 140](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=140) created by orwant on 2009-11-30T21:45:51.000Z:
<b>What would you like to see us add to this API?</b>
A feature where you can compare values to different time-periods would be
very helpful.
E.G. How developed my sales in the last month to the same month in the
previous year ? or how was the quarter 3/09 to 3/08. Perhaps with green.
orange and red arrows, pointing up, side or down...
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
I Guess it would be a new component which appears like a data-Table.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| priority | comparing of different time periods original created by orwant on what would you like to see us add to this api a feature where you can compare values to different time periods would be very helpful e g how developed my sales in the last month to the same month in the previous year or how was the quarter to perhaps with green orange and red arrows pointing up side or down what component is this issue related to piechart linechart datatable query etc i guess it would be a new component which appears like a data table for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 1 |
353,970 | 10,561,380,709 | IssuesEvent | 2019-10-04 15:45:01 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | API projectFromPoint function is wrong | bug-type: unexpected behavior priority: low problem: bug type: corrective | API `XY.projectFromPoint()` has a few things off with it.
It's marking the result with the incoming projection instead of the outcome projection.
Also allows you to default to the XY object's co-ords, which makes no sense as they are already in latlong.
The function itself really doesn't need to be on the XY class as it's a stateless function, but will keep there to avoid breaking the interface. | 1.0 | API projectFromPoint function is wrong - API `XY.projectFromPoint()` has a few things off with it.
It's marking the result with the incoming projection instead of the outcome projection.
Also allows you to default to the XY object's co-ords, which makes no sense as they are already in latlong.
The function itself really doesn't need to be on the XY class as it's a stateless function, but will keep there to avoid breaking the interface. | priority | api projectfrompoint function is wrong api xy projectfrompoint has a few things off with it it s marking the result with the incoming projection instead of the outcome projection also allows you to default to the xy object s co ords which makes no sense as they are already in latlong the function itself really doesn t need to be on the xy class as it s a stateless function but will keep there to avoid breaking the interface | 1 |
414,706 | 12,110,534,493 | IssuesEvent | 2020-04-21 10:35:06 | Icyr/DnDApp | https://api.github.com/repos/Icyr/DnDApp | closed | Add loading indicators. | change request priority: low | There is no delay on add (thanks to Firestore).
EDIT: Room has a delay on create, we need to add loading for it too.
But there is a delay on initial load.
We need to display loading indicator when list is updated. | 1.0 | Add loading indicators. - There is no delay on add (thanks to Firestore).
EDIT: Room has a delay on create, we need to add loading for it too.
But there is a delay on initial load.
We need to display loading indicator when list is updated. | priority | add loading indicators there is no delay on add thanks to firestore edit room has a delay on create we need to add loading for it too but there is a delay on initial load we need to display loading indicator when list is updated | 1 |
467,203 | 13,443,404,207 | IssuesEvent | 2020-09-08 08:18:47 | input-output-hk/cardano-wallet | https://api.github.com/repos/input-output-hk/cardano-wallet | closed | Listing transaction when node is still in the Byron era may fail with 500 | BUG PRIORITY:LOW SEVERITY:HIGH | # Context
<!-- WHEN CREATED
Any information that is useful to understand the bug and the subsystem
it evolves in. References to documentation and or other tickets are
welcome.
-->
| Information | - |
| --- | --- |
| Version | Close to Shelley Mainnet release |
| Platform | <!-- Windows, Mac OS, Linux, Docker, All --> |
| Installation | <!-- From Source? From Github Release? --> |
With #1869, we use a `TimeInterpreter` from the node for time and slotting conversions. The conversions fails if "Past the Horizon". From genesis to close to the fork, the node will have a limited horizon.
# Steps to Reproduce
<!-- WHEN CREATED
Steps to reproduce the behavior.
-->
1. Have a node and wallet which is running Byron;Shelley, but which has not yet reached Shelley. (E.g. on Cardano mainnet when first syncing from genesis)
1. List transactions with a time range were one of the dates are far ahead of the node tip (which is continuously syncing).
## Expected behavior
<!-- WHEN CREATED
A clear and concise description of what you expected to happen.
-->
1. I get a response, or a helpful error
## Actual behavior
1. (Theorized) `500: Something went wrong`
*Note*: This shouldn't happen when the node is in Shelley.
<!-- WHEN CREATED
A clear and concise description of what you observe instead. If applicable add
screenshots to help explain your problem.
-->
---
# Resolution
<!-- WHEN IN PROGRESS
What is happening? How is this going to be fixed? Detail the approach and give,
in the form of a TODO list steps toward the resolution of the bug. Attach a PR to
each item in the list.
This may be refined as the investigation progresses.
-->
---
# QA
<!-- WHEN IN PROGRESS
How do we make sure the bug has been fixed? Give here manual steps or tests to
verify the fix. How/why could this bug slip through testing?
-->
| 1.0 | Listing transaction when node is still in the Byron era may fail with 500 - # Context
<!-- WHEN CREATED
Any information that is useful to understand the bug and the subsystem
it evolves in. References to documentation and or other tickets are
welcome.
-->
| Information | - |
| --- | --- |
| Version | Close to Shelley Mainnet release |
| Platform | <!-- Windows, Mac OS, Linux, Docker, All --> |
| Installation | <!-- From Source? From Github Release? --> |
With #1869, we use a `TimeInterpreter` from the node for time and slotting conversions. The conversions fails if "Past the Horizon". From genesis to close to the fork, the node will have a limited horizon.
# Steps to Reproduce
<!-- WHEN CREATED
Steps to reproduce the behavior.
-->
1. Have a node and wallet which is running Byron;Shelley, but which has not yet reached Shelley. (E.g. on Cardano mainnet when first syncing from genesis)
1. List transactions with a time range were one of the dates are far ahead of the node tip (which is continuously syncing).
## Expected behavior
<!-- WHEN CREATED
A clear and concise description of what you expected to happen.
-->
1. I get a response, or a helpful error
## Actual behavior
1. (Theorized) `500: Something went wrong`
*Note*: This shouldn't happen when the node is in Shelley.
<!-- WHEN CREATED
A clear and concise description of what you observe instead. If applicable add
screenshots to help explain your problem.
-->
---
# Resolution
<!-- WHEN IN PROGRESS
What is happening? How is this going to be fixed? Detail the approach and give,
in the form of a TODO list steps toward the resolution of the bug. Attach a PR to
each item in the list.
This may be refined as the investigation progresses.
-->
---
# QA
<!-- WHEN IN PROGRESS
How do we make sure the bug has been fixed? Give here manual steps or tests to
verify the fix. How/why could this bug slip through testing?
-->
| priority | listing transaction when node is still in the byron era may fail with context when created any information that is useful to understand the bug and the subsystem it evolves in references to documentation and or other tickets are welcome information version close to shelley mainnet release platform installation with we use a timeinterpreter from the node for time and slotting conversions the conversions fails if past the horizon from genesis to close to the fork the node will have a limited horizon steps to reproduce when created steps to reproduce the behavior have a node and wallet which is running byron shelley but which has not yet reached shelley e g on cardano mainnet when first syncing from genesis list transactions with a time range were one of the dates are far ahead of the node tip which is continuously syncing expected behavior when created a clear and concise description of what you expected to happen i get a response or a helpful error actual behavior theorized something went wrong note this shouldn t happen when the node is in shelley when created a clear and concise description of what you observe instead if applicable add screenshots to help explain your problem resolution when in progress what is happening how is this going to be fixed detail the approach and give in the form of a todo list steps toward the resolution of the bug attach a pr to each item in the list this may be refined as the investigation progresses qa when in progress how do we make sure the bug has been fixed give here manual steps or tests to verify the fix how why could this bug slip through testing | 1 |
646,526 | 21,051,694,848 | IssuesEvent | 2022-03-31 21:08:36 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | Vending Table store wrong coordinate if merchant warped during vending state. | status:confirmed component:core priority:low mode:renewal mode:prerenewal type:bug | <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
https://github.com/rathena/rathena/commit/73a8d1365e84252e37185f8c579a8825abd925c4
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**:
2018-06-20
<!-- Please specify the client date you used. -->
* **Server Mode**:
RE
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->

* Expected Result: <!-- Describe what you would expect to happen in detail. -->
- If vending shop is active, shouldn't allow warp.
- or database should reflect latest warp location (or at least re-update upon used `@autotrade`?)
- or remove vending shop upon warped.
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
- Create a merchant with vending skill, add an item to vending shop (but I was testing using a lv99 GM merchant).
- Open the shop (locationed stored to db i guess), then type `@warp` or `@jump` with vending shop still active.
- After warped, vending shop still remain active, but actual location changed from `(256, 143)` to ` (104,316)`, but db remained as `(256, 143)`.
- Type `@autotrade`
- Login another user to check the location, using `@whomap3` it actually reflect the correct location which is ` (104,316)` and search for the vendor, but db storing the old location `(256, 143)` (before merchant used `@warp`).
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
None | 1.0 | Vending Table store wrong coordinate if merchant warped during vending state. - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
https://github.com/rathena/rathena/commit/73a8d1365e84252e37185f8c579a8825abd925c4
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**:
2018-06-20
<!-- Please specify the client date you used. -->
* **Server Mode**:
RE
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->

* Expected Result: <!-- Describe what you would expect to happen in detail. -->
- If vending shop is active, shouldn't allow warp.
- or database should reflect latest warp location (or at least re-update upon used `@autotrade`?)
- or remove vending shop upon warped.
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
- Create a merchant with vending skill, add an item to vending shop (but I was testing using a lv99 GM merchant).
- Open the shop (locationed stored to db i guess), then type `@warp` or `@jump` with vending shop still active.
- After warped, vending shop still remain active, but actual location changed from `(256, 143)` to ` (104,316)`, but db remained as `(256, 143)`.
- Type `@autotrade`
- Login another user to check the location, using `@whomap3` it actually reflect the correct location which is ` (104,316)` and search for the vendor, but db storing the old location `(256, 143)` (before merchant used `@warp`).
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
None | priority | vending table store wrong coordinate if merchant warped during vending state rathena hash please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode re description of issue result expected result if vending shop is active shouldn t allow warp or database should reflect latest warp location or at least re update upon used autotrade or remove vending shop upon warped how to reproduce create a merchant with vending skill add an item to vending shop but i was testing using a gm merchant open the shop locationed stored to db i guess then type warp or jump with vending shop still active after warped vending shop still remain active but actual location changed from to but db remained as type autotrade login another user to check the location using it actually reflect the correct location which is and search for the vendor but db storing the old location before merchant used warp official information modifications that may affect results none | 1 |
674,664 | 23,061,359,382 | IssuesEvent | 2022-07-25 10:13:29 | SeldonIO/alibi-detect | https://api.github.com/repos/SeldonIO/alibi-detect | opened | Review content of metadata in detector config files | Priority: Low Type: Engineering | Our detector `self.meta` attribute currently contains the following information:
```python
DEFAULT_META = {
"name": None,
"online": None, # true or false
"data_type": None, # tabular, image or time-series
"version": None,
"detector_type": None # drift, outlier or adversarial
} # type: Dict
```
Whereas the metadata in detector config files contains:
```toml
[meta]
version = "0.10.0rc1"
config_spec = "0.1"
version_warning = false
```
We should review what content it makes sense to include in detector config files. Fields such as `detector_type` and `online` could be useful to view without instantiating a detector. In https://github.com/SeldonIO/alibi-detect/pull/564 it was decided not to include the entirety of `self.meta` in config files since this leads to excessive duplication (e.g. `name`, `backend` and `data_type`). | 1.0 | Review content of metadata in detector config files - Our detector `self.meta` attribute currently contains the following information:
```python
DEFAULT_META = {
"name": None,
"online": None, # true or false
"data_type": None, # tabular, image or time-series
"version": None,
"detector_type": None # drift, outlier or adversarial
} # type: Dict
```
Whereas the metadata in detector config files contains:
```toml
[meta]
version = "0.10.0rc1"
config_spec = "0.1"
version_warning = false
```
We should review what content it makes sense to include in detector config files. Fields such as `detector_type` and `online` could be useful to view without instantiating a detector. In https://github.com/SeldonIO/alibi-detect/pull/564 it was decided not to include the entirety of `self.meta` in config files since this leads to excessive duplication (e.g. `name`, `backend` and `data_type`). | priority | review content of metadata in detector config files our detector self meta attribute currently contains the following information python default meta name none online none true or false data type none tabular image or time series version none detector type none drift outlier or adversarial type dict whereas the metadata in detector config files contains toml version config spec version warning false we should review what content it makes sense to include in detector config files fields such as detector type and online could be useful to view without instantiating a detector in it was decided not to include the entirety of self meta in config files since this leads to excessive duplication e g name backend and data type | 1 |
226,261 | 7,516,866,513 | IssuesEvent | 2018-04-12 00:12:56 | atlassian/react-beautiful-dnd | https://api.github.com/repos/atlassian/react-beautiful-dnd | closed | Add a real browser test | engineering health good first issue idea priority: low | I am not sure if we need this - but I thought I would raise it anyway
Generally I am quite against browser tests. However, I think it could be worthwhile having one smoke test that runs in a real browser. I am fairly keen to use [puppeteer](https://github.com/GoogleChrome/puppeteer) for this one. It would not need to run in all browsers. This is purely just a small smoke test.
Criteria
- Would need to be able to run on travis (I think if we run chrome in headless mode this should be fine)
- Needs to be crazy simple with minimal config - I do not want to spend heaps of time maintaining this. | 1.0 | Add a real browser test - I am not sure if we need this - but I thought I would raise it anyway
Generally I am quite against browser tests. However, I think it could be worthwhile having one smoke test that runs in a real browser. I am fairly keen to use [puppeteer](https://github.com/GoogleChrome/puppeteer) for this one. It would not need to run in all browsers. This is purely just a small smoke test.
Criteria
- Would need to be able to run on travis (I think if we run chrome in headless mode this should be fine)
- Needs to be crazy simple with minimal config - I do not want to spend heaps of time maintaining this. | priority | add a real browser test i am not sure if we need this but i thought i would raise it anyway generally i am quite against browser tests however i think it could be worthwhile having one smoke test that runs in a real browser i am fairly keen to use for this one it would not need to run in all browsers this is purely just a small smoke test criteria would need to be able to run on travis i think if we run chrome in headless mode this should be fine needs to be crazy simple with minimal config i do not want to spend heaps of time maintaining this | 1 |
416,897 | 12,152,421,797 | IssuesEvent | 2020-04-24 22:13:32 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Support authorisation time based ( business-hours / business-days / etc.) | Priority: Low customer request enhancement wontfix | **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
No option to have time based access control to apis.
Use case - backend services are responding only during biz hours and are shutdown or going through maintenance outside this slot.
**What is the expected behavior?**
We can have a definition on the api level - for instance, biz hours of the api. Outside the permitted time, a predefined mock response (with configured error code, payload and headers) will be returned.
**Going forward:**
Add it on the policy level as well, in a similar way to rate limit - we have it on the key level, i.e. per user but also on the api level to protect the backend service.
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.7 | 1.0 | Support authorisation time based ( business-hours / business-days / etc.) - **Do you want to request a *feature* or report a *bug*?**
Feature
**What is the current behavior?**
No option to have time based access control to apis.
Use case - backend services are responding only during biz hours and are shutdown or going through maintenance outside this slot.
**What is the expected behavior?**
We can have a definition on the api level - for instance, biz hours of the api. Outside the permitted time, a predefined mock response (with configured error code, payload and headers) will be returned.
**Going forward:**
Add it on the policy level as well, in a similar way to rate limit - we have it on the key level, i.e. per user but also on the api level to protect the backend service.
**If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem**
**Which versions of Tyk affected by this issue? Did this work in previous versions of Tyk?**
2.7 | priority | support authorisation time based business hours business days etc do you want to request a feature or report a bug feature what is the current behavior no option to have time based access control to apis use case backend services are responding only during biz hours and are shutdown or going through maintenance outside this slot what is the expected behavior we can have a definition on the api level for instance biz hours of the api outside the permitted time a predefined mock response with configured error code payload and headers will be returned going forward add it on the policy level as well in a similar way to rate limit we have it on the key level i e per user but also on the api level to protect the backend service if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem which versions of tyk affected by this issue did this work in previous versions of tyk | 1 |
450,597 | 13,016,984,889 | IssuesEvent | 2020-07-26 09:53:17 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | duplicated workspace role alert msg should be displayed before saving role | area/console kind/bug kind/need-to-verify priority/low | ## English only!
**注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。**
**General remarks**
> Please delete this section including header before submitting
>
> This form is to report bugs. For general usage questions refer to our Slack channel
> [KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTdkNTc3OTdmNzdiODViZjViNTU5ZDY3M2I2MzY4MTI4OGZlOTJmMDg5ZTFiMDAwYzNlZDY5NjA0NzZlNDU5NmY)
**Describe the Bug**
duplicated workspace role alert msg should be displayed before saving role
**Versions Used**
KubeSphere: 3.0 alpha 1
**How To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Workspace settings' - 'Workspace roles'
2. Create a new role which identifier is duplicated with existing workspace role
3. It prompts 'duplicated' when saving this role at the last step

**Expected behavior**
It should prompt when finishing the input of the identifier name.
| 1.0 | duplicated workspace role alert msg should be displayed before saving role - ## English only!
**注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。**
**General remarks**
> Please delete this section including header before submitting
>
> This form is to report bugs. For general usage questions refer to our Slack channel
> [KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTdkNTc3OTdmNzdiODViZjViNTU5ZDY3M2I2MzY4MTI4OGZlOTJmMDg5ZTFiMDAwYzNlZDY5NjA0NzZlNDU5NmY)
**Describe the Bug**
duplicated workspace role alert msg should be displayed before saving role
**Versions Used**
KubeSphere: 3.0 alpha 1
**How To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Workspace settings' - 'Workspace roles'
2. Create a new role which identifier is duplicated with existing workspace role
3. It prompts 'duplicated' when saving this role at the last step

**Expected behavior**
It should prompt when finishing the input of the identifier name.
| priority | duplicated workspace role alert msg should be displayed before saving role english only 注意!github issue 仅支持英文,中文 issue 请在 提交。 general remarks please delete this section including header before submitting this form is to report bugs for general usage questions refer to our slack channel describe the bug duplicated workspace role alert msg should be displayed before saving role versions used kubesphere alpha how to reproduce steps to reproduce the behavior go to workspace settings workspace roles create a new role which identifier is duplicated with existing workspace role it prompts duplicated when saving this role at the last step expected behavior it should prompt when finishing the input of the identifier name | 1 |
31,452 | 2,732,910,563 | IssuesEvent | 2015-04-17 10:10:13 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | Layout Management | auto-migrated Priority-Low Schedule-LongTerm Type-Enhancement | ```
The current layouting possibilities are very rudimentary. The goal is to
offer process model layout manager. Because the layouts of different
stencil sets differ, layouts must be stencil set specific.
```
Original issue reported on code.google.com by `NicoPete...@gmail.com` on 17 Apr 2008 at 1:44 | 1.0 | Layout Management - ```
The current layouting possibilities are very rudimentary. The goal is to
offer process model layout manager. Because the layouts of different
stencil sets differ, layouts must be stencil set specific.
```
Original issue reported on code.google.com by `NicoPete...@gmail.com` on 17 Apr 2008 at 1:44 | priority | layout management the current layouting possibilities are very rudimentary the goal is to offer process model layout manager because the layouts of different stencil sets differ layouts must be stencil set specific original issue reported on code google com by nicopete gmail com on apr at | 1 |
491,192 | 14,146,644,219 | IssuesEvent | 2020-11-10 19:32:17 | powercomm/PCM-Dashboard | https://api.github.com/repos/powercomm/PCM-Dashboard | opened | When a new PCM is added, history chart is blank, user experience needs improvement | Deficiency Priority: Low | Type 30 Edit form needs to fill history chart when no history points are returned from PCM_AC, a straight line will make it easier to see the threshold alignments. | 1.0 | When a new PCM is added, history chart is blank, user experience needs improvement - Type 30 Edit form needs to fill history chart when no history points are returned from PCM_AC, a straight line will make it easier to see the threshold alignments. | priority | when a new pcm is added history chart is blank user experience needs improvement type edit form needs to fill history chart when no history points are returned from pcm ac a straight line will make it easier to see the threshold alignments | 1 |
493,989 | 14,243,036,089 | IssuesEvent | 2020-11-19 03:20:16 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Update images that have never been launched as a container | Priority: Low Status: Available Status: Stale Type: Enhancement | I have a use case where I want images updated on a machine but I haven't necessarily ran the image yet as a container. In this case, `--include-stopped` takes no effect (a container never ran), and there are no options to look at images only.
Expanded problem:
I have a CI system. The worker instances, when created, pull images that we've approved to run. When a new worker starts up it uses the local version of the image (never reaches out to a repository for images). Our system will periodically recycle the worker instances and may load up older versions of images on them.
We'd like something in place that looks at the images, determines they are older, then updates them. This way when the first job comes into a machine it'll have the latest version of that image instead of the one we had specified when the image was built. Only certain images should be able to pull the "latest" while others are ignored and pinned to the version we specified.
**Describe the solution you'd like**
I would like a flag that allows watchtower to look at images, not containers, on a machine and determine if a new version needs to be pulled.
**Describe alternatives you've considered**
Given our scenario we could run all the images on startup then shut them down. The `--include-stopped` flag would likely work for us then.
**Additional context**
I'm willing to do the work. I was already looking at where the `--include-stopped` flag is implemented and am proficient with go. The biggest hurdle is that this relies on images and not containers. I want to know if this is something the community would want before I go putting in the effort to open a PR with this functionality. If it's not valuable, or is too specific to my use case, I will consider other workarounds or roll my own image update process. I'd like to build something more useful though :).
| 1.0 | Update images that have never been launched as a container - I have a use case where I want images updated on a machine but I haven't necessarily ran the image yet as a container. In this case, `--include-stopped` takes no effect (a container never ran), and there are no options to look at images only.
Expanded problem:
I have a CI system. The worker instances, when created, pull images that we've approved to run. When a new worker starts up it uses the local version of the image (never reaches out to a repository for images). Our system will periodically recycle the worker instances and may load up older versions of images on them.
We'd like something in place that looks at the images, determines they are older, then updates them. This way when the first job comes into a machine it'll have the latest version of that image instead of the one we had specified when the image was built. Only certain images should be able to pull the "latest" while others are ignored and pinned to the version we specified.
**Describe the solution you'd like**
I would like a flag that allows watchtower to look at images, not containers, on a machine and determine if a new version needs to be pulled.
**Describe alternatives you've considered**
Given our scenario we could run all the images on startup then shut them down. The `--include-stopped` flag would likely work for us then.
**Additional context**
I'm willing to do the work. I was already looking at where the `--include-stopped` flag is implemented and am proficient with go. The biggest hurdle is that this relies on images and not containers. I want to know if this is something the community would want before I go putting in the effort to open a PR with this functionality. If it's not valuable, or is too specific to my use case, I will consider other workarounds or roll my own image update process. I'd like to build something more useful though :).
| priority | update images that have never been launched as a container i have a use case where i want images updated on a machine but i haven t necessarily ran the image yet as a container in this case include stopped takes no effect a container never ran and there are no options to look at images only expanded problem i have a ci system the worker instances when created pull images that we ve approved to run when a new worker starts up it uses the local version of the image never reaches out to a repository for images our system will periodically recycle the worker instances and may load up older versions of images on them we d like something in place that looks at the images determines they are older then updates them this way when the first job comes into a machine it ll have the latest version of that image instead of the one we had specified when the image was built only certain images should be able to pull the latest while others are ignored and pinned to the version we specified describe the solution you d like i would like a flag that allows watchtower to look at images not containers on a machine and determine if a new version needs to be pulled describe alternatives you ve considered given our scenario we could run all the images on startup then shut them down the include stopped flag would likely work for us then additional context i m willing to do the work i was already looking at where the include stopped flag is implemented and am proficient with go the biggest hurdle is that this relies on images and not containers i want to know if this is something the community would want before i go putting in the effort to open a pr with this functionality if it s not valuable or is too specific to my use case i will consider other workarounds or roll my own image update process i d like to build something more useful though | 1 |
664,399 | 22,268,774,705 | IssuesEvent | 2022-06-10 10:05:27 | mozilla/perfcompare | https://api.github.com/repos/mozilla/perfcompare | closed | export store creating function for tests | priority: low | from Julien:
> I believe the next step would be to export a store creating function instead of the store object itself, and create an empty store for each test, so that tests are always fully isolated. Happy to discuss about it more later! I was thinking that as a first step, if you don't want to change existing tests, you could export both `store` and a `createStore()` function, so that new tests would use the creator, but old tests would still work with the object directly. That's up to you really :-) | 1.0 | export store creating function for tests - from Julien:
> I believe the next step would be to export a store creating function instead of the store object itself, and create an empty store for each test, so that tests are always fully isolated. Happy to discuss about it more later! I was thinking that as a first step, if you don't want to change existing tests, you could export both `store` and a `createStore()` function, so that new tests would use the creator, but old tests would still work with the object directly. That's up to you really :-) | priority | export store creating function for tests from julien i believe the next step would be to export a store creating function instead of the store object itself and create an empty store for each test so that tests are always fully isolated happy to discuss about it more later i was thinking that as a first step if you don t want to change existing tests you could export both store and a createstore function so that new tests would use the creator but old tests would still work with the object directly that s up to you really | 1 |
41,341 | 2,868,999,142 | IssuesEvent | 2015-06-05 22:28:46 | dart-lang/args | https://api.github.com/repos/dart-lang/args | opened | zsh/bash autocompletion generator from ArgParser | enhancement Priority-Low | <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8389_
----
zsh and bash both have a shell autocomplete capability, so that you can quickly complete flags, arguments, etc.
It'd be amazing if we could ingest an ArgParser and spit out zsh/bash completion files so that any dart script can autocomplete in this way. | 1.0 | zsh/bash autocompletion generator from ArgParser - <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8389_
----
zsh and bash both have a shell autocomplete capability, so that you can quickly complete flags, arguments, etc.
It'd be amazing if we could ingest an ArgParser and spit out zsh/bash completion files so that any dart script can autocomplete in this way. | priority | zsh bash autocompletion generator from argparser issue by originally opened as dart lang sdk zsh and bash both have a shell autocomplete capability so that you can quickly complete flags arguments etc it d be amazing if we could ingest an argparser and spit out zsh bash completion files so that any dart script can autocomplete in this way | 1 |
306,404 | 9,392,654,150 | IssuesEvent | 2019-04-07 03:18:03 | generative-music/generative.fm | https://api.github.com/repos/generative-music/generative.fm | closed | Update indicator animation | enhancement high priority low effort | Animate the update indicator above the "About" tab label so users click on it. | 1.0 | Update indicator animation - Animate the update indicator above the "About" tab label so users click on it. | priority | update indicator animation animate the update indicator above the about tab label so users click on it | 1 |
827,297 | 31,765,057,160 | IssuesEvent | 2023-09-12 08:16:30 | filamentphp/filament | https://api.github.com/repos/filamentphp/filament | opened | Test | bug unconfirmed low priority | ### Package
filament/filament
### Package Version
vNothing
### Laravel Version
vNothing
### Livewire Version
vNothing
### PHP Version
vNothing
### Problem description
N/A
### Expected behavior
N/A
### Steps to reproduce
N/A
### Reproduction repository
N/A
### Relevant log output
```shell
N/A
```
| 1.0 | Test - ### Package
filament/filament
### Package Version
vNothing
### Laravel Version
vNothing
### Livewire Version
vNothing
### PHP Version
vNothing
### Problem description
N/A
### Expected behavior
N/A
### Steps to reproduce
N/A
### Reproduction repository
N/A
### Relevant log output
```shell
N/A
```
| priority | test package filament filament package version vnothing laravel version vnothing livewire version vnothing php version vnothing problem description n a expected behavior n a steps to reproduce n a reproduction repository n a relevant log output shell n a | 1 |
35,609 | 2,791,519,588 | IssuesEvent | 2015-05-10 06:42:47 | minj/foxtrick | https://api.github.com/repos/minj/foxtrick | closed | Consider adding 'force English flags' in CountryList | feature Misc needs-feedback Priority-Low | Original [issue 1217](https://code.google.com/p/foxtrick/issues/detail?id=1217) created by [minj](mailto:4mr.minj@gmail.com) on 2014-07-14T08:32:13.000Z:
**From:** knullig
**PostID:** [(16412248.177)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D177%26v%3D4)
**Reply:** [(16412248.1)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D1%26v%3D4)
**To:** Everyone
**Datetime:** 2014-06-26 1at 3:30
**Message:**
I like to request 2 options:
1. country names everywhere in same language
When i check the map or flaglist of the countries i have played against.
the names of the countries are all in a different language.
for example: spain is displayed as España
it would be great if the same name (if possible always in english) would be used everywhere.
| 1.0 | Consider adding 'force English flags' in CountryList - Original [issue 1217](https://code.google.com/p/foxtrick/issues/detail?id=1217) created by [minj](mailto:4mr.minj@gmail.com) on 2014-07-14T08:32:13.000Z:
**From:** knullig
**PostID:** [(16412248.177)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D177%26v%3D4)
**Reply:** [(16412248.1)](https://www.hattrick.org/goto.ashx?path=/Forum/Read.aspx%3Ft%3D16412248%26n%3D1%26v%3D4)
**To:** Everyone
**Datetime:** 2014-06-26 1at 3:30
**Message:**
I like to request 2 options:
1. country names everywhere in same language
When i check the map or flaglist of the countries i have played against.
the names of the countries are all in a different language.
for example: spain is displayed as España
it would be great if the same name (if possible always in english) would be used everywhere.
| priority | consider adding force english flags in countrylist original created by mailto minj gmail com on from knullig postid reply to everyone datetime message i like to request options country names everywhere in same language when i check the map or flaglist of the countries i have played against the names of the countries are all in a different language for example spain is displayed as españa it would be great if the same name if possible always in english would be used everywhere | 1 |
196,323 | 6,926,823,400 | IssuesEvent | 2017-11-30 20:32:44 | vmware/vic-ui | https://api.github.com/repos/vmware/vic-ui | closed | Create VCH wizard says "VM endpoint" instead of "Endpoint VM" | area/ui priority/low team/lifecycle | In the Compute page of the VCH deployment wizard, the text "VM endpoint settings" should be "VCH endpoint VM settings:

| 1.0 | Create VCH wizard says "VM endpoint" instead of "Endpoint VM" - In the Compute page of the VCH deployment wizard, the text "VM endpoint settings" should be "VCH endpoint VM settings:

| priority | create vch wizard says vm endpoint instead of endpoint vm in the compute page of the vch deployment wizard the text vm endpoint settings should be vch endpoint vm settings | 1 |
396,031 | 11,700,391,601 | IssuesEvent | 2020-03-06 17:23:50 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Investigate cache job runtime increase, part 2! | Priority: Low Team: Echo 🐬 Type: Investigation Type: Tech-Improvement | Runtimes for [the `UpdateCachedAppealsAttributesJob`](https://github.com/department-of-veterans-affairs/caseflow/blob/master/app/jobs/update_cached_appeals_attributes_job.rb) increased from significantly after [recent changes](https://github.com/department-of-veterans-affairs/caseflow/pull/13285). Short runtimes allow us to run the job more frequently and keep the cache more up-to-date. This ticket exists to determine why runtimes increased and address it if possible.

_[Source: Datadog](https://app.datadoghq.com/dashboard/t5t-exj-m32/cronjob-statistics)_
## Acceptance criteria
* [ ] Determine the cause of the increased runtime length
* [ ] Determine if the runtime length can be easily optimized to reduce runtime
* [ ] Make the optimization | 1.0 | Investigate cache job runtime increase, part 2! - Runtimes for [the `UpdateCachedAppealsAttributesJob`](https://github.com/department-of-veterans-affairs/caseflow/blob/master/app/jobs/update_cached_appeals_attributes_job.rb) increased from significantly after [recent changes](https://github.com/department-of-veterans-affairs/caseflow/pull/13285). Short runtimes allow us to run the job more frequently and keep the cache more up-to-date. This ticket exists to determine why runtimes increased and address it if possible.

_[Source: Datadog](https://app.datadoghq.com/dashboard/t5t-exj-m32/cronjob-statistics)_
## Acceptance criteria
* [ ] Determine the cause of the increased runtime length
* [ ] Determine if the runtime length can be easily optimized to reduce runtime
* [ ] Make the optimization | priority | investigate cache job runtime increase part runtimes for increased from significantly after short runtimes allow us to run the job more frequently and keep the cache more up to date this ticket exists to determine why runtimes increased and address it if possible acceptance criteria determine the cause of the increased runtime length determine if the runtime length can be easily optimized to reduce runtime make the optimization | 1 |
298,069 | 9,195,552,186 | IssuesEvent | 2019-03-07 02:54:58 | gw2efficiency/issues | https://api.github.com/repos/gw2efficiency/issues | closed | Miscalculating Deaths/hour and/or Playtime | 1-Type: Bug 2-Priority: C 3-Complexity: Low 4-Impact: Low 5-Area: Account 9-Status: For next release 9-Status: Ready for Release | I have 4,142h 33m Playtime and 7,666 deaths as per the https://gw2efficiency.com/account/statistics page.
In the https://gw2efficiency.com/account/statistics/statistics.deathCountPerHour Page i have 1.85 deaths/hour which seems correct.
In the https://gw2efficiency.com/account/overview Bottom it shows:
You played a total of 4,038 hours across all characters
During that time you died a total of 7,666 times. That's 1.90 deaths per hour
This is using your account data from 4:01:42 PM.
which is also correct for the given data.
But why is there a difference in the Playtime?
Does the first include deleted characters and the other not?
But if so my deleted characters also had a few deaths, wich don't seem to count as both death values are the same.
| 1.0 | Miscalculating Deaths/hour and/or Playtime - I have 4,142h 33m Playtime and 7,666 deaths as per the https://gw2efficiency.com/account/statistics page.
In the https://gw2efficiency.com/account/statistics/statistics.deathCountPerHour Page i have 1.85 deaths/hour which seems correct.
In the https://gw2efficiency.com/account/overview Bottom it shows:
You played a total of 4,038 hours across all characters
During that time you died a total of 7,666 times. That's 1.90 deaths per hour
This is using your account data from 4:01:42 PM.
which is also correct for the given data.
But why is there a difference in the Playtime?
Does the first include deleted characters and the other not?
But if so my deleted characters also had a few deaths, wich don't seem to count as both death values are the same.
| priority | miscalculating deaths hour and or playtime i have playtime and deaths as per the page in the page i have deaths hour which seems correct in the bottom it shows you played a total of hours across all characters during that time you died a total of times that s deaths per hour this is using your account data from pm which is also correct for the given data but why is there a difference in the playtime does the first include deleted characters and the other not but if so my deleted characters also had a few deaths wich don t seem to count as both death values are the same | 1 |
712,032 | 24,482,597,367 | IssuesEvent | 2022-10-09 02:30:26 | chaotic-aur/packages | https://api.github.com/repos/chaotic-aur/packages | closed | [Request] openssl3-git | request:new-pkg priority:low | ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/openssl3-git
### Utility this package has for you
provides library file for ruffle, a flash emulator
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | 1.0 | [Request] openssl3-git - ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/openssl3-git
### Utility this package has for you
provides library file for ruffle, a flash emulator
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | priority | git link to the package s in the aur utility this package has for you provides library file for ruffle a flash emulator do you consider the package s to be useful for every chaotic aur user no but for a few do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information no response | 1 |
185,817 | 6,730,853,842 | IssuesEvent | 2017-10-18 03:48:51 | wireservice/csvkit | https://api.github.com/repos/wireservice/csvkit | closed | Install with Homebrew | feature Low Priority | [Homebrew](http://brew.sh)
Tried with `brew diy` but
```
Error: Couldn't determine build system
```
| 1.0 | Install with Homebrew - [Homebrew](http://brew.sh)
Tried with `brew diy` but
```
Error: Couldn't determine build system
```
| priority | install with homebrew tried with brew diy but error couldn t determine build system | 1 |
485,031 | 13,960,082,618 | IssuesEvent | 2020-10-24 19:20:47 | learnweb/moodle-mod_ratingallocate | https://api.github.com/repos/learnweb/moodle-mod_ratingallocate | closed | Adapt the plugin's code style to Moodle's code style | Effort: Very High Priority: Low enhancement | This is definitely not the most important issue, but should be tackled sometime:
If the code style would conform to Moodle's code style, the plugin would (hopefully) be
- easier to understand (especially for those who are new to the plugin but have experience with Moodle)
- easier to maintain (as a result from the first bullet point).
Furthermore, code quality of future additions could be checked more easily by a continuous integration server by invoking `moodle-plugin-ci codechecker` (using the awesome Moodle Plugin for Travis CI: https://github.com/moodlerooms/moodle-plugin-ci). But this can only happen if the current codebase already adheres to the code style.
| 1.0 | Adapt the plugin's code style to Moodle's code style - This is definitely not the most important issue, but should be tackled sometime:
If the code style would conform to Moodle's code style, the plugin would (hopefully) be
- easier to understand (especially for those who are new to the plugin but have experience with Moodle)
- easier to maintain (as a result from the first bullet point).
Furthermore, code quality of future additions could be checked more easily by a continuous integration server by invoking `moodle-plugin-ci codechecker` (using the awesome Moodle Plugin for Travis CI: https://github.com/moodlerooms/moodle-plugin-ci). But this can only happen if the current codebase already adheres to the code style.
| priority | adapt the plugin s code style to moodle s code style this is definitely not the most important issue but should be tackled sometime if the code style would conform to moodle s code style the plugin would hopefully be easier to understand especially for those who are new to the plugin but have experience with moodle easier to maintain as a result from the first bullet point furthermore code quality of future additions could be checked more easily by a continuous integration server by invoking moodle plugin ci codechecker using the awesome moodle plugin for travis ci but this can only happen if the current codebase already adheres to the code style | 1 |
539,924 | 15,797,128,845 | IssuesEvent | 2021-04-02 16:07:58 | neurostuff/NiMARE | https://api.github.com/repos/neurostuff/NiMARE | closed | Generic data type check and initialization function | effort: low impact: low priority: low refactoring | ## Summary
<!--What would you like changed/added and why?-->
There are a number of methods where we expect an object of a specific class (e.g., a KernelTransformer), either initialized or not. The check we have (e.g., see below) is pretty straightforward, and could easily be abstracted out to its own function, which would make it easier to apply across a range of methods.
https://github.com/neurostuff/NiMARE/blob/5b965f23d8a4e5e7670fb896d657552e77d37bb9/nimare/meta/cbma/base.py#L42-L60
## Additional details
<!--What would be the benefit?-->
I think we'd want the function to take (1) the object, (2) the expected class, and (3) optional keyword arguments. It would then initialize the object as needed (with the keyword args) and check that it's the right type, raising exceptions/warnings as needed. Then it would return the initialized object.
## Next steps
<!--Do you have any ideas about the implementation?-->
1. Implement new function.
2. Add tests.
3. Incorporate new function into existing classes as necessary. | 1.0 | Generic data type check and initialization function - ## Summary
<!--What would you like changed/added and why?-->
There are a number of methods where we expect an object of a specific class (e.g., a KernelTransformer), either initialized or not. The check we have (e.g., see below) is pretty straightforward, and could easily be abstracted out to its own function, which would make it easier to apply across a range of methods.
https://github.com/neurostuff/NiMARE/blob/5b965f23d8a4e5e7670fb896d657552e77d37bb9/nimare/meta/cbma/base.py#L42-L60
## Additional details
<!--What would be the benefit?-->
I think we'd want the function to take (1) the object, (2) the expected class, and (3) optional keyword arguments. It would then initialize the object as needed (with the keyword args) and check that it's the right type, raising exceptions/warnings as needed. Then it would return the initialized object.
## Next steps
<!--Do you have any ideas about the implementation?-->
1. Implement new function.
2. Add tests.
3. Incorporate new function into existing classes as necessary. | priority | generic data type check and initialization function summary there are a number of methods where we expect an object of a specific class e g a kerneltransformer either initialized or not the check we have e g see below is pretty straightforward and could easily be abstracted out to its own function which would make it easier to apply across a range of methods additional details i think we d want the function to take the object the expected class and optional keyword arguments it would then initialize the object as needed with the keyword args and check that it s the right type raising exceptions warnings as needed then it would return the initialized object next steps implement new function add tests incorporate new function into existing classes as necessary | 1 |
355,314 | 10,579,271,950 | IssuesEvent | 2019-10-08 02:00:25 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | closed | OSX: map selector color inconsistencies | Priority: Low Size: Small Type: Bug | (not sure if it only happens on osx)
1. right click map haven't downloaded > add to library (map red) > download map (map green) > restart game > map turns red again (not intentional) > map will appear as not downloaded > right click to "download" map again > map turns white but doesn't download because it's already installed > stays white regardless of what I do afterwards
2. remove a (downloaded) "green" map from library (map turns white) > readd to library > map stays white (not intentional) regardless of what I do afterwards (until it turns red on game restart like above)
I assume removing the maps locally will restart the whole process
| 1.0 | OSX: map selector color inconsistencies - (not sure if it only happens on osx)
1. right click map haven't downloaded > add to library (map red) > download map (map green) > restart game > map turns red again (not intentional) > map will appear as not downloaded > right click to "download" map again > map turns white but doesn't download because it's already installed > stays white regardless of what I do afterwards
2. remove a (downloaded) "green" map from library (map turns white) > readd to library > map stays white (not intentional) regardless of what I do afterwards (until it turns red on game restart like above)
I assume removing the maps locally will restart the whole process
| priority | osx map selector color inconsistencies not sure if it only happens on osx right click map haven t downloaded add to library map red download map map green restart game map turns red again not intentional map will appear as not downloaded right click to download map again map turns white but doesn t download because it s already installed stays white regardless of what i do afterwards remove a downloaded green map from library map turns white readd to library map stays white not intentional regardless of what i do afterwards until it turns red on game restart like above i assume removing the maps locally will restart the whole process | 1 |
552,959 | 16,331,928,287 | IssuesEvent | 2021-05-12 10:16:10 | stackabletech/t2 | https://api.github.com/repos/stackabletech/t2 | closed | configure firewalld properly | priority/low size/M type/enhancement | On the host t2.stackable.tech, we disabled firewalld (host firewall) because it stood in our way and we didn't take the time to configure it properly. As we have an enabled firewall in the network, this isn't risky as of now. Nevertheless, we should take the time to configure it properly "some time". | 1.0 | configure firewalld properly - On the host t2.stackable.tech, we disabled firewalld (host firewall) because it stood in our way and we didn't take the time to configure it properly. As we have an enabled firewall in the network, this isn't risky as of now. Nevertheless, we should take the time to configure it properly "some time". | priority | configure firewalld properly on the host stackable tech we disabled firewalld host firewall because it stood in our way and we didn t take the time to configure it properly as we have an enabled firewall in the network this isn t risky as of now nevertheless we should take the time to configure it properly some time | 1 |
208,438 | 7,154,496,884 | IssuesEvent | 2018-01-26 08:47:53 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | closed | view-source: scheme doesn't load qutebrowser's internal JavaScript | component: QtWebEngine priority: 2 - low qt | As stated in the title, keybindings for scrolling - gg, G, ctrl+d and ctrl+u - don't work in view-source mode, though 'j' and 'k' work normally. I understand that most users don't scroll that much through source, but it would still be a nice thing to have, and doesn't seem much different from ordinary page scrolling.
***
qutebrowser v1.0.4
Git commit:
Backend: QtWebEngine (Chromium 61.0.3163.140)
CPython: 3.6.4
Qt: 5.10.0
PyQt: 5.9.2
sip: 4.19.6
colorama: no
pypeg2: 2.15
jinja2: 2.10
pygments: 2.2.0
yaml: 3.12
cssutils: no
attr: 17.4.0
PyQt5.QtWebEngineWidgets: yes
PyQt5.QtWebKitWidgets: yes
pdf.js: no
sqlite: 3.21.0
QtNetwork SSL: OpenSSL 1.1.0g 2 Nov 2017
Style: QFusionStyle
Platform: Linux-4.14.12-1-ARCH-x86_64-with-arch-Arch-Linux, 64bit
Linux distribution: Arch Linux (arch)
Frozen: False
Imported from /usr/lib/python3.6/site-packages/qutebrowser
Qt library executable path: /usr/lib/qt/libexec, data path: /usr/share/qt
Paths:
cache: /home/lark/.cache/qutebrowser
config: /home/lark/.config/qutebrowser
data: /home/lark/.local/share/qutebrowser
runtime: /run/user/1000/qutebrowser
system data: /usr/share/qutebrowser | 1.0 | view-source: scheme doesn't load qutebrowser's internal JavaScript - As stated in the title, keybindings for scrolling - gg, G, ctrl+d and ctrl+u - don't work in view-source mode, though 'j' and 'k' work normally. I understand that most users don't scroll that much through source, but it would still be a nice thing to have, and doesn't seem much different from ordinary page scrolling.
***
qutebrowser v1.0.4
Git commit:
Backend: QtWebEngine (Chromium 61.0.3163.140)
CPython: 3.6.4
Qt: 5.10.0
PyQt: 5.9.2
sip: 4.19.6
colorama: no
pypeg2: 2.15
jinja2: 2.10
pygments: 2.2.0
yaml: 3.12
cssutils: no
attr: 17.4.0
PyQt5.QtWebEngineWidgets: yes
PyQt5.QtWebKitWidgets: yes
pdf.js: no
sqlite: 3.21.0
QtNetwork SSL: OpenSSL 1.1.0g 2 Nov 2017
Style: QFusionStyle
Platform: Linux-4.14.12-1-ARCH-x86_64-with-arch-Arch-Linux, 64bit
Linux distribution: Arch Linux (arch)
Frozen: False
Imported from /usr/lib/python3.6/site-packages/qutebrowser
Qt library executable path: /usr/lib/qt/libexec, data path: /usr/share/qt
Paths:
cache: /home/lark/.cache/qutebrowser
config: /home/lark/.config/qutebrowser
data: /home/lark/.local/share/qutebrowser
runtime: /run/user/1000/qutebrowser
system data: /usr/share/qutebrowser | priority | view source scheme doesn t load qutebrowser s internal javascript as stated in the title keybindings for scrolling gg g ctrl d and ctrl u don t work in view source mode though j and k work normally i understand that most users don t scroll that much through source but it would still be a nice thing to have and doesn t seem much different from ordinary page scrolling qutebrowser git commit backend qtwebengine chromium cpython qt pyqt sip colorama no pygments yaml cssutils no attr qtwebenginewidgets yes qtwebkitwidgets yes pdf js no sqlite qtnetwork ssl openssl nov style qfusionstyle platform linux arch with arch arch linux linux distribution arch linux arch frozen false imported from usr lib site packages qutebrowser qt library executable path usr lib qt libexec data path usr share qt paths cache home lark cache qutebrowser config home lark config qutebrowser data home lark local share qutebrowser runtime run user qutebrowser system data usr share qutebrowser | 1 |
531,371 | 15,496,195,473 | IssuesEvent | 2021-03-11 02:14:07 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Note Editor: Add Icons to Multimedia Menu | Enhancement Good First Issue! Help Wanted Multimedia Note Editor Priority-Low Stale UI | Good first issue - need to replace icons with XML drawables (Via Add - Vector Icon), then set a style/tint on the XML icons.
https://github.com/ankidroid/Anki-Android/blob/71fdd0445f020ec7f798e29bc08dc13d658d9234/AnkiDroid/src/main/java/com/ichi2/anki/NoteEditor.java#L1405-L1406
Menu currently looks like:
 | 1.0 | Note Editor: Add Icons to Multimedia Menu - Good first issue - need to replace icons with XML drawables (Via Add - Vector Icon), then set a style/tint on the XML icons.
https://github.com/ankidroid/Anki-Android/blob/71fdd0445f020ec7f798e29bc08dc13d658d9234/AnkiDroid/src/main/java/com/ichi2/anki/NoteEditor.java#L1405-L1406
Menu currently looks like:
 | priority | note editor add icons to multimedia menu good first issue need to replace icons with xml drawables via add vector icon then set a style tint on the xml icons menu currently looks like | 1 |
81,406 | 3,590,557,716 | IssuesEvent | 2016-02-01 06:52:18 | jamesmontemagno/Xamarin.Plugins | https://api.github.com/repos/jamesmontemagno/Xamarin.Plugins | closed | [Feature Request] (Media/Android) Add ability to push MediaIntent so that returning doesn't force application resume with Xamarin.Forms | enhancement Media priority-low | Currently on Android, media is taken using an intent. The intent is always launched with the "NewTask" flag, which means that when the user is done getting/taking media, a Xamarin.Forms application will go through the OnResume method. Is it possible to keep the intent in the current application's Task to prevent the OnResume method from being called?
It would be convenient to be able to configure this, at app start or on a per call basis, so either behavior could be used.
(The reason I wish to avoid the OnResume method is because of a pin page being presented during OnResume. The pin page is required for security so I can't remove it. My plan for now is to add an extra flag to avoid the pin page if the developer says to skip it this time, but it's a hack.) | 1.0 | [Feature Request] (Media/Android) Add ability to push MediaIntent so that returning doesn't force application resume with Xamarin.Forms - Currently on Android, media is taken using an intent. The intent is always launched with the "NewTask" flag, which means that when the user is done getting/taking media, a Xamarin.Forms application will go through the OnResume method. Is it possible to keep the intent in the current application's Task to prevent the OnResume method from being called?
It would be convenient to be able to configure this, at app start or on a per call basis, so either behavior could be used.
(The reason I wish to avoid the OnResume method is because of a pin page being presented during OnResume. The pin page is required for security so I can't remove it. My plan for now is to add an extra flag to avoid the pin page if the developer says to skip it this time, but it's a hack.) | priority | media android add ability to push mediaintent so that returning doesn t force application resume with xamarin forms currently on android media is taken using an intent the intent is always launched with the newtask flag which means that when the user is done getting taking media a xamarin forms application will go through the onresume method is it possible to keep the intent in the current application s task to prevent the onresume method from being called it would be convenient to be able to configure this at app start or on a per call basis so either behavior could be used the reason i wish to avoid the onresume method is because of a pin page being presented during onresume the pin page is required for security so i can t remove it my plan for now is to add an extra flag to avoid the pin page if the developer says to skip it this time but it s a hack | 1 |
216,394 | 7,307,618,145 | IssuesEvent | 2018-02-28 03:47:22 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Update CMSSW in the json templates | Low Priority | The following workflows are giving failures when ran on singularity nodes:
TaskChain_LumiMask_multiRun
ReDigi_AllFlag
ReDigi_cmsRun2
StepChain_ReDigi3
StepChain_LumiMask
ReDigi_cmsRun3
ReDigi_LumiMask
at least...
Just in case, the error is:
```
An exception of category 'PluginLibraryLoadError' occurred while
[0] Constructing the EventProcessor
Exception Message:
unable to load /cvmfs/cms.cern.ch/slc6_amd64_gcc481/cms/cmssw/CMSSW_7_2_0/lib/slc6_amd64_gcc481/pluginCondToolsHcalPlugin.so because libaio.so.1: cannot open shared object file: No such file or directory
```
which seems to be coming from an old bug in singularity/CMSSW that hasn't been backported to 7x releases. | 1.0 | Update CMSSW in the json templates - The following workflows are giving failures when ran on singularity nodes:
TaskChain_LumiMask_multiRun
ReDigi_AllFlag
ReDigi_cmsRun2
StepChain_ReDigi3
StepChain_LumiMask
ReDigi_cmsRun3
ReDigi_LumiMask
at least...
Just in case, the error is:
```
An exception of category 'PluginLibraryLoadError' occurred while
[0] Constructing the EventProcessor
Exception Message:
unable to load /cvmfs/cms.cern.ch/slc6_amd64_gcc481/cms/cmssw/CMSSW_7_2_0/lib/slc6_amd64_gcc481/pluginCondToolsHcalPlugin.so because libaio.so.1: cannot open shared object file: No such file or directory
```
which seems to be coming from an old bug in singularity/CMSSW that hasn't been backported to 7x releases. | priority | update cmssw in the json templates the following workflows are giving failures when ran on singularity nodes taskchain lumimask multirun redigi allflag redigi stepchain stepchain lumimask redigi redigi lumimask at least just in case the error is an exception of category pluginlibraryloaderror occurred while constructing the eventprocessor exception message unable to load cvmfs cms cern ch cms cmssw cmssw lib plugincondtoolshcalplugin so because libaio so cannot open shared object file no such file or directory which seems to be coming from an old bug in singularity cmssw that hasn t been backported to releases | 1 |
798,548 | 28,289,559,052 | IssuesEvent | 2023-04-09 02:44:23 | AY2223S2-CS2103T-W09-4/tp | https://api.github.com/repos/AY2223S2-CS2103T-W09-4/tp | closed | [PE-D][Tester D] [Minor] days instead of day | priority.Low type.FeatureFlaw | 
<!--session: 1680242808805-81f07706-8824-452e-a98e-ad93d1f47077--><!--Version: Web v3.4.7-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: rockman007372/ped#9 | 1.0 | [PE-D][Tester D] [Minor] days instead of day - 
<!--session: 1680242808805-81f07706-8824-452e-a98e-ad93d1f47077--><!--Version: Web v3.4.7-->
-------------
Labels: `severity.VeryLow` `type.FunctionalityBug`
original: rockman007372/ped#9 | priority | days instead of day labels severity verylow type functionalitybug original ped | 1 |
644,285 | 20,972,829,348 | IssuesEvent | 2022-03-28 12:58:05 | mito-ds/monorepo | https://api.github.com/repos/mito-ds/monorepo | closed | Pivot table aggregation methods like `median` should be disabled for non-number columns | type: mitosheet effort: 2 priority: low | **Describe the bug**
Add a row to a pivot table, and then a value that is a string column. Set the aggregation method to `mean`, and get an error.
**Expected behavior**
The invalid options for the value aggregation should be disabled (with an error message) on non-number columns (e.g. everything but count, and count unique).
| 1.0 | Pivot table aggregation methods like `median` should be disabled for non-number columns - **Describe the bug**
Add a row to a pivot table, and then a value that is a string column. Set the aggregation method to `mean`, and get an error.
**Expected behavior**
The invalid options for the value aggregation should be disabled (with an error message) on non-number columns (e.g. everything but count, and count unique).
| priority | pivot table aggregation methods like median should be disabled for non number columns describe the bug add a row to a pivot table and then a value that is a string column set the aggregation method to mean and get an error expected behavior the invalid options for the value aggregation should be disabled with an error message on non number columns e g everything but count and count unique | 1 |
56,099 | 3,078,216,071 | IssuesEvent | 2015-08-21 08:43:38 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Папки с медиафайлами в шаре имеют "качество звука" | enhancement imported Priority-Low | _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on March 11, 2011 19:29:43_
Скрин ниже
**Attachment:** [Безымянный.JPG](http://code.google.com/p/flylinkdc/issues/detail?id=392)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=392_ | 1.0 | Папки с медиафайлами в шаре имеют "качество звука" - _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on March 11, 2011 19:29:43_
Скрин ниже
**Attachment:** [Безымянный.JPG](http://code.google.com/p/flylinkdc/issues/detail?id=392)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=392_ | priority | папки с медиафайлами в шаре имеют качество звука from on march скрин ниже attachment original issue | 1 |
491,427 | 14,163,820,106 | IssuesEvent | 2020-11-12 03:20:55 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | closed | info_particle_systems with "start_active 1" don't appear on map load | Priority: Low Size: Medium Type: Bug | **Describe the bug**
Any info_particle_system with the "start_active" keyvalue doesn't appear unless restarted via entity inputs.
**To Reproduce**
Steps to reproduce the behavior:
- Load a [map](https://github.com/momentum-mod/game/files/4366426/firetest.zip) with the aforementioned info_particle_system.
- Do `ent_fire particle stop` followed by `ent_fire particle start`, now the particle appears.
| 1.0 | info_particle_systems with "start_active 1" don't appear on map load - **Describe the bug**
Any info_particle_system with the "start_active" keyvalue doesn't appear unless restarted via entity inputs.
**To Reproduce**
Steps to reproduce the behavior:
- Load a [map](https://github.com/momentum-mod/game/files/4366426/firetest.zip) with the aforementioned info_particle_system.
- Do `ent_fire particle stop` followed by `ent_fire particle start`, now the particle appears.
| priority | info particle systems with start active don t appear on map load describe the bug any info particle system with the start active keyvalue doesn t appear unless restarted via entity inputs to reproduce steps to reproduce the behavior load a with the aforementioned info particle system do ent fire particle stop followed by ent fire particle start now the particle appears | 1 |
477,684 | 13,766,373,003 | IssuesEvent | 2020-10-07 14:29:47 | ansible/awx | https://api.github.com/repos/ansible/awx | opened | Host details variables showing JSON format without indentation on load | component:ui_next priority:low state:needs_devel type:bug | <!-- Issues are for **concrete, actionable bugs and feature requests** only - if you're just asking for debugging help or technical support, please use:
- http://webchat.freenode.net/?channels=ansible-awx
- https://groups.google.com/forum/#!forum/awx-project
We have to limit this because of limited volunteer time to respond to issues! -->
##### ISSUE TYPE
- Bug Report
##### SUMMARY
Host details variables showing JSON format without indentation on load
##### ENVIRONMENT
* AWX version: 15.0.0
* AWX install method:docker for mac
* Ansible version: 2.9.13
* Operating System: Catalina
* Web Browser: Chrome
##### STEPS TO REPRODUCE
1. Navigate to Hosts
2. Select a host which has variables
##### EXPECTED RESULTS
The CodeMirror should show the JSON indented
##### ACTUAL RESULTS
The JSON is displayed without indentation
##### ADDITIONAL INFORMATION

| 1.0 | Host details variables showing JSON format without indentation on load - <!-- Issues are for **concrete, actionable bugs and feature requests** only - if you're just asking for debugging help or technical support, please use:
- http://webchat.freenode.net/?channels=ansible-awx
- https://groups.google.com/forum/#!forum/awx-project
We have to limit this because of limited volunteer time to respond to issues! -->
##### ISSUE TYPE
- Bug Report
##### SUMMARY
Host details variables showing JSON format without indentation on load
##### ENVIRONMENT
* AWX version: 15.0.0
* AWX install method:docker for mac
* Ansible version: 2.9.13
* Operating System: Catalina
* Web Browser: Chrome
##### STEPS TO REPRODUCE
1. Navigate to Hosts
2. Select a host which has variables
##### EXPECTED RESULTS
The CodeMirror should show the JSON indented
##### ACTUAL RESULTS
The JSON is displayed without indentation
##### ADDITIONAL INFORMATION

| priority | host details variables showing json format without indentation on load issues are for concrete actionable bugs and feature requests only if you re just asking for debugging help or technical support please use we have to limit this because of limited volunteer time to respond to issues issue type bug report summary host details variables showing json format without indentation on load environment awx version awx install method docker for mac ansible version operating system catalina web browser chrome steps to reproduce navigate to hosts select a host which has variables expected results the codemirror should show the json indented actual results the json is displayed without indentation additional information | 1 |
709,032 | 24,365,653,485 | IssuesEvent | 2022-10-03 14:58:49 | NCAR/wrfcloud | https://api.github.com/repos/NCAR/wrfcloud | opened | Add workflow manager to runtime components | priority: low alert: NEED MORE DEFINITION component: NWP components | ## Describe the New Feature ##
Consider adding a workflow manager, e.g. ecflow or rocoto, to the runtime components to help manage the monitoring and job dependencies.
### Acceptance Testing ###
*List input data types and sources.*
*Describe tests required for new functionality.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the new feature down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
### Projects and Milestone ###
- [ ] Select **Project**
- [ ] Select **Milestone** as the next official version or **Backlog of Development Ideas**
## New Feature Checklist ##
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>/<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project**, and **Development** issue
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Add workflow manager to runtime components - ## Describe the New Feature ##
Consider adding a workflow manager, e.g. ecflow or rocoto, to the runtime components to help manage the monitoring and job dependencies.
### Acceptance Testing ###
*List input data types and sources.*
*Describe tests required for new functionality.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the new feature down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
### Projects and Milestone ###
- [ ] Select **Project**
- [ ] Select **Milestone** as the next official version or **Backlog of Development Ideas**
## New Feature Checklist ##
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>/<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project**, and **Development** issue
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | add workflow manager to runtime components describe the new feature consider adding a workflow manager e g ecflow or rocoto to the runtime components to help manage the monitoring and job dependencies acceptance testing list input data types and sources describe tests required for new functionality time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the new feature down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority projects and milestone select project select milestone as the next official version or backlog of development ideas new feature checklist complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project and development issue select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
157,577 | 6,008,757,547 | IssuesEvent | 2017-06-06 08:46:16 | lxde/lxqt | https://api.github.com/repos/lxde/lxqt | closed | Some widgets can not be themed | low-priority lxqt-panel qss/themes | Maybe Im missing something, but this are some little issues I have found while theming:
**LXQT Runner**
The searh results (QListView or "commandList") use the system theme, but still refuses to be styled with the lxqt-runner.qss. The last update did not change this, or I did not notice how to do it.
**Panel Plugins**
Show Desktop, Screen Saver, Volume and Mount seems to have a fixed size. If you change the size of the panel to 60px you will see that this widgets stay in a same size while all the others expand with the panel. This leads to them having non clickable areas. This happen also with the old themes.

I have been lurking the code searching for a way to change this behavior. From the stylesheets the only thing that seems to work is to put "height: 100%" but this is not necesary with other widgets (and this lead to other glitches), so I think that there must be something wrong with this plugins.
| 1.0 | Some widgets can not be themed - Maybe Im missing something, but this are some little issues I have found while theming:
**LXQT Runner**
The searh results (QListView or "commandList") use the system theme, but still refuses to be styled with the lxqt-runner.qss. The last update did not change this, or I did not notice how to do it.
**Panel Plugins**
Show Desktop, Screen Saver, Volume and Mount seems to have a fixed size. If you change the size of the panel to 60px you will see that this widgets stay in a same size while all the others expand with the panel. This leads to them having non clickable areas. This happen also with the old themes.

I have been lurking the code searching for a way to change this behavior. From the stylesheets the only thing that seems to work is to put "height: 100%" but this is not necesary with other widgets (and this lead to other glitches), so I think that there must be something wrong with this plugins.
| priority | some widgets can not be themed maybe im missing something but this are some little issues i have found while theming lxqt runner the searh results qlistview or commandlist use the system theme but still refuses to be styled with the lxqt runner qss the last update did not change this or i did not notice how to do it panel plugins show desktop screen saver volume and mount seems to have a fixed size if you change the size of the panel to you will see that this widgets stay in a same size while all the others expand with the panel this leads to them having non clickable areas this happen also with the old themes i have been lurking the code searching for a way to change this behavior from the stylesheets the only thing that seems to work is to put height but this is not necesary with other widgets and this lead to other glitches so i think that there must be something wrong with this plugins | 1 |
194,819 | 6,899,517,119 | IssuesEvent | 2017-11-24 14:08:07 | highcharts/highcharts | https://api.github.com/repos/highcharts/highcharts | closed | One data label not showing. | Bug Priority:Low | When I create a stack bar chart, and choose the stacked bars to be horizontally, along with the labels showing on the stacked bars, one data label is not being displayed.
As shown in the snapshot,

The data label(value) is not shown on stacked bar .
If any other legend is hidden, then the data label on the stacked bar will be displayed correctly.The snapshot is given below.

After again making all the legends enable, the data label will still be displayed.Please refer to the snapshot below.

But after saving the chart, that particular data label is again not displayed on stacked bar
If the legend values are also displayed along with the legends, then the datalabel is displayed.Please refer to the attachment below.

| 1.0 | One data label not showing. - When I create a stack bar chart, and choose the stacked bars to be horizontally, along with the labels showing on the stacked bars, one data label is not being displayed.
As shown in the snapshot,

The data label(value) is not shown on stacked bar .
If any other legend is hidden, then the data label on the stacked bar will be displayed correctly.The snapshot is given below.

After again making all the legends enable, the data label will still be displayed.Please refer to the snapshot below.

But after saving the chart, that particular data label is again not displayed on stacked bar
If the legend values are also displayed along with the legends, then the datalabel is displayed.Please refer to the attachment below.

| priority | one data label not showing when i create a stack bar chart and choose the stacked bars to be horizontally along with the labels showing on the stacked bars one data label is not being displayed as shown in the snapshot the data label value is not shown on stacked bar if any other legend is hidden then the data label on the stacked bar will be displayed correctly the snapshot is given below after again making all the legends enable the data label will still be displayed please refer to the snapshot below but after saving the chart that particular data label is again not displayed on stacked bar if the legend values are also displayed along with the legends then the datalabel is displayed please refer to the attachment below | 1 |
789,482 | 27,791,678,161 | IssuesEvent | 2023-03-17 09:24:13 | VeriFIT/mata | https://api.github.com/repos/VeriFIT/mata | opened | Complementation over non-existent states | For:library Module:nfa Type:discussion Priority:low | In classical complement algorithm implemented in `Mata::Nfa::complement_classical`, we call `Mata::Nfa::Nfa::size()` which returns domain size of initial/final state sets. Those might contain “deleted” states (final/initial at some point but not now, yet still allocated with `false` value in `NumberPredicate`), so when we complement such automaton, non-existent states become final.
As of now, we agreed on leaving it up to the user to call `Mata::Nfa::Nfa::trim` before complementing. We might want to reconsider this decision in the future. | 1.0 | Complementation over non-existent states - In classical complement algorithm implemented in `Mata::Nfa::complement_classical`, we call `Mata::Nfa::Nfa::size()` which returns domain size of initial/final state sets. Those might contain “deleted” states (final/initial at some point but not now, yet still allocated with `false` value in `NumberPredicate`), so when we complement such automaton, non-existent states become final.
As of now, we agreed on leaving it up to the user to call `Mata::Nfa::Nfa::trim` before complementing. We might want to reconsider this decision in the future. | priority | complementation over non existent states in classical complement algorithm implemented in mata nfa complement classical we call mata nfa nfa size which returns domain size of initial final state sets those might contain “deleted” states final initial at some point but not now yet still allocated with false value in numberpredicate so when we complement such automaton non existent states become final as of now we agreed on leaving it up to the user to call mata nfa nfa trim before complementing we might want to reconsider this decision in the future | 1 |
585,245 | 17,483,511,838 | IssuesEvent | 2021-08-09 07:53:18 | chaosblade-io/chaosblade | https://api.github.com/repos/chaosblade-io/chaosblade | closed | Does the dd command in the mac system not support the inflag attribute? There will be problems when simulating the disk fill scene | priority/low type/feature chaosblade-exec-os | 我在mac上执行下面命令会报错:
blade create disk fill -d --mount-point /home --size 1024
错误为:
{"code":604,"success":false,"error":"dd: unknown operand iflag\n exit status 1 exit status 1"}
看了一下上面的讨论,我直接执行下面的命令也会报错
命令为:
dd if=/dev/zero of=/home/chaos_filldisk.log.dat bs=1b count=1 iflag=fullblock
错误为:
dd: unknown operand iflag | 1.0 | Does the dd command in the mac system not support the inflag attribute? There will be problems when simulating the disk fill scene - 我在mac上执行下面命令会报错:
blade create disk fill -d --mount-point /home --size 1024
错误为:
{"code":604,"success":false,"error":"dd: unknown operand iflag\n exit status 1 exit status 1"}
看了一下上面的讨论,我直接执行下面的命令也会报错
命令为:
dd if=/dev/zero of=/home/chaos_filldisk.log.dat bs=1b count=1 iflag=fullblock
错误为:
dd: unknown operand iflag | priority | does the dd command in the mac system not support the inflag attribute there will be problems when simulating the disk fill scene 我在mac上执行下面命令会报错: blade create disk fill d mount point home size 错误为: code success false error dd unknown operand iflag n exit status exit status 看了一下上面的讨论,我直接执行下面的命令也会报错 命令为: dd if dev zero of home chaos filldisk log dat bs count iflag fullblock 错误为: dd unknown operand iflag | 1 |
59,326 | 3,105,473,842 | IssuesEvent | 2015-08-31 21:06:11 | UniVR/GolfVR | https://api.github.com/repos/UniVR/GolfVR | opened | Explain rules introduction | priority:low type:idea | Explain the rules before the game begin
_Watch the ball to shoot (the time the ball is watched will define the power)
_... (to be defined) | 1.0 | Explain rules introduction - Explain the rules before the game begin
_Watch the ball to shoot (the time the ball is watched will define the power)
_... (to be defined) | priority | explain rules introduction explain the rules before the game begin watch the ball to shoot the time the ball is watched will define the power to be defined | 1 |
233,092 | 7,693,577,662 | IssuesEvent | 2018-05-18 04:38:28 | ElektraInitiative/libelektra | https://api.github.com/repos/ElektraInitiative/libelektra | opened | doc: changes to icheck script | low priority | Forgot to add section to doc/news/_preparation_next_release.md mentioning that the icheck script now no longer leaves the base directory. | 1.0 | doc: changes to icheck script - Forgot to add section to doc/news/_preparation_next_release.md mentioning that the icheck script now no longer leaves the base directory. | priority | doc changes to icheck script forgot to add section to doc news preparation next release md mentioning that the icheck script now no longer leaves the base directory | 1 |
744,867 | 25,958,769,527 | IssuesEvent | 2022-12-18 15:51:19 | CosmosOS/Cosmos | https://api.github.com/repos/CosmosOS/Cosmos | closed | Deleting file and directory not working | Bug Up for Grabs Complexity: Medium Priority: Low Area: File System | #### Area of Cosmos - What area of Cosmos are we dealing with?
File system
#### Expected Behaviour - What do you think that should happen?
I should be able to delete files and directories when I want.
#### Actual Behaviour - What unexpectedly happens?
I can only delete files and directories after I created them, after I close VMware and I re-open it (via the Start button in VS 2019) I can't delete them anymore. **However this only seems to happen to folders and files I create, as I was able to delete the directory TEST, though it got created automatically.**
#### Reproduction - How did you get this error to appear?
Initialize a CosmosVFS, then create a file or directory, close and open VMware again (from VS 2019 via the Start button, or F5) and try to delete the file or directory you've created.
#### Version - Were you using the User Kit or Dev Kit? And what User Kit version or Dev Kit commit (Cosmos, IL2CPU, X#)?
Latest User Kit (20200708) | 1.0 | Deleting file and directory not working - #### Area of Cosmos - What area of Cosmos are we dealing with?
File system
#### Expected Behaviour - What do you think that should happen?
I should be able to delete files and directories when I want.
#### Actual Behaviour - What unexpectedly happens?
I can only delete files and directories after I created them, after I close VMware and I re-open it (via the Start button in VS 2019) I can't delete them anymore. **However this only seems to happen to folders and files I create, as I was able to delete the directory TEST, though it got created automatically.**
#### Reproduction - How did you get this error to appear?
Initialize a CosmosVFS, then create a file or directory, close and open VMware again (from VS 2019 via the Start button, or F5) and try to delete the file or directory you've created.
#### Version - Were you using the User Kit or Dev Kit? And what User Kit version or Dev Kit commit (Cosmos, IL2CPU, X#)?
Latest User Kit (20200708) | priority | deleting file and directory not working area of cosmos what area of cosmos are we dealing with file system expected behaviour what do you think that should happen i should be able to delete files and directories when i want actual behaviour what unexpectedly happens i can only delete files and directories after i created them after i close vmware and i re open it via the start button in vs i can t delete them anymore however this only seems to happen to folders and files i create as i was able to delete the directory test though it got created automatically reproduction how did you get this error to appear initialize a cosmosvfs then create a file or directory close and open vmware again from vs via the start button or and try to delete the file or directory you ve created version were you using the user kit or dev kit and what user kit version or dev kit commit cosmos x latest user kit | 1 |
626,133 | 19,785,976,500 | IssuesEvent | 2022-01-18 06:47:32 | Valdes-Tresanco-MS/gmx_MMPBSA | https://api.github.com/repos/Valdes-Tresanco-MS/gmx_MMPBSA | closed | Improve gmx_MMPBSA_test | low priority improvement | - [ ] Add a flag to overwrite or not the folder where the repository with the examples was cloned.
It would be something like `gmx_MMPBSA_test` [`-nr` / `-noreuse`] (Default = True)
`-nr` will store a 'store_false' action
- [ ] Add a flag to delete the test folder once finished
- [x] Add the use of MPI for the examples if possible | 1.0 | Improve gmx_MMPBSA_test - - [ ] Add a flag to overwrite or not the folder where the repository with the examples was cloned.
It would be something like `gmx_MMPBSA_test` [`-nr` / `-noreuse`] (Default = True)
`-nr` will store a 'store_false' action
- [ ] Add a flag to delete the test folder once finished
- [x] Add the use of MPI for the examples if possible | priority | improve gmx mmpbsa test add a flag to overwrite or not the folder where the repository with the examples was cloned it would be something like gmx mmpbsa test default true nr will store a store false action add a flag to delete the test folder once finished add the use of mpi for the examples if possible | 1 |
387,374 | 11,460,299,217 | IssuesEvent | 2020-02-07 09:27:36 | nhsconnect/gpconnect-demonstrator | https://api.github.com/repos/nhsconnect/gpconnect-demonstrator | closed | Add representative clinical data to the demonstrator | 1.2.3 Low Priority Structured | Placeholder note for now, while clinical data is produced by Pete and colleagues. | 1.0 | Add representative clinical data to the demonstrator - Placeholder note for now, while clinical data is produced by Pete and colleagues. | priority | add representative clinical data to the demonstrator placeholder note for now while clinical data is produced by pete and colleagues | 1 |
264,547 | 8,316,517,518 | IssuesEvent | 2018-09-25 09:16:24 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | Command with template throws an error when command.visible handler is added | Bug C: Grid Kendo1 Priority 1 SEV: Low | ### Reproduction of the problem
1. Run [this dojo](https://dojo.telerik.com/OHeXajOF)
1. Error occurs and the content of the grid does not render
### Environment
* **Kendo UI version:** 2018.3.911
| 1.0 | Command with template throws an error when command.visible handler is added - ### Reproduction of the problem
1. Run [this dojo](https://dojo.telerik.com/OHeXajOF)
1. Error occurs and the content of the grid does not render
### Environment
* **Kendo UI version:** 2018.3.911
| priority | command with template throws an error when command visible handler is added reproduction of the problem run error occurs and the content of the grid does not render environment kendo ui version | 1 |
621,911 | 19,599,346,374 | IssuesEvent | 2022-01-05 22:18:01 | softmatterlab/Braph-2.0-Matlab | https://api.github.com/repos/softmatterlab/Braph-2.0-Matlab | opened | Minimize and Close Plot | low priority GUI | Write functions to minimize and close figures as in cb_bring_to_front():
- [ ] base functions cb_minimize() and cb_close() in Plot
- [ ] follow-up functions where needed (e.g., GUI, PlotElement, PPBrainAtlas_Surf, PPGroup_SUBDIct, ...)
- [ ] change callbacks for closing windows to these functions. | 1.0 | Minimize and Close Plot - Write functions to minimize and close figures as in cb_bring_to_front():
- [ ] base functions cb_minimize() and cb_close() in Plot
- [ ] follow-up functions where needed (e.g., GUI, PlotElement, PPBrainAtlas_Surf, PPGroup_SUBDIct, ...)
- [ ] change callbacks for closing windows to these functions. | priority | minimize and close plot write functions to minimize and close figures as in cb bring to front base functions cb minimize and cb close in plot follow up functions where needed e g gui plotelement ppbrainatlas surf ppgroup subdict change callbacks for closing windows to these functions | 1 |
271,581 | 8,485,379,223 | IssuesEvent | 2018-10-26 07:36:01 | swe-ms-boun/2018fall-swe574-g1 | https://api.github.com/repos/swe-ms-boun/2018fall-swe574-g1 | opened | Annotations page | module.annotation priority.low tier.ui type.feature | * Development of annotations page. Associated requirements:
* The system should allow registered users to view their annotations.
* The system should allow the registered users to go to the memories by using their annotations.
* Registered users should be able to view their all tags they use when annotating the memories on the annotations page.
* The system shall display a total number of annotations on the annotations page. | 1.0 | Annotations page - * Development of annotations page. Associated requirements:
* The system should allow registered users to view their annotations.
* The system should allow the registered users to go to the memories by using their annotations.
* Registered users should be able to view their all tags they use when annotating the memories on the annotations page.
* The system shall display a total number of annotations on the annotations page. | priority | annotations page development of annotations page associated requirements the system should allow registered users to view their annotations the system should allow the registered users to go to the memories by using their annotations registered users should be able to view their all tags they use when annotating the memories on the annotations page the system shall display a total number of annotations on the annotations page | 1 |
748,998 | 26,146,893,299 | IssuesEvent | 2022-12-30 07:01:04 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [Quest] The Gift That Keeps On Giving | Quest - Cataclysm (1-60) Quest - Event Animation Priority: Low Status: Confirmed Bug Report from Discord | Unintendet
OP
— 12/10/2022 6:43 PM
Quest = The Gift That Keeps On Giving only turns one Miner into an Ghoul, it should turn everyone in that Radius into one. The Miners in General aren't moving theyr Carts, they just walk around it.
| 1.0 | [Quest] The Gift That Keeps On Giving - Unintendet
OP
— 12/10/2022 6:43 PM
Quest = The Gift That Keeps On Giving only turns one Miner into an Ghoul, it should turn everyone in that Radius into one. The Miners in General aren't moving theyr Carts, they just walk around it.
| priority | the gift that keeps on giving unintendet op — pm quest the gift that keeps on giving only turns one miner into an ghoul it should turn everyone in that radius into one the miners in general aren t moving theyr carts they just walk around it | 1 |
711,811 | 24,476,250,628 | IssuesEvent | 2022-10-08 07:41:39 | borgbase/vorta | https://api.github.com/repos/borgbase/vorta | closed | Error when adding locked repo | priority:low type:bug | **To Reproduce**
Steps to reproduce the behavior:
1. Init a repository to a remote ssh using borg commandline 1.1.15
```
borg init user@host:/home/user/borg_backup --encryption=keyfile --remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"
```
2. Create a backup from command line but cancel it in the middle
```
borg create --stats user@host:/home/user/borg_backup::First ~/ --remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"
```
3. Try to add this repo to Vorta (using extra args `--remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"`
4. Get error below
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Vorta version: Vorta 0.7.7, Borg 1.1.15
- Installed from: pipx
Vorta and Borg versions can be found in Main Window > Misc Tab.
**Additional context**
```
2021-08-19 17:01:49,558 - root - DEBUG - Found 0 passwords matching repo URL.
2021-08-19 17:01:49,558 - vorta.borg.borg_thread - DEBUG - Password not found in primary keyring. Falling back to VortaDBKeyring.
2021-08-19 17:01:49,568 - vorta.borg.borg_thread - INFO - Running command /usr/bin/borg info --remote-path=/usr/bin/borg-1.1.14/borg-1.1.14 --info --json --log-json user@host:/home/user/borg_backup
2021-08-19 17:04:06,167 - vorta.borg.borg_thread - ERROR - Failed to create/acquire the lock /home/user/borg_backup/lock.exclusive (timeout).
2021-08-19 17:04:06,171 - root - CRITICAL - Uncaught exception, file a report at https://github.com/borgbase/vorta/issues/new
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6884, in get
return clone.execute(database)[0]
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 4275, in __getitem__
return self.row_cache[item]
IndexError: list index out of range
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/vorta/application.py", line 221, in react_to_log
profile = BackupProfileModel.get(name=context['profile_name'])
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6438, in get
return sq.get()
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6887, in get
raise self.model.DoesNotExist('%s instance matching query does '
vorta.models.BackupProfileModelDoesNotExist: <Model: BackupProfileModel> instance matching query does not exist:
SQL: SELECT "t1"."id", "t1"."name", "t1"."added_at", "t1"."repo_id", "t1"."ssh_key", "t1"."compression", "t1"."exclude_patterns", "t1"."exclude_if_present", "t1"."schedule_mode", "t1"."schedule_interval_hours", "t1"."schedule_interval_minutes", "t1"."schedule_fixed_hour", "t1"."schedule_fixed_minute", "t1"."validation_on", "t1"."validation_weeks", "t1"."prune_on", "t1"."prune_hour", "t1"."prune_day", "t1"."prune_week", "t1"."prune_month", "t1"."prune_year", "t1"."prune_keep_within", "t1"."new_archive_name", "t1"."prune_prefix", "t1"."pre_backup_cmd", "t1"."post_backup_cmd", "t1"."dont_run_on_metered_networks" FROM "backupprofilemodel" AS "t1" WHERE ("t1"."name" = ?) LIMIT ? OFFSET ?
Params: ['New Repo', 1, 0]
2021-08-19 17:07:22,981 - vorta.keyring.abc - DEBUG - Only available on macOS
```
```
Uncaught exception, please file a report with this text at
https://github.com/borgbase/vorta/issues/new
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6884, in get
return clone.execute(database)[0]
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 4275, in __getitem__
return self.row_cache[item]
IndexError: list index out of range
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/vorta/application.py", line 221, in react_to_log
profile = BackupProfileModel.get(name=context['profile_name'])
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6438, in get
return sq.get()
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6887, in get
raise self.model.DoesNotExist('%s instance matching query does '
vorta.models.BackupProfileModelDoesNotExist: <Model: BackupProfileModel> instance matching query does not exist:
SQL: SELECT "t1"."id", "t1"."name", "t1"."added_at", "t1"."repo_id", "t1"."ssh_key", "t1"."compression", "t1"."exclude_patterns", "t1"."exclude_if_present", "t1"."schedule_mode", "t1"."schedule_interval_hours", "t1"."schedule_interval_minutes", "t1"."schedule_fixed_hour", "t1"."schedule_fixed_minute", "t1"."validation_on", "t1"."validation_weeks", "t1"."prune_on", "t1"."prune_hour", "t1"."prune_day", "t1"."prune_week", "t1"."prune_month", "t1"."prune_year", "t1"."prune_keep_within", "t1"."new_archive_name", "t1"."prune_prefix", "t1"."pre_backup_cmd", "t1"."post_backup_cmd", "t1"."dont_run_on_metered_networks" FROM "backupprofilemodel" AS "t1" WHERE ("t1"."name" = ?) LIMIT ? OFFSET ?
Params: ['New Repo', 1, 0]
``` | 1.0 | Error when adding locked repo - **To Reproduce**
Steps to reproduce the behavior:
1. Init a repository to a remote ssh using borg commandline 1.1.15
```
borg init user@host:/home/user/borg_backup --encryption=keyfile --remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"
```
2. Create a backup from command line but cancel it in the middle
```
borg create --stats user@host:/home/user/borg_backup::First ~/ --remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"
```
3. Try to add this repo to Vorta (using extra args `--remote-path="/usr/bin/borg-1.1.14/borg-1.1.14"`
4. Get error below
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Vorta version: Vorta 0.7.7, Borg 1.1.15
- Installed from: pipx
Vorta and Borg versions can be found in Main Window > Misc Tab.
**Additional context**
```
2021-08-19 17:01:49,558 - root - DEBUG - Found 0 passwords matching repo URL.
2021-08-19 17:01:49,558 - vorta.borg.borg_thread - DEBUG - Password not found in primary keyring. Falling back to VortaDBKeyring.
2021-08-19 17:01:49,568 - vorta.borg.borg_thread - INFO - Running command /usr/bin/borg info --remote-path=/usr/bin/borg-1.1.14/borg-1.1.14 --info --json --log-json user@host:/home/user/borg_backup
2021-08-19 17:04:06,167 - vorta.borg.borg_thread - ERROR - Failed to create/acquire the lock /home/user/borg_backup/lock.exclusive (timeout).
2021-08-19 17:04:06,171 - root - CRITICAL - Uncaught exception, file a report at https://github.com/borgbase/vorta/issues/new
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6884, in get
return clone.execute(database)[0]
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 4275, in __getitem__
return self.row_cache[item]
IndexError: list index out of range
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/vorta/application.py", line 221, in react_to_log
profile = BackupProfileModel.get(name=context['profile_name'])
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6438, in get
return sq.get()
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6887, in get
raise self.model.DoesNotExist('%s instance matching query does '
vorta.models.BackupProfileModelDoesNotExist: <Model: BackupProfileModel> instance matching query does not exist:
SQL: SELECT "t1"."id", "t1"."name", "t1"."added_at", "t1"."repo_id", "t1"."ssh_key", "t1"."compression", "t1"."exclude_patterns", "t1"."exclude_if_present", "t1"."schedule_mode", "t1"."schedule_interval_hours", "t1"."schedule_interval_minutes", "t1"."schedule_fixed_hour", "t1"."schedule_fixed_minute", "t1"."validation_on", "t1"."validation_weeks", "t1"."prune_on", "t1"."prune_hour", "t1"."prune_day", "t1"."prune_week", "t1"."prune_month", "t1"."prune_year", "t1"."prune_keep_within", "t1"."new_archive_name", "t1"."prune_prefix", "t1"."pre_backup_cmd", "t1"."post_backup_cmd", "t1"."dont_run_on_metered_networks" FROM "backupprofilemodel" AS "t1" WHERE ("t1"."name" = ?) LIMIT ? OFFSET ?
Params: ['New Repo', 1, 0]
2021-08-19 17:07:22,981 - vorta.keyring.abc - DEBUG - Only available on macOS
```
```
Uncaught exception, please file a report with this text at
https://github.com/borgbase/vorta/issues/new
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6884, in get
return clone.execute(database)[0]
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 4275, in __getitem__
return self.row_cache[item]
IndexError: list index out of range
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/vorta/application.py", line 221, in react_to_log
profile = BackupProfileModel.get(name=context['profile_name'])
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6438, in get
return sq.get()
File "/home/guru/.local/pipx/venvs/vorta/lib/python3.8/site-packages/peewee.py", line 6887, in get
raise self.model.DoesNotExist('%s instance matching query does '
vorta.models.BackupProfileModelDoesNotExist: <Model: BackupProfileModel> instance matching query does not exist:
SQL: SELECT "t1"."id", "t1"."name", "t1"."added_at", "t1"."repo_id", "t1"."ssh_key", "t1"."compression", "t1"."exclude_patterns", "t1"."exclude_if_present", "t1"."schedule_mode", "t1"."schedule_interval_hours", "t1"."schedule_interval_minutes", "t1"."schedule_fixed_hour", "t1"."schedule_fixed_minute", "t1"."validation_on", "t1"."validation_weeks", "t1"."prune_on", "t1"."prune_hour", "t1"."prune_day", "t1"."prune_week", "t1"."prune_month", "t1"."prune_year", "t1"."prune_keep_within", "t1"."new_archive_name", "t1"."prune_prefix", "t1"."pre_backup_cmd", "t1"."post_backup_cmd", "t1"."dont_run_on_metered_networks" FROM "backupprofilemodel" AS "t1" WHERE ("t1"."name" = ?) LIMIT ? OFFSET ?
Params: ['New Repo', 1, 0]
``` | priority | error when adding locked repo to reproduce steps to reproduce the behavior init a repository to a remote ssh using borg commandline borg init user host home user borg backup encryption keyfile remote path usr bin borg borg create a backup from command line but cancel it in the middle borg create stats user host home user borg backup first remote path usr bin borg borg try to add this repo to vorta using extra args remote path usr bin borg borg get error below environment please complete the following information os ubuntu vorta version vorta borg installed from pipx vorta and borg versions can be found in main window misc tab additional context root debug found passwords matching repo url vorta borg borg thread debug password not found in primary keyring falling back to vortadbkeyring vorta borg borg thread info running command usr bin borg info remote path usr bin borg borg info json log json user host home user borg backup vorta borg borg thread error failed to create acquire the lock home user borg backup lock exclusive timeout root critical uncaught exception file a report at traceback most recent call last file home guru local pipx venvs vorta lib site packages peewee py line in get return clone execute database file home guru local pipx venvs vorta lib site packages peewee py line in getitem return self row cache indexerror list index out of range during handling of the above exception another exception occurred traceback most recent call last file home guru local pipx venvs vorta lib site packages vorta application py line in react to log profile backupprofilemodel get name context file home guru local pipx venvs vorta lib site packages peewee py line in get return sq get file home guru local pipx venvs vorta lib site packages peewee py line in get raise self model doesnotexist s instance matching query does vorta models backupprofilemodeldoesnotexist instance matching query does not exist sql select id name added at repo id ssh key compression exclude patterns exclude if present schedule mode schedule interval hours schedule interval minutes schedule fixed hour schedule fixed minute validation on validation weeks prune on prune hour prune day prune week prune month prune year prune keep within new archive name prune prefix pre backup cmd post backup cmd dont run on metered networks from backupprofilemodel as where name limit offset params vorta keyring abc debug only available on macos uncaught exception please file a report with this text at traceback most recent call last file home guru local pipx venvs vorta lib site packages peewee py line in get return clone execute database file home guru local pipx venvs vorta lib site packages peewee py line in getitem return self row cache indexerror list index out of range during handling of the above exception another exception occurred traceback most recent call last file home guru local pipx venvs vorta lib site packages vorta application py line in react to log profile backupprofilemodel get name context file home guru local pipx venvs vorta lib site packages peewee py line in get return sq get file home guru local pipx venvs vorta lib site packages peewee py line in get raise self model doesnotexist s instance matching query does vorta models backupprofilemodeldoesnotexist instance matching query does not exist sql select id name added at repo id ssh key compression exclude patterns exclude if present schedule mode schedule interval hours schedule interval minutes schedule fixed hour schedule fixed minute validation on validation weeks prune on prune hour prune day prune week prune month prune year prune keep within new archive name prune prefix pre backup cmd post backup cmd dont run on metered networks from backupprofilemodel as where name limit offset params | 1 |
767,387 | 26,922,289,844 | IssuesEvent | 2023-02-07 11:19:22 | AFM-SPM/TopoStats | https://api.github.com/repos/AFM-SPM/TopoStats | closed | Remove extraneous plotting from grainstats.py | good first issue Low Priority GrainStats | `GrainStats()` currently has lines that are used to plot the data, but no image is saved (see lines [158-160](https://github.com/AFM-SPM/TopoStats/blob/main/topostats/grainstats.py#L158)).
These are likely from development/debugging and weren't picked up during review of the PR. | 1.0 | Remove extraneous plotting from grainstats.py - `GrainStats()` currently has lines that are used to plot the data, but no image is saved (see lines [158-160](https://github.com/AFM-SPM/TopoStats/blob/main/topostats/grainstats.py#L158)).
These are likely from development/debugging and weren't picked up during review of the PR. | priority | remove extraneous plotting from grainstats py grainstats currently has lines that are used to plot the data but no image is saved see lines these are likely from development debugging and weren t picked up during review of the pr | 1 |
177,225 | 6,575,802,913 | IssuesEvent | 2017-09-11 17:23:10 | suluke/hfg-particles | https://api.github.com/repos/suluke/hfg-particles | closed | [Effect] Windblow | low priority | Hit all particles with a one-directional force, but with random strength (and also weakend to the edges?) | 1.0 | [Effect] Windblow - Hit all particles with a one-directional force, but with random strength (and also weakend to the edges?) | priority | windblow hit all particles with a one directional force but with random strength and also weakend to the edges | 1 |
441,463 | 12,718,460,860 | IssuesEvent | 2020-06-24 07:34:32 | frontity/frontity | https://api.github.com/repos/frontity/frontity | closed | Filter the `when` prop from the Switch component children | bug good first issue priority: low | <!-- Not following the template might result in your issue being closed -->
**_Bug report_**
- I've updated both my project and my packages to the latest Frontity versions: https://docs.frontity.org/guides/keep-frontity-updated
- I've searched for existing issues.
- I've checked the documentation: https://docs.frontity.org
- I've checked the community forum: https://community.frontity.org
**Expected behavior**
The Switch component uses a `when` prop to determine which component it should render:
```js
<Switch>
<Loading when={data.isFetching} />
<Home when={data.isHome} />
<Archive when={data.isArchive} />
<Post when={data.isPostType} />
<ErrorPage /> {/* rendered by default */}
</Switch>
```
More info on: https://docs.frontity.org/api-reference-1/frontity-components#switch
The `when` prop is only needed for `Switch`, but it shouldn't be passed down to the rendered component.
**Observed behavior**
The `when` prop is sent to the rendered component.
**Possible solution**
`Switch` can remove the `when` prop from the props of the child that is going to render. I guess it would need to get access to the props object and do something like:
```js
const { when, ...otherProps } = component.props;
// ...
return <component {...otherProps} />;
```
The code is here:
https://github.com/frontity/frontity/blob/dev/packages/components/switch.tsx#L28-L32
A test for this can be a modification of this test, where the prop `when` is missing: https://github.com/frontity/frontity/blob/add-test-for-switch-props/packages/components/__tests__/switch.tests.tsx#L55-L66 | 1.0 | Filter the `when` prop from the Switch component children - <!-- Not following the template might result in your issue being closed -->
**_Bug report_**
- I've updated both my project and my packages to the latest Frontity versions: https://docs.frontity.org/guides/keep-frontity-updated
- I've searched for existing issues.
- I've checked the documentation: https://docs.frontity.org
- I've checked the community forum: https://community.frontity.org
**Expected behavior**
The Switch component uses a `when` prop to determine which component it should render:
```js
<Switch>
<Loading when={data.isFetching} />
<Home when={data.isHome} />
<Archive when={data.isArchive} />
<Post when={data.isPostType} />
<ErrorPage /> {/* rendered by default */}
</Switch>
```
More info on: https://docs.frontity.org/api-reference-1/frontity-components#switch
The `when` prop is only needed for `Switch`, but it shouldn't be passed down to the rendered component.
**Observed behavior**
The `when` prop is sent to the rendered component.
**Possible solution**
`Switch` can remove the `when` prop from the props of the child that is going to render. I guess it would need to get access to the props object and do something like:
```js
const { when, ...otherProps } = component.props;
// ...
return <component {...otherProps} />;
```
The code is here:
https://github.com/frontity/frontity/blob/dev/packages/components/switch.tsx#L28-L32
A test for this can be a modification of this test, where the prop `when` is missing: https://github.com/frontity/frontity/blob/add-test-for-switch-props/packages/components/__tests__/switch.tests.tsx#L55-L66 | priority | filter the when prop from the switch component children bug report i ve updated both my project and my packages to the latest frontity versions i ve searched for existing issues i ve checked the documentation i ve checked the community forum expected behavior the switch component uses a when prop to determine which component it should render js rendered by default more info on the when prop is only needed for switch but it shouldn t be passed down to the rendered component observed behavior the when prop is sent to the rendered component possible solution switch can remove the when prop from the props of the child that is going to render i guess it would need to get access to the props object and do something like js const when otherprops component props return the code is here a test for this can be a modification of this test where the prop when is missing | 1 |
755,461 | 26,429,806,030 | IssuesEvent | 2023-01-14 17:17:04 | OffprintStudios/Sailfish | https://api.github.com/repos/OffprintStudios/Sailfish | closed | Setting cover art for work displays "Updating avatar" text | bug low priority | When setting the cover art for a work, the popup that appears says "updating avatar" rather than more appropriate text. | 1.0 | Setting cover art for work displays "Updating avatar" text - When setting the cover art for a work, the popup that appears says "updating avatar" rather than more appropriate text. | priority | setting cover art for work displays updating avatar text when setting the cover art for a work the popup that appears says updating avatar rather than more appropriate text | 1 |
214,032 | 7,262,746,067 | IssuesEvent | 2018-02-19 08:01:38 | pmem/issues | https://api.github.com/repos/pmem/issues | closed | test: obj_tx_locks_abort/TEST0: SETUP (all/pmem/debug/memcheck) | Exposure: Low OS: Linux Priority: 4 low State: To be verified Type: Bug | Found on dd622819dd4ee97d3920f913c70be371b5ee215f
```
==25613== Memcheck, a memory error detector
==25613== Copyright (C) 2002-2015, and GNU GPL'd, by Julian Seward et al.
==25613== Using Valgrind-3.12.0 and LibVEX; rerun with -h for copyright info
==25613== Command: ./obj_tx_locks_abort.static-nondebug /dev/shm/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/test_obj_tx_locks_abort0😘⠝⠧⠍⠇ɗPMDKӜ⥺🙋/testfile1
==25613== Parent PID: 25591
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC05D0,0x6DC0628> (<offset:0x3c05d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC05D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404412: do_nested_tx (obj_tx_locks_abort.c:61)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc05d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404B4D: main (obj_tx_locks_abort.c:140)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0650,0x6DC06A8> (<offset:0x3c0650,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0654
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc0654 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC06D0,0x6DC0728> (<offset:0x3c06d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC06D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x4044CF: do_nested_tx (obj_tx_locks_abort.c:69)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc06d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0750,0x6DC07A8> (<offset:0x3c0750,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0754
**25613** Snapshotting uninitialized data in range <0x6DC05D0,0x6DC0628> (<offset:0x3c05d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC05D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404674: do_aborted_nested_tx (obj_tx_locks_abort.c:83)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc05d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404B4D: main (obj_tx_locks_abort.c:140)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0650,0x6DC06A8> (<offset:0x3c0650,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0654
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x4046FB: do_aborted_nested_tx (obj_tx_locks_abort.c:90)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc0654 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC06D0,0x6DC0728> (<offset:0x3c06d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC06D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x404731: do_aborted_nested_tx (obj_tx_locks_abort.c:91)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc06d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0750,0x6DC07A8> (<offset:0x3c0750,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0754
==25613==
==25613== HEAP SUMMARY:
==25613== in use at exit: 3,372 bytes in 42 blocks
==25613== total heap usage: 12,305 allocs, 12,257 frees, 8,161,681 bytes allocated
==25613==
==25613== LEAK SUMMARY:
==25613== definitely lost: 0 bytes in 0 blocks
==25613== indirectly lost: 0 bytes in 0 blocks
==25613== possibly lost: 0 bytes in 0 blocks
==25613== still reachable: 0 bytes in 0 blocks
==25613== suppressed: 3,372 bytes in 42 blocks
==25613==
==25613== For counts of detected and suppressed errors, rerun with: -v
==25613== Use --track-origins=yes to see where uninitialised values come from
==25613== ERROR SUMMARY: 8 errors from 6 contexts (suppressed: 3 from 3)
``` | 1.0 | test: obj_tx_locks_abort/TEST0: SETUP (all/pmem/debug/memcheck) - Found on dd622819dd4ee97d3920f913c70be371b5ee215f
```
==25613== Memcheck, a memory error detector
==25613== Copyright (C) 2002-2015, and GNU GPL'd, by Julian Seward et al.
==25613== Using Valgrind-3.12.0 and LibVEX; rerun with -h for copyright info
==25613== Command: ./obj_tx_locks_abort.static-nondebug /dev/shm/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/LoremipsumdolorsitametconsecteturadipiscingelitVivamuslacinianibhattortordictumsollicitudinNullamvariusvestibulumligulaetegestaselitsemperidMaurisultriciesligulaeuipsumtinciduntluctusMorbimaximusvariusdolorid/test_obj_tx_locks_abort0😘⠝⠧⠍⠇ɗPMDKӜ⥺🙋/testfile1
==25613== Parent PID: 25591
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC05D0,0x6DC0628> (<offset:0x3c05d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC05D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404412: do_nested_tx (obj_tx_locks_abort.c:61)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc05d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404B4D: main (obj_tx_locks_abort.c:140)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0650,0x6DC06A8> (<offset:0x3c0650,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0654
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc0654 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC06D0,0x6DC0728> (<offset:0x3c06d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC06D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x4044CF: do_nested_tx (obj_tx_locks_abort.c:69)
==25613== by 0x404E38: main (obj_tx_locks_abort.c:158)
==25613== Address 0x6dc06d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0750,0x6DC07A8> (<offset:0x3c0750,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0754
**25613** Snapshotting uninitialized data in range <0x6DC05D0,0x6DC0628> (<offset:0x3c05d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC05D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404674: do_aborted_nested_tx (obj_tx_locks_abort.c:83)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc05d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404B4D: main (obj_tx_locks_abort.c:140)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0650,0x6DC06A8> (<offset:0x3c0650,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0654
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x4046FB: do_aborted_nested_tx (obj_tx_locks_abort.c:90)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc0654 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC06D0,0x6DC0728> (<offset:0x3c06d0,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC06D4
==25613== Uninitialised byte(s) found during client check request
==25613== at 0x425A98: pmemobj_tx_add_common.part.25.constprop.29 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426CCC: pmemobj_tx_add_range (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404499: do_nested_tx (obj_tx_locks_abort.c:68)
==25613== by 0x404731: do_aborted_nested_tx (obj_tx_locks_abort.c:91)
==25613== by 0x404EC2: main (obj_tx_locks_abort.c:162)
==25613== Address 0x6dc06d4 is 4 bytes inside a block of size 112 client-defined
==25613== at 0x41E55E: palloc_reservation_create (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x425DC7: tx_alloc_common.isra.26.constprop.33 (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x426E35: pmemobj_tx_alloc (in /home/jenkins/workspace/pmdk/src/test/obj_tx_locks_abort/obj_tx_locks_abort.static-nondebug)
==25613== by 0x404C38: main (obj_tx_locks_abort.c:146)
==25613==
**25613** Snapshotting uninitialized data in range <0x6DC0750,0x6DC07A8> (<offset:0x3c0750,size:0x58>)
**25613** Uninitialized data starts at: 0x6DC0754
==25613==
==25613== HEAP SUMMARY:
==25613== in use at exit: 3,372 bytes in 42 blocks
==25613== total heap usage: 12,305 allocs, 12,257 frees, 8,161,681 bytes allocated
==25613==
==25613== LEAK SUMMARY:
==25613== definitely lost: 0 bytes in 0 blocks
==25613== indirectly lost: 0 bytes in 0 blocks
==25613== possibly lost: 0 bytes in 0 blocks
==25613== still reachable: 0 bytes in 0 blocks
==25613== suppressed: 3,372 bytes in 42 blocks
==25613==
==25613== For counts of detected and suppressed errors, rerun with: -v
==25613== Use --track-origins=yes to see where uninitialised values come from
==25613== ERROR SUMMARY: 8 errors from 6 contexts (suppressed: 3 from 3)
``` | priority | test obj tx locks abort setup all pmem debug memcheck found on memcheck a memory error detector copyright c and gnu gpl d by julian seward et al using valgrind and libvex rerun with h for copyright info command obj tx locks abort static nondebug dev shm loremipsumdolorsitametconsecteturadipiscingelitvivamuslacinianibhattortordictumsollicitudinnullamvariusvestibulumligulaetegestaselitsemperidmaurisultriciesligulaeuipsumtinciduntluctusmorbimaximusvariusdolorid loremipsumdolorsitametconsecteturadipiscingelitvivamuslacinianibhattortordictumsollicitudinnullamvariusvestibulumligulaetegestaselitsemperidmaurisultriciesligulaeuipsumtinciduntluctusmorbimaximusvariusdolorid loremipsumdolorsitametconsecteturadipiscingelitvivamuslacinianibhattortordictumsollicitudinnullamvariusvestibulumligulaetegestaselitsemperidmaurisultriciesligulaeuipsumtinciduntluctusmorbimaximusvariusdolorid loremipsumdolorsitametconsecteturadipiscingelitvivamuslacinianibhattortordictumsollicitudinnullamvariusvestibulumligulaetegestaselitsemperidmaurisultriciesligulaeuipsumtinciduntluctusmorbimaximusvariusdolorid loremipsumdolorsitametconsecteturadipiscingelitvivamuslacinianibhattortordictumsollicitudinnullamvariusvestibulumligulaetegestaselitsemperidmaurisultriciesligulaeuipsumtinciduntluctusmorbimaximusvariusdolorid test obj tx locks 😘⠝⠧⠍⠇ɗpmdkӝ⥺🙋 parent pid snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do nested tx obj tx locks abort c by do nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do aborted nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do aborted nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at uninitialised byte s found during client check request at pmemobj tx add common part constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx add range in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by do nested tx obj tx locks abort c by do aborted nested tx obj tx locks abort c by main obj tx locks abort c address is bytes inside a block of size client defined at palloc reservation create in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by tx alloc common isra constprop in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by pmemobj tx alloc in home jenkins workspace pmdk src test obj tx locks abort obj tx locks abort static nondebug by main obj tx locks abort c snapshotting uninitialized data in range uninitialized data starts at heap summary in use at exit bytes in blocks total heap usage allocs frees bytes allocated leak summary definitely lost bytes in blocks indirectly lost bytes in blocks possibly lost bytes in blocks still reachable bytes in blocks suppressed bytes in blocks for counts of detected and suppressed errors rerun with v use track origins yes to see where uninitialised values come from error summary errors from contexts suppressed from | 1 |
469,258 | 13,504,368,040 | IssuesEvent | 2020-09-13 17:42:02 | shahednasser/sbuttons | https://api.github.com/repos/shahednasser/sbuttons | closed | Disable selection of logo text | Priority: Low first-timers-only good first issue help wanted up-for-grabs | Add a rule to `.logo h1` to disable user selection. Make sure to add the changes to `assets/css/index.css`. Check out the screenshot for an example

| 1.0 | Disable selection of logo text - Add a rule to `.logo h1` to disable user selection. Make sure to add the changes to `assets/css/index.css`. Check out the screenshot for an example

| priority | disable selection of logo text add a rule to logo to disable user selection make sure to add the changes to assets css index css check out the screenshot for an example | 1 |
798,151 | 28,238,169,224 | IssuesEvent | 2023-04-06 03:45:45 | AY2223S2-CS2103T-F12-4/tp | https://api.github.com/repos/AY2223S2-CS2103T-F12-4/tp | closed | [PE-D][Tester D] Capitalisation for jar file name | priority.Low severity.Low | UG states that jar file is Ultron.jar but the actual jar file is ultron.jar. However, it does not affect functionality as jar file names are case insensitive.

<!--session: 1680242296715-29bb0fea-05f1-41ae-b4d9-71039571a17e-->
<!--Version: Web v3.4.7-->
-------------
Labels: `type.DocumentationBug` `severity.VeryLow`
original: gyulong1/ped#9 | 1.0 | [PE-D][Tester D] Capitalisation for jar file name - UG states that jar file is Ultron.jar but the actual jar file is ultron.jar. However, it does not affect functionality as jar file names are case insensitive.

<!--session: 1680242296715-29bb0fea-05f1-41ae-b4d9-71039571a17e-->
<!--Version: Web v3.4.7-->
-------------
Labels: `type.DocumentationBug` `severity.VeryLow`
original: gyulong1/ped#9 | priority | capitalisation for jar file name ug states that jar file is ultron jar but the actual jar file is ultron jar however it does not affect functionality as jar file names are case insensitive labels type documentationbug severity verylow original ped | 1 |
298,659 | 9,200,638,507 | IssuesEvent | 2019-03-07 17:32:10 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | name of layer can't be seen when it's created | Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **pascal-ferrand-tiscali-fr -** (pascal-ferrand-tiscali-fr -)
Original Redmine Issue: 224, https://issues.qgis.org/issues/224
Original Assignee: Gary Sherman
---
on GQIS-0.8.0_preview-1_source working on linux
we can't see name of vector layer at its creation in "légende". it appear only after have been delete and call.
it's the same with qgis-0.8.0-win32-060724
| 1.0 | name of layer can't be seen when it's created - ---
Author Name: **pascal-ferrand-tiscali-fr -** (pascal-ferrand-tiscali-fr -)
Original Redmine Issue: 224, https://issues.qgis.org/issues/224
Original Assignee: Gary Sherman
---
on GQIS-0.8.0_preview-1_source working on linux
we can't see name of vector layer at its creation in "légende". it appear only after have been delete and call.
it's the same with qgis-0.8.0-win32-060724
| priority | name of layer can t be seen when it s created author name pascal ferrand tiscali fr pascal ferrand tiscali fr original redmine issue original assignee gary sherman on gqis preview source working on linux we can t see name of vector layer at its creation in légende it appear only after have been delete and call it s the same with qgis | 1 |
227,356 | 7,531,292,867 | IssuesEvent | 2018-04-15 03:37:50 | MrBlizzard/RCAdmins-Tracker | https://api.github.com/repos/MrBlizzard/RCAdmins-Tracker | opened | [Perms] Rank perms required to create Vote Crates | awaiting information priority:low | Admin is able to create vote, vip, PokemonEgg and FossilEgg crates, however Deity and Challenger cannot. No confirmed information on other ranks, further information needed. | 1.0 | [Perms] Rank perms required to create Vote Crates - Admin is able to create vote, vip, PokemonEgg and FossilEgg crates, however Deity and Challenger cannot. No confirmed information on other ranks, further information needed. | priority | rank perms required to create vote crates admin is able to create vote vip pokemonegg and fossilegg crates however deity and challenger cannot no confirmed information on other ranks further information needed | 1 |
702,872 | 24,139,590,139 | IssuesEvent | 2022-09-21 13:56:52 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | opened | Switch is not rendered as expected when used inside a popup editor | Bug SEV: Low C: Grid jQuery2 Priority 5 C: Switch | ### Bug report
When the Switch is used inside a Kendo Grid popup it is not rendered as expected.
### Reproduction of the problem
1. Open the Dojo - https://dojo.telerik.com/@NeliKondova/uRaJITug
2. Click on the Edit button
### Current behavior
The Switch is rendered incorrectly:

### Expected/desired behavior
The Switch thumb should be rendered inside the Switch track.
#### The issue is a regression starting width Kendo 2022.3.913
### Environment
* **Kendo UI version:** 2022.3.913
* **Browser:** [all ]
| 1.0 | Switch is not rendered as expected when used inside a popup editor - ### Bug report
When the Switch is used inside a Kendo Grid popup it is not rendered as expected.
### Reproduction of the problem
1. Open the Dojo - https://dojo.telerik.com/@NeliKondova/uRaJITug
2. Click on the Edit button
### Current behavior
The Switch is rendered incorrectly:

### Expected/desired behavior
The Switch thumb should be rendered inside the Switch track.
#### The issue is a regression starting width Kendo 2022.3.913
### Environment
* **Kendo UI version:** 2022.3.913
* **Browser:** [all ]
| priority | switch is not rendered as expected when used inside a popup editor bug report when the switch is used inside a kendo grid popup it is not rendered as expected reproduction of the problem open the dojo click on the edit button current behavior the switch is rendered incorrectly expected desired behavior the switch thumb should be rendered inside the switch track the issue is a regression starting width kendo environment kendo ui version browser | 1 |
718,617 | 24,725,763,796 | IssuesEvent | 2022-10-20 13:57:31 | GIScience/oshdb | https://api.github.com/repos/GIScience/oshdb | closed | add aggregateByGeometry method with callback function | enhancement priority:low | from https://github.com/GIScience/oshdb/issues/165#issuecomment-487129692:
> I believe it wouldn't be too hard to add a second `aggregateByGeometry` method that not only accepts the map of geometries but also a function that returns a "feature" geometry. Then one could use the `aggregateByGeometry` functionality also in even more use cases, e.g. where the geometry needs to be modified before the aggregate step (e.g. computing a buffer around a snapshot geometry, or calculating the mean centroid of all contributions happening on a single osm entity, etc.). | 1.0 | add aggregateByGeometry method with callback function - from https://github.com/GIScience/oshdb/issues/165#issuecomment-487129692:
> I believe it wouldn't be too hard to add a second `aggregateByGeometry` method that not only accepts the map of geometries but also a function that returns a "feature" geometry. Then one could use the `aggregateByGeometry` functionality also in even more use cases, e.g. where the geometry needs to be modified before the aggregate step (e.g. computing a buffer around a snapshot geometry, or calculating the mean centroid of all contributions happening on a single osm entity, etc.). | priority | add aggregatebygeometry method with callback function from i believe it wouldn t be too hard to add a second aggregatebygeometry method that not only accepts the map of geometries but also a function that returns a feature geometry then one could use the aggregatebygeometry functionality also in even more use cases e g where the geometry needs to be modified before the aggregate step e g computing a buffer around a snapshot geometry or calculating the mean centroid of all contributions happening on a single osm entity etc | 1 |
396,409 | 11,709,009,071 | IssuesEvent | 2020-03-08 16:26:53 | jrabbit/pyborg-1up | https://api.github.com/repos/jrabbit/pyborg-1up | closed | mastodon/twitter parity | Low-Priority/Wishlist mastodon | should respect (ignore) cws, #nobot tags in profiles and other fediverse politenesses | 1.0 | mastodon/twitter parity - should respect (ignore) cws, #nobot tags in profiles and other fediverse politenesses | priority | mastodon twitter parity should respect ignore cws nobot tags in profiles and other fediverse politenesses | 1 |
67,576 | 3,275,274,648 | IssuesEvent | 2015-10-26 14:58:39 | PhenotypeFoundation/PhenotypeDatabase | https://api.github.com/repos/PhenotypeFoundation/PhenotypeDatabase | closed | Wrong redirect after I delete a study from My Studies-list | enhancement Priority low | When I view one of my studies in the study list, and then 'Delete WHOLE STUDY', I'm redirected to the page 'Browse all studies'.
However, I came from My studies.
This is confusing, since these lists are different.
(On hyve server, gscf 0.9.1.1) | 1.0 | Wrong redirect after I delete a study from My Studies-list - When I view one of my studies in the study list, and then 'Delete WHOLE STUDY', I'm redirected to the page 'Browse all studies'.
However, I came from My studies.
This is confusing, since these lists are different.
(On hyve server, gscf 0.9.1.1) | priority | wrong redirect after i delete a study from my studies list when i view one of my studies in the study list and then delete whole study i m redirected to the page browse all studies however i came from my studies this is confusing since these lists are different on hyve server gscf | 1 |
296,420 | 9,115,465,861 | IssuesEvent | 2019-02-22 05:07:54 | WeAreDevs/material-clicker | https://api.github.com/repos/WeAreDevs/material-clicker | opened | Achievements | enhancement priority: low | When doing certain things, you can trigger an achievement. I propose an API Something similar to promises.
```js
registerAchievement({
id: 'win-universe',
name: 'Win the Universe'
description: 'Example to show how cool the world is.'
}, (award) => {
// add event handles and stuff to detect if they get the achievement
award(); // call award to give the achievement.
});
``` | 1.0 | Achievements - When doing certain things, you can trigger an achievement. I propose an API Something similar to promises.
```js
registerAchievement({
id: 'win-universe',
name: 'Win the Universe'
description: 'Example to show how cool the world is.'
}, (award) => {
// add event handles and stuff to detect if they get the achievement
award(); // call award to give the achievement.
});
``` | priority | achievements when doing certain things you can trigger an achievement i propose an api something similar to promises js registerachievement id win universe name win the universe description example to show how cool the world is award add event handles and stuff to detect if they get the achievement award call award to give the achievement | 1 |
249,702 | 7,964,659,940 | IssuesEvent | 2018-07-13 22:38:25 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Display a tool-tip asking user to change the default admin password | Priority: Low | Users aren't changing the default password for admin. We could add a tool-tip for instances with default password so they are inclined to do so. | 1.0 | Display a tool-tip asking user to change the default admin password - Users aren't changing the default password for admin. We could add a tool-tip for instances with default password so they are inclined to do so. | priority | display a tool tip asking user to change the default admin password users aren t changing the default password for admin we could add a tool tip for instances with default password so they are inclined to do so | 1 |
72,050 | 3,371,408,300 | IssuesEvent | 2015-11-23 19:00:13 | aaroneiche/do-want | https://api.github.com/repos/aaroneiche/do-want | closed | Implement OAuth | enhancement imported Priority-Low | _Original author: aaron.ei...@gmail.com (September 06, 2012 16:41:34)_
Do-Want should support OAuth authentication through popular services like google, facebook, and twitter.
_Original issue: http://code.google.com/p/do-want/issues/detail?id=16_ | 1.0 | Implement OAuth - _Original author: aaron.ei...@gmail.com (September 06, 2012 16:41:34)_
Do-Want should support OAuth authentication through popular services like google, facebook, and twitter.
_Original issue: http://code.google.com/p/do-want/issues/detail?id=16_ | priority | implement oauth original author aaron ei gmail com september do want should support oauth authentication through popular services like google facebook and twitter original issue | 1 |
350,781 | 10,508,486,341 | IssuesEvent | 2019-09-27 08:46:08 | woocommerce/woocommerce-gateway-paypal-express-checkout | https://api.github.com/repos/woocommerce/woocommerce-gateway-paypal-express-checkout | closed | Synchronise Renewals Not Supported ATM | Priority: Low [Type] Enhancement | Hello,
I am delighted to see that PayPal Subscriptions for download/virtual products are now supported. However, I noticed that apparently Synchronise renewals are not a thing with PayPal? Is this something that can be added?
I tried this with the same product and switching the setting to sync to 1st of the months and not sync at all and in the checkout process, it offered PayPal only when Sync Renewals were turned off for the subscription I was trying to make.
PayPal CHeckout Version: 1.6.16
WooCommerce Version: 3.6.5
WooCommerce Subscriptions Version: 2.5.7
Thanks to all putting in the effort for the Subscription support in the first place,
Chris | 1.0 | Synchronise Renewals Not Supported ATM - Hello,
I am delighted to see that PayPal Subscriptions for download/virtual products are now supported. However, I noticed that apparently Synchronise renewals are not a thing with PayPal? Is this something that can be added?
I tried this with the same product and switching the setting to sync to 1st of the months and not sync at all and in the checkout process, it offered PayPal only when Sync Renewals were turned off for the subscription I was trying to make.
PayPal CHeckout Version: 1.6.16
WooCommerce Version: 3.6.5
WooCommerce Subscriptions Version: 2.5.7
Thanks to all putting in the effort for the Subscription support in the first place,
Chris | priority | synchronise renewals not supported atm hello i am delighted to see that paypal subscriptions for download virtual products are now supported however i noticed that apparently synchronise renewals are not a thing with paypal is this something that can be added i tried this with the same product and switching the setting to sync to of the months and not sync at all and in the checkout process it offered paypal only when sync renewals were turned off for the subscription i was trying to make paypal checkout version woocommerce version woocommerce subscriptions version thanks to all putting in the effort for the subscription support in the first place chris | 1 |
742,340 | 25,849,995,158 | IssuesEvent | 2022-12-13 09:44:08 | l7mp/stunner | https://api.github.com/repos/l7mp/stunner | closed | stunnerd stops responding to filesystem events | priority: low status: cannot reproduce type: bug | The `stunnerd` pod seems to hang, with the following logs:
```
14:09:00.748041 main.go:148: stunnerd WARNING: unhnadled notify op on config file "/etc/stunnerd/stunnerd.conf" (ignoring): CHMOD
14:09:00.748077 main.go:133: stunnerd WARNING: config file deleted "REMOVE", disabling watcher
14:09:00.748090 main.go:138: stunnerd WARNING: could not remove config file "/etc/stunnerd/stunnerd.conf" from watcher: can't remove non-existent inotify watch for: /etc/stunnerd/stunnerd.conf
14:09:01.400101 reconcile.go:145: stunner INFO: reconciliation ready: new objects: 0, changed objects: 3, deleted objects: 0
```
After this point, config file updates are not picked up by `stunnerd` any more | 1.0 | stunnerd stops responding to filesystem events - The `stunnerd` pod seems to hang, with the following logs:
```
14:09:00.748041 main.go:148: stunnerd WARNING: unhnadled notify op on config file "/etc/stunnerd/stunnerd.conf" (ignoring): CHMOD
14:09:00.748077 main.go:133: stunnerd WARNING: config file deleted "REMOVE", disabling watcher
14:09:00.748090 main.go:138: stunnerd WARNING: could not remove config file "/etc/stunnerd/stunnerd.conf" from watcher: can't remove non-existent inotify watch for: /etc/stunnerd/stunnerd.conf
14:09:01.400101 reconcile.go:145: stunner INFO: reconciliation ready: new objects: 0, changed objects: 3, deleted objects: 0
```
After this point, config file updates are not picked up by `stunnerd` any more | priority | stunnerd stops responding to filesystem events the stunnerd pod seems to hang with the following logs main go stunnerd warning unhnadled notify op on config file etc stunnerd stunnerd conf ignoring chmod main go stunnerd warning config file deleted remove disabling watcher main go stunnerd warning could not remove config file etc stunnerd stunnerd conf from watcher can t remove non existent inotify watch for etc stunnerd stunnerd conf reconcile go stunner info reconciliation ready new objects changed objects deleted objects after this point config file updates are not picked up by stunnerd any more | 1 |
162,317 | 6,150,799,408 | IssuesEvent | 2017-06-27 23:50:02 | chaos/pdsh | https://api.github.com/repos/chaos/pdsh | closed | Support user defined variables in pdsh | auto-migrated Priority-Low Type-Enhancement | ```
Hi!
It would be nice having user defined variables per host in pdsh.
This could be the first "execute different command by host (I see a wishlist
bug here talking about)" implementation:
Like %h is hostname.
We could define:
$ cat ~/.pdsh/vars/host01:
ports="8080 8081"
$ cat ~/.pdsh/vars/host02:
ports="8080"
Then:
pdsh -w ^filewithhosts "
for port in %ports
do
wget -O - --user user --password pass localhost:\$port/manager/list
done
"
Do you think it could be possible in the future ?
Thank you developping this awesome tool !
```
Original issue reported on code.google.com by `javibarr...@gmail.com` on 9 Sep 2011 at 6:32
| 1.0 | Support user defined variables in pdsh - ```
Hi!
It would be nice having user defined variables per host in pdsh.
This could be the first "execute different command by host (I see a wishlist
bug here talking about)" implementation:
Like %h is hostname.
We could define:
$ cat ~/.pdsh/vars/host01:
ports="8080 8081"
$ cat ~/.pdsh/vars/host02:
ports="8080"
Then:
pdsh -w ^filewithhosts "
for port in %ports
do
wget -O - --user user --password pass localhost:\$port/manager/list
done
"
Do you think it could be possible in the future ?
Thank you developping this awesome tool !
```
Original issue reported on code.google.com by `javibarr...@gmail.com` on 9 Sep 2011 at 6:32
| priority | support user defined variables in pdsh hi it would be nice having user defined variables per host in pdsh this could be the first execute different command by host i see a wishlist bug here talking about implementation like h is hostname we could define cat pdsh vars ports cat pdsh vars ports then pdsh w filewithhosts for port in ports do wget o user user password pass localhost port manager list done do you think it could be possible in the future thank you developping this awesome tool original issue reported on code google com by javibarr gmail com on sep at | 1 |
526 | 2,497,833,277 | IssuesEvent | 2015-01-07 11:25:42 | HubTurbo/HubTurbo | https://api.github.com/repos/HubTurbo/HubTurbo | opened | Make '<' the default 'updated' qualifier | aspect-ui feature-filters priority.low status.accepted type.enhancement | I guess this will be the most used one, so it helps if it can be the default | 1.0 | Make '<' the default 'updated' qualifier - I guess this will be the most used one, so it helps if it can be the default | priority | make the default updated qualifier i guess this will be the most used one so it helps if it can be the default | 1 |
214,528 | 7,274,151,125 | IssuesEvent | 2018-02-21 08:59:00 | STEP-tw/battleship-phoenix | https://api.github.com/repos/STEP-tw/battleship-phoenix | closed | View settings | Low Priority small | As a _player_
I want to _have settings_
So that I may personalize language and settings_
**Additional Details**
User has landed on homepage.
Settings should be below about game option
**Acceptance Criteria**
- [x] Criteria 1
- Given _settings option_
- When _I click on settings_
- Then _I should see settings box with music, sound, language and rate_
- [x] Criteria 2
- Given _settings option_
- When _I have done setting_
- Then _I should see a cancel button at right top corner of settings box_
| 1.0 | View settings - As a _player_
I want to _have settings_
So that I may personalize language and settings_
**Additional Details**
User has landed on homepage.
Settings should be below about game option
**Acceptance Criteria**
- [x] Criteria 1
- Given _settings option_
- When _I click on settings_
- Then _I should see settings box with music, sound, language and rate_
- [x] Criteria 2
- Given _settings option_
- When _I have done setting_
- Then _I should see a cancel button at right top corner of settings box_
| priority | view settings as a player i want to have settings so that i may personalize language and settings additional details user has landed on homepage settings should be below about game option acceptance criteria criteria given settings option when i click on settings then i should see settings box with music sound language and rate criteria given settings option when i have done setting then i should see a cancel button at right top corner of settings box | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.