Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
660,007 | 21,948,305,781 | IssuesEvent | 2022-05-24 04:44:35 | DeFiCh/wallet | https://api.github.com/repos/DeFiCh/wallet | closed | Price Rates - order of token display | triage/accepted kind/feature area/ui-ux priority/low | <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
The order of the token displayed in price rates should be aligned with https://defiscan.live/
Follow the order of the pairing. If it's dBTC-DFI, then dBTC should be displayed first.
Align with Scan and go with the non-DUSD/DFI token as the “primary” token. Some examples to illustrate:
**dBTC-DFI**
```
Price Rates:
dBTC = ...
DFI = ...
```
**dTLSA-DUSD**
```
Price Rates:
dTSLA = ...
DUSD = ...
```
**DUSD-DFI**
```
Price Rates:
DUSD = ...
DFI = ...
```

- [ ] Main DEX Screen
- [ ] Check on Swap page if applicable
- [ ] Check on Add Liquidity page if applicable
- [ ] Check on Remove Liquidity page if applicable
#### Why is this needed:
This works better if we want to move towards a proper DEX trading experience
| 1.0 | Price Rates - order of token display - <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
The order of the token displayed in price rates should be aligned with https://defiscan.live/
Follow the order of the pairing. If it's dBTC-DFI, then dBTC should be displayed first.
Align with Scan and go with the non-DUSD/DFI token as the “primary” token. Some examples to illustrate:
**dBTC-DFI**
```
Price Rates:
dBTC = ...
DFI = ...
```
**dTLSA-DUSD**
```
Price Rates:
dTSLA = ...
DUSD = ...
```
**DUSD-DFI**
```
Price Rates:
DUSD = ...
DFI = ...
```

- [ ] Main DEX Screen
- [ ] Check on Swap page if applicable
- [ ] Check on Add Liquidity page if applicable
- [ ] Check on Remove Liquidity page if applicable
#### Why is this needed:
This works better if we want to move towards a proper DEX trading experience
| priority | price rates order of token display what would you like to be added the order of the token displayed in price rates should be aligned with follow the order of the pairing if it s dbtc dfi then dbtc should be displayed first align with scan and go with the non dusd dfi token as the “primary” token some examples to illustrate dbtc dfi price rates dbtc dfi dtlsa dusd price rates dtsla dusd dusd dfi price rates dusd dfi main dex screen check on swap page if applicable check on add liquidity page if applicable check on remove liquidity page if applicable why is this needed this works better if we want to move towards a proper dex trading experience | 1 |
257,355 | 8,136,301,283 | IssuesEvent | 2018-08-20 07:56:31 | ow2-proactive/scheduling-portal | https://api.github.com/repos/ow2-proactive/scheduling-portal | opened | When a user doesn't have the right to execute scripts from the Resource Manager portal, a Http 500 error is reported | priority:low severity:minor type:bug | When a user doesn't have the right to execute scripts from the RM Portal, the portal reports a Http 500 error. This error message is confusing. Better replace it by a message like "You are not authorised to execute scripts on this node, please contact the node's administrator." | 1.0 | When a user doesn't have the right to execute scripts from the Resource Manager portal, a Http 500 error is reported - When a user doesn't have the right to execute scripts from the RM Portal, the portal reports a Http 500 error. This error message is confusing. Better replace it by a message like "You are not authorised to execute scripts on this node, please contact the node's administrator." | priority | when a user doesn t have the right to execute scripts from the resource manager portal a http error is reported when a user doesn t have the right to execute scripts from the rm portal the portal reports a http error this error message is confusing better replace it by a message like you are not authorised to execute scripts on this node please contact the node s administrator | 1 |
542,779 | 15,866,617,088 | IssuesEvent | 2021-04-08 15:56:33 | ESCOMP/CTSM | https://api.github.com/repos/ESCOMP/CTSM | closed | external munging with cdeps and fox | priority: low tag: next type: -external type: enhancement | ### Brief summary of bug
The following generates an unclean state in components/cdeps that I'm not quite sure how to clean out:
```
git clone git@github.com:ESCOMP/CTSM.git ctsm-test-cdeps
cd ctsm-test-cdeps/
git remote add ckoven_repo git@github.com:ckoven/CTSM.git
git fetch ckoven_repo
./manage_externals/checkout_externals
git checkout -b snow_occlusion_ctsm ckoven_repo/snow_occlusion_ctsm
./manage_externals/checkout_externals
git checkout master
./manage_externals/checkout_externals
```
generates this message:
```
cheyenne4 rgknox/ctsm-test-cdeps> ./manage_externals/checkout_externals
Processing externals description file : Externals.cfg
Processing externals description file : Externals_CLM.cfg
Processing externals description file : Externals_CISM.cfg
Processing externals description file : .gitmodules
Processing submodules description file : .gitmodules
Processing externals description file : .gitmodules
Processing submodules description file : .gitmodules
Checking status of externals: clm, fates, ptclm, mosart, mizuroute, cime, rtm, cism, source_cism, cdeps, fox, cmeps, nems/lib/parallelio, nems/lib/genf90, doc-builder,
s ./cime
s ./cime/src/drivers/nuopc/
./cime/src/drivers/nuopc/nems/lib/ParallelIO
./cime/src/drivers/nuopc/nems/lib/genf90
sM ./components/cdeps
./components/cdeps/fox
s ./components/cism
./components/cism/source_cism
./components/mizuRoute
s ./components/mosart
s ./components/rtm
e-o ./doc/doc-builder
./src/fates
./tools/PTCLM
----------------------------------------------------------------------
The external repositories labeled with 'M' above are not in a clean state.
The following are three options for how to proceed:
(1) Go into each external that is not in a clean state and issue either a 'git status' or
an 'svn status' command (depending on whether the external is managed by git or
svn). Either revert or commit your changes so that all externals are in a clean
state. (To revert changes in git, follow the instructions given when you run 'git
status'.) (Note, though, that it is okay to have untracked files in your working
directory.) Then rerun checkout_externals.
(2) Alternatively, you do not have to rely on checkout_externals. Instead, you can manually
update out-of-sync externals (labeled with 's' above) as described in the
configuration file Externals.cfg. (For example, run 'git fetch' and 'git checkout'
commands to checkout the appropriate tags for each external, as given in
Externals.cfg.)
(3) You can also use checkout_externals to manage most, but not all externals: You can specify
one or more externals to ignore using the '-x' or '--exclude' argument to
checkout_externals. Excluding externals labeled with 'M' will allow checkout_externals to
update the other, non-excluded externals.
The external repositories labeled with '?' above are not under version
control using the expected protocol. If you are sure you want to switch
protocols, and you don't have any work you need to save from this
directory, then run "rm -rf [directory]" before re-running the
checkout_externals tool.
----------------------------------------------------------------------
```
Here is the status of cdeps:
```
cheyenne4 components/cdeps> git status
HEAD detached at 45b7a85
Changes not staged for commit:
(use "git add <file>..." to update what will be committed)
(use "git checkout -- <file>..." to discard changes in working directory)
modified: fox (new commits)
Untracked files:
(use "git add <file>..." to include in what will be committed)
share/genf90/
no changes added to commit (use "git add" and/or "git commit -a")
```
I step into fox and here is the status:
```
cheyenne4 cdeps/fox> git status
HEAD detached at 4ff17b4
nothing to commit, working tree clean
```
| 1.0 | external munging with cdeps and fox - ### Brief summary of bug
The following generates an unclean state in components/cdeps that I'm not quite sure how to clean out:
```
git clone git@github.com:ESCOMP/CTSM.git ctsm-test-cdeps
cd ctsm-test-cdeps/
git remote add ckoven_repo git@github.com:ckoven/CTSM.git
git fetch ckoven_repo
./manage_externals/checkout_externals
git checkout -b snow_occlusion_ctsm ckoven_repo/snow_occlusion_ctsm
./manage_externals/checkout_externals
git checkout master
./manage_externals/checkout_externals
```
generates this message:
```
cheyenne4 rgknox/ctsm-test-cdeps> ./manage_externals/checkout_externals
Processing externals description file : Externals.cfg
Processing externals description file : Externals_CLM.cfg
Processing externals description file : Externals_CISM.cfg
Processing externals description file : .gitmodules
Processing submodules description file : .gitmodules
Processing externals description file : .gitmodules
Processing submodules description file : .gitmodules
Checking status of externals: clm, fates, ptclm, mosart, mizuroute, cime, rtm, cism, source_cism, cdeps, fox, cmeps, nems/lib/parallelio, nems/lib/genf90, doc-builder,
s ./cime
s ./cime/src/drivers/nuopc/
./cime/src/drivers/nuopc/nems/lib/ParallelIO
./cime/src/drivers/nuopc/nems/lib/genf90
sM ./components/cdeps
./components/cdeps/fox
s ./components/cism
./components/cism/source_cism
./components/mizuRoute
s ./components/mosart
s ./components/rtm
e-o ./doc/doc-builder
./src/fates
./tools/PTCLM
----------------------------------------------------------------------
The external repositories labeled with 'M' above are not in a clean state.
The following are three options for how to proceed:
(1) Go into each external that is not in a clean state and issue either a 'git status' or
an 'svn status' command (depending on whether the external is managed by git or
svn). Either revert or commit your changes so that all externals are in a clean
state. (To revert changes in git, follow the instructions given when you run 'git
status'.) (Note, though, that it is okay to have untracked files in your working
directory.) Then rerun checkout_externals.
(2) Alternatively, you do not have to rely on checkout_externals. Instead, you can manually
update out-of-sync externals (labeled with 's' above) as described in the
configuration file Externals.cfg. (For example, run 'git fetch' and 'git checkout'
commands to checkout the appropriate tags for each external, as given in
Externals.cfg.)
(3) You can also use checkout_externals to manage most, but not all externals: You can specify
one or more externals to ignore using the '-x' or '--exclude' argument to
checkout_externals. Excluding externals labeled with 'M' will allow checkout_externals to
update the other, non-excluded externals.
The external repositories labeled with '?' above are not under version
control using the expected protocol. If you are sure you want to switch
protocols, and you don't have any work you need to save from this
directory, then run "rm -rf [directory]" before re-running the
checkout_externals tool.
----------------------------------------------------------------------
```
Here is the status of cdeps:
```
cheyenne4 components/cdeps> git status
HEAD detached at 45b7a85
Changes not staged for commit:
(use "git add <file>..." to update what will be committed)
(use "git checkout -- <file>..." to discard changes in working directory)
modified: fox (new commits)
Untracked files:
(use "git add <file>..." to include in what will be committed)
share/genf90/
no changes added to commit (use "git add" and/or "git commit -a")
```
I step into fox and here is the status:
```
cheyenne4 cdeps/fox> git status
HEAD detached at 4ff17b4
nothing to commit, working tree clean
```
| priority | external munging with cdeps and fox brief summary of bug the following generates an unclean state in components cdeps that i m not quite sure how to clean out git clone git github com escomp ctsm git ctsm test cdeps cd ctsm test cdeps git remote add ckoven repo git github com ckoven ctsm git git fetch ckoven repo manage externals checkout externals git checkout b snow occlusion ctsm ckoven repo snow occlusion ctsm manage externals checkout externals git checkout master manage externals checkout externals generates this message rgknox ctsm test cdeps manage externals checkout externals processing externals description file externals cfg processing externals description file externals clm cfg processing externals description file externals cism cfg processing externals description file gitmodules processing submodules description file gitmodules processing externals description file gitmodules processing submodules description file gitmodules checking status of externals clm fates ptclm mosart mizuroute cime rtm cism source cism cdeps fox cmeps nems lib parallelio nems lib doc builder s cime s cime src drivers nuopc cime src drivers nuopc nems lib parallelio cime src drivers nuopc nems lib sm components cdeps components cdeps fox s components cism components cism source cism components mizuroute s components mosart s components rtm e o doc doc builder src fates tools ptclm the external repositories labeled with m above are not in a clean state the following are three options for how to proceed go into each external that is not in a clean state and issue either a git status or an svn status command depending on whether the external is managed by git or svn either revert or commit your changes so that all externals are in a clean state to revert changes in git follow the instructions given when you run git status note though that it is okay to have untracked files in your working directory then rerun checkout externals alternatively you do not have to rely on checkout externals instead you can manually update out of sync externals labeled with s above as described in the configuration file externals cfg for example run git fetch and git checkout commands to checkout the appropriate tags for each external as given in externals cfg you can also use checkout externals to manage most but not all externals you can specify one or more externals to ignore using the x or exclude argument to checkout externals excluding externals labeled with m will allow checkout externals to update the other non excluded externals the external repositories labeled with above are not under version control using the expected protocol if you are sure you want to switch protocols and you don t have any work you need to save from this directory then run rm rf before re running the checkout externals tool here is the status of cdeps components cdeps git status head detached at changes not staged for commit use git add to update what will be committed use git checkout to discard changes in working directory modified fox new commits untracked files use git add to include in what will be committed share no changes added to commit use git add and or git commit a i step into fox and here is the status cdeps fox git status head detached at nothing to commit working tree clean | 1 |
135,879 | 5,266,289,552 | IssuesEvent | 2017-02-04 10:46:23 | japanesemediamanager/ShokoServer | https://api.github.com/repos/japanesemediamanager/ShokoServer | closed | multiple db entries linking to same file | Bug - Low Priority Most Likely Fixed - Need Confirmation | I have file renaming, but it's failing sometimes.
1.Is there a way to avoid creating multiple database entries which link to same file? Check before overwriting a file?
2.Where does CRC come from? Anidb? Can it be calculated locally when missing?
JMM Desktop and Server 3.6.1.0

| 1.0 | multiple db entries linking to same file - I have file renaming, but it's failing sometimes.
1.Is there a way to avoid creating multiple database entries which link to same file? Check before overwriting a file?
2.Where does CRC come from? Anidb? Can it be calculated locally when missing?
JMM Desktop and Server 3.6.1.0

| priority | multiple db entries linking to same file i have file renaming but it s failing sometimes is there a way to avoid creating multiple database entries which link to same file check before overwriting a file where does crc come from anidb can it be calculated locally when missing jmm desktop and server | 1 |
628,680 | 20,010,627,887 | IssuesEvent | 2022-02-01 05:42:57 | geoff-maddock/events-tracker | https://api.github.com/repos/geoff-maddock/events-tracker | closed | Design - visual flourish | style low priority | Look at codepen.io for some ideas of how to build svg or other icons or animations for the site | 1.0 | Design - visual flourish - Look at codepen.io for some ideas of how to build svg or other icons or animations for the site | priority | design visual flourish look at codepen io for some ideas of how to build svg or other icons or animations for the site | 1 |
634,652 | 20,368,274,501 | IssuesEvent | 2022-02-21 08:43:17 | MeAmAnUsername/pie | https://api.github.com/repos/MeAmAnUsername/pie | opened | Enable renaming | Type: enhancement Status: specified Priority: low Component: editor | **Summary**
Enable renaming
**Todo**
- [ ] Figure out how it interacts with imports
- [ ] Implement
**Reason**
Renaming is a fairly common refactoring, and Spoofax has a language parameteric renaming which hopfully allows us to include a few strategies and set up a menu item and then it works
**Detailed info**
Not entirely sure how this interacts with imports. Right now references reference the import first, which then references the actual declaration. #248 proposes to skip the reference to the import, which might break renaming entirely.
**Related issues**
- #248 proposes to change reference resolution for imports, which likely affects this feature | 1.0 | Enable renaming - **Summary**
Enable renaming
**Todo**
- [ ] Figure out how it interacts with imports
- [ ] Implement
**Reason**
Renaming is a fairly common refactoring, and Spoofax has a language parameteric renaming which hopfully allows us to include a few strategies and set up a menu item and then it works
**Detailed info**
Not entirely sure how this interacts with imports. Right now references reference the import first, which then references the actual declaration. #248 proposes to skip the reference to the import, which might break renaming entirely.
**Related issues**
- #248 proposes to change reference resolution for imports, which likely affects this feature | priority | enable renaming summary enable renaming todo figure out how it interacts with imports implement reason renaming is a fairly common refactoring and spoofax has a language parameteric renaming which hopfully allows us to include a few strategies and set up a menu item and then it works detailed info not entirely sure how this interacts with imports right now references reference the import first which then references the actual declaration proposes to skip the reference to the import which might break renaming entirely related issues proposes to change reference resolution for imports which likely affects this feature | 1 |
130,492 | 5,117,100,167 | IssuesEvent | 2017-01-07 13:00:11 | RevivalPMMP/PureEntitiesX | https://api.github.com/repos/RevivalPMMP/PureEntitiesX | closed | What's wrong? | Category: Bug Priority: Low Status: Confirmed | Warning: Declaration of revivalpmmp\pureentities\entity\BaseEntity::entityBaseTick($tickDiff = 1) should be compatible with pocketmine\entity\Living::entityBaseTick($tickDiff = 1, $EnchantL = 0) in C:\Users\UserX\Desktop\PM-MP\plugins\PureEntitiesX-1.0.0-bug-fixes\src\revivalpmmp\pureentities\entity\BaseEntity.php on line 0 | 1.0 | What's wrong? - Warning: Declaration of revivalpmmp\pureentities\entity\BaseEntity::entityBaseTick($tickDiff = 1) should be compatible with pocketmine\entity\Living::entityBaseTick($tickDiff = 1, $EnchantL = 0) in C:\Users\UserX\Desktop\PM-MP\plugins\PureEntitiesX-1.0.0-bug-fixes\src\revivalpmmp\pureentities\entity\BaseEntity.php on line 0 | priority | what s wrong? warning declaration of revivalpmmp pureentities entity baseentity entitybasetick tickdiff should be compatible with pocketmine entity living entitybasetick tickdiff enchantl in c users userx desktop pm mp plugins pureentitiesx bug fixes src revivalpmmp pureentities entity baseentity php on line | 1 |
405,993 | 11,885,812,326 | IssuesEvent | 2020-03-27 20:25:52 | Apexal/late | https://api.github.com/repos/Apexal/late | opened | Bad wording when there are no current polls | Area: Front End Difficulty: Trivial Priority: Low | **Describe the bug**
When there are no current polls, the polls tab has text "There no current polls"
**To Reproduce**
Steps to reproduce the behavior:
*Example:*
1. Go to 'Go to the polls tab when there are no polls'
**Expected behavior**
This could either say "No current polls" or "There are no current polls"
**Screenshots** *optional*
<img width="404" alt="Screen Shot 2020-03-27 at 4 21 43 PM" src="https://user-images.githubusercontent.com/18222508/77797455-7835c480-7047-11ea-81b5-22e86e7ea689.png">
| 1.0 | Bad wording when there are no current polls - **Describe the bug**
When there are no current polls, the polls tab has text "There no current polls"
**To Reproduce**
Steps to reproduce the behavior:
*Example:*
1. Go to 'Go to the polls tab when there are no polls'
**Expected behavior**
This could either say "No current polls" or "There are no current polls"
**Screenshots** *optional*
<img width="404" alt="Screen Shot 2020-03-27 at 4 21 43 PM" src="https://user-images.githubusercontent.com/18222508/77797455-7835c480-7047-11ea-81b5-22e86e7ea689.png">
| priority | bad wording when there are no current polls describe the bug when there are no current polls the polls tab has text there no current polls to reproduce steps to reproduce the behavior example go to go to the polls tab when there are no polls expected behavior this could either say no current polls or there are no current polls screenshots optional img width alt screen shot at pm src | 1 |
596,436 | 18,104,283,600 | IssuesEvent | 2021-09-22 17:23:33 | NOAA-GSL/VxLegacyIngest | https://api.github.com/repos/NOAA-GSL/VxLegacyIngest | closed | Problem with Surface Verification Retro | Status: Doing Type: Bug Priority: Low | ---
Author Name: **jeffrey.a.hamilton** (jeffrey.a.hamilton)
Original Redmine Issue: 22901, https://vlab.ncep.noaa.gov/redmine/issues/22901
Original Date: 2016-10-06
Original Assignee: jeffrey.a.hamilton
---
Eric has noticed a problem with one of his Surface verification retro runs. Investigate and correct once the source of the problem has been found.
| 1.0 | Problem with Surface Verification Retro - ---
Author Name: **jeffrey.a.hamilton** (jeffrey.a.hamilton)
Original Redmine Issue: 22901, https://vlab.ncep.noaa.gov/redmine/issues/22901
Original Date: 2016-10-06
Original Assignee: jeffrey.a.hamilton
---
Eric has noticed a problem with one of his Surface verification retro runs. Investigate and correct once the source of the problem has been found.
| priority | problem with surface verification retro author name jeffrey a hamilton jeffrey a hamilton original redmine issue original date original assignee jeffrey a hamilton eric has noticed a problem with one of his surface verification retro runs investigate and correct once the source of the problem has been found | 1 |
515,702 | 14,967,646,142 | IssuesEvent | 2021-01-27 15:55:04 | kobotoolbox/kpi | https://api.github.com/repos/kobotoolbox/kpi | opened | ENTER doesn't always work in the New Library Item and New Project modals | bug frontend low priority | ## Description
On the first time you use it, ENTER key submits the modal form. It doesn't work again until you reload page. | 1.0 | ENTER doesn't always work in the New Library Item and New Project modals - ## Description
On the first time you use it, ENTER key submits the modal form. It doesn't work again until you reload page. | priority | enter doesn t always work in the new library item and new project modals description on the first time you use it enter key submits the modal form it doesn t work again until you reload page | 1 |
475,651 | 13,724,269,745 | IssuesEvent | 2020-10-03 13:30:48 | rism-ch/verovio | https://api.github.com/repos/rism-ch/verovio | opened | Tie + augmentation dot + system break interaction | enhancement low priority | When a tie crosses a system break, the tie start will collide with augmentation dots on the starting note for the tie. In other words, the dot-avoidance code for ties is not being called when the tie goes over a system break:
<img width="378" alt="Screen Shot 2020-10-03 at 6 21 40 AM" src="https://user-images.githubusercontent.com/3487289/94992777-10ee5500-0541-11eb-88e6-ea24302a5b0c.png">
MEI test data:
```xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://relaxng.org/ns/structure/1.0"?>
<?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://purl.oclc.org/dsdl/schematron"?>
<mei xmlns="http://www.music-encoding.org/ns/mei" meiversion="4.0.0">
<meiHead>
<fileDesc>
<titleStmt>
<title />
</titleStmt>
<pubStmt />
</fileDesc>
<encodingDesc>
<appInfo>
<application isodate="2020-10-03T06:24:27" version="3.0.0-dev-f0996ca-dirty">
<name>Verovio</name>
<p>Transcoded from Humdrum</p>
</application>
</appInfo>
</encodingDesc>
<workList>
<work>
<title />
</work>
</workList>
</meiHead>
<music>
<body>
<mdiv xml:id="mdiv-0000001384622065">
<score xml:id="score-0000001262513604">
<scoreDef xml:id="scoredef-0000000830027336" midi.bpm="400">
<staffGrp xml:id="staffgrp-0000000003742673">
<staffDef xml:id="staffdef-0000002084676771" n="1" lines="5">
<clef xml:id="clef-0000000718382857" shape="G" line="2" />
</staffDef>
</staffGrp>
</scoreDef>
<section xml:id="section-L1F1">
<measure xml:id="measure-L1" n="1">
<staff xml:id="staff-0000001596716717" n="1">
<layer xml:id="layer-L1F1N1" n="1">
<note xml:id="note-L3F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L3F1-L5F1" startid="#note-L3F1" endid="#note-L5F1" />
</measure>
<measure xml:id="measure-L4" n="2">
<staff xml:id="staff-L4F1N1" n="1">
<layer xml:id="layer-L4F1N1" n="1">
<note xml:id="note-L5F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L5F1-L7F1" startid="#note-L5F1" endid="#note-L7F1" />
</measure>
<measure xml:id="measure-L6" n="3">
<staff xml:id="staff-L6F1N1" n="1">
<layer xml:id="layer-L6F1N1" n="1">
<note xml:id="note-L7F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L7F1-L9F1" startid="#note-L7F1" endid="#note-L9F1" />
</measure>
<measure xml:id="measure-L8" n="4">
<staff xml:id="staff-L8F1N1" n="1">
<layer xml:id="layer-L8F1N1" n="1">
<note xml:id="note-L9F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L9F1-L11F1" startid="#note-L9F1" endid="#note-L11F1" />
</measure>
<measure xml:id="measure-L10" n="5">
<staff xml:id="staff-L10F1N1" n="1">
<layer xml:id="layer-L10F1N1" n="1">
<note xml:id="note-L11F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L11F1-L13F1" startid="#note-L11F1" endid="#note-L13F1" />
</measure>
<measure xml:id="measure-L12" n="6">
<staff xml:id="staff-L12F1N1" n="1">
<layer xml:id="layer-L12F1N1" n="1">
<note xml:id="note-L13F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L13F1-L15F1" startid="#note-L13F1" endid="#note-L15F1" />
</measure>
<measure xml:id="measure-L14" n="7">
<staff xml:id="staff-L14F1N1" n="1">
<layer xml:id="layer-L14F1N1" n="1">
<note xml:id="note-L15F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L15F1-L17F1" startid="#note-L15F1" endid="#note-L17F1" />
</measure>
<measure xml:id="measure-L16" n="8">
<staff xml:id="staff-L16F1N1" n="1">
<layer xml:id="layer-L16F1N1" n="1">
<note xml:id="note-L17F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L17F1-L19F1" startid="#note-L17F1" endid="#note-L19F1" />
</measure>
<measure xml:id="measure-L18" right="end" n="9">
<staff xml:id="staff-L18F1N1" n="1">
<layer xml:id="layer-L18F1N1" n="1">
<note xml:id="note-L19F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
</measure>
</section>
</score>
</mdiv>
</body>
</music>
</mei>
```
| 1.0 | Tie + augmentation dot + system break interaction - When a tie crosses a system break, the tie start will collide with augmentation dots on the starting note for the tie. In other words, the dot-avoidance code for ties is not being called when the tie goes over a system break:
<img width="378" alt="Screen Shot 2020-10-03 at 6 21 40 AM" src="https://user-images.githubusercontent.com/3487289/94992777-10ee5500-0541-11eb-88e6-ea24302a5b0c.png">
MEI test data:
```xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://relaxng.org/ns/structure/1.0"?>
<?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://purl.oclc.org/dsdl/schematron"?>
<mei xmlns="http://www.music-encoding.org/ns/mei" meiversion="4.0.0">
<meiHead>
<fileDesc>
<titleStmt>
<title />
</titleStmt>
<pubStmt />
</fileDesc>
<encodingDesc>
<appInfo>
<application isodate="2020-10-03T06:24:27" version="3.0.0-dev-f0996ca-dirty">
<name>Verovio</name>
<p>Transcoded from Humdrum</p>
</application>
</appInfo>
</encodingDesc>
<workList>
<work>
<title />
</work>
</workList>
</meiHead>
<music>
<body>
<mdiv xml:id="mdiv-0000001384622065">
<score xml:id="score-0000001262513604">
<scoreDef xml:id="scoredef-0000000830027336" midi.bpm="400">
<staffGrp xml:id="staffgrp-0000000003742673">
<staffDef xml:id="staffdef-0000002084676771" n="1" lines="5">
<clef xml:id="clef-0000000718382857" shape="G" line="2" />
</staffDef>
</staffGrp>
</scoreDef>
<section xml:id="section-L1F1">
<measure xml:id="measure-L1" n="1">
<staff xml:id="staff-0000001596716717" n="1">
<layer xml:id="layer-L1F1N1" n="1">
<note xml:id="note-L3F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L3F1-L5F1" startid="#note-L3F1" endid="#note-L5F1" />
</measure>
<measure xml:id="measure-L4" n="2">
<staff xml:id="staff-L4F1N1" n="1">
<layer xml:id="layer-L4F1N1" n="1">
<note xml:id="note-L5F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L5F1-L7F1" startid="#note-L5F1" endid="#note-L7F1" />
</measure>
<measure xml:id="measure-L6" n="3">
<staff xml:id="staff-L6F1N1" n="1">
<layer xml:id="layer-L6F1N1" n="1">
<note xml:id="note-L7F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L7F1-L9F1" startid="#note-L7F1" endid="#note-L9F1" />
</measure>
<measure xml:id="measure-L8" n="4">
<staff xml:id="staff-L8F1N1" n="1">
<layer xml:id="layer-L8F1N1" n="1">
<note xml:id="note-L9F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L9F1-L11F1" startid="#note-L9F1" endid="#note-L11F1" />
</measure>
<measure xml:id="measure-L10" n="5">
<staff xml:id="staff-L10F1N1" n="1">
<layer xml:id="layer-L10F1N1" n="1">
<note xml:id="note-L11F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L11F1-L13F1" startid="#note-L11F1" endid="#note-L13F1" />
</measure>
<measure xml:id="measure-L12" n="6">
<staff xml:id="staff-L12F1N1" n="1">
<layer xml:id="layer-L12F1N1" n="1">
<note xml:id="note-L13F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L13F1-L15F1" startid="#note-L13F1" endid="#note-L15F1" />
</measure>
<measure xml:id="measure-L14" n="7">
<staff xml:id="staff-L14F1N1" n="1">
<layer xml:id="layer-L14F1N1" n="1">
<note xml:id="note-L15F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L15F1-L17F1" startid="#note-L15F1" endid="#note-L17F1" />
</measure>
<measure xml:id="measure-L16" n="8">
<staff xml:id="staff-L16F1N1" n="1">
<layer xml:id="layer-L16F1N1" n="1">
<note xml:id="note-L17F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
<tie xml:id="tie-L17F1-L19F1" startid="#note-L17F1" endid="#note-L19F1" />
</measure>
<measure xml:id="measure-L18" right="end" n="9">
<staff xml:id="staff-L18F1N1" n="1">
<layer xml:id="layer-L18F1N1" n="1">
<note xml:id="note-L19F1" dots="1" dur="2" oct="5" pname="f" accid.ges="n" />
</layer>
</staff>
</measure>
</section>
</score>
</mdiv>
</body>
</music>
</mei>
```
| priority | tie augmentation dot system break interaction when a tie crosses a system break the tie start will collide with augmentation dots on the starting note for the tie in other words the dot avoidance code for ties is not being called when the tie goes over a system break img width alt screen shot at am src mei test data xml xml model href type application xml schematypens xml model href type application xml schematypens verovio transcoded from humdrum | 1 |
754,690 | 26,398,513,125 | IssuesEvent | 2023-01-12 21:56:45 | GamerSafer/ABBA | https://api.github.com/repos/GamerSafer/ABBA | closed | Migrate to a command framework. | enhancement low priority | A command framework like [cloud](https://github.com/Incendo/cloud) would make things like command suggestions easier.
Blocks #21 | 1.0 | Migrate to a command framework. - A command framework like [cloud](https://github.com/Incendo/cloud) would make things like command suggestions easier.
Blocks #21 | priority | migrate to a command framework a command framework like would make things like command suggestions easier blocks | 1 |
422,189 | 12,267,740,300 | IssuesEvent | 2020-05-07 11:14:50 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | opened | JS prompt doesn't accept input when launched from select element | bug: behavior component: QtWebEngine component: keyinput priority: 2 - low | With this HTML snippet:
```html
<select onchange="prompt('test')">
<option>one</option>
<option>two</option>
</select>
```
the prompt input doesn't work (only enter/esc does). Logs:
```
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: False --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keyrelease:293 filter: False
13:13:35 DEBUG mouse eventfilter:eventFilter:302 Ignoring QKeyEvent to <PyQt5.QtWidgets.QWidget object at 0x7f615c614280>
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: False --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keyrelease:293 filter: False
```
Not sure what qutebrowser can do here - might be something Qt needs to fix. | 1.0 | JS prompt doesn't accept input when launched from select element - With this HTML snippet:
```html
<select onchange="prompt('test')">
<option>one</option>
<option>two</option>
</select>
```
the prompt input doesn't work (only enter/esc does). Logs:
```
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: False --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keyrelease:293 filter: False
13:13:35 DEBUG mouse eventfilter:eventFilter:302 Ignoring QKeyEvent to <PyQt5.QtWidgets.QWidget object at 0x7f615c614280>
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: True --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keypress:245 got keypress in mode KeyMode.prompt - delegating to <qutebrowser.keyinput.modeparsers.PassthroughKeyParser mode=<KeyMode.prompt: 5>>
13:13:35 DEBUG modes modeman:_handle_keypress:268 match: 0, forward_unbound_keys: auto, passthrough: True, is_non_alnum: False, dry_run: False --> filter: False (focused: <qutebrowser.mainwindow.prompt.LineEdit>)
13:13:35 DEBUG modes modeman:_handle_keyrelease:293 filter: False
```
Not sure what qutebrowser can do here - might be something Qt needs to fix. | priority | js prompt doesn t accept input when launched from select element with this html snippet html one two the prompt input doesn t work only enter esc does logs debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run true filter false focused debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run true filter false focused debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run false filter false focused debug modes modeman handle keyrelease filter false debug mouse eventfilter eventfilter ignoring qkeyevent to debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run true filter false focused debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run true filter false focused debug modes modeman handle keypress got keypress in mode keymode prompt delegating to debug modes modeman handle keypress match forward unbound keys auto passthrough true is non alnum false dry run false filter false focused debug modes modeman handle keyrelease filter false not sure what qutebrowser can do here might be something qt needs to fix | 1 |
108,801 | 4,350,780,463 | IssuesEvent | 2016-07-31 13:38:13 | lampepfl/dotty | https://api.github.com/repos/lampepfl/dotty | closed | Singleton types should not be inferred for final vars | Fix available In progress Low hanging fruit Low Priority Typer | ```scala
class Test {
final var x = false
x = true
}
```
```scala
error: type mismatch:
found : Boolean(true)
required: Boolean(false)
x = true
^
``` | 1.0 | Singleton types should not be inferred for final vars - ```scala
class Test {
final var x = false
x = true
}
```
```scala
error: type mismatch:
found : Boolean(true)
required: Boolean(false)
x = true
^
``` | priority | singleton types should not be inferred for final vars scala class test final var x false x true scala error type mismatch found boolean true required boolean false x true | 1 |
53,665 | 3,043,670,474 | IssuesEvent | 2015-08-10 00:10:26 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Add 'global' keyword to selections | Component-Selections enhancement Priority-Low | When performing atom selections from an AtomGroup instance all subexpressions are always evaluated within the set of the atoms of that instance. However, that isn't always intuitive:
```python
complex_preselected_solvent_group.selectAtoms("around 10 protein")
# This returns an empty AtomGroup because there is
# no protein in complex_preselected_solvent_group
# It can always be worked around...
universe.selectAtoms("group solv and around 10 protein", solv=complex_preselected_solvent_group)
```
A `global`keyword could be added to make the subsequent subexpressions be evaluated in the universe context, instead of the calling AtomGroup instance:
```python
complex_preselected_solvent_group.selectAtoms("around 10 global protein")
# This would work as expected
```
This construct would also eliminate the need for the `fullgroup` keyword, as we could now use `global group` instead, with equal legibility.
It also makes more sense now that we have the `same ... as` construct:
```python
# Suppose we have multiple protein copies in our system and want the active-site residues
# of those copies that are close to a POPC membrane
active_site_residues.selectAtoms("same fragment as global (protein and around 10 resname POPC)")
```
This is not a serious problem, since it can always be worked around by issuing the selection from the universe object, and using the `group` keyword to bring in the AtomGroup's atoms. But it makes for more readable code, and easier to write.
Opinions? | 1.0 | Add 'global' keyword to selections - When performing atom selections from an AtomGroup instance all subexpressions are always evaluated within the set of the atoms of that instance. However, that isn't always intuitive:
```python
complex_preselected_solvent_group.selectAtoms("around 10 protein")
# This returns an empty AtomGroup because there is
# no protein in complex_preselected_solvent_group
# It can always be worked around...
universe.selectAtoms("group solv and around 10 protein", solv=complex_preselected_solvent_group)
```
A `global`keyword could be added to make the subsequent subexpressions be evaluated in the universe context, instead of the calling AtomGroup instance:
```python
complex_preselected_solvent_group.selectAtoms("around 10 global protein")
# This would work as expected
```
This construct would also eliminate the need for the `fullgroup` keyword, as we could now use `global group` instead, with equal legibility.
It also makes more sense now that we have the `same ... as` construct:
```python
# Suppose we have multiple protein copies in our system and want the active-site residues
# of those copies that are close to a POPC membrane
active_site_residues.selectAtoms("same fragment as global (protein and around 10 resname POPC)")
```
This is not a serious problem, since it can always be worked around by issuing the selection from the universe object, and using the `group` keyword to bring in the AtomGroup's atoms. But it makes for more readable code, and easier to write.
Opinions? | priority | add global keyword to selections when performing atom selections from an atomgroup instance all subexpressions are always evaluated within the set of the atoms of that instance however that isn t always intuitive python complex preselected solvent group selectatoms around protein this returns an empty atomgroup because there is no protein in complex preselected solvent group it can always be worked around universe selectatoms group solv and around protein solv complex preselected solvent group a global keyword could be added to make the subsequent subexpressions be evaluated in the universe context instead of the calling atomgroup instance python complex preselected solvent group selectatoms around global protein this would work as expected this construct would also eliminate the need for the fullgroup keyword as we could now use global group instead with equal legibility it also makes more sense now that we have the same as construct python suppose we have multiple protein copies in our system and want the active site residues of those copies that are close to a popc membrane active site residues selectatoms same fragment as global protein and around resname popc this is not a serious problem since it can always be worked around by issuing the selection from the universe object and using the group keyword to bring in the atomgroup s atoms but it makes for more readable code and easier to write opinions | 1 |
15,752 | 2,611,514,221 | IssuesEvent | 2015-02-27 05:50:01 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Greater weapon customization | auto-migrated Priority-Low Type-Enhancement | ```
What steps will reproduce the problem?
1. Only 4 types of customization: Ammo, power, crate, and delay
2. Can't change Deagle ammo type
3. Can't make kamikaze controllable
Enchant hedge-wars
What is the expected output? What do you see instead?
More customization
Like, Worms 2-style customization
What version of the product are you using? On what operating system?
0.9.18
Windows
Please provide any additional information below.
Allow more customization, like worms 2-style customization
After all, it's freely customizable in files, so why not allow more
customization for weapons?
```
Original issue reported on code.google.com by `Openwor...@gmail.com` on 2 Jan 2013 at 8:01 | 1.0 | Greater weapon customization - ```
What steps will reproduce the problem?
1. Only 4 types of customization: Ammo, power, crate, and delay
2. Can't change Deagle ammo type
3. Can't make kamikaze controllable
Enchant hedge-wars
What is the expected output? What do you see instead?
More customization
Like, Worms 2-style customization
What version of the product are you using? On what operating system?
0.9.18
Windows
Please provide any additional information below.
Allow more customization, like worms 2-style customization
After all, it's freely customizable in files, so why not allow more
customization for weapons?
```
Original issue reported on code.google.com by `Openwor...@gmail.com` on 2 Jan 2013 at 8:01 | priority | greater weapon customization what steps will reproduce the problem only types of customization ammo power crate and delay can t change deagle ammo type can t make kamikaze controllable enchant hedge wars what is the expected output what do you see instead more customization like worms style customization what version of the product are you using on what operating system windows please provide any additional information below allow more customization like worms style customization after all it s freely customizable in files so why not allow more customization for weapons original issue reported on code google com by openwor gmail com on jan at | 1 |
21,964 | 2,643,595,633 | IssuesEvent | 2015-03-12 12:12:33 | Araq/Nim | https://api.github.com/repos/Araq/Nim | closed | Strange issues with `$`[tuple|object] when using it on a TTable without tables being imported | Low Priority | Consider two files:
File foo.nim:
```nimrod
import tables, strtabs, asyncio
type
TRequest* = object
formData*: TTable[string, tuple[fields: PStringTable, body: string]]
proc test*(): TRequest =
var x = TRequest()
x.formData = initTable[string, tuple[fields: PStringTable, body: string]]()
x.formData["asd"] = (newStringTable(), "asdas")
result = x
```
and
```nimrod
import foo
var x = test()
echo($x.formData)
```
This results in an odd error:
```
a21.nim(4, 6) Info: instantiation from here
lib/system.nim(1605, 21) Error: undeclared field: 'data'
```
Importing ``tables`` inside the second file stops the error from happening. | 1.0 | Strange issues with `$`[tuple|object] when using it on a TTable without tables being imported - Consider two files:
File foo.nim:
```nimrod
import tables, strtabs, asyncio
type
TRequest* = object
formData*: TTable[string, tuple[fields: PStringTable, body: string]]
proc test*(): TRequest =
var x = TRequest()
x.formData = initTable[string, tuple[fields: PStringTable, body: string]]()
x.formData["asd"] = (newStringTable(), "asdas")
result = x
```
and
```nimrod
import foo
var x = test()
echo($x.formData)
```
This results in an odd error:
```
a21.nim(4, 6) Info: instantiation from here
lib/system.nim(1605, 21) Error: undeclared field: 'data'
```
Importing ``tables`` inside the second file stops the error from happening. | priority | strange issues with when using it on a ttable without tables being imported consider two files file foo nim nimrod import tables strtabs asyncio type trequest object formdata ttable proc test trequest var x trequest x formdata inittable x formdata newstringtable asdas result x and nimrod import foo var x test echo x formdata this results in an odd error nim info instantiation from here lib system nim error undeclared field data importing tables inside the second file stops the error from happening | 1 |
72,339 | 3,384,428,667 | IssuesEvent | 2015-11-27 02:17:10 | mctaggaj/SolidArc_AdminPortal | https://api.github.com/repos/mctaggaj/SolidArc_AdminPortal | closed | Edit Team Details has no functionality | Difficulty: Easy Priority: Low Type: Enhancement | The page is empty other than the title. Maybe add a text box where we can change the name or something. Put things we would change if it worked with the database.
Its not going to work, but we can just say its the database teams fault. | 1.0 | Edit Team Details has no functionality - The page is empty other than the title. Maybe add a text box where we can change the name or something. Put things we would change if it worked with the database.
Its not going to work, but we can just say its the database teams fault. | priority | edit team details has no functionality the page is empty other than the title maybe add a text box where we can change the name or something put things we would change if it worked with the database its not going to work but we can just say its the database teams fault | 1 |
22,063 | 2,644,956,286 | IssuesEvent | 2015-03-12 19:48:41 | acardona/CATMAID | https://api.github.com/repos/acardona/CATMAID | closed | Selection table: add way to hide only arbor | context: 3d-viewer priority: low type: enhancement | should hide whole actor, if shift pressed, only hide arbor.
does not hide soma spheres | 1.0 | Selection table: add way to hide only arbor - should hide whole actor, if shift pressed, only hide arbor.
does not hide soma spheres | priority | selection table add way to hide only arbor should hide whole actor if shift pressed only hide arbor does not hide soma spheres | 1 |
137,595 | 5,312,835,584 | IssuesEvent | 2017-02-13 10:17:53 | pmem/issues | https://api.github.com/repos/pmem/issues | opened | unit tests: obj_bucket/TEST0: SETUP (all/pmem/debug/pmemcheck) fails | Exposure: Low OS: Linux Priority: 4 low Type: Bug | Revision: 0fd509d73382160069b98525597976e20d98f1ea
> obj_bucket/TEST0: SETUP (all/pmem/debug/pmemcheck)
> obj_bucket/TEST0: START: obj_bucket
> obj_bucket/TEST0 crashed (signal 11). err0.log below.
> {ut_backtrace.c:203 ut_sighandler} obj_bucket/TEST0:
>
> {ut_backtrace.c:204 ut_sighandler} obj_bucket/TEST0: Signal 11, backtrace:
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 0: ./obj_bucket (ut_sighandler+0x52) [0x436915] [0x36915]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 1: /lib/x86_64-linux-gnu/libc.so.6 (killpg+0x40) [0x54aa4ef] [0x354ef]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 2: ./obj_bucket (bucket_insert_block+0x3b) [0x404558] [0x4558]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 3: ./obj_bucket (test_bucket_insert_get+0xfb) [0x40322c] [0x322c]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 4: ./obj_bucket (main+0x44) [0x4034fb] [0x34fb]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 5: /lib/x86_64-linux-gnu/libc.so.6 (__libc_start_main+0xf0) [0x5495830] [0x20830]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 6: ./obj_bucket (_start+0x29) [0x402f39] [0x2f39]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 7: ? (?+0x29) [0x29] [0x0]
> {ut_backtrace.c:206 ut_sighandler} obj_bucket/TEST0:
>
> pmemcheck0.log below.
> obj_bucket/TEST0 pmemcheck0.log ==27933== pmemcheck-0.2, a simple persistent store checker
> obj_bucket/TEST0 pmemcheck0.log ==27933== Copyright (c) 2014-2016, Intel Corporation
> obj_bucket/TEST0 pmemcheck0.log ==27933== Using Valgrind-3.10.1 and LibVEX; rerun with -h for copyright info
> obj_bucket/TEST0 pmemcheck0.log ==27933== Command: ./obj_bucket
> obj_bucket/TEST0 pmemcheck0.log ==27933== Parent PID: 27884
> obj_bucket/TEST0 pmemcheck0.log ==27933==
> obj_bucket/TEST0 pmemcheck0.log ==27933==
> obj_bucket/TEST0 pmemcheck0.log ==27933== Number of stores not made persistent: 0
> obj_bucket/TEST0 pmemcheck0.log ==27933== ERROR SUMMARY: 0 errors
>
> out0.log below.
> obj_bucket/TEST0 out0.log obj_bucket/TEST0: START: obj_bucket
> obj_bucket/TEST0 out0.log ./obj_bucket
>
> pmem0.log below.
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:244 out_init] pid 27933: program: /home/jenkins/workspace/nvml_fork_tests_valgrind_force_enable_ubuntu/src/test/obj_bucket/obj_bucket
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:246 out_init] libpmem version 1.0
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:247 out_init] src version SRCVERSION:1.2+wtp1-241-g0fd509d
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind pmemcheck
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind helgrind
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind memcheck
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:270 out_init] compiled with support for Valgrind drd
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [mmap.c:59 util_mmap_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [libpmem.c:56 libpmem_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1197 pmem_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1163 pmem_get_cpuinfo] clflush supported
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1141 pmem_log_cpuinfo] using clflush
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1148 pmem_log_cpuinfo] using movnt
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [libpmem.c:69 libpmem_fini]
>
> pmemobj0.log below.
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:244 out_init] pid 27933: program: /home/jenkins/workspace/nvml_fork_tests_valgrind_force_enable_ubuntu/src/test/obj_bucket/obj_bucket
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:246 out_init] libpmemobj version 2.0
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:247 out_init] src version SRCVERSION:1.2+wtp1-241-g0fd509d
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:255 out_init] compiled with support for Valgrind pmemcheck
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:260 out_init] compiled with support for Valgrind helgrind
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:265 out_init] compiled with support for Valgrind memcheck
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:270 out_init] compiled with support for Valgrind drd
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [mmap.c:59 util_mmap_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [libpmemobj.c:52 libpmemobj_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [obj.c:180 obj_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [set.c:95 util_remote_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [obj.c:209 obj_fini]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [set.c:107 util_remote_fini]
>
> RUNTESTS: stopping: obj_bucket//TEST0 failed, TEST=all FS=pmem BUILD=debug
> | 1.0 | unit tests: obj_bucket/TEST0: SETUP (all/pmem/debug/pmemcheck) fails - Revision: 0fd509d73382160069b98525597976e20d98f1ea
> obj_bucket/TEST0: SETUP (all/pmem/debug/pmemcheck)
> obj_bucket/TEST0: START: obj_bucket
> obj_bucket/TEST0 crashed (signal 11). err0.log below.
> {ut_backtrace.c:203 ut_sighandler} obj_bucket/TEST0:
>
> {ut_backtrace.c:204 ut_sighandler} obj_bucket/TEST0: Signal 11, backtrace:
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 0: ./obj_bucket (ut_sighandler+0x52) [0x436915] [0x36915]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 1: /lib/x86_64-linux-gnu/libc.so.6 (killpg+0x40) [0x54aa4ef] [0x354ef]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 2: ./obj_bucket (bucket_insert_block+0x3b) [0x404558] [0x4558]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 3: ./obj_bucket (test_bucket_insert_get+0xfb) [0x40322c] [0x322c]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 4: ./obj_bucket (main+0x44) [0x4034fb] [0x34fb]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 5: /lib/x86_64-linux-gnu/libc.so.6 (__libc_start_main+0xf0) [0x5495830] [0x20830]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 6: ./obj_bucket (_start+0x29) [0x402f39] [0x2f39]
> {ut_backtrace.c:120 ut_dump_backtrace} obj_bucket/TEST0: 7: ? (?+0x29) [0x29] [0x0]
> {ut_backtrace.c:206 ut_sighandler} obj_bucket/TEST0:
>
> pmemcheck0.log below.
> obj_bucket/TEST0 pmemcheck0.log ==27933== pmemcheck-0.2, a simple persistent store checker
> obj_bucket/TEST0 pmemcheck0.log ==27933== Copyright (c) 2014-2016, Intel Corporation
> obj_bucket/TEST0 pmemcheck0.log ==27933== Using Valgrind-3.10.1 and LibVEX; rerun with -h for copyright info
> obj_bucket/TEST0 pmemcheck0.log ==27933== Command: ./obj_bucket
> obj_bucket/TEST0 pmemcheck0.log ==27933== Parent PID: 27884
> obj_bucket/TEST0 pmemcheck0.log ==27933==
> obj_bucket/TEST0 pmemcheck0.log ==27933==
> obj_bucket/TEST0 pmemcheck0.log ==27933== Number of stores not made persistent: 0
> obj_bucket/TEST0 pmemcheck0.log ==27933== ERROR SUMMARY: 0 errors
>
> out0.log below.
> obj_bucket/TEST0 out0.log obj_bucket/TEST0: START: obj_bucket
> obj_bucket/TEST0 out0.log ./obj_bucket
>
> pmem0.log below.
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:244 out_init] pid 27933: program: /home/jenkins/workspace/nvml_fork_tests_valgrind_force_enable_ubuntu/src/test/obj_bucket/obj_bucket
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:246 out_init] libpmem version 1.0
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:247 out_init] src version SRCVERSION:1.2+wtp1-241-g0fd509d
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind pmemcheck
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind helgrind
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind memcheck
> obj_bucket/TEST0 pmem0.log <libpmem>: <1> [out.c:270 out_init] compiled with support for Valgrind drd
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [mmap.c:59 util_mmap_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [libpmem.c:56 libpmem_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1197 pmem_init]
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1163 pmem_get_cpuinfo] clflush supported
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1141 pmem_log_cpuinfo] using clflush
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [pmem.c:1148 pmem_log_cpuinfo] using movnt
> obj_bucket/TEST0 pmem0.log <libpmem>: <3> [libpmem.c:69 libpmem_fini]
>
> pmemobj0.log below.
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:244 out_init] pid 27933: program: /home/jenkins/workspace/nvml_fork_tests_valgrind_force_enable_ubuntu/src/test/obj_bucket/obj_bucket
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:246 out_init] libpmemobj version 2.0
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:247 out_init] src version SRCVERSION:1.2+wtp1-241-g0fd509d
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:255 out_init] compiled with support for Valgrind pmemcheck
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:260 out_init] compiled with support for Valgrind helgrind
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:265 out_init] compiled with support for Valgrind memcheck
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <1> [out.c:270 out_init] compiled with support for Valgrind drd
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [mmap.c:59 util_mmap_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [libpmemobj.c:52 libpmemobj_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [obj.c:180 obj_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [set.c:95 util_remote_init]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [libpmemobj.c:65 libpmemobj_fini]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [obj.c:209 obj_fini]
> obj_bucket/TEST0 pmemobj0.log <libpmemobj>: <3> [set.c:107 util_remote_fini]
>
> RUNTESTS: stopping: obj_bucket//TEST0 failed, TEST=all FS=pmem BUILD=debug
> | priority | unit tests obj bucket setup all pmem debug pmemcheck fails revision obj bucket setup all pmem debug pmemcheck obj bucket start obj bucket obj bucket crashed signal log below ut backtrace c ut sighandler obj bucket ut backtrace c ut sighandler obj bucket signal backtrace ut backtrace c ut dump backtrace obj bucket obj bucket ut sighandler ut backtrace c ut dump backtrace obj bucket lib linux gnu libc so killpg ut backtrace c ut dump backtrace obj bucket obj bucket bucket insert block ut backtrace c ut dump backtrace obj bucket obj bucket test bucket insert get ut backtrace c ut dump backtrace obj bucket obj bucket main ut backtrace c ut dump backtrace obj bucket lib linux gnu libc so libc start main ut backtrace c ut dump backtrace obj bucket obj bucket start ut backtrace c ut dump backtrace obj bucket ut backtrace c ut sighandler obj bucket log below obj bucket log pmemcheck a simple persistent store checker obj bucket log copyright c intel corporation obj bucket log using valgrind and libvex rerun with h for copyright info obj bucket log command obj bucket obj bucket log parent pid obj bucket log obj bucket log obj bucket log number of stores not made persistent obj bucket log error summary errors log below obj bucket log obj bucket start obj bucket obj bucket log obj bucket log below obj bucket log pid program home jenkins workspace nvml fork tests valgrind force enable ubuntu src test obj bucket obj bucket obj bucket log libpmem version obj bucket log src version srcversion obj bucket log compiled with support for valgrind pmemcheck obj bucket log compiled with support for valgrind helgrind obj bucket log compiled with support for valgrind memcheck obj bucket log compiled with support for valgrind drd obj bucket log obj bucket log obj bucket log obj bucket log clflush supported obj bucket log using clflush obj bucket log using movnt obj bucket log log below obj bucket log pid program home jenkins workspace nvml fork tests valgrind force enable ubuntu src test obj bucket obj bucket obj bucket log libpmemobj version obj bucket log src version srcversion obj bucket log compiled with support for valgrind pmemcheck obj bucket log compiled with support for valgrind helgrind obj bucket log compiled with support for valgrind memcheck obj bucket log compiled with support for valgrind drd obj bucket log obj bucket log obj bucket log obj bucket log obj bucket log obj bucket log obj bucket log runtests stopping obj bucket failed test all fs pmem build debug | 1 |
197,716 | 6,963,191,618 | IssuesEvent | 2017-12-08 16:28:14 | Parabot/Parabot | https://api.github.com/repos/Parabot/Parabot | opened | Read version from pom.xml for Travis | priority:low status:accepted status:under consideration type:improvement | Currently we have to adjust both the pom.xml and the .travis.yml file, if a new version gets released.
Could we maybe put the pom.xml property into the environment configuration and read it with Travis? | 1.0 | Read version from pom.xml for Travis - Currently we have to adjust both the pom.xml and the .travis.yml file, if a new version gets released.
Could we maybe put the pom.xml property into the environment configuration and read it with Travis? | priority | read version from pom xml for travis currently we have to adjust both the pom xml and the travis yml file if a new version gets released could we maybe put the pom xml property into the environment configuration and read it with travis | 1 |
516,652 | 14,985,703,887 | IssuesEvent | 2021-01-28 20:17:42 | kymckay/f21as-project | https://api.github.com/repos/kymckay/f21as-project | closed | Set up JUnit CI | priority|low type|task | I'd like to get JUnit tests set up to run via GitHub Actions (will run any time pushes are made for easy test automation).
Seems like a good run through of one possible setup: https://dev.to/ewefie/getting-started-with-github-actions-run-junit-5-tests-in-a-java-project-with-maven-20g4 | 1.0 | Set up JUnit CI - I'd like to get JUnit tests set up to run via GitHub Actions (will run any time pushes are made for easy test automation).
Seems like a good run through of one possible setup: https://dev.to/ewefie/getting-started-with-github-actions-run-junit-5-tests-in-a-java-project-with-maven-20g4 | priority | set up junit ci i d like to get junit tests set up to run via github actions will run any time pushes are made for easy test automation seems like a good run through of one possible setup | 1 |
224,685 | 7,472,053,348 | IssuesEvent | 2018-04-03 11:20:50 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | Suite P - Emails View Relationship popup collapses on second use | Fix Proposed Low Priority Resolved: Next Release bug | #### Issue
Using the Suite P theme in version 7.7.9. In the Emails component right click on an email and select the 'View Relationships' option. The 'Email Record' popup displays with sufficient height to view the contents (see attachment 1).
If the popup is closed and then the action is repeated (right click on an email and select the 'View Relationships' option) the popup reappears but only the popup header is visible (none of the content is visible) rendering the popup useless (see attachment 2).
Refreshing the page works around the issue but is very inconvenient to have to do so every time.
[This forum post](https://suitecrm.com/forum/suite-themes/12065-suitep-quick-create-from-email-multiselect-dropdown-box-size#40926) describes the same bad behaviour for the 'Quick Create' -> 'Contact' popup
#### Expected Behavior
The popup should open every time with at least adequate minimum height to make the content view-able.
#### Actual Behavior
The popup opens up the first time with adequate height to make the content view-able, but all subsequent openings result in the popup having far too little height to view any contents.
#### Possible Fix
Looking at the attachments you can see the height of the container is calculated the first time at 381px but the second time the height is calculated at only 10px!
I suggest that when dynamically calculating the height of a popup's content there should be a lower limit (say 150px) that the height will not be set lower than ever. This would at least allow the content to be viewed.
#### Steps to Reproduce
1. Visit [Live Demo](http://demo.suiteondemand.com/index.php?module=Emails&action=index&parentTab=All)
2. Got to Emails module and select 'My Sent Emails'
3. Right click on an email and select the 'View Relationships' option.
4. Close popup
5. Right click on an email and select the 'View Relationships' option.
#### Context
Our users spend most of there time in the Emails module so this is a very annoying and time wasting bug.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.7.9
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Chrome Version 55.0.2883.87 (64-bit)
* Environment name and version (e.g. MySQL, PHP 7): MySQL 5.6, PHP 5.4.36
* Operating System and version (e.g Ubuntu 16.04): Ubuntu 14.04.1 LTS
| 1.0 | Suite P - Emails View Relationship popup collapses on second use - #### Issue
Using the Suite P theme in version 7.7.9. In the Emails component right click on an email and select the 'View Relationships' option. The 'Email Record' popup displays with sufficient height to view the contents (see attachment 1).
If the popup is closed and then the action is repeated (right click on an email and select the 'View Relationships' option) the popup reappears but only the popup header is visible (none of the content is visible) rendering the popup useless (see attachment 2).
Refreshing the page works around the issue but is very inconvenient to have to do so every time.
[This forum post](https://suitecrm.com/forum/suite-themes/12065-suitep-quick-create-from-email-multiselect-dropdown-box-size#40926) describes the same bad behaviour for the 'Quick Create' -> 'Contact' popup
#### Expected Behavior
The popup should open every time with at least adequate minimum height to make the content view-able.
#### Actual Behavior
The popup opens up the first time with adequate height to make the content view-able, but all subsequent openings result in the popup having far too little height to view any contents.
#### Possible Fix
Looking at the attachments you can see the height of the container is calculated the first time at 381px but the second time the height is calculated at only 10px!
I suggest that when dynamically calculating the height of a popup's content there should be a lower limit (say 150px) that the height will not be set lower than ever. This would at least allow the content to be viewed.
#### Steps to Reproduce
1. Visit [Live Demo](http://demo.suiteondemand.com/index.php?module=Emails&action=index&parentTab=All)
2. Got to Emails module and select 'My Sent Emails'
3. Right click on an email and select the 'View Relationships' option.
4. Close popup
5. Right click on an email and select the 'View Relationships' option.
#### Context
Our users spend most of there time in the Emails module so this is a very annoying and time wasting bug.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.7.9
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Chrome Version 55.0.2883.87 (64-bit)
* Environment name and version (e.g. MySQL, PHP 7): MySQL 5.6, PHP 5.4.36
* Operating System and version (e.g Ubuntu 16.04): Ubuntu 14.04.1 LTS
| priority | suite p emails view relationship popup collapses on second use issue using the suite p theme in version in the emails component right click on an email and select the view relationships option the email record popup displays with sufficient height to view the contents see attachment if the popup is closed and then the action is repeated right click on an email and select the view relationships option the popup reappears but only the popup header is visible none of the content is visible rendering the popup useless see attachment refreshing the page works around the issue but is very inconvenient to have to do so every time describes the same bad behaviour for the quick create contact popup expected behavior the popup should open every time with at least adequate minimum height to make the content view able actual behavior the popup opens up the first time with adequate height to make the content view able but all subsequent openings result in the popup having far too little height to view any contents possible fix looking at the attachments you can see the height of the container is calculated the first time at but the second time the height is calculated at only i suggest that when dynamically calculating the height of a popup s content there should be a lower limit say that the height will not be set lower than ever this would at least allow the content to be viewed steps to reproduce visit got to emails module and select my sent emails right click on an email and select the view relationships option close popup right click on an email and select the view relationships option context our users spend most of there time in the emails module so this is a very annoying and time wasting bug your environment suitecrm version used version browser name and version e g chrome version bit chrome version bit environment name and version e g mysql php mysql php operating system and version e g ubuntu ubuntu lts | 1 |
148,215 | 5,661,034,852 | IssuesEvent | 2017-04-10 16:23:40 | askmike/gekko | https://api.github.com/repos/askmike/gekko | closed | Can you add whaleclub to supported exchanges | enhancement exchange low-priority open-for-pulls | Is there a way to add whaleclub to supported exchanges since they just released their api? | 1.0 | Can you add whaleclub to supported exchanges - Is there a way to add whaleclub to supported exchanges since they just released their api? | priority | can you add whaleclub to supported exchanges is there a way to add whaleclub to supported exchanges since they just released their api | 1 |
328,851 | 10,000,760,830 | IssuesEvent | 2019-07-12 14:07:52 | zenika-open-source/FAQ | https://api.github.com/repos/zenika-open-source/FAQ | closed | Crash on home page following specific steps | bug good first issue priority/low | **Describe the bug**
When I do the following steps, the page will crash showing a blank page
**To Reproduce**
Steps to reproduce the behavior:
1. Open 2 browser tabs
2. On the first tab, create a new question (ex: "test123")
3. On the second tab, go the homepage and make a search that will show this question (ex: "test")
4. On the first tab, answer the question
5. On the second tab, _without refreshing_, click on the question
6. On the second tab, go back to homepage
7. Observe crash on the second tab
**Expected behavior**
It shouldn't crash
**Screenshots**
<img width="748" alt="Screenshot 2019-07-10 at 17 45 42" src="https://user-images.githubusercontent.com/6669733/60983635-91bca200-a33a-11e9-8983-a846ae68da05.png">
**Additional context**
I think it is some form of Apollo caching issue
Note: Since those are pretty specific steps to reproduce this and we simply need to refresh the tab to restore the app, I added the tag "priority/low" | 1.0 | Crash on home page following specific steps - **Describe the bug**
When I do the following steps, the page will crash showing a blank page
**To Reproduce**
Steps to reproduce the behavior:
1. Open 2 browser tabs
2. On the first tab, create a new question (ex: "test123")
3. On the second tab, go the homepage and make a search that will show this question (ex: "test")
4. On the first tab, answer the question
5. On the second tab, _without refreshing_, click on the question
6. On the second tab, go back to homepage
7. Observe crash on the second tab
**Expected behavior**
It shouldn't crash
**Screenshots**
<img width="748" alt="Screenshot 2019-07-10 at 17 45 42" src="https://user-images.githubusercontent.com/6669733/60983635-91bca200-a33a-11e9-8983-a846ae68da05.png">
**Additional context**
I think it is some form of Apollo caching issue
Note: Since those are pretty specific steps to reproduce this and we simply need to refresh the tab to restore the app, I added the tag "priority/low" | priority | crash on home page following specific steps describe the bug when i do the following steps the page will crash showing a blank page to reproduce steps to reproduce the behavior open browser tabs on the first tab create a new question ex on the second tab go the homepage and make a search that will show this question ex test on the first tab answer the question on the second tab without refreshing click on the question on the second tab go back to homepage observe crash on the second tab expected behavior it shouldn t crash screenshots img width alt screenshot at src additional context i think it is some form of apollo caching issue note since those are pretty specific steps to reproduce this and we simply need to refresh the tab to restore the app i added the tag priority low | 1 |
727,395 | 25,034,009,723 | IssuesEvent | 2022-11-04 14:37:31 | chaotic-aur/packages | https://api.github.com/repos/chaotic-aur/packages | closed | [Request] Include palemoon-gtk instead of binary | request:new-pkg rejected:nature request:remove-pkg priority:low | ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/palemoon-gtk3
https://aur.archlinux.org/packages/palemoon-gtk3-bin
### Utility this package has for you
You're currently creating `palemoon-gtk3-bin` (linked above). Not sure why you package a binary, I mean it has no benefits.
Could you please include `palemoon-gtk3` instead? This would make sure the package will be built with the latest build tools arch provides, instead of the build server the project is using. :)
### Related requests
https://github.com/chaotic-aur/packages/issues/573
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | 1.0 | [Request] Include palemoon-gtk instead of binary - ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/palemoon-gtk3
https://aur.archlinux.org/packages/palemoon-gtk3-bin
### Utility this package has for you
You're currently creating `palemoon-gtk3-bin` (linked above). Not sure why you package a binary, I mean it has no benefits.
Could you please include `palemoon-gtk3` instead? This would make sure the package will be built with the latest build tools arch provides, instead of the build server the project is using. :)
### Related requests
https://github.com/chaotic-aur/packages/issues/573
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | priority | include palemoon gtk instead of binary link to the package s in the aur utility this package has for you you re currently creating palemoon bin linked above not sure why you package a binary i mean it has no benefits could you please include palemoon instead this would make sure the package will be built with the latest build tools arch provides instead of the build server the project is using related requests do you consider the package s to be useful for every chaotic aur user no but for a few do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information no response | 1 |
296,737 | 9,125,653,180 | IssuesEvent | 2019-02-24 15:31:39 | pixijs/pixi.js | https://api.github.com/repos/pixijs/pixi.js | closed | Prefiltered line technique to dramatically improve the rendering of lines an shapes | Difficulty: Medium Domain: API Plugin: Graphics Priority: Low Resolution: Won't Fix Status: Accepting PRs Type: Feature Request Version: v5.x | My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA)
I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI.
http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
| 1.0 | Prefiltered line technique to dramatically improve the rendering of lines an shapes - My understanding is that currently PIXI uses the stencil buffer to draw graphics objects like lines and shapes. Since the stencil buffer does not support alpha all lines are very jagged/aliased. This is fine with AA or FXAA on but both of these come with other sets of issues that may or may not be acceptable for a given project.(GPU/browser support/performance for AA and text rendering issues for FXAA)
I've stumbled across a different technique which appears to complete solve this issue and if my understanding is right, it should be faster than the current method. I thought I would share the link in case it is something you guys might want to look into for a future version of PIXI.
http://http.developer.nvidia.com/GPUGems2/gpugems2_chapter22.html
| priority | prefiltered line technique to dramatically improve the rendering of lines an shapes my understanding is that currently pixi uses the stencil buffer to draw graphics objects like lines and shapes since the stencil buffer does not support alpha all lines are very jagged aliased this is fine with aa or fxaa on but both of these come with other sets of issues that may or may not be acceptable for a given project gpu browser support performance for aa and text rendering issues for fxaa i ve stumbled across a different technique which appears to complete solve this issue and if my understanding is right it should be faster than the current method i thought i would share the link in case it is something you guys might want to look into for a future version of pixi | 1 |
41,828 | 2,869,084,429 | IssuesEvent | 2015-06-05 23:12:26 | dart-lang/polymer-dart | https://api.github.com/repos/dart-lang/polymer-dart | opened | improve handling of private methods in polymer code generation | enhancement PolymerMilestone-Later Priority-Low | <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#18490_
----
See: https://code.google.com/p/dart/issues/detail?id=18481
Options:
(a) give better error messages in cases like \_highlightChanged
(b) add support for private methods. This would require:
- switching smoke to use strings instead of symbols (#_foo is not supported)
- switching codegeneration to generate code on each library (requires issue dart-lang/code_transformers#17), we then can generate the getters for these private symbols in the context of the same library
| 1.0 | improve handling of private methods in polymer code generation - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#18490_
----
See: https://code.google.com/p/dart/issues/detail?id=18481
Options:
(a) give better error messages in cases like \_highlightChanged
(b) add support for private methods. This would require:
- switching smoke to use strings instead of symbols (#_foo is not supported)
- switching codegeneration to generate code on each library (requires issue dart-lang/code_transformers#17), we then can generate the getters for these private symbols in the context of the same library
| priority | improve handling of private methods in polymer code generation issue by originally opened as dart lang sdk see options a give better error messages in cases like highlightchanged b add support for private methods this would require nbsp nbsp switching smoke to use strings instead of symbols foo is not supported nbsp nbsp switching codegeneration to generate code on each library requires issue dart lang code transformers we then can generate the getters for these private symbols in the context of the same library | 1 |
90,844 | 3,833,319,606 | IssuesEvent | 2016-04-01 02:29:50 | yanxijian/rapidjson | https://api.github.com/repos/yanxijian/rapidjson | closed | perftest or unittest should be set as startup project in Visual Studio solutions | auto-migrated Priority-Low Type-Enhancement | ```
If you open vs2010/test.sln, the startup project is set to gtest, which gives
an error if you try to debug the solution. It would be more user friendly if
either perftest or unittest was set as startup project for all Visual Studio
solutions.
There should be some setting in premake4 to do this, but not sure what it is.
```
Original issue reported on code.google.com by `keith3...@rocketmail.com` on 22 Jan 2013 at 1:57 | 1.0 | perftest or unittest should be set as startup project in Visual Studio solutions - ```
If you open vs2010/test.sln, the startup project is set to gtest, which gives
an error if you try to debug the solution. It would be more user friendly if
either perftest or unittest was set as startup project for all Visual Studio
solutions.
There should be some setting in premake4 to do this, but not sure what it is.
```
Original issue reported on code.google.com by `keith3...@rocketmail.com` on 22 Jan 2013 at 1:57 | priority | perftest or unittest should be set as startup project in visual studio solutions if you open test sln the startup project is set to gtest which gives an error if you try to debug the solution it would be more user friendly if either perftest or unittest was set as startup project for all visual studio solutions there should be some setting in to do this but not sure what it is original issue reported on code google com by rocketmail com on jan at | 1 |
191,186 | 6,826,750,587 | IssuesEvent | 2017-11-08 15:04:44 | forcecraft/aion | https://api.github.com/repos/forcecraft/aion | closed | fix(pre-commit): Fix dialyzer reports | 0 - Backlog Priority: Low Type: Bug | **Type**
Bug
**Current behaviour**
If dialyzer fails during running of the pre-commit hook right now, it doesn't print any error messages
**Expected behaviour**
It would be good not to have to run mix dialyzer to get to know what's the problem.
| 1.0 | fix(pre-commit): Fix dialyzer reports - **Type**
Bug
**Current behaviour**
If dialyzer fails during running of the pre-commit hook right now, it doesn't print any error messages
**Expected behaviour**
It would be good not to have to run mix dialyzer to get to know what's the problem.
| priority | fix pre commit fix dialyzer reports type bug current behaviour if dialyzer fails during running of the pre commit hook right now it doesn t print any error messages expected behaviour it would be good not to have to run mix dialyzer to get to know what s the problem | 1 |
147,031 | 5,632,500,303 | IssuesEvent | 2017-04-05 16:43:15 | Esri/data-assistant | https://api.github.com/repos/Esri/data-assistant | closed | Error if you do not establish an initial connection | bug Known Issue priority: low Verified | If you open the Pro project and it contains data in an enterprise geodatabase, you need to establish a connection via the sde connection before you can create an DA config file.
If you open the pro project and try to make a config file right away, you get the following error:
```
Running script SourceTargetMapping...
Source: Transformer
Target: Electric Device
File: C:\Users\Win10\Downloads\UN_SampleData\DataAssistantFIles_EGDB\test.xml
Used data source property
Transformer
Used data source property
Electric Device
Failed script SourceTargetMapping...
Traceback (most recent call last): File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 331, in <module> main() File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 72, in main createDlaFile(source,target,xmlFileName) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 83, in createDlaFile targetPath = dla.getLayerPath(target) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 798, in getLayerPath pth = repairLayerSourceUrl(pth,layer) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 870, in repairLayerSourceUrl connfile = getConnectionFile(lyr.connectionProperties) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 1001, in getConnectionFile date=getcp(cp,'date')) File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\management.py", line 17073, in CreateDatabaseConnection raise e File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\management.py", line 17070, in CreateDatabaseConnection retval = convertArcObjectToPythonObject(gp.CreateDatabaseConnection_management(*gp_fixargs((out_folder_path, out_name, database_platform, instance, account_authentication, username, password, save_user_pass, database, schema, version_type, version, date), True))) File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\geoprocessing\_base.py", line 506, in <lambda> return lambda *args: val(*gp_fixargs(args, True))arcgisscripting.ExecuteError: ERROR 999999: Error executing function.Failed to execute (CreateDatabaseConnection).
Failed to execute (SourceTargetMapping).
``` | 1.0 | Error if you do not establish an initial connection - If you open the Pro project and it contains data in an enterprise geodatabase, you need to establish a connection via the sde connection before you can create an DA config file.
If you open the pro project and try to make a config file right away, you get the following error:
```
Running script SourceTargetMapping...
Source: Transformer
Target: Electric Device
File: C:\Users\Win10\Downloads\UN_SampleData\DataAssistantFIles_EGDB\test.xml
Used data source property
Transformer
Used data source property
Electric Device
Failed script SourceTargetMapping...
Traceback (most recent call last): File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 331, in <module> main() File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 72, in main createDlaFile(source,target,xmlFileName) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dlaCreateSourceTarget.py", line 83, in createDlaFile targetPath = dla.getLayerPath(target) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 798, in getLayerPath pth = repairLayerSourceUrl(pth,layer) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 870, in repairLayerSourceUrl connfile = getConnectionFile(lyr.connectionProperties) File "C:\Users\Win10\AppData\Local\ESRI\ArcGISPro\AssemblyCache\{28142961-b645-420f-ba2a-72bcf8212558}\GPTools\arcpy\dla.py", line 1001, in getConnectionFile date=getcp(cp,'date')) File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\management.py", line 17073, in CreateDatabaseConnection raise e File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\management.py", line 17070, in CreateDatabaseConnection retval = convertArcObjectToPythonObject(gp.CreateDatabaseConnection_management(*gp_fixargs((out_folder_path, out_name, database_platform, instance, account_authentication, username, password, save_user_pass, database, schema, version_type, version, date), True))) File "c:\program files\arcgis\pro\Resources\arcpy\arcpy\geoprocessing\_base.py", line 506, in <lambda> return lambda *args: val(*gp_fixargs(args, True))arcgisscripting.ExecuteError: ERROR 999999: Error executing function.Failed to execute (CreateDatabaseConnection).
Failed to execute (SourceTargetMapping).
``` | priority | error if you do not establish an initial connection if you open the pro project and it contains data in an enterprise geodatabase you need to establish a connection via the sde connection before you can create an da config file if you open the pro project and try to make a config file right away you get the following error running script sourcetargetmapping source transformer target electric device file c users downloads un sampledata dataassistantfiles egdb test xml used data source property transformer used data source property electric device failed script sourcetargetmapping traceback most recent call last file c users appdata local esri arcgispro assemblycache gptools arcpy dlacreatesourcetarget py line in main file c users appdata local esri arcgispro assemblycache gptools arcpy dlacreatesourcetarget py line in main createdlafile source target xmlfilename file c users appdata local esri arcgispro assemblycache gptools arcpy dlacreatesourcetarget py line in createdlafile targetpath dla getlayerpath target file c users appdata local esri arcgispro assemblycache gptools arcpy dla py line in getlayerpath pth repairlayersourceurl pth layer file c users appdata local esri arcgispro assemblycache gptools arcpy dla py line in repairlayersourceurl connfile getconnectionfile lyr connectionproperties file c users appdata local esri arcgispro assemblycache gptools arcpy dla py line in getconnectionfile date getcp cp date file c program files arcgis pro resources arcpy arcpy management py line in createdatabaseconnection raise e file c program files arcgis pro resources arcpy arcpy management py line in createdatabaseconnection retval convertarcobjecttopythonobject gp createdatabaseconnection management gp fixargs out folder path out name database platform instance account authentication username password save user pass database schema version type version date true file c program files arcgis pro resources arcpy arcpy geoprocessing base py line in return lambda args val gp fixargs args true arcgisscripting executeerror error error executing function failed to execute createdatabaseconnection failed to execute sourcetargetmapping | 1 |
785,090 | 27,597,935,085 | IssuesEvent | 2023-03-09 08:02:52 | svthalia/Reaxit | https://api.github.com/repos/svthalia/Reaxit | opened | Support caching of S3 images | priority: low | <!--
Please add the appropriate label for what change should be made:
docs: changes to the documentation)
refactor: refactoring production code, eg. renaming a variable or rewriting a function
test: adding missing tests, refactoring tests; no production code change
chore: updating poetry etc; no production code change
-->
### What
On the website we're going to replace images served with a get_thumbnail url (that would either serve an image directly from SSD or redirect to an S3 cdn.thalia.nu url) with directly serving the relevant S3 url. I'm not yet entirely sure about the details, but in any case, if cdn.thalia.nu urls are returned by the api, we currently would not cache properly: the cache key would include the signature, i.e. we only use the cache as long as the signature hasn't changed.
### Why
S3 will be nice, cheap and scalable, but still caching is better than not caching.
### How
Details are somewhat TBD. I expect simply also removing query params from cdn urls when determining cache keys would probably work. | 1.0 | Support caching of S3 images - <!--
Please add the appropriate label for what change should be made:
docs: changes to the documentation)
refactor: refactoring production code, eg. renaming a variable or rewriting a function
test: adding missing tests, refactoring tests; no production code change
chore: updating poetry etc; no production code change
-->
### What
On the website we're going to replace images served with a get_thumbnail url (that would either serve an image directly from SSD or redirect to an S3 cdn.thalia.nu url) with directly serving the relevant S3 url. I'm not yet entirely sure about the details, but in any case, if cdn.thalia.nu urls are returned by the api, we currently would not cache properly: the cache key would include the signature, i.e. we only use the cache as long as the signature hasn't changed.
### Why
S3 will be nice, cheap and scalable, but still caching is better than not caching.
### How
Details are somewhat TBD. I expect simply also removing query params from cdn urls when determining cache keys would probably work. | priority | support caching of images please add the appropriate label for what change should be made docs changes to the documentation refactor refactoring production code eg renaming a variable or rewriting a function test adding missing tests refactoring tests no production code change chore updating poetry etc no production code change what on the website we re going to replace images served with a get thumbnail url that would either serve an image directly from ssd or redirect to an cdn thalia nu url with directly serving the relevant url i m not yet entirely sure about the details but in any case if cdn thalia nu urls are returned by the api we currently would not cache properly the cache key would include the signature i e we only use the cache as long as the signature hasn t changed why will be nice cheap and scalable but still caching is better than not caching how details are somewhat tbd i expect simply also removing query params from cdn urls when determining cache keys would probably work | 1 |
587,220 | 17,607,513,828 | IssuesEvent | 2021-08-17 19:09:22 | canonical-web-and-design/jaas-dashboard | https://api.github.com/repos/canonical-web-and-design/jaas-dashboard | opened | Update TestRoute to check routes | Priority: Low | With the change to the model details nested routing the TestRoute component needed to be gutted. This will need to be updated again to check that any route being used in a test is a valid route in the app. | 1.0 | Update TestRoute to check routes - With the change to the model details nested routing the TestRoute component needed to be gutted. This will need to be updated again to check that any route being used in a test is a valid route in the app. | priority | update testroute to check routes with the change to the model details nested routing the testroute component needed to be gutted this will need to be updated again to check that any route being used in a test is a valid route in the app | 1 |
617,090 | 19,342,441,618 | IssuesEvent | 2021-12-15 07:01:37 | matrixorigin/matrixone | https://api.github.com/repos/matrixorigin/matrixone | closed | print matrixone and matrixcube version and commit info in log | priority/low kind/feature needs-triage severity/minor | ### Is there an existing issue for the same bug?
- [X] I have checked the existing issues.
### Environment
```markdown
- Version or commit-id (e.g. v0.1.0 or 8b23a93):
- Hardware parameters:
- OS type:
- Others:
```
### Actual Behavior
print matrixone and matrixcube version and commit info in log
### Expected Behavior
_No response_
### Steps to Reproduce
_No response_
### Additional information
_No response_ | 1.0 | print matrixone and matrixcube version and commit info in log - ### Is there an existing issue for the same bug?
- [X] I have checked the existing issues.
### Environment
```markdown
- Version or commit-id (e.g. v0.1.0 or 8b23a93):
- Hardware parameters:
- OS type:
- Others:
```
### Actual Behavior
print matrixone and matrixcube version and commit info in log
### Expected Behavior
_No response_
### Steps to Reproduce
_No response_
### Additional information
_No response_ | priority | print matrixone and matrixcube version and commit info in log is there an existing issue for the same bug i have checked the existing issues environment markdown version or commit id e g or hardware parameters os type others actual behavior print matrixone and matrixcube version and commit info in log expected behavior no response steps to reproduce no response additional information no response | 1 |
192,683 | 6,876,403,394 | IssuesEvent | 2017-11-20 00:10:30 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | opened | make sunstone animations configurable | Category: Sunstone Priority: Low Status: Pending Tracker: Backlog | ---
Author Name: **Rolandas Naujikas** (Rolandas Naujikas)
Original Redmine Issue: 2018, https://dev.opennebula.org/issues/2018
Original Date: 2013-05-10
---
Many dialog windows reveals with default animation (fadeAndPop).
Please make some way to configure it as user settings, with system default also.
As a workaround this patch works (after sunstone server restart):
```
diff --git a/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js b/src/sunstone/public/vendor/4.0/foundation/jquery.foundation
index 4972931..ae63a52 100644
--- a/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js
+++ b/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js
@@ -54,7 +54,7 @@
* @type {String}
* @default fadeAndPop
*/
- animation: 'fadeAndPop',
+ animation: 'none',
/**
* Speed at which the reveal should show. How fast animtions are.
*
```
| 1.0 | make sunstone animations configurable - ---
Author Name: **Rolandas Naujikas** (Rolandas Naujikas)
Original Redmine Issue: 2018, https://dev.opennebula.org/issues/2018
Original Date: 2013-05-10
---
Many dialog windows reveals with default animation (fadeAndPop).
Please make some way to configure it as user settings, with system default also.
As a workaround this patch works (after sunstone server restart):
```
diff --git a/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js b/src/sunstone/public/vendor/4.0/foundation/jquery.foundation
index 4972931..ae63a52 100644
--- a/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js
+++ b/src/sunstone/public/vendor/4.0/foundation/jquery.foundation.reveal.js
@@ -54,7 +54,7 @@
* @type {String}
* @default fadeAndPop
*/
- animation: 'fadeAndPop',
+ animation: 'none',
/**
* Speed at which the reveal should show. How fast animtions are.
*
```
| priority | make sunstone animations configurable author name rolandas naujikas rolandas naujikas original redmine issue original date many dialog windows reveals with default animation fadeandpop please make some way to configure it as user settings with system default also as a workaround this patch works after sunstone server restart diff git a src sunstone public vendor foundation jquery foundation reveal js b src sunstone public vendor foundation jquery foundation index a src sunstone public vendor foundation jquery foundation reveal js b src sunstone public vendor foundation jquery foundation reveal js type string default fadeandpop animation fadeandpop animation none speed at which the reveal should show how fast animtions are | 1 |
290,707 | 8,903,077,673 | IssuesEvent | 2019-01-17 09:32:53 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | TreeView collapse method does not work if called right after calling expand method and vise versa | Bug C: TreeView Kendo2 Priority 2 SEV: Low Triaged Won't Fix | ### Bug report - scenario 1
The collapse method does not work as expected if it is called after calling expand and the TreeView's animation is enabled.
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/Ixeve) example.
1. Run the example.
2. **Expected**: the "foo" node should be collapsed.
3. **Actual**: the "foo" node remains expanded.
### Bug report - scneario 2
The expandTo method does not work as expected if it is called after calling collapse.
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/eWeXOVIF) example.
1. Run the example.
2. Click 'Collapse then expand to node'
3. **Expected**: The node should collapse and then expand again
4. **Actual**: The node get collapsed, but does not expand again.
If setTimeout is used, then the node is expanded correctly (click 'Collapse then expand to node (with delay)'
### Workaround
Wrap the second method in setTimeout function.
`setTimeout(function(){
treeView.expandTo(dataItem)
}, 500);`
### Environment
* **Kendo UI version:** 2018.1.221
* **jQuery version:** x.y
* **Browser:** [all ]
| 1.0 | TreeView collapse method does not work if called right after calling expand method and vise versa - ### Bug report - scenario 1
The collapse method does not work as expected if it is called after calling expand and the TreeView's animation is enabled.
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/Ixeve) example.
1. Run the example.
2. **Expected**: the "foo" node should be collapsed.
3. **Actual**: the "foo" node remains expanded.
### Bug report - scneario 2
The expandTo method does not work as expected if it is called after calling collapse.
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/eWeXOVIF) example.
1. Run the example.
2. Click 'Collapse then expand to node'
3. **Expected**: The node should collapse and then expand again
4. **Actual**: The node get collapsed, but does not expand again.
If setTimeout is used, then the node is expanded correctly (click 'Collapse then expand to node (with delay)'
### Workaround
Wrap the second method in setTimeout function.
`setTimeout(function(){
treeView.expandTo(dataItem)
}, 500);`
### Environment
* **Kendo UI version:** 2018.1.221
* **jQuery version:** x.y
* **Browser:** [all ]
| priority | treeview collapse method does not work if called right after calling expand method and vise versa bug report scenario the collapse method does not work as expected if it is called after calling expand and the treeview s animation is enabled reproduction of the problem example run the example expected the foo node should be collapsed actual the foo node remains expanded bug report scneario the expandto method does not work as expected if it is called after calling collapse reproduction of the problem example run the example click collapse then expand to node expected the node should collapse and then expand again actual the node get collapsed but does not expand again if settimeout is used then the node is expanded correctly click collapse then expand to node with delay workaround wrap the second method in settimeout function settimeout function treeview expandto dataitem environment kendo ui version jquery version x y browser | 1 |
289,306 | 8,868,676,121 | IssuesEvent | 2019-01-11 01:07:40 | SamuelMoriarty/YARP2 | https://api.github.com/repos/SamuelMoriarty/YARP2 | opened | Implement Chat Log | low priority | The chat log will be largely the same as in YARP1, unless I come up with a better idea.
Instead of having entirely separate chat logs, there will be "filters" for IC and OOC chat, which will be toggleable with a command.
In addition, I might add some extra commands for configuring the size of the chat log box. | 1.0 | Implement Chat Log - The chat log will be largely the same as in YARP1, unless I come up with a better idea.
Instead of having entirely separate chat logs, there will be "filters" for IC and OOC chat, which will be toggleable with a command.
In addition, I might add some extra commands for configuring the size of the chat log box. | priority | implement chat log the chat log will be largely the same as in unless i come up with a better idea instead of having entirely separate chat logs there will be filters for ic and ooc chat which will be toggleable with a command in addition i might add some extra commands for configuring the size of the chat log box | 1 |
268,897 | 8,415,583,356 | IssuesEvent | 2018-10-13 16:11:39 | chanzuckerberg/cellxgene | https://api.github.com/repos/chanzuckerberg/cellxgene | closed | splat when doing differential expression compute on large data set | Priority Low bug frontend | using pbmc33k data set, computed differential expression where both data sets were all cells. Generated an error.
```POST http://pbmc33k.cxg.czi.technology/api/v0.1/expression 400 (BAD REQUEST)```
```
d3.js:127 Uncaught (in promise) TypeError: Cannot read property 'length' of undefined
at Object.extent (d3.js:127)
at t.maybeSetupScalesAndDrawAxes ((index):1)
at t.componentWillReceiveProps ((index):1)
at c.updateComponent ((index):1)
at c.receiveComponent ((index):1)
at Object.receiveComponent ((index):1)
at c._updateRenderedComponent ((index):1)
at c._performComponentUpdate ((index):1)
at c.updateComponent ((index):1)
at c.performUpdateIfNecessary ((index):1)
at Object.performUpdateIfNecessary ((index):1)
at a ((index):1)
at r.perform ((index):1)
at o.perform ((index):1)
at o.perform ((index):1)
at Object.w [as flushBatchedUpdates] ((index):1)
at r.closeAll ((index):1)
at r.perform ((index):1)
at Object.batchedUpdates ((index):1)
at Object.e [as enqueueUpdate] ((index):1)
at r ((index):1)
at Object.enqueueSetState ((index):1)
at i.r.setState ((index):1)
at i.onStateChange ((index):1)
at Object.notify ((index):1)
at e.notifyNestedSubs ((index):1)
at i.onStateChange ((index):1)
at p ((index):1)
at (index):1
at (index):1
at (index):1
at (index):1
at dispatch ((index):1)
at (index):1
at <anonymous>
``` | 1.0 | splat when doing differential expression compute on large data set - using pbmc33k data set, computed differential expression where both data sets were all cells. Generated an error.
```POST http://pbmc33k.cxg.czi.technology/api/v0.1/expression 400 (BAD REQUEST)```
```
d3.js:127 Uncaught (in promise) TypeError: Cannot read property 'length' of undefined
at Object.extent (d3.js:127)
at t.maybeSetupScalesAndDrawAxes ((index):1)
at t.componentWillReceiveProps ((index):1)
at c.updateComponent ((index):1)
at c.receiveComponent ((index):1)
at Object.receiveComponent ((index):1)
at c._updateRenderedComponent ((index):1)
at c._performComponentUpdate ((index):1)
at c.updateComponent ((index):1)
at c.performUpdateIfNecessary ((index):1)
at Object.performUpdateIfNecessary ((index):1)
at a ((index):1)
at r.perform ((index):1)
at o.perform ((index):1)
at o.perform ((index):1)
at Object.w [as flushBatchedUpdates] ((index):1)
at r.closeAll ((index):1)
at r.perform ((index):1)
at Object.batchedUpdates ((index):1)
at Object.e [as enqueueUpdate] ((index):1)
at r ((index):1)
at Object.enqueueSetState ((index):1)
at i.r.setState ((index):1)
at i.onStateChange ((index):1)
at Object.notify ((index):1)
at e.notifyNestedSubs ((index):1)
at i.onStateChange ((index):1)
at p ((index):1)
at (index):1
at (index):1
at (index):1
at (index):1
at dispatch ((index):1)
at (index):1
at <anonymous>
``` | priority | splat when doing differential expression compute on large data set using data set computed differential expression where both data sets were all cells generated an error post bad request js uncaught in promise typeerror cannot read property length of undefined at object extent js at t maybesetupscalesanddrawaxes index at t componentwillreceiveprops index at c updatecomponent index at c receivecomponent index at object receivecomponent index at c updaterenderedcomponent index at c performcomponentupdate index at c updatecomponent index at c performupdateifnecessary index at object performupdateifnecessary index at a index at r perform index at o perform index at o perform index at object w index at r closeall index at r perform index at object batchedupdates index at object e index at r index at object enqueuesetstate index at i r setstate index at i onstatechange index at object notify index at e notifynestedsubs index at i onstatechange index at p index at index at index at index at index at dispatch index at index at | 1 |
718,712 | 24,729,544,831 | IssuesEvent | 2022-10-20 16:23:32 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | [COT/HPOS]: Add corresponding hooks for `untrash_post` for COT/HPOS implementation | type: enhancement priority: low needs: developer feedback plugin: woocommerce | ### Describe the solution you'd like
I am working on the Deposits extension to add support for trash, untrash, and delete orders for COT/HPOS implementation. To handle, when a parent order is trashed related scheduled orders should get trashed, when parent order is untrashed/restored, related scheduled orders should get untrashed/restored and finally when parent order deleted, related parent order should get deleted as well.
We have added corresponding hooks for `wp_trash_post` and `delete_post` for COT implementation in PR https://github.com/woocommerce/woocommerce/pull/34858, but we still don't have support for `untrash_post` which I needed for restored related scheduled payment orders when parent order restored/untrashed.
### Describe alternatives you've considered
Alternatively, we may use the `woocommerce_order_status_changed` hook to check if the order gets updated from `trash` status to something else.
### Additional context
Deposits issue https://github.com/woocommerce/woocommerce-deposits/issues/467 | 1.0 | [COT/HPOS]: Add corresponding hooks for `untrash_post` for COT/HPOS implementation - ### Describe the solution you'd like
I am working on the Deposits extension to add support for trash, untrash, and delete orders for COT/HPOS implementation. To handle, when a parent order is trashed related scheduled orders should get trashed, when parent order is untrashed/restored, related scheduled orders should get untrashed/restored and finally when parent order deleted, related parent order should get deleted as well.
We have added corresponding hooks for `wp_trash_post` and `delete_post` for COT implementation in PR https://github.com/woocommerce/woocommerce/pull/34858, but we still don't have support for `untrash_post` which I needed for restored related scheduled payment orders when parent order restored/untrashed.
### Describe alternatives you've considered
Alternatively, we may use the `woocommerce_order_status_changed` hook to check if the order gets updated from `trash` status to something else.
### Additional context
Deposits issue https://github.com/woocommerce/woocommerce-deposits/issues/467 | priority | add corresponding hooks for untrash post for cot hpos implementation describe the solution you d like i am working on the deposits extension to add support for trash untrash and delete orders for cot hpos implementation to handle when a parent order is trashed related scheduled orders should get trashed when parent order is untrashed restored related scheduled orders should get untrashed restored and finally when parent order deleted related parent order should get deleted as well we have added corresponding hooks for wp trash post and delete post for cot implementation in pr but we still don t have support for untrash post which i needed for restored related scheduled payment orders when parent order restored untrashed describe alternatives you ve considered alternatively we may use the woocommerce order status changed hook to check if the order gets updated from trash status to something else additional context deposits issue | 1 |
584,160 | 17,407,922,250 | IssuesEvent | 2021-08-03 08:35:24 | translate-tools/linguist | https://api.github.com/repos/translate-tools/linguist | closed | Improve size style for options | low priority refactor | Change size between sections for options to visually separate segments (obey the principle of proximity) | 1.0 | Improve size style for options - Change size between sections for options to visually separate segments (obey the principle of proximity) | priority | improve size style for options change size between sections for options to visually separate segments obey the principle of proximity | 1 |
690,168 | 23,647,705,625 | IssuesEvent | 2022-08-26 01:13:33 | ploomber/ploomber | https://api.github.com/repos/ploomber/ploomber | closed | delete `python -m ploomber.onboarding` | med priority low effort | nobody is using this so we should delete it: delete the code, the references in the README (and in the docs, if any), and the tests | 1.0 | delete `python -m ploomber.onboarding` - nobody is using this so we should delete it: delete the code, the references in the README (and in the docs, if any), and the tests | priority | delete python m ploomber onboarding nobody is using this so we should delete it delete the code the references in the readme and in the docs if any and the tests | 1 |
591,895 | 17,864,571,907 | IssuesEvent | 2021-09-06 07:52:02 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Редагування профілю] Photo size validation does not work | bug Priority: Low | **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Log in
**Steps to reproduce**
1. Go to 'Мій Профіль' page
2. Click on 'Редагувати профіль'
3. Click on 'Завантажити фото' link
4. Select photo with 1) wrong resolution ( 100x200 pixels

), 2) wrong size ( 7.4 MB

)
5. Click on 'Зберегти зміни'
**Actual result**
'Зберегти зміни' button is enabled. Photo is saved in the profile.
1)

2)

**Expected result**
'Зберегти зміни' button is disabled and

**Postconditions**
Return edited data to the initial state
**User story link**
User story #164
| 1.0 | [Редагування профілю] Photo size validation does not work - **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Log in
**Steps to reproduce**
1. Go to 'Мій Профіль' page
2. Click on 'Редагувати профіль'
3. Click on 'Завантажити фото' link
4. Select photo with 1) wrong resolution ( 100x200 pixels

), 2) wrong size ( 7.4 MB

)
5. Click on 'Зберегти зміни'
**Actual result**
'Зберегти зміни' button is enabled. Photo is saved in the profile.
1)

2)

**Expected result**
'Зберегти зміни' button is disabled and

**Postconditions**
Return edited data to the initial state
**User story link**
User story #164
| priority | photo size validation does not work environment windows google chrome version official build bit reproducible always build found last commit preconditions go to log in steps to reproduce go to мій профіль page click on редагувати профіль click on завантажити фото link select photo with wrong resolution pixels wrong size mb click on зберегти зміни actual result зберегти зміни button is enabled photo is saved in the profile expected result зберегти зміни button is disabled and postconditions return edited data to the initial state user story link user story | 1 |
62,904 | 3,193,604,011 | IssuesEvent | 2015-09-30 06:55:44 | fusioninventory/fusioninventory-for-glpi | https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi | closed | Know the imports refused | Component: For junior contributor Priority: Low Status: Rejected Tracker: Feature | ---
Author Name: **DuyLong LE** (DuyLong LE)
Original Redmine Issue: 1295, http://forge.fusioninventory.org/issues/1295
Original Date: 2011-10-26
---
With a new page..., not need to check the history tasks everyday.
| 1.0 | Know the imports refused - ---
Author Name: **DuyLong LE** (DuyLong LE)
Original Redmine Issue: 1295, http://forge.fusioninventory.org/issues/1295
Original Date: 2011-10-26
---
With a new page..., not need to check the history tasks everyday.
| priority | know the imports refused author name duylong le duylong le original redmine issue original date with a new page not need to check the history tasks everyday | 1 |
678,690 | 23,207,237,562 | IssuesEvent | 2022-08-02 06:55:45 | TencentBlueKing/bk-nodeman | https://api.github.com/repos/TencentBlueKing/bk-nodeman | closed | [FEATURE] Agent 安装 DHCP 场景适配 | kind/optimization module/backend module/saas version/V2.2.X priority/low | **你想要什么功能**
Agent 安装 DHCP 场景适配
**如何实现**
- [ ] 同 IP + 云区域: 有一台 Agent 在线时禁止安装
- [ ] 都下线:同业务下选其中一台同 IP + 云区域 的机器进行 Agent 重装
- [ ] 不存在:新装
| 1.0 | [FEATURE] Agent 安装 DHCP 场景适配 - **你想要什么功能**
Agent 安装 DHCP 场景适配
**如何实现**
- [ ] 同 IP + 云区域: 有一台 Agent 在线时禁止安装
- [ ] 都下线:同业务下选其中一台同 IP + 云区域 的机器进行 Agent 重装
- [ ] 不存在:新装
| priority | agent 安装 dhcp 场景适配 你想要什么功能 agent 安装 dhcp 场景适配 如何实现 同 ip 云区域: 有一台 agent 在线时禁止安装 都下线:同业务下选其中一台同 ip 云区域 的机器进行 agent 重装 不存在:新装 | 1 |
793,337 | 27,991,100,842 | IssuesEvent | 2023-03-27 03:53:16 | fastcampus-final/go-together-be | https://api.github.com/repos/fastcampus-final/go-together-be | opened | 인기 도시 관리 기능 | For: Backend For: Frontend Priority: Low Status: Available Type: Idea | ## Title
메인 페이지 인기 도시 노출과 , 검색 페이지 인기 도시 노출을 위한 관리 시스템
## Description
예상 필요 컬럼
PK
도시 이름
도시 이미지
노출 중요도
## Tasks
- [ ] 인기도시 Entity 설계
- [ ] 인기도시 DB table 추가
- [ ] 인기도시 추가 기능
- [ ] 인기도시 리스트 보기 기능
- [ ] 인기도시 수정 기능
- [ ] 인기도시 삭제 기능 | 1.0 | 인기 도시 관리 기능 - ## Title
메인 페이지 인기 도시 노출과 , 검색 페이지 인기 도시 노출을 위한 관리 시스템
## Description
예상 필요 컬럼
PK
도시 이름
도시 이미지
노출 중요도
## Tasks
- [ ] 인기도시 Entity 설계
- [ ] 인기도시 DB table 추가
- [ ] 인기도시 추가 기능
- [ ] 인기도시 리스트 보기 기능
- [ ] 인기도시 수정 기능
- [ ] 인기도시 삭제 기능 | priority | 인기 도시 관리 기능 title 메인 페이지 인기 도시 노출과 검색 페이지 인기 도시 노출을 위한 관리 시스템 description 예상 필요 컬럼 pk 도시 이름 도시 이미지 노출 중요도 tasks 인기도시 entity 설계 인기도시 db table 추가 인기도시 추가 기능 인기도시 리스트 보기 기능 인기도시 수정 기능 인기도시 삭제 기능 | 1 |
789,649 | 27,799,695,742 | IssuesEvent | 2023-03-17 15:00:04 | storj/storj | https://api.github.com/repos/storj/storj | closed | Explore batch insert into billing table for token payments | Priority: Low Needs Estimation | This ticket in an enhancement to the insertion of billing transactions into the billing table. Right now the records are inserted one at a time, and subsequently the user balance is updated accordingly.
It may be possible to use a batch insert for these transactions to improve performance when there are lots of token payments being received, however the balance must be updated carefully to ensure the users balance does not become negative (i.e. a user has spent more on an invoice than what they have deposited in tokens.)
In previous attempts, it appeared that batch would not work because order could not be guaranteed when batching inserts, however this may not be the case. (See comments here https://review.dev.storj.io/c/storj/storj/+/8257) | 1.0 | Explore batch insert into billing table for token payments - This ticket in an enhancement to the insertion of billing transactions into the billing table. Right now the records are inserted one at a time, and subsequently the user balance is updated accordingly.
It may be possible to use a batch insert for these transactions to improve performance when there are lots of token payments being received, however the balance must be updated carefully to ensure the users balance does not become negative (i.e. a user has spent more on an invoice than what they have deposited in tokens.)
In previous attempts, it appeared that batch would not work because order could not be guaranteed when batching inserts, however this may not be the case. (See comments here https://review.dev.storj.io/c/storj/storj/+/8257) | priority | explore batch insert into billing table for token payments this ticket in an enhancement to the insertion of billing transactions into the billing table right now the records are inserted one at a time and subsequently the user balance is updated accordingly it may be possible to use a batch insert for these transactions to improve performance when there are lots of token payments being received however the balance must be updated carefully to ensure the users balance does not become negative i e a user has spent more on an invoice than what they have deposited in tokens in previous attempts it appeared that batch would not work because order could not be guaranteed when batching inserts however this may not be the case see comments here | 1 |
408,324 | 11,945,049,736 | IssuesEvent | 2020-04-03 04:31:13 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | Python - cv2.resize with INTER_NEAREST return corrupted slice for channels > 4 | bug category: imgproc priority: low | ##### System information (version)
OpenCV => 4.1.0
Operating System / Platform => Ubuntu 16.04.1, Linux 4.13.0
Python => 3.6.8
numpy => 1.16.4
##### Detailed description
cv2.resize with an array having a 3rd-dimension > 4 (must be an odd number) and nearest-neighbor interpolation returns corrupted data in the last channel. It is not deterministic and looks like memory corruption.
##### Steps to reproduce
```python
import numpy as np
import cv2
# up to DIM=4 is OK
DIM = 5
T = np.zeros((12, 12,DIM)).astype(np.uint8)
T[2, 2, DIM-1] = 1
print("Sum: ", T.sum())
print(T[:, :, DIM-1])
R = cv2.resize(T, None, fx=0.9, fy=1,interpolation=cv2.INTER_NEAREST)
# it does not help to supply a pre-allocated array either
#R = np.zeros((12, 11, DIM)).astype(np.uint8)
#R = cv2.resize(T, dsize=R.shape[:2], dst=R, interpolation=cv2.INTER_NEAREST)
print("Sum: ", R.sum())
print(R[:, :, DIM-1])
```
Output:
```python
Sum: 1
[[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 1 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]]
Sum: 836
[[248 113 0 214 176 85 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]]
``` | 1.0 | Python - cv2.resize with INTER_NEAREST return corrupted slice for channels > 4 - ##### System information (version)
OpenCV => 4.1.0
Operating System / Platform => Ubuntu 16.04.1, Linux 4.13.0
Python => 3.6.8
numpy => 1.16.4
##### Detailed description
cv2.resize with an array having a 3rd-dimension > 4 (must be an odd number) and nearest-neighbor interpolation returns corrupted data in the last channel. It is not deterministic and looks like memory corruption.
##### Steps to reproduce
```python
import numpy as np
import cv2
# up to DIM=4 is OK
DIM = 5
T = np.zeros((12, 12,DIM)).astype(np.uint8)
T[2, 2, DIM-1] = 1
print("Sum: ", T.sum())
print(T[:, :, DIM-1])
R = cv2.resize(T, None, fx=0.9, fy=1,interpolation=cv2.INTER_NEAREST)
# it does not help to supply a pre-allocated array either
#R = np.zeros((12, 11, DIM)).astype(np.uint8)
#R = cv2.resize(T, dsize=R.shape[:2], dst=R, interpolation=cv2.INTER_NEAREST)
print("Sum: ", R.sum())
print(R[:, :, DIM-1])
```
Output:
```python
Sum: 1
[[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 1 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]
[0 0 0 0 0 0 0 0 0 0 0 0]]
Sum: 836
[[248 113 0 214 176 85 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]
[ 0 0 0 0 0 0 0 0 0 0 0]]
``` | priority | python resize with inter nearest return corrupted slice for channels system information version opencv operating system platform ubuntu linux python numpy detailed description resize with an array having a dimension must be an odd number and nearest neighbor interpolation returns corrupted data in the last channel it is not deterministic and looks like memory corruption steps to reproduce python import numpy as np import up to dim is ok dim t np zeros dim astype np t print sum t sum print t r resize t none fx fy interpolation inter nearest it does not help to supply a pre allocated array either r np zeros dim astype np r resize t dsize r shape dst r interpolation inter nearest print sum r sum print r output python sum sum | 1 |
280,740 | 8,686,063,895 | IssuesEvent | 2018-12-03 09:46:51 | Text-Mining/Persian-NER | https://api.github.com/repos/Text-Mining/Persian-NER | opened | اصلاح رنگ پس زمینه برچسبها | Effort: Low Priority: Low Product: Web App Type: enhancement | <div dir="rtl">
این پیشنهاد در این توییت داده شده است:
https://twitter.com/iMojtaba/status/1068788379158962177
بنفش شروع زمان رو یه مقدار متمایز تر از بنفش زمان بذارید
</div> | 1.0 | اصلاح رنگ پس زمینه برچسبها - <div dir="rtl">
این پیشنهاد در این توییت داده شده است:
https://twitter.com/iMojtaba/status/1068788379158962177
بنفش شروع زمان رو یه مقدار متمایز تر از بنفش زمان بذارید
</div> | priority | اصلاح رنگ پس زمینه برچسبها این پیشنهاد در این توییت داده شده است بنفش شروع زمان رو یه مقدار متمایز تر از بنفش زمان بذارید | 1 |
607,763 | 18,790,047,939 | IssuesEvent | 2021-11-08 15:55:19 | ChainSafe/chainbridge-ui | https://api.github.com/repos/ChainSafe/chainbridge-ui | closed | Autoupdate for transfer page at the explorer UI | explorer Type: Feature Priority: P3 Low | The page should show spinner at the end of Transfer Timeline if Deposits status is NOT Executed or Canceled and update data until the transfer is finished or canceled.
| 1.0 | Autoupdate for transfer page at the explorer UI - The page should show spinner at the end of Transfer Timeline if Deposits status is NOT Executed or Canceled and update data until the transfer is finished or canceled.
| priority | autoupdate for transfer page at the explorer ui the page should show spinner at the end of transfer timeline if deposits status is not executed or canceled and update data until the transfer is finished or canceled | 1 |
169,316 | 6,399,261,408 | IssuesEvent | 2017-08-04 23:25:58 | UGXaero/UGXrealms | https://api.github.com/repos/UGXaero/UGXrealms | closed | [ugx] lava priv | low priority new feature | Replace this code https://github.com/UGXaero/UGXrealms/blob/master/worldmods/ugx/init.lua#L1 with code that registers lava priv and then checks if player has lava priv when trying to place lava above 0 height. If no lava priv then prevent player from placing lava and send chat message explaining lava priv is required above sea level.
:large_orange_diamond: | 1.0 | [ugx] lava priv - Replace this code https://github.com/UGXaero/UGXrealms/blob/master/worldmods/ugx/init.lua#L1 with code that registers lava priv and then checks if player has lava priv when trying to place lava above 0 height. If no lava priv then prevent player from placing lava and send chat message explaining lava priv is required above sea level.
:large_orange_diamond: | priority | lava priv replace this code with code that registers lava priv and then checks if player has lava priv when trying to place lava above height if no lava priv then prevent player from placing lava and send chat message explaining lava priv is required above sea level large orange diamond | 1 |
663,358 | 22,191,593,119 | IssuesEvent | 2022-06-07 00:00:36 | woocommerce/woocommerce-blocks | https://api.github.com/repos/woocommerce/woocommerce-blocks | opened | `SearchListControl` shows product names in the same case the user entered. | type: bug good first issue 🔹 block-type: reviews 🔹 block-type: product priority: low | ## Describe the bug
When searching for a product, the product name is displayed in the same case you type it in the input.
## To reproduce
1. Add a Hand Picked Products Block.
2. Search for `hOoDiE`.
3. Observe the list of products returned contains the same case you entered.
## Expected behavior
I think it should show the product name as entered by the merchant, i.e. `Hoodie`
## Screenshots
<img width="747" alt="image" src="https://user-images.githubusercontent.com/5656702/172268138-7445fbf1-ad25-4716-8b5a-ee4b463af54b.png">
## Additional context
Code responsible for highlighting the part of the product name is here: https://github.com/woocommerce/woocommerce-blocks/blob/trunk/assets/js/editor-components/search-list-control/utils.tsx
| 1.0 | `SearchListControl` shows product names in the same case the user entered. - ## Describe the bug
When searching for a product, the product name is displayed in the same case you type it in the input.
## To reproduce
1. Add a Hand Picked Products Block.
2. Search for `hOoDiE`.
3. Observe the list of products returned contains the same case you entered.
## Expected behavior
I think it should show the product name as entered by the merchant, i.e. `Hoodie`
## Screenshots
<img width="747" alt="image" src="https://user-images.githubusercontent.com/5656702/172268138-7445fbf1-ad25-4716-8b5a-ee4b463af54b.png">
## Additional context
Code responsible for highlighting the part of the product name is here: https://github.com/woocommerce/woocommerce-blocks/blob/trunk/assets/js/editor-components/search-list-control/utils.tsx
| priority | searchlistcontrol shows product names in the same case the user entered describe the bug when searching for a product the product name is displayed in the same case you type it in the input to reproduce add a hand picked products block search for hoodie observe the list of products returned contains the same case you entered expected behavior i think it should show the product name as entered by the merchant i e hoodie screenshots img width alt image src additional context code responsible for highlighting the part of the product name is here | 1 |
763,951 | 26,779,251,738 | IssuesEvent | 2023-01-31 19:41:10 | CommunityFreeTalk/FreeTalk-Beta | https://api.github.com/repos/CommunityFreeTalk/FreeTalk-Beta | closed | Solve Merge Conflict | Priority: Low For: Backend | ## Description
Merge하는 과정에서 생긴 불필요 코드를 정리한다.
<!--
당신이 할 일에 대한 설명.
주의사항
- 당신이 하지 않는다고 가정하고 설명할것.
-->
## Tasks
- [x] Merge하는 과정에서 생긴 불필요 코드를 정리
## References
-[src](https://github.com/CommunityFreeTalk/FreeTalk-Beta/tree/master/src)
| 1.0 | Solve Merge Conflict - ## Description
Merge하는 과정에서 생긴 불필요 코드를 정리한다.
<!--
당신이 할 일에 대한 설명.
주의사항
- 당신이 하지 않는다고 가정하고 설명할것.
-->
## Tasks
- [x] Merge하는 과정에서 생긴 불필요 코드를 정리
## References
-[src](https://github.com/CommunityFreeTalk/FreeTalk-Beta/tree/master/src)
| priority | solve merge conflict description merge하는 과정에서 생긴 불필요 코드를 정리한다 당신이 할 일에 대한 설명 주의사항 당신이 하지 않는다고 가정하고 설명할것 tasks merge하는 과정에서 생긴 불필요 코드를 정리 references | 1 |
395,154 | 11,672,269,095 | IssuesEvent | 2020-03-04 06:06:48 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | Buy participation modal 'Gas' fixes | Add post v2 launch Priority: Low | Screen grab from build:

'Gas' should be 'Estimated Gas Fee' and number should have 'ETH' after it | 1.0 | Buy participation modal 'Gas' fixes - Screen grab from build:

'Gas' should be 'Estimated Gas Fee' and number should have 'ETH' after it | priority | buy participation modal gas fixes screen grab from build gas should be estimated gas fee and number should have eth after it | 1 |
201,279 | 7,028,639,975 | IssuesEvent | 2017-12-25 12:11:24 | silviabooks/scala-project | https://api.github.com/repos/silviabooks/scala-project | closed | Missing required field: errore nell'unmarshalling se l'oggetto json nel payload manca di qualche field (magari opzionale perche' generato da backend successivamente, es. _id) | bug low-priority | Riprodurre il bug:
```bash
aleskandro@hexaflake scala-project % curl -XPOST -H "Content-Type: application/json" \
http://localhost:8080/events \
--data '{ "name": "EventName", "date": "123123", "category": "asd", "description": "saasas" }'
The request content was malformed:
Object is missing required member '_id'
``` | 1.0 | Missing required field: errore nell'unmarshalling se l'oggetto json nel payload manca di qualche field (magari opzionale perche' generato da backend successivamente, es. _id) - Riprodurre il bug:
```bash
aleskandro@hexaflake scala-project % curl -XPOST -H "Content-Type: application/json" \
http://localhost:8080/events \
--data '{ "name": "EventName", "date": "123123", "category": "asd", "description": "saasas" }'
The request content was malformed:
Object is missing required member '_id'
``` | priority | missing required field errore nell unmarshalling se l oggetto json nel payload manca di qualche field magari opzionale perche generato da backend successivamente es id riprodurre il bug bash aleskandro hexaflake scala project curl xpost h content type application json data name eventname date category asd description saasas the request content was malformed object is missing required member id | 1 |
240,663 | 7,804,342,820 | IssuesEvent | 2018-06-11 07:03:20 | BuiltBrokenModding/Atomic-Science | https://api.github.com/repos/BuiltBrokenModding/Atomic-Science | opened | [1.7] Radiation zones | enhancement low priority | This will be a mechanic to allow using the radiation system for world customization. In which a world can be set with background radiation or a default radiation level.
How this will work is that a JSON config will exist per world. This config will contain
- [ ] Default radiation level
- [ ] Source of radiation (ground, sun, air)
- [ ] Radiation areas (3D cube)
- [ ] Radiation area source type (ground, sun, air)
Radiation source will control the logic for the areas.
Ground -> works the same way as the current system. Each tile will have a default value and will path radiation. Will cause contamination to spread overtime but can be cleaned up.
Sun -> any block connected to the top of the map by line of sight will have radiation. This can be blocked by anything between the player and the sun. Is a cached value just like the normal map. However, will be stored in a separate system and added to getRadiation(x, y, z) as a percentage of current sun exposure.
air -> any air block connected to the sky will have radiation. Can be stopped by normal mechanics and will not contaminate blocks.
| 1.0 | [1.7] Radiation zones - This will be a mechanic to allow using the radiation system for world customization. In which a world can be set with background radiation or a default radiation level.
How this will work is that a JSON config will exist per world. This config will contain
- [ ] Default radiation level
- [ ] Source of radiation (ground, sun, air)
- [ ] Radiation areas (3D cube)
- [ ] Radiation area source type (ground, sun, air)
Radiation source will control the logic for the areas.
Ground -> works the same way as the current system. Each tile will have a default value and will path radiation. Will cause contamination to spread overtime but can be cleaned up.
Sun -> any block connected to the top of the map by line of sight will have radiation. This can be blocked by anything between the player and the sun. Is a cached value just like the normal map. However, will be stored in a separate system and added to getRadiation(x, y, z) as a percentage of current sun exposure.
air -> any air block connected to the sky will have radiation. Can be stopped by normal mechanics and will not contaminate blocks.
| priority | radiation zones this will be a mechanic to allow using the radiation system for world customization in which a world can be set with background radiation or a default radiation level how this will work is that a json config will exist per world this config will contain default radiation level source of radiation ground sun air radiation areas cube radiation area source type ground sun air radiation source will control the logic for the areas ground works the same way as the current system each tile will have a default value and will path radiation will cause contamination to spread overtime but can be cleaned up sun any block connected to the top of the map by line of sight will have radiation this can be blocked by anything between the player and the sun is a cached value just like the normal map however will be stored in a separate system and added to getradiation x y z as a percentage of current sun exposure air any air block connected to the sky will have radiation can be stopped by normal mechanics and will not contaminate blocks | 1 |
314,688 | 9,601,626,613 | IssuesEvent | 2019-05-10 12:41:30 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | Item Bonus Limit -10000 ~ +10000 | component:core mode:prerenewal mode:renewal priority:low type:bug | <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**: https://github.com/rathena/rathena/commit/ce1508a01c68dee80544242b470987742f3b3ece
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: 20180621
<!-- Please specify the client date you used. -->
* **Server Mode**: RE
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->
Some items need values greater than this set range. For examples:
https://www.divine-pride.net/database/item/29584
https://www.divine-pride.net/database/item/2582
[Warning]: pc_bonus_itembonus: Item bonus val 30000 exceeds -10000~10000 range, capping.
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
The problem begins after this:
https://github.com/rathena/rathena/commit/a0bcb92b8a39a45005d111f21697af380aa6f1a1
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
* Official Information:<!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| 1.0 | Item Bonus Limit -10000 ~ +10000 - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**: https://github.com/rathena/rathena/commit/ce1508a01c68dee80544242b470987742f3b3ece
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: 20180621
<!-- Please specify the client date you used. -->
* **Server Mode**: RE
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->
Some items need values greater than this set range. For examples:
https://www.divine-pride.net/database/item/29584
https://www.divine-pride.net/database/item/2582
[Warning]: pc_bonus_itembonus: Item bonus val 30000 exceeds -10000~10000 range, capping.
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
The problem begins after this:
https://github.com/rathena/rathena/commit/a0bcb92b8a39a45005d111f21697af380aa6f1a1
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->
* Official Information:<!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| priority | item bonus limit rathena hash please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode re description of issue result some items need values greater than this set range for examples pc bonus itembonus item bonus val exceeds range capping expected result the problem begins after this how to reproduce official information modifications that may affect results | 1 |
787,093 | 27,705,932,752 | IssuesEvent | 2023-03-14 11:11:25 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | closed | Introduce Throttling for GraphQL services | Priority/Low Type/NewFeature module/graphql Later Team/PCM Area/Security Deferred | **Description:**
As a part of adding security features, the GraphQL services should have a way to throttle incoming queries, based on the time and the complexity.
Related: #1357, #1358 | 1.0 | Introduce Throttling for GraphQL services - **Description:**
As a part of adding security features, the GraphQL services should have a way to throttle incoming queries, based on the time and the complexity.
Related: #1357, #1358 | priority | introduce throttling for graphql services description as a part of adding security features the graphql services should have a way to throttle incoming queries based on the time and the complexity related | 1 |
499,827 | 14,479,943,426 | IssuesEvent | 2020-12-10 10:29:05 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | opened | Add C++ FetchDMX and RegisterUniverse Examples | Difficulty-Easy Language-C++ OpSys-All Priority-Low enhancement | To mirror ola_set_dmx/ola_streaming_client to either get the current data once, or get a feed of fresh data. | 1.0 | Add C++ FetchDMX and RegisterUniverse Examples - To mirror ola_set_dmx/ola_streaming_client to either get the current data once, or get a feed of fresh data. | priority | add c fetchdmx and registeruniverse examples to mirror ola set dmx ola streaming client to either get the current data once or get a feed of fresh data | 1 |
412,446 | 12,042,644,376 | IssuesEvent | 2020-04-14 10:57:24 | wix/wix-style-react | https://api.github.com/repos/wix/wix-style-react | closed | `<Slider/>` - pushable handlers get stuck when positioned at the start and end. | <Slider/> Bug Priority: Low | # 🐛 Bug Report
Pushable Handlers" example feels broken when controlled with a mouse. It gets stuck when knobs gets positioned at the start and end.
### 🏗 Relevant Components
`<Slider/>`
### 😯 Current Behavior
Sometimes while dragging a slider with multiple handlers - one of the knobs get stuck.
### 🤔 Expected Behavior
Should be smooth.
### 👣 Steps to Reproduce
Open storybook API docs for Slider and try play with `Pushable Handlers` example.
### 👀 Severity
- Major
| 1.0 | `<Slider/>` - pushable handlers get stuck when positioned at the start and end. - # 🐛 Bug Report
Pushable Handlers" example feels broken when controlled with a mouse. It gets stuck when knobs gets positioned at the start and end.
### 🏗 Relevant Components
`<Slider/>`
### 😯 Current Behavior
Sometimes while dragging a slider with multiple handlers - one of the knobs get stuck.
### 🤔 Expected Behavior
Should be smooth.
### 👣 Steps to Reproduce
Open storybook API docs for Slider and try play with `Pushable Handlers` example.
### 👀 Severity
- Major
| priority | pushable handlers get stuck when positioned at the start and end 🐛 bug report pushable handlers example feels broken when controlled with a mouse it gets stuck when knobs gets positioned at the start and end 🏗 relevant components 😯 current behavior sometimes while dragging a slider with multiple handlers one of the knobs get stuck 🤔 expected behavior should be smooth 👣 steps to reproduce open storybook api docs for slider and try play with pushable handlers example 👀 severity major | 1 |
119,862 | 4,776,970,455 | IssuesEvent | 2016-10-27 15:08:56 | jpppina/migracion-galeno-art-forms11g | https://api.github.com/repos/jpppina/migracion-galeno-art-forms11g | closed | COBRA – cbcarres – Contratos a Rescindir | Aplicación-COBRA DONE Error Priority-Low | Ambiente: COBRA
Usuario: APARICIOJ/Desaa001
Opción de menú: Procesos > Cartas > Rescisión de contratos
Form: cbcarres
Error: No se abre la pantalla de dicho menú.
Observaciones:
| 1.0 | COBRA – cbcarres – Contratos a Rescindir - Ambiente: COBRA
Usuario: APARICIOJ/Desaa001
Opción de menú: Procesos > Cartas > Rescisión de contratos
Form: cbcarres
Error: No se abre la pantalla de dicho menú.
Observaciones:
| priority | cobra – cbcarres – contratos a rescindir ambiente cobra usuario aparicioj opción de menú procesos cartas rescisión de contratos form cbcarres error no se abre la pantalla de dicho menú observaciones | 1 |
675,352 | 23,091,397,178 | IssuesEvent | 2022-07-26 15:29:24 | jahirfiquitiva/Kuper | https://api.github.com/repos/jahirfiquitiva/Kuper | closed | Bottom navigation bar | Status: Accepted Type: Feature Request Priority: Low | ### Kuper Version
2.3.8
### Describe the feature
Please add the ability to edit the bottom navigation bar.. the icons and the options to show..
### Describe alternatives you have considered
_No response_
### Additional context or info
_No response_ | 1.0 | Bottom navigation bar - ### Kuper Version
2.3.8
### Describe the feature
Please add the ability to edit the bottom navigation bar.. the icons and the options to show..
### Describe alternatives you have considered
_No response_
### Additional context or info
_No response_ | priority | bottom navigation bar kuper version describe the feature please add the ability to edit the bottom navigation bar the icons and the options to show describe alternatives you have considered no response additional context or info no response | 1 |
113,803 | 4,569,196,057 | IssuesEvent | 2016-09-15 16:29:53 | HazyResearch/snorkel | https://api.github.com/repos/HazyResearch/snorkel | opened | Update text in Intro Tutorial | bug low priority | The first page of the [Intro Tutorial](https://github.com/HazyResearch/snorkel/blob/41914615d11e95e76b090b51242fd12ed77f41e1/tutorials/intro/Intro_Tutorial_1.ipynb) in the `tables_merge` branch has a typo.
The intro walks through finding spouses in the news, but contains the text
> We'll start by defining a DocParser class to read in Pubmed abstracts from Pubtator), where they are stored along with "gold" (i.e. hand-annotated by experts) chemical and disease mention annotations. We'll use the XMLDocParser class, which allows us to use XPath queries to specify the relevant sections of the XML format.
which references PubMed and a different task. Update this to reflect the task of finding spouses in the news. | 1.0 | Update text in Intro Tutorial - The first page of the [Intro Tutorial](https://github.com/HazyResearch/snorkel/blob/41914615d11e95e76b090b51242fd12ed77f41e1/tutorials/intro/Intro_Tutorial_1.ipynb) in the `tables_merge` branch has a typo.
The intro walks through finding spouses in the news, but contains the text
> We'll start by defining a DocParser class to read in Pubmed abstracts from Pubtator), where they are stored along with "gold" (i.e. hand-annotated by experts) chemical and disease mention annotations. We'll use the XMLDocParser class, which allows us to use XPath queries to specify the relevant sections of the XML format.
which references PubMed and a different task. Update this to reflect the task of finding spouses in the news. | priority | update text in intro tutorial the first page of the in the tables merge branch has a typo the intro walks through finding spouses in the news but contains the text we ll start by defining a docparser class to read in pubmed abstracts from pubtator where they are stored along with gold i e hand annotated by experts chemical and disease mention annotations we ll use the xmldocparser class which allows us to use xpath queries to specify the relevant sections of the xml format which references pubmed and a different task update this to reflect the task of finding spouses in the news | 1 |
557,131 | 16,501,477,093 | IssuesEvent | 2021-05-25 14:53:15 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | UI improvements for reordering columns in the data grid. | experience: ux improvements: quality needs: estimate priority: low type: perfective | Tested with build:v3.0.0-b4
URL used for test : http://fgpv.cloudapp.net/demo/v3.0.0-b4/prod/samples/index-fgp-en.html?keys=JOSM
To improve the user interface, the cell of the column selected for reordering in the data grid should always be in focus. The cell must be in focus, including the column label and reorder buttons. In the current version only some of the reordered cells can be visible, or the cell can be completely hidden during reordering, and the user must use the scroll bar to get to the cell selected for reordering.
| 1.0 | UI improvements for reordering columns in the data grid. - Tested with build:v3.0.0-b4
URL used for test : http://fgpv.cloudapp.net/demo/v3.0.0-b4/prod/samples/index-fgp-en.html?keys=JOSM
To improve the user interface, the cell of the column selected for reordering in the data grid should always be in focus. The cell must be in focus, including the column label and reorder buttons. In the current version only some of the reordered cells can be visible, or the cell can be completely hidden during reordering, and the user must use the scroll bar to get to the cell selected for reordering.
| priority | ui improvements for reordering columns in the data grid tested with build url used for test to improve the user interface the cell of the column selected for reordering in the data grid should always be in focus the cell must be in focus including the column label and reorder buttons in the current version only some of the reordered cells can be visible or the cell can be completely hidden during reordering and the user must use the scroll bar to get to the cell selected for reordering | 1 |
629,047 | 20,021,997,965 | IssuesEvent | 2022-02-01 17:13:02 | onicagroup/runway | https://api.github.com/repos/onicagroup/runway | closed | [BUG] runway incorrectly shows parameters to be removed in "plan" | bug priority:low | ### Bug Description
When running a cloudformation plan on an existing stack, the output states that parameters are to be removed:
```console
phil4079@L4CLYSQ2:~/runway-cloudformation-bug-test$ DEPLOY_ENVIRONMENT=common runway plan
[runway] could not find runway.variables.yml or runway.variables.yaml in the current directory; continuing without a variables file
[runway] deploy environment "common" is explicitly defined in the environment
[runway] if not correct, update the value or unset it to fall back to the name of the current git branch or parent directory
[runway]
[runway]
[runway] deployment_1:processing deployment (in progress)
[runway] deployment_1:processing regions sequentially...
[runway]
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (in progress)
[runway] found environment file: /home/phil4079/runway-cloudformation-bug-test/runway-cfn-test.cfn/common.env
[runway] stack-definition.yaml:plan (in progress)
[runway] diffing stacks: runway-cfn-test-stack
[runway] philcorp-dev-runway-cfn-test-stack changes:
Parameters Removed: InstanceType, SSHLocation, WebServerCount
Replacements:
- Modify EC2SRTA14Y5Q (AWS::EC2::SubnetRouteTableAssociation)
- Modify PublicSubnet (AWS::EC2::Subnet)
Changes:
- Modify ASGPolicy (AWS::AutoScaling::ScalingPolicy)
- Modify PublicElasticLoadBalancer (AWS::ElasticLoadBalancing::LoadBalancer)
- Modify WebServerFleet (AWS::AutoScaling::AutoScalingGroup)
Show full change set? [y/n] n
[runway] RawTemplateBlueprint.get_output_definitions is deprecated and will be removed in a future release
[runway] runway-cfn-test-stack:complete
[runway] stack-definition.yaml:plan (complete)
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (complete)
[runway] deployment_1:processing deployment (complete)
```
However, when running the same stack in deploy mode, these parameters are NOT listed for removal (which is the correct behavior):
```console
phil4079@L4CLYSQ2:~/runway-cloudformation-bug-test$ DEPLOY_ENVIRONMENT=common runway deploy
[runway] could not find runway.variables.yml or runway.variables.yaml in the current directory; continuing without a variables file
[runway] deploy environment "common" is explicitly defined in the environment
[runway] if not correct, update the value or unset it to fall back to the name of the current git branch or parent directory
[runway]
[runway]
[runway] deployment_1:processing deployment (in progress)
[runway] deployment_1:processing regions sequentially...
[runway]
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (in progress)
[runway] found environment file: /home/phil4079/runway-cloudformation-bug-test/runway-cfn-test.cfn/common.env
[runway] stack-definition.yaml:init (in progress)
[runway] cfngin_bucket cfngin-philcorp-dev-us-east-1 already exists
[runway] stack-definition.yaml:init (complete)
[runway] stack-definition.yaml:deploy (in progress)
[runway] philcorp-dev-runway-cfn-test-stack changes:
Replacements:
- Modify EC2SRTA14Y5Q (AWS::EC2::SubnetRouteTableAssociation)
- Modify PublicSubnet (AWS::EC2::Subnet)
Changes:
- Modify ASGPolicy (AWS::AutoScaling::ScalingPolicy)
- Modify PublicElasticLoadBalancer (AWS::ElasticLoadBalancing::LoadBalancer)
- Modify WebServerFleet (AWS::AutoScaling::AutoScalingGroup)
Execute the above changes? [y/n/v] n
[runway] runway-cfn-test-stack:skipped (canceled execution)
[runway] stack-definition.yaml:deploy (complete)
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (complete)
[runway] deployment_1:processing deployment (complete)
```
VIewing the change set in AWS console does not show any parameters to be removed (see attached).
[runway-cloudformation-bug-test-stack-20220128.txt](https://github.com/onicagroup/runway/files/7961947/runway-cloudformation-bug-test-stack-20220128.txt)
[runway-cloudformation-bug-test-change-set-20220128.txt](https://github.com/onicagroup/runway/files/7961967/runway-cloudformation-bug-test-change-set-20220128.txt)
### Expected Behavior
Both "plan" and "deploy" should not show any parameters being removed.
### Steps To Reproduce
Example project: ...
1. Deploy cloudformation stack with runway
2. Make a change to the stack
3. Perform a runway "plan" and "deploy" to see the output discrepancy
### Runway version
2.4.4
### Installation Type
pypi (pip, pipenv, poetry, etc)
### OS / Environment
- OS: Windows 10
- runway: 2.4.4
### Anything else?
_No response_ | 1.0 | [BUG] runway incorrectly shows parameters to be removed in "plan" - ### Bug Description
When running a cloudformation plan on an existing stack, the output states that parameters are to be removed:
```console
phil4079@L4CLYSQ2:~/runway-cloudformation-bug-test$ DEPLOY_ENVIRONMENT=common runway plan
[runway] could not find runway.variables.yml or runway.variables.yaml in the current directory; continuing without a variables file
[runway] deploy environment "common" is explicitly defined in the environment
[runway] if not correct, update the value or unset it to fall back to the name of the current git branch or parent directory
[runway]
[runway]
[runway] deployment_1:processing deployment (in progress)
[runway] deployment_1:processing regions sequentially...
[runway]
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (in progress)
[runway] found environment file: /home/phil4079/runway-cloudformation-bug-test/runway-cfn-test.cfn/common.env
[runway] stack-definition.yaml:plan (in progress)
[runway] diffing stacks: runway-cfn-test-stack
[runway] philcorp-dev-runway-cfn-test-stack changes:
Parameters Removed: InstanceType, SSHLocation, WebServerCount
Replacements:
- Modify EC2SRTA14Y5Q (AWS::EC2::SubnetRouteTableAssociation)
- Modify PublicSubnet (AWS::EC2::Subnet)
Changes:
- Modify ASGPolicy (AWS::AutoScaling::ScalingPolicy)
- Modify PublicElasticLoadBalancer (AWS::ElasticLoadBalancing::LoadBalancer)
- Modify WebServerFleet (AWS::AutoScaling::AutoScalingGroup)
Show full change set? [y/n] n
[runway] RawTemplateBlueprint.get_output_definitions is deprecated and will be removed in a future release
[runway] runway-cfn-test-stack:complete
[runway] stack-definition.yaml:plan (complete)
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (complete)
[runway] deployment_1:processing deployment (complete)
```
However, when running the same stack in deploy mode, these parameters are NOT listed for removal (which is the correct behavior):
```console
phil4079@L4CLYSQ2:~/runway-cloudformation-bug-test$ DEPLOY_ENVIRONMENT=common runway deploy
[runway] could not find runway.variables.yml or runway.variables.yaml in the current directory; continuing without a variables file
[runway] deploy environment "common" is explicitly defined in the environment
[runway] if not correct, update the value or unset it to fall back to the name of the current git branch or parent directory
[runway]
[runway]
[runway] deployment_1:processing deployment (in progress)
[runway] deployment_1:processing regions sequentially...
[runway]
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (in progress)
[runway] found environment file: /home/phil4079/runway-cloudformation-bug-test/runway-cfn-test.cfn/common.env
[runway] stack-definition.yaml:init (in progress)
[runway] cfngin_bucket cfngin-philcorp-dev-us-east-1 already exists
[runway] stack-definition.yaml:init (complete)
[runway] stack-definition.yaml:deploy (in progress)
[runway] philcorp-dev-runway-cfn-test-stack changes:
Replacements:
- Modify EC2SRTA14Y5Q (AWS::EC2::SubnetRouteTableAssociation)
- Modify PublicSubnet (AWS::EC2::Subnet)
Changes:
- Modify ASGPolicy (AWS::AutoScaling::ScalingPolicy)
- Modify PublicElasticLoadBalancer (AWS::ElasticLoadBalancing::LoadBalancer)
- Modify WebServerFleet (AWS::AutoScaling::AutoScalingGroup)
Execute the above changes? [y/n/v] n
[runway] runway-cfn-test-stack:skipped (canceled execution)
[runway] stack-definition.yaml:deploy (complete)
[runway] deployment_1.runway-cfn-test.cfn:processing module in us-east-1 (complete)
[runway] deployment_1:processing deployment (complete)
```
VIewing the change set in AWS console does not show any parameters to be removed (see attached).
[runway-cloudformation-bug-test-stack-20220128.txt](https://github.com/onicagroup/runway/files/7961947/runway-cloudformation-bug-test-stack-20220128.txt)
[runway-cloudformation-bug-test-change-set-20220128.txt](https://github.com/onicagroup/runway/files/7961967/runway-cloudformation-bug-test-change-set-20220128.txt)
### Expected Behavior
Both "plan" and "deploy" should not show any parameters being removed.
### Steps To Reproduce
Example project: ...
1. Deploy cloudformation stack with runway
2. Make a change to the stack
3. Perform a runway "plan" and "deploy" to see the output discrepancy
### Runway version
2.4.4
### Installation Type
pypi (pip, pipenv, poetry, etc)
### OS / Environment
- OS: Windows 10
- runway: 2.4.4
### Anything else?
_No response_ | priority | runway incorrectly shows parameters to be removed in plan bug description when running a cloudformation plan on an existing stack the output states that parameters are to be removed console runway cloudformation bug test deploy environment common runway plan could not find runway variables yml or runway variables yaml in the current directory continuing without a variables file deploy environment common is explicitly defined in the environment if not correct update the value or unset it to fall back to the name of the current git branch or parent directory deployment processing deployment in progress deployment processing regions sequentially deployment runway cfn test cfn processing module in us east in progress found environment file home runway cloudformation bug test runway cfn test cfn common env stack definition yaml plan in progress diffing stacks runway cfn test stack philcorp dev runway cfn test stack changes parameters removed instancetype sshlocation webservercount replacements modify aws subnetroutetableassociation modify publicsubnet aws subnet changes modify asgpolicy aws autoscaling scalingpolicy modify publicelasticloadbalancer aws elasticloadbalancing loadbalancer modify webserverfleet aws autoscaling autoscalinggroup show full change set n rawtemplateblueprint get output definitions is deprecated and will be removed in a future release runway cfn test stack complete stack definition yaml plan complete deployment runway cfn test cfn processing module in us east complete deployment processing deployment complete however when running the same stack in deploy mode these parameters are not listed for removal which is the correct behavior console runway cloudformation bug test deploy environment common runway deploy could not find runway variables yml or runway variables yaml in the current directory continuing without a variables file deploy environment common is explicitly defined in the environment if not correct update the value or unset it to fall back to the name of the current git branch or parent directory deployment processing deployment in progress deployment processing regions sequentially deployment runway cfn test cfn processing module in us east in progress found environment file home runway cloudformation bug test runway cfn test cfn common env stack definition yaml init in progress cfngin bucket cfngin philcorp dev us east already exists stack definition yaml init complete stack definition yaml deploy in progress philcorp dev runway cfn test stack changes replacements modify aws subnetroutetableassociation modify publicsubnet aws subnet changes modify asgpolicy aws autoscaling scalingpolicy modify publicelasticloadbalancer aws elasticloadbalancing loadbalancer modify webserverfleet aws autoscaling autoscalinggroup execute the above changes n runway cfn test stack skipped canceled execution stack definition yaml deploy complete deployment runway cfn test cfn processing module in us east complete deployment processing deployment complete viewing the change set in aws console does not show any parameters to be removed see attached expected behavior both plan and deploy should not show any parameters being removed steps to reproduce example project deploy cloudformation stack with runway make a change to the stack perform a runway plan and deploy to see the output discrepancy runway version installation type pypi pip pipenv poetry etc os environment os windows runway anything else no response | 1 |
730,914 | 25,194,363,396 | IssuesEvent | 2022-11-12 10:05:58 | tsutaj/statements-manager | https://api.github.com/repos/tsutaj/statements-manager | closed | 出力する PDF をいろいろ装飾したい | Priority: Low | とりあえず必要なのはこのへん?
- 表紙
- header
- コンテスト名とか付けたいよね
- footer
- 現在のページ / 全ページ みたいなページャーとか | 1.0 | 出力する PDF をいろいろ装飾したい - とりあえず必要なのはこのへん?
- 表紙
- header
- コンテスト名とか付けたいよね
- footer
- 現在のページ / 全ページ みたいなページャーとか | priority | 出力する pdf をいろいろ装飾したい とりあえず必要なのはこのへん? 表紙 header コンテスト名とか付けたいよね footer 現在のページ 全ページ みたいなページャーとか | 1 |
197,964 | 6,967,234,913 | IssuesEvent | 2017-12-10 05:42:21 | facelessuser/pymdown-extensions | https://api.github.com/repos/facelessuser/pymdown-extensions | closed | Take link styling a bit further | Feature Maybe Pass On Priority - Low | So repository links have specific classes attached to them, but it may be nice to have an option that takes things a bit further. If we add an option that strips out the special character (`#`, `!`, and `@`) and wraps content like so:
```html
<a href"#" class="magiclink magiclink-github magiclink-mention"><span class="magiclink-username">name</span></a>
<a href"#" class="magiclink magiclink-gtihub magiclink-repository"><span class="magiclink-reponame">name/repo</span></a>
<a href"#" class="magiclink magiclink-github magiclink-issue"><span class="magiclink-reponame">name/repo</span><span class="magiclink-value">33</span></a>
```
Then you could easily use `:before` attributes to add a bug icon to issue numbers, the corresponding provider icon for reponames, maybe even use something like a user icon instead of `@`. The possibilities open up. Wish I'd thought of this before the last release.
| 1.0 | Take link styling a bit further - So repository links have specific classes attached to them, but it may be nice to have an option that takes things a bit further. If we add an option that strips out the special character (`#`, `!`, and `@`) and wraps content like so:
```html
<a href"#" class="magiclink magiclink-github magiclink-mention"><span class="magiclink-username">name</span></a>
<a href"#" class="magiclink magiclink-gtihub magiclink-repository"><span class="magiclink-reponame">name/repo</span></a>
<a href"#" class="magiclink magiclink-github magiclink-issue"><span class="magiclink-reponame">name/repo</span><span class="magiclink-value">33</span></a>
```
Then you could easily use `:before` attributes to add a bug icon to issue numbers, the corresponding provider icon for reponames, maybe even use something like a user icon instead of `@`. The possibilities open up. Wish I'd thought of this before the last release.
| priority | take link styling a bit further so repository links have specific classes attached to them but it may be nice to have an option that takes things a bit further if we add an option that strips out the special character and and wraps content like so html name name repo name repo then you could easily use before attributes to add a bug icon to issue numbers the corresponding provider icon for reponames maybe even use something like a user icon instead of the possibilities open up wish i d thought of this before the last release | 1 |
805,407 | 29,519,400,646 | IssuesEvent | 2023-06-04 22:00:47 | khoulihan/godot4-cutscene-graph-editor | https://api.github.com/repos/khoulihan/godot4-cutscene-graph-editor | opened | Implement the arrange nodes function | enhancement priority:low | There is a new function in the graph editor in Godot 4 to request that the graph be rearranged automatically. Would be nice to implement this. There should be a warning. | 1.0 | Implement the arrange nodes function - There is a new function in the graph editor in Godot 4 to request that the graph be rearranged automatically. Would be nice to implement this. There should be a warning. | priority | implement the arrange nodes function there is a new function in the graph editor in godot to request that the graph be rearranged automatically would be nice to implement this there should be a warning | 1 |
427,509 | 12,396,155,945 | IssuesEvent | 2020-05-20 19:59:59 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User | Priority: Medium Product: caseflow-queue Team: Echo 🐬 Type: Bug | ## Description
The Assigned To field should display the correct user ID on the search cases page
## Acceptance criteria
- [ ] The assigned to field on the search cases page displays the correct user assigned to the case
## Background/context/resources
During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again
Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots
## Technical notes
Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search.
## Screenshots

| 1.0 | Case Movement_Non-Blocking Tasks | Assigned To Field Not Displaying Correct User - ## Description
The Assigned To field should display the correct user ID on the search cases page
## Acceptance criteria
- [ ] The assigned to field on the search cases page displays the correct user assigned to the case
## Background/context/resources
During Dogfooding: Assigning a case to an judge (Legacy), reassigning to Burnham assigned case to ID4 instead of CSS_ID29 when searching the case again
Check the [dogfooding google doc](https://docs.google.com/document/d/1JbCt-TYW3wmLDpFbLoxFxMeX8YAlUauq3viaJm_A_J4/edit?usp=sharing) for screenshots
## Technical notes
Reproduce before digging into. Hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs. This would leave "Assigned to" unchanged in case search.
## Screenshots

| priority | case movement non blocking tasks assigned to field not displaying correct user description the assigned to field should display the correct user id on the search cases page acceptance criteria the assigned to field on the search cases page displays the correct user assigned to the case background context resources during dogfooding assigning a case to an judge legacy reassigning to burnham assigned case to instead of css when searching the case again check the for screenshots technical notes reproduce before digging into hunter believes that this is due to being unable to assign legacy cases in demo and the lack of an error shown when this occurs this would leave assigned to unchanged in case search screenshots | 1 |
582,148 | 17,354,313,282 | IssuesEvent | 2021-07-29 12:44:35 | netdata/netdata | https://api.github.com/repos/netdata/netdata | closed | Add support for CRUX with kickstart.sh / install-required-packages.sh | area/ci area/packaging feature request priority/low | > I know there are very likely few folks that actually run [CRUX](https://crux.nu) in production environments; but hey I do on my NAS 😀
##### Feature idea summary
As per title.
##### Expected behavior
- Kickstart installer works.
- Install required packages works.
----
Currently this fails with:
```sh
prologic@Jamess-iMac
Thu Jan 30 14:23:45
~/NetData/netdata
(centos8_deps_2) 255
$ dki -t --rm -v $PWD:/netdata -w /netdata crux /bin/sh -c './packaging/installer/install-required-packages.sh --dont-wait --non-interactive'
Cannot find /etc/os-release
Cannot find valid distribution with lsb-release
I NEED YOUR HELP
It seems I cannot detect your system automatically.
And it seems I cannot find a known package manager in this system.
Please open a github issue to help us support your system too.
```
FYI:
- Package manager on CRUX is: `ports` and `prt-get`
- Arch Linux is a direct derivative of CRUX (history lesson_) 😀-- Which is why `PKGBUILD`(s) are so similar to `Pkgfile`(s) 🤣 | 1.0 | Add support for CRUX with kickstart.sh / install-required-packages.sh - > I know there are very likely few folks that actually run [CRUX](https://crux.nu) in production environments; but hey I do on my NAS 😀
##### Feature idea summary
As per title.
##### Expected behavior
- Kickstart installer works.
- Install required packages works.
----
Currently this fails with:
```sh
prologic@Jamess-iMac
Thu Jan 30 14:23:45
~/NetData/netdata
(centos8_deps_2) 255
$ dki -t --rm -v $PWD:/netdata -w /netdata crux /bin/sh -c './packaging/installer/install-required-packages.sh --dont-wait --non-interactive'
Cannot find /etc/os-release
Cannot find valid distribution with lsb-release
I NEED YOUR HELP
It seems I cannot detect your system automatically.
And it seems I cannot find a known package manager in this system.
Please open a github issue to help us support your system too.
```
FYI:
- Package manager on CRUX is: `ports` and `prt-get`
- Arch Linux is a direct derivative of CRUX (history lesson_) 😀-- Which is why `PKGBUILD`(s) are so similar to `Pkgfile`(s) 🤣 | priority | add support for crux with kickstart sh install required packages sh i know there are very likely few folks that actually run in production environments but hey i do on my nas 😀 feature idea summary as per title expected behavior kickstart installer works install required packages works currently this fails with sh prologic jamess imac thu jan netdata netdata deps dki t rm v pwd netdata w netdata crux bin sh c packaging installer install required packages sh dont wait non interactive cannot find etc os release cannot find valid distribution with lsb release i need your help it seems i cannot detect your system automatically and it seems i cannot find a known package manager in this system please open a github issue to help us support your system too fyi package manager on crux is ports and prt get arch linux is a direct derivative of crux history lesson 😀 which is why pkgbuild s are so similar to pkgfile s 🤣 | 1 |
733,641 | 25,315,569,169 | IssuesEvent | 2022-11-17 21:18:00 | NoppesTheFolf/Fluffle | https://api.github.com/repos/NoppesTheFolf/Fluffle | opened | Find more artists on Twitter by scraping for furry hashtags | feature priority:low | It would be possible to find more accounts posting furry art if we were to look for hashtags related to the furry fandom (like #furryart). | 1.0 | Find more artists on Twitter by scraping for furry hashtags - It would be possible to find more accounts posting furry art if we were to look for hashtags related to the furry fandom (like #furryart). | priority | find more artists on twitter by scraping for furry hashtags it would be possible to find more accounts posting furry art if we were to look for hashtags related to the furry fandom like furryart | 1 |
566,268 | 16,817,125,260 | IssuesEvent | 2021-06-17 08:43:05 | oceanprotocol/ocean-faucet | https://api.github.com/repos/oceanprotocol/ocean-faucet | opened | Migrate to Typescript | Priority: Low Type: Enhancement | At some point we should move this repo to Typescript to make it consistent with all of our other repositories. | 1.0 | Migrate to Typescript - At some point we should move this repo to Typescript to make it consistent with all of our other repositories. | priority | migrate to typescript at some point we should move this repo to typescript to make it consistent with all of our other repositories | 1 |
343,203 | 10,326,285,493 | IssuesEvent | 2019-09-02 01:04:04 | tiagoalvesdulce/pi-ui | https://api.github.com/repos/tiagoalvesdulce/pi-ui | closed | Message: align left icon and the first line of text | component-improvement priority: low type: display | Currently, the icon is positioned a bit higher than the first line. They should be vertically-center aligned. | 1.0 | Message: align left icon and the first line of text - Currently, the icon is positioned a bit higher than the first line. They should be vertically-center aligned. | priority | message align left icon and the first line of text currently the icon is positioned a bit higher than the first line they should be vertically center aligned | 1 |
111,915 | 4,494,695,115 | IssuesEvent | 2016-08-31 07:26:40 | TrinityCore/TrinityCore | https://api.github.com/repos/TrinityCore/TrinityCore | closed | [6.x] Who list not working (opcode disabled) | Branch-6.x Priority-Low | When someone clicks on Refresh in the Who list nothing happens.
Worldserver.exe Console shows:
Prevented sending disabled opcode [SMSG_WHO 0x2BAE (11182)] to [Player: Mike GUID Full: 0x08000400000000000000000000000002 Type: Player Entry: 0 Low: 2, Account: 2]
53aecd6
windows 10
tdb 6.04 with updates from trinitycore autoupdate | 1.0 | [6.x] Who list not working (opcode disabled) - When someone clicks on Refresh in the Who list nothing happens.
Worldserver.exe Console shows:
Prevented sending disabled opcode [SMSG_WHO 0x2BAE (11182)] to [Player: Mike GUID Full: 0x08000400000000000000000000000002 Type: Player Entry: 0 Low: 2, Account: 2]
53aecd6
windows 10
tdb 6.04 with updates from trinitycore autoupdate | priority | who list not working opcode disabled when someone clicks on refresh in the who list nothing happens worldserver exe console shows prevented sending disabled opcode to windows tdb with updates from trinitycore autoupdate | 1 |
673,666 | 23,026,459,415 | IssuesEvent | 2022-07-22 09:40:02 | apache/incubator-devlake | https://api.github.com/repos/apache/incubator-devlake | closed | Extract file-level data from git repository | type/feature-request priority/medium need-follow-up | ## User story
As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution.
## Goal
The key is to match `bug` to `component`.
- If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`;
- If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`;
## Proposed Solution
- [ ] Add a `commit_files` table in the domain layer
| field | type | length | description | key |
| ------------- | -------- | -------- | ----------- | -----|
| `commit_sha` | char | 40 | | FK |
| `file_name` | varchar | 255 | | |
| `status` | varchar | 255 | | |
| `additions` | int | | | |
| `deletions` | int | | | |
| `changes` | int | | | |
| `file_component` | varchar | 255 | | |
a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.**
b. how to assign `file_component` value to a commit file? It varies in different programming languages.
- In Go, a component can be found by folder names;
- while in other languages, it might need user input to define the `file_component`.
- [ ] Allow users to define the `file_component` in `.env` by REGEX
## An example of commit files
For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files:

## Belongs to issue
This is a sub-issue of both #949 and #1771
## Additional context
N/A
| 1.0 | Extract file-level data from git repository - ## User story
As a tech lead, I want to see the bug distribution in low-level components (from code files), as a guide for the future framework evolution.
## Goal
The key is to match `bug` to `component`.
- If we want to find the overall distribution of all bugs, we can do: `issues(bug)` join `prs` join `commits` join `commit_files(component)`;
- If we want to find the overall distribution of bugs fixed in a specific tag, we can do: `ref_diff_issues(bug)` join `prs` join `commits` join `commit_files(component)`;
## Proposed Solution
- [ ] Add a `commit_files` table in the domain layer
| field | type | length | description | key |
| ------------- | -------- | -------- | ----------- | -----|
| `commit_sha` | char | 40 | | FK |
| `file_name` | varchar | 255 | | |
| `status` | varchar | 255 | | |
| `additions` | int | | | |
| `deletions` | int | | | |
| `changes` | int | | | |
| `file_component` | varchar | 255 | | |
a. all the columns except `file_component` can be retrieved from [gitExtractor](https://github.com/apache/incubator-devlake/tree/main/plugins/refdiff). **Note: not from [Github API](https://docs.github.com/en/rest/commits/commits#get-a-commit) or Gitlab API.**
b. how to assign `file_component` value to a commit file? It varies in different programming languages.
- In Go, a component can be found by folder names;
- while in other languages, it might need user input to define the `file_component`.
- [ ] Allow users to define the `file_component` in `.env` by REGEX
## An example of commit files
For instance, this [commit](https://github.com/apache/incubator-devlake/pull/2039/commits/aeec784efc4c21e9e92a414e3008987cb7f791fa) touches 4 files:

## Belongs to issue
This is a sub-issue of both #949 and #1771
## Additional context
N/A
| priority | extract file level data from git repository user story as a tech lead i want to see the bug distribution in low level components from code files as a guide for the future framework evolution goal the key is to match bug to component if we want to find the overall distribution of all bugs we can do issues bug join prs join commits join commit files component if we want to find the overall distribution of bugs fixed in a specific tag we can do ref diff issues bug join prs join commits join commit files component proposed solution add a commit files table in the domain layer field type length description key commit sha char fk file name varchar status varchar additions int deletions int changes int file component varchar a all the columns except file component can be retrieved from note not from or gitlab api b how to assign file component value to a commit file it varies in different programming languages in go a component can be found by folder names while in other languages it might need user input to define the file component allow users to define the file component in env by regex an example of commit files for instance this touches files belongs to issue this is a sub issue of both and additional context n a | 1 |
124,012 | 4,890,577,516 | IssuesEvent | 2016-11-18 14:20:00 | reminisceme/app | https://api.github.com/repos/reminisceme/app | opened | [Feature proposition] Let the users delete games | client enhancement priority-low server | Having old games that will never be finished (because they are bugged against a bot or because someone stopped playing) clutters the interface. Letting a user delete (which would make it invisible to them but we can keep a trace of it in the database) seems to me like a possible approach to solve the problem. | 1.0 | [Feature proposition] Let the users delete games - Having old games that will never be finished (because they are bugged against a bot or because someone stopped playing) clutters the interface. Letting a user delete (which would make it invisible to them but we can keep a trace of it in the database) seems to me like a possible approach to solve the problem. | priority | let the users delete games having old games that will never be finished because they are bugged against a bot or because someone stopped playing clutters the interface letting a user delete which would make it invisible to them but we can keep a trace of it in the database seems to me like a possible approach to solve the problem | 1 |
532,859 | 15,572,193,827 | IssuesEvent | 2021-03-17 06:35:46 | Azure/static-web-apps-cli | https://api.github.com/repos/Azure/static-web-apps-cli | reopened | Warning about workflow not found can be confusing | priority: low | **Describe the bug**
When there's no workflow file in the repo, the CLI prints a warning.
```
INFO: Azure Static Web Apps configuration not found under ".github/workflows/"
```
This might mislead the user to believe that a workflow is required. Suggest removing the message. Instead, show a message that config was loaded from the workflow (it's already doing this). | 1.0 | Warning about workflow not found can be confusing - **Describe the bug**
When there's no workflow file in the repo, the CLI prints a warning.
```
INFO: Azure Static Web Apps configuration not found under ".github/workflows/"
```
This might mislead the user to believe that a workflow is required. Suggest removing the message. Instead, show a message that config was loaded from the workflow (it's already doing this). | priority | warning about workflow not found can be confusing describe the bug when there s no workflow file in the repo the cli prints a warning info azure static web apps configuration not found under github workflows this might mislead the user to believe that a workflow is required suggest removing the message instead show a message that config was loaded from the workflow it s already doing this | 1 |
509,943 | 14,762,511,470 | IssuesEvent | 2021-01-09 04:06:26 | PazerOP/tf2_bot_detector | https://api.github.com/repos/PazerOP/tf2_bot_detector | opened | Create a tag specifically for bots, separeted from cheating players. | Priority: Low Type: Enhancement | **A clear and concise description of what the problem is.**
As [discussed](https://discord.com/channels/716525494421553243/716525708771328040/797301570043510784) in the tool's discord, some users have objections to using this tool against human cheaters because they might not be cheating at the moment. Also, even when desired, kicking a cheater as soon as it enters might be hard in some situations, specially if the cheater protests, which does not happen with bots.
**Describe the solution you'd like**
Separate "cheaters" into "cheaters" and "bots", where cheaters will only contain human players.
**Describe alternatives you've considered**
I've also considered doing away with the "exploiter" tag, as most users barely use it and I don't see exploits as a big issue currently.
I've made a separate thread to discuss this change here: **(THREAD STILL BEING WRITTEN)**
| 1.0 | Create a tag specifically for bots, separeted from cheating players. - **A clear and concise description of what the problem is.**
As [discussed](https://discord.com/channels/716525494421553243/716525708771328040/797301570043510784) in the tool's discord, some users have objections to using this tool against human cheaters because they might not be cheating at the moment. Also, even when desired, kicking a cheater as soon as it enters might be hard in some situations, specially if the cheater protests, which does not happen with bots.
**Describe the solution you'd like**
Separate "cheaters" into "cheaters" and "bots", where cheaters will only contain human players.
**Describe alternatives you've considered**
I've also considered doing away with the "exploiter" tag, as most users barely use it and I don't see exploits as a big issue currently.
I've made a separate thread to discuss this change here: **(THREAD STILL BEING WRITTEN)**
| priority | create a tag specifically for bots separeted from cheating players a clear and concise description of what the problem is as in the tool s discord some users have objections to using this tool against human cheaters because they might not be cheating at the moment also even when desired kicking a cheater as soon as it enters might be hard in some situations specially if the cheater protests which does not happen with bots describe the solution you d like separate cheaters into cheaters and bots where cheaters will only contain human players describe alternatives you ve considered i ve also considered doing away with the exploiter tag as most users barely use it and i don t see exploits as a big issue currently i ve made a separate thread to discuss this change here thread still being written | 1 |
495,632 | 14,285,435,322 | IssuesEvent | 2020-11-23 13:55:09 | geocollections/sarv-edit | https://api.github.com/repos/geocollections/sarv-edit | closed | Drillcore box alternative list views | Difficulty: Medium Priority: Low Source: App Status: Available Type: Feature Request | Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared).
A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):
 

| 1.0 | Drillcore box alternative list views - Drillcore box filter/list view (https://edit2.geocollections.info/drillcore_box) should have separate view with box images, one for each box (if no image, placeholder should be shown), similar to file list view where images are shown within table. When clicking on box image, large image should be shown in separate popup window (that will remain open so that two or more box images can be compared).
A different view is needed for printing box images into pdf (two landscape images per portrait A4 sheet):
 

| priority | drillcore box alternative list views drillcore box filter list view should have separate view with box images one for each box if no image placeholder should be shown similar to file list view where images are shown within table when clicking on box image large image should be shown in separate popup window that will remain open so that two or more box images can be compared a different view is needed for printing box images into pdf two landscape images per portrait sheet   | 1 |
3,382 | 2,537,868,461 | IssuesEvent | 2015-01-26 23:41:05 | embolalia/willie | https://api.github.com/repos/embolalia/willie | closed | .seen shows messages from private channels | Bug Low Priority | Submitted by: Embolalia
From channel: #Embo
At 2013-02-21 21:58:58 UTC | 1.0 | .seen shows messages from private channels - Submitted by: Embolalia
From channel: #Embo
At 2013-02-21 21:58:58 UTC | priority | seen shows messages from private channels submitted by embolalia from channel embo at utc | 1 |
380,023 | 11,253,345,107 | IssuesEvent | 2020-01-11 15:40:33 | William-Lake/NLP-API | https://api.github.com/repos/William-Lake/NLP-API | closed | Provide a history of the path actions in the response. | High Priority Low Urgency | E.g. if the entire requests fails or only one endpoint, return the appropriate feedback to the user. | 1.0 | Provide a history of the path actions in the response. - E.g. if the entire requests fails or only one endpoint, return the appropriate feedback to the user. | priority | provide a history of the path actions in the response e g if the entire requests fails or only one endpoint return the appropriate feedback to the user | 1 |
439,134 | 12,678,095,250 | IssuesEvent | 2020-06-19 09:08:25 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | AutoFitColumn in multi-columns headers grid with locked column is not working | Bug C: Grid FP: Completed Kendo1 Next LIB Priority 1 SEV: Low | ### Bug report
AutoFitColumn in multi-columns headers grid with locked column is not working
### Reproduction of the problem
https://dojo.telerik.com/ehURUgIW/2
### Current behavior
autoFitColumn is not working in this scenario.
### Expected/desired behavior
autoFitColumn to be working
### Environment
* **Kendo UI version:** 2020.1.406
* **Browser:** [all]
| 1.0 | AutoFitColumn in multi-columns headers grid with locked column is not working - ### Bug report
AutoFitColumn in multi-columns headers grid with locked column is not working
### Reproduction of the problem
https://dojo.telerik.com/ehURUgIW/2
### Current behavior
autoFitColumn is not working in this scenario.
### Expected/desired behavior
autoFitColumn to be working
### Environment
* **Kendo UI version:** 2020.1.406
* **Browser:** [all]
| priority | autofitcolumn in multi columns headers grid with locked column is not working bug report autofitcolumn in multi columns headers grid with locked column is not working reproduction of the problem current behavior autofitcolumn is not working in this scenario expected desired behavior autofitcolumn to be working environment kendo ui version browser | 1 |
148,033 | 5,657,915,189 | IssuesEvent | 2017-04-10 08:40:38 | HGustavs/LenaSYS | https://api.github.com/repos/HGustavs/LenaSYS | closed | Code cleanup in resulted.js | lowPriority | Fix code formatting in resulted.js to make it more consistent and easier to understand. **Attention:** This issue has been changed. | 1.0 | Code cleanup in resulted.js - Fix code formatting in resulted.js to make it more consistent and easier to understand. **Attention:** This issue has been changed. | priority | code cleanup in resulted js fix code formatting in resulted js to make it more consistent and easier to understand attention this issue has been changed | 1 |
293,064 | 8,972,370,694 | IssuesEvent | 2019-01-29 18:06:19 | TeamTwilight/twilightforest | https://api.github.com/repos/TeamTwilight/twilightforest | closed | Biome Weather Effects persist with tfEnforcedProgression set to False | bug priority-low | Forge version: 14.23.4.2759
Twilight Forest version: 3.8.716
Steps to reproduce:
1. Set `tfEnforcedProgression` to `false`
2. Run around the biomes that are affected by Progression, making sure they are locked
What I expected to happen:
Weather effects shouldn't render.
What happened instead:
It seems that the weather effects continue to play out, despite having set `tfEnforcedProgression` to `false`. The structure protection still plays out normally (as tested with a Lich Tower before having slain a Naga). The weather effects do not induce their effects when walked in to, so it appears to just be a visual discrepancy.
| 1.0 | Biome Weather Effects persist with tfEnforcedProgression set to False - Forge version: 14.23.4.2759
Twilight Forest version: 3.8.716
Steps to reproduce:
1. Set `tfEnforcedProgression` to `false`
2. Run around the biomes that are affected by Progression, making sure they are locked
What I expected to happen:
Weather effects shouldn't render.
What happened instead:
It seems that the weather effects continue to play out, despite having set `tfEnforcedProgression` to `false`. The structure protection still plays out normally (as tested with a Lich Tower before having slain a Naga). The weather effects do not induce their effects when walked in to, so it appears to just be a visual discrepancy.
| priority | biome weather effects persist with tfenforcedprogression set to false forge version twilight forest version steps to reproduce set tfenforcedprogression to false run around the biomes that are affected by progression making sure they are locked what i expected to happen weather effects shouldn t render what happened instead it seems that the weather effects continue to play out despite having set tfenforcedprogression to false the structure protection still plays out normally as tested with a lich tower before having slain a naga the weather effects do not induce their effects when walked in to so it appears to just be a visual discrepancy | 1 |
382,191 | 11,301,950,027 | IssuesEvent | 2020-01-17 16:35:45 | bubblesub/bubblesub | https://api.github.com/repos/bubblesub/bubblesub | closed | Aliasing on the bubblesub icon | bug low priority | On some small screens, the bubblesub icon presents an aliasing effect.
It would be useful to provide icons at different dimensions, for example:
- 16x16
- 32x32
- 64x64
- 256x256 | 1.0 | Aliasing on the bubblesub icon - On some small screens, the bubblesub icon presents an aliasing effect.
It would be useful to provide icons at different dimensions, for example:
- 16x16
- 32x32
- 64x64
- 256x256 | priority | aliasing on the bubblesub icon on some small screens the bubblesub icon presents an aliasing effect it would be useful to provide icons at different dimensions for example | 1 |
93,668 | 3,907,583,083 | IssuesEvent | 2016-04-19 13:21:52 | softdevteam/krun | https://api.github.com/repos/softdevteam/krun | closed | Improving the "benchmark started" email | enhancement low priority (nice to have) | Since users will start a benchmark run and not want to ssh into the b/marking box again, it would be nice for the benchmark-started email to contain a few more details. This would make it easier to catch config mistakes after a run has started.
An easy fix for this issue would just be to email out the contents of the config file. | 1.0 | Improving the "benchmark started" email - Since users will start a benchmark run and not want to ssh into the b/marking box again, it would be nice for the benchmark-started email to contain a few more details. This would make it easier to catch config mistakes after a run has started.
An easy fix for this issue would just be to email out the contents of the config file. | priority | improving the benchmark started email since users will start a benchmark run and not want to ssh into the b marking box again it would be nice for the benchmark started email to contain a few more details this would make it easier to catch config mistakes after a run has started an easy fix for this issue would just be to email out the contents of the config file | 1 |
713,473 | 24,528,995,288 | IssuesEvent | 2022-10-11 15:01:49 | MTASA-TFF/The-Favoured-Few | https://api.github.com/repos/MTASA-TFF/The-Favoured-Few | closed | Idea/request: VenaDe - LOTTO For Money 1-100 | enhancement to be uploaded low priority | Hello again lotto script have add for Money number 1-100 write 1 hours in again lotto everything good day TFF! | 1.0 | Idea/request: VenaDe - LOTTO For Money 1-100 - Hello again lotto script have add for Money number 1-100 write 1 hours in again lotto everything good day TFF! | priority | idea request venade lotto for money hello again lotto script have add for money number write hours in again lotto everything good day tff | 1 |
202,221 | 7,045,430,767 | IssuesEvent | 2018-01-01 19:37:46 | xcodeswift/xcproj | https://api.github.com/repos/xcodeswift/xcproj | reopened | Improve parsing errors | difficulty:moderate priority:low status:wont-fix type:enhancement | ## Context 🕵️♀️
Right now, if the parsing fails for some reason, it throws an error saying that a property that was expected, wasn't there or it was in the wrong format. It'd be great to throw something more explicit like:
> We've found an element XXXX with a reference XXXX that is missing the property XXXX.
## What 🌱
Improve the error message.
## Proposal 🎉
> We've found an element XXXX with a reference XXXX that is missing the property XXXX. Your project might have got into an invalid state after a merging conflict. We recommend you to open it with Xcode, identify the issue and fix it before trying to open it with xcproj.
| 1.0 | Improve parsing errors - ## Context 🕵️♀️
Right now, if the parsing fails for some reason, it throws an error saying that a property that was expected, wasn't there or it was in the wrong format. It'd be great to throw something more explicit like:
> We've found an element XXXX with a reference XXXX that is missing the property XXXX.
## What 🌱
Improve the error message.
## Proposal 🎉
> We've found an element XXXX with a reference XXXX that is missing the property XXXX. Your project might have got into an invalid state after a merging conflict. We recommend you to open it with Xcode, identify the issue and fix it before trying to open it with xcproj.
| priority | improve parsing errors context 🕵️♀️ right now if the parsing fails for some reason it throws an error saying that a property that was expected wasn t there or it was in the wrong format it d be great to throw something more explicit like we ve found an element xxxx with a reference xxxx that is missing the property xxxx what 🌱 improve the error message proposal 🎉 we ve found an element xxxx with a reference xxxx that is missing the property xxxx your project might have got into an invalid state after a merging conflict we recommend you to open it with xcode identify the issue and fix it before trying to open it with xcproj | 1 |
733,104 | 25,288,092,316 | IssuesEvent | 2022-11-16 21:09:33 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Update the `Empty` blueprint Homepage text | enhancement priority: low | ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
The empty bp home page uses `site` instead of `project`
### Describe the solution you'd like
Update empty bp home page to use `project` in the text | 1.0 | [studio] Update the `Empty` blueprint Homepage text - ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
The empty bp home page uses `site` instead of `project`
### Describe the solution you'd like
Update empty bp home page to use `project` in the text | priority | update the empty blueprint homepage text duplicates i have searched the existing issues is your feature request related to a problem please describe the empty bp home page uses site instead of project describe the solution you d like update empty bp home page to use project in the text | 1 |
613,625 | 19,095,154,472 | IssuesEvent | 2021-11-29 15:58:13 | axelarnetwork/tofnd | https://api.github.com/repos/axelarnetwork/tofnd | opened | Better handling for multisig `KeyPresence` | low priority | Currently `KeyPresence` is blindly returning `Present` to maintain compatibility with axelar-core. That's a little unorthodox because the keys are not stored in the kv-store. If we must remain on having `KeyPresence` for multisig, consider adding a new response type for `KeyPresence` that indicate that keys are created on the fly. | 1.0 | Better handling for multisig `KeyPresence` - Currently `KeyPresence` is blindly returning `Present` to maintain compatibility with axelar-core. That's a little unorthodox because the keys are not stored in the kv-store. If we must remain on having `KeyPresence` for multisig, consider adding a new response type for `KeyPresence` that indicate that keys are created on the fly. | priority | better handling for multisig keypresence currently keypresence is blindly returning present to maintain compatibility with axelar core that s a little unorthodox because the keys are not stored in the kv store if we must remain on having keypresence for multisig consider adding a new response type for keypresence that indicate that keys are created on the fly | 1 |
121,340 | 4,807,761,529 | IssuesEvent | 2016-11-02 22:30:13 | LatvianModder/FTBUtilities | https://api.github.com/repos/LatvianModder/FTBUtilities | closed | [Request] Ability to disable sethome/back commands | Priority: Low Type: Suggestion / ToDo | Greetings,
I would like to humbly request the config option to disable all FTBLib/FTBUtils methods to /home, /set_home, /back commands.. basically anything that will teleport a non-admin player. Is this possible? It looks like /warp is easily taken care of by not setting any warps but I don't see a way to prevent the others.
Thanks so much.
| 1.0 | [Request] Ability to disable sethome/back commands - Greetings,
I would like to humbly request the config option to disable all FTBLib/FTBUtils methods to /home, /set_home, /back commands.. basically anything that will teleport a non-admin player. Is this possible? It looks like /warp is easily taken care of by not setting any warps but I don't see a way to prevent the others.
Thanks so much.
| priority | ability to disable sethome back commands greetings i would like to humbly request the config option to disable all ftblib ftbutils methods to home set home back commands basically anything that will teleport a non admin player is this possible it looks like warp is easily taken care of by not setting any warps but i don t see a way to prevent the others thanks so much | 1 |
525,945 | 15,269,527,976 | IssuesEvent | 2021-02-22 12:55:24 | AY2021S2-CS2103T-W13-2/tp | https://api.github.com/repos/AY2021S2-CS2103T-W13-2/tp | opened | Remind readers to return books overdue | priority.Low type.Story | As an individual operating a private book loaning service, I can remind readers to return their books which are nearly overdue, they will be more likely to return the books on time. | 1.0 | Remind readers to return books overdue - As an individual operating a private book loaning service, I can remind readers to return their books which are nearly overdue, they will be more likely to return the books on time. | priority | remind readers to return books overdue as an individual operating a private book loaning service i can remind readers to return their books which are nearly overdue they will be more likely to return the books on time | 1 |
812,766 | 30,351,490,450 | IssuesEvent | 2023-07-11 19:20:32 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Coupons are possible to be created without a coupon code name | type: enhancement priority: low focus: coupon [team:Proton] plugin: woocommerce | **Prerequisites (mark completed items with an [x]):**
- [x] I have have carried out troubleshooting steps and I believe I have found a bug.
- [x] I have searched for similar bugs in both open and closed issues and cannot find a duplicate.
**Describe the bug**
Coupons are possible to be created without a coupon code name.
**Expected behavior**
I expected to be shown a warning letting me know that the coupon code name is required in order to create a coupon.
**Steps to reproduce the bug**
Steps to reproduce the bug:
1. Go to `Marketing/Coupons`;
2. Click on the `Add coupon` button;
3. Fill in coupon details but leave the coupon code filed empty;
4. Click on the `Publish` button to create a coupon;
5. Note that the coupon had been created without a coupon code.
It is not a huge deal but it will be nice to let merchants know that they need to add a coupon code name as well.
**Screenshots**
See gif below:

<!-- Please try testing your site for theme and plugins conflict. To do that deactivate all plugins except for WooCommerce and switch to a default WordPress theme or [Storefront](https://en-gb.wordpress.org/themes/storefront/). Then test again. If the issue is resolved with the default theme and all plugins deactivated, it means that one of your plugins or a theme is causing the issue. You will then need to enable it one by one and test every time you do that in order to figure out which plugin is causing the issue. -->
**Isolating the problem (mark completed items with an [x]):**
- [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active.
- [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/).
- [x] I can reproduce this bug consistently using the steps above.
**WordPress Environment**
We use the [WooCommerce System Status Report](https://docs.woocommerce.com/document/understanding-the-woocommerce-system-status-report/) to help us evaluate the issue.
Without this report we won't be able to fully evaluate this issue.
<details>
```
`
### WordPress Environment ###
WordPress address (URL): https://ephemeral-jamosova-20201109.atomicsites.blog
Site address (URL): https://ephemeral-jamosova-20201109.atomicsites.blog
WC Version: 4.7.0
REST API Version: ✔ 4.7.0
WC Blocks Version: ✔ 3.6.0
Action Scheduler Version: ✔ 3.1.6
WC Admin Version: ✔ 1.6.3
Log Directory Writable: ✔
WP Version: 5.5.3
WP Multisite: –
WP Memory Limit: 256 MB
WP Debug Mode: ✔
WP Cron: ✔
Language: en_US
External object cache: ✔
### Server Environment ###
Server Info: nginx
PHP Version: 7.4.11
PHP Post Max Size: 2 GB
PHP Time Limit: 1200
PHP Max Input Vars: 6144
cURL Version: 7.72.0
OpenSSL/1.1.1d
SUHOSIN Installed: –
MySQL Version: 5.5.5-10.3.25-MariaDB-log
Max Upload Size: 2 GB
Default Timezone is UTC: ✔
fsockopen/cURL: ✔
SoapClient: ✔
DOMDocument: ✔
GZip: ✔
Multibyte String: ✔
Remote Post: ✔
Remote Get: ✔
### Database ###
WC Database Version: 4.7.0
WC Database Prefix: wp_
Total Database Size: 2.97MB
Database Data Size: 1.36MB
Database Index Size: 1.61MB
wp_woocommerce_sessions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_woocommerce_order_items: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_order_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_actions: Data: 0.02MB + Index: 0.11MB + Engine InnoDB
wp_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_logs: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_comments: Data: 0.02MB + Index: 0.09MB + Engine InnoDB
wp_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_mailchimp_carts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_mailchimp_jobs: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_options: Data: 0.23MB + Index: 0.06MB + Engine InnoDB
wp_postmeta: Data: 0.13MB + Index: 0.09MB + Engine InnoDB
wp_posts: Data: 0.06MB + Index: 0.06MB + Engine InnoDB
wp_termmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_terms: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_term_relationships: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_term_taxonomy: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_usermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB
wp_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_bookings_availability: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_bookings_availabilitymeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_booking_relationships: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB
wp_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_product_meta_lookup: Data: 0.02MB + Index: 0.09MB + Engine InnoDB
wp_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_bundled_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_bundled_items: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
### Post Type Counts ###
attachment: 25
page: 8
post: 2
product: 21
product_variation: 12
revision: 1
shop_coupon: 3
### Security ###
Secure connection (HTTPS): ✔
Hide errors from visitors: ✔
### Active Plugins (13) ###
Akismet Anti-Spam: by Automattic – 4.1.7
Facebook for WooCommerce: by Facebook – 2.1.4 – Installed version not tested with active version of WooCommerce 4.7.0
Gutenberg: by Gutenberg Team – 9.3.0
Jetpack by WordPress.com: by Automattic – 9.0.2
Google Ads & Marketing by Kliken: by Kliken – 1.0.6 – Installed version not tested with active version of WooCommerce 4.7.0
Mailchimp for WooCommerce: by Mailchimp – 2.4.7 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Bookings: by WooCommerce – 1.15.29 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Memberships: by SkyVerge – 1.19.2 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Payments: by Automattic – 1.6.0 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Product Add-ons: by WooCommerce – 3.1.0 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Product Bundles: by SomewhereWarm – 6.5.0
WooCommerce Shipping & Tax: by Automattic – 1.25.1 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce: by Automattic – 4.7.0
### Inactive Plugins (1) ###
WooCommerce Subscriptions: by WooCommerce – 3.0.9 – Installed version not tested with active version of WooCommerce 4.7.0
### Dropin Plugins (2) ###
advanced-cache.php: advanced-cache.php
object-cache.php: Memcached
### Settings ###
API Enabled: –
Force SSL: –
Currency: USD ($)
Currency Position: left
Thousand Separator: ,
Decimal Separator: .
Number of Decimals: 2
Taxonomies: Product Types: bundle (bundle)
external (external)
grouped (grouped)
simple (simple)
subscription (subscription)
variable (variable)
variable subscription (variable-subscription)
Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog)
exclude-from-search (exclude-from-search)
featured (featured)
outofstock (outofstock)
rated-1 (rated-1)
rated-2 (rated-2)
rated-3 (rated-3)
rated-4 (rated-4)
rated-5 (rated-5)
Connected to WooCommerce.com: ✔
### WC Pages ###
Shop base: #7 - /shop/
Cart: #8 - /cart/
Checkout: #9 - /checkout/
My account: #10 - /my-account/
Terms and conditions: ❌ Page not set
### Theme ###
Name: Storefront
Version: 2.8.0
Author URL: https://woocommerce.com/
Child Theme: ❌ – If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme
WooCommerce Support: ✔
### Templates ###
Overrides: –
### Memberships ###
Restriction mode:: Hide content only
Public content:: –
Excluded from discounts:: –
Membership plans:: ❌ No published plans
### Product Bundles ###
Database Version: 6.5.0
Loopback Test: ✔
Template Overrides: –
### Action Scheduler ###
Complete: 22
Oldest: 2020-11-09 20:36:57 +0000
Newest: 2020-11-09 21:16:27 +0000
Pending: 1
Oldest: 2020-11-09 21:17:27 +0000
Newest: 2020-11-09 21:17:27 +0000
Canceled: 1
Oldest: 2020-11-09 21:11:26 +0000
Newest: 2020-11-09 21:11:26 +0000
`
```
</details>
| 1.0 | Coupons are possible to be created without a coupon code name - **Prerequisites (mark completed items with an [x]):**
- [x] I have have carried out troubleshooting steps and I believe I have found a bug.
- [x] I have searched for similar bugs in both open and closed issues and cannot find a duplicate.
**Describe the bug**
Coupons are possible to be created without a coupon code name.
**Expected behavior**
I expected to be shown a warning letting me know that the coupon code name is required in order to create a coupon.
**Steps to reproduce the bug**
Steps to reproduce the bug:
1. Go to `Marketing/Coupons`;
2. Click on the `Add coupon` button;
3. Fill in coupon details but leave the coupon code filed empty;
4. Click on the `Publish` button to create a coupon;
5. Note that the coupon had been created without a coupon code.
It is not a huge deal but it will be nice to let merchants know that they need to add a coupon code name as well.
**Screenshots**
See gif below:

<!-- Please try testing your site for theme and plugins conflict. To do that deactivate all plugins except for WooCommerce and switch to a default WordPress theme or [Storefront](https://en-gb.wordpress.org/themes/storefront/). Then test again. If the issue is resolved with the default theme and all plugins deactivated, it means that one of your plugins or a theme is causing the issue. You will then need to enable it one by one and test every time you do that in order to figure out which plugin is causing the issue. -->
**Isolating the problem (mark completed items with an [x]):**
- [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active.
- [x] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/).
- [x] I can reproduce this bug consistently using the steps above.
**WordPress Environment**
We use the [WooCommerce System Status Report](https://docs.woocommerce.com/document/understanding-the-woocommerce-system-status-report/) to help us evaluate the issue.
Without this report we won't be able to fully evaluate this issue.
<details>
```
`
### WordPress Environment ###
WordPress address (URL): https://ephemeral-jamosova-20201109.atomicsites.blog
Site address (URL): https://ephemeral-jamosova-20201109.atomicsites.blog
WC Version: 4.7.0
REST API Version: ✔ 4.7.0
WC Blocks Version: ✔ 3.6.0
Action Scheduler Version: ✔ 3.1.6
WC Admin Version: ✔ 1.6.3
Log Directory Writable: ✔
WP Version: 5.5.3
WP Multisite: –
WP Memory Limit: 256 MB
WP Debug Mode: ✔
WP Cron: ✔
Language: en_US
External object cache: ✔
### Server Environment ###
Server Info: nginx
PHP Version: 7.4.11
PHP Post Max Size: 2 GB
PHP Time Limit: 1200
PHP Max Input Vars: 6144
cURL Version: 7.72.0
OpenSSL/1.1.1d
SUHOSIN Installed: –
MySQL Version: 5.5.5-10.3.25-MariaDB-log
Max Upload Size: 2 GB
Default Timezone is UTC: ✔
fsockopen/cURL: ✔
SoapClient: ✔
DOMDocument: ✔
GZip: ✔
Multibyte String: ✔
Remote Post: ✔
Remote Get: ✔
### Database ###
WC Database Version: 4.7.0
WC Database Prefix: wp_
Total Database Size: 2.97MB
Database Data Size: 1.36MB
Database Index Size: 1.61MB
wp_woocommerce_sessions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_woocommerce_order_items: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_order_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_actions: Data: 0.02MB + Index: 0.11MB + Engine InnoDB
wp_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_actionscheduler_logs: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_comments: Data: 0.02MB + Index: 0.09MB + Engine InnoDB
wp_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_mailchimp_carts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_mailchimp_jobs: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_options: Data: 0.23MB + Index: 0.06MB + Engine InnoDB
wp_postmeta: Data: 0.13MB + Index: 0.09MB + Engine InnoDB
wp_posts: Data: 0.06MB + Index: 0.06MB + Engine InnoDB
wp_termmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_terms: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_term_relationships: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_term_taxonomy: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_usermeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB
wp_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_bookings_availability: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_bookings_availabilitymeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_booking_relationships: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB
wp_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB
wp_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_wc_product_meta_lookup: Data: 0.02MB + Index: 0.09MB + Engine InnoDB
wp_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB
wp_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB
wp_woocommerce_bundled_itemmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
wp_woocommerce_bundled_items: Data: 0.02MB + Index: 0.03MB + Engine InnoDB
### Post Type Counts ###
attachment: 25
page: 8
post: 2
product: 21
product_variation: 12
revision: 1
shop_coupon: 3
### Security ###
Secure connection (HTTPS): ✔
Hide errors from visitors: ✔
### Active Plugins (13) ###
Akismet Anti-Spam: by Automattic – 4.1.7
Facebook for WooCommerce: by Facebook – 2.1.4 – Installed version not tested with active version of WooCommerce 4.7.0
Gutenberg: by Gutenberg Team – 9.3.0
Jetpack by WordPress.com: by Automattic – 9.0.2
Google Ads & Marketing by Kliken: by Kliken – 1.0.6 – Installed version not tested with active version of WooCommerce 4.7.0
Mailchimp for WooCommerce: by Mailchimp – 2.4.7 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Bookings: by WooCommerce – 1.15.29 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Memberships: by SkyVerge – 1.19.2 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Payments: by Automattic – 1.6.0 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Product Add-ons: by WooCommerce – 3.1.0 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce Product Bundles: by SomewhereWarm – 6.5.0
WooCommerce Shipping & Tax: by Automattic – 1.25.1 – Installed version not tested with active version of WooCommerce 4.7.0
WooCommerce: by Automattic – 4.7.0
### Inactive Plugins (1) ###
WooCommerce Subscriptions: by WooCommerce – 3.0.9 – Installed version not tested with active version of WooCommerce 4.7.0
### Dropin Plugins (2) ###
advanced-cache.php: advanced-cache.php
object-cache.php: Memcached
### Settings ###
API Enabled: –
Force SSL: –
Currency: USD ($)
Currency Position: left
Thousand Separator: ,
Decimal Separator: .
Number of Decimals: 2
Taxonomies: Product Types: bundle (bundle)
external (external)
grouped (grouped)
simple (simple)
subscription (subscription)
variable (variable)
variable subscription (variable-subscription)
Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog)
exclude-from-search (exclude-from-search)
featured (featured)
outofstock (outofstock)
rated-1 (rated-1)
rated-2 (rated-2)
rated-3 (rated-3)
rated-4 (rated-4)
rated-5 (rated-5)
Connected to WooCommerce.com: ✔
### WC Pages ###
Shop base: #7 - /shop/
Cart: #8 - /cart/
Checkout: #9 - /checkout/
My account: #10 - /my-account/
Terms and conditions: ❌ Page not set
### Theme ###
Name: Storefront
Version: 2.8.0
Author URL: https://woocommerce.com/
Child Theme: ❌ – If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme
WooCommerce Support: ✔
### Templates ###
Overrides: –
### Memberships ###
Restriction mode:: Hide content only
Public content:: –
Excluded from discounts:: –
Membership plans:: ❌ No published plans
### Product Bundles ###
Database Version: 6.5.0
Loopback Test: ✔
Template Overrides: –
### Action Scheduler ###
Complete: 22
Oldest: 2020-11-09 20:36:57 +0000
Newest: 2020-11-09 21:16:27 +0000
Pending: 1
Oldest: 2020-11-09 21:17:27 +0000
Newest: 2020-11-09 21:17:27 +0000
Canceled: 1
Oldest: 2020-11-09 21:11:26 +0000
Newest: 2020-11-09 21:11:26 +0000
`
```
</details>
| priority | coupons are possible to be created without a coupon code name prerequisites mark completed items with an i have have carried out troubleshooting steps and i believe i have found a bug i have searched for similar bugs in both open and closed issues and cannot find a duplicate describe the bug coupons are possible to be created without a coupon code name expected behavior i expected to be shown a warning letting me know that the coupon code name is required in order to create a coupon steps to reproduce the bug steps to reproduce the bug go to marketing coupons click on the add coupon button fill in coupon details but leave the coupon code filed empty click on the publish button to create a coupon note that the coupon had been created without a coupon code it is not a huge deal but it will be nice to let merchants know that they need to add a coupon code name as well screenshots see gif below isolating the problem mark completed items with an i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above wordpress environment we use the to help us evaluate the issue without this report we won t be able to fully evaluate this issue wordpress environment wordpress address url site address url wc version rest api version ✔ wc blocks version ✔ action scheduler version ✔ wc admin version ✔ log directory writable ✔ wp version wp multisite – wp memory limit mb wp debug mode ✔ wp cron ✔ language en us external object cache ✔ server environment server info nginx php version php post max size gb php time limit php max input vars curl version openssl suhosin installed – mysql version mariadb log max upload size gb default timezone is utc ✔ fsockopen curl ✔ soapclient ✔ domdocument ✔ gzip ✔ multibyte string ✔ remote post ✔ remote get ✔ database wc database version wc database prefix wp total database size database data size database index size wp woocommerce sessions data index engine innodb wp woocommerce api keys data index engine innodb wp woocommerce attribute taxonomies data index engine innodb wp woocommerce downloadable product permissions data index engine innodb wp woocommerce order items data index engine innodb wp woocommerce order itemmeta data index engine innodb wp woocommerce tax rates data index engine innodb wp woocommerce tax rate locations data index engine innodb wp woocommerce shipping zones data index engine innodb wp woocommerce shipping zone locations data index engine innodb wp woocommerce shipping zone methods data index engine innodb wp woocommerce payment tokens data index engine innodb wp woocommerce payment tokenmeta data index engine innodb wp woocommerce log data index engine innodb wp actionscheduler actions data index engine innodb wp actionscheduler claims data index engine innodb wp actionscheduler groups data index engine innodb wp actionscheduler logs data index engine innodb wp commentmeta data index engine innodb wp comments data index engine innodb wp links data index engine innodb wp mailchimp carts data index engine innodb wp mailchimp jobs data index engine innodb wp options data index engine innodb wp postmeta data index engine innodb wp posts data index engine innodb wp termmeta data index engine innodb wp terms data index engine innodb wp term relationships data index engine innodb wp term taxonomy data index engine innodb wp usermeta data index engine innodb wp users data index engine innodb wp wc admin notes data index engine innodb wp wc admin note actions data index engine innodb wp wc bookings availability data index engine innodb wp wc bookings availabilitymeta data index engine innodb wp wc booking relationships data index engine innodb wp wc category lookup data index engine innodb wp wc customer lookup data index engine innodb wp wc download log data index engine innodb wp wc order coupon lookup data index engine innodb wp wc order product lookup data index engine innodb wp wc order stats data index engine innodb wp wc order tax lookup data index engine innodb wp wc product meta lookup data index engine innodb wp wc reserved stock data index engine innodb wp wc tax rate classes data index engine innodb wp wc webhooks data index engine innodb wp woocommerce bundled itemmeta data index engine innodb wp woocommerce bundled items data index engine innodb post type counts attachment page post product product variation revision shop coupon security secure connection https ✔ hide errors from visitors ✔ active plugins akismet anti spam by automattic – facebook for woocommerce by facebook – – installed version not tested with active version of woocommerce gutenberg by gutenberg team – jetpack by wordpress com by automattic – google ads marketing by kliken by kliken – – installed version not tested with active version of woocommerce mailchimp for woocommerce by mailchimp – – installed version not tested with active version of woocommerce woocommerce bookings by woocommerce – – installed version not tested with active version of woocommerce woocommerce memberships by skyverge – – installed version not tested with active version of woocommerce woocommerce payments by automattic – – installed version not tested with active version of woocommerce woocommerce product add ons by woocommerce – – installed version not tested with active version of woocommerce woocommerce product bundles by somewherewarm – woocommerce shipping tax by automattic – – installed version not tested with active version of woocommerce woocommerce by automattic – inactive plugins woocommerce subscriptions by woocommerce – – installed version not tested with active version of woocommerce dropin plugins advanced cache php advanced cache php object cache php memcached settings api enabled – force ssl – currency usd currency position left thousand separator decimal separator number of decimals taxonomies product types bundle bundle external external grouped grouped simple simple subscription subscription variable variable variable subscription variable subscription taxonomies product visibility exclude from catalog exclude from catalog exclude from search exclude from search featured featured outofstock outofstock rated rated rated rated rated rated rated rated rated rated connected to woocommerce com ✔ wc pages shop base shop cart cart checkout checkout my account my account terms and conditions ❌ page not set theme name storefront version author url child theme ❌ – if you are modifying woocommerce on a parent theme that you did not build personally we recommend using a child theme see how to create a child theme woocommerce support ✔ templates overrides – memberships restriction mode hide content only public content – excluded from discounts – membership plans ❌ no published plans product bundles database version loopback test ✔ template overrides – action scheduler complete oldest newest pending oldest newest canceled oldest newest | 1 |
715,917 | 24,615,398,880 | IssuesEvent | 2022-10-15 08:37:39 | okfde/froide | https://api.github.com/repos/okfde/froide | closed | Better visualization of responses by multiple public bodies | low priority | sometimes a request for information gets forwarded internally to more then one public body if that happens the number of responses by many different institutions gets unorganized.
A better way to display this requests would be nice (also for the person asking for information as it helps them to see where there is still work to be done)
| 1.0 | Better visualization of responses by multiple public bodies - sometimes a request for information gets forwarded internally to more then one public body if that happens the number of responses by many different institutions gets unorganized.
A better way to display this requests would be nice (also for the person asking for information as it helps them to see where there is still work to be done)
| priority | better visualization of responses by multiple public bodies sometimes a request for information gets forwarded internally to more then one public body if that happens the number of responses by many different institutions gets unorganized a better way to display this requests would be nice also for the person asking for information as it helps them to see where there is still work to be done | 1 |
559,183 | 16,552,009,836 | IssuesEvent | 2021-05-28 09:38:46 | sopra-fs21-group-09/sopra-fs21-group-09-client | https://api.github.com/repos/sopra-fs21-group-09/sopra-fs21-group-09-client | closed | Create a Intro message for new users | Frontend low priority task | * Due to feedback, we want to give some information at the start of using the app.
Estimate: 1h, priority: 1 | 1.0 | Create a Intro message for new users - * Due to feedback, we want to give some information at the start of using the app.
Estimate: 1h, priority: 1 | priority | create a intro message for new users due to feedback we want to give some information at the start of using the app estimate priority | 1 |
650,800 | 21,417,663,614 | IssuesEvent | 2022-04-22 12:37:05 | MeAmAnUsername/pie | https://api.github.com/repos/MeAmAnUsername/pie | opened | Annotation to suppress messages | Type: enhancement Component: DSL Priority: low Component: editor Status: suggestion | **Summary**
Add annotations to suppress messages
**Todo**
- [ ] Figure out if there is ever a valid reason to suppress an error (and a warning?)
- [ ] Create SPT tests
- [ ] Implement
- [ ] Add to documentation
- [ ] Make it clear that the comment field in the suppression is optional but highly recommended.
This means that it should not just be mentioned in the description of that field, but also in the main description of suppressions.
**Reason**
Not every message can be fixed. In particular, notes may not be fixable. Having notes that have already been deemed unfixable is noise, so it would be good if these can be suppressed.
**Description**
Annotations to suppress messages. This may make it possible to add notes when code does not follow language convention: if you don't want that, you can simply suppress them (possibly project wide, with `suppress { PROJECT, conventionBreak }`)
Parameters:
`CaptureCount` (optional, option name can be omitted): `SINGLE`, `OPT`, `ALL`, `ANY`, `MIN(int)`, `MAX(int|inf)`, `EXACT(int)`, `RANGE(int, int|inf)`. All of the options are sugar for a RANGE. `CaptureCount` is the number of messages that this suppression can capture. If the suppression matches outside the range of `CaptureCount`, there is a warning on `suppress`.
`MessageName` (required, option name can be omitted): the name of the message to suppress. Message names form taxonomies, so you could e.g. use `suppress { shadowing }` to suppress all messages related to shadowing
`match`: a filter that is matched against the text of messages. This suppression only matches messages for which the text matches. Takes a string (`"hello"`) or a regex (`/he..o/`).
`priority`: the priority of the suppression (any int). Higher priority suppressions capture messages earlier.
`scope`: can be either `PROJECT`, `FILE`, `TARGET` (default) or a lexical scope created by adding a second set of brackets to the annotation. The second set can be either just `{` and `}`, or `{x|` and `|x}`, where `x` must be a non-negative integer (opening and closing bracket must use the same integer). `PROJECT` applies the suppression to the whole project. `FILE` applies the suppression to the whole file. `TARGET` applies the thing to
`comment` (optional, option name can be omitted): a comment describing why this message is suppressed. It is a string which does not require enclosing brackets.
**Example**
```
module org:example:suppress_messages
// No notes unless specified in comment
data E1[E1] = foreign java Example {} // note on type parameter name E1.
data E2[suppress { shadowingTypeParameter, example to show how suppression on name works } E2] = foreign java Example {}
suppress { shadowingTypeParameter, example to show how suppression on data definition works } // annotation applies only to next definition, i.e. E3
data E3[E3] = foreign java Example {}
suppress { shadowingTypeParameter, example to show what happens if no message to suppress } // warning on `suppress`: "No shadowingTypeParameter messages to suppress. Remove this suppression or use option `OPT` to allow useless suppressions"
data E4[E] = foreign java Example {}
suppress { OPT, shadowingTypeParameter, example to show how optional suppression works }
data E5[E] = foreign java Example {}
suppress { shadowingTypeParameter, match: "E6", example to show how suppression with match works }
data E6[E6, E5] = foreign java Example {} // note on E5
suppress { shadowingTypeParameter, match: "E5", example to show how suppression with non-matching match works } // warning on `suppress`: "No messages to suppress. There were shadowingTypeParameter messages but they did not match "E5". Remove this suppression or use option `OPT` to allow useless suppressions"
data E7[E7, E6] = foreign java Example {} // note on type parameters E7, E6
suppress { shadowingTypeParameter, example to show what happens with multiple matching messages } // warning on `suppress`: "Multiple matching messages. Use option `ALL` to suppress them all, use a match to filter on the message text, or add a specific suppression for each message }
data E8[E8, E9] = foreign java Example {}
suppress { ALL, shadowingTypeParameter, example to show the ALL option }
data E9[E9, E8] = foreign java Example {}
suppress { shadowingTypeParameter } // no comment
data E10[E10] = foreign java Example {}
suppress { ALL, shadowingTypeParameter, example to show how suppression within a scope works}{
data E11[E11] = foreign java Example {}
data E12[E12] = foreign java Example {}
}
suppress { ALL, shadowingTypeParameter, example to show how custom brackets for annotation scope works}{1|
data E13[E13] = foreign java Example {}
data E14[E14] = foreign java Example {}
|1}
suppress { ALL, shadowingTypeParameter, example to show how suppression of something that was already suppressed works}{
suppress { shadowingTypeParameter } // applies only to E15. warning on `suppress`: "Matching shadowingTypeParameter message "[message]" is already suppressed in surrounding scope. Add a priority higher than 0 to override suppression in surrounding scope. Alternatively, remove this annotation." Warnings like these are only emitted when this suppression captures nothing that is not already captured by suppressions in higher scopes.
data E15[E15] = foreign java Example {}
data E16[E16] = foreign java Example {}
}
suppress { ALL, shadowingTypeParameter, example to show how suppression of something that was already suppressed works}{
suppress { shadowingTypeParameter, priority: 1 } // applies only to E17.
data E17[E17] = foreign java Example {}
data E18[E18] = foreign java Example {}
}
```
**Extensions**
Create a way to view/filter all suppressions. Already possible by just searching for `suppress`, but that may turn up some false positives. Such a view could show the comments (i.e. reason) as well. Might be useful to find suppressions without comments.
**Related issues**
- Annotations: #120
**Fun fact**
You can upgrade a note to a warning by doing `suppress { EXACT(0), ... }`, which will show a warning on this suppression if such a note is found. | 1.0 | Annotation to suppress messages - **Summary**
Add annotations to suppress messages
**Todo**
- [ ] Figure out if there is ever a valid reason to suppress an error (and a warning?)
- [ ] Create SPT tests
- [ ] Implement
- [ ] Add to documentation
- [ ] Make it clear that the comment field in the suppression is optional but highly recommended.
This means that it should not just be mentioned in the description of that field, but also in the main description of suppressions.
**Reason**
Not every message can be fixed. In particular, notes may not be fixable. Having notes that have already been deemed unfixable is noise, so it would be good if these can be suppressed.
**Description**
Annotations to suppress messages. This may make it possible to add notes when code does not follow language convention: if you don't want that, you can simply suppress them (possibly project wide, with `suppress { PROJECT, conventionBreak }`)
Parameters:
`CaptureCount` (optional, option name can be omitted): `SINGLE`, `OPT`, `ALL`, `ANY`, `MIN(int)`, `MAX(int|inf)`, `EXACT(int)`, `RANGE(int, int|inf)`. All of the options are sugar for a RANGE. `CaptureCount` is the number of messages that this suppression can capture. If the suppression matches outside the range of `CaptureCount`, there is a warning on `suppress`.
`MessageName` (required, option name can be omitted): the name of the message to suppress. Message names form taxonomies, so you could e.g. use `suppress { shadowing }` to suppress all messages related to shadowing
`match`: a filter that is matched against the text of messages. This suppression only matches messages for which the text matches. Takes a string (`"hello"`) or a regex (`/he..o/`).
`priority`: the priority of the suppression (any int). Higher priority suppressions capture messages earlier.
`scope`: can be either `PROJECT`, `FILE`, `TARGET` (default) or a lexical scope created by adding a second set of brackets to the annotation. The second set can be either just `{` and `}`, or `{x|` and `|x}`, where `x` must be a non-negative integer (opening and closing bracket must use the same integer). `PROJECT` applies the suppression to the whole project. `FILE` applies the suppression to the whole file. `TARGET` applies the thing to
`comment` (optional, option name can be omitted): a comment describing why this message is suppressed. It is a string which does not require enclosing brackets.
**Example**
```
module org:example:suppress_messages
// No notes unless specified in comment
data E1[E1] = foreign java Example {} // note on type parameter name E1.
data E2[suppress { shadowingTypeParameter, example to show how suppression on name works } E2] = foreign java Example {}
suppress { shadowingTypeParameter, example to show how suppression on data definition works } // annotation applies only to next definition, i.e. E3
data E3[E3] = foreign java Example {}
suppress { shadowingTypeParameter, example to show what happens if no message to suppress } // warning on `suppress`: "No shadowingTypeParameter messages to suppress. Remove this suppression or use option `OPT` to allow useless suppressions"
data E4[E] = foreign java Example {}
suppress { OPT, shadowingTypeParameter, example to show how optional suppression works }
data E5[E] = foreign java Example {}
suppress { shadowingTypeParameter, match: "E6", example to show how suppression with match works }
data E6[E6, E5] = foreign java Example {} // note on E5
suppress { shadowingTypeParameter, match: "E5", example to show how suppression with non-matching match works } // warning on `suppress`: "No messages to suppress. There were shadowingTypeParameter messages but they did not match "E5". Remove this suppression or use option `OPT` to allow useless suppressions"
data E7[E7, E6] = foreign java Example {} // note on type parameters E7, E6
suppress { shadowingTypeParameter, example to show what happens with multiple matching messages } // warning on `suppress`: "Multiple matching messages. Use option `ALL` to suppress them all, use a match to filter on the message text, or add a specific suppression for each message }
data E8[E8, E9] = foreign java Example {}
suppress { ALL, shadowingTypeParameter, example to show the ALL option }
data E9[E9, E8] = foreign java Example {}
suppress { shadowingTypeParameter } // no comment
data E10[E10] = foreign java Example {}
suppress { ALL, shadowingTypeParameter, example to show how suppression within a scope works}{
data E11[E11] = foreign java Example {}
data E12[E12] = foreign java Example {}
}
suppress { ALL, shadowingTypeParameter, example to show how custom brackets for annotation scope works}{1|
data E13[E13] = foreign java Example {}
data E14[E14] = foreign java Example {}
|1}
suppress { ALL, shadowingTypeParameter, example to show how suppression of something that was already suppressed works}{
suppress { shadowingTypeParameter } // applies only to E15. warning on `suppress`: "Matching shadowingTypeParameter message "[message]" is already suppressed in surrounding scope. Add a priority higher than 0 to override suppression in surrounding scope. Alternatively, remove this annotation." Warnings like these are only emitted when this suppression captures nothing that is not already captured by suppressions in higher scopes.
data E15[E15] = foreign java Example {}
data E16[E16] = foreign java Example {}
}
suppress { ALL, shadowingTypeParameter, example to show how suppression of something that was already suppressed works}{
suppress { shadowingTypeParameter, priority: 1 } // applies only to E17.
data E17[E17] = foreign java Example {}
data E18[E18] = foreign java Example {}
}
```
**Extensions**
Create a way to view/filter all suppressions. Already possible by just searching for `suppress`, but that may turn up some false positives. Such a view could show the comments (i.e. reason) as well. Might be useful to find suppressions without comments.
**Related issues**
- Annotations: #120
**Fun fact**
You can upgrade a note to a warning by doing `suppress { EXACT(0), ... }`, which will show a warning on this suppression if such a note is found. | priority | annotation to suppress messages summary add annotations to suppress messages todo figure out if there is ever a valid reason to suppress an error and a warning create spt tests implement add to documentation make it clear that the comment field in the suppression is optional but highly recommended this means that it should not just be mentioned in the description of that field but also in the main description of suppressions reason not every message can be fixed in particular notes may not be fixable having notes that have already been deemed unfixable is noise so it would be good if these can be suppressed description annotations to suppress messages this may make it possible to add notes when code does not follow language convention if you don t want that you can simply suppress them possibly project wide with suppress project conventionbreak parameters capturecount optional option name can be omitted single opt all any min int max int inf exact int range int int inf all of the options are sugar for a range capturecount is the number of messages that this suppression can capture if the suppression matches outside the range of capturecount there is a warning on suppress messagename required option name can be omitted the name of the message to suppress message names form taxonomies so you could e g use suppress shadowing to suppress all messages related to shadowing match a filter that is matched against the text of messages this suppression only matches messages for which the text matches takes a string hello or a regex he o priority the priority of the suppression any int higher priority suppressions capture messages earlier scope can be either project file target default or a lexical scope created by adding a second set of brackets to the annotation the second set can be either just and or x and x where x must be a non negative integer opening and closing bracket must use the same integer project applies the suppression to the whole project file applies the suppression to the whole file target applies the thing to comment optional option name can be omitted a comment describing why this message is suppressed it is a string which does not require enclosing brackets example module org example suppress messages no notes unless specified in comment data foreign java example note on type parameter name data foreign java example suppress shadowingtypeparameter example to show how suppression on data definition works annotation applies only to next definition i e data foreign java example suppress shadowingtypeparameter example to show what happens if no message to suppress warning on suppress no shadowingtypeparameter messages to suppress remove this suppression or use option opt to allow useless suppressions data foreign java example suppress opt shadowingtypeparameter example to show how optional suppression works data foreign java example suppress shadowingtypeparameter match example to show how suppression with match works data foreign java example note on suppress shadowingtypeparameter match example to show how suppression with non matching match works warning on suppress no messages to suppress there were shadowingtypeparameter messages but they did not match remove this suppression or use option opt to allow useless suppressions data foreign java example note on type parameters suppress shadowingtypeparameter example to show what happens with multiple matching messages warning on suppress multiple matching messages use option all to suppress them all use a match to filter on the message text or add a specific suppression for each message data foreign java example suppress all shadowingtypeparameter example to show the all option data foreign java example suppress shadowingtypeparameter no comment data foreign java example suppress all shadowingtypeparameter example to show how suppression within a scope works data foreign java example data foreign java example suppress all shadowingtypeparameter example to show how custom brackets for annotation scope works data foreign java example data foreign java example suppress all shadowingtypeparameter example to show how suppression of something that was already suppressed works suppress shadowingtypeparameter applies only to warning on suppress matching shadowingtypeparameter message is already suppressed in surrounding scope add a priority higher than to override suppression in surrounding scope alternatively remove this annotation warnings like these are only emitted when this suppression captures nothing that is not already captured by suppressions in higher scopes data foreign java example data foreign java example suppress all shadowingtypeparameter example to show how suppression of something that was already suppressed works suppress shadowingtypeparameter priority applies only to data foreign java example data foreign java example extensions create a way to view filter all suppressions already possible by just searching for suppress but that may turn up some false positives such a view could show the comments i e reason as well might be useful to find suppressions without comments related issues annotations fun fact you can upgrade a note to a warning by doing suppress exact which will show a warning on this suppression if such a note is found | 1 |
562,198 | 16,653,488,162 | IssuesEvent | 2021-06-05 04:40:30 | code-gambit/VT-WebClient | https://api.github.com/repos/code-gambit/VT-WebClient | closed | [BUG/Enhancement] enhancements in file listing page | Priority: Low Type: Enhancement UI | **Describe the bug**
On the file listing page, currently, the user has to tap on the small circle icon on the right to view its details but clicking on the file item itself doesn't allow the user to view the file details. Instead, considering the user's point of view it's better to remove the extra interaction of that button and allow the user to view file details on click of the entire file list item.
Area Where Bug Occurred
- [x] UI/UX
- [ ] Business Logic
**Screenshot**

**Before submitting an issue check the below**
- [x] I have synced all my node versions as mentioned in the project"
- [x] I am using the same version of npm as is the project
- [x] My current branch is in sync with the development branch
| 1.0 | [BUG/Enhancement] enhancements in file listing page - **Describe the bug**
On the file listing page, currently, the user has to tap on the small circle icon on the right to view its details but clicking on the file item itself doesn't allow the user to view the file details. Instead, considering the user's point of view it's better to remove the extra interaction of that button and allow the user to view file details on click of the entire file list item.
Area Where Bug Occurred
- [x] UI/UX
- [ ] Business Logic
**Screenshot**

**Before submitting an issue check the below**
- [x] I have synced all my node versions as mentioned in the project"
- [x] I am using the same version of npm as is the project
- [x] My current branch is in sync with the development branch
| priority | enhancements in file listing page describe the bug on the file listing page currently the user has to tap on the small circle icon on the right to view its details but clicking on the file item itself doesn t allow the user to view the file details instead considering the user s point of view it s better to remove the extra interaction of that button and allow the user to view file details on click of the entire file list item area where bug occurred ui ux business logic screenshot before submitting an issue check the below i have synced all my node versions as mentioned in the project i am using the same version of npm as is the project my current branch is in sync with the development branch | 1 |
747,476 | 26,086,032,527 | IssuesEvent | 2022-12-26 03:03:01 | ImranR98/Obtainium | https://api.github.com/repos/ImranR98/Obtainium | closed | Show no of apps selected | enhancement low priority | Add how many apps are selected in apps section, like 1 apps selected. | 1.0 | Show no of apps selected - Add how many apps are selected in apps section, like 1 apps selected. | priority | show no of apps selected add how many apps are selected in apps section like apps selected | 1 |
398,254 | 11,739,402,438 | IssuesEvent | 2020-03-11 17:38:54 | thaliawww/ThaliApp | https://api.github.com/repos/thaliawww/ThaliApp | closed | "Register" button in Login screen | easy and fun feature priority: low | In GitLab by @gijshendriksen on Aug 13, 2018, 15:36
### One-sentence description
Add button to register as a member in the Login screen, which can open the website when pressed.
### Desired behaviour
The only current options in the Login screen are "login" and "forgot password". For new students, it could be useful to add a button showing where they can register as a member. | 1.0 | "Register" button in Login screen - In GitLab by @gijshendriksen on Aug 13, 2018, 15:36
### One-sentence description
Add button to register as a member in the Login screen, which can open the website when pressed.
### Desired behaviour
The only current options in the Login screen are "login" and "forgot password". For new students, it could be useful to add a button showing where they can register as a member. | priority | register button in login screen in gitlab by gijshendriksen on aug one sentence description add button to register as a member in the login screen which can open the website when pressed desired behaviour the only current options in the login screen are login and forgot password for new students it could be useful to add a button showing where they can register as a member | 1 |
299,561 | 9,205,615,496 | IssuesEvent | 2019-03-08 11:05:59 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | Legend name too long when executing "qgis vectorfile.shp" | Category: GUI Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **cgs_bob -** (cgs_bob -)
Original Redmine Issue: 936, https://issues.qgis.org/issues/936
Original Assignee: Redmine Admin
---
Displaying a vector layer by specifying a vector file on the command line displays the vector and the complete file path as the name in the legend. If I work in a deeply nest directory, this name can be very long. It would be great if the legend name is the basename of this long path.
| 1.0 | Legend name too long when executing "qgis vectorfile.shp" - ---
Author Name: **cgs_bob -** (cgs_bob -)
Original Redmine Issue: 936, https://issues.qgis.org/issues/936
Original Assignee: Redmine Admin
---
Displaying a vector layer by specifying a vector file on the command line displays the vector and the complete file path as the name in the legend. If I work in a deeply nest directory, this name can be very long. It would be great if the legend name is the basename of this long path.
| priority | legend name too long when executing qgis vectorfile shp author name cgs bob cgs bob original redmine issue original assignee redmine admin displaying a vector layer by specifying a vector file on the command line displays the vector and the complete file path as the name in the legend if i work in a deeply nest directory this name can be very long it would be great if the legend name is the basename of this long path | 1 |
44,518 | 2,906,620,405 | IssuesEvent | 2015-06-19 11:23:23 | wchill/steamSummerMinigame | https://api.github.com/repos/wchill/steamSummerMinigame | closed | Disable/Fix Health Easing | 3 - Low Priority | The game has built in code to "ease" the health based on estimations. This is often inaccurate. One example is when you die prior to your health indicating you're low.
I'm not sure how it's implemented, but I'd like to see the numbers as soon as I get them from the server instead of having the game "ease" down to them.
Alternatively, show two health bars for mobs and players - one of the actual health and one of the eased health? Or have them overlapped with different colors or something. | 1.0 | Disable/Fix Health Easing - The game has built in code to "ease" the health based on estimations. This is often inaccurate. One example is when you die prior to your health indicating you're low.
I'm not sure how it's implemented, but I'd like to see the numbers as soon as I get them from the server instead of having the game "ease" down to them.
Alternatively, show two health bars for mobs and players - one of the actual health and one of the eased health? Or have them overlapped with different colors or something. | priority | disable fix health easing the game has built in code to ease the health based on estimations this is often inaccurate one example is when you die prior to your health indicating you re low i m not sure how it s implemented but i d like to see the numbers as soon as i get them from the server instead of having the game ease down to them alternatively show two health bars for mobs and players one of the actual health and one of the eased health or have them overlapped with different colors or something | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.