Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
54,906 | 3,071,587,715 | IssuesEvent | 2015-08-19 13:02:49 | sunpy/sunpy | https://api.github.com/repos/sunpy/sunpy | opened | Add a shift() method to GenericMap | effort-low Feature Request Map priority-medium | In order to easily fix any translation errors in the map location. | 1.0 | Add a shift() method to GenericMap - In order to easily fix any translation errors in the map location. | priority | add a shift method to genericmap in order to easily fix any translation errors in the map location | 1 |
46,642 | 2,963,660,513 | IssuesEvent | 2015-07-10 12:07:09 | joeha480/dotify | https://api.github.com/repos/joeha480/dotify | closed | Use factories and separate jars for input interpreters | auto-migrated Component-Input Maintainability Priority-Medium Type-Task | ```
Use java factories and separate jars for input interpreters.
The reason for this is that these have a tendency to change while the system
remains the same, and it can generate a lot of new versions.
If separated, those interested in a package of input interpreters can update
only these, and the system does not have to change version.
```
Original issue reported on code.google.com by `joel.hak...@mtm.se` on 26 Sep 2011 at 7:46
* Blocked on: #60 | 1.0 | Use factories and separate jars for input interpreters - ```
Use java factories and separate jars for input interpreters.
The reason for this is that these have a tendency to change while the system
remains the same, and it can generate a lot of new versions.
If separated, those interested in a package of input interpreters can update
only these, and the system does not have to change version.
```
Original issue reported on code.google.com by `joel.hak...@mtm.se` on 26 Sep 2011 at 7:46
* Blocked on: #60 | priority | use factories and separate jars for input interpreters use java factories and separate jars for input interpreters the reason for this is that these have a tendency to change while the system remains the same and it can generate a lot of new versions if separated those interested in a package of input interpreters can update only these and the system does not have to change version original issue reported on code google com by joel hak mtm se on sep at blocked on | 1 |
29,600 | 2,716,621,621 | IssuesEvent | 2015-04-10 20:17:32 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | Alert window appears unexpectedly | bug Component-UI imported Milestone-M14-C3 Module-CruxWidgets Priority-Medium TargetVersion-5.2.0 | _From [alexan...@triggolabs.com](https://code.google.com/u/114384922929308053156/) on August 21, 2014 11:48:34_
Alert window appears unexpectedly during use of an compiled system. The system use crux-widgets (MessageDialog class).
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=486_ | 1.0 | Alert window appears unexpectedly - _From [alexan...@triggolabs.com](https://code.google.com/u/114384922929308053156/) on August 21, 2014 11:48:34_
Alert window appears unexpectedly during use of an compiled system. The system use crux-widgets (MessageDialog class).
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=486_ | priority | alert window appears unexpectedly from on august alert window appears unexpectedly during use of an compiled system the system use crux widgets messagedialog class original issue | 1 |
153,531 | 5,893,800,318 | IssuesEvent | 2017-05-17 23:08:24 | Radarr/Radarr | https://api.github.com/repos/Radarr/Radarr | reopened | Chmod the osx binary with executable | bug priority:medium under investigation | **Description:**
There's currently and issue where the OSX package gives "The application 'Radarr' cannot be opened", which is fixed by **chmod +x** on the **Contents/MacOS/Sonarr** file.
I will update the cake build script to setup the permissions correctly. | 1.0 | Chmod the osx binary with executable - **Description:**
There's currently and issue where the OSX package gives "The application 'Radarr' cannot be opened", which is fixed by **chmod +x** on the **Contents/MacOS/Sonarr** file.
I will update the cake build script to setup the permissions correctly. | priority | chmod the osx binary with executable description there s currently and issue where the osx package gives the application radarr cannot be opened which is fixed by chmod x on the contents macos sonarr file i will update the cake build script to setup the permissions correctly | 1 |
463 | 2,497,112,986 | IssuesEvent | 2015-01-07 01:26:32 | leonbloy/pngj | https://api.github.com/repos/leonbloy/pngj | closed | Make method invalidateRaw() of class PngChunk public | bug imported Priority-Medium | _From [gijsb...@gmail.com](https://code.google.com/u/101485543047692581958/) on December 17, 2013 08:36:48_
In order to edit, for instance, text chucks, the ChunkRaw attribute has to be cleared before writing a new png file. Otherwise, it still uses the old value instead. Therefore the invalidateRaw() method should be declared public, but is currently declared protected. What version of the product are you using? On what operating system? 2.0.0, Debian Wheezy
_Original issue: http://code.google.com/p/pngj/issues/detail?id=27_ | 1.0 | Make method invalidateRaw() of class PngChunk public - _From [gijsb...@gmail.com](https://code.google.com/u/101485543047692581958/) on December 17, 2013 08:36:48_
In order to edit, for instance, text chucks, the ChunkRaw attribute has to be cleared before writing a new png file. Otherwise, it still uses the old value instead. Therefore the invalidateRaw() method should be declared public, but is currently declared protected. What version of the product are you using? On what operating system? 2.0.0, Debian Wheezy
_Original issue: http://code.google.com/p/pngj/issues/detail?id=27_ | priority | make method invalidateraw of class pngchunk public from on december in order to edit for instance text chucks the chunkraw attribute has to be cleared before writing a new png file otherwise it still uses the old value instead therefore the invalidateraw method should be declared public but is currently declared protected what version of the product are you using on what operating system debian wheezy original issue | 1 |
57,807 | 3,083,990,703 | IssuesEvent | 2015-08-24 12:45:07 | StefanIsidorovic/salira | https://api.github.com/repos/StefanIsidorovic/salira | closed | G-Code creation | auto-migrated Priority-Medium Type-Task | ```
For established skeleton, with parser and expressions, modify evals so it can
create G-CODE.
```
Original issue reported on code.google.com by `teva....@gmail.com` on 24 May 2015 at 1:02 | 1.0 | G-Code creation - ```
For established skeleton, with parser and expressions, modify evals so it can
create G-CODE.
```
Original issue reported on code.google.com by `teva....@gmail.com` on 24 May 2015 at 1:02 | priority | g code creation for established skeleton with parser and expressions modify evals so it can create g code original issue reported on code google com by teva gmail com on may at | 1 |
718,678 | 24,728,514,916 | IssuesEvent | 2022-10-20 15:39:36 | ooni/ooni.org | https://api.github.com/repos/ooni/ooni.org | closed | Prepare OONI session for OTF Shutdown Getdown event (Oxford) | priority/medium workshop | I'll be facilitating an OONI MAT session as part of this event, and I need to prepare the session. | 1.0 | Prepare OONI session for OTF Shutdown Getdown event (Oxford) - I'll be facilitating an OONI MAT session as part of this event, and I need to prepare the session. | priority | prepare ooni session for otf shutdown getdown event oxford i ll be facilitating an ooni mat session as part of this event and i need to prepare the session | 1 |
603,699 | 18,670,550,096 | IssuesEvent | 2021-10-30 16:20:33 | theseion/Fuel | https://api.github.com/repos/theseion/Fuel | closed | Support for both, text and binary streams | Type-Task Priority-Medium auto-migrated stale | ```
Text or binary: There are serializers like "pickle" in Python or Google
Protocol Buffers, that let you choose between text and binary representation.
For debugging or while developing one can just use text based, which is easy to
see, inspect and modify, and then, at production time, you can switch to a
binary format. This is really cool and I would love to have it in Fuel.
Maybe we can do it after issue 3
```
Original issue reported on code.google.com by `marianopeck` on 24 Mar 2011 at 9:56
| 1.0 | Support for both, text and binary streams - ```
Text or binary: There are serializers like "pickle" in Python or Google
Protocol Buffers, that let you choose between text and binary representation.
For debugging or while developing one can just use text based, which is easy to
see, inspect and modify, and then, at production time, you can switch to a
binary format. This is really cool and I would love to have it in Fuel.
Maybe we can do it after issue 3
```
Original issue reported on code.google.com by `marianopeck` on 24 Mar 2011 at 9:56
| priority | support for both text and binary streams text or binary there are serializers like pickle in python or google protocol buffers that let you choose between text and binary representation for debugging or while developing one can just use text based which is easy to see inspect and modify and then at production time you can switch to a binary format this is really cool and i would love to have it in fuel maybe we can do it after issue original issue reported on code google com by marianopeck on mar at | 1 |
677,654 | 23,169,575,068 | IssuesEvent | 2022-07-30 13:49:03 | dnd-side-project/dnd-7th-7-frontend | https://api.github.com/repos/dnd-side-project/dnd-7th-7-frontend | opened | [feature] Home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์ ์์
| ๐ง Type: Feature ๐ Priority: Medium | ## ๐ฅ ์ค๋ช
(์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์!)
- ๋๋ต์ ์ธ ์์ด์ดํ๋ ์์ ๋ฐ๋ผ Home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์๋ง ์ก์๋๊ฒ ์ต๋๋ค.
<br>
## ๐ฅ ETC (๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์!)
<img width="352" alt="image" src="https://user-images.githubusercontent.com/80577900/181917347-e750df32-bb28-44cc-aa30-421c73f7f690.png">
<br>
>#### types
>- feat : ์๋ก์ด ๊ธฐ๋ฅ
>- update : ๋ก์ง ๋ณ๊ฒฝ
>- bug : ๋ฒ๊ทธ ์์
>- refactor : ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง
>- style : ์ฝ๋ ํ์ ๋ณ๊ฒฝ
>- docs : ๋ฌธ์ ๊ด๋ จ
>- test : ํ
์คํธ ์ฝ๋
>- etc : ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ
>
| 1.0 | [feature] Home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์ ์์
- ## ๐ฅ ์ค๋ช
(์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์!)
- ๋๋ต์ ์ธ ์์ด์ดํ๋ ์์ ๋ฐ๋ผ Home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์๋ง ์ก์๋๊ฒ ์ต๋๋ค.
<br>
## ๐ฅ ETC (๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์!)
<img width="352" alt="image" src="https://user-images.githubusercontent.com/80577900/181917347-e750df32-bb28-44cc-aa30-421c73f7f690.png">
<br>
>#### types
>- feat : ์๋ก์ด ๊ธฐ๋ฅ
>- update : ๋ก์ง ๋ณ๊ฒฝ
>- bug : ๋ฒ๊ทธ ์์
>- refactor : ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง
>- style : ์ฝ๋ ํ์ ๋ณ๊ฒฝ
>- docs : ๋ฌธ์ ๊ด๋ จ
>- test : ํ
์คํธ ์ฝ๋
>- etc : ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ
>
| priority | home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์ ์์
๐ฅ ์ค๋ช
์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์ ๋๋ต์ ์ธ ์์ด์ดํ๋ ์์ ๋ฐ๋ผ home ์ปดํฌ๋ํธ ์์ด์ด ํ๋ ์๋ง ์ก์๋๊ฒ ์ต๋๋ค ๐ฅ etc ๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์ img width alt image src types feat ์๋ก์ด ๊ธฐ๋ฅ update ๋ก์ง ๋ณ๊ฒฝ bug ๋ฒ๊ทธ ์์ refactor ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง style ์ฝ๋ ํ์ ๋ณ๊ฒฝ docs ๋ฌธ์ ๊ด๋ จ test ํ
์คํธ ์ฝ๋ etc ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ | 1 |
555,334 | 16,451,932,507 | IssuesEvent | 2021-05-21 07:14:48 | scality/metalk8s | https://api.github.com/repos/scality/metalk8s | opened | Add a pre check to avoid deploying a node using the IP of an already deployed node | priority:medium topic:operations topic:ui | **Component**:
'salt', 'ui'
**Why this is needed**:
The administrator may (by mistake) create a Nde specification using the IP of an already deployed node. When deploying such node, this leads to failure and may have an impact on salt-minion and key negotiation
**What should be done**:
When adding a node, from the UI or from the CLI, we should check that the IP to reach the node is not already referenced in the Nodes that are already part of the cluster.
**Implementation proposal** (strongly recommended):
**Test plan**:
| 1.0 | Add a pre check to avoid deploying a node using the IP of an already deployed node - **Component**:
'salt', 'ui'
**Why this is needed**:
The administrator may (by mistake) create a Nde specification using the IP of an already deployed node. When deploying such node, this leads to failure and may have an impact on salt-minion and key negotiation
**What should be done**:
When adding a node, from the UI or from the CLI, we should check that the IP to reach the node is not already referenced in the Nodes that are already part of the cluster.
**Implementation proposal** (strongly recommended):
**Test plan**:
| priority | add a pre check to avoid deploying a node using the ip of an already deployed node component salt ui why this is needed the administrator may by mistake create a nde specification using the ip of an already deployed node when deploying such node this leads to failure and may have an impact on salt minion and key negotiation what should be done when adding a node from the ui or from the cli we should check that the ip to reach the node is not already referenced in the nodes that are already part of the cluster implementation proposal strongly recommended test plan | 1 |
808,286 | 30,054,268,792 | IssuesEvent | 2023-06-28 05:02:35 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] Skip adding global limit when possible | kind/enhancement area/docdb priority/medium | Jira Link: [DB-6725](https://yugabyte.atlassian.net/browse/DB-6725)
### Description
IntentAwareIterator adds global limit to the key during seek.
It is necessary to avoid iterating over too new records.
But we already have logic to filter out such records in SkipFutureRecords.
So no reason to add global limit when global limit is greater than current safe time.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-6725]: https://yugabyte.atlassian.net/browse/DB-6725?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [DocDB] Skip adding global limit when possible - Jira Link: [DB-6725](https://yugabyte.atlassian.net/browse/DB-6725)
### Description
IntentAwareIterator adds global limit to the key during seek.
It is necessary to avoid iterating over too new records.
But we already have logic to filter out such records in SkipFutureRecords.
So no reason to add global limit when global limit is greater than current safe time.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-6725]: https://yugabyte.atlassian.net/browse/DB-6725?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | skip adding global limit when possible jira link description intentawareiterator adds global limit to the key during seek it is necessary to avoid iterating over too new records but we already have logic to filter out such records in skipfuturerecords so no reason to add global limit when global limit is greater than current safe time warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information | 1 |
502,210 | 14,542,390,082 | IssuesEvent | 2020-12-15 15:39:34 | thehyve/ohdsi-etl-template | https://api.github.com/repos/thehyve/ohdsi-etl-template | reopened | Update logging code | medium priority refactoring | - [ ] remove obsolete logging code
- [ ] Add title lines to be able to easily distinguish different stages of the pipeline
I found and marked additional log messages in the template wrapper and main which could be integrated in the wrapper package.
Figure which ones are superfluous and which useful, and remove/refactor accordingly.
in `main.py`:
https://github.com/thehyve/ohdsi-etl-template/blob/d99b63f3cfe034ad84ad75bffd8708ee981e0245/main.py#L49-L54
in `wrapper.py`: (`run()`)
https://github.com/thehyve/ohdsi-etl-template/blob/d99b63f3cfe034ad84ad75bffd8708ee981e0245/src/main/python/wrapper.py#L54-L90
| 1.0 | Update logging code - - [ ] remove obsolete logging code
- [ ] Add title lines to be able to easily distinguish different stages of the pipeline
I found and marked additional log messages in the template wrapper and main which could be integrated in the wrapper package.
Figure which ones are superfluous and which useful, and remove/refactor accordingly.
in `main.py`:
https://github.com/thehyve/ohdsi-etl-template/blob/d99b63f3cfe034ad84ad75bffd8708ee981e0245/main.py#L49-L54
in `wrapper.py`: (`run()`)
https://github.com/thehyve/ohdsi-etl-template/blob/d99b63f3cfe034ad84ad75bffd8708ee981e0245/src/main/python/wrapper.py#L54-L90
| priority | update logging code remove obsolete logging code add title lines to be able to easily distinguish different stages of the pipeline i found and marked additional log messages in the template wrapper and main which could be integrated in the wrapper package figure which ones are superfluous and which useful and remove refactor accordingly in main py in wrapper py run | 1 |
803,018 | 29,088,271,711 | IssuesEvent | 2023-05-16 03:05:59 | masastack/MASA.Alert | https://api.github.com/repos/masastack/MASA.Alert | closed | Process the alarm history of the state, add the "modify handler" function | type/demand severity/medium status/resolved site/staging priority/p3 | ๅค็ไธญ็ถๆ็ๅ่ญฆๅๅฒ๏ผๅขๅ โไฟฎๆนๅค็ไบบโๅ่ฝ๏ผ็นๅปๅ๏ผๆๅผ่ฟๅ
ฅๅค็ๅ่ญฆ้กต้ข๏ผๅฏไฟฎๆน.

| 1.0 | Process the alarm history of the state, add the "modify handler" function - ๅค็ไธญ็ถๆ็ๅ่ญฆๅๅฒ๏ผๅขๅ โไฟฎๆนๅค็ไบบโๅ่ฝ๏ผ็นๅปๅ๏ผๆๅผ่ฟๅ
ฅๅค็ๅ่ญฆ้กต้ข๏ผๅฏไฟฎๆน.

| priority | process the alarm history of the state add the modify handler function ๅค็ไธญ็ถๆ็ๅ่ญฆๅๅฒ๏ผๅขๅ โไฟฎๆนๅค็ไบบโๅ่ฝ๏ผ็นๅปๅ๏ผๆๅผ่ฟๅ
ฅๅค็ๅ่ญฆ้กต้ข๏ผๅฏไฟฎๆน | 1 |
55,962 | 3,075,584,467 | IssuesEvent | 2015-08-20 14:23:38 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | Solo not gettin initiated after opening other Activity | bug imported invalid Priority-Medium | _From [reneprei...@gmail.com](https://code.google.com/u/100831329948317499403/) on May 08, 2014 06:48:54_
What steps will reproduce the problem? 1. Run two tests for an application
2. The first test starts a function which opens another application
(for example writing a message, opening a picture or the browser)
3. When the first test is finished, the second test is only getting to the setUp() method and is not creating a new Solo (the test is doing nothing and only finishes when I stop the test manually) What is the expected output? What do you see instead? I expect a normal test run (especially when both tests test the same activity) What version of the product are you using? On what operating system? Android: 4.0.2 (Emulator)
Robotium: 5.0.1 Please provide any additional information below. This problem accures everytime and even if I run one test two times with a test suite. (The first test succeeds and opens the browser and the second test
does start but is doing nothing since it hangs up at the point where Solo gets instantiated)
Thank you very much for your time!
_Original issue: http://code.google.com/p/robotium/issues/detail?id=603_ | 1.0 | Solo not gettin initiated after opening other Activity - _From [reneprei...@gmail.com](https://code.google.com/u/100831329948317499403/) on May 08, 2014 06:48:54_
What steps will reproduce the problem? 1. Run two tests for an application
2. The first test starts a function which opens another application
(for example writing a message, opening a picture or the browser)
3. When the first test is finished, the second test is only getting to the setUp() method and is not creating a new Solo (the test is doing nothing and only finishes when I stop the test manually) What is the expected output? What do you see instead? I expect a normal test run (especially when both tests test the same activity) What version of the product are you using? On what operating system? Android: 4.0.2 (Emulator)
Robotium: 5.0.1 Please provide any additional information below. This problem accures everytime and even if I run one test two times with a test suite. (The first test succeeds and opens the browser and the second test
does start but is doing nothing since it hangs up at the point where Solo gets instantiated)
Thank you very much for your time!
_Original issue: http://code.google.com/p/robotium/issues/detail?id=603_ | priority | solo not gettin initiated after opening other activity from on may what steps will reproduce the problem run two tests for an application the first test starts a function which opens another application for example writing a message opening a picture or the browser when the first test is finished the second test is only getting to the setup method and is not creating a new solo the test is doing nothing and only finishes when i stop the test manually what is the expected output what do you see instead i expect a normal test run especially when both tests test the same activity what version of the product are you using on what operating system android emulator robotium please provide any additional information below this problem accures everytime and even if i run one test two times with a test suite the first test succeeds and opens the browser and the second test does start but is doing nothing since it hangs up at the point where solo gets instantiated thank you very much for your time original issue | 1 |
449,179 | 12,964,605,220 | IssuesEvent | 2020-07-20 20:47:01 | radical-cybertools/radical.pilot | https://api.github.com/repos/radical-cybertools/radical.pilot | closed | Incorrect # cores calculation? | comp:agent:bootstrapper layer:saga priority:medium topic:resource type:bug | Asking 272+ CPUs (68 cores * 4 hardware threads) in EnTK launches 2 KNL nodes on Stampede2, not 68+ CPUs. This also raises ValueError from base.py in lrms like:
```
(base) login2(1192)$ more agent_0.err
Traceback (most recent call last):
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/bin/radical-pil
ot-agent", line 71, in <module>
bootstrap_3(sys.argv[1])
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/bin/radical-pil
ot-agent", line 42, in bootstrap_3
if agent_name == 'agent_0': agent = rpa.Agent_0(agent_name)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/agent_0.py", line 146, in __init__
session=self._session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/base.py", line 242, in create
return impl(cfg, session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/slurm.py", line 22, in __init__
LRMS.__init__(self, cfg, session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/base.py", line 171, in __init__
% (str(cores_avail), str(self.requested_cores)))
ValueError: Not enough cores available (136 < 340).
```
Tried to enable `RADICAL_DEBUG` as it says the ValueError is triggered when it's not enabled but the error is persistent. | 1.0 | Incorrect # cores calculation? - Asking 272+ CPUs (68 cores * 4 hardware threads) in EnTK launches 2 KNL nodes on Stampede2, not 68+ CPUs. This also raises ValueError from base.py in lrms like:
```
(base) login2(1192)$ more agent_0.err
Traceback (most recent call last):
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/bin/radical-pil
ot-agent", line 71, in <module>
bootstrap_3(sys.argv[1])
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/bin/radical-pil
ot-agent", line 42, in bootstrap_3
if agent_name == 'agent_0': agent = rpa.Agent_0(agent_name)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/agent_0.py", line 146, in __init__
session=self._session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/base.py", line 242, in create
return impl(cfg, session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/slurm.py", line 22, in __init__
LRMS.__init__(self, cfg, session)
File "/work/06079/tg853783/stampede2/radical.pilot.sandbox/re.session.js-17-248.jetstream-cloud.org.hrlee.018124.0014/pilot.0000/rp_install/lib/python2.7/s
ite-packages/radical/pilot/agent/rm/base.py", line 171, in __init__
% (str(cores_avail), str(self.requested_cores)))
ValueError: Not enough cores available (136 < 340).
```
Tried to enable `RADICAL_DEBUG` as it says the ValueError is triggered when it's not enabled but the error is persistent. | priority | incorrect cores calculation asking cpus cores hardware threads in entk launches knl nodes on not cpus this also raises valueerror from base py in lrms like base more agent err traceback most recent call last file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install bin radical pil ot agent line in bootstrap sys argv file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install bin radical pil ot agent line in bootstrap if agent name agent agent rpa agent agent name file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install lib s ite packages radical pilot agent agent py line in init session self session file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install lib s ite packages radical pilot agent rm base py line in create return impl cfg session file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install lib s ite packages radical pilot agent rm slurm py line in init lrms init self cfg session file work radical pilot sandbox re session js jetstream cloud org hrlee pilot rp install lib s ite packages radical pilot agent rm base py line in init str cores avail str self requested cores valueerror not enough cores available tried to enable radical debug as it says the valueerror is triggered when it s not enabled but the error is persistent | 1 |
77,572 | 3,507,160,948 | IssuesEvent | 2016-01-08 11:36:38 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Where MMAPS (BB #556) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 16.03.2014 20:16:45 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/556
<hr>
Please where download MMAPS for OREGON CORE please? thx
| 1.0 | Where MMAPS (BB #556) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 16.03.2014 20:16:45 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/556
<hr>
Please where download MMAPS for OREGON CORE please? thx
| priority | where mmaps bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state invalid direct link please where download mmaps for oregon core please thx | 1 |
653,430 | 21,582,038,664 | IssuesEvent | 2022-05-02 19:50:47 | vdjagilev/nmap-formatter | https://api.github.com/repos/vdjagilev/nmap-formatter | closed | Move scan summary `Services` field under the table | type/bug priority/medium tech/markdown | `Services` field often contains a big list of ports which is problematic to display in markdown table (it has to be hidden using `<details><summary>...` html tags). Move this part in a sub-chapter, enclosing it in ` ``` ` would be better. | 1.0 | Move scan summary `Services` field under the table - `Services` field often contains a big list of ports which is problematic to display in markdown table (it has to be hidden using `<details><summary>...` html tags). Move this part in a sub-chapter, enclosing it in ` ``` ` would be better. | priority | move scan summary services field under the table services field often contains a big list of ports which is problematic to display in markdown table it has to be hidden using html tags move this part in a sub chapter enclosing it in would be better | 1 |
721,909 | 24,843,211,458 | IssuesEvent | 2022-10-26 14:08:44 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] speedup split child tablet local bootstrap by copying in-memory retryable requests structure from parent tablet | kind/enhancement area/docdb priority/medium 2.14 Backport Required | Jira Link: [DB-3717](https://yugabyte.atlassian.net/browse/DB-3717)
### Description
Split child tablets need to replay 660 secs worth of WAL entries to recover retryable requests structure, causing a huge unavailability window. We can copy the structure directly from its parent (if still available on the current tserver) to speedup the local bootstrap. | 1.0 | [DocDB] speedup split child tablet local bootstrap by copying in-memory retryable requests structure from parent tablet - Jira Link: [DB-3717](https://yugabyte.atlassian.net/browse/DB-3717)
### Description
Split child tablets need to replay 660 secs worth of WAL entries to recover retryable requests structure, causing a huge unavailability window. We can copy the structure directly from its parent (if still available on the current tserver) to speedup the local bootstrap. | priority | speedup split child tablet local bootstrap by copying in memory retryable requests structure from parent tablet jira link description split child tablets need to replay secs worth of wal entries to recover retryable requests structure causing a huge unavailability window we can copy the structure directly from its parent if still available on the current tserver to speedup the local bootstrap | 1 |
426,762 | 12,378,978,387 | IssuesEvent | 2020-05-19 11:40:58 | pingcap/chaos-mesh | https://api.github.com/repos/pingcap/chaos-mesh | closed | Use templates to simplify the complexity of configuring sidecar ConfigMap | component/operator difficulty/medium priority/P1 | ## Feature Request
**Is your feature request related to a problem? Please describe:**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
When we want to inject IOChaos, we need to define a ConfigMap for your application before starting chaos experiment and the ConfigMap have many fields that need to be full. And we found that most of the field values in different configmaps do not need to be changed
**Describe the feature you'd like:**
<!-- A clear and concise description of what you want to happen. -->
We can use a template to help define ConfigMap and the users just to set the PV path and the data directory of the application.
For example, we can define a template named `chaosfs-configmap` and we can use it to define a sidecar ConfigMap for tikv:
```
xxxx
template: chaosfs
arguments:
dataPath: /var/lib/tikv/data
pvName: tikv
pvPath: xxx
```
| 1.0 | Use templates to simplify the complexity of configuring sidecar ConfigMap - ## Feature Request
**Is your feature request related to a problem? Please describe:**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
When we want to inject IOChaos, we need to define a ConfigMap for your application before starting chaos experiment and the ConfigMap have many fields that need to be full. And we found that most of the field values in different configmaps do not need to be changed
**Describe the feature you'd like:**
<!-- A clear and concise description of what you want to happen. -->
We can use a template to help define ConfigMap and the users just to set the PV path and the data directory of the application.
For example, we can define a template named `chaosfs-configmap` and we can use it to define a sidecar ConfigMap for tikv:
```
xxxx
template: chaosfs
arguments:
dataPath: /var/lib/tikv/data
pvName: tikv
pvPath: xxx
```
| priority | use templates to simplify the complexity of configuring sidecar configmap feature request is your feature request related to a problem please describe when we want to inject iochaos we need to define a configmap for your application before starting chaos experiment and the configmap have many fields that need to be full and we found that most of the field values in different configmaps do not need to be changed describe the feature you d like we can use a template to help define configmap and the users just to set the pv path and the data directory of the application for example we can define a template named chaosfs configmap and we can use it to define a sidecar configmap for tikv xxxx template chaosfs arguments datapath var lib tikv data pvname tikv pvpath xxx | 1 |
802,213 | 28,807,726,624 | IssuesEvent | 2023-05-03 00:04:31 | Radarr/Radarr | https://api.github.com/repos/Radarr/Radarr | closed | Unable to search for movies from movie list and "Search" tab of movie crashing | Type: Bug Priority: Medium Status: Confirmed Type: Regression | ### Is there an existing issue for this?
- [X] I have searched the existing open and closed issues
### Current Behavior
If you click on the little magnifying glass of a movie, search fails. So does the search tab from a movie view
### Expected Behavior
Search, search, search
### Steps To Reproduce
Not sure if it's Linux and/or docker specific, but I presume it's related to the latest version
### Environment
```markdown
- OS: 23.04
- Radarr: 4.5.1.7175
- Docker Install: Yes
- Using Reverse Proxy: No
- Browser: Any
- Database: Sqlite 3.40.1
```
### What branch are you running?
Nightly
### Trace Logs?
https://gist.github.com/PartTheSeas/496089c345d7b854dc6716f2b8fc3e84 | 1.0 | Unable to search for movies from movie list and "Search" tab of movie crashing - ### Is there an existing issue for this?
- [X] I have searched the existing open and closed issues
### Current Behavior
If you click on the little magnifying glass of a movie, search fails. So does the search tab from a movie view
### Expected Behavior
Search, search, search
### Steps To Reproduce
Not sure if it's Linux and/or docker specific, but I presume it's related to the latest version
### Environment
```markdown
- OS: 23.04
- Radarr: 4.5.1.7175
- Docker Install: Yes
- Using Reverse Proxy: No
- Browser: Any
- Database: Sqlite 3.40.1
```
### What branch are you running?
Nightly
### Trace Logs?
https://gist.github.com/PartTheSeas/496089c345d7b854dc6716f2b8fc3e84 | priority | unable to search for movies from movie list and search tab of movie crashing is there an existing issue for this i have searched the existing open and closed issues current behavior if you click on the little magnifying glass of a movie search fails so does the search tab from a movie view expected behavior search search search steps to reproduce not sure if it s linux and or docker specific but i presume it s related to the latest version environment markdown os radarr docker install yes using reverse proxy no browser any database sqlite what branch are you running nightly trace logs | 1 |
551,072 | 16,137,020,708 | IssuesEvent | 2021-04-29 13:08:03 | geosolutions-it/geonode | https://api.github.com/repos/geosolutions-it/geonode | closed | Old layers and maps should be delete automatically on "master.demo" | Priority: Medium | ## Expected Behavior
Every Layer, Map and Document should be automatically deleted every Sunday @midnight
## Actual Behavior
Resources are not automatically removed anymore
## Steps to Reproduce the Problem
1.
2.
3.
## Specifications
- GeoNode version:
- Installation method (manual, GeoNode Docker, SPCGeoNode Docker):
- Platform:
- Additional details:
| 1.0 | Old layers and maps should be delete automatically on "master.demo" - ## Expected Behavior
Every Layer, Map and Document should be automatically deleted every Sunday @midnight
## Actual Behavior
Resources are not automatically removed anymore
## Steps to Reproduce the Problem
1.
2.
3.
## Specifications
- GeoNode version:
- Installation method (manual, GeoNode Docker, SPCGeoNode Docker):
- Platform:
- Additional details:
| priority | old layers and maps should be delete automatically on master demo expected behavior every layer map and document should be automatically deleted every sunday midnight actual behavior resources are not automatically removed anymore steps to reproduce the problem specifications geonode version installation method manual geonode docker spcgeonode docker platform additional details | 1 |
24,762 | 2,672,809,046 | IssuesEvent | 2015-03-24 15:57:50 | 52North/SOS | https://api.github.com/repos/52North/SOS | opened | Settings Export: Operations disabled stated not restored correctly | bug medium priority | The current implementation of settings export and import does not correctly restore the enabled/disabled state of operations.
In addition, it should include the datasource definition, too. | 1.0 | Settings Export: Operations disabled stated not restored correctly - The current implementation of settings export and import does not correctly restore the enabled/disabled state of operations.
In addition, it should include the datasource definition, too. | priority | settings export operations disabled stated not restored correctly the current implementation of settings export and import does not correctly restore the enabled disabled state of operations in addition it should include the datasource definition too | 1 |
1,671 | 2,517,431,818 | IssuesEvent | 2015-01-16 14:47:18 | Metaswitch/crest | https://api.github.com/repos/Metaswitch/crest | closed | Bulk provisioning scripts do not create the `impi_mapping` table in `homestead_cache` | bug medium-priority | This applies both to the cassandra-cli and sstable provisioner. Looks like they were not updated following the addition of registration termination support. | 1.0 | Bulk provisioning scripts do not create the `impi_mapping` table in `homestead_cache` - This applies both to the cassandra-cli and sstable provisioner. Looks like they were not updated following the addition of registration termination support. | priority | bulk provisioning scripts do not create the impi mapping table in homestead cache this applies both to the cassandra cli and sstable provisioner looks like they were not updated following the addition of registration termination support | 1 |
76,480 | 3,488,456,318 | IssuesEvent | 2016-01-02 23:58:51 | jasonbellamy/git-label | https://api.github.com/repos/jasonbellamy/git-label | opened | Accept filepath OR directory path of packages | Priority: Medium Status: Available Type: Enhancement | Instead of only accepting an array of file paths, it might be beneficial to accept an array of both files and/or directories of files. | 1.0 | Accept filepath OR directory path of packages - Instead of only accepting an array of file paths, it might be beneficial to accept an array of both files and/or directories of files. | priority | accept filepath or directory path of packages instead of only accepting an array of file paths it might be beneficial to accept an array of both files and or directories of files | 1 |
474,190 | 13,653,982,107 | IssuesEvent | 2020-09-27 15:18:07 | rubyforgood/diaper | https://api.github.com/repos/rubyforgood/diaper | closed | Integrate brakeman security scans to our Travis CI checks | In Progress Medium Priority โก๏ธ Power โก๏ธ | # Summary
We should run security scans to our TravisCI checks to better ensure we don't introduce security vulnerabilities. Let's update our Travis configuration to also run these checks and let us know if we should fix any vulnerabilities.
https://brakemanscanner.org/docs/quickstart/
## Things to Consider
We should add this also to the partner application https://github.com/rubyforgood/partner
# Criteria for Completion
- [ ] TravisCI tests should fail if a pull request has a security vulnerability | 1.0 | Integrate brakeman security scans to our Travis CI checks - # Summary
We should run security scans to our TravisCI checks to better ensure we don't introduce security vulnerabilities. Let's update our Travis configuration to also run these checks and let us know if we should fix any vulnerabilities.
https://brakemanscanner.org/docs/quickstart/
## Things to Consider
We should add this also to the partner application https://github.com/rubyforgood/partner
# Criteria for Completion
- [ ] TravisCI tests should fail if a pull request has a security vulnerability | priority | integrate brakeman security scans to our travis ci checks summary we should run security scans to our travisci checks to better ensure we don t introduce security vulnerabilities let s update our travis configuration to also run these checks and let us know if we should fix any vulnerabilities things to consider we should add this also to the partner application criteria for completion travisci tests should fail if a pull request has a security vulnerability | 1 |
54,917 | 3,071,607,058 | IssuesEvent | 2015-08-19 13:09:08 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | Click on text not working | bug imported Priority-Medium wontfix | _From [glenview...@gmail.com](https://code.google.com/u/110087215095127878251/) on October 15, 2010 21:46:13_
I've been using Robotium now quite successfully to run around 30 different solo method calls in various tests. I was using 1.7.1, and tried the new 1.8.0, as well as the suggestions to modify the manifest. Nothing changes this behavior.
I'm having some trouble coaxing it to click a textview though. It's a nested view structure with a scrollview that has a textview and a bunch of buttons inside.
I tried solo.clickOnText(".*Show.*") but Solo responds that it couldn't find the text even though it's clearly there on screen. I even stepped through the test to confirm.
I tried then using clickOnScreen as a kludge. It clicked on the screen and enabled the buttons; however then, the clickOnButton method exhibited the same behavior.
By the way, it would be nice if you posted a snapshot of the source to the downloads from now on with each version so I don't have to use git and try to figure out which is the right source version. That way I can try to debug it myself.
Thanks!
_Original issue: http://code.google.com/p/robotium/issues/detail?id=23_ | 1.0 | Click on text not working - _From [glenview...@gmail.com](https://code.google.com/u/110087215095127878251/) on October 15, 2010 21:46:13_
I've been using Robotium now quite successfully to run around 30 different solo method calls in various tests. I was using 1.7.1, and tried the new 1.8.0, as well as the suggestions to modify the manifest. Nothing changes this behavior.
I'm having some trouble coaxing it to click a textview though. It's a nested view structure with a scrollview that has a textview and a bunch of buttons inside.
I tried solo.clickOnText(".*Show.*") but Solo responds that it couldn't find the text even though it's clearly there on screen. I even stepped through the test to confirm.
I tried then using clickOnScreen as a kludge. It clicked on the screen and enabled the buttons; however then, the clickOnButton method exhibited the same behavior.
By the way, it would be nice if you posted a snapshot of the source to the downloads from now on with each version so I don't have to use git and try to figure out which is the right source version. That way I can try to debug it myself.
Thanks!
_Original issue: http://code.google.com/p/robotium/issues/detail?id=23_ | priority | click on text not working from on october i ve been using robotium now quite successfully to run around different solo method calls in various tests i was using and tried the new as well as the suggestions to modify the manifest nothing changes this behavior i m having some trouble coaxing it to click a textview though it s a nested view structure with a scrollview that has a textview and a bunch of buttons inside i tried solo clickontext show but solo responds that it couldn t find the text even though it s clearly there on screen i even stepped through the test to confirm i tried then using clickonscreen as a kludge it clicked on the screen and enabled the buttons however then the clickonbutton method exhibited the same behavior by the way it would be nice if you posted a snapshot of the source to the downloads from now on with each version so i don t have to use git and try to figure out which is the right source version that way i can try to debug it myself thanks original issue | 1 |
84,275 | 3,656,070,232 | IssuesEvent | 2016-02-17 18:29:26 | jakev/dtfmods-core | https://api.github.com/repos/jakev/dtfmods-core | opened | Add capablities for analyzing backups | enhancement priority-medium Status: In Progress | This will be a component update:
* Add "allow_backup" to `sysapps.db` schema (modifications to `sysappdb` and `AppDb`)
* Add Nikolay Elenkov's `abe.jar` as a binary + module wrapper for creating / uploading backups
* Add module for finding apps that are backup-able: `findbackups`. This should be more than just allowBackup, and should look for a backupagent
| 1.0 | Add capablities for analyzing backups - This will be a component update:
* Add "allow_backup" to `sysapps.db` schema (modifications to `sysappdb` and `AppDb`)
* Add Nikolay Elenkov's `abe.jar` as a binary + module wrapper for creating / uploading backups
* Add module for finding apps that are backup-able: `findbackups`. This should be more than just allowBackup, and should look for a backupagent
| priority | add capablities for analyzing backups this will be a component update add allow backup to sysapps db schema modifications to sysappdb and appdb add nikolay elenkov s abe jar as a binary module wrapper for creating uploading backups add module for finding apps that are backup able findbackups this should be more than just allowbackup and should look for a backupagent | 1 |
57,047 | 3,081,234,135 | IssuesEvent | 2015-08-22 14:22:36 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Mysterious "empty vertex problem" | 019 bug imported Priority-Medium | _From [bkcon...@gmail.com](https://code.google.com/u/112584966209893525339/) on October 31, 2013 00:52:14_
Wasn't able to make a reproduction case, but I caught this crash in gdb while hovering over a certain polywall
Assert: Empty vertex problem in /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp line 2651
Program received signal SIGTRAP, Trace/breakpoint trap.
Zap::EditorUserInterface::checkForEdgeHit (this=0x12b3530, point=..., object=
0x145d160) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2653
2653 bool loop = (object->getGeomType() == geomPolygon);
(gdb) bt
\#0 Zap::EditorUserInterface::checkForEdgeHit (this=0x12b3530, point=...,
object=0x145d160) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2653
\#1 0x0000000000aad012 in Zap::EditorUserInterface::findHitItemAndEdge (
this=0x12b3530) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2599
\#2 0x0000000000aadce6 in Zap::EditorUserInterface::onMouseMoved (
this=0x12b3530) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2816
\#3 0x0000000000a05bb0 in Zap::Event::onMouseMoved (currentUI=0x12b3530,
x=720, y=443, mode=Zap::DISPLAY_MODE_WINDOWED)
at /home/kaen/code/bitfighter-clean/zap/Event.cpp:393
\#4 0x0000000000a05254 in Zap::Event::onEvent (game=0x11f7370,
event=0x7fffffffdd20) at /home/kaen/code/bitfighter-clean/zap/Event.cpp:216
\#5 0x0000000000b20576 in Zap::idle ()
at /home/kaen/code/bitfighter-clean/zap/main.cpp:405
\#6 0x0000000000b20611 in Zap::dedicatedServerLoop ()
at /home/kaen/code/bitfighter-clean/zap/main.cpp:427
\#7 0x0000000000b22f49 in main (argc=5, argv=0x7fffffffdfc8)
at /home/kaen/code/bitfighter-clean/zap/main.cpp:1217
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=277_ | 1.0 | Mysterious "empty vertex problem" - _From [bkcon...@gmail.com](https://code.google.com/u/112584966209893525339/) on October 31, 2013 00:52:14_
Wasn't able to make a reproduction case, but I caught this crash in gdb while hovering over a certain polywall
Assert: Empty vertex problem in /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp line 2651
Program received signal SIGTRAP, Trace/breakpoint trap.
Zap::EditorUserInterface::checkForEdgeHit (this=0x12b3530, point=..., object=
0x145d160) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2653
2653 bool loop = (object->getGeomType() == geomPolygon);
(gdb) bt
\#0 Zap::EditorUserInterface::checkForEdgeHit (this=0x12b3530, point=...,
object=0x145d160) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2653
\#1 0x0000000000aad012 in Zap::EditorUserInterface::findHitItemAndEdge (
this=0x12b3530) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2599
\#2 0x0000000000aadce6 in Zap::EditorUserInterface::onMouseMoved (
this=0x12b3530) at /home/kaen/code/bitfighter-clean/zap/UIEditor.cpp:2816
\#3 0x0000000000a05bb0 in Zap::Event::onMouseMoved (currentUI=0x12b3530,
x=720, y=443, mode=Zap::DISPLAY_MODE_WINDOWED)
at /home/kaen/code/bitfighter-clean/zap/Event.cpp:393
\#4 0x0000000000a05254 in Zap::Event::onEvent (game=0x11f7370,
event=0x7fffffffdd20) at /home/kaen/code/bitfighter-clean/zap/Event.cpp:216
\#5 0x0000000000b20576 in Zap::idle ()
at /home/kaen/code/bitfighter-clean/zap/main.cpp:405
\#6 0x0000000000b20611 in Zap::dedicatedServerLoop ()
at /home/kaen/code/bitfighter-clean/zap/main.cpp:427
\#7 0x0000000000b22f49 in main (argc=5, argv=0x7fffffffdfc8)
at /home/kaen/code/bitfighter-clean/zap/main.cpp:1217
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=277_ | priority | mysterious empty vertex problem from on october wasn t able to make a reproduction case but i caught this crash in gdb while hovering over a certain polywall assert empty vertex problem in home kaen code bitfighter clean zap uieditor cpp line program received signal sigtrap trace breakpoint trap zap editoruserinterface checkforedgehit this point object at home kaen code bitfighter clean zap uieditor cpp bool loop object getgeomtype geompolygon gdb bt zap editoruserinterface checkforedgehit this point object at home kaen code bitfighter clean zap uieditor cpp in zap editoruserinterface findhititemandedge this at home kaen code bitfighter clean zap uieditor cpp in zap editoruserinterface onmousemoved this at home kaen code bitfighter clean zap uieditor cpp in zap event onmousemoved currentui x y mode zap display mode windowed at home kaen code bitfighter clean zap event cpp in zap event onevent game event at home kaen code bitfighter clean zap event cpp in zap idle at home kaen code bitfighter clean zap main cpp in zap dedicatedserverloop at home kaen code bitfighter clean zap main cpp in main argc argv at home kaen code bitfighter clean zap main cpp original issue | 1 |
588,199 | 17,649,789,034 | IssuesEvent | 2021-08-20 11:34:54 | Javacord/Javacord | https://api.github.com/repos/Javacord/Javacord | closed | Make UncachedMessageUtil support multiple embeds and use "embeds" instead of "embed" | medium priority | The "embed" field on message has been depreacted in favor of "embeds", but the UncachedMessageUtil is still using the old one in some cases. | 1.0 | Make UncachedMessageUtil support multiple embeds and use "embeds" instead of "embed" - The "embed" field on message has been depreacted in favor of "embeds", but the UncachedMessageUtil is still using the old one in some cases. | priority | make uncachedmessageutil support multiple embeds and use embeds instead of embed the embed field on message has been depreacted in favor of embeds but the uncachedmessageutil is still using the old one in some cases | 1 |
691,879 | 23,714,928,322 | IssuesEvent | 2022-08-30 10:57:01 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | closed | Preload doesnโt stop when asked to | type: bug module: preload priority: medium status: blocked effort: [M] future severity: major | **Describe the bug**
The "Stop preload" button doesnโt stop the preload when sitemap preloading is enabled.
**To Reproduce**
Steps to reproduce the behavior:
1. Choose a site with a "large" number of pages
2. In the Preload section from WPRโs settings, enable preloading + sitemap-based cache preloading, and fill in the sitemap URL.
3. Once the settings saved, hit the "Preload cache" button from the admin bar or in the quick actions sidebar from WPRโs settings page.
4. The banner displaying the number of paged that have been preloaded will appear.
5. Hit the "Stop preload" button.
6. The same banner should still be displayed with the same button: preloading didnโt stop.
**Expected behavior**
Preloading should stop and the result banner should appear.
**Source of the bug**
The reason of this bug is quite simple: the "stop" action only cancels the homepage preloading, not the sitemap preloading. See `WP_Rocket\Engine\Preload\PreloadSubscriber->do_admin_post_stop_preload()`:
https://github.com/wp-media/wp-rocket/blob/e154e170c205534398247e935cc5eaef070eaeeb/inc/Engine/Preload/PreloadSubscriber.php#L360-L374
**Backlog Grooming**
- [x] Reproduce the problem
- [x] Identify the root cause
- [x] Scope a solution
- [x] Estimate the effort | 1.0 | Preload doesnโt stop when asked to - **Describe the bug**
The "Stop preload" button doesnโt stop the preload when sitemap preloading is enabled.
**To Reproduce**
Steps to reproduce the behavior:
1. Choose a site with a "large" number of pages
2. In the Preload section from WPRโs settings, enable preloading + sitemap-based cache preloading, and fill in the sitemap URL.
3. Once the settings saved, hit the "Preload cache" button from the admin bar or in the quick actions sidebar from WPRโs settings page.
4. The banner displaying the number of paged that have been preloaded will appear.
5. Hit the "Stop preload" button.
6. The same banner should still be displayed with the same button: preloading didnโt stop.
**Expected behavior**
Preloading should stop and the result banner should appear.
**Source of the bug**
The reason of this bug is quite simple: the "stop" action only cancels the homepage preloading, not the sitemap preloading. See `WP_Rocket\Engine\Preload\PreloadSubscriber->do_admin_post_stop_preload()`:
https://github.com/wp-media/wp-rocket/blob/e154e170c205534398247e935cc5eaef070eaeeb/inc/Engine/Preload/PreloadSubscriber.php#L360-L374
**Backlog Grooming**
- [x] Reproduce the problem
- [x] Identify the root cause
- [x] Scope a solution
- [x] Estimate the effort | priority | preload doesnโt stop when asked to describe the bug the stop preload button doesnโt stop the preload when sitemap preloading is enabled to reproduce steps to reproduce the behavior choose a site with a large number of pages in the preload section from wprโs settings enable preloading sitemap based cache preloading and fill in the sitemap url once the settings saved hit the preload cache button from the admin bar or in the quick actions sidebar from wprโs settings page the banner displaying the number of paged that have been preloaded will appear hit the stop preload button the same banner should still be displayed with the same button preloading didnโt stop expected behavior preloading should stop and the result banner should appear source of the bug the reason of this bug is quite simple the stop action only cancels the homepage preloading not the sitemap preloading see wp rocket engine preload preloadsubscriber do admin post stop preload backlog grooming reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
795,844 | 28,088,788,505 | IssuesEvent | 2023-03-30 11:40:51 | kdt-final-3/salarying-be | https://api.github.com/repos/kdt-final-3/salarying-be | opened | feat: ์ธ์ฌ๊ด๋ฆฌ ์๋น์ค - ์กฐํ ๊ด๋ จ ๊ธฐ๋ฅ ๊ตฌํ | For: API Priority: Medium Status: In Progress Type: Feature | ## Description(์ค๋ช
)
์ธ์ฌ๊ด๋ฆฌ ์๋น์ค
- ์ ํํ ์ฑ์ฉํผ์ ๋ง๋ ์ ์ฒด ์ง์์ ์กฐํ
- ์ ํํ ์ฑ์ฉํผ์์ ์ ํํ ์ฑ์ฉ๋จ๊ณ์ ์๋ ์ง์์๋ฅผ ์กฐํ
## Tasks(New feature)
- [ ] ์ ์ฒด ์ง์์ ์กฐํ
- [ ] ๋จ๊ณ๋ณ ์ง์์ ์กฐํ | 1.0 | feat: ์ธ์ฌ๊ด๋ฆฌ ์๋น์ค - ์กฐํ ๊ด๋ จ ๊ธฐ๋ฅ ๊ตฌํ - ## Description(์ค๋ช
)
์ธ์ฌ๊ด๋ฆฌ ์๋น์ค
- ์ ํํ ์ฑ์ฉํผ์ ๋ง๋ ์ ์ฒด ์ง์์ ์กฐํ
- ์ ํํ ์ฑ์ฉํผ์์ ์ ํํ ์ฑ์ฉ๋จ๊ณ์ ์๋ ์ง์์๋ฅผ ์กฐํ
## Tasks(New feature)
- [ ] ์ ์ฒด ์ง์์ ์กฐํ
- [ ] ๋จ๊ณ๋ณ ์ง์์ ์กฐํ | priority | feat ์ธ์ฌ๊ด๋ฆฌ ์๋น์ค ์กฐํ ๊ด๋ จ ๊ธฐ๋ฅ ๊ตฌํ description ์ค๋ช
์ธ์ฌ๊ด๋ฆฌ ์๋น์ค ์ ํํ ์ฑ์ฉํผ์ ๋ง๋ ์ ์ฒด ์ง์์ ์กฐํ ์ ํํ ์ฑ์ฉํผ์์ ์ ํํ ์ฑ์ฉ๋จ๊ณ์ ์๋ ์ง์์๋ฅผ ์กฐํ tasks new feature ์ ์ฒด ์ง์์ ์กฐํ ๋จ๊ณ๋ณ ์ง์์ ์กฐํ | 1 |
170,603 | 6,460,131,259 | IssuesEvent | 2017-08-16 02:36:38 | ponylang/ponyc | https://api.github.com/repos/ponylang/ponyc | closed | Arrays of lambdas build a union typearg; could infer a common interface | difficulty: 2 - medium enhancement: 1 - needs discussion priority: 1 - low | The construct:
```
fun algorithms(): Array[{(I128): I128} val] ref^ => [
lambda(n: I128): I128 => Factorial.iterative(n) end
]
```
leads to the error message:
```
{(I128): I128} val is not a subtype of lambda(I128): I128 end val
```
which is not only a wee bit incomprehensible, but (I am reliably informed) is a compiler error that should be reported. Hence this issue.
| 1.0 | Arrays of lambdas build a union typearg; could infer a common interface - The construct:
```
fun algorithms(): Array[{(I128): I128} val] ref^ => [
lambda(n: I128): I128 => Factorial.iterative(n) end
]
```
leads to the error message:
```
{(I128): I128} val is not a subtype of lambda(I128): I128 end val
```
which is not only a wee bit incomprehensible, but (I am reliably informed) is a compiler error that should be reported. Hence this issue.
| priority | arrays of lambdas build a union typearg could infer a common interface the construct fun algorithms array ref lambda n factorial iterative n end leads to the error message val is not a subtype of lambda end val which is not only a wee bit incomprehensible but i am reliably informed is a compiler error that should be reported hence this issue | 1 |
77,271 | 3,506,331,807 | IssuesEvent | 2016-01-08 05:49:07 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | missing /win/ folder (BB #319) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 06.10.2010 20:49:05 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/319
<hr>
like in title, since few days, win folder is missing. can someone upload it? | 1.0 | missing /win/ folder (BB #319) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 06.10.2010 20:49:05 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/319
<hr>
like in title, since few days, win folder is missing. can someone upload it? | priority | missing win folder bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state invalid direct link like in title since few days win folder is missing can someone upload it | 1 |
27,791 | 2,696,178,370 | IssuesEvent | 2015-04-02 12:24:11 | jilm/control4j | https://api.github.com/repos/jilm/control4j | closed | Expand nested Use elements | enhancement Priority-Medium | It belongs to the block expansion #47
Use depth first search algorithm with cycle detection. | 1.0 | Expand nested Use elements - It belongs to the block expansion #47
Use depth first search algorithm with cycle detection. | priority | expand nested use elements it belongs to the block expansion use depth first search algorithm with cycle detection | 1 |
149,717 | 5,724,397,220 | IssuesEvent | 2017-04-20 14:27:36 | certificate-helper/TLS-Inspector | https://api.github.com/repos/certificate-helper/TLS-Inspector | closed | Tapping on a certificate in the list on iPad loads in side view | bug CHCertificate medium priority merged | **Affected Version:**
1.5 (22)
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
iPad only
**What steps will reproduce the problem?**
1. Inspect a domain
2. Tap any certificate in the list
**What is the expected output?**
Certificate loads on right hand view
**What do you see instead?**
Certificate loads on left hand view
**Please provide any additional information below.**
| 1.0 | Tapping on a certificate in the list on iPad loads in side view - **Affected Version:**
1.5 (22)
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
iPad only
**What steps will reproduce the problem?**
1. Inspect a domain
2. Tap any certificate in the list
**What is the expected output?**
Certificate loads on right hand view
**What do you see instead?**
Certificate loads on left hand view
**Please provide any additional information below.**
| priority | tapping on a certificate in the list on ipad loads in side view affected version is this a test flight version or the app store version app store device and ios version ipad only what steps will reproduce the problem inspect a domain tap any certificate in the list what is the expected output certificate loads on right hand view what do you see instead certificate loads on left hand view please provide any additional information below | 1 |
56,458 | 3,079,860,021 | IssuesEvent | 2015-08-21 18:40:07 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | ะะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั, ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
| bug duplicate imported Priority-Medium | _From [Destract...@pochta.ru](https://code.google.com/u/104951531913696676932/) on January 22, 2012 16:45:04_
ะะพั ะตัะต ะพะดะฝะฐ ะฝะตะธัะฟัะฐะฒะปะตะฝะฝะฐั ะฑะฐะณะฐ, ะฒะธัะธั ัะถะต ะฟะพะผะพะตะผั ั 500ะน ะฒะตััะธะธ..
ะะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั, ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
http://s018.radikal.ru/i522/1201/11/f7d2dea0f6d5.png ัะบัะธะฝ ะฒััะต
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=666_ | 1.0 | ะะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั, ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
- _From [Destract...@pochta.ru](https://code.google.com/u/104951531913696676932/) on January 22, 2012 16:45:04_
ะะพั ะตัะต ะพะดะฝะฐ ะฝะตะธัะฟัะฐะฒะปะตะฝะฝะฐั ะฑะฐะณะฐ, ะฒะธัะธั ัะถะต ะฟะพะผะพะตะผั ั 500ะน ะฒะตััะธะธ..
ะะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั, ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
http://s018.radikal.ru/i522/1201/11/f7d2dea0f6d5.png ัะบัะธะฝ ะฒััะต
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=666_ | priority | ะฝะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
from on january ะฒะพั ะตัะต ะพะดะฝะฐ ะฝะตะธัะฟัะฐะฒะปะตะฝะฝะฐั ะฑะฐะณะฐ ะฒะธัะธั ัะถะต ะฟะพะผะพะตะผั ั ะฒะตััะธะธ ะฝะต ัะฐะฑะพัะฐะตั ะฟัะฐะฒะธะปัะฝะพ ััะฝะบัะธั ะฟะพ ะบะปะธะบั ะฟัะฐะฒะพะน ะบะฝะพะฟะบะธ ะฝะฐ ะฟะพะปัะทะพะฒะฐัะตะปั ะธะฝัะพัะผะฐัะธั ะพ ะฟะพะปัะทะพะฒะฐัะตะปะต ะบะฐะบ ะฒ ัะฐะฝะฝะธั
ะฒะตััะธัั
ัะบัะธะฝ ะฒััะต original issue | 1 |
208,740 | 7,157,919,191 | IssuesEvent | 2018-01-26 21:50:50 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Placed logs rapidly with right click while carrying food, caused a client interaction desync | Medium Priority | The placed logs showed up, but couldn't be picked up again. Then the carried slot icon was out of sync (showed empty when it should have had more logs that I picked up). Re-logging fixed it. | 1.0 | Placed logs rapidly with right click while carrying food, caused a client interaction desync - The placed logs showed up, but couldn't be picked up again. Then the carried slot icon was out of sync (showed empty when it should have had more logs that I picked up). Re-logging fixed it. | priority | placed logs rapidly with right click while carrying food caused a client interaction desync the placed logs showed up but couldn t be picked up again then the carried slot icon was out of sync showed empty when it should have had more logs that i picked up re logging fixed it | 1 |
806,650 | 29,866,639,783 | IssuesEvent | 2023-06-20 04:52:28 | tamalCodes/Milan | https://api.github.com/repos/tamalCodes/Milan | closed | [BUG] Club Page cards redirecting to the bottom of the new page | ๐น aspect: interface ๐ goal: fix ๐จ priority: medium ๐ status: under development gssoc23 level1 | ### Description
The cards on the club page when clicked, redirect to the bottom of their details page rather than showing the page from the top.
### Screenshots
<img width="960" alt="image" src="https://github.com/IAmTamal/Milan/assets/70560812/5f65550b-c94d-4dc5-9f1a-5b77e50c80cd">
<img width="960" alt="image" src="https://github.com/IAmTamal/Milan/assets/70560812/1c6143a4-f0ef-423f-84c6-c92eaf91f1d4">
### Browser ๐ฅฆ
Google Chrome
### Checklist โ
- [X] I checked and didn't find similar issue
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
- [X] I am willing to work on this issue (blank for no) | 1.0 | [BUG] Club Page cards redirecting to the bottom of the new page - ### Description
The cards on the club page when clicked, redirect to the bottom of their details page rather than showing the page from the top.
### Screenshots
<img width="960" alt="image" src="https://github.com/IAmTamal/Milan/assets/70560812/5f65550b-c94d-4dc5-9f1a-5b77e50c80cd">
<img width="960" alt="image" src="https://github.com/IAmTamal/Milan/assets/70560812/1c6143a4-f0ef-423f-84c6-c92eaf91f1d4">
### Browser ๐ฅฆ
Google Chrome
### Checklist โ
- [X] I checked and didn't find similar issue
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
- [X] I am willing to work on this issue (blank for no) | priority | club page cards redirecting to the bottom of the new page description the cards on the club page when clicked redirect to the bottom of their details page rather than showing the page from the top screenshots img width alt image src img width alt image src browser ๐ฅฆ google chrome checklist โ
i checked and didn t find similar issue i have read the i am willing to work on this issue blank for no | 1 |
166,836 | 6,312,765,021 | IssuesEvent | 2017-07-24 05:29:22 | vuejs/rollup-plugin-vue | https://api.github.com/repos/vuejs/rollup-plugin-vue | closed | Fail to import a LESS sheet inside a vue file | Priority: Medium Status: In Progress Type: Bug | ### Expected behavior
These code should work correctly:
```
<style lang="less">
@import "../assets/block";
</style>
```
### Actual behavior
Compiling the vue file failed with these words:
```
๏ฟฝ (vue plugin) '../assets/block.less' wasn't found. Tried - ../assets/block.less,..\assets\block.less
src\components\block.vue
```
The file "../assets/block.less" really exists. I have tried importing the file in the form below, but it still doesn't work:
```
<style lang="less">
@import "..\assets\block";
</style>
```
Just such thing is returned:
```
๏ฟฝ (vue plugin) '..\assets\block.less' wasn't found. Tried - ..\assets\block.less,..\assets\block.less
src\components\block.vue
```
### Steps to reproduce the behavior
| 1.0 | Fail to import a LESS sheet inside a vue file - ### Expected behavior
These code should work correctly:
```
<style lang="less">
@import "../assets/block";
</style>
```
### Actual behavior
Compiling the vue file failed with these words:
```
๏ฟฝ (vue plugin) '../assets/block.less' wasn't found. Tried - ../assets/block.less,..\assets\block.less
src\components\block.vue
```
The file "../assets/block.less" really exists. I have tried importing the file in the form below, but it still doesn't work:
```
<style lang="less">
@import "..\assets\block";
</style>
```
Just such thing is returned:
```
๏ฟฝ (vue plugin) '..\assets\block.less' wasn't found. Tried - ..\assets\block.less,..\assets\block.less
src\components\block.vue
```
### Steps to reproduce the behavior
| priority | fail to import a less sheet inside a vue file expected behavior these code should work correctly import assets block actual behavior compiling the vue file failed with these words ๏ฟฝ vue plugin assets block less wasn t found tried assets block less assets block less src components block vue the file assets block less really exists i have tried importing the file in the form below but it still doesn t work import assets block just such thing is returned ๏ฟฝ vue plugin assets block less wasn t found tried assets block less assets block less src components block vue steps to reproduce the behavior | 1 |
700,366 | 24,058,488,474 | IssuesEvent | 2022-09-16 19:23:21 | rstudio/gt | https://api.github.com/repos/rstudio/gt | closed | suggest: align on a character | Difficulty: [2] Intermediate Effort: [3] High Priority: [2] Medium Type: โ
Enhancement | Nice presentation at rstudio::conf2019.
Is it possible to align by decimal point (period or comma, by locale)? That is, your output rendered something like the left, where it might be nice to see it aligned as on the right:
```
raw dot_align
10.234 x 10^1 10.234 x 10^1
1.234 x 10^2 1.234 x 10^2
2.222 2.222
1.234 x 10^3 1.234 x 10^3
```
It might be nice to generalize a little so that we may align on, say, `=`:
```
right_align char_align
x = 3.14159 x = 3.14159
another_x = 7 another_x = 7
```
This might be something like `cols_align(align=".")` or `cols_align(align="=")`. | 1.0 | suggest: align on a character - Nice presentation at rstudio::conf2019.
Is it possible to align by decimal point (period or comma, by locale)? That is, your output rendered something like the left, where it might be nice to see it aligned as on the right:
```
raw dot_align
10.234 x 10^1 10.234 x 10^1
1.234 x 10^2 1.234 x 10^2
2.222 2.222
1.234 x 10^3 1.234 x 10^3
```
It might be nice to generalize a little so that we may align on, say, `=`:
```
right_align char_align
x = 3.14159 x = 3.14159
another_x = 7 another_x = 7
```
This might be something like `cols_align(align=".")` or `cols_align(align="=")`. | priority | suggest align on a character nice presentation at rstudio is it possible to align by decimal point period or comma by locale that is your output rendered something like the left where it might be nice to see it aligned as on the right raw dot align x x x x x x it might be nice to generalize a little so that we may align on say right align char align x x another x another x this might be something like cols align align or cols align align | 1 |
87,457 | 3,754,892,037 | IssuesEvent | 2016-03-12 08:33:44 | enwikipedia-acc/waca | https://api.github.com/repos/enwikipedia-acc/waca | closed | No confirmation pop-up for 'COI risk' | bug live config Priority: Medium | I didn't get that "confirmation"-pop up when I picked "COI risk", so I never got to see the "Did you do all the steps"-window...which caused me to send an email to #163150 before creating the account (did the wrong order, which I would hav been reminded of, if I got the pop-up. _I think this is what happened_, can't figure out why I would go through with the email, if I got the pop-up stopping me) so..untested... | 1.0 | No confirmation pop-up for 'COI risk' - I didn't get that "confirmation"-pop up when I picked "COI risk", so I never got to see the "Did you do all the steps"-window...which caused me to send an email to #163150 before creating the account (did the wrong order, which I would hav been reminded of, if I got the pop-up. _I think this is what happened_, can't figure out why I would go through with the email, if I got the pop-up stopping me) so..untested... | priority | no confirmation pop up for coi risk i didn t get that confirmation pop up when i picked coi risk so i never got to see the did you do all the steps window which caused me to send an email to before creating the account did the wrong order which i would hav been reminded of if i got the pop up i think this is what happened can t figure out why i would go through with the email if i got the pop up stopping me so untested | 1 |
387,563 | 11,463,357,541 | IssuesEvent | 2020-02-07 15:51:38 | canonical-web-and-design/tutorials.ubuntu.com | https://api.github.com/repos/canonical-web-and-design/tutorials.ubuntu.com | closed | Step 3 (Verify release screenshot) includes non-working instructions for OS X | Bug ๐ Priority: Medium Tutorials Content | https://tutorials.ubuntu.com/tutorial/tutorial-how-to-verify-ubuntu?_ga=2.53919659.265699136.1505613246-1906034191.1505613246#2
It includes spaces in the key spec information in the command which must be removed on OS X before the keys can be imported.
See this askubuntu answer for more details: https://askubuntu.com/a/796638 | 1.0 | Step 3 (Verify release screenshot) includes non-working instructions for OS X - https://tutorials.ubuntu.com/tutorial/tutorial-how-to-verify-ubuntu?_ga=2.53919659.265699136.1505613246-1906034191.1505613246#2
It includes spaces in the key spec information in the command which must be removed on OS X before the keys can be imported.
See this askubuntu answer for more details: https://askubuntu.com/a/796638 | priority | step verify release screenshot includes non working instructions for os x it includes spaces in the key spec information in the command which must be removed on os x before the keys can be imported see this askubuntu answer for more details | 1 |
556,107 | 16,474,607,505 | IssuesEvent | 2021-05-24 02:12:40 | SupremeObsidian/ProjectManager | https://api.github.com/repos/SupremeObsidian/ProjectManager | reopened | Update schematics for Human castle 1-10 to actually be human castle not elf and orc | Medium Priority bug | 1) Figure out what schematics are incorrect
2) Copy and paste correct Schematics to the right building | 1.0 | Update schematics for Human castle 1-10 to actually be human castle not elf and orc - 1) Figure out what schematics are incorrect
2) Copy and paste correct Schematics to the right building | priority | update schematics for human castle to actually be human castle not elf and orc figure out what schematics are incorrect copy and paste correct schematics to the right building | 1 |
474,129 | 13,653,122,342 | IssuesEvent | 2020-09-27 11:07:28 | kiwicom/pytest-recording | https://api.github.com/repos/kiwicom/pytest-recording | opened | [FEATURE] Add release automation | Priority: Medium Type: CI | To automate releases we could use something similar to what I have in [this repo](https://github.com/Stranger6667/pytest-click/blob/master/.github/workflows/release.yml) | 1.0 | [FEATURE] Add release automation - To automate releases we could use something similar to what I have in [this repo](https://github.com/Stranger6667/pytest-click/blob/master/.github/workflows/release.yml) | priority | add release automation to automate releases we could use something similar to what i have in | 1 |
280,027 | 8,677,181,709 | IssuesEvent | 2018-11-30 16:05:32 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | logging: "log_strdup pool empty!" is confusing | area: Logging bug priority: medium | I've enabled CONFIG_NET_DEFAULT_LOG_LEVEL_DBG=y (with echo_server), and here's what I see:
~~~
[00:00:06.554,027] <dbg> net_ctx.net_context_bind: (0x20002510): Context 0x2000072c binding to UDP [::]:4242 iface 0x2000c5c0
[00:00:06.555,215] <dbg> net_conn.net_conn_register: (0x20002510): [3/10/17/0x05] remote 0x00000000/log_strdup pool empty!/0
[00:00:06.556,034] <dbg> net_conn.net_conn_register: (0x20002510): local 0x2000a7b8/log_strdup pool empty!/4242 cb 0x0000894d ud 0x2000005c
~~~
| 1.0 | logging: "log_strdup pool empty!" is confusing - I've enabled CONFIG_NET_DEFAULT_LOG_LEVEL_DBG=y (with echo_server), and here's what I see:
~~~
[00:00:06.554,027] <dbg> net_ctx.net_context_bind: (0x20002510): Context 0x2000072c binding to UDP [::]:4242 iface 0x2000c5c0
[00:00:06.555,215] <dbg> net_conn.net_conn_register: (0x20002510): [3/10/17/0x05] remote 0x00000000/log_strdup pool empty!/0
[00:00:06.556,034] <dbg> net_conn.net_conn_register: (0x20002510): local 0x2000a7b8/log_strdup pool empty!/4242 cb 0x0000894d ud 0x2000005c
~~~
| priority | logging log strdup pool empty is confusing i ve enabled config net default log level dbg y with echo server and here s what i see net ctx net context bind context binding to udp iface net conn net conn register remote log strdup pool empty net conn net conn register local log strdup pool empty cb ud | 1 |
314,230 | 9,594,183,965 | IssuesEvent | 2019-05-09 13:27:05 | linux-audit/audit-kernel | https://api.github.com/repos/linux-audit/audit-kernel | closed | RFE: audit time changes via the adjtimex(2) syscall | difficulty/medium enhancement priority/low | The adjtimex syscall takes a pointer to a structure as its argument. We have to be able to audit when someone or something changes the system clock because that affects correlation of events. Auditing this syscall floods the audit trail with status requests.
| 1.0 | RFE: audit time changes via the adjtimex(2) syscall - The adjtimex syscall takes a pointer to a structure as its argument. We have to be able to audit when someone or something changes the system clock because that affects correlation of events. Auditing this syscall floods the audit trail with status requests.
| priority | rfe audit time changes via the adjtimex syscall the adjtimex syscall takes a pointer to a structure as its argument we have to be able to audit when someone or something changes the system clock because that affects correlation of events auditing this syscall floods the audit trail with status requests | 1 |
332,251 | 10,090,038,706 | IssuesEvent | 2019-07-26 10:27:15 | nuetzliches/nuts | https://api.github.com/repos/nuetzliches/nuts | closed | Add builder for necessary configurations | Priority: Medium Type: Enhancement | Instead of only supporting config files from appsettings.json there should also be (as original intended) support for the builder pattern where it is necessary (e.g. GlobalExceptionHandler). | 1.0 | Add builder for necessary configurations - Instead of only supporting config files from appsettings.json there should also be (as original intended) support for the builder pattern where it is necessary (e.g. GlobalExceptionHandler). | priority | add builder for necessary configurations instead of only supporting config files from appsettings json there should also be as original intended support for the builder pattern where it is necessary e g globalexceptionhandler | 1 |
523,767 | 15,189,525,857 | IssuesEvent | 2021-02-15 16:32:14 | plotly/Plotly.NET | https://api.github.com/repos/plotly/Plotly.NET | closed | Change legend position? | Area: MissingAbstraction Priority: Medium Type: Enhancement | ### Description
How do I change the legend position? For instance, to put the legend below the chart in this sample? I can't see this documented anywhere.
### Repro steps
```fsharp
open Plotly.NET
let x = [1.; 2.; 3.; 4.; 5.; 6.; 7.; 8.; 9.; 10.; ]
let y = [2.; 1.5; 5.; 1.5; 3.; 2.5; 2.5; 1.5; 3.5; 1.]
let combinedChart =
[
Chart.Line(x,y,Name="first")
Chart.Line(y,x,Name="second")
]
|> Chart.Combine
// has legend to the right of the chart. I'd like it below the chart.
combinedChart.Show
```
| 1.0 | Change legend position? - ### Description
How do I change the legend position? For instance, to put the legend below the chart in this sample? I can't see this documented anywhere.
### Repro steps
```fsharp
open Plotly.NET
let x = [1.; 2.; 3.; 4.; 5.; 6.; 7.; 8.; 9.; 10.; ]
let y = [2.; 1.5; 5.; 1.5; 3.; 2.5; 2.5; 1.5; 3.5; 1.]
let combinedChart =
[
Chart.Line(x,y,Name="first")
Chart.Line(y,x,Name="second")
]
|> Chart.Combine
// has legend to the right of the chart. I'd like it below the chart.
combinedChart.Show
```
| priority | change legend position description how do i change the legend position for instance to put the legend below the chart in this sample i can t see this documented anywhere repro steps fsharp open plotly net let x let y let combinedchart chart line x y name first chart line y x name second chart combine has legend to the right of the chart i d like it below the chart combinedchart show | 1 |
1,428 | 2,514,392,009 | IssuesEvent | 2015-01-15 10:58:24 | 29th/personnel | https://api.github.com/repos/29th/personnel | opened | Create activity feed | medium-priority personnel-app | Think about how comments could work...similar to activity on forum?
Also think about whether we should simply pull from the existing service record type queries or if there should be an `activities` table where all activity is aggregated like in the old system. @swomma thoughts? | 1.0 | Create activity feed - Think about how comments could work...similar to activity on forum?
Also think about whether we should simply pull from the existing service record type queries or if there should be an `activities` table where all activity is aggregated like in the old system. @swomma thoughts? | priority | create activity feed think about how comments could work similar to activity on forum also think about whether we should simply pull from the existing service record type queries or if there should be an activities table where all activity is aggregated like in the old system swomma thoughts | 1 |
829,879 | 31,927,497,551 | IssuesEvent | 2023-09-19 03:49:01 | oceanbase/odc | https://api.github.com/repos/oceanbase/odc | closed | [Bug]: select the word, ctrl+c, then ctrl+f, then ctrl+v๏ผthe word that select is cut | type-bug priority-medium | ### ODC version
4.2.0
### OB version
all
### What happened?
as the title says,
select the word, ctrl+c, then ctrl+f, then ctrl+v๏ผthe word that select is cut


### What did you expect to happen?
the word that select is reserved
### How can we reproduce it (as minimally and precisely as possible)?
...
### Anything else we need to know?
_No response_
### Cloud
_No response_ | 1.0 | [Bug]: select the word, ctrl+c, then ctrl+f, then ctrl+v๏ผthe word that select is cut - ### ODC version
4.2.0
### OB version
all
### What happened?
as the title says,
select the word, ctrl+c, then ctrl+f, then ctrl+v๏ผthe word that select is cut


### What did you expect to happen?
the word that select is reserved
### How can we reproduce it (as minimally and precisely as possible)?
...
### Anything else we need to know?
_No response_
### Cloud
_No response_ | priority | select the word ctrl c then ctrl f then ctrl v๏ผthe word that select is cut odc version ob version all what happened as the title says select the word ctrl c then ctrl f then ctrl v๏ผthe word that select is cut what did you expect to happen the word that select is reserved how can we reproduce it as minimally and precisely as possible anything else we need to know no response cloud no response | 1 |
370,384 | 10,931,276,033 | IssuesEvent | 2019-11-23 08:47:20 | bounswe/bounswe2019group10 | https://api.github.com/repos/bounswe/bounswe2019group10 | closed | Editing profile page Android | Priority: Medium Relation: Android Status: Completed | Edit and create missing parts of the profile page.
Deadline: 17.11.19 | 1.0 | Editing profile page Android - Edit and create missing parts of the profile page.
Deadline: 17.11.19 | priority | editing profile page android edit and create missing parts of the profile page deadline | 1 |
207,554 | 7,131,093,999 | IssuesEvent | 2018-01-22 09:45:19 | webpack/webpack-cli | https://api.github.com/repos/webpack/webpack-cli | opened | Remove local fork of Yeoman | Priority: Medium bug | **Do you want to request a *feature* or report a *bug*?**
Bug
**What is the current behavior?**
The `cli` uses a local fork of yeoman generator
**If the current behavior is a bug, please provide the steps to reproduce.**
<!-- A great way to do this is to provide your configuration via a GitHub gist. -->
It works as expected but npm will install the local fork from github, which _**won't work if the user is behind a corporate proxy or they use a package manager like Artifactory**_
**What is the expected behavior?**
Install all the deps from npm
Something important that I want to highlight is that `webpack v4` will have a dependency from the `cli`, which means that if a user will try to update `webpack`, they won't be able to run some commands because yeoman won't be installed. I think this is something that should be fixed before webpack v4 will be officially released.
As I work in a company where they have a corporate proxy, they won't allow an installation from github because it's not a package manager as github, What if we would do a PR to `yeoman` repository with the changes made inside the local fork?
@ev1stensberg what's your opinion about this topic?
| 1.0 | Remove local fork of Yeoman - **Do you want to request a *feature* or report a *bug*?**
Bug
**What is the current behavior?**
The `cli` uses a local fork of yeoman generator
**If the current behavior is a bug, please provide the steps to reproduce.**
<!-- A great way to do this is to provide your configuration via a GitHub gist. -->
It works as expected but npm will install the local fork from github, which _**won't work if the user is behind a corporate proxy or they use a package manager like Artifactory**_
**What is the expected behavior?**
Install all the deps from npm
Something important that I want to highlight is that `webpack v4` will have a dependency from the `cli`, which means that if a user will try to update `webpack`, they won't be able to run some commands because yeoman won't be installed. I think this is something that should be fixed before webpack v4 will be officially released.
As I work in a company where they have a corporate proxy, they won't allow an installation from github because it's not a package manager as github, What if we would do a PR to `yeoman` repository with the changes made inside the local fork?
@ev1stensberg what's your opinion about this topic?
| priority | remove local fork of yeoman do you want to request a feature or report a bug bug what is the current behavior the cli uses a local fork of yeoman generator if the current behavior is a bug please provide the steps to reproduce it works as expected but npm will install the local fork from github which won t work if the user is behind a corporate proxy or they use a package manager like artifactory what is the expected behavior install all the deps from npm something important that i want to highlight is that webpack will have a dependency from the cli which means that if a user will try to update webpack they won t be able to run some commands because yeoman won t be installed i think this is something that should be fixed before webpack will be officially released as i work in a company where they have a corporate proxy they won t allow an installation from github because it s not a package manager as github what if we would do a pr to yeoman repository with the changes made inside the local fork what s your opinion about this topic | 1 |
515,369 | 14,961,036,250 | IssuesEvent | 2021-01-27 07:02:24 | mit-cml/appinventor-sources | https://api.github.com/repos/mit-cml/appinventor-sources | opened | Backport help icon for properties from PUNYA | affects: master enhancement issue: accepted priority: medium | Backport the help icon for designer properties implemented for PUNYA to App Inventor. | 1.0 | Backport help icon for properties from PUNYA - Backport the help icon for designer properties implemented for PUNYA to App Inventor. | priority | backport help icon for properties from punya backport the help icon for designer properties implemented for punya to app inventor | 1 |
57,160 | 3,081,245,018 | IssuesEvent | 2015-08-22 14:35:37 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | ะคะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ | bug imported Priority-Medium | _From [kostya....@gmail.com](https://code.google.com/u/115975207375694284232/) on December 06, 2012 06:08:30_
ะััั ะฝะตะพะฑั
ะพะดะธะผะพััั ะฝะฐัััะพะธัั ัะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ.
ะะฐะบ ััะพ ัะดะตะปะฐัั?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=874_ | 1.0 | ะคะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ - _From [kostya....@gmail.com](https://code.google.com/u/115975207375694284232/) on December 06, 2012 06:08:30_
ะััั ะฝะตะพะฑั
ะพะดะธะผะพััั ะฝะฐัััะพะธัั ัะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ.
ะะฐะบ ััะพ ัะดะตะปะฐัั?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=874_ | priority | ัะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ from on december ะตััั ะฝะตะพะฑั
ะพะดะธะผะพััั ะฝะฐัััะพะธัั ัะธะปัััะฐัะธั ะบะพะผะฐะฝะด ะพั ั
ะฐะฑะพะฒ ะบะฐะบ ััะพ ัะดะตะปะฐัั original issue | 1 |
372,792 | 11,028,296,475 | IssuesEvent | 2019-12-06 11:20:02 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [FEATURE] Add support for JavaDoc {@inheritDoc}-tags | Priority: Medium Status: Accepted Time: ~1 hour Type: Enhancement wontfix | ##### **Overview of the issue**
It would be great to have support for something similiar to JavaDocs {@inheritDoc}. Even better would be to autolink the interfaces or extended classes description if no comment is explicitly given.
| 1.0 | [FEATURE] Add support for JavaDoc {@inheritDoc}-tags - ##### **Overview of the issue**
It would be great to have support for something similiar to JavaDocs {@inheritDoc}. Even better would be to autolink the interfaces or extended classes description if no comment is explicitly given.
| priority | add support for javadoc inheritdoc tags overview of the issue it would be great to have support for something similiar to javadocs inheritdoc even better would be to autolink the interfaces or extended classes description if no comment is explicitly given | 1 |
430,348 | 12,451,813,297 | IssuesEvent | 2020-05-27 11:09:29 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | closed | Select multiple tasks for validation | Component: Frontend Difficulty: Medium Priority: High Status: Review needed Type: Enhancement | Feedback from the testing with validators about the need to validate with more than one task (this was possible in TM3):
* Option to select multiple adjoining tasks to download (using a polygon) in JOSM/iD.
* Eventually as clicks or with the option to draw and select a square/rectangular over tasks to be validated.
| 1.0 | Select multiple tasks for validation - Feedback from the testing with validators about the need to validate with more than one task (this was possible in TM3):
* Option to select multiple adjoining tasks to download (using a polygon) in JOSM/iD.
* Eventually as clicks or with the option to draw and select a square/rectangular over tasks to be validated.
| priority | select multiple tasks for validation feedback from the testing with validators about the need to validate with more than one task this was possible in option to select multiple adjoining tasks to download using a polygon in josm id eventually as clicks or with the option to draw and select a square rectangular over tasks to be validated | 1 |
506,463 | 14,665,727,162 | IssuesEvent | 2020-12-29 14:51:05 | dita-ot/dita-ot | https://api.github.com/repos/dita-ot/dita-ot | closed | Abbreviated-form in topic title not handled consistently [DOT 1.8 and 2.0] | bug plugin/xhtml priority/medium stale | Let's say I use an abbreviated-form in a DITA topic title:
```
<title>This is my product:<abbreviated-form keyref="ABBS"/></title>
```
When generating XHTML output the table of contents will not contain the proper expanded title. Also links to the topic (both in XHTML-based and PDF outputs) does not contain the proper expanded title.
I tested with both DITA OT 1.8 and 2.0
| 1.0 | Abbreviated-form in topic title not handled consistently [DOT 1.8 and 2.0] - Let's say I use an abbreviated-form in a DITA topic title:
```
<title>This is my product:<abbreviated-form keyref="ABBS"/></title>
```
When generating XHTML output the table of contents will not contain the proper expanded title. Also links to the topic (both in XHTML-based and PDF outputs) does not contain the proper expanded title.
I tested with both DITA OT 1.8 and 2.0
| priority | abbreviated form in topic title not handled consistently let s say i use an abbreviated form in a dita topic title this is my product when generating xhtml output the table of contents will not contain the proper expanded title also links to the topic both in xhtml based and pdf outputs does not contain the proper expanded title i tested with both dita ot and | 1 |
78,589 | 3,511,365,453 | IssuesEvent | 2016-01-10 06:42:28 | ESAPI/esapi-java-legacy | https://api.github.com/repos/ESAPI/esapi-java-legacy | closed | HTTPParameterValue | bug imported Priority-Medium | _From [parashar...@gmail.com](https://code.google.com/u/108634640519800304581/) on August 27, 2014 00:30:51_
I need to allow alphanumeric, new line, \t\n\r, single space and all special character except (< > double space)
What would be the excepted expression for configure Validator.HTTPParameterValue.
Ex. "02-Aug-2014 at 21:05#%#%";
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=335_ | 1.0 | HTTPParameterValue - _From [parashar...@gmail.com](https://code.google.com/u/108634640519800304581/) on August 27, 2014 00:30:51_
I need to allow alphanumeric, new line, \t\n\r, single space and all special character except (< > double space)
What would be the excepted expression for configure Validator.HTTPParameterValue.
Ex. "02-Aug-2014 at 21:05#%#%";
_Original issue: http://code.google.com/p/owasp-esapi-java/issues/detail?id=335_ | priority | httpparametervalue from on august i need to allow alphanumeric new line t n r single space and all special character except double space what would be the excepted expression for configure validator httpparametervalue ex aug at original issue | 1 |
822,595 | 30,879,159,971 | IssuesEvent | 2023-08-03 16:13:36 | opendcs/opendcs | https://api.github.com/repos/opendcs/opendcs | opened | Create tests for SQL over XML. | priority-medium feature request | The current integration tests will be modified to be a junit5 "test-engine" for looking up and actually engaging the tests. The dynamic test mechanism works okay, but junit5 was designed to be extended in this way and I think it will easier to maintain long term.
The SQL over XML implementation for now will remain an independent "Database" implementation. Hence the initial ticket in this change of doing the Database SPI implementation. | 1.0 | Create tests for SQL over XML. - The current integration tests will be modified to be a junit5 "test-engine" for looking up and actually engaging the tests. The dynamic test mechanism works okay, but junit5 was designed to be extended in this way and I think it will easier to maintain long term.
The SQL over XML implementation for now will remain an independent "Database" implementation. Hence the initial ticket in this change of doing the Database SPI implementation. | priority | create tests for sql over xml the current integration tests will be modified to be a test engine for looking up and actually engaging the tests the dynamic test mechanism works okay but was designed to be extended in this way and i think it will easier to maintain long term the sql over xml implementation for now will remain an independent database implementation hence the initial ticket in this change of doing the database spi implementation | 1 |
409,467 | 11,963,006,521 | IssuesEvent | 2020-04-05 14:28:52 | OrangeJuice7/SDL-OpenGL-Game-Framework | https://api.github.com/repos/OrangeJuice7/SDL-OpenGL-Game-Framework | opened | Add Player system | area.Model priority.medium work.medium | - [ ] Track player data (e.g. score, money) somewhere
- [ ] Entities may possibly belong to a player / have a faction system
| 1.0 | Add Player system - - [ ] Track player data (e.g. score, money) somewhere
- [ ] Entities may possibly belong to a player / have a faction system
| priority | add player system track player data e g score money somewhere entities may possibly belong to a player have a faction system | 1 |
126,956 | 5,008,266,406 | IssuesEvent | 2016-12-12 19:01:49 | RevivalPMMP/PureEntitiesX | https://api.github.com/repos/RevivalPMMP/PureEntitiesX | closed | AutoSpawnTask | Category: Enhancement Category: TODO Priority: Medium Status: Fixed | This task, which auto spawns mobs (Like Vanilla Minecraft), has another `//ToDo` under it. It is a huge part of the plugin, so we need to implement it | 1.0 | AutoSpawnTask - This task, which auto spawns mobs (Like Vanilla Minecraft), has another `//ToDo` under it. It is a huge part of the plugin, so we need to implement it | priority | autospawntask this task which auto spawns mobs like vanilla minecraft has another todo under it it is a huge part of the plugin so we need to implement it | 1 |
315,537 | 9,621,958,265 | IssuesEvent | 2019-05-14 11:58:13 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | opened | v9-gui: display of access level for users | Priority: Medium Type: Bug | When access level is set to NONE for a user, **new** GUI doesn't display this information in `Actions` tab of the user. Not the case with `ALL`.
As I mentioned in #4395, after a new creation of a user, access level is displayed. If I clear my browser cache, I lost this information. | 1.0 | v9-gui: display of access level for users - When access level is set to NONE for a user, **new** GUI doesn't display this information in `Actions` tab of the user. Not the case with `ALL`.
As I mentioned in #4395, after a new creation of a user, access level is displayed. If I clear my browser cache, I lost this information. | priority | gui display of access level for users when access level is set to none for a user new gui doesn t display this information in actions tab of the user not the case with all as i mentioned in after a new creation of a user access level is displayed if i clear my browser cache i lost this information | 1 |
61,154 | 3,141,342,860 | IssuesEvent | 2015-09-12 14:09:28 | ruleant/getback_gps | https://api.github.com/repos/ruleant/getback_gps | closed | save altitude when storing a location | enhancement priority medium | - [x] save altitude in StoredDestination
- [x] calculate height difference between destination and current location
- [ ] display height difference on Main and Details activity | 1.0 | save altitude when storing a location - - [x] save altitude in StoredDestination
- [x] calculate height difference between destination and current location
- [ ] display height difference on Main and Details activity | priority | save altitude when storing a location save altitude in storeddestination calculate height difference between destination and current location display height difference on main and details activity | 1 |
17,155 | 2,615,132,290 | IssuesEvent | 2015-03-01 06:02:31 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Drive API and Spreadsheet authenticated with OAuth2 in one android example. | auto-migrated Priority-Medium Type-Sample | ```
I searched searched and searched but there is nothing that auth Drive API and
Spreadsheet scopes together in one example using OAuth 2.0. This is very common
use case for people to fully leverage the power of Drive.
Which Google API and version (e.g. Google Calendar API version 3)?
Latest Spreadsheet API for android
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Android 4.2.2
External references, such as API reference guide?
Please provide any additional information below.
```
Original issue reported on code.google.com by `paracha3` on 30 Apr 2013 at 3:35 | 1.0 | Drive API and Spreadsheet authenticated with OAuth2 in one android example. - ```
I searched searched and searched but there is nothing that auth Drive API and
Spreadsheet scopes together in one example using OAuth 2.0. This is very common
use case for people to fully leverage the power of Drive.
Which Google API and version (e.g. Google Calendar API version 3)?
Latest Spreadsheet API for android
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Android 4.2.2
External references, such as API reference guide?
Please provide any additional information below.
```
Original issue reported on code.google.com by `paracha3` on 30 Apr 2013 at 3:35 | priority | drive api and spreadsheet authenticated with in one android example i searched searched and searched but there is nothing that auth drive api and spreadsheet scopes together in one example using oauth this is very common use case for people to fully leverage the power of drive which google api and version e g google calendar api version latest spreadsheet api for android java environment e g java android app engine android external references such as api reference guide please provide any additional information below original issue reported on code google com by on apr at | 1 |
186,795 | 6,742,713,472 | IssuesEvent | 2017-10-20 08:57:38 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Migrate existing Grunt tasks to Gulp | help wanted priority: medium | Gulp is better and also we don't like having 2 build tools
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/26131762-migrate-existing-grunt-tasks-to-gulp?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| 1.0 | Migrate existing Grunt tasks to Gulp - Gulp is better and also we don't like having 2 build tools
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/26131762-migrate-existing-grunt-tasks-to-gulp?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| priority | migrate existing grunt tasks to gulp gulp is better and also we don t like having build tools want to back this issue we accept bounties via | 1 |
672,242 | 22,796,958,606 | IssuesEvent | 2022-07-10 21:27:15 | projectdiscovery/mapcidr | https://api.github.com/repos/projectdiscovery/mapcidr | closed | Match and Filter IP/CIDR support | Priority: Medium Status: Completed Type: Enhancement | **Is your feature request related to a problem? Please describe.**
For the cases, where we have known list of **IPs/CIDRs** that I wanted to match or filter to have desired list of **IPs/CIDRs** to process/scan as output and would be good addition for the project.
**Describe the solution you'd like**
This can be done by introducing 2 additional CLI flag to accept IP/CIDR to perform said action, for example:
```yaml
-mi, -match-ip string[] IP/CIDR/FILE containing list of IP/CIDR to match (comma-separated, file input)
-fi, -filter-ip string[] IP/CIDR/FILE containing list of IP/CIDR to filter (comma-separated, file input)
```
**Match IP/CIDR**
> Display IPs from input that belongs to given IP/CIDR ranges.
```bash
$ mapcidr -cidr 192.168.1.0/24 -mi 192.168.1.253,192.168.1.252
192.168.1.252
192.168.1.253
```
More input examples:
```bash
$ mapcidr -cidr 192.168.1.0/24 -mi 192.168.1.0/25
$ mapcidr -cidr 192.168.1.0/24 -mi cidrs.txt
```
**Filter IP/CIDR**
> Filter IPs from input that belongs to given IP/CIDR ranges.
```bash
$ mapcidr -cidr 192.168.1.224/28 -fi 192.168.1.233,192.168.1.234
192.168.1.224
192.168.1.225
192.168.1.226
192.168.1.227
192.168.1.228
192.168.1.229
192.168.1.230
192.168.1.231
192.168.1.232
192.168.1.235
192.168.1.236
192.168.1.237
192.168.1.238
192.168.1.239
```
More input examples:
```bash
$ mapcidr -cidr 192.168.1.0/24 -fi 192.168.1.0/25
$ mapcidr -cidr 192.168.1.0/24 -fi cidrs.txt
```
Note:
Currently available `ip/il` cli flag was introduced as part of the feature request in https://github.com/projectdiscovery/mapcidr/issues/4 to match the IPs, which is same as `m/match` feature, as such as per this change `ip/il` can be deprecated to avoid any possible confusion. | 1.0 | Match and Filter IP/CIDR support - **Is your feature request related to a problem? Please describe.**
For the cases, where we have known list of **IPs/CIDRs** that I wanted to match or filter to have desired list of **IPs/CIDRs** to process/scan as output and would be good addition for the project.
**Describe the solution you'd like**
This can be done by introducing 2 additional CLI flag to accept IP/CIDR to perform said action, for example:
```yaml
-mi, -match-ip string[] IP/CIDR/FILE containing list of IP/CIDR to match (comma-separated, file input)
-fi, -filter-ip string[] IP/CIDR/FILE containing list of IP/CIDR to filter (comma-separated, file input)
```
**Match IP/CIDR**
> Display IPs from input that belongs to given IP/CIDR ranges.
```bash
$ mapcidr -cidr 192.168.1.0/24 -mi 192.168.1.253,192.168.1.252
192.168.1.252
192.168.1.253
```
More input examples:
```bash
$ mapcidr -cidr 192.168.1.0/24 -mi 192.168.1.0/25
$ mapcidr -cidr 192.168.1.0/24 -mi cidrs.txt
```
**Filter IP/CIDR**
> Filter IPs from input that belongs to given IP/CIDR ranges.
```bash
$ mapcidr -cidr 192.168.1.224/28 -fi 192.168.1.233,192.168.1.234
192.168.1.224
192.168.1.225
192.168.1.226
192.168.1.227
192.168.1.228
192.168.1.229
192.168.1.230
192.168.1.231
192.168.1.232
192.168.1.235
192.168.1.236
192.168.1.237
192.168.1.238
192.168.1.239
```
More input examples:
```bash
$ mapcidr -cidr 192.168.1.0/24 -fi 192.168.1.0/25
$ mapcidr -cidr 192.168.1.0/24 -fi cidrs.txt
```
Note:
Currently available `ip/il` cli flag was introduced as part of the feature request in https://github.com/projectdiscovery/mapcidr/issues/4 to match the IPs, which is same as `m/match` feature, as such as per this change `ip/il` can be deprecated to avoid any possible confusion. | priority | match and filter ip cidr support is your feature request related to a problem please describe for the cases where we have known list of ips cidrs that i wanted to match or filter to have desired list of ips cidrs to process scan as output and would be good addition for the project describe the solution you d like this can be done by introducing additional cli flag to accept ip cidr to perform said action for example yaml mi match ip string ip cidr file containing list of ip cidr to match comma separated file input fi filter ip string ip cidr file containing list of ip cidr to filter comma separated file input match ip cidr display ips from input that belongs to given ip cidr ranges bash mapcidr cidr mi more input examples bash mapcidr cidr mi mapcidr cidr mi cidrs txt filter ip cidr filter ips from input that belongs to given ip cidr ranges bash mapcidr cidr fi more input examples bash mapcidr cidr fi mapcidr cidr fi cidrs txt note currently available ip il cli flag was introduced as part of the feature request in to match the ips which is same as m match feature as such as per this change ip il can be deprecated to avoid any possible confusion | 1 |
784,555 | 27,575,824,073 | IssuesEvent | 2023-03-08 12:56:07 | tahowallet/extension | https://api.github.com/repos/tahowallet/extension | opened | Abilities page contains abilities from unimported accounts | Type: Bug Status: Pending Priority: Medium | Current: If you have an ability on one of your ledger accounts but you decide to not import that account then ability will be imported anyway for this address.
Expected: abilities are fetched only for Ledger accounts that were selected by user during onboarding
Should be fixed together with: https://github.com/tahowallet/extension/issues/3100
| 1.0 | Abilities page contains abilities from unimported accounts - Current: If you have an ability on one of your ledger accounts but you decide to not import that account then ability will be imported anyway for this address.
Expected: abilities are fetched only for Ledger accounts that were selected by user during onboarding
Should be fixed together with: https://github.com/tahowallet/extension/issues/3100
| priority | abilities page contains abilities from unimported accounts current if you have an ability on one of your ledger accounts but you decide to not import that account then ability will be imported anyway for this address expected abilities are fetched only for ledger accounts that were selected by user during onboarding should be fixed together with | 1 |
204,698 | 7,089,601,298 | IssuesEvent | 2018-01-12 03:49:44 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Prevent overshoot of job estimates when acquiring WQE | Medium Priority | Work acquisition in the agent is basically a nested for loop:
```
for wqe in work_list:
for site in site_list:
... see if site is under threshold and can acquire wqe ...
```
If the site is at least 1 job under threshold, it will acquire the WQE, regardless of how many jobs there are in the WQE. If there is a big request with a large site whitelist, then each site can potentially pull in a WQE in a given iteration.
This is particularly problematic if the WQE element has a large blow-up factor (i.e., there's hundreds of jobs per WQE): the agent can overshoot its targets by thousands of jobs per iteration.
I propose a twist in the algorithm to help the agent acquire work less aggressively when close to the limits:
- If the site is under threshold by at least one job, then we still accept the WQE.
- Instead of moving on to the next WQE immediately, try to assign the other jobs in the WQE to other sites in the whitelist which are below threshold.
This will mean we no longer overshoot once per-site but by one WQE (in the case where the site whitelist is all sites).
@amaltaro
| 1.0 | Prevent overshoot of job estimates when acquiring WQE - Work acquisition in the agent is basically a nested for loop:
```
for wqe in work_list:
for site in site_list:
... see if site is under threshold and can acquire wqe ...
```
If the site is at least 1 job under threshold, it will acquire the WQE, regardless of how many jobs there are in the WQE. If there is a big request with a large site whitelist, then each site can potentially pull in a WQE in a given iteration.
This is particularly problematic if the WQE element has a large blow-up factor (i.e., there's hundreds of jobs per WQE): the agent can overshoot its targets by thousands of jobs per iteration.
I propose a twist in the algorithm to help the agent acquire work less aggressively when close to the limits:
- If the site is under threshold by at least one job, then we still accept the WQE.
- Instead of moving on to the next WQE immediately, try to assign the other jobs in the WQE to other sites in the whitelist which are below threshold.
This will mean we no longer overshoot once per-site but by one WQE (in the case where the site whitelist is all sites).
@amaltaro
| priority | prevent overshoot of job estimates when acquiring wqe work acquisition in the agent is basically a nested for loop for wqe in work list for site in site list see if site is under threshold and can acquire wqe if the site is at least job under threshold it will acquire the wqe regardless of how many jobs there are in the wqe if there is a big request with a large site whitelist then each site can potentially pull in a wqe in a given iteration this is particularly problematic if the wqe element has a large blow up factor i e there s hundreds of jobs per wqe the agent can overshoot its targets by thousands of jobs per iteration i propose a twist in the algorithm to help the agent acquire work less aggressively when close to the limits if the site is under threshold by at least one job then we still accept the wqe instead of moving on to the next wqe immediately try to assign the other jobs in the wqe to other sites in the whitelist which are below threshold this will mean we no longer overshoot once per site but by one wqe in the case where the site whitelist is all sites amaltaro | 1 |
482,229 | 13,903,145,544 | IssuesEvent | 2020-10-20 06:45:56 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.1.0 beta staging-1813]Civics: UX UI tweaks | Category: UI Priority: Medium | - [ ] Chat visible inside while editing a civics object - allow the UI to be resized / moved?
Currently there is no way to talk to people while making civic objects/edits because the UI does not allow resizing and is primary
and overlaps everything. | 1.0 | [0.9.1.0 beta staging-1813]Civics: UX UI tweaks - - [ ] Chat visible inside while editing a civics object - allow the UI to be resized / moved?
Currently there is no way to talk to people while making civic objects/edits because the UI does not allow resizing and is primary
and overlaps everything. | priority | civics ux ui tweaks chat visible inside while editing a civics object allow the ui to be resized moved currently there is no way to talk to people while making civic objects edits because the ui does not allow resizing and is primary and overlaps everything | 1 |
725,598 | 24,967,681,082 | IssuesEvent | 2022-11-01 20:56:45 | FlutterFlow/flutterflow-issues | https://api.github.com/repos/FlutterFlow/flutterflow-issues | opened | AudioPlayer not working on web with deep linking enabled for asset audios | status: confirmed priority: medium | Issue tracker is **ONLY** used for reporting bugs. New feature suggestions and questions should be discussed on Community or submitted through our user feedback form.
Your issue may already be reported! Please search in the [issue tracker](../) before creating one.
Please **thumbs up** this issue if you have also experienced it. You may also add more information if there is something relevant that was not mentioned. However, please refrain from comments that are not constructive, like "I have this problem too", etc.
## Expected behavior (required)
<!-- A clear and concise description of what you expected to happen. -->
Audio player loads "asset" audios.
## Current behavior (required)
<!-- What happens instead of the expected behavior. -->
Audio player can't play asset audios on web in the scenario below.
## To Reproduce (required)
<!-- Please be detailed as possible here so we can help diagnose the issue. Issues cannot be accepted if they are too vague. For example, "project fails to build" could be better reported as:
1. Create new page
2. Add container widget
3. Set width = 123
4. Click Run
5. Observe that project doesnโt build
Code can be included in this section if it is relevant to reproducing the bug.
-->
Steps to reproduce the behavior:
1. Create new project
2. Add new page, e.g. Page2
3. Add link from HomePage to Page2.
4. Add an audio to project assets.
5. On Page2, create an audio player.
6. Select audio type = Asset and the uploaded audio as Asset Audio.
7. Publish to web.
8. Open the published link.
9. Navigate from HomePage to Page2.
10. Observe that the audio widget does not show.
11. Open browser dev tools/network.
12. Observe that it tried to load the audio from `https://<project>/page2/assets/assets/audios/<filename>.mp3`. Expected: `https://<project>.flutterflow.app/assets/assets/audios/<filename>.mp3`
Project: https://app.flutterflow.io/project/audio-asset-63344x
## Context (required)
<!-- How has this issue affected you? What are you trying to accomplish? -->
Can't use audio player on web.
## Screenshots / recordings
<!-- If applicable, add screenshots to help explain your problem. -->
N/A
## Your environment
<!--- Include relevant details about the environment you experienced the bug in -->
* **Bug Report Code:**
* Version of FlutterFlow used: FlutterFlow v3.0, released October 28, 2022
* Platform (e.g. Web, MacOS Desktop): Web
* Browser name and version: Chrome, Version 107.0.5304.87 (Official Build) (x86_64)
* Operating system and version (desktop or mobile): MacOS 12.5.1 Monterey.
| 1.0 | AudioPlayer not working on web with deep linking enabled for asset audios - Issue tracker is **ONLY** used for reporting bugs. New feature suggestions and questions should be discussed on Community or submitted through our user feedback form.
Your issue may already be reported! Please search in the [issue tracker](../) before creating one.
Please **thumbs up** this issue if you have also experienced it. You may also add more information if there is something relevant that was not mentioned. However, please refrain from comments that are not constructive, like "I have this problem too", etc.
## Expected behavior (required)
<!-- A clear and concise description of what you expected to happen. -->
Audio player loads "asset" audios.
## Current behavior (required)
<!-- What happens instead of the expected behavior. -->
Audio player can't play asset audios on web in the scenario below.
## To Reproduce (required)
<!-- Please be detailed as possible here so we can help diagnose the issue. Issues cannot be accepted if they are too vague. For example, "project fails to build" could be better reported as:
1. Create new page
2. Add container widget
3. Set width = 123
4. Click Run
5. Observe that project doesnโt build
Code can be included in this section if it is relevant to reproducing the bug.
-->
Steps to reproduce the behavior:
1. Create new project
2. Add new page, e.g. Page2
3. Add link from HomePage to Page2.
4. Add an audio to project assets.
5. On Page2, create an audio player.
6. Select audio type = Asset and the uploaded audio as Asset Audio.
7. Publish to web.
8. Open the published link.
9. Navigate from HomePage to Page2.
10. Observe that the audio widget does not show.
11. Open browser dev tools/network.
12. Observe that it tried to load the audio from `https://<project>/page2/assets/assets/audios/<filename>.mp3`. Expected: `https://<project>.flutterflow.app/assets/assets/audios/<filename>.mp3`
Project: https://app.flutterflow.io/project/audio-asset-63344x
## Context (required)
<!-- How has this issue affected you? What are you trying to accomplish? -->
Can't use audio player on web.
## Screenshots / recordings
<!-- If applicable, add screenshots to help explain your problem. -->
N/A
## Your environment
<!--- Include relevant details about the environment you experienced the bug in -->
* **Bug Report Code:**
* Version of FlutterFlow used: FlutterFlow v3.0, released October 28, 2022
* Platform (e.g. Web, MacOS Desktop): Web
* Browser name and version: Chrome, Version 107.0.5304.87 (Official Build) (x86_64)
* Operating system and version (desktop or mobile): MacOS 12.5.1 Monterey.
| priority | audioplayer not working on web with deep linking enabled for asset audios issue tracker is only used for reporting bugs new feature suggestions and questions should be discussed on community or submitted through our user feedback form your issue may already be reported please search in the before creating one please thumbs up this issue if you have also experienced it you may also add more information if there is something relevant that was not mentioned however please refrain from comments that are not constructive like i have this problem too etc expected behavior required audio player loads asset audios current behavior required audio player can t play asset audios on web in the scenario below to reproduce required please be detailed as possible here so we can help diagnose the issue issues cannot be accepted if they are too vague for example project fails to build could be better reported as create new page add container widget set width click run observe that project doesnโt build code can be included in this section if it is relevant to reproducing the bug steps to reproduce the behavior create new project add new page e g add link from homepage to add an audio to project assets on create an audio player select audio type asset and the uploaded audio as asset audio publish to web open the published link navigate from homepage to observe that the audio widget does not show open browser dev tools network observe that it tried to load the audio from expected project context required can t use audio player on web screenshots recordings n a your environment bug report code version of flutterflow used flutterflow released october platform e g web macos desktop web browser name and version chrome version official build operating system and version desktop or mobile macos monterey | 1 |
109,321 | 4,385,562,177 | IssuesEvent | 2016-08-08 09:24:58 | bazingatechnologies/FSharp.Data.GraphQL | https://api.github.com/repos/bazingatechnologies/FSharp.Data.GraphQL | opened | Introduce query complexity analysis | enhancement medium-priority | Right now we don't have a way to protect the server from executing too complex queries, that could possibly be dangerous (i.e. could consume all memory or CPU power). For this we should introduce complexity analysis: this have already been done in other GraphQL implementations such as [Sangria](http://sangria-graphql.org/learn/#protection-against-malicious-queries).
The core idea is an ability to add weight (complexity) value to each field, then compute the overall complexity of the query to decide what to do next before executing it. Complexity could be a property/method added to `ExecutionPlan`/`ExecutionPlanInfo` records. | 1.0 | Introduce query complexity analysis - Right now we don't have a way to protect the server from executing too complex queries, that could possibly be dangerous (i.e. could consume all memory or CPU power). For this we should introduce complexity analysis: this have already been done in other GraphQL implementations such as [Sangria](http://sangria-graphql.org/learn/#protection-against-malicious-queries).
The core idea is an ability to add weight (complexity) value to each field, then compute the overall complexity of the query to decide what to do next before executing it. Complexity could be a property/method added to `ExecutionPlan`/`ExecutionPlanInfo` records. | priority | introduce query complexity analysis right now we don t have a way to protect the server from executing too complex queries that could possibly be dangerous i e could consume all memory or cpu power for this we should introduce complexity analysis this have already been done in other graphql implementations such as the core idea is an ability to add weight complexity value to each field then compute the overall complexity of the query to decide what to do next before executing it complexity could be a property method added to executionplan executionplaninfo records | 1 |
467,598 | 13,451,207,764 | IssuesEvent | 2020-09-08 19:51:06 | silentium-labs/merlin-gql | https://api.github.com/repos/silentium-labs/merlin-gql | opened | Clean playground or create a new repository to serve as boilerplate for the new command | Priority: Medium Status: Pending Type: Enhancement | Looking forward for the upcoming release candidate, we should create either a new repository or clean the current playground and make sure that we include the bare minimum code that is necessary and that we want whenever a user of merlin-gql starts a new project.
My proposed pieces would be:
- Fully working graphql api
- GraphQL playground enabled if environment is not production
- Maybe 2 or 3 typeorm models with fields exposed to the GraphQL Schema layer, we could use in-memory database as default and let the user change it within the code or we could ask in the new command which type of database the user wants us to generate
- Example Data Generator
- Basic security (i guess login only is good enough, we could discuss if registering would be required although i don't feel it should)
- A field resolver that gets some data from an external source, could be a public API or a json file, the intention here is to provide a working example of getting data from another source than typeorm
- A custom resolver with some type-graphql decorators to provide a working example of a custom resolver doing some non standard database related operations | 1.0 | Clean playground or create a new repository to serve as boilerplate for the new command - Looking forward for the upcoming release candidate, we should create either a new repository or clean the current playground and make sure that we include the bare minimum code that is necessary and that we want whenever a user of merlin-gql starts a new project.
My proposed pieces would be:
- Fully working graphql api
- GraphQL playground enabled if environment is not production
- Maybe 2 or 3 typeorm models with fields exposed to the GraphQL Schema layer, we could use in-memory database as default and let the user change it within the code or we could ask in the new command which type of database the user wants us to generate
- Example Data Generator
- Basic security (i guess login only is good enough, we could discuss if registering would be required although i don't feel it should)
- A field resolver that gets some data from an external source, could be a public API or a json file, the intention here is to provide a working example of getting data from another source than typeorm
- A custom resolver with some type-graphql decorators to provide a working example of a custom resolver doing some non standard database related operations | priority | clean playground or create a new repository to serve as boilerplate for the new command looking forward for the upcoming release candidate we should create either a new repository or clean the current playground and make sure that we include the bare minimum code that is necessary and that we want whenever a user of merlin gql starts a new project my proposed pieces would be fully working graphql api graphql playground enabled if environment is not production maybe or typeorm models with fields exposed to the graphql schema layer we could use in memory database as default and let the user change it within the code or we could ask in the new command which type of database the user wants us to generate example data generator basic security i guess login only is good enough we could discuss if registering would be required although i don t feel it should a field resolver that gets some data from an external source could be a public api or a json file the intention here is to provide a working example of getting data from another source than typeorm a custom resolver with some type graphql decorators to provide a working example of a custom resolver doing some non standard database related operations | 1 |
30,308 | 2,723,444,346 | IssuesEvent | 2015-04-14 12:40:14 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | ProjectGenerator must generate a Crux.properties telling Crux to use the CruxHtmlTagsScreenResolver | bug CruxCore imported Milestone-2.2.1 Priority-Medium | _From [tr_busta...@yahoo.com.br](https://code.google.com/u/115454294030253308352/) on April 07, 2010 10:49:07_
ProjectGenerator must generate a Crux.properties telling Crux to use the
CruxHtmlTagsScreenResolver
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=58_ | 1.0 | ProjectGenerator must generate a Crux.properties telling Crux to use the CruxHtmlTagsScreenResolver - _From [tr_busta...@yahoo.com.br](https://code.google.com/u/115454294030253308352/) on April 07, 2010 10:49:07_
ProjectGenerator must generate a Crux.properties telling Crux to use the
CruxHtmlTagsScreenResolver
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=58_ | priority | projectgenerator must generate a crux properties telling crux to use the cruxhtmltagsscreenresolver from on april projectgenerator must generate a crux properties telling crux to use the cruxhtmltagsscreenresolver original issue | 1 |
442,280 | 12,743,030,135 | IssuesEvent | 2020-06-26 09:35:58 | DIAGNijmegen/website-content | https://api.github.com/repos/DIAGNijmegen/website-content | closed | Change formulation and capitalisation positions | Priority: Medium content | All positions should be capitalized eg. Assistant Professor and PhD Student should be changed to PhD Candidate. | 1.0 | Change formulation and capitalisation positions - All positions should be capitalized eg. Assistant Professor and PhD Student should be changed to PhD Candidate. | priority | change formulation and capitalisation positions all positions should be capitalized eg assistant professor and phd student should be changed to phd candidate | 1 |
398,572 | 11,741,882,833 | IssuesEvent | 2020-03-11 22:56:01 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Generen van thumbnail van niet-plaatjes levert OSError op | bug photos priority: medium | In GitLab by @joostrijneveld on Jan 10, 2017, 14:58
Dit vindt bijvoorbeeld plaats wanneer je https://thalia.nu/photos/2015-09-17-bowlen/ bezoekt, aangezien IMG_3299.JPG geen goed-geformatte JPG lijkt te zijn (ik kan 'm lokaal ook niet openen in een imageviewer). Iets vergelijkbaars gebeurt met videobestanden.
Gerelateerd aan #205 | 1.0 | Generen van thumbnail van niet-plaatjes levert OSError op - In GitLab by @joostrijneveld on Jan 10, 2017, 14:58
Dit vindt bijvoorbeeld plaats wanneer je https://thalia.nu/photos/2015-09-17-bowlen/ bezoekt, aangezien IMG_3299.JPG geen goed-geformatte JPG lijkt te zijn (ik kan 'm lokaal ook niet openen in een imageviewer). Iets vergelijkbaars gebeurt met videobestanden.
Gerelateerd aan #205 | priority | generen van thumbnail van niet plaatjes levert oserror op in gitlab by joostrijneveld on jan dit vindt bijvoorbeeld plaats wanneer je bezoekt aangezien img jpg geen goed geformatte jpg lijkt te zijn ik kan m lokaal ook niet openen in een imageviewer iets vergelijkbaars gebeurt met videobestanden gerelateerd aan | 1 |
275,014 | 8,570,550,624 | IssuesEvent | 2018-11-11 21:10:23 | CS2103-AY1819S1-W13-3/main | https://api.github.com/repos/CS2103-AY1819S1-W13-3/main | closed | Add "uses" for Marks | feature.Mark priority:Medium status.Ongoing type.Enhancement | For other features such as attendance, edit, delete, email etc, implement the IMarkExecutable interface if the feature is supports execution with marks (e.g. batch attendance updating, batch emailing)
@preethi-d suggested format for `email`:
```
email m/MARK b/BODY
```
@yogtew suggested format for `attendance`:
```
attendance m/MARK at/ATTENDANCE
```
Suggested implementation details:
1. Update `YourCommandParser` to accept a `PREFIX_MARK` in the tokenizer
2. Check if `argumentMultiMap` contains a value for the `PREFIX_MARK` attribute
3. If present, call `executeMark` method from `execute` and return the `CommandResult` from that
4. Otherwise, run the rest of `execute` as normal
Can refer to the `SetTagCommand` for inspiration | 1.0 | Add "uses" for Marks - For other features such as attendance, edit, delete, email etc, implement the IMarkExecutable interface if the feature is supports execution with marks (e.g. batch attendance updating, batch emailing)
@preethi-d suggested format for `email`:
```
email m/MARK b/BODY
```
@yogtew suggested format for `attendance`:
```
attendance m/MARK at/ATTENDANCE
```
Suggested implementation details:
1. Update `YourCommandParser` to accept a `PREFIX_MARK` in the tokenizer
2. Check if `argumentMultiMap` contains a value for the `PREFIX_MARK` attribute
3. If present, call `executeMark` method from `execute` and return the `CommandResult` from that
4. Otherwise, run the rest of `execute` as normal
Can refer to the `SetTagCommand` for inspiration | priority | add uses for marks for other features such as attendance edit delete email etc implement the imarkexecutable interface if the feature is supports execution with marks e g batch attendance updating batch emailing preethi d suggested format for email email m mark b body yogtew suggested format for attendance attendance m mark at attendance suggested implementation details update yourcommandparser to accept a prefix mark in the tokenizer check if argumentmultimap contains a value for the prefix mark attribute if present call executemark method from execute and return the commandresult from that otherwise run the rest of execute as normal can refer to the settagcommand for inspiration | 1 |
164,704 | 6,254,380,767 | IssuesEvent | 2017-07-14 01:51:26 | tootsuite/mastodon | https://api.github.com/repos/tootsuite/mastodon | closed | Make "remove remote media attachments" time period adjustable | admin support enhancement fit for beginners help wanted priority - medium | There is a rake command to drop remote media attachments after a week: `rails run media:remove_remote` ([code](https://github.com/tootsuite/mastodon/blob/75c6513c678148e9a6adfcda57cfd5c8a16c804a/lib/tasks/mastodon.rake#L57-L64)).
This is a great feature to keep the asset directory from growing indefinitely, but a week seems awfully short to be hardcoded.
Let's make this adjustable, either in `.env.production` or otherwise?
* * * *
- [x] I searched or browsed the repoโs other issues to ensure this is not a duplicate.
- [x] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
| 1.0 | Make "remove remote media attachments" time period adjustable - There is a rake command to drop remote media attachments after a week: `rails run media:remove_remote` ([code](https://github.com/tootsuite/mastodon/blob/75c6513c678148e9a6adfcda57cfd5c8a16c804a/lib/tasks/mastodon.rake#L57-L64)).
This is a great feature to keep the asset directory from growing indefinitely, but a week seems awfully short to be hardcoded.
Let's make this adjustable, either in `.env.production` or otherwise?
* * * *
- [x] I searched or browsed the repoโs other issues to ensure this is not a duplicate.
- [x] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
| priority | make remove remote media attachments time period adjustable there is a rake command to drop remote media attachments after a week rails run media remove remote this is a great feature to keep the asset directory from growing indefinitely but a week seems awfully short to be hardcoded let s make this adjustable either in env production or otherwise i searched or browsed the repoโs other issues to ensure this is not a duplicate this bug happens on a and not on master if you re a user don t worry about this | 1 |
763,021 | 26,740,899,820 | IssuesEvent | 2023-01-30 12:55:10 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [Driver] YSQL server list refresh sequence/interval | kind/bug priority/medium area/ecosystem | Jira Link: [DB-4782](https://yugabyte.atlassian.net/browse/DB-4782)
While testing the client-side resiliency from a driver standpoint, encountered the following case
**Initial configuration:**
- Connections per node limited to - 'X' (let's say 50)
- Steady-state would generate roughly 60-70% of the traffic continuously (60-70% out of the max connections would be active). Let's say in a cluster with 3 nodes - out of 150 connections, roughly 80-90 connections would be pre-initialized and kept active.
(assumed 3 node cluster here, but it could vary)
**Use case execution flow**
1. Steady-state
- all 3 nodes (node#1, node#2, node#3) were running
- connections were distributed between 3 nodes
2. Isolate node#3
- connections served by node#3 became invalid
- new compensating connections were distributed to the surviving nodes (node#1 and node#2). Since the two nodes can accommodate up to 100 connections, 80 was still within the upper threshold limit.
3. node#3 joins the cluster in a few mins (in 1-2 mins). Trigger a graceful shutdown of node#2.
Though we have 2 active nodes (node#1 and node#3), the smart driver active server list has only node#1 as it has already invalidated node#3 (when it got separated from the cluster) and node#2 (when it went down). This list takes some time to get refreshed. In this case, a single node couldn't accommodate the entire client traffic and ended up with a lot of connection exceptions for some time.
The exception encountered in this case didn't signify the real issue as it was momentarily and got eventually resolved after the next refresh.
It would be nice to refresh the active server list whenever there is a connection exception to validate whether it is to do with the outdated list or the cluster itself has really hit the max limit. | 1.0 | [Driver] YSQL server list refresh sequence/interval - Jira Link: [DB-4782](https://yugabyte.atlassian.net/browse/DB-4782)
While testing the client-side resiliency from a driver standpoint, encountered the following case
**Initial configuration:**
- Connections per node limited to - 'X' (let's say 50)
- Steady-state would generate roughly 60-70% of the traffic continuously (60-70% out of the max connections would be active). Let's say in a cluster with 3 nodes - out of 150 connections, roughly 80-90 connections would be pre-initialized and kept active.
(assumed 3 node cluster here, but it could vary)
**Use case execution flow**
1. Steady-state
- all 3 nodes (node#1, node#2, node#3) were running
- connections were distributed between 3 nodes
2. Isolate node#3
- connections served by node#3 became invalid
- new compensating connections were distributed to the surviving nodes (node#1 and node#2). Since the two nodes can accommodate up to 100 connections, 80 was still within the upper threshold limit.
3. node#3 joins the cluster in a few mins (in 1-2 mins). Trigger a graceful shutdown of node#2.
Though we have 2 active nodes (node#1 and node#3), the smart driver active server list has only node#1 as it has already invalidated node#3 (when it got separated from the cluster) and node#2 (when it went down). This list takes some time to get refreshed. In this case, a single node couldn't accommodate the entire client traffic and ended up with a lot of connection exceptions for some time.
The exception encountered in this case didn't signify the real issue as it was momentarily and got eventually resolved after the next refresh.
It would be nice to refresh the active server list whenever there is a connection exception to validate whether it is to do with the outdated list or the cluster itself has really hit the max limit. | priority | ysql server list refresh sequence interval jira link while testing the client side resiliency from a driver standpoint encountered the following case initial configuration connections per node limited to x let s say steady state would generate roughly of the traffic continuously out of the max connections would be active let s say in a cluster with nodes out of connections roughly connections would be pre initialized and kept active assumed node cluster here but it could vary use case execution flow steady state all nodes node node node were running connections were distributed between nodes isolate node connections served by node became invalid new compensating connections were distributed to the surviving nodes node and node since the two nodes can accommodate up to connections was still within the upper threshold limit node joins the cluster in a few mins in mins trigger a graceful shutdown of node though we have active nodes node and node the smart driver active server list has only node as it has already invalidated node when it got separated from the cluster and node when it went down this list takes some time to get refreshed in this case a single node couldn t accommodate the entire client traffic and ended up with a lot of connection exceptions for some time the exception encountered in this case didn t signify the real issue as it was momentarily and got eventually resolved after the next refresh it would be nice to refresh the active server list whenever there is a connection exception to validate whether it is to do with the outdated list or the cluster itself has really hit the max limit | 1 |
637,226 | 20,623,589,332 | IssuesEvent | 2022-03-07 19:58:23 | LLK/scratchjr | https://api.github.com/repos/LLK/scratchjr | closed | Should clicking checkmark on blank project act like cancel? | priority 3 Medium Impact Medium Severity needs discussion | From Ben: If you open the character library, select the blank character and then click the checkmark it acts the same as clicking the 'x' cancel. Should it act as if you really clicked the paint button. | 1.0 | Should clicking checkmark on blank project act like cancel? - From Ben: If you open the character library, select the blank character and then click the checkmark it acts the same as clicking the 'x' cancel. Should it act as if you really clicked the paint button. | priority | should clicking checkmark on blank project act like cancel from ben if you open the character library select the blank character and then click the checkmark it acts the same as clicking the x cancel should it act as if you really clicked the paint button | 1 |
709,229 | 24,371,150,831 | IssuesEvent | 2022-10-03 19:24:56 | AMastryukov/LD51 | https://api.github.com/repos/AMastryukov/LD51 | closed | 3D Environment/Character Shader and Texture | Medium Priority Art | **Description**
Add unique texture to character, unify all 3d assets with appropriate shader (possibly clay)
**References**
https://dribbble.com/shots/18434463-Clay-doh-Shader-Study
| 1.0 | 3D Environment/Character Shader and Texture - **Description**
Add unique texture to character, unify all 3d assets with appropriate shader (possibly clay)
**References**
https://dribbble.com/shots/18434463-Clay-doh-Shader-Study
| priority | environment character shader and texture description add unique texture to character unify all assets with appropriate shader possibly clay references | 1 |
716,092 | 24,620,845,825 | IssuesEvent | 2022-10-15 22:41:24 | MLVETDevelopers/mlvet | https://api.github.com/repos/MLVETDevelopers/mlvet | closed | Entire feedback button should be clickable | minor-change Medium-priority | Current behaviour: only icon is clickable
<img width="103" alt="Screen Shot 2022-10-02 at 9 19 58 pm" src="https://user-images.githubusercontent.com/6735055/193449250-430f9eb0-5bbf-4c17-aa9a-1ba282f2da35.png">
Expected behaviour: the entire area around the icon and the text are a single click target
Reasoning: I went to click it to test it out and assumed that clicking the text would make the popup appear, but you actually have to click the icon itself - and there is also almost no padding, so you have to click it exactly. There is basically no downside to making the whole area clickable since there are no other click targets nearby to get confused with. | 1.0 | Entire feedback button should be clickable - Current behaviour: only icon is clickable
<img width="103" alt="Screen Shot 2022-10-02 at 9 19 58 pm" src="https://user-images.githubusercontent.com/6735055/193449250-430f9eb0-5bbf-4c17-aa9a-1ba282f2da35.png">
Expected behaviour: the entire area around the icon and the text are a single click target
Reasoning: I went to click it to test it out and assumed that clicking the text would make the popup appear, but you actually have to click the icon itself - and there is also almost no padding, so you have to click it exactly. There is basically no downside to making the whole area clickable since there are no other click targets nearby to get confused with. | priority | entire feedback button should be clickable current behaviour only icon is clickable img width alt screen shot at pm src expected behaviour the entire area around the icon and the text are a single click target reasoning i went to click it to test it out and assumed that clicking the text would make the popup appear but you actually have to click the icon itself and there is also almost no padding so you have to click it exactly there is basically no downside to making the whole area clickable since there are no other click targets nearby to get confused with | 1 |
217,888 | 7,328,758,398 | IssuesEvent | 2018-03-04 23:46:27 | JianweiCxyz/UIL_webpages | https://api.github.com/repos/JianweiCxyz/UIL_webpages | closed | Browser compatibility | Longer Term Medium Priority bug | Desktop
- Firefox: all fine.
- Chrome: all fine.
- IE: cannot display legends
- Edge: cannot display legends
Mobile (ios)
- Safari / Chrome: cannot display maps
Mobile (Android)
- Unknown | 1.0 | Browser compatibility - Desktop
- Firefox: all fine.
- Chrome: all fine.
- IE: cannot display legends
- Edge: cannot display legends
Mobile (ios)
- Safari / Chrome: cannot display maps
Mobile (Android)
- Unknown | priority | browser compatibility desktop firefox all fine chrome all fine ie cannot display legends edge cannot display legends mobile ios safari chrome cannot display maps mobile android unknown | 1 |
400,228 | 11,770,716,997 | IssuesEvent | 2020-03-15 20:28:49 | Extum/material | https://api.github.com/repos/Extum/material | closed | Device Compatibility | enhancement mobile priority: medium | - [ ] Mobile Compatiblity (Devices with max-width view-port of 360px - Extra Small)
- [ ] Mobile and Tablet Compatiblity (Devices with min-width view-port of 361px and max-width of 768px - Small)
- [ ] Tablet Compatiblity (Devices with min-width view-port of 769px and max-width of 1024px - Medium)
- [ ] Laptop Compatiblity (Devices with min-width view-port of 1025px and max-width of 1280px - Large)
- [ ] Desktop Compatiblity (Devices with min-width view-port of 1281px - Extra Large)
**This is expected to be released in the middle of `1.0.0-beta.4` but, will be completely finished by the end of `1.0.0-beta.5`** (Do **NOT** remove the Milestone for `1.0.0-beta.5`) | 1.0 | Device Compatibility - - [ ] Mobile Compatiblity (Devices with max-width view-port of 360px - Extra Small)
- [ ] Mobile and Tablet Compatiblity (Devices with min-width view-port of 361px and max-width of 768px - Small)
- [ ] Tablet Compatiblity (Devices with min-width view-port of 769px and max-width of 1024px - Medium)
- [ ] Laptop Compatiblity (Devices with min-width view-port of 1025px and max-width of 1280px - Large)
- [ ] Desktop Compatiblity (Devices with min-width view-port of 1281px - Extra Large)
**This is expected to be released in the middle of `1.0.0-beta.4` but, will be completely finished by the end of `1.0.0-beta.5`** (Do **NOT** remove the Milestone for `1.0.0-beta.5`) | priority | device compatibility mobile compatiblity devices with max width view port of extra small mobile and tablet compatiblity devices with min width view port of and max width of small tablet compatiblity devices with min width view port of and max width of medium laptop compatiblity devices with min width view port of and max width of large desktop compatiblity devices with min width view port of extra large this is expected to be released in the middle of beta but will be completely finished by the end of beta do not remove the milestone for beta | 1 |
525,006 | 15,227,278,049 | IssuesEvent | 2021-02-18 09:59:38 | google/flax | https://api.github.com/repos/google/flax | closed | Mutable / Unmutable state when training | Priority: P2 - medium | Hi Flax team,
I'm working on a model with an internal state which gets updated during training. When calling the model during validation, I do not want to update these variables. I could technically introduce a training variable, but I feel it could be more elegantly (and much simpler!) by simply checking if a state is mutable:
1 ) Is there an easy way to check if a variable is mutable, so that my code only updates when it is? e.g.
`if is_initialized and var.is_mutable(): var.value = f(....)`
2 ) If I set mutable=False, I only get back the output. Is there a way to get back the state regardless? e.g.
`output, updated_state = model.apply(inputs, mutable=False, return_state=True)`
My usecase is that for my validation metrics I call my loss function with the testdata and extract the metrics, so that I can use the same code for both training and validation.
Thanks! | 1.0 | Mutable / Unmutable state when training - Hi Flax team,
I'm working on a model with an internal state which gets updated during training. When calling the model during validation, I do not want to update these variables. I could technically introduce a training variable, but I feel it could be more elegantly (and much simpler!) by simply checking if a state is mutable:
1 ) Is there an easy way to check if a variable is mutable, so that my code only updates when it is? e.g.
`if is_initialized and var.is_mutable(): var.value = f(....)`
2 ) If I set mutable=False, I only get back the output. Is there a way to get back the state regardless? e.g.
`output, updated_state = model.apply(inputs, mutable=False, return_state=True)`
My usecase is that for my validation metrics I call my loss function with the testdata and extract the metrics, so that I can use the same code for both training and validation.
Thanks! | priority | mutable unmutable state when training hi flax team i m working on a model with an internal state which gets updated during training when calling the model during validation i do not want to update these variables i could technically introduce a training variable but i feel it could be more elegantly and much simpler by simply checking if a state is mutable is there an easy way to check if a variable is mutable so that my code only updates when it is e g if is initialized and var is mutable var value f if i set mutable false i only get back the output is there a way to get back the state regardless e g output updated state model apply inputs mutable false return state true my usecase is that for my validation metrics i call my loss function with the testdata and extract the metrics so that i can use the same code for both training and validation thanks | 1 |
322,854 | 9,829,363,771 | IssuesEvent | 2019-06-15 20:03:09 | gitblit/gitblit | https://api.github.com/repos/gitblit/gitblit | closed | Feature request: Support nested groups on the Repositories page | Priority-Medium Status-Done Type-Enhancement | Originally reported on Google Code with ID 429
```
In my organization we have a lot of Git repositories and would like to be able to group
the repositories in a nested fashion, e.g.
Managers
- Manager X
- Server
- Client
- Manager Y
- Server
- Database
- Client
etc.
Here https://code.google.com/p/gitblit/issues/detail?id=406#c1 it is mentioned that
it's possible to nest repos with '/', so what I'm suggesting is really just that the
interface reflects this tree structure more clearly than now. Furthermore the GUI change
here https://code.google.com/p/gitblit/issues/detail?id=231&q=collapse#makechanges,
will fit nicely into this feature.
```
Reported by `jakob@jboysen.dk` on 2014-05-20 11:04:35
| 1.0 | Feature request: Support nested groups on the Repositories page - Originally reported on Google Code with ID 429
```
In my organization we have a lot of Git repositories and would like to be able to group
the repositories in a nested fashion, e.g.
Managers
- Manager X
- Server
- Client
- Manager Y
- Server
- Database
- Client
etc.
Here https://code.google.com/p/gitblit/issues/detail?id=406#c1 it is mentioned that
it's possible to nest repos with '/', so what I'm suggesting is really just that the
interface reflects this tree structure more clearly than now. Furthermore the GUI change
here https://code.google.com/p/gitblit/issues/detail?id=231&q=collapse#makechanges,
will fit nicely into this feature.
```
Reported by `jakob@jboysen.dk` on 2014-05-20 11:04:35
| priority | feature request support nested groups on the repositories page originally reported on google code with id in my organization we have a lot of git repositories and would like to be able to group the repositories in a nested fashion e g managers manager x server client manager y server database client etc here it is mentioned that it s possible to nest repos with so what i m suggesting is really just that the interface reflects this tree structure more clearly than now furthermore the gui change here will fit nicely into this feature reported by jakob jboysen dk on | 1 |
316,664 | 9,653,037,863 | IssuesEvent | 2019-05-18 23:11:10 | nimona/go-nimona | https://api.github.com/repos/nimona/go-nimona | closed | Add logger | Priority: Medium Type: Enhancement | A structured logger should be implemented for use across the project. Preferably over zap. | 1.0 | Add logger - A structured logger should be implemented for use across the project. Preferably over zap. | priority | add logger a structured logger should be implemented for use across the project preferably over zap | 1 |
7,037 | 2,596,636,551 | IssuesEvent | 2015-02-20 22:06:42 | uic-evl/omegalib | https://api.github.com/repos/uic-evl/omegalib | closed | Z buffer precision | priority:MEDIUM status:DROPPED type:BUG | setting the near plane to fractional values usually causes Z fighting. Issue seems to be related with Z buffer precision (which should be set to 24 bits by default but does not behave like that) | 1.0 | Z buffer precision - setting the near plane to fractional values usually causes Z fighting. Issue seems to be related with Z buffer precision (which should be set to 24 bits by default but does not behave like that) | priority | z buffer precision setting the near plane to fractional values usually causes z fighting issue seems to be related with z buffer precision which should be set to bits by default but does not behave like that | 1 |
202,937 | 7,056,541,620 | IssuesEvent | 2018-01-04 13:12:55 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | opened | Multitenant: ip4log issue while restarting iptables | Priority: Medium Type: Bug | Getting 2 errors while restarting iptables on the latest feature/multitenant commit
Updated the schema to the latest version
```
Jan 04 08:04:24 pf-julien.inverse packetfence[30275]: ERROR -e(30275): Database query failed with non retryable error: Unknown column 'type' in 'field list' (errno: 1054) [SELECT `mac`, `ip`, `type`, `start_time`, `end_time` FROM `ip4log` WHERE ( ( ( `end_time` = ? OR `end_time` > NOW() ) AND `ip4log`.`tenant_id` = ? ) )]{0000-00-00 00:00:00, 1} (pf::dal::db_execute)
``` | 1.0 | Multitenant: ip4log issue while restarting iptables - Getting 2 errors while restarting iptables on the latest feature/multitenant commit
Updated the schema to the latest version
```
Jan 04 08:04:24 pf-julien.inverse packetfence[30275]: ERROR -e(30275): Database query failed with non retryable error: Unknown column 'type' in 'field list' (errno: 1054) [SELECT `mac`, `ip`, `type`, `start_time`, `end_time` FROM `ip4log` WHERE ( ( ( `end_time` = ? OR `end_time` > NOW() ) AND `ip4log`.`tenant_id` = ? ) )]{0000-00-00 00:00:00, 1} (pf::dal::db_execute)
``` | priority | multitenant issue while restarting iptables getting errors while restarting iptables on the latest feature multitenant commit updated the schema to the latest version jan pf julien inverse packetfence error e database query failed with non retryable error unknown column type in field list errno pf dal db execute | 1 |
286,538 | 8,789,621,512 | IssuesEvent | 2018-12-21 04:52:08 | ilmtest/search-engine | https://api.github.com/repos/ilmtest/search-engine | closed | Add a feature to reply to the last post on Telegram | feature priority/medium server usability user-interface | This way it makes it easy to publish tafaseer | 1.0 | Add a feature to reply to the last post on Telegram - This way it makes it easy to publish tafaseer | priority | add a feature to reply to the last post on telegram this way it makes it easy to publish tafaseer | 1 |
254,403 | 8,073,699,982 | IssuesEvent | 2018-08-06 20:11:43 | minio/minio | https://api.github.com/repos/minio/minio | reopened | Distributed 3-node Minio stops working when 1 of 3 nodes is down | priority: medium triage working as intended | <!--- Provide a general summary of the issue in the Title above -->
On FreeBSD 11.1 amd64 we have 3 nodes for Minio Distributed cluster - an asymetric setup as follows:
node1
[01/16] http://node1:9000/data0 - 246 GiB online
[02/16] http://node1:9000/data1 - 246 GiB online
[03/16] http://node1:9000/data2 - 246 GiB online
[04/16] http://node1:9000/data3 - 246 GiB online
[05/16] http://node1:9000/data4 - 246 GiB online
[06/16] http://node1:9000/data5 - 246 GiB online
[07/16] http://node1:9000/data6 - 246 GiB online
node2
[08/16] http://node2:9000/data0 - 246 GiB online
[09/16] http://node2:9000/data1 - 246 GiB online
[10/16] http://node2:9000/data2 - 246 GiB online
[11/16] http://node2:9000/data3 - 246 GiB online
[12/16] http://node2:9000/data4 - 246 GiB online
[13/16] http://node2:9000/data5 - 246 GiB online
[14/16] http://node2:9000/data6 - 246 GiB online
node3
[15/16] http://node3:9000/data0 - 77 GiB online
[16/16] http://node3:9000/data1 - 77 GiB online
## Expected Behavior
In theory if I powerdown only ONE of them, Minio should work just fine for both writes and read, yet if I pwoerdown node2 for example them Minio becomes unusable - not able to show files in the web interface for example.
## Current Behavior
... and I do not know whay I am doing wrong.
## Your Environment
**os**
```
# uname -a
FreeBSD c003s7303.gkpge.pl 11.1-RELEASE-p4 FreeBSD 11.1-RELEASE-p4 #0: Tue Nov 14 06:12:40 UTC 2017 root@amd64-builder.daemonology.net:/usr/obj/usr/src/sys/GENERIC amd64
```
**version**
```
# pkg info | grep minio
minio-2017.11.22.19.55.46 Amazon S3 compatible object storage server
minio-client-2017.02.06.20.16.19_1 Replacement for ls, cp, mkdir, diff and rsync commands for filesystems
```
**node1 | node2**
```
# zfs list -r s3/storage
NAME USED AVAIL REFER MOUNTPOINT
s3/storage 512M 246G 23K none
s3/storage/data0 73.7M 246G 73.7M /data0
s3/storage/data1 73.2M 246G 73.2M /data1
s3/storage/data2 73.4M 246G 73.4M /data2
s3/storage/data3 72.7M 246G 72.7M /data3
s3/storage/data4 72.8M 246G 72.8M /data4
s3/storage/data5 73.0M 246G 73.0M /data5
s3/storage/data6 73.5M 246G 73.5M /data6
```
**node3**
```
# zfs list -r s3/storage
NAME USED AVAIL REFER MOUNTPOINT
s3/storage 108K 77.0G 23K none
s3/storage/data0 43K 77.0G 43K /data0
s3/storage/data1 42K 77.0G 42K /data1
```
**command to start**
```
# minio server --config-dir /usr/local/etc/minio http://node3/data0 http://node3/data1 http://node1/data0 http://node1/data1 http://node1/data2 http://node1/data3 http://node1/data4 http://node1/data5 http://node1/data6 http://node2/data0 http://node2/data1 http://node2/data2 http://node2/data3 http://node2/data4 http://node2/data5 http://node2/data6
Initializing data volume.
[01/16] http://node1:9000/data0 - 246 GiB online
[02/16] http://node1:9000/data1 - 246 GiB online
[03/16] http://node1:9000/data2 - 246 GiB online
[04/16] http://node1:9000/data3 - 246 GiB online
[05/16] http://node1:9000/data4 - 246 GiB online
[06/16] http://node1:9000/data5 - 246 GiB online
[07/16] http://node1:9000/data6 - 246 GiB online
[08/16] http://node2:9000/data0 - 246 GiB online
[09/16] http://node2:9000/data1 - 246 GiB online
[10/16] http://node2:9000/data2 - 246 GiB online
[11/16] http://node2:9000/data3 - 246 GiB online
[12/16] http://node2:9000/data4 - 246 GiB online
[13/16] http://node2:9000/data5 - 246 GiB online
[14/16] http://node2:9000/data6 - 246 GiB online
[15/16] http://node3:9000/data0 - 77 GiB online
[16/16] http://node3:9000/data1 - 77 GiB online
Endpoint: http://IP3:9000 http://127.0.0.1:9000
AccessKey: AccessKey
SecretKey: SecretKey
Browser Access:
http://node2:9000 http://127.0.0.1:9000
Command-line Access: https://docs.minio.io/docs/minio-client-quickstart-guide
$ mc config host add myminio http://node2:9000 AccessKey SecretKey
Object API (Amazon S3 compatible):
Go: https://docs.minio.io/docs/golang-client-quickstart-guide
Java: https://docs.minio.io/docs/java-client-quickstart-guide
Python: https://docs.minio.io/docs/python-client-quickstart-guide
JavaScript: https://docs.minio.io/docs/javascript-client-quickstart-guide
.NET: https://docs.minio.io/docs/dotnet-client-quickstart-guide
Drive Capacity: 616 GiB Free, 616 GiB Total
Status: 16 Online, 0 Offline. We can withstand [8] drive failure(s).
```
**log**
```
# cat /var/log/minio.log
time="2018-07-13T10:17:42+02:00" level=error msg="Unable to shutdown http server" cause="timed out. some connections are still active. doing abnormal shutdown" source="[signals.go:42:handleSignals.func2()]"
```
**config**
```
# cat /usr/local/etc/minio/config.json
{
"version": "20",
"credential": {
"accessKey": "AccessKey",
"secretKey": "SecretKey"
},
"region": "",
"browser": "on",
"domain": "",
"logger": {
"console": {
"enable": true
},
"file": {
"enable": false,
"filename": ""
}
},
"notify": {
"amqp": {
"1": {
"enable": false,
"url": "",
"exchange": "",
"routingKey": "",
"exchangeType": "",
"deliveryMode": 0,
"mandatory": false,
"immediate": false,
"durable": false,
"internal": false,
"noWait": false,
"autoDeleted": false
}
},
"nats": {
"1": {
"enable": false,
"address": "",
"subject": "",
"username": "",
"password": "",
"token": "",
"secure": false,
"pingInterval": 0,
"streaming": {
"enable": false,
"clusterID": "",
"clientID": "",
"async": false,
"maxPubAcksInflight": 0
}
}
},
"elasticsearch": {
"1": {
"enable": false,
"format": "",
"url": "",
"index": ""
}
},
"redis": {
"1": {
"enable": false,
"format": "",
"address": "",
"password": "",
"key": ""
}
},
"postgresql": {
"1": {
"enable": false,
"format": "",
"connectionString": "",
"table": "",
"host": "",
"port": "",
"user": "",
"password": "",
"database": ""
}
},
"kafka": {
"1": {
"enable": false,
"brokers": null,
"topic": ""
}
},
"webhook": {
"1": {
"enable": false,
"endpoint": ""
}
},
"mysql": {
"1": {
"enable": false,
"format": "",
"dsnString": "",
"table": "",
"host": "",
"port": "",
"user": "",
"password": "",
"database": ""
}
},
"mqtt": {
"1": {
"enable": false,
"broker": "",
"topic": "",
"qos": 0,
"clientId": "",
"username": "",
"password": ""
}
}
}
}
```
Thanks in advance.
Regards,
vermaden
| 1.0 | Distributed 3-node Minio stops working when 1 of 3 nodes is down - <!--- Provide a general summary of the issue in the Title above -->
On FreeBSD 11.1 amd64 we have 3 nodes for Minio Distributed cluster - an asymetric setup as follows:
node1
[01/16] http://node1:9000/data0 - 246 GiB online
[02/16] http://node1:9000/data1 - 246 GiB online
[03/16] http://node1:9000/data2 - 246 GiB online
[04/16] http://node1:9000/data3 - 246 GiB online
[05/16] http://node1:9000/data4 - 246 GiB online
[06/16] http://node1:9000/data5 - 246 GiB online
[07/16] http://node1:9000/data6 - 246 GiB online
node2
[08/16] http://node2:9000/data0 - 246 GiB online
[09/16] http://node2:9000/data1 - 246 GiB online
[10/16] http://node2:9000/data2 - 246 GiB online
[11/16] http://node2:9000/data3 - 246 GiB online
[12/16] http://node2:9000/data4 - 246 GiB online
[13/16] http://node2:9000/data5 - 246 GiB online
[14/16] http://node2:9000/data6 - 246 GiB online
node3
[15/16] http://node3:9000/data0 - 77 GiB online
[16/16] http://node3:9000/data1 - 77 GiB online
## Expected Behavior
In theory if I powerdown only ONE of them, Minio should work just fine for both writes and read, yet if I pwoerdown node2 for example them Minio becomes unusable - not able to show files in the web interface for example.
## Current Behavior
... and I do not know whay I am doing wrong.
## Your Environment
**os**
```
# uname -a
FreeBSD c003s7303.gkpge.pl 11.1-RELEASE-p4 FreeBSD 11.1-RELEASE-p4 #0: Tue Nov 14 06:12:40 UTC 2017 root@amd64-builder.daemonology.net:/usr/obj/usr/src/sys/GENERIC amd64
```
**version**
```
# pkg info | grep minio
minio-2017.11.22.19.55.46 Amazon S3 compatible object storage server
minio-client-2017.02.06.20.16.19_1 Replacement for ls, cp, mkdir, diff and rsync commands for filesystems
```
**node1 | node2**
```
# zfs list -r s3/storage
NAME USED AVAIL REFER MOUNTPOINT
s3/storage 512M 246G 23K none
s3/storage/data0 73.7M 246G 73.7M /data0
s3/storage/data1 73.2M 246G 73.2M /data1
s3/storage/data2 73.4M 246G 73.4M /data2
s3/storage/data3 72.7M 246G 72.7M /data3
s3/storage/data4 72.8M 246G 72.8M /data4
s3/storage/data5 73.0M 246G 73.0M /data5
s3/storage/data6 73.5M 246G 73.5M /data6
```
**node3**
```
# zfs list -r s3/storage
NAME USED AVAIL REFER MOUNTPOINT
s3/storage 108K 77.0G 23K none
s3/storage/data0 43K 77.0G 43K /data0
s3/storage/data1 42K 77.0G 42K /data1
```
**command to start**
```
# minio server --config-dir /usr/local/etc/minio http://node3/data0 http://node3/data1 http://node1/data0 http://node1/data1 http://node1/data2 http://node1/data3 http://node1/data4 http://node1/data5 http://node1/data6 http://node2/data0 http://node2/data1 http://node2/data2 http://node2/data3 http://node2/data4 http://node2/data5 http://node2/data6
Initializing data volume.
[01/16] http://node1:9000/data0 - 246 GiB online
[02/16] http://node1:9000/data1 - 246 GiB online
[03/16] http://node1:9000/data2 - 246 GiB online
[04/16] http://node1:9000/data3 - 246 GiB online
[05/16] http://node1:9000/data4 - 246 GiB online
[06/16] http://node1:9000/data5 - 246 GiB online
[07/16] http://node1:9000/data6 - 246 GiB online
[08/16] http://node2:9000/data0 - 246 GiB online
[09/16] http://node2:9000/data1 - 246 GiB online
[10/16] http://node2:9000/data2 - 246 GiB online
[11/16] http://node2:9000/data3 - 246 GiB online
[12/16] http://node2:9000/data4 - 246 GiB online
[13/16] http://node2:9000/data5 - 246 GiB online
[14/16] http://node2:9000/data6 - 246 GiB online
[15/16] http://node3:9000/data0 - 77 GiB online
[16/16] http://node3:9000/data1 - 77 GiB online
Endpoint: http://IP3:9000 http://127.0.0.1:9000
AccessKey: AccessKey
SecretKey: SecretKey
Browser Access:
http://node2:9000 http://127.0.0.1:9000
Command-line Access: https://docs.minio.io/docs/minio-client-quickstart-guide
$ mc config host add myminio http://node2:9000 AccessKey SecretKey
Object API (Amazon S3 compatible):
Go: https://docs.minio.io/docs/golang-client-quickstart-guide
Java: https://docs.minio.io/docs/java-client-quickstart-guide
Python: https://docs.minio.io/docs/python-client-quickstart-guide
JavaScript: https://docs.minio.io/docs/javascript-client-quickstart-guide
.NET: https://docs.minio.io/docs/dotnet-client-quickstart-guide
Drive Capacity: 616 GiB Free, 616 GiB Total
Status: 16 Online, 0 Offline. We can withstand [8] drive failure(s).
```
**log**
```
# cat /var/log/minio.log
time="2018-07-13T10:17:42+02:00" level=error msg="Unable to shutdown http server" cause="timed out. some connections are still active. doing abnormal shutdown" source="[signals.go:42:handleSignals.func2()]"
```
**config**
```
# cat /usr/local/etc/minio/config.json
{
"version": "20",
"credential": {
"accessKey": "AccessKey",
"secretKey": "SecretKey"
},
"region": "",
"browser": "on",
"domain": "",
"logger": {
"console": {
"enable": true
},
"file": {
"enable": false,
"filename": ""
}
},
"notify": {
"amqp": {
"1": {
"enable": false,
"url": "",
"exchange": "",
"routingKey": "",
"exchangeType": "",
"deliveryMode": 0,
"mandatory": false,
"immediate": false,
"durable": false,
"internal": false,
"noWait": false,
"autoDeleted": false
}
},
"nats": {
"1": {
"enable": false,
"address": "",
"subject": "",
"username": "",
"password": "",
"token": "",
"secure": false,
"pingInterval": 0,
"streaming": {
"enable": false,
"clusterID": "",
"clientID": "",
"async": false,
"maxPubAcksInflight": 0
}
}
},
"elasticsearch": {
"1": {
"enable": false,
"format": "",
"url": "",
"index": ""
}
},
"redis": {
"1": {
"enable": false,
"format": "",
"address": "",
"password": "",
"key": ""
}
},
"postgresql": {
"1": {
"enable": false,
"format": "",
"connectionString": "",
"table": "",
"host": "",
"port": "",
"user": "",
"password": "",
"database": ""
}
},
"kafka": {
"1": {
"enable": false,
"brokers": null,
"topic": ""
}
},
"webhook": {
"1": {
"enable": false,
"endpoint": ""
}
},
"mysql": {
"1": {
"enable": false,
"format": "",
"dsnString": "",
"table": "",
"host": "",
"port": "",
"user": "",
"password": "",
"database": ""
}
},
"mqtt": {
"1": {
"enable": false,
"broker": "",
"topic": "",
"qos": 0,
"clientId": "",
"username": "",
"password": ""
}
}
}
}
```
Thanks in advance.
Regards,
vermaden
| priority | distributed node minio stops working when of nodes is down on freebsd we have nodes for minio distributed cluster an asymetric setup as follows gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online expected behavior in theory if i powerdown only one of them minio should work just fine for both writes and read yet if i pwoerdown for example them minio becomes unusable not able to show files in the web interface for example current behavior and i do not know whay i am doing wrong your environment os uname a freebsd gkpge pl release freebsd release tue nov utc root builder daemonology net usr obj usr src sys generic version pkg info grep minio minio amazon compatible object storage server minio client replacement for ls cp mkdir diff and rsync commands for filesystems zfs list r storage name used avail refer mountpoint storage none storage storage storage storage storage storage storage zfs list r storage name used avail refer mountpoint storage none storage storage command to start minio server config dir usr local etc minio initializing data volume gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online gib online endpoint accesskey accesskey secretkey secretkey browser access command line access mc config host add myminio accesskey secretkey object api amazon compatible go java python javascript net drive capacity gib free gib total status online offline we can withstand drive failure s log cat var log minio log time level error msg unable to shutdown http server cause timed out some connections are still active doing abnormal shutdown source config cat usr local etc minio config json version credential accesskey accesskey secretkey secretkey region browser on domain logger console enable true file enable false filename notify amqp enable false url exchange routingkey exchangetype deliverymode mandatory false immediate false durable false internal false nowait false autodeleted false nats enable false address subject username password token secure false pinginterval streaming enable false clusterid clientid async false maxpubacksinflight elasticsearch enable false format url index redis enable false format address password key postgresql enable false format connectionstring table host port user password database kafka enable false brokers null topic webhook enable false endpoint mysql enable false format dsnstring table host port user password database mqtt enable false broker topic qos clientid username password thanks in advance regards vermaden | 1 |
576,104 | 17,079,183,542 | IssuesEvent | 2021-07-08 00:47:06 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | reopened | Unable to use `azure_rm_aduser_info` or `azure_rm_adgroup_info`: Insufficient privileges to complete the operation. | medium_priority | ##### SUMMARY
I am unable to use the `azure_rm_adgroup_info` or `azure_rm_aduser_info` modules. They report:
> "msg": "failed to get ad group info Insufficient privileges to complete the operation."
I am using a service principal which has Owner access for the subscription and looks like it has sufficient API access for user information:
<img width="1223" alt="Screen Shot 2021-07-03 at 2 17 15 PM" src="https://user-images.githubusercontent.com/2243641/124364829-8b035580-dc09-11eb-8105-584e9c26c3f2.png">
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_adgroup_info
azure_rm_aduser_info
##### ANSIBLE VERSION
```bash
-> ansible --version
ansible [core 2.11.2]
config file = /Users/ross/repos/azure-config/ansible.cfg
configured module search path = ['/Users/ross/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /Users/ross/repos/azure-config/venv/lib/python3.9/site-packages/ansible
ansible collection location = /Users/ross/.ansible/collections:/usr/share/ansible/collections
executable location = /Users/ross/repos/azure-config/venv/bin/ansible
python version = 3.9.1 (default, Dec 10 2020, 10:36:35) [Clang 12.0.0 (clang-1200.0.32.27)]
jinja version = 3.0.1
libyaml = True
```
##### COLLECTION VERSION
```bash
-> ansible-galaxy collection list azure.azcollection
# /Users/ross/repos/azure-config/venv/lib/python3.9/site-packages/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.7.0
# /Users/ross/.ansible/collections/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.7.0
```
##### CONFIGURATION
```bash
-> ansible-config dump --only-changed
DISPLAY_SKIPPED_HOSTS(/Users/ross/repos/azure-config/ansible.cfg) = False
```
##### OS / ENVIRONMENT
Mac OS Big Sur 11.4.
##### STEPS TO REPRODUCE
I am testing using the following playbook:
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: "Test loading AD group info"
hosts: localhost
connection: local
gather_facts: false
tasks:
- name: "Get group info."
azure.azcollection.azure_rm_adgroup_info:
tenant: "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx"
all: true
register: group_info
```
##### EXPECTED RESULTS
I would be able to retrieve results of groups in the tenant.
##### ACTUAL RESULTS
Module fails with error.
```bash
-> ansible-playbook sandbox.yml
[WARNING]: No inventory was parsed, only implicit localhost is available
[WARNING]: provided hosts list is empty, only localhost is available. Note that the implicit localhost does not match 'all'
PLAY [Test loading AD group info] *******************************************************************************************************************************************************
TASK [Get group info.] ******************************************************************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "failed to get ad group info Insufficient privileges to complete the operation."}
PLAY RECAP ******************************************************************************************************************************************************************************
localhost : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0
```
Reference: https://github.com/ansible-collections/azure/pull/423. | 1.0 | Unable to use `azure_rm_aduser_info` or `azure_rm_adgroup_info`: Insufficient privileges to complete the operation. - ##### SUMMARY
I am unable to use the `azure_rm_adgroup_info` or `azure_rm_aduser_info` modules. They report:
> "msg": "failed to get ad group info Insufficient privileges to complete the operation."
I am using a service principal which has Owner access for the subscription and looks like it has sufficient API access for user information:
<img width="1223" alt="Screen Shot 2021-07-03 at 2 17 15 PM" src="https://user-images.githubusercontent.com/2243641/124364829-8b035580-dc09-11eb-8105-584e9c26c3f2.png">
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_adgroup_info
azure_rm_aduser_info
##### ANSIBLE VERSION
```bash
-> ansible --version
ansible [core 2.11.2]
config file = /Users/ross/repos/azure-config/ansible.cfg
configured module search path = ['/Users/ross/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /Users/ross/repos/azure-config/venv/lib/python3.9/site-packages/ansible
ansible collection location = /Users/ross/.ansible/collections:/usr/share/ansible/collections
executable location = /Users/ross/repos/azure-config/venv/bin/ansible
python version = 3.9.1 (default, Dec 10 2020, 10:36:35) [Clang 12.0.0 (clang-1200.0.32.27)]
jinja version = 3.0.1
libyaml = True
```
##### COLLECTION VERSION
```bash
-> ansible-galaxy collection list azure.azcollection
# /Users/ross/repos/azure-config/venv/lib/python3.9/site-packages/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.7.0
# /Users/ross/.ansible/collections/ansible_collections
Collection Version
------------------ -------
azure.azcollection 1.7.0
```
##### CONFIGURATION
```bash
-> ansible-config dump --only-changed
DISPLAY_SKIPPED_HOSTS(/Users/ross/repos/azure-config/ansible.cfg) = False
```
##### OS / ENVIRONMENT
Mac OS Big Sur 11.4.
##### STEPS TO REPRODUCE
I am testing using the following playbook:
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: "Test loading AD group info"
hosts: localhost
connection: local
gather_facts: false
tasks:
- name: "Get group info."
azure.azcollection.azure_rm_adgroup_info:
tenant: "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx"
all: true
register: group_info
```
##### EXPECTED RESULTS
I would be able to retrieve results of groups in the tenant.
##### ACTUAL RESULTS
Module fails with error.
```bash
-> ansible-playbook sandbox.yml
[WARNING]: No inventory was parsed, only implicit localhost is available
[WARNING]: provided hosts list is empty, only localhost is available. Note that the implicit localhost does not match 'all'
PLAY [Test loading AD group info] *******************************************************************************************************************************************************
TASK [Get group info.] ******************************************************************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "failed to get ad group info Insufficient privileges to complete the operation."}
PLAY RECAP ******************************************************************************************************************************************************************************
localhost : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0
```
Reference: https://github.com/ansible-collections/azure/pull/423. | priority | unable to use azure rm aduser info or azure rm adgroup info insufficient privileges to complete the operation summary i am unable to use the azure rm adgroup info or azure rm aduser info modules they report msg failed to get ad group info insufficient privileges to complete the operation i am using a service principal which has owner access for the subscription and looks like it has sufficient api access for user information img width alt screen shot at pm src issue type bug report component name azure rm adgroup info azure rm aduser info ansible version bash ansible version ansible config file users ross repos azure config ansible cfg configured module search path ansible python module location users ross repos azure config venv lib site packages ansible ansible collection location users ross ansible collections usr share ansible collections executable location users ross repos azure config venv bin ansible python version default dec jinja version libyaml true collection version bash ansible galaxy collection list azure azcollection users ross repos azure config venv lib site packages ansible collections collection version azure azcollection users ross ansible collections ansible collections collection version azure azcollection configuration bash ansible config dump only changed display skipped hosts users ross repos azure config ansible cfg false os environment mac os big sur steps to reproduce i am testing using the following playbook yaml name test loading ad group info hosts localhost connection local gather facts false tasks name get group info azure azcollection azure rm adgroup info tenant xxxxxxxx xxxx xxxx xxxx xxxxxxxxxxxx all true register group info expected results i would be able to retrieve results of groups in the tenant actual results module fails with error bash ansible playbook sandbox yml no inventory was parsed only implicit localhost is available provided hosts list is empty only localhost is available note that the implicit localhost does not match all play task fatal failed changed false msg failed to get ad group info insufficient privileges to complete the operation play recap localhost ok changed unreachable failed skipped rescued ignored reference | 1 |
751,892 | 26,263,394,055 | IssuesEvent | 2023-01-06 10:08:14 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | Patients: Not able to create a patients | Effort: medium Priority: release blocker Bug: development | ## Describe the bug
Not able to create a new patient.
### To reproduce
Steps to reproduce the behaviour:
1. Go to Dispensing.
2. Click on New patients.
3. Fill all mandatory field.
4. See error
### Expected behaviour
Must allow creation of new patients.
### Proposed Solution
Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase.
### Version and device info
- App version: 8.6.2 rc03
- Tablet model: lenovo
- OS version:
### Additional context
Add any other context about the problem here.
| 1.0 | Patients: Not able to create a patients - ## Describe the bug
Not able to create a new patient.
### To reproduce
Steps to reproduce the behaviour:
1. Go to Dispensing.
2. Click on New patients.
3. Fill all mandatory field.
4. See error
### Expected behaviour
Must allow creation of new patients.
### Proposed Solution
Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase.
### Version and device info
- App version: 8.6.2 rc03
- Tablet model: lenovo
- OS version:
### Additional context
Add any other context about the problem here.
| priority | patients not able to create a patients describe the bug not able to create a new patient to reproduce steps to reproduce the behaviour go to dispensing click on new patients fill all mandatory field see error expected behaviour must allow creation of new patients proposed solution leave if you don t know how to fix implement edit this issue description and explain here if you know the best path of implementing the fix within the codebase version and device info app version tablet model lenovo os version additional context add any other context about the problem here | 1 |
30,522 | 2,723,996,227 | IssuesEvent | 2015-04-14 15:31:25 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | Screen onLoad event being called more than once | bug CruxCore imported Milestone-3.0.1 Priority-Medium | _From [gessedafe@gmail.com](https://code.google.com/u/gessedafe@gmail.com/) on November 09, 2010 12:48:04_
If screen contains any lazy, invisible or collapsed panel, the onLoad event is fired once for each panel.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=221_ | 1.0 | Screen onLoad event being called more than once - _From [gessedafe@gmail.com](https://code.google.com/u/gessedafe@gmail.com/) on November 09, 2010 12:48:04_
If screen contains any lazy, invisible or collapsed panel, the onLoad event is fired once for each panel.
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=221_ | priority | screen onload event being called more than once from on november if screen contains any lazy invisible or collapsed panel the onload event is fired once for each panel original issue | 1 |
778,405 | 27,314,781,568 | IssuesEvent | 2023-02-24 14:55:19 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | opened | [studio-ui] Update the Sidebar navigators | new feature priority: medium triage | ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
The PathNav is not working as well as we had hoped. The TreeNav has some limitations for search and pagination.
### Describe the solution you'd like
Please update the Sidebar widgets as follows:
- TreeNav
- Update pagination to be configurable
- Same as today
- Same as PathNav (with pagination)
- Update the magnifying glass to be always visible, but, it's very faint when not hovered over and fully visible on hover
- PathNav
- Update the magnifying glass to be always visible, but, it's very faint when not hovered over and fully visible on hover
- Don't show ellipsis for paths in the breadcrumb, always have them expanded
- Add a "back" or "up" button to allow the user to navigate up a level
- Allow the user to see the breadcrumb when a filter is enabled
- Update all blueprints to have TreeNav as the default nav for all Sidebar cabinets by default | 1.0 | [studio-ui] Update the Sidebar navigators - ### Duplicates
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe.
The PathNav is not working as well as we had hoped. The TreeNav has some limitations for search and pagination.
### Describe the solution you'd like
Please update the Sidebar widgets as follows:
- TreeNav
- Update pagination to be configurable
- Same as today
- Same as PathNav (with pagination)
- Update the magnifying glass to be always visible, but, it's very faint when not hovered over and fully visible on hover
- PathNav
- Update the magnifying glass to be always visible, but, it's very faint when not hovered over and fully visible on hover
- Don't show ellipsis for paths in the breadcrumb, always have them expanded
- Add a "back" or "up" button to allow the user to navigate up a level
- Allow the user to see the breadcrumb when a filter is enabled
- Update all blueprints to have TreeNav as the default nav for all Sidebar cabinets by default | priority | update the sidebar navigators duplicates i have searched the existing issues is your feature request related to a problem please describe the pathnav is not working as well as we had hoped the treenav has some limitations for search and pagination describe the solution you d like please update the sidebar widgets as follows treenav update pagination to be configurable same as today same as pathnav with pagination update the magnifying glass to be always visible but it s very faint when not hovered over and fully visible on hover pathnav update the magnifying glass to be always visible but it s very faint when not hovered over and fully visible on hover don t show ellipsis for paths in the breadcrumb always have them expanded add a back or up button to allow the user to navigate up a level allow the user to see the breadcrumb when a filter is enabled update all blueprints to have treenav as the default nav for all sidebar cabinets by default | 1 |
758,964 | 26,575,565,271 | IssuesEvent | 2023-01-21 19:21:50 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | opened | Metrics not being updated | Type: Bug Priority: Medium Status: Available | ### Describe the bug
After a first successful run of updating containers, I see the log `"Session done" Failed=0 Scanned=18 Updated=10 notify=no`, and then poll the `/v1/metrics` endpoint, but still see all zeros for the `watchtower_*` metrics.
```
time="2023-01-21T14:05:57-05:00" level=info msg="Session done" Failed=0 Scanned=19 Updated=10 notify=no
```
### Steps to reproduce
1. Force an update run (via http call)
2. Let update run finish
3. See "Session done" with update stats in logs
4. Curl `/v1/metrics` endpoint and see metrics for containers_* not incremented.
### Expected behavior
`watchtower_containers_*` metrics should be incremented based on update run.
### Screenshots
_No response_
### Environment
- Platform: ubuntu-20.04.1
- Architecture: linux/amd64
- Docker Version: 20.10.12
- Watchtower 1.5.1
Compose File:
```
version: "3"
services:
watchtower:
image: containrrr/watchtower
ports:
- 8111:8080
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- /etc/localtime:/etc/localtime:ro
command: --cleanup --schedule "0 0 4 * * *" --http-api-update --http-api-token "watchtower1" --http-api-periodic-polls --http-api-metrics
```
### Your logs
```text
metrics excerpt:
# HELP promhttp_metric_handler_requests_total Total number of scrapes by HTTP status code.
# TYPE promhttp_metric_handler_requests_total counter
promhttp_metric_handler_requests_total{code="200"} 83
promhttp_metric_handler_requests_total{code="500"} 0
promhttp_metric_handler_requests_total{code="503"} 0
# HELP watchtower_containers_failed Number of containers where update failed during the last scan
# TYPE watchtower_containers_failed gauge
watchtower_containers_failed 0
# HELP watchtower_containers_scanned Number of containers scanned for changes by watchtower during the last scan
# TYPE watchtower_containers_scanned gauge
watchtower_containers_scanned 0
# HELP watchtower_containers_updated Number of containers updated by watchtower during the last scan
# TYPE watchtower_containers_updated gauge
watchtower_containers_updated 0
# HELP watchtower_scans_skipped Number of skipped scans since watchtower started
# TYPE watchtower_scans_skipped counter
watchtower_scans_skipped 0
# HELP watchtower_scans_total Number of scans since the watchtower started
# TYPE watchtower_scans_total counter
watchtower_scans_total 0
```
Successful run:
```
watchtower_1 | time="2023-01-21T14:16:29-05:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="Making sure everything is sane before starting"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="Retrieving running containers"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="There are no additional watchtower containers"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Watchtower 1.5.1"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Using no notifications"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Checking all containers (except explicitly disabled with label)"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Scheduling first run: 2023-01-22 04:00:00 -0500 EST"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Note that the first check will be performed in 13 hours, 43 minutes, 29 seconds"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="The HTTP API is enabled at :8080."
watchtower_1 | time="2023-01-21T14:16:31-05:00" level=debug msg="Valid token found."
watchtower_1 | time="2023-01-21T14:16:37-05:00" level=debug msg="Valid token found."
watchtower_1 | time="2023-01-21T14:16:37-05:00" level=info msg="Updates triggered by HTTP API request."
.......................snip...................
level=info msg="Session done" Failed=0 Scanned=18 Updated=0 notify=no
```
```
### Additional context
_No response_ | 1.0 | Metrics not being updated - ### Describe the bug
After a first successful run of updating containers, I see the log `"Session done" Failed=0 Scanned=18 Updated=10 notify=no`, and then poll the `/v1/metrics` endpoint, but still see all zeros for the `watchtower_*` metrics.
```
time="2023-01-21T14:05:57-05:00" level=info msg="Session done" Failed=0 Scanned=19 Updated=10 notify=no
```
### Steps to reproduce
1. Force an update run (via http call)
2. Let update run finish
3. See "Session done" with update stats in logs
4. Curl `/v1/metrics` endpoint and see metrics for containers_* not incremented.
### Expected behavior
`watchtower_containers_*` metrics should be incremented based on update run.
### Screenshots
_No response_
### Environment
- Platform: ubuntu-20.04.1
- Architecture: linux/amd64
- Docker Version: 20.10.12
- Watchtower 1.5.1
Compose File:
```
version: "3"
services:
watchtower:
image: containrrr/watchtower
ports:
- 8111:8080
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- /etc/localtime:/etc/localtime:ro
command: --cleanup --schedule "0 0 4 * * *" --http-api-update --http-api-token "watchtower1" --http-api-periodic-polls --http-api-metrics
```
### Your logs
```text
metrics excerpt:
# HELP promhttp_metric_handler_requests_total Total number of scrapes by HTTP status code.
# TYPE promhttp_metric_handler_requests_total counter
promhttp_metric_handler_requests_total{code="200"} 83
promhttp_metric_handler_requests_total{code="500"} 0
promhttp_metric_handler_requests_total{code="503"} 0
# HELP watchtower_containers_failed Number of containers where update failed during the last scan
# TYPE watchtower_containers_failed gauge
watchtower_containers_failed 0
# HELP watchtower_containers_scanned Number of containers scanned for changes by watchtower during the last scan
# TYPE watchtower_containers_scanned gauge
watchtower_containers_scanned 0
# HELP watchtower_containers_updated Number of containers updated by watchtower during the last scan
# TYPE watchtower_containers_updated gauge
watchtower_containers_updated 0
# HELP watchtower_scans_skipped Number of skipped scans since watchtower started
# TYPE watchtower_scans_skipped counter
watchtower_scans_skipped 0
# HELP watchtower_scans_total Number of scans since the watchtower started
# TYPE watchtower_scans_total counter
watchtower_scans_total 0
```
Successful run:
```
watchtower_1 | time="2023-01-21T14:16:29-05:00" level=debug msg="Sleeping for a second to ensure the docker api client has been properly initialized."
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="Making sure everything is sane before starting"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="Retrieving running containers"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=debug msg="There are no additional watchtower containers"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Watchtower 1.5.1"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Using no notifications"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Checking all containers (except explicitly disabled with label)"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Scheduling first run: 2023-01-22 04:00:00 -0500 EST"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="Note that the first check will be performed in 13 hours, 43 minutes, 29 seconds"
watchtower_1 | time="2023-01-21T14:16:30-05:00" level=info msg="The HTTP API is enabled at :8080."
watchtower_1 | time="2023-01-21T14:16:31-05:00" level=debug msg="Valid token found."
watchtower_1 | time="2023-01-21T14:16:37-05:00" level=debug msg="Valid token found."
watchtower_1 | time="2023-01-21T14:16:37-05:00" level=info msg="Updates triggered by HTTP API request."
.......................snip...................
level=info msg="Session done" Failed=0 Scanned=18 Updated=0 notify=no
```
```
### Additional context
_No response_ | priority | metrics not being updated describe the bug after a first successful run of updating containers i see the log session done failed scanned updated notify no and then poll the metrics endpoint but still see all zeros for the watchtower metrics time level info msg session done failed scanned updated notify no steps to reproduce force an update run via http call let update run finish see session done with update stats in logs curl metrics endpoint and see metrics for containers not incremented expected behavior watchtower containers metrics should be incremented based on update run screenshots no response environment platform ubuntu architecture linux docker version watchtower compose file version services watchtower image containrrr watchtower ports volumes var run docker sock var run docker sock etc localtime etc localtime ro command cleanup schedule http api update http api token http api periodic polls http api metrics your logs text metrics excerpt help promhttp metric handler requests total total number of scrapes by http status code type promhttp metric handler requests total counter promhttp metric handler requests total code promhttp metric handler requests total code promhttp metric handler requests total code help watchtower containers failed number of containers where update failed during the last scan type watchtower containers failed gauge watchtower containers failed help watchtower containers scanned number of containers scanned for changes by watchtower during the last scan type watchtower containers scanned gauge watchtower containers scanned help watchtower containers updated number of containers updated by watchtower during the last scan type watchtower containers updated gauge watchtower containers updated help watchtower scans skipped number of skipped scans since watchtower started type watchtower scans skipped counter watchtower scans skipped help watchtower scans total number of scans since the watchtower started type watchtower scans total counter watchtower scans total successful run watchtower time level debug msg sleeping for a second to ensure the docker api client has been properly initialized watchtower time level debug msg making sure everything is sane before starting watchtower time level debug msg retrieving running containers watchtower time level debug msg there are no additional watchtower containers watchtower time level info msg watchtower watchtower time level info msg using no notifications watchtower time level info msg checking all containers except explicitly disabled with label watchtower time level info msg scheduling first run est watchtower time level info msg note that the first check will be performed in hours minutes seconds watchtower time level info msg the http api is enabled at watchtower time level debug msg valid token found watchtower time level debug msg valid token found watchtower time level info msg updates triggered by http api request snip level info msg session done failed scanned updated notify no additional context no response | 1 |
583,764 | 17,397,987,324 | IssuesEvent | 2021-08-02 15:37:25 | airshipit/airshipctl | https://api.github.com/repos/airshipit/airshipctl | closed | Upgrade Kustomize to the latest version | enhancement priority/medium | **Problem description**
In an effort to keep Airship current with the latest updates of its supporting components, upgrade Kustomize to the latest version (currently v4.2.0)
Kustomize information can be found here:
- https://kustomize.io/
- https://github.com/kubernetes-sigs/kustomize/
- https://github.com/kubernetes-sigs/kustomize/releases
**Proposed change**
- Upgrade Kustomize to v4.2.0
- Validate Kustomize still functional & resolve any breaking changes | 1.0 | Upgrade Kustomize to the latest version - **Problem description**
In an effort to keep Airship current with the latest updates of its supporting components, upgrade Kustomize to the latest version (currently v4.2.0)
Kustomize information can be found here:
- https://kustomize.io/
- https://github.com/kubernetes-sigs/kustomize/
- https://github.com/kubernetes-sigs/kustomize/releases
**Proposed change**
- Upgrade Kustomize to v4.2.0
- Validate Kustomize still functional & resolve any breaking changes | priority | upgrade kustomize to the latest version problem description in an effort to keep airship current with the latest updates of its supporting components upgrade kustomize to the latest version currently kustomize information can be found here proposed change upgrade kustomize to validate kustomize still functional resolve any breaking changes | 1 |
317,034 | 9,660,103,596 | IssuesEvent | 2019-05-20 14:49:03 | comunica/sparqlee | https://api.github.com/repos/comunica/sparqlee | closed | Plain literal behaviour | difficulty:challenging priority:medium spec | ### Context
In the SPARQL spec, more spefically [operand data types](https://www.w3.org/TR/sparql11-query/#operandDataTypes) and [operator mapping] there exist the concepts of:
- plain literals: one of
- simple literals
- language tagged literals
- typed literals with `xsd:string type`
### Problem
The distinction is important, as the string operators only apply to apply to typed literals, and not the plain ones. For example, following expressions are **NOT** equivalent:
`"foo"^^xsd:string != "bar"^^xsd:string` and `"foo" != "bar"`.
The first one will evaluate true, the second one should error, since the operator is not defined. A solution to this is to allow simple literals in problematic operators.
A bigger problem however is that the return type the some functions (mainly functions on strings) is dependent on the original type, e.g. a simple literal should be returned when the (first) argument was a simple literal. This is however impossible.
However: the [RDF.js](https://github.com/rdfjs/representation-task-force/blob/master/interface-spec.md) spec hides the difference between **simple** and **typed** literals, as it automatically adds the `xsd:string` data type.
### Proposal
Concluded from the discussion on the [RDF.js repo](https://github.com/rdfjs/representation-task-force/issues/127), we want to consider simple literals as being typed with `xsd:string`. By extension, the same goes for language tagged literals and `rdf:langString`.
Any untyped literals we get as input we could type ourselves.
This will give issues with spec tests.
| 1.0 | Plain literal behaviour - ### Context
In the SPARQL spec, more spefically [operand data types](https://www.w3.org/TR/sparql11-query/#operandDataTypes) and [operator mapping] there exist the concepts of:
- plain literals: one of
- simple literals
- language tagged literals
- typed literals with `xsd:string type`
### Problem
The distinction is important, as the string operators only apply to apply to typed literals, and not the plain ones. For example, following expressions are **NOT** equivalent:
`"foo"^^xsd:string != "bar"^^xsd:string` and `"foo" != "bar"`.
The first one will evaluate true, the second one should error, since the operator is not defined. A solution to this is to allow simple literals in problematic operators.
A bigger problem however is that the return type the some functions (mainly functions on strings) is dependent on the original type, e.g. a simple literal should be returned when the (first) argument was a simple literal. This is however impossible.
However: the [RDF.js](https://github.com/rdfjs/representation-task-force/blob/master/interface-spec.md) spec hides the difference between **simple** and **typed** literals, as it automatically adds the `xsd:string` data type.
### Proposal
Concluded from the discussion on the [RDF.js repo](https://github.com/rdfjs/representation-task-force/issues/127), we want to consider simple literals as being typed with `xsd:string`. By extension, the same goes for language tagged literals and `rdf:langString`.
Any untyped literals we get as input we could type ourselves.
This will give issues with spec tests.
| priority | plain literal behaviour context in the sparql spec more spefically and there exist the concepts of plain literals one of simple literals language tagged literals typed literals with xsd string type problem the distinction is important as the string operators only apply to apply to typed literals and not the plain ones for example following expressions are not equivalent foo xsd string bar xsd string and foo bar the first one will evaluate true the second one should error since the operator is not defined a solution to this is to allow simple literals in problematic operators a bigger problem however is that the return type the some functions mainly functions on strings is dependent on the original type e g a simple literal should be returned when the first argument was a simple literal this is however impossible however the spec hides the difference between simple and typed literals as it automatically adds the xsd string data type proposal concluded from the discussion on the we want to consider simple literals as being typed with xsd string by extension the same goes for language tagged literals and rdf langstring any untyped literals we get as input we could type ourselves this will give issues with spec tests | 1 |
367,390 | 10,853,183,701 | IssuesEvent | 2019-11-13 14:16:25 | kiwicom/schemathesis | https://api.github.com/repos/kiwicom/schemathesis | closed | Add verbose messages to the CLI output regarding the reasons of failed checks | Priority: Medium Type: Enhancement | For example, it will be nice to see that the response code is 502 for not_server_error / status code conformance checks or which content type exactly was received on the content type conformance check. In a way that it is clear what was the input and what was expected. | 1.0 | Add verbose messages to the CLI output regarding the reasons of failed checks - For example, it will be nice to see that the response code is 502 for not_server_error / status code conformance checks or which content type exactly was received on the content type conformance check. In a way that it is clear what was the input and what was expected. | priority | add verbose messages to the cli output regarding the reasons of failed checks for example it will be nice to see that the response code is for not server error status code conformance checks or which content type exactly was received on the content type conformance check in a way that it is clear what was the input and what was expected | 1 |
292,000 | 8,952,153,523 | IssuesEvent | 2019-01-25 15:49:13 | vanilla-framework/vanilla-framework | https://api.github.com/repos/vanilla-framework/vanilla-framework | closed | The tables of icons donโt include the icons | Priority: Medium Status: Review | _Original comment by @matthewpaulthomas posted [here](https://github.com/canonical-websites/vanillaframework.io/issues/183)_
[The โIconsโ page](https://docs.vanillaframework.io/en/patterns/icons) contains two tables of icons: the first for all other icons, and the second for social icons.
Each table has two columns: the first gives the โCLASS NAMEโ for each icon, and the second โฆ is completely empty. Neither column actually contains the icon. The icons are shown, unlabelled, in a box below. This makes it gratuitously difficult to tell which is which.

What should happen: Each icon is shown before its name in the table. | 1.0 | The tables of icons donโt include the icons - _Original comment by @matthewpaulthomas posted [here](https://github.com/canonical-websites/vanillaframework.io/issues/183)_
[The โIconsโ page](https://docs.vanillaframework.io/en/patterns/icons) contains two tables of icons: the first for all other icons, and the second for social icons.
Each table has two columns: the first gives the โCLASS NAMEโ for each icon, and the second โฆ is completely empty. Neither column actually contains the icon. The icons are shown, unlabelled, in a box below. This makes it gratuitously difficult to tell which is which.

What should happen: Each icon is shown before its name in the table. | priority | the tables of icons donโt include the icons original comment by matthewpaulthomas posted contains two tables of icons the first for all other icons and the second for social icons each table has two columns the first gives the โclass nameโ for each icon and the second โฆ is completely empty neither column actually contains the icon the icons are shown unlabelled in a box below this makes it gratuitously difficult to tell which is which what should happen each icon is shown before its name in the table | 1 |
496,721 | 14,353,400,123 | IssuesEvent | 2020-11-30 06:48:45 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Prevent Duplication Bugs due to internal Exceptions | Category: Tech Priority: Medium | Version: 0.9.1.4 beta release-116
Exception:
```
[02:46:50] System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Components.ElevatorCallPostComponent.OnDestroyElevator()
at Eco.Shared.Utils.ListExtensions.ForEach[T](IList`1 list, Action`1 action)
at Eco.Gameplay.Objects.WorldObject.Destroy()
at Eco.Gameplay.Objects.WorldObject.<>c__DisplayClass111_0.<TryPickUp>b__1()
at Eco.Gameplay.GameActions.GameActionPack.<>c__DisplayClass13_0.<AddPostEffect>b__0()
at Eco.Gameplay.GameActions.GameActionManager.TryPerformActions(User user, List`1 actions, IEnumerable`1 preTests, IEnumerable`1 postEffects)
at Eco.Gameplay.GameActions.GameActionManager.TryPerformActions(User user, GameActionPack actionPack, Boolean notify)
at Eco.Gameplay.GameActions.GameActionManager.TryPerform(GameActionPack pack, Boolean notify)
at Eco.Gameplay.Objects.WorldObject.TryPickUp(Player player, Single caloriesNeeded)
at HammerItem.PickupDialog(WorldObject obj, Player player) in /home/container/Mods/Tools/HammerItem.cs:line 125
```
Now unfortunately I cannot tell what has led to this duplication, only that picking up the elevator with a hammer caused the item to appear in the inventory without being deconstructed in the world, apparently due to the NRE in the ElevatorCallPostComponent.
Apart from this particular issue, it might be worth to catch Exceptions in WorldObject#Destroy and don't add the item to the inventory in this case. The downside could be, that in case of a partial destruction the object is already gone from the world but not awarded to the user, effectively stealing them from the user, which is also not really desirable.
On the other hand if we're at the stage of Exceptions in Destroy we could end up with a half corrupted world state anyway.
Talking about this ElevatorCallPost issue, it was fixed by restarting the server. | 1.0 | Prevent Duplication Bugs due to internal Exceptions - Version: 0.9.1.4 beta release-116
Exception:
```
[02:46:50] System.NullReferenceException: Object reference not set to an instance of an object.
at Eco.Gameplay.Components.ElevatorCallPostComponent.OnDestroyElevator()
at Eco.Shared.Utils.ListExtensions.ForEach[T](IList`1 list, Action`1 action)
at Eco.Gameplay.Objects.WorldObject.Destroy()
at Eco.Gameplay.Objects.WorldObject.<>c__DisplayClass111_0.<TryPickUp>b__1()
at Eco.Gameplay.GameActions.GameActionPack.<>c__DisplayClass13_0.<AddPostEffect>b__0()
at Eco.Gameplay.GameActions.GameActionManager.TryPerformActions(User user, List`1 actions, IEnumerable`1 preTests, IEnumerable`1 postEffects)
at Eco.Gameplay.GameActions.GameActionManager.TryPerformActions(User user, GameActionPack actionPack, Boolean notify)
at Eco.Gameplay.GameActions.GameActionManager.TryPerform(GameActionPack pack, Boolean notify)
at Eco.Gameplay.Objects.WorldObject.TryPickUp(Player player, Single caloriesNeeded)
at HammerItem.PickupDialog(WorldObject obj, Player player) in /home/container/Mods/Tools/HammerItem.cs:line 125
```
Now unfortunately I cannot tell what has led to this duplication, only that picking up the elevator with a hammer caused the item to appear in the inventory without being deconstructed in the world, apparently due to the NRE in the ElevatorCallPostComponent.
Apart from this particular issue, it might be worth to catch Exceptions in WorldObject#Destroy and don't add the item to the inventory in this case. The downside could be, that in case of a partial destruction the object is already gone from the world but not awarded to the user, effectively stealing them from the user, which is also not really desirable.
On the other hand if we're at the stage of Exceptions in Destroy we could end up with a half corrupted world state anyway.
Talking about this ElevatorCallPost issue, it was fixed by restarting the server. | priority | prevent duplication bugs due to internal exceptions version beta release exception system nullreferenceexception object reference not set to an instance of an object at eco gameplay components elevatorcallpostcomponent ondestroyelevator at eco shared utils listextensions foreach ilist list action action at eco gameplay objects worldobject destroy at eco gameplay objects worldobject c b at eco gameplay gameactions gameactionpack c b at eco gameplay gameactions gameactionmanager tryperformactions user user list actions ienumerable pretests ienumerable posteffects at eco gameplay gameactions gameactionmanager tryperformactions user user gameactionpack actionpack boolean notify at eco gameplay gameactions gameactionmanager tryperform gameactionpack pack boolean notify at eco gameplay objects worldobject trypickup player player single caloriesneeded at hammeritem pickupdialog worldobject obj player player in home container mods tools hammeritem cs line now unfortunately i cannot tell what has led to this duplication only that picking up the elevator with a hammer caused the item to appear in the inventory without being deconstructed in the world apparently due to the nre in the elevatorcallpostcomponent apart from this particular issue it might be worth to catch exceptions in worldobject destroy and don t add the item to the inventory in this case the downside could be that in case of a partial destruction the object is already gone from the world but not awarded to the user effectively stealing them from the user which is also not really desirable on the other hand if we re at the stage of exceptions in destroy we could end up with a half corrupted world state anyway talking about this elevatorcallpost issue it was fixed by restarting the server | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.