Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
172,660
| 6,515,025,566
|
IssuesEvent
|
2017-08-26 09:40:13
|
BuckleScript/bucklescript
|
https://api.github.com/repos/BuckleScript/bucklescript
|
closed
|
Userland bettererror handling
|
discussion PRIORITY:HIGH
|
I think I've found a process that lets us iterate on BetterErrors in userland while upstreaming the stable, generalizable better error reporting into BS in the future. Upon compilation error, bsb gets the path of the file and find the dependency (third/first-party)'s customError.ml/re file and invoke it while passing the file name, the error location, message, etc. The script returns the properly explained error which the compile displays (or the default, if none).
- That single file is a nice default and ensures that we don't get too crazy with dragging in 10 deps into. file to analyze error, as to stay fast. Helps us dogfood the future bs stdlib stuff too.
- The third-party is also able to provide much more specific message than the default compiler ones. E.g. ReasonReact has some "variable cannot be generalized", "foo would escape its scope" pitfalls which are intimidating to newcomers but which can be very trivially explained away using ReasonReact-specific explanations.
- This way we get to iterate on a clean version of BetterErrors in userland, including better error report, etc., inside ReasonReact itself. E.g. we could show a visual diff of type. Other dependencies interested in such thing can copy paste our logic. Then upstream if relevant.
Implementation-wise, I think this is feasible with only a few changes to the generated ninja file. Correct me if I'm missing something.
cc @jaredly @sanderspies @glennsl
_Btw, there's been some positive internal changes (that we'll discuss in private) that makes it so that better error reporting is bumped all the way to the top of our priorities_.
|
1.0
|
Userland bettererror handling - I think I've found a process that lets us iterate on BetterErrors in userland while upstreaming the stable, generalizable better error reporting into BS in the future. Upon compilation error, bsb gets the path of the file and find the dependency (third/first-party)'s customError.ml/re file and invoke it while passing the file name, the error location, message, etc. The script returns the properly explained error which the compile displays (or the default, if none).
- That single file is a nice default and ensures that we don't get too crazy with dragging in 10 deps into. file to analyze error, as to stay fast. Helps us dogfood the future bs stdlib stuff too.
- The third-party is also able to provide much more specific message than the default compiler ones. E.g. ReasonReact has some "variable cannot be generalized", "foo would escape its scope" pitfalls which are intimidating to newcomers but which can be very trivially explained away using ReasonReact-specific explanations.
- This way we get to iterate on a clean version of BetterErrors in userland, including better error report, etc., inside ReasonReact itself. E.g. we could show a visual diff of type. Other dependencies interested in such thing can copy paste our logic. Then upstream if relevant.
Implementation-wise, I think this is feasible with only a few changes to the generated ninja file. Correct me if I'm missing something.
cc @jaredly @sanderspies @glennsl
_Btw, there's been some positive internal changes (that we'll discuss in private) that makes it so that better error reporting is bumped all the way to the top of our priorities_.
|
priority
|
userland bettererror handling i think i ve found a process that lets us iterate on bettererrors in userland while upstreaming the stable generalizable better error reporting into bs in the future upon compilation error bsb gets the path of the file and find the dependency third first party s customerror ml re file and invoke it while passing the file name the error location message etc the script returns the properly explained error which the compile displays or the default if none that single file is a nice default and ensures that we don t get too crazy with dragging in deps into file to analyze error as to stay fast helps us dogfood the future bs stdlib stuff too the third party is also able to provide much more specific message than the default compiler ones e g reasonreact has some variable cannot be generalized foo would escape its scope pitfalls which are intimidating to newcomers but which can be very trivially explained away using reasonreact specific explanations this way we get to iterate on a clean version of bettererrors in userland including better error report etc inside reasonreact itself e g we could show a visual diff of type other dependencies interested in such thing can copy paste our logic then upstream if relevant implementation wise i think this is feasible with only a few changes to the generated ninja file correct me if i m missing something cc jaredly sanderspies glennsl btw there s been some positive internal changes that we ll discuss in private that makes it so that better error reporting is bumped all the way to the top of our priorities
| 1
|
491,236
| 14,147,479,140
|
IssuesEvent
|
2020-11-10 20:52:44
|
re-vault/practical-revault
|
https://api.github.com/repos/re-vault/practical-revault
|
reopened
|
Unvault tx fixed feerate of 253perkw isn't a good idea
|
High priority
|
Not crucial security-wise as it's not a revocation tx (to which we attach fee-bumping inputs so it's fine), but if the rolling feerate of nodes' mempools get above that (as it is at the time i'm writing this) then the unvault transaction won't be relayed (feefilter protocol message).
TL;DR: unvault tx fees should not be static (or at least not that low but it does not count as a *solution*)
|
1.0
|
Unvault tx fixed feerate of 253perkw isn't a good idea - Not crucial security-wise as it's not a revocation tx (to which we attach fee-bumping inputs so it's fine), but if the rolling feerate of nodes' mempools get above that (as it is at the time i'm writing this) then the unvault transaction won't be relayed (feefilter protocol message).
TL;DR: unvault tx fees should not be static (or at least not that low but it does not count as a *solution*)
|
priority
|
unvault tx fixed feerate of isn t a good idea not crucial security wise as it s not a revocation tx to which we attach fee bumping inputs so it s fine but if the rolling feerate of nodes mempools get above that as it is at the time i m writing this then the unvault transaction won t be relayed feefilter protocol message tl dr unvault tx fees should not be static or at least not that low but it does not count as a solution
| 1
|
55,807
| 3,074,836,106
|
IssuesEvent
|
2015-08-20 09:53:24
|
agda/agda
|
https://api.github.com/repos/agda/agda
|
closed
|
Caching: Internal error in getConstructorData
|
auto-migrated bug Caching master Priority-Highest
|
```
Recipe to reproduce in emacs:
0. start with empty .agda file
1. restart Agda (clearing the cache)
2. paste the following code
record R : Set1 where
field
f : Set
3. load
4. append the following code
foo = record { f = {!!} }
5. load
An internal error has occurred. Please report this as a bug.
Location of the error: src/full/Agda/TypeChecking/Datatypes.hs:59
```
Original issue reported on code.google.com by `andreas....@gmail.com` on 24 Feb 2015 at 7:35
|
1.0
|
Caching: Internal error in getConstructorData - ```
Recipe to reproduce in emacs:
0. start with empty .agda file
1. restart Agda (clearing the cache)
2. paste the following code
record R : Set1 where
field
f : Set
3. load
4. append the following code
foo = record { f = {!!} }
5. load
An internal error has occurred. Please report this as a bug.
Location of the error: src/full/Agda/TypeChecking/Datatypes.hs:59
```
Original issue reported on code.google.com by `andreas....@gmail.com` on 24 Feb 2015 at 7:35
|
priority
|
caching internal error in getconstructordata recipe to reproduce in emacs start with empty agda file restart agda clearing the cache paste the following code record r where field f set load append the following code foo record f load an internal error has occurred please report this as a bug location of the error src full agda typechecking datatypes hs original issue reported on code google com by andreas gmail com on feb at
| 1
|
258,937
| 8,181,058,202
|
IssuesEvent
|
2018-08-28 21:29:13
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[studio] Updated Global Menu config so that it uses the correct permissions
|
bug priority: high
|
After https://github.com/craftercms/craftercms/issues/2387 and the correct user/group permissions are defined, please update the `global-menu-config.xml` with the right permissions.
|
1.0
|
[studio] Updated Global Menu config so that it uses the correct permissions - After https://github.com/craftercms/craftercms/issues/2387 and the correct user/group permissions are defined, please update the `global-menu-config.xml` with the right permissions.
|
priority
|
updated global menu config so that it uses the correct permissions after and the correct user group permissions are defined please update the global menu config xml with the right permissions
| 1
|
761,549
| 26,685,423,913
|
IssuesEvent
|
2023-01-26 21:28:58
|
nikkistorme/shelf
|
https://api.github.com/repos/nikkistorme/shelf
|
closed
|
Improve shelf sorting
|
high priority 2 hours Shelves
|
# Sort by:
- [x] Page count
- [x] Date added to Library
- [x] Name
- [x] Author
- [x] Finished date (hide this on the Unread shelf)

|
1.0
|
Improve shelf sorting - # Sort by:
- [x] Page count
- [x] Date added to Library
- [x] Name
- [x] Author
- [x] Finished date (hide this on the Unread shelf)

|
priority
|
improve shelf sorting sort by page count date added to library name author finished date hide this on the unread shelf
| 1
|
523,750
| 15,189,045,427
|
IssuesEvent
|
2021-02-15 15:53:28
|
MapColonies/discrete-layer-client
|
https://api.github.com/repos/MapColonies/discrete-layer-client
|
closed
|
Write CSW catalog service API for JSON
|
enhancement priority: high
|
as csw catalog supports XML protocol and we need to send a JSON from the client we need to write NPM package that will support transforming the JSON to csw/xml protocol
|
1.0
|
Write CSW catalog service API for JSON - as csw catalog supports XML protocol and we need to send a JSON from the client we need to write NPM package that will support transforming the JSON to csw/xml protocol
|
priority
|
write csw catalog service api for json as csw catalog supports xml protocol and we need to send a json from the client we need to write npm package that will support transforming the json to csw xml protocol
| 1
|
451,998
| 13,044,641,660
|
IssuesEvent
|
2020-07-29 05:20:05
|
WaifuHarem/waifu-server
|
https://api.github.com/repos/WaifuHarem/waifu-server
|
opened
|
[tests/parsertest.js] Implement parsertest
|
High Priority
|
parsertest.js is currently an empty file. A test suite needs to be created for the module which will test all implemented parser operations with valid and invalid inputs.
|
1.0
|
[tests/parsertest.js] Implement parsertest - parsertest.js is currently an empty file. A test suite needs to be created for the module which will test all implemented parser operations with valid and invalid inputs.
|
priority
|
implement parsertest parsertest js is currently an empty file a test suite needs to be created for the module which will test all implemented parser operations with valid and invalid inputs
| 1
|
543,306
| 15,879,709,393
|
IssuesEvent
|
2021-04-09 12:49:34
|
Adyen/adyen-magento2
|
https://api.github.com/repos/Adyen/adyen-magento2
|
closed
|
[PW-2029] Charged currency is incorrect
|
Bug report Issue: Confirmed Priority: high Progress: in progress
|
**Magento version**: 2.3.1
**Plugin version**: 2.4.0
**Description**
When I configure Magento to display a local currency, this currency is sent to Adyen as the currency the customer has to complete the payment with.
This behaviour is different than other (built in) payment providers in Magento.
Magento itself also displays a notice that the currency that the customer will be charged for will be different:


If I complete this payment with a payment method provided by the Adyen plugin, the customer will be charged for £23.98 instead of €25.94
|
1.0
|
[PW-2029] Charged currency is incorrect - **Magento version**: 2.3.1
**Plugin version**: 2.4.0
**Description**
When I configure Magento to display a local currency, this currency is sent to Adyen as the currency the customer has to complete the payment with.
This behaviour is different than other (built in) payment providers in Magento.
Magento itself also displays a notice that the currency that the customer will be charged for will be different:


If I complete this payment with a payment method provided by the Adyen plugin, the customer will be charged for £23.98 instead of €25.94
|
priority
|
charged currency is incorrect magento version plugin version description when i configure magento to display a local currency this currency is sent to adyen as the currency the customer has to complete the payment with this behaviour is different than other built in payment providers in magento magento itself also displays a notice that the currency that the customer will be charged for will be different if i complete this payment with a payment method provided by the adyen plugin the customer will be charged for £ instead of €
| 1
|
660,221
| 21,957,294,584
|
IssuesEvent
|
2022-05-24 13:10:57
|
ut-issl/s2e-core
|
https://api.github.com/repos/ut-issl/s2e-core
|
closed
|
Improve updating time in SimTime
|
priority::high
|
## Overview
Improve implementation of updating simulation time to not skip time for realtime simulation.
## Details
Relevant code: [SimTime.cpp](https://github.com/ut-issl/s2e-core/blob/13d443f2806d983e2288290add1e5c8d0f0a9d56/src/Environment/Global/SimTime.cpp#L85)
Relevant discussion: https://arkedge-space.slack.com/archives/C02JRCY0A5V/p1646816850264289
### Cause
When running realtime (or faster, step_sec >= 1) simulation, normally your computer is fast enough.
But sometimes it needs more calculation than normal steps, then computing time may exceeds realtime.
In that case S2E assesses your computer performance is not enough and forcibly skips time.
### Problem
- S2E's time became faster and faster than C2A's time because [this code](https://github.com/ut-issl/s2e-core/blob/13d443f2806d983e2288290add1e5c8d0f0a9d56/src/Environment/Global/SimTime.cpp#L92) skips time, and we cannot properly compare S2E's logged value and C2A's telemetry value.
- Error message "Error: the specified step_sec is too small for this computer." is continuously displayed dispite your computer is fast enough.
### Possible improvement
Use average wait time over a short term (one second or something) instead of single step to check the computer is able to compute simulation steps in realtime.
## Conditions for close
When the implementation is complete or another solution is found.
## Supplement
NA
## Note
NA
|
1.0
|
Improve updating time in SimTime - ## Overview
Improve implementation of updating simulation time to not skip time for realtime simulation.
## Details
Relevant code: [SimTime.cpp](https://github.com/ut-issl/s2e-core/blob/13d443f2806d983e2288290add1e5c8d0f0a9d56/src/Environment/Global/SimTime.cpp#L85)
Relevant discussion: https://arkedge-space.slack.com/archives/C02JRCY0A5V/p1646816850264289
### Cause
When running realtime (or faster, step_sec >= 1) simulation, normally your computer is fast enough.
But sometimes it needs more calculation than normal steps, then computing time may exceeds realtime.
In that case S2E assesses your computer performance is not enough and forcibly skips time.
### Problem
- S2E's time became faster and faster than C2A's time because [this code](https://github.com/ut-issl/s2e-core/blob/13d443f2806d983e2288290add1e5c8d0f0a9d56/src/Environment/Global/SimTime.cpp#L92) skips time, and we cannot properly compare S2E's logged value and C2A's telemetry value.
- Error message "Error: the specified step_sec is too small for this computer." is continuously displayed dispite your computer is fast enough.
### Possible improvement
Use average wait time over a short term (one second or something) instead of single step to check the computer is able to compute simulation steps in realtime.
## Conditions for close
When the implementation is complete or another solution is found.
## Supplement
NA
## Note
NA
|
priority
|
improve updating time in simtime overview improve implementation of updating simulation time to not skip time for realtime simulation details relevant code relevant discussion cause when running realtime or faster step sec simulation normally your computer is fast enough but sometimes it needs more calculation than normal steps then computing time may exceeds realtime in that case assesses your computer performance is not enough and forcibly skips time problem s time became faster and faster than s time because skips time and we cannot properly compare s logged value and s telemetry value error message error the specified step sec is too small for this computer is continuously displayed dispite your computer is fast enough possible improvement use average wait time over a short term one second or something instead of single step to check the computer is able to compute simulation steps in realtime conditions for close when the implementation is complete or another solution is found supplement na note na
| 1
|
98,465
| 4,021,731,188
|
IssuesEvent
|
2016-05-16 23:18:06
|
shelljs/shelljs
|
https://api.github.com/repos/shelljs/shelljs
|
opened
|
Big picture shelljs@1.0.0 discussion
|
chore feat help wanted high priority meeting question
|
Hi Everyone,
As discussed in #443, I'd like to start discussing `shelljs@1.0.0` features on a high level.
`v1.0.0` features [@shelljs/contributors, feel free to edit]:
- [ ] Async versions of all commands (#402, #387)
- [ ] Plugin System (#391)
- [ ] It would be quite nice to impliment all built in commands as plugins.
- [ ] Code Style Guidelines (#317)
- [ ] Miscellaneous cleanup/refactor of everything.
Thoughts?
|
1.0
|
Big picture shelljs@1.0.0 discussion - Hi Everyone,
As discussed in #443, I'd like to start discussing `shelljs@1.0.0` features on a high level.
`v1.0.0` features [@shelljs/contributors, feel free to edit]:
- [ ] Async versions of all commands (#402, #387)
- [ ] Plugin System (#391)
- [ ] It would be quite nice to impliment all built in commands as plugins.
- [ ] Code Style Guidelines (#317)
- [ ] Miscellaneous cleanup/refactor of everything.
Thoughts?
|
priority
|
big picture shelljs discussion hi everyone as discussed in i d like to start discussing shelljs features on a high level features async versions of all commands plugin system it would be quite nice to impliment all built in commands as plugins code style guidelines miscellaneous cleanup refactor of everything thoughts
| 1
|
814,072
| 30,485,052,458
|
IssuesEvent
|
2023-07-18 00:57:36
|
calcom/cal.com
|
https://api.github.com/repos/calcom/cal.com
|
closed
|
[CAL-2170] Wrong slot starting times for timezones that are not full hours
|
🐛 bug High priority
|
The slot starting times aren't correct for timezones like Asia/Kolkata (+5:30). Hourly events would start at half-hour slots instead of full-hour slots.
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-2170](https://linear.app/calcom/issue/CAL-2170/wrong-slot-starting-times-for-timezones-that-are-not-a-full-hour)</sub>
|
1.0
|
[CAL-2170] Wrong slot starting times for timezones that are not full hours - The slot starting times aren't correct for timezones like Asia/Kolkata (+5:30). Hourly events would start at half-hour slots instead of full-hour slots.
<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-2170](https://linear.app/calcom/issue/CAL-2170/wrong-slot-starting-times-for-timezones-that-are-not-a-full-hour)</sub>
|
priority
|
wrong slot starting times for timezones that are not full hours the slot starting times aren t correct for timezones like asia kolkata hourly events would start at half hour slots instead of full hour slots from
| 1
|
284,178
| 8,736,331,295
|
IssuesEvent
|
2018-12-11 19:14:36
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
Huge memory leaks in GUI adding, deleting, hiding, showing plots
|
bug likelihood medium priority reviewed severity high
|
Allen Sanderson reported a huge memory leak in the VisIt GUI when he was adding, deleting, hiding and showing plots. He had about 400 variables in his file and the memory jumped by about 100Mb each time such an operation was performed. This could cause the GUI to quickly exceed several Gb.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 3108
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Immediate
Subject: Huge memory leaks in GUI adding, deleting, hiding, showing plots
Assigned to: Mark Miller
Category:
Target version: 2.13.2
Author: Eric Brugger
Start: 05/16/2018
Due date:
% Done: 0
Estimated time: 6.0
Created: 05/16/2018 12:51 pm
Updated: 05/22/2018 07:02 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.13.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Allen Sanderson reported a huge memory leak in the VisIt GUI when he was adding, deleting, hiding and showing plots. He had about 400 variables in his file and the memory jumped by about 100Mb each time such an operation was performed. This could cause the GUI to quickly exceed several Gb.
Comments:
|
1.0
|
Huge memory leaks in GUI adding, deleting, hiding, showing plots - Allen Sanderson reported a huge memory leak in the VisIt GUI when he was adding, deleting, hiding and showing plots. He had about 400 variables in his file and the memory jumped by about 100Mb each time such an operation was performed. This could cause the GUI to quickly exceed several Gb.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 3108
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Immediate
Subject: Huge memory leaks in GUI adding, deleting, hiding, showing plots
Assigned to: Mark Miller
Category:
Target version: 2.13.2
Author: Eric Brugger
Start: 05/16/2018
Due date:
% Done: 0
Estimated time: 6.0
Created: 05/16/2018 12:51 pm
Updated: 05/22/2018 07:02 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.13.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Allen Sanderson reported a huge memory leak in the VisIt GUI when he was adding, deleting, hiding and showing plots. He had about 400 variables in his file and the memory jumped by about 100Mb each time such an operation was performed. This could cause the GUI to quickly exceed several Gb.
Comments:
|
priority
|
huge memory leaks in gui adding deleting hiding showing plots allen sanderson reported a huge memory leak in the visit gui when he was adding deleting hiding and showing plots he had about variables in his file and the memory jumped by about each time such an operation was performed this could cause the gui to quickly exceed several gb redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority immediate subject huge memory leaks in gui adding deleting hiding showing plots assigned to mark miller category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version impact expected use os all support group any description allen sanderson reported a huge memory leak in the visit gui when he was adding deleting hiding and showing plots he had about variables in his file and the memory jumped by about each time such an operation was performed this could cause the gui to quickly exceed several gb comments
| 1
|
272,777
| 8,517,643,032
|
IssuesEvent
|
2018-11-01 08:50:24
|
egoist/bili
|
https://api.github.com/repos/egoist/bili
|
closed
|
Not working when using top-level node_modules (like Yarn Workspaces)
|
bug priority: high
|
Hi,
Bili 3.1.2 complains about `× Cannot find plugin "rollup-plugin-typescript2" in current directory!`, but I do have it in my `package.json`. Indeed, I see [the `require()` is done relative to the current dir](https://github.com/egoist/bili/blob/bafeef9e26a192f2c887067e4dfd16c5cbfdc50b/src/index.js#L647).
In my case, I have the following tree:
```
node_modules/
packages/
--a
----package.json
--b
----package.json
...
```
All dependencies are installed by in the parent `node_modules/` by [Yarn Workspaces](https://yarnpkg.com/lang/en/docs/workspaces/), which is perfectly valid:
> Node will look for your modules in special folders named node_modules. A node_modules folder can be on the same level as the current file, or higher up in the directory chain. Node will walk up the directory chain, looking through each node_modules until it finds the module you tried to load.
|
1.0
|
Not working when using top-level node_modules (like Yarn Workspaces) - Hi,
Bili 3.1.2 complains about `× Cannot find plugin "rollup-plugin-typescript2" in current directory!`, but I do have it in my `package.json`. Indeed, I see [the `require()` is done relative to the current dir](https://github.com/egoist/bili/blob/bafeef9e26a192f2c887067e4dfd16c5cbfdc50b/src/index.js#L647).
In my case, I have the following tree:
```
node_modules/
packages/
--a
----package.json
--b
----package.json
...
```
All dependencies are installed by in the parent `node_modules/` by [Yarn Workspaces](https://yarnpkg.com/lang/en/docs/workspaces/), which is perfectly valid:
> Node will look for your modules in special folders named node_modules. A node_modules folder can be on the same level as the current file, or higher up in the directory chain. Node will walk up the directory chain, looking through each node_modules until it finds the module you tried to load.
|
priority
|
not working when using top level node modules like yarn workspaces hi bili complains about × cannot find plugin rollup plugin in current directory but i do have it in my package json indeed i see in my case i have the following tree node modules packages a package json b package json all dependencies are installed by in the parent node modules by which is perfectly valid node will look for your modules in special folders named node modules a node modules folder can be on the same level as the current file or higher up in the directory chain node will walk up the directory chain looking through each node modules until it finds the module you tried to load
| 1
|
791,229
| 27,856,527,629
|
IssuesEvent
|
2023-03-20 23:41:56
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
opened
|
other identifier search on numeric is wonky
|
Priority-High (Needed for work) Function-SearchOrDownload Bug
|
See the search and results below--
search is for identifier = BSI + numeric= 184
results includes BSI values that include the last 2 digits? even with the ```=``` qualifier
So what is up?

|
1.0
|
other identifier search on numeric is wonky - See the search and results below--
search is for identifier = BSI + numeric= 184
results includes BSI values that include the last 2 digits? even with the ```=``` qualifier
So what is up?

|
priority
|
other identifier search on numeric is wonky see the search and results below search is for identifier bsi numeric results includes bsi values that include the last digits even with the qualifier so what is up
| 1
|
382,600
| 11,308,876,281
|
IssuesEvent
|
2020-01-19 09:15:46
|
troop-370/troop370
|
https://api.github.com/repos/troop-370/troop370
|
closed
|
New River Trek date has the wrong month
|
priority: high update
|
From the committee chair:
> Please correct dates from June 4-11 to July 4-11
|
1.0
|
New River Trek date has the wrong month - From the committee chair:
> Please correct dates from June 4-11 to July 4-11
|
priority
|
new river trek date has the wrong month from the committee chair please correct dates from june to july
| 1
|
363,783
| 10,755,115,746
|
IssuesEvent
|
2019-10-31 08:23:53
|
onaio/rdt-standard
|
https://api.github.com/repos/onaio/rdt-standard
|
closed
|
Update Bahasa Translation Hamburger Menu and Name
|
Android Client Priority - high bug
|
Please see translations for hamburger menu below:
<string name="drawer_menu_item_sync">Sinkronisasi</string>
<string name="drawer_menu_item_logout">Keluar</string>
<string name="drawer_menu_item_image_sync">Sinkronisasi gambar</string>
<string name="search_hint">Cari</string>
</resources>
|
1.0
|
Update Bahasa Translation Hamburger Menu and Name - Please see translations for hamburger menu below:
<string name="drawer_menu_item_sync">Sinkronisasi</string>
<string name="drawer_menu_item_logout">Keluar</string>
<string name="drawer_menu_item_image_sync">Sinkronisasi gambar</string>
<string name="search_hint">Cari</string>
</resources>
|
priority
|
update bahasa translation hamburger menu and name please see translations for hamburger menu below sinkronisasi keluar sinkronisasi gambar cari
| 1
|
635,754
| 20,508,166,138
|
IssuesEvent
|
2022-03-01 01:34:32
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
Data Entry - make fields in Identification and Identification Extras identical
|
Priority-High (Needed for work) Enhancement Display/Interface
|
Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Is your feature request related to a problem? Please describe.**
The fields in the Identification section are not the same in the Identification Extra section.
<img width="804" alt="Screen Shot 2022-02-20 at 11 54 59 AM" src="https://user-images.githubusercontent.com/15368365/154859338-9a9ca7ff-be23-4aeb-a66e-eeb5c7c6953f.png">
<img width="1038" alt="Screen Shot 2022-02-20 at 12 01 00 PM" src="https://user-images.githubusercontent.com/15368365/154859544-366ff67f-4434-4dcc-a130-706e95c208d7.png">
**Describe what you're trying to accomplish**
1. We need to be able to "build" taxonomic names in the Identification Extra section.
2. We need to be able to leave the date field blank as in the primary identification.
**Describe the solution you'd like**
Duplicate the Identification fields in the Identification Extras section.
**Describe alternatives you've considered**
We have to correct the record after it is uploaded which is a wonderful opportunity create errors and omissions.
**Additional context**
It would also be nice if the fields had the same names in each section and were in the same order.
Agent 1 vs. Identifying Agent
NatureofID vs. Nature of ID
ID Date vs. MadeDate
ID Confidence vs. Confidence
Remarks vs. ID Remarks
Volunteers who database for a few hours a week really need UI consistency.
**Priority**
We're running into problems daily, so it's high priority for us.
|
1.0
|
Data Entry - make fields in Identification and Identification Extras identical - Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Is your feature request related to a problem? Please describe.**
The fields in the Identification section are not the same in the Identification Extra section.
<img width="804" alt="Screen Shot 2022-02-20 at 11 54 59 AM" src="https://user-images.githubusercontent.com/15368365/154859338-9a9ca7ff-be23-4aeb-a66e-eeb5c7c6953f.png">
<img width="1038" alt="Screen Shot 2022-02-20 at 12 01 00 PM" src="https://user-images.githubusercontent.com/15368365/154859544-366ff67f-4434-4dcc-a130-706e95c208d7.png">
**Describe what you're trying to accomplish**
1. We need to be able to "build" taxonomic names in the Identification Extra section.
2. We need to be able to leave the date field blank as in the primary identification.
**Describe the solution you'd like**
Duplicate the Identification fields in the Identification Extras section.
**Describe alternatives you've considered**
We have to correct the record after it is uploaded which is a wonderful opportunity create errors and omissions.
**Additional context**
It would also be nice if the fields had the same names in each section and were in the same order.
Agent 1 vs. Identifying Agent
NatureofID vs. Nature of ID
ID Date vs. MadeDate
ID Confidence vs. Confidence
Remarks vs. ID Remarks
Volunteers who database for a few hours a week really need UI consistency.
**Priority**
We're running into problems daily, so it's high priority for us.
|
priority
|
data entry make fields in identification and identification extras identical issue documentation is is your feature request related to a problem please describe the fields in the identification section are not the same in the identification extra section img width alt screen shot at am src img width alt screen shot at pm src describe what you re trying to accomplish we need to be able to build taxonomic names in the identification extra section we need to be able to leave the date field blank as in the primary identification describe the solution you d like duplicate the identification fields in the identification extras section describe alternatives you ve considered we have to correct the record after it is uploaded which is a wonderful opportunity create errors and omissions additional context it would also be nice if the fields had the same names in each section and were in the same order agent vs identifying agent natureofid vs nature of id id date vs madedate id confidence vs confidence remarks vs id remarks volunteers who database for a few hours a week really need ui consistency priority we re running into problems daily so it s high priority for us
| 1
|
804,559
| 29,492,948,892
|
IssuesEvent
|
2023-06-02 14:43:04
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
opened
|
[YCQL] Add some of tablet, table, partition hash, and tserver UUID to the slow query log
|
kind/new-feature area/docdb priority/high area/ycql status/awaiting-triage jira-originated
|
Jira Link: [DB-6781](https://yugabyte.atlassian.net/browse/DB-6781)
|
1.0
|
[YCQL] Add some of tablet, table, partition hash, and tserver UUID to the slow query log - Jira Link: [DB-6781](https://yugabyte.atlassian.net/browse/DB-6781)
|
priority
|
add some of tablet table partition hash and tserver uuid to the slow query log jira link
| 1
|
444,599
| 12,814,866,806
|
IssuesEvent
|
2020-07-04 21:36:41
|
GeyserMC/Geyser
|
https://api.github.com/repos/GeyserMC/Geyser
|
closed
|
Memory leak w/ ping passthrough
|
Confirmed Bug Priority: High
|
**Describe the bug**
<!--- A clear and concise description of what the bug is. -->
Enabling any feature that enables ping passthrough will cause the proxy to memory leak.
**To Reproduce**
Enable `passthrough-motd` or `passthrough-player-counts` in the config.
**Geyser Version**
Standalone git-master-e81c6f2
**Additional Context**
Heap dump: https://www.mediafire.com/file/bqyuz2sum1mpn2u/java_pid30423.hprof/file
|
1.0
|
Memory leak w/ ping passthrough - **Describe the bug**
<!--- A clear and concise description of what the bug is. -->
Enabling any feature that enables ping passthrough will cause the proxy to memory leak.
**To Reproduce**
Enable `passthrough-motd` or `passthrough-player-counts` in the config.
**Geyser Version**
Standalone git-master-e81c6f2
**Additional Context**
Heap dump: https://www.mediafire.com/file/bqyuz2sum1mpn2u/java_pid30423.hprof/file
|
priority
|
memory leak w ping passthrough describe the bug enabling any feature that enables ping passthrough will cause the proxy to memory leak to reproduce enable passthrough motd or passthrough player counts in the config geyser version standalone git master additional context heap dump
| 1
|
72,332
| 3,381,326,705
|
IssuesEvent
|
2015-11-26 01:52:43
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
closed
|
harmonize flags
|
priority: high
|
Please refer to: https://github.com/neuropoly/spinalcordtoolbox/wiki/flags
Important: if you make change to a flag, please use deprecated mode and indicate in this thread.
If not possible to use deprecated, please list potential conflicts for each file.
If no parser in the script, add it.
**Sara**
- [x] sct_apply_transfo
- "-c" changed to "-crop"
- [x] sct_average_data_within_mask
- added parser
- "-t" was used to number of volume to use if 4D. Changed it to "-nvol".
- [x] sct_check_atlas_integrity
- added parser
- "-m" was used for GM segmentation. Changed it to "-gm"
- "-t" was use for atlas threshold. Changed it to "-thr"
- "-g" was use for GM threshold. Changed it to "-thrgm" (to match with "-thr")
- [x] sct_check_dependences
- added parser
- "-l" was used for log file option. Changed it to "-log"
- [x] sct_compute_ernst_angle
- added get_parser() method.
- "-d" was used for display option. Changed it to "-v"
- [x] sct_compute_hausdorff_distance
- added get_parser() method.
- "-t" was used for thinning. Changed it to "-thinning"
- "-r" was used for second input (reference). Changed it to "-d"
- [x] sct_compute_mtr
- added parser
- changed "-i" and "-j" to "-mt0" and "mt1" to be more explicit.
- [x] sct_concat_transfo
- added parser
- [x] sct_convert (already up to date with flags convention)
- [x] sct_segment_graymatter
- "-l" was for vertebral level file. Changed to "-vert"
- [x] sct_register_to_template
- '-p' replaced by '-param'
- [x] sct_resample
- added get_parser() function
- [x] sct_smooth_spinalcord
- added parser
- changed flag '-s' to '-smooth' **WARNING: -s is not deprecated to -smooth**
- changed "-c" (centerline or seg) to "-s"
- [x] sct_straighten_spinalcord
- changed flag '-p' to '-pad'
- changed flag '-params' to '-param'
- changed flag "-c" to "-s"
- [x] sct_testing
- added parser
- [x] sct_warp_template
- changed flag -o to -ofolder
**Simon**
- [x] sct_create_mask
- added parser
- "-m" was for method. Changed to "-p"
- "-s" was for size, changed for "-size"
- [x] sct_crop_image
- added get_parser function.
- [x] sct_dice_coefficient
- did a warper for the sct_dice_coefficient binary (now isct_dice...)
- input files now have flags -i and -d **WARNING: you will need to change this in your scripts, there was no flags before so there is no deprecated option ...**
- [x] sct_dmri_compute_dti (already up to date with flags convention)
- [x] sct_dmri_concat_bvals (already up to date with flags convention)
- [x] sct_dmri_concat_bvecs (already up to date with flags convention)
- [x] sct_dmri_get_bvalue
- added parser
- [x] sct_dmri_moco
- added parser
- '-b' was for bvecs file, changed to '-bvec'
- '-p' was for param, changed to '-param'
- '-t' was for an otsu threshold, changed to '-thr'
- '-o' was for output folder, changed to '-ofolder'
**Benjamin**
- [x] sct_dmri_separate_b0_and_dwi
- "-m" was used for bvals. Changed it to "-bval".
- "-b" was used for bvecs. Changed it to "-bvec".
- "-o" was used for output folder. Changed it to "-ofolder".
- [x] sct_dmri_transpose_bvecs
- "-i" was used for bvec file. Changed it to "-bvec". "-i" is now deprecated.
- [x] sct_download_data
- [x] sct_extract_metric
- No parser.
- '-p' changed to '-param'
- [x] sct_flatten_sagittal
- No parser.
- "-c" changed to "-s"
- "-s" changed to "-x" **WARNING: -s is not depredated to -x**
- [x] sct_fmri_compute_tsnr
- added get_parser function.
- [x] sct_fmri_moco
- No parser.
- '-p' changed to '-param'
- [x] sct_get_centerline
- "-t" option was used for contrast. It is now deprecated by "-c".
- "-p" option was used for a file containing a point. It is now deprecated by "-point".
**Olivier**
- [x] sct_image
- [x] sct_label_utils
- "-r" was used to reference an image. Now deprecated by "-ref"
- "-level" was used for vertebral level. Now deprecated by "-vert"
- "-t" --> changed for -p (julien)
- [x] sct_label_vertebrae
- "-t" was used for processes, now using -p
- "-seg" for segmentation file is now "-s"
- [x] sct_maths
- [x] sct_process_segmentation
- "-l" switched to "-vert" for vertebral levels
- "-s" replaced by "-size"
- No Parser
- [x] sct_propseg
- "-t" option is for contrast
- [x] sct_register_multimodal
- "-p" was used for params, now -param
|
1.0
|
harmonize flags - Please refer to: https://github.com/neuropoly/spinalcordtoolbox/wiki/flags
Important: if you make change to a flag, please use deprecated mode and indicate in this thread.
If not possible to use deprecated, please list potential conflicts for each file.
If no parser in the script, add it.
**Sara**
- [x] sct_apply_transfo
- "-c" changed to "-crop"
- [x] sct_average_data_within_mask
- added parser
- "-t" was used to number of volume to use if 4D. Changed it to "-nvol".
- [x] sct_check_atlas_integrity
- added parser
- "-m" was used for GM segmentation. Changed it to "-gm"
- "-t" was use for atlas threshold. Changed it to "-thr"
- "-g" was use for GM threshold. Changed it to "-thrgm" (to match with "-thr")
- [x] sct_check_dependences
- added parser
- "-l" was used for log file option. Changed it to "-log"
- [x] sct_compute_ernst_angle
- added get_parser() method.
- "-d" was used for display option. Changed it to "-v"
- [x] sct_compute_hausdorff_distance
- added get_parser() method.
- "-t" was used for thinning. Changed it to "-thinning"
- "-r" was used for second input (reference). Changed it to "-d"
- [x] sct_compute_mtr
- added parser
- changed "-i" and "-j" to "-mt0" and "mt1" to be more explicit.
- [x] sct_concat_transfo
- added parser
- [x] sct_convert (already up to date with flags convention)
- [x] sct_segment_graymatter
- "-l" was for vertebral level file. Changed to "-vert"
- [x] sct_register_to_template
- '-p' replaced by '-param'
- [x] sct_resample
- added get_parser() function
- [x] sct_smooth_spinalcord
- added parser
- changed flag '-s' to '-smooth' **WARNING: -s is not deprecated to -smooth**
- changed "-c" (centerline or seg) to "-s"
- [x] sct_straighten_spinalcord
- changed flag '-p' to '-pad'
- changed flag '-params' to '-param'
- changed flag "-c" to "-s"
- [x] sct_testing
- added parser
- [x] sct_warp_template
- changed flag -o to -ofolder
**Simon**
- [x] sct_create_mask
- added parser
- "-m" was for method. Changed to "-p"
- "-s" was for size, changed for "-size"
- [x] sct_crop_image
- added get_parser function.
- [x] sct_dice_coefficient
- did a warper for the sct_dice_coefficient binary (now isct_dice...)
- input files now have flags -i and -d **WARNING: you will need to change this in your scripts, there was no flags before so there is no deprecated option ...**
- [x] sct_dmri_compute_dti (already up to date with flags convention)
- [x] sct_dmri_concat_bvals (already up to date with flags convention)
- [x] sct_dmri_concat_bvecs (already up to date with flags convention)
- [x] sct_dmri_get_bvalue
- added parser
- [x] sct_dmri_moco
- added parser
- '-b' was for bvecs file, changed to '-bvec'
- '-p' was for param, changed to '-param'
- '-t' was for an otsu threshold, changed to '-thr'
- '-o' was for output folder, changed to '-ofolder'
**Benjamin**
- [x] sct_dmri_separate_b0_and_dwi
- "-m" was used for bvals. Changed it to "-bval".
- "-b" was used for bvecs. Changed it to "-bvec".
- "-o" was used for output folder. Changed it to "-ofolder".
- [x] sct_dmri_transpose_bvecs
- "-i" was used for bvec file. Changed it to "-bvec". "-i" is now deprecated.
- [x] sct_download_data
- [x] sct_extract_metric
- No parser.
- '-p' changed to '-param'
- [x] sct_flatten_sagittal
- No parser.
- "-c" changed to "-s"
- "-s" changed to "-x" **WARNING: -s is not depredated to -x**
- [x] sct_fmri_compute_tsnr
- added get_parser function.
- [x] sct_fmri_moco
- No parser.
- '-p' changed to '-param'
- [x] sct_get_centerline
- "-t" option was used for contrast. It is now deprecated by "-c".
- "-p" option was used for a file containing a point. It is now deprecated by "-point".
**Olivier**
- [x] sct_image
- [x] sct_label_utils
- "-r" was used to reference an image. Now deprecated by "-ref"
- "-level" was used for vertebral level. Now deprecated by "-vert"
- "-t" --> changed for -p (julien)
- [x] sct_label_vertebrae
- "-t" was used for processes, now using -p
- "-seg" for segmentation file is now "-s"
- [x] sct_maths
- [x] sct_process_segmentation
- "-l" switched to "-vert" for vertebral levels
- "-s" replaced by "-size"
- No Parser
- [x] sct_propseg
- "-t" option is for contrast
- [x] sct_register_multimodal
- "-p" was used for params, now -param
|
priority
|
harmonize flags please refer to important if you make change to a flag please use deprecated mode and indicate in this thread if not possible to use deprecated please list potential conflicts for each file if no parser in the script add it sara sct apply transfo c changed to crop sct average data within mask added parser t was used to number of volume to use if changed it to nvol sct check atlas integrity added parser m was used for gm segmentation changed it to gm t was use for atlas threshold changed it to thr g was use for gm threshold changed it to thrgm to match with thr sct check dependences added parser l was used for log file option changed it to log sct compute ernst angle added get parser method d was used for display option changed it to v sct compute hausdorff distance added get parser method t was used for thinning changed it to thinning r was used for second input reference changed it to d sct compute mtr added parser changed i and j to and to be more explicit sct concat transfo added parser sct convert already up to date with flags convention sct segment graymatter l was for vertebral level file changed to vert sct register to template p replaced by param sct resample added get parser function sct smooth spinalcord added parser changed flag s to smooth warning s is not deprecated to smooth changed c centerline or seg to s sct straighten spinalcord changed flag p to pad changed flag params to param changed flag c to s sct testing added parser sct warp template changed flag o to ofolder simon sct create mask added parser m was for method changed to p s was for size changed for size sct crop image added get parser function sct dice coefficient did a warper for the sct dice coefficient binary now isct dice input files now have flags i and d warning you will need to change this in your scripts there was no flags before so there is no deprecated option sct dmri compute dti already up to date with flags convention sct dmri concat bvals already up to date with flags convention sct dmri concat bvecs already up to date with flags convention sct dmri get bvalue added parser sct dmri moco added parser b was for bvecs file changed to bvec p was for param changed to param t was for an otsu threshold changed to thr o was for output folder changed to ofolder benjamin sct dmri separate and dwi m was used for bvals changed it to bval b was used for bvecs changed it to bvec o was used for output folder changed it to ofolder sct dmri transpose bvecs i was used for bvec file changed it to bvec i is now deprecated sct download data sct extract metric no parser p changed to param sct flatten sagittal no parser c changed to s s changed to x warning s is not depredated to x sct fmri compute tsnr added get parser function sct fmri moco no parser p changed to param sct get centerline t option was used for contrast it is now deprecated by c p option was used for a file containing a point it is now deprecated by point olivier sct image sct label utils r was used to reference an image now deprecated by ref level was used for vertebral level now deprecated by vert t changed for p julien sct label vertebrae t was used for processes now using p seg for segmentation file is now s sct maths sct process segmentation l switched to vert for vertebral levels s replaced by size no parser sct propseg t option is for contrast sct register multimodal p was used for params now param
| 1
|
653,369
| 21,580,637,830
|
IssuesEvent
|
2022-05-02 18:19:22
|
WordPress/openverse-frontend
|
https://api.github.com/repos/WordPress/openverse-frontend
|
opened
|
Cannot close global audio player after clicking on the waveform
|
🟧 priority: high 🚦 status: awaiting triage 🛠 goal: fix 🕹 aspect: interface
|
## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
Global audio player close button becomes not interactive after you've clicked on the waveform.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. <!-- Step 1 ... -->Open the global audio player.
2. <!-- Step 2 ... -->Click on the waveform.
3. <!-- Step 3 ... -->Try closing the global audio player by clicking on the x button. It doesn't work.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
## Environment
<!-- Please complete this, unless you are certain the problem is not environment specific. -->
- Device: <!-- (_eg._ iPhone Xs; laptop) -->
- OS: <!-- (_eg._ iOS 13.5; Fedora 32) -->
- Browser: <!-- (_eg._ Safari; Firefox) -->
- Version: <!-- (_eg._ 13; 73) -->
- Other info: <!-- (_eg._ display resolution, ease-of-access settings) -->
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in resolving this bug.
|
1.0
|
Cannot close global audio player after clicking on the waveform - ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
Global audio player close button becomes not interactive after you've clicked on the waveform.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. <!-- Step 1 ... -->Open the global audio player.
2. <!-- Step 2 ... -->Click on the waveform.
3. <!-- Step 3 ... -->Try closing the global audio player by clicking on the x button. It doesn't work.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
## Environment
<!-- Please complete this, unless you are certain the problem is not environment specific. -->
- Device: <!-- (_eg._ iPhone Xs; laptop) -->
- OS: <!-- (_eg._ iOS 13.5; Fedora 32) -->
- Browser: <!-- (_eg._ Safari; Firefox) -->
- Version: <!-- (_eg._ 13; 73) -->
- Other info: <!-- (_eg._ display resolution, ease-of-access settings) -->
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in resolving this bug.
|
priority
|
cannot close global audio player after clicking on the waveform description global audio player close button becomes not interactive after you ve clicked on the waveform reproduction open the global audio player click on the waveform try closing the global audio player by clicking on the x button it doesn t work screenshots environment device os browser version other info additional context resolution 🙋 i would be interested in resolving this bug
| 1
|
634,122
| 20,326,542,082
|
IssuesEvent
|
2022-02-18 06:27:08
|
avneesh0612/react-nextjs-snippets
|
https://api.github.com/repos/avneesh0612/react-nextjs-snippets
|
closed
|
[FEATURE] Add imports and hooks to Typescript in React
|
good first issue 🕹 aspect: interface 🏁 status: ready for dev ⭐ goal: addition 🟧 priority: high
|
### Description
The react imports and hooks don't work in typescript `tsx` files. So, we just need to add these imports in ts snippets JSON and add types if and where needed
### Screenshots
_No response_
### Additional information
_No response_
|
1.0
|
[FEATURE] Add imports and hooks to Typescript in React - ### Description
The react imports and hooks don't work in typescript `tsx` files. So, we just need to add these imports in ts snippets JSON and add types if and where needed
### Screenshots
_No response_
### Additional information
_No response_
|
priority
|
add imports and hooks to typescript in react description the react imports and hooks don t work in typescript tsx files so we just need to add these imports in ts snippets json and add types if and where needed screenshots no response additional information no response
| 1
|
229,829
| 7,595,693,532
|
IssuesEvent
|
2018-04-27 06:52:53
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
opened
|
Setting local claims for federated users is not handled
|
Component/Auth Framework Priority/High Type/Bug
|
When calling user.localClaims["<claim_uri>"] = "value"; this should check for the remote claim from dialect or claim mapping and set the corresponding claim value.
|
1.0
|
Setting local claims for federated users is not handled - When calling user.localClaims["<claim_uri>"] = "value"; this should check for the remote claim from dialect or claim mapping and set the corresponding claim value.
|
priority
|
setting local claims for federated users is not handled when calling user localclaims value this should check for the remote claim from dialect or claim mapping and set the corresponding claim value
| 1
|
328,474
| 9,995,182,280
|
IssuesEvent
|
2019-07-11 19:35:21
|
rstudio/gt
|
https://api.github.com/repos/rstudio/gt
|
closed
|
Rename the `cells_styles()` function as `cell_text()`; create `cell_fill()` fcn
|
Difficulty: ① Novice Effort: ① Low Priority: ③ High Type: ★ Enhancement
|
Rename the `cells_styles()` function as `cell_text()`. This helper function should only be concerned with cell text so all `text_*` args should be renamed to lose the `text_` part.
|
1.0
|
Rename the `cells_styles()` function as `cell_text()`; create `cell_fill()` fcn - Rename the `cells_styles()` function as `cell_text()`. This helper function should only be concerned with cell text so all `text_*` args should be renamed to lose the `text_` part.
|
priority
|
rename the cells styles function as cell text create cell fill fcn rename the cells styles function as cell text this helper function should only be concerned with cell text so all text args should be renamed to lose the text part
| 1
|
482,488
| 13,907,692,671
|
IssuesEvent
|
2020-10-20 12:56:28
|
ctm/mb2-doc
|
https://api.github.com/repos/ctm/mb2-doc
|
closed
|
Mexican poker log strings are incorrect
|
chore easy high priority
|
When up and down are chosen in Mexican poker, the text that gets printed talks about discarding. Ugh.
That may be a little tricky to solve cleanly, but I can almost definitely hack in something that's better even if it's not a general solution.
|
1.0
|
Mexican poker log strings are incorrect - When up and down are chosen in Mexican poker, the text that gets printed talks about discarding. Ugh.
That may be a little tricky to solve cleanly, but I can almost definitely hack in something that's better even if it's not a general solution.
|
priority
|
mexican poker log strings are incorrect when up and down are chosen in mexican poker the text that gets printed talks about discarding ugh that may be a little tricky to solve cleanly but i can almost definitely hack in something that s better even if it s not a general solution
| 1
|
615,216
| 19,250,014,456
|
IssuesEvent
|
2021-12-09 03:17:19
|
matrixorigin/matrixone
|
https://api.github.com/repos/matrixorigin/matrixone
|
opened
|
add AOE RFC documents
|
component/aoe priority/high kind/feature severity/critical
|
1. Overall architecture
2. WAL
3. Buffer manager
4. Metadata
5. Data and index management
6. MVCC
7. Logstore
|
1.0
|
add AOE RFC documents - 1. Overall architecture
2. WAL
3. Buffer manager
4. Metadata
5. Data and index management
6. MVCC
7. Logstore
|
priority
|
add aoe rfc documents overall architecture wal buffer manager metadata data and index management mvcc logstore
| 1
|
591,959
| 17,866,597,354
|
IssuesEvent
|
2021-09-06 10:10:54
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Add an API for finding symbol references in a given document only
|
Type/Improvement Priority/High Team/CompilerFETools Points/4 Area/SemanticAPI SwanLakeDump
|
**Description:**
Currently, in order to find the references we access all the constructs in a given BLangPackage and when it comes to a larger project, we iterate through around hundreds of constructs in a single run.
Since we use the references API to implement the semantic token feature, we spend cycles to find the references in unnecessary documents and related constructs as well. Exposing an API to capture the references in a given document we can improve the performance.
**Describe your solution(s)**
Introduce an API to find the references in a single document
|
1.0
|
Add an API for finding symbol references in a given document only - **Description:**
Currently, in order to find the references we access all the constructs in a given BLangPackage and when it comes to a larger project, we iterate through around hundreds of constructs in a single run.
Since we use the references API to implement the semantic token feature, we spend cycles to find the references in unnecessary documents and related constructs as well. Exposing an API to capture the references in a given document we can improve the performance.
**Describe your solution(s)**
Introduce an API to find the references in a single document
|
priority
|
add an api for finding symbol references in a given document only description currently in order to find the references we access all the constructs in a given blangpackage and when it comes to a larger project we iterate through around hundreds of constructs in a single run since we use the references api to implement the semantic token feature we spend cycles to find the references in unnecessary documents and related constructs as well exposing an api to capture the references in a given document we can improve the performance describe your solution s introduce an api to find the references in a single document
| 1
|
602,369
| 18,467,769,356
|
IssuesEvent
|
2021-10-17 07:22:29
|
SmallMolecules/small-molecules
|
https://api.github.com/repos/SmallMolecules/small-molecules
|
closed
|
Create particle object
|
Priority: High Size: Medium
|
Need to create a basic particle object following the mind map that was made earlier.

|
1.0
|
Create particle object - Need to create a basic particle object following the mind map that was made earlier.

|
priority
|
create particle object need to create a basic particle object following the mind map that was made earlier
| 1
|
209,937
| 7,181,373,248
|
IssuesEvent
|
2018-02-01 04:33:51
|
Mawerh/OOPP
|
https://api.github.com/repos/Mawerh/OOPP
|
closed
|
Function for weather api
|
Complexity High Priority High
|
As a person who is out most of the time, i wish for there to be some way my windows will close and open based on the weather.
|
1.0
|
Function for weather api - As a person who is out most of the time, i wish for there to be some way my windows will close and open based on the weather.
|
priority
|
function for weather api as a person who is out most of the time i wish for there to be some way my windows will close and open based on the weather
| 1
|
281,105
| 8,690,956,575
|
IssuesEvent
|
2018-12-03 23:15:56
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
closed
|
send_email: Add support for passing in additional `To` addresses
|
area: emails area: refactoring help wanted in progress priority: high
|
For the recent "realm reactivation request" feature (https://github.com/zulip/zulip/pull/10816) and likely several similar features we may add in the future where we want to email all organization administrators, we ideally want to off a `send_email` wrapper/variant called maybe `send_email_to_admins` that take a realm and sends the provided email to all the organization administrators of the realm, but in a single email (which means that e.g. those users replying to each other works in a good way).
|
1.0
|
send_email: Add support for passing in additional `To` addresses - For the recent "realm reactivation request" feature (https://github.com/zulip/zulip/pull/10816) and likely several similar features we may add in the future where we want to email all organization administrators, we ideally want to off a `send_email` wrapper/variant called maybe `send_email_to_admins` that take a realm and sends the provided email to all the organization administrators of the realm, but in a single email (which means that e.g. those users replying to each other works in a good way).
|
priority
|
send email add support for passing in additional to addresses for the recent realm reactivation request feature and likely several similar features we may add in the future where we want to email all organization administrators we ideally want to off a send email wrapper variant called maybe send email to admins that take a realm and sends the provided email to all the organization administrators of the realm but in a single email which means that e g those users replying to each other works in a good way
| 1
|
797,869
| 28,182,081,430
|
IssuesEvent
|
2023-04-04 03:56:36
|
Snapmaker/Luban
|
https://api.github.com/repos/Snapmaker/Luban
|
closed
|
Bug: Variable line width skeleton production error due to polygon problem
|
Type: Bug/Bug Fix Priority: High Software: Slicer
|
## Description
This is a slicing engine problem.

In the slicing of the above model there is a certain possibility of slicing failure at different layer thickness parameters. This issue was also found in Cura Engine and was not fully resolved in 5.3.
|
1.0
|
Bug: Variable line width skeleton production error due to polygon problem - ## Description
This is a slicing engine problem.

In the slicing of the above model there is a certain possibility of slicing failure at different layer thickness parameters. This issue was also found in Cura Engine and was not fully resolved in 5.3.
|
priority
|
bug variable line width skeleton production error due to polygon problem description this is a slicing engine problem in the slicing of the above model there is a certain possibility of slicing failure at different layer thickness parameters this issue was also found in cura engine and was not fully resolved in
| 1
|
276,649
| 8,607,013,169
|
IssuesEvent
|
2018-11-17 17:58:31
|
josephroqueca/bowling-companion
|
https://api.github.com/repos/josephroqueca/bowling-companion
|
opened
|
Update Kotlin to 1.3
|
high priority
|
New stable release of Kotlin, with coroutines hitting version 1.0.0 means this will need more than just a simple version bump
|
1.0
|
Update Kotlin to 1.3 - New stable release of Kotlin, with coroutines hitting version 1.0.0 means this will need more than just a simple version bump
|
priority
|
update kotlin to new stable release of kotlin with coroutines hitting version means this will need more than just a simple version bump
| 1
|
196,479
| 6,928,382,417
|
IssuesEvent
|
2017-12-01 04:29:42
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[studio-ui] Insert component, layout and stub RTE dropdowns wrap early and do not allow scroll
|
bug priority: high
|
### Expected behavior
* Dropdown should resize the width to match the text.
* Dropdown should be scrollable after it reaches 80% of the length of the window.
### Actual behavior

### Steps to reproduce the problem
* configure a lot of components to the RTE insert stubs
* Open RTE
Applies to 2.x and 3.x
|
1.0
|
[studio-ui] Insert component, layout and stub RTE dropdowns wrap early and do not allow scroll - ### Expected behavior
* Dropdown should resize the width to match the text.
* Dropdown should be scrollable after it reaches 80% of the length of the window.
### Actual behavior

### Steps to reproduce the problem
* configure a lot of components to the RTE insert stubs
* Open RTE
Applies to 2.x and 3.x
|
priority
|
insert component layout and stub rte dropdowns wrap early and do not allow scroll expected behavior dropdown should resize the width to match the text dropdown should be scrollable after it reaches of the length of the window actual behavior steps to reproduce the problem configure a lot of components to the rte insert stubs open rte applies to x and x
| 1
|
514,911
| 14,946,811,036
|
IssuesEvent
|
2021-01-26 07:33:26
|
ProjectSidewalk/SidewalkWebpage
|
https://api.github.com/repos/ProjectSidewalk/SidewalkWebpage
|
closed
|
Sidewalk Gallery: Cards in "All" should be shuffled randomly
|
Priority: High bug sidewalkgallery
|
Currently, while the cards of each label type are chosen randomly, they appear on the "All" page in a grouped order (i.e., Others, Occlusions, No Sidewalk...). This is not ideal and we probably should shuffle them more.
A quick fix is just to shuffle the list of labels queried in the `getAssortedLabels()` method in `LabelTable.scala` before returning.
|
1.0
|
Sidewalk Gallery: Cards in "All" should be shuffled randomly - Currently, while the cards of each label type are chosen randomly, they appear on the "All" page in a grouped order (i.e., Others, Occlusions, No Sidewalk...). This is not ideal and we probably should shuffle them more.
A quick fix is just to shuffle the list of labels queried in the `getAssortedLabels()` method in `LabelTable.scala` before returning.
|
priority
|
sidewalk gallery cards in all should be shuffled randomly currently while the cards of each label type are chosen randomly they appear on the all page in a grouped order i e others occlusions no sidewalk this is not ideal and we probably should shuffle them more a quick fix is just to shuffle the list of labels queried in the getassortedlabels method in labeltable scala before returning
| 1
|
150,959
| 5,794,236,336
|
IssuesEvent
|
2017-05-02 14:29:56
|
GSA/fpki-guides
|
https://api.github.com/repos/GSA/fpki-guides
|
closed
|
Trust Store Management Guide
|
Audience - Engineers general overview Priority - High
|
@dasgituser (Dave Silver) and @tkpk (Giuseppe Cimmino) are converting the FPKI Management Authority''s Trust Store Management Guide to a playbook. The Federal Public Key Infrastructure Management Authority designed and created the Trust Store Management Guide as an education resource for Department, Agency, corporate, and other organizational system level administrators and managers who use the Federal Public Key Infrastructure (FPKI) as part of regular business practices.
|
1.0
|
Trust Store Management Guide - @dasgituser (Dave Silver) and @tkpk (Giuseppe Cimmino) are converting the FPKI Management Authority''s Trust Store Management Guide to a playbook. The Federal Public Key Infrastructure Management Authority designed and created the Trust Store Management Guide as an education resource for Department, Agency, corporate, and other organizational system level administrators and managers who use the Federal Public Key Infrastructure (FPKI) as part of regular business practices.
|
priority
|
trust store management guide dasgituser dave silver and tkpk giuseppe cimmino are converting the fpki management authority s trust store management guide to a playbook the federal public key infrastructure management authority designed and created the trust store management guide as an education resource for department agency corporate and other organizational system level administrators and managers who use the federal public key infrastructure fpki as part of regular business practices
| 1
|
773,108
| 27,146,514,137
|
IssuesEvent
|
2023-02-16 20:24:18
|
IDAES/idaes-pse
|
https://api.github.com/repos/IDAES/idaes-pse
|
closed
|
Cubic EoS model fails to initialize at high pressures.
|
Priority:High
|
An external user brought a use case to my attention where the Cubic EoS model failed to converge at high pressures.
The issue appears to be that the deviation between the ideal initial guess and the actual solution for bubble temperature is large enough to cause the solver to fail (scaling might also be an issue).
Test case files attached.
[C1C2C3_CEOS.txt](https://github.com/IDAES/idaes-dev/files/5189179/C1C2C3_CEOS.txt)
[mainfile_test_cubic_eos_PR.txt](https://github.com/IDAES/idaes-dev/files/5189180/mainfile_test_cubic_eos_PR.txt)
|
1.0
|
Cubic EoS model fails to initialize at high pressures. - An external user brought a use case to my attention where the Cubic EoS model failed to converge at high pressures.
The issue appears to be that the deviation between the ideal initial guess and the actual solution for bubble temperature is large enough to cause the solver to fail (scaling might also be an issue).
Test case files attached.
[C1C2C3_CEOS.txt](https://github.com/IDAES/idaes-dev/files/5189179/C1C2C3_CEOS.txt)
[mainfile_test_cubic_eos_PR.txt](https://github.com/IDAES/idaes-dev/files/5189180/mainfile_test_cubic_eos_PR.txt)
|
priority
|
cubic eos model fails to initialize at high pressures an external user brought a use case to my attention where the cubic eos model failed to converge at high pressures the issue appears to be that the deviation between the ideal initial guess and the actual solution for bubble temperature is large enough to cause the solver to fail scaling might also be an issue test case files attached
| 1
|
715,207
| 24,590,176,487
|
IssuesEvent
|
2022-10-14 00:53:05
|
Krenbot/body-of-cards
|
https://api.github.com/repos/Krenbot/body-of-cards
|
closed
|
Implement game rules section for MVP
|
enhancement high priority MVP
|
## Main Points
- Card value indicates the number of reps
- 2 through 10 are their face values
- Jack is 11
- Queen is 12
- King is 13
- Ace is 14
- Joker is ?? TBD
- You can swap each card only once
- Each card has a linked exercise that appears once the card is shown
|
1.0
|
Implement game rules section for MVP - ## Main Points
- Card value indicates the number of reps
- 2 through 10 are their face values
- Jack is 11
- Queen is 12
- King is 13
- Ace is 14
- Joker is ?? TBD
- You can swap each card only once
- Each card has a linked exercise that appears once the card is shown
|
priority
|
implement game rules section for mvp main points card value indicates the number of reps through are their face values jack is queen is king is ace is joker is tbd you can swap each card only once each card has a linked exercise that appears once the card is shown
| 1
|
45,042
| 2,919,845,579
|
IssuesEvent
|
2015-06-24 16:00:46
|
andresriancho/w3af
|
https://api.github.com/repos/andresriancho/w3af
|
closed
|
export requests / import results
|
improvement plugin priority:high
|
## User story
The export requests plugin writes the fuzzable request in a crappy format that doesn't allow me to export multipart, json, etc.
The import results doesn't allow me to import requests which are not post-data / QS.
## Conditions of satisfaction
- [x] A new export requests plugin replaces the current one. We save the raw HTTP request to the file
- [x] A new import results plugin replaces the current one. We load the HTTP request form the output file using the http request parser
|
1.0
|
export requests / import results - ## User story
The export requests plugin writes the fuzzable request in a crappy format that doesn't allow me to export multipart, json, etc.
The import results doesn't allow me to import requests which are not post-data / QS.
## Conditions of satisfaction
- [x] A new export requests plugin replaces the current one. We save the raw HTTP request to the file
- [x] A new import results plugin replaces the current one. We load the HTTP request form the output file using the http request parser
|
priority
|
export requests import results user story the export requests plugin writes the fuzzable request in a crappy format that doesn t allow me to export multipart json etc the import results doesn t allow me to import requests which are not post data qs conditions of satisfaction a new export requests plugin replaces the current one we save the raw http request to the file a new import results plugin replaces the current one we load the http request form the output file using the http request parser
| 1
|
179,713
| 6,627,965,219
|
IssuesEvent
|
2017-09-23 11:39:53
|
inf3rno/dataflower
|
https://api.github.com/repos/inf3rno/dataflower
|
closed
|
Should I continue the project?
|
priority:high undetermined
|
I am not sure whether I should continue the project now that async functions are supported in node and in browsers with babel.
|
1.0
|
Should I continue the project? - I am not sure whether I should continue the project now that async functions are supported in node and in browsers with babel.
|
priority
|
should i continue the project i am not sure whether i should continue the project now that async functions are supported in node and in browsers with babel
| 1
|
700,891
| 24,077,214,098
|
IssuesEvent
|
2022-09-18 23:46:45
|
naia-lib/naia
|
https://api.github.com/repos/naia-lib/naia
|
closed
|
Client should emit an AuthFailed event
|
naia bug feature high priority
|
If a Client fails to Auth, it currently will just keep trying to connect to the Server with the same Auth payload ... forever ...
Instead, the Server should repeatedly send back an "AuthFailed" message to the Client, and the Client should stop the handshaking process until you call `Client.connect()` again (hopefully with a new and working Auth payload)
|
1.0
|
Client should emit an AuthFailed event - If a Client fails to Auth, it currently will just keep trying to connect to the Server with the same Auth payload ... forever ...
Instead, the Server should repeatedly send back an "AuthFailed" message to the Client, and the Client should stop the handshaking process until you call `Client.connect()` again (hopefully with a new and working Auth payload)
|
priority
|
client should emit an authfailed event if a client fails to auth it currently will just keep trying to connect to the server with the same auth payload forever instead the server should repeatedly send back an authfailed message to the client and the client should stop the handshaking process until you call client connect again hopefully with a new and working auth payload
| 1
|
185,595
| 6,725,677,368
|
IssuesEvent
|
2017-10-17 06:56:14
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
OAuth Implicit Flow fails with exception
|
Component/OAuth Priority/High Type/Bug
|
**Description:**
<!-- Give a brief description of the issue -->
When testing the WSO2 OAuth2 Playground application I get an exception:
**Suggested Labels:**
<!-- For non-committers only. Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels. Labels can be found here - https://github.com/wso2/product-is/labels -->
**Suggested Assignees:**
<!--For non-committers only. Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
wso2is-5.4.0-alpha3
**OS, DB, other environment details and versions:**
Windows 10
**Steps to reproduce:**
I modified the identity.xml
```xml
<Begin File>
<OAuth>
<AppInfoCacheTimeout>-1</AppInfoCacheTimeout>
<AuthorizationGrantCacheTimeout>-1</AuthorizationGrantCacheTimeout>
<SessionDataCacheTimeout>-1</SessionDataCacheTimeout>
<ClaimCacheTimeout>-1</ClaimCacheTimeout>
<IdentityOAuthTokenGenerator>org.wso2.carbon.identity.oauth2.token.JWTTokenIssuer</IdentityOAuthTokenGenerator>
<AccessTokenValueGenerator>org.wso2.carbon.identity.oauth.tokenvaluegenerator.SHA256Generator</AccessTokenValueGenerator>
<EndFile>
```
(File is attached)
In wso2 console create a simple identity provider (only name given, all values default)
In wso2 console create a service provider
Add "OAuth/OpenID Connect Configuration"
with callback url http://wso2is.local:8080/playground2/oauth2client
start playground app
Use Implicit flow with
Callback URL : | http://wso2is.local:8080/playground2/oauth2client
Authorize Endpoint : https://localhost:9443/oauth2/authorize
sign in
Press Approve
In the console an error is logged:
```
[2017-09-27 13:52:18,377] ERROR {org.wso2.carbon.identity.oauth2.OAuth2Service} - Error occurred when processing the authorization request. Returning an error back to client.
org.wso2.carbon.identity.oauth2.IdentityOAuth2Exception: Error occurred while storing new access token :
....
Value too long for column "ACCESS_TOKEN VARCHAR(512)": "
....
```
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
[identity.zip](https://github.com/wso2/product-is/files/1336979/identity.zip)
|
1.0
|
OAuth Implicit Flow fails with exception - **Description:**
<!-- Give a brief description of the issue -->
When testing the WSO2 OAuth2 Playground application I get an exception:
**Suggested Labels:**
<!-- For non-committers only. Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels. Labels can be found here - https://github.com/wso2/product-is/labels -->
**Suggested Assignees:**
<!--For non-committers only. Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
wso2is-5.4.0-alpha3
**OS, DB, other environment details and versions:**
Windows 10
**Steps to reproduce:**
I modified the identity.xml
```xml
<Begin File>
<OAuth>
<AppInfoCacheTimeout>-1</AppInfoCacheTimeout>
<AuthorizationGrantCacheTimeout>-1</AuthorizationGrantCacheTimeout>
<SessionDataCacheTimeout>-1</SessionDataCacheTimeout>
<ClaimCacheTimeout>-1</ClaimCacheTimeout>
<IdentityOAuthTokenGenerator>org.wso2.carbon.identity.oauth2.token.JWTTokenIssuer</IdentityOAuthTokenGenerator>
<AccessTokenValueGenerator>org.wso2.carbon.identity.oauth.tokenvaluegenerator.SHA256Generator</AccessTokenValueGenerator>
<EndFile>
```
(File is attached)
In wso2 console create a simple identity provider (only name given, all values default)
In wso2 console create a service provider
Add "OAuth/OpenID Connect Configuration"
with callback url http://wso2is.local:8080/playground2/oauth2client
start playground app
Use Implicit flow with
Callback URL : | http://wso2is.local:8080/playground2/oauth2client
Authorize Endpoint : https://localhost:9443/oauth2/authorize
sign in
Press Approve
In the console an error is logged:
```
[2017-09-27 13:52:18,377] ERROR {org.wso2.carbon.identity.oauth2.OAuth2Service} - Error occurred when processing the authorization request. Returning an error back to client.
org.wso2.carbon.identity.oauth2.IdentityOAuth2Exception: Error occurred while storing new access token :
....
Value too long for column "ACCESS_TOKEN VARCHAR(512)": "
....
```
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
[identity.zip](https://github.com/wso2/product-is/files/1336979/identity.zip)
|
priority
|
oauth implicit flow fails with exception description when testing the playground application i get an exception suggested labels suggested assignees affected product version os db other environment details and versions windows steps to reproduce i modified the identity xml xml org carbon identity token jwttokenissuer org carbon identity oauth tokenvaluegenerator file is attached in console create a simple identity provider only name given all values default in console create a service provider add oauth openid connect configuration with callback url start playground app use implicit flow with callback url authorize endpoint sign in press approve in the console an error is logged error org carbon identity error occurred when processing the authorization request returning an error back to client org carbon identity error occurred while storing new access token value too long for column access token varchar related issues
| 1
|
395,834
| 11,697,156,895
|
IssuesEvent
|
2020-03-06 11:11:59
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Unable to login when the only authentication step is fido
|
Affected/5.10.0-RC1 Priority/Highest Severity/Blocker
|
**Steps to reproduce:**
1. Create an application configure fido as the authenticator (single factor - fido - username less)
2. Try to login to the application.
3. When prompted, tap the fido device and complete the authentication.
The below exception is printed in the backend and user is redirected to error page..
```
[2020-03-06 14:46:46,443] [05202b49-9629-48cf-8336-07273b9bd7ea] ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Exception in Authentication Framework java.lang.NullPointerException
at org.wso2.carbon.identity.application.authentication.framework.util.FrameworkUtils.getUserStoreManager(FrameworkUtils.java:2489)
at org.wso2.carbon.identity.application.authentication.framework.util.FrameworkUtils.resolveUserIdFromUsername(FrameworkUtils.java:2439)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.storeSessionData(DefaultAuthenticationRequestHandler.java:612)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.concludeFlow(DefaultAuthenticationRequestHandler.java:476)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.handle(DefaultAuthenticationRequestHandler.java:168)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator.handle(DefaultRequestCoordinator.java:240)
at org.wso2.carbon.identity.application.authentication.framework.servlet.CommonAuthenticationServlet.doPost(CommonAuthenticationServlet.java:53)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:741)
at org.eclipse.equinox.http.helper.ContextPathServletAdaptor.service(ContextPathServletAdaptor.java:37)
at org.eclipse.equinox.http.servlet.internal.ServletRegistration.service(ServletRegistration.java:61)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.processAlias(ProxyServlet.java:128)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.service(ProxyServlet.java:60)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:741)
at org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.service(DelegationServlet.java:68)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.wso2.carbon.identity.captcha.filter.CaptchaFilter.doFilter(CaptchaFilter.java:66)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.owasp.csrfguard.CsrfGuardFilter.doFilter(CsrfGuardFilter.java:72)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.wso2.carbon.tomcat.ext.filter.CharacterSetFilter.doFilter(CharacterSetFilter.java:65)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:126)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92)
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86)
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110)
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99)
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62)
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145)
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688)
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57)
at org.wso2.carbon.tomcat.ext.valves.RequestEncodingValve.invoke(RequestEncodingValve.java:49)
at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343)
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367)
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65)
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639)
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:748)
```
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
1.0
|
Unable to login when the only authentication step is fido - **Steps to reproduce:**
1. Create an application configure fido as the authenticator (single factor - fido - username less)
2. Try to login to the application.
3. When prompted, tap the fido device and complete the authentication.
The below exception is printed in the backend and user is redirected to error page..
```
[2020-03-06 14:46:46,443] [05202b49-9629-48cf-8336-07273b9bd7ea] ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Exception in Authentication Framework java.lang.NullPointerException
at org.wso2.carbon.identity.application.authentication.framework.util.FrameworkUtils.getUserStoreManager(FrameworkUtils.java:2489)
at org.wso2.carbon.identity.application.authentication.framework.util.FrameworkUtils.resolveUserIdFromUsername(FrameworkUtils.java:2439)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.storeSessionData(DefaultAuthenticationRequestHandler.java:612)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.concludeFlow(DefaultAuthenticationRequestHandler.java:476)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultAuthenticationRequestHandler.handle(DefaultAuthenticationRequestHandler.java:168)
at org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator.handle(DefaultRequestCoordinator.java:240)
at org.wso2.carbon.identity.application.authentication.framework.servlet.CommonAuthenticationServlet.doPost(CommonAuthenticationServlet.java:53)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:741)
at org.eclipse.equinox.http.helper.ContextPathServletAdaptor.service(ContextPathServletAdaptor.java:37)
at org.eclipse.equinox.http.servlet.internal.ServletRegistration.service(ServletRegistration.java:61)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.processAlias(ProxyServlet.java:128)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.service(ProxyServlet.java:60)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:741)
at org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.service(DelegationServlet.java:68)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.wso2.carbon.identity.captcha.filter.CaptchaFilter.doFilter(CaptchaFilter.java:66)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.owasp.csrfguard.CsrfGuardFilter.doFilter(CsrfGuardFilter.java:72)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.wso2.carbon.tomcat.ext.filter.CharacterSetFilter.doFilter(CharacterSetFilter.java:65)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.catalina.filters.HttpHeaderSecurityFilter.doFilter(HttpHeaderSecurityFilter.java:126)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92)
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86)
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110)
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99)
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49)
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62)
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145)
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688)
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57)
at org.wso2.carbon.tomcat.ext.valves.RequestEncodingValve.invoke(RequestEncodingValve.java:49)
at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343)
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367)
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65)
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639)
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:748)
```
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
priority
|
unable to login when the only authentication step is fido steps to reproduce create an application configure fido as the authenticator single factor fido username less try to login to the application when prompted tap the fido device and complete the authentication the below exception is printed in the backend and user is redirected to error page error org carbon identity application authentication framework handler request impl defaultrequestcoordinator exception in authentication framework java lang nullpointerexception at org carbon identity application authentication framework util frameworkutils getuserstoremanager frameworkutils java at org carbon identity application authentication framework util frameworkutils resolveuseridfromusername frameworkutils java at org carbon identity application authentication framework handler request impl defaultauthenticationrequesthandler storesessiondata defaultauthenticationrequesthandler java at org carbon identity application authentication framework handler request impl defaultauthenticationrequesthandler concludeflow defaultauthenticationrequesthandler java at org carbon identity application authentication framework handler request impl defaultauthenticationrequesthandler handle defaultauthenticationrequesthandler java at org carbon identity application authentication framework handler request impl defaultrequestcoordinator handle defaultrequestcoordinator java at org carbon identity application authentication framework servlet commonauthenticationservlet dopost commonauthenticationservlet java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at org eclipse equinox http helper contextpathservletadaptor service contextpathservletadaptor java at org eclipse equinox http servlet internal servletregistration service servletregistration java at org eclipse equinox http servlet internal proxyservlet processalias proxyservlet java at org eclipse equinox http servlet internal proxyservlet service proxyservlet java at javax servlet http httpservlet service httpservlet java at org carbon tomcat ext servlet delegationservlet service delegationservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org carbon identity captcha filter captchafilter dofilter captchafilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org owasp csrfguard csrfguardfilter dofilter csrfguardfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org carbon tomcat ext filter charactersetfilter dofilter charactersetfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina filters httpheadersecurityfilter dofilter httpheadersecurityfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org carbon identity context rewrite valve tenantcontextrewritevalve invoke tenantcontextrewritevalve java at org carbon identity authz valve authorizationvalve invoke authorizationvalve java at org carbon identity auth valve authenticationvalve invoke authenticationvalve java at org carbon tomcat ext valves compositevalve continueinvocation compositevalve java at org carbon tomcat ext valves tomcatvalvecontainer invokevalves tomcatvalvecontainer java at org carbon tomcat ext valves compositevalve invoke compositevalve java at org carbon tomcat ext valves carbonstuckthreaddetectionvalve invoke carbonstuckthreaddetectionvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org carbon tomcat ext valves carboncontextcreatorvalve invoke carboncontextcreatorvalve java at org carbon tomcat ext valves requestencodingvalve invoke requestencodingvalve java at org carbon tomcat ext valves requestcorrelationidvalve invoke requestcorrelationidvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java related issues
| 1
|
667,635
| 22,495,124,873
|
IssuesEvent
|
2022-06-23 06:49:33
|
ballerina-platform/ballerina-standard-library
|
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
|
closed
|
Should a resource with POST action respond with 201 Created by default?
|
Points/3 Priority/High Type/Improvement module/http Team/PCM
|
The following resource responds with 200 OK at the moment.
```ballerina
resource function post albums(@http:Payload Album album) returns Album {
....
}
```
Should it respond with 201 instead? Post is used to create or append new resources. In that case, responding with 201 make sense to me.
|
1.0
|
Should a resource with POST action respond with 201 Created by default? - The following resource responds with 200 OK at the moment.
```ballerina
resource function post albums(@http:Payload Album album) returns Album {
....
}
```
Should it respond with 201 instead? Post is used to create or append new resources. In that case, responding with 201 make sense to me.
|
priority
|
should a resource with post action respond with created by default the following resource responds with ok at the moment ballerina resource function post albums http payload album album returns album should it respond with instead post is used to create or append new resources in that case responding with make sense to me
| 1
|
445,502
| 12,831,960,575
|
IssuesEvent
|
2020-07-07 06:42:46
|
gambitph/Stackable
|
https://api.github.com/repos/gambitph/Stackable
|
opened
|
Icon Spacing is not working on frontend
|
[block] icon bug high priority
|
Icon Spacing is not working on frontend
frontend:

backend:

|
1.0
|
Icon Spacing is not working on frontend - Icon Spacing is not working on frontend
frontend:

backend:

|
priority
|
icon spacing is not working on frontend icon spacing is not working on frontend frontend backend
| 1
|
479,899
| 13,807,140,310
|
IssuesEvent
|
2020-10-11 20:41:56
|
raceintospace/raceintospace
|
https://api.github.com/repos/raceintospace/raceintospace
|
closed
|
Not getting Photo Recon for lunar flybys!
|
High Priority bug
|
The last few games I've gotten no photo reconnaissance from doing lunar flybys. That can't be right.


|
1.0
|
Not getting Photo Recon for lunar flybys! - The last few games I've gotten no photo reconnaissance from doing lunar flybys. That can't be right.


|
priority
|
not getting photo recon for lunar flybys the last few games i ve gotten no photo reconnaissance from doing lunar flybys that can t be right
| 1
|
697,461
| 23,940,407,478
|
IssuesEvent
|
2022-09-11 20:29:56
|
Capitalisk/ldpos-wallet
|
https://api.github.com/repos/Capitalisk/ldpos-wallet
|
closed
|
Handle error in `DetailedData.vue`
|
high priority
|
Error is of type Object and we want to display it accurately, maybe even in red.
```json
{
"name": "VoterAlreadyVotedForDelegateError",
"message": "Voter clskbc84c66b69477c2160ef2eb5c92579840fa0ef3f has already voted for delegate clskbde48efe4fb34f91f3bfad4e4e8feb081b302b0f"
}
```
See [this transaction](http://localhost:8080/#/transactions/e5af4645037dcad2f73298cc2e9683d85f511a23) on the testnet where I encountered the problem.
|
1.0
|
Handle error in `DetailedData.vue` - Error is of type Object and we want to display it accurately, maybe even in red.
```json
{
"name": "VoterAlreadyVotedForDelegateError",
"message": "Voter clskbc84c66b69477c2160ef2eb5c92579840fa0ef3f has already voted for delegate clskbde48efe4fb34f91f3bfad4e4e8feb081b302b0f"
}
```
See [this transaction](http://localhost:8080/#/transactions/e5af4645037dcad2f73298cc2e9683d85f511a23) on the testnet where I encountered the problem.
|
priority
|
handle error in detaileddata vue error is of type object and we want to display it accurately maybe even in red json name voteralreadyvotedfordelegateerror message voter has already voted for delegate see on the testnet where i encountered the problem
| 1
|
474,919
| 13,684,424,668
|
IssuesEvent
|
2020-09-30 04:51:38
|
wso2/docker-is
|
https://api.github.com/repos/wso2/docker-is
|
closed
|
[5.11.0] Avoid Packaging MySQL JDBC Driver in WSO2 Product Docker Images
|
Priority/High Type/Improvement
|
**Description:**
It has been suggested to avoid packaging the MySQL JDBC Driver in the WSO2 product Docker images.
The original purpose of packaging this binary in Docker images was to support evaluatory deployments in container platforms such as, Kubernetes with externalized MySQL RDBMS support.
For a user who intends to use a different version of the particular connector (as opposed to what is packaged inside the image) may not be able to use the current, official WSO2 product Docker images as the base since, he/she may experience version conflicts.
**Affected Product Version:**
Docker resources for WSO2 IAM version `v5.10.0.3` and below
|
1.0
|
[5.11.0] Avoid Packaging MySQL JDBC Driver in WSO2 Product Docker Images - **Description:**
It has been suggested to avoid packaging the MySQL JDBC Driver in the WSO2 product Docker images.
The original purpose of packaging this binary in Docker images was to support evaluatory deployments in container platforms such as, Kubernetes with externalized MySQL RDBMS support.
For a user who intends to use a different version of the particular connector (as opposed to what is packaged inside the image) may not be able to use the current, official WSO2 product Docker images as the base since, he/she may experience version conflicts.
**Affected Product Version:**
Docker resources for WSO2 IAM version `v5.10.0.3` and below
|
priority
|
avoid packaging mysql jdbc driver in product docker images description it has been suggested to avoid packaging the mysql jdbc driver in the product docker images the original purpose of packaging this binary in docker images was to support evaluatory deployments in container platforms such as kubernetes with externalized mysql rdbms support for a user who intends to use a different version of the particular connector as opposed to what is packaged inside the image may not be able to use the current official product docker images as the base since he she may experience version conflicts affected product version docker resources for iam version and below
| 1
|
698,637
| 23,987,690,130
|
IssuesEvent
|
2022-09-13 20:42:27
|
responsible-ai-collaborative/aiid
|
https://api.github.com/repos/responsible-ai-collaborative/aiid
|
closed
|
Latest Incident Report image broken on mobile
|
Type:Bug Priority:High
|
### Issue:
When loading the Home page on mobile, the "Latest Incident Report" image breaks on iPhone sizes
### Evidence

### Steps to reproduce
1. Open https://incidentdatabase.ai/ in chrome with iPhone dimensions
2. First image appears broken in section "latest Incident Report"
|
1.0
|
Latest Incident Report image broken on mobile - ### Issue:
When loading the Home page on mobile, the "Latest Incident Report" image breaks on iPhone sizes
### Evidence

### Steps to reproduce
1. Open https://incidentdatabase.ai/ in chrome with iPhone dimensions
2. First image appears broken in section "latest Incident Report"
|
priority
|
latest incident report image broken on mobile issue when loading the home page on mobile the latest incident report image breaks on iphone sizes evidence steps to reproduce open in chrome with iphone dimensions first image appears broken in section latest incident report
| 1
|
606,823
| 18,768,908,018
|
IssuesEvent
|
2021-11-06 13:07:41
|
AesaraB/contextual-wiktionary
|
https://api.github.com/repos/AesaraB/contextual-wiktionary
|
closed
|
Unicode issues
|
bug high priority
|
* đã
* đã
These two words are displayed the same, but their unicode chars are different:
* %C4%91a%CC%83
* %C4%91%C3%A3
On wiktionary, both:
* http://en.wiktionary.org/wiki/%C4%91a%CC%83
* http://en.wiktionary.org/wiki/%C4%91%C3%A3
work and lead to the same page.
With this extension, only the second one works. Is it possible to fix this issue ?
|
1.0
|
Unicode issues - * đã
* đã
These two words are displayed the same, but their unicode chars are different:
* %C4%91a%CC%83
* %C4%91%C3%A3
On wiktionary, both:
* http://en.wiktionary.org/wiki/%C4%91a%CC%83
* http://en.wiktionary.org/wiki/%C4%91%C3%A3
work and lead to the same page.
With this extension, only the second one works. Is it possible to fix this issue ?
|
priority
|
unicode issues đã đã these two words are displayed the same but their unicode chars are different cc on wiktionary both work and lead to the same page with this extension only the second one works is it possible to fix this issue
| 1
|
745,391
| 25,982,405,692
|
IssuesEvent
|
2022-12-19 20:08:04
|
encorelab/ck-board
|
https://api.github.com/repos/encorelab/ck-board
|
closed
|
Non-fixed nav bar in CK Monitor
|
bug high priority
|
In the nav bar, once "View by TODOs" (see image) is selected and there are a long list of TODOs, scrolling through the TODO items will result in the nav bar scrolling out of view
<img width="447" alt="Screen Shot 2022-12-16 at 3 54 05 PM" src="https://user-images.githubusercontent.com/6416247/208187212-eb24b9a9-46a8-44ed-a527-46407250d148.png">
|
1.0
|
Non-fixed nav bar in CK Monitor - In the nav bar, once "View by TODOs" (see image) is selected and there are a long list of TODOs, scrolling through the TODO items will result in the nav bar scrolling out of view
<img width="447" alt="Screen Shot 2022-12-16 at 3 54 05 PM" src="https://user-images.githubusercontent.com/6416247/208187212-eb24b9a9-46a8-44ed-a527-46407250d148.png">
|
priority
|
non fixed nav bar in ck monitor in the nav bar once view by todos see image is selected and there are a long list of todos scrolling through the todo items will result in the nav bar scrolling out of view img width alt screen shot at pm src
| 1
|
344,004
| 10,339,099,907
|
IssuesEvent
|
2019-09-03 18:30:27
|
netdata/netdata
|
https://api.github.com/repos/netdata/netdata
|
closed
|
no clear notifications with repeat feature
|
bug needs triage priority/high size:1
|
Hi all,
##### Bug report summary
I'm not receiving `CLEAR` (green) notifications any more after the upgrade to netdata 1.16 and enabling the repeat notification feature. Also, I don't see the alarms in web ui Alarms Log.
Config change to enable the repeat feature:
```
[health]
stock health configuration directory = /etc/netdata/conf.d/health.d
default repeat warning = 2h
default repeat critical = 1h
```
##### OS / Environment
```
# /usr/libexec/netdata/plugins.d/system-info.sh
NETDATA_SYSTEM_OS_NAME="CentOS Linux"
NETDATA_SYSTEM_OS_ID=centos
NETDATA_SYSTEM_OS_ID_LIKE=rhel fedora
NETDATA_SYSTEM_OS_VERSION=7 (Core)
NETDATA_SYSTEM_OS_VERSION_ID=7
NETDATA_SYSTEM_OS_DETECTION=/etc/os-release
NETDATA_SYSTEM_KERNEL_NAME=Linux
NETDATA_SYSTEM_KERNEL_VERSION=3.10.0-957.21.3.el7.x86_64
NETDATA_SYSTEM_ARCHITECTURE=x86_64
NETDATA_SYSTEM_VIRTUALIZATION=kvm
NETDATA_SYSTEM_VIRT_DETECTION=systemd-detect-virt
NETDATA_SYSTEM_CONTAINER=none
NETDATA_SYSTEM_CONTAINER_DETECTION=systemd-detect-virt
```
##### Netdata version (ouput of `netdata -V`)
```
# netdata -V
netdata v1.16.0
```
##### Component Name
health
##### Steps To Reproduce
Upgrade to netdata 1.16, globally enable repeat notification feature as described above, wait for alarm. You should receive notification with raised alarm, and once the the alarm is gone, clear notification. I didn't receive repeated notification nor clear notification.
##### Expected behavior
Receive clear notification if alarm is gone. Also it should be visible in Alarm Log in web ui, currently it's not.
|
1.0
|
no clear notifications with repeat feature - Hi all,
##### Bug report summary
I'm not receiving `CLEAR` (green) notifications any more after the upgrade to netdata 1.16 and enabling the repeat notification feature. Also, I don't see the alarms in web ui Alarms Log.
Config change to enable the repeat feature:
```
[health]
stock health configuration directory = /etc/netdata/conf.d/health.d
default repeat warning = 2h
default repeat critical = 1h
```
##### OS / Environment
```
# /usr/libexec/netdata/plugins.d/system-info.sh
NETDATA_SYSTEM_OS_NAME="CentOS Linux"
NETDATA_SYSTEM_OS_ID=centos
NETDATA_SYSTEM_OS_ID_LIKE=rhel fedora
NETDATA_SYSTEM_OS_VERSION=7 (Core)
NETDATA_SYSTEM_OS_VERSION_ID=7
NETDATA_SYSTEM_OS_DETECTION=/etc/os-release
NETDATA_SYSTEM_KERNEL_NAME=Linux
NETDATA_SYSTEM_KERNEL_VERSION=3.10.0-957.21.3.el7.x86_64
NETDATA_SYSTEM_ARCHITECTURE=x86_64
NETDATA_SYSTEM_VIRTUALIZATION=kvm
NETDATA_SYSTEM_VIRT_DETECTION=systemd-detect-virt
NETDATA_SYSTEM_CONTAINER=none
NETDATA_SYSTEM_CONTAINER_DETECTION=systemd-detect-virt
```
##### Netdata version (ouput of `netdata -V`)
```
# netdata -V
netdata v1.16.0
```
##### Component Name
health
##### Steps To Reproduce
Upgrade to netdata 1.16, globally enable repeat notification feature as described above, wait for alarm. You should receive notification with raised alarm, and once the the alarm is gone, clear notification. I didn't receive repeated notification nor clear notification.
##### Expected behavior
Receive clear notification if alarm is gone. Also it should be visible in Alarm Log in web ui, currently it's not.
|
priority
|
no clear notifications with repeat feature hi all bug report summary i m not receiving clear green notifications any more after the upgrade to netdata and enabling the repeat notification feature also i don t see the alarms in web ui alarms log config change to enable the repeat feature stock health configuration directory etc netdata conf d health d default repeat warning default repeat critical os environment usr libexec netdata plugins d system info sh netdata system os name centos linux netdata system os id centos netdata system os id like rhel fedora netdata system os version core netdata system os version id netdata system os detection etc os release netdata system kernel name linux netdata system kernel version netdata system architecture netdata system virtualization kvm netdata system virt detection systemd detect virt netdata system container none netdata system container detection systemd detect virt netdata version ouput of netdata v netdata v netdata component name health steps to reproduce upgrade to netdata globally enable repeat notification feature as described above wait for alarm you should receive notification with raised alarm and once the the alarm is gone clear notification i didn t receive repeated notification nor clear notification expected behavior receive clear notification if alarm is gone also it should be visible in alarm log in web ui currently it s not
| 1
|
233,415
| 7,697,763,625
|
IssuesEvent
|
2018-05-18 20:02:43
|
arescentral/antares
|
https://api.github.com/repos/arescentral/antares
|
closed
|
Introduce text-based plugin format
|
Complexity:High Data Maintainability Priority:High Project:Plugins Type:Enhancement
|
The current plugin format, where entries of many types (objects, actions, scenarios…) are packed into a single binary blob and addressed by index, is brittle. It’s impossible to edit without an editor program, and the complexity of assigning and reassigning indexes makes it difficult to write an editor.
Additionally, entries of certain types (actions, conditions, briefings, and initials) exist mainly to be referenced from containing objects, and it would be nice to have them inlined in their container.
None of this is a problem in a text-based format. Let’s do that.
|
1.0
|
Introduce text-based plugin format - The current plugin format, where entries of many types (objects, actions, scenarios…) are packed into a single binary blob and addressed by index, is brittle. It’s impossible to edit without an editor program, and the complexity of assigning and reassigning indexes makes it difficult to write an editor.
Additionally, entries of certain types (actions, conditions, briefings, and initials) exist mainly to be referenced from containing objects, and it would be nice to have them inlined in their container.
None of this is a problem in a text-based format. Let’s do that.
|
priority
|
introduce text based plugin format the current plugin format where entries of many types objects actions scenarios… are packed into a single binary blob and addressed by index is brittle it’s impossible to edit without an editor program and the complexity of assigning and reassigning indexes makes it difficult to write an editor additionally entries of certain types actions conditions briefings and initials exist mainly to be referenced from containing objects and it would be nice to have them inlined in their container none of this is a problem in a text based format let’s do that
| 1
|
291,419
| 8,924,992,047
|
IssuesEvent
|
2019-01-21 20:49:04
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
food/room rate displayed not relative to # of displayed points
|
Semi-High Priority
|

Seems like we should either display skillpoints or profession points... mixing them is confusing
|
1.0
|
food/room rate displayed not relative to # of displayed points - 
Seems like we should either display skillpoints or profession points... mixing them is confusing
|
priority
|
food room rate displayed not relative to of displayed points seems like we should either display skillpoints or profession points mixing them is confusing
| 1
|
97,820
| 4,006,892,033
|
IssuesEvent
|
2016-05-12 16:15:41
|
isawnyu/pleiades-gazetteer
|
https://api.github.com/repos/isawnyu/pleiades-gazetteer
|
closed
|
16.04 staging version: history link throws error
|
bug priority: high
|
steps to reproduce:
- As an anonymous user, on the 16.04 staging version, navigate to ```/places/442473```
- Select "History" link in the byline
- Note following error message, vel sim:
> We’re sorry, but there seems to be an error…
> The error has been logged as entry number 1461692857.870.350016951558.
An authenticated user also experiences the error, but gets a traceback:
```
Traceback (innermost last):
Module ZPublisher.Publish, line 138, in publish
Module ZPublisher.mapply, line 77, in mapply
Module ZPublisher.Publish, line 48, in call_object
Module Products.Five.browser.metaconfigure, line 485, in __call__
Module Products.Five.browser.pagetemplatefile, line 125, in __call__
Module Products.Five.browser.pagetemplatefile, line 59, in __call__
Module zope.pagetemplate.pagetemplate, line 132, in pt_render
Module zope.pagetemplate.pagetemplate, line 240, in __call__
Module collective.newrelic.patches.talinterpreter, line 17, in monkeypatch
Module newrelic.api.function_trace, line 110, in literal_wrapper
Module zope.tal.talinterpreter, line 271, in __call__
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 888, in do_useMacro
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 954, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 858, in do_defineMacro
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 954, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 946, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 531, in do_optTag_tal
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 742, in do_insertStructure_tal
Module Products.PageTemplates.Expressions, line 218, in evaluateStructure
Module zope.tales.tales, line 696, in evaluate
URL: /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/history_view.pt
Line 18, Column 12
Expression: <PathExpr standard:u'here/@@contenthistory'>
Names:
{'args': (), 'container': <Place at /plone/places/442473>, 'context': <Place at /plone/places/442473>, 'default': <object object at 0x7f232f5eebc0>, 'here': <Place at /plone/places/442473>, 'loop': {}, 'nothing': None, 'options': {}, 'repeat': <Products.PageTemplates.Expressions.SafeMapping object at 0x7f230cde9788>, 'request': <HTTPRequest, URL=http://52.38.6.139/places/442473/@@historyview>, 'root': <Application at >, 'template': <Products.Five.browser.pagetemplatefile.ViewPageTemplateFile object at 0x7f230d11fe50>, 'traverse_subpath': [], 'user': <PloneUser 'thomase'>, 'view': <Products.Five.metaclass.SimpleViewClass from /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/history_view.pt object at 0x7f230d11fcd0>, 'views': <Products.Five.browser.pagetemplatefile.ViewMapper object at 0x7f230ceb8e50>}
Module zope.tales.expressions, line 217, in __call__
Module Products.PageTemplates.Expressions, line 155, in _eval
Module Products.PageTemplates.Expressions, line 117, in render
Module Products.Five.browser.metaconfigure, line 485, in __call__
Module Products.Five.browser.pagetemplatefile, line 125, in __call__
Module Products.Five.browser.pagetemplatefile, line 59, in __call__
Module zope.pagetemplate.pagetemplate, line 132, in pt_render
Module zope.pagetemplate.pagetemplate, line 240, in __call__
Module collective.newrelic.patches.talinterpreter, line 17, in monkeypatch
Module newrelic.api.function_trace, line 110, in literal_wrapper
Module zope.tal.talinterpreter, line 271, in __call__
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 583, in do_setLocal_tal
Module zope.tales.tales, line 696, in evaluate
URL: /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/content_history.pt
Line 1, Column 0
Expression: <PathExpr standard:u'view/fullHistory'>
Names:
{'args': (), 'container': <Place at /plone/places/442473>, 'context': <Place at /plone/places/442473>, 'default': <object object at 0x7f232f5eebc0>, 'here': <Place at /plone/places/442473>, 'loop': {}, 'nothing': None, 'options': {}, 'repeat': <Products.PageTemplates.Expressions.SafeMapping object at 0x7f230cde9890>, 'request': <HTTPRequest, URL=http://52.38.6.139/places/442473/@@historyview>, 'root': <Application at >, 'template': <Products.Five.browser.pagetemplatefile.ViewPageTemplateFile object at 0x7f230c99ff10>, 'traverse_subpath': [], 'user': <PloneUser 'thomase'>, 'view': <Products.Five.metaclass.contenthistory object at 0x7f230cddea90>, 'views': <Products.Five.browser.pagetemplatefile.ViewMapper object at 0x7f230ce57f90>}
Module zope.tales.expressions, line 217, in __call__
Module Products.PageTemplates.Expressions, line 155, in _eval
Module Products.PageTemplates.Expressions, line 117, in render
Module plone.app.layout.viewlets.content, line 325, in fullHistory
Module plone.app.layout.viewlets.content, line 320, in revisionHistory
Module plone.app.layout.viewlets.content, line 299, in morphVersionDataToHistoryFormat
Module Products.CMFEditions.CopyModifyMergeRepositoryTool, line 421, in isUpToDate
Module Products.CMFEditions.ArchivistTool, line 398, in isUpToDate
Module Products.CMFEditions.ZVCStorageTool, line 307, in getModificationDate
Module Products.CMFEditions.ZVCStorageTool, line 265, in retrieve
Module Products.ZopeVersionControl.Repository, line 463, in getVersionOfResource
Module Products.ZopeVersionControl.Version, line 103, in copyState
Module Products.ZopeVersionControl.Version, line 112, in stateCopy
Module Products.ZopeVersionControl.Version, line 60, in cloneByPickle
Module ZODB.broken, line 321, in __reduce__
BrokenModified: <persistent broken contentratings.storage.UserRatingStorage instance '\x00\x00\x00\x00\x01\x96ty'>
```
So, apparently, we have still not completely succeeded in eliminating the content ratings cruft as envisioned in #89?
|
1.0
|
16.04 staging version: history link throws error - steps to reproduce:
- As an anonymous user, on the 16.04 staging version, navigate to ```/places/442473```
- Select "History" link in the byline
- Note following error message, vel sim:
> We’re sorry, but there seems to be an error…
> The error has been logged as entry number 1461692857.870.350016951558.
An authenticated user also experiences the error, but gets a traceback:
```
Traceback (innermost last):
Module ZPublisher.Publish, line 138, in publish
Module ZPublisher.mapply, line 77, in mapply
Module ZPublisher.Publish, line 48, in call_object
Module Products.Five.browser.metaconfigure, line 485, in __call__
Module Products.Five.browser.pagetemplatefile, line 125, in __call__
Module Products.Five.browser.pagetemplatefile, line 59, in __call__
Module zope.pagetemplate.pagetemplate, line 132, in pt_render
Module zope.pagetemplate.pagetemplate, line 240, in __call__
Module collective.newrelic.patches.talinterpreter, line 17, in monkeypatch
Module newrelic.api.function_trace, line 110, in literal_wrapper
Module zope.tal.talinterpreter, line 271, in __call__
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 888, in do_useMacro
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 954, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 858, in do_defineMacro
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 954, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 946, in do_defineSlot
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 533, in do_optTag_tal
Module zope.tal.talinterpreter, line 518, in do_optTag
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 531, in do_optTag_tal
Module zope.tal.talinterpreter, line 513, in no_tag
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 742, in do_insertStructure_tal
Module Products.PageTemplates.Expressions, line 218, in evaluateStructure
Module zope.tales.tales, line 696, in evaluate
URL: /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/history_view.pt
Line 18, Column 12
Expression: <PathExpr standard:u'here/@@contenthistory'>
Names:
{'args': (), 'container': <Place at /plone/places/442473>, 'context': <Place at /plone/places/442473>, 'default': <object object at 0x7f232f5eebc0>, 'here': <Place at /plone/places/442473>, 'loop': {}, 'nothing': None, 'options': {}, 'repeat': <Products.PageTemplates.Expressions.SafeMapping object at 0x7f230cde9788>, 'request': <HTTPRequest, URL=http://52.38.6.139/places/442473/@@historyview>, 'root': <Application at >, 'template': <Products.Five.browser.pagetemplatefile.ViewPageTemplateFile object at 0x7f230d11fe50>, 'traverse_subpath': [], 'user': <PloneUser 'thomase'>, 'view': <Products.Five.metaclass.SimpleViewClass from /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/history_view.pt object at 0x7f230d11fcd0>, 'views': <Products.Five.browser.pagetemplatefile.ViewMapper object at 0x7f230ceb8e50>}
Module zope.tales.expressions, line 217, in __call__
Module Products.PageTemplates.Expressions, line 155, in _eval
Module Products.PageTemplates.Expressions, line 117, in render
Module Products.Five.browser.metaconfigure, line 485, in __call__
Module Products.Five.browser.pagetemplatefile, line 125, in __call__
Module Products.Five.browser.pagetemplatefile, line 59, in __call__
Module zope.pagetemplate.pagetemplate, line 132, in pt_render
Module zope.pagetemplate.pagetemplate, line 240, in __call__
Module collective.newrelic.patches.talinterpreter, line 17, in monkeypatch
Module newrelic.api.function_trace, line 110, in literal_wrapper
Module zope.tal.talinterpreter, line 271, in __call__
Module zope.tal.talinterpreter, line 343, in interpret
Module zope.tal.talinterpreter, line 583, in do_setLocal_tal
Module zope.tales.tales, line 696, in evaluate
URL: /srv/python27-apps/pleiades4/eggs/plone.app.layout-2.3.13-py2.7.egg/plone/app/layout/viewlets/content_history.pt
Line 1, Column 0
Expression: <PathExpr standard:u'view/fullHistory'>
Names:
{'args': (), 'container': <Place at /plone/places/442473>, 'context': <Place at /plone/places/442473>, 'default': <object object at 0x7f232f5eebc0>, 'here': <Place at /plone/places/442473>, 'loop': {}, 'nothing': None, 'options': {}, 'repeat': <Products.PageTemplates.Expressions.SafeMapping object at 0x7f230cde9890>, 'request': <HTTPRequest, URL=http://52.38.6.139/places/442473/@@historyview>, 'root': <Application at >, 'template': <Products.Five.browser.pagetemplatefile.ViewPageTemplateFile object at 0x7f230c99ff10>, 'traverse_subpath': [], 'user': <PloneUser 'thomase'>, 'view': <Products.Five.metaclass.contenthistory object at 0x7f230cddea90>, 'views': <Products.Five.browser.pagetemplatefile.ViewMapper object at 0x7f230ce57f90>}
Module zope.tales.expressions, line 217, in __call__
Module Products.PageTemplates.Expressions, line 155, in _eval
Module Products.PageTemplates.Expressions, line 117, in render
Module plone.app.layout.viewlets.content, line 325, in fullHistory
Module plone.app.layout.viewlets.content, line 320, in revisionHistory
Module plone.app.layout.viewlets.content, line 299, in morphVersionDataToHistoryFormat
Module Products.CMFEditions.CopyModifyMergeRepositoryTool, line 421, in isUpToDate
Module Products.CMFEditions.ArchivistTool, line 398, in isUpToDate
Module Products.CMFEditions.ZVCStorageTool, line 307, in getModificationDate
Module Products.CMFEditions.ZVCStorageTool, line 265, in retrieve
Module Products.ZopeVersionControl.Repository, line 463, in getVersionOfResource
Module Products.ZopeVersionControl.Version, line 103, in copyState
Module Products.ZopeVersionControl.Version, line 112, in stateCopy
Module Products.ZopeVersionControl.Version, line 60, in cloneByPickle
Module ZODB.broken, line 321, in __reduce__
BrokenModified: <persistent broken contentratings.storage.UserRatingStorage instance '\x00\x00\x00\x00\x01\x96ty'>
```
So, apparently, we have still not completely succeeded in eliminating the content ratings cruft as envisioned in #89?
|
priority
|
staging version history link throws error steps to reproduce as an anonymous user on the staging version navigate to places select history link in the byline note following error message vel sim we’re sorry but there seems to be an error… the error has been logged as entry number an authenticated user also experiences the error but gets a traceback traceback innermost last module zpublisher publish line in publish module zpublisher mapply line in mapply module zpublisher publish line in call object module products five browser metaconfigure line in call module products five browser pagetemplatefile line in call module products five browser pagetemplatefile line in call module zope pagetemplate pagetemplate line in pt render module zope pagetemplate pagetemplate line in call module collective newrelic patches talinterpreter line in monkeypatch module newrelic api function trace line in literal wrapper module zope tal talinterpreter line in call module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do usemacro module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do opttag tal module zope tal talinterpreter line in do opttag module zope tal talinterpreter line in no tag module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do defineslot module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do opttag tal module zope tal talinterpreter line in do opttag module zope tal talinterpreter line in no tag module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do definemacro module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do defineslot module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do opttag tal module zope tal talinterpreter line in do opttag module zope tal talinterpreter line in no tag module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do defineslot module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do opttag tal module zope tal talinterpreter line in do opttag module zope tal talinterpreter line in no tag module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do opttag tal module zope tal talinterpreter line in no tag module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do insertstructure tal module products pagetemplates expressions line in evaluatestructure module zope tales tales line in evaluate url srv apps eggs plone app layout egg plone app layout viewlets history view pt line column expression names args container context default here loop nothing none options repeat request template traverse subpath user view views module zope tales expressions line in call module products pagetemplates expressions line in eval module products pagetemplates expressions line in render module products five browser metaconfigure line in call module products five browser pagetemplatefile line in call module products five browser pagetemplatefile line in call module zope pagetemplate pagetemplate line in pt render module zope pagetemplate pagetemplate line in call module collective newrelic patches talinterpreter line in monkeypatch module newrelic api function trace line in literal wrapper module zope tal talinterpreter line in call module zope tal talinterpreter line in interpret module zope tal talinterpreter line in do setlocal tal module zope tales tales line in evaluate url srv apps eggs plone app layout egg plone app layout viewlets content history pt line column expression names args container context default here loop nothing none options repeat request template traverse subpath user view views module zope tales expressions line in call module products pagetemplates expressions line in eval module products pagetemplates expressions line in render module plone app layout viewlets content line in fullhistory module plone app layout viewlets content line in revisionhistory module plone app layout viewlets content line in morphversiondatatohistoryformat module products cmfeditions copymodifymergerepositorytool line in isuptodate module products cmfeditions archivisttool line in isuptodate module products cmfeditions zvcstoragetool line in getmodificationdate module products cmfeditions zvcstoragetool line in retrieve module products zopeversioncontrol repository line in getversionofresource module products zopeversioncontrol version line in copystate module products zopeversioncontrol version line in statecopy module products zopeversioncontrol version line in clonebypickle module zodb broken line in reduce brokenmodified so apparently we have still not completely succeeded in eliminating the content ratings cruft as envisioned in
| 1
|
236,245
| 7,747,818,113
|
IssuesEvent
|
2018-05-30 05:49:25
|
llmhyy/microbat
|
https://api.github.com/repos/llmhyy/microbat
|
closed
|
[Instrumentation] Agent Crash
|
high priority
|
hi @lylytran
The agent crash with the new features. Please kindly check the following cmd:
```
E:\linyun\bug_repo\jdk1.7.0_800\bin\java -Xmx30g -XX:+UseG1GC -ea -noverify -javaagent:E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\instrumentator.jar=precheck=true,excludes=java.util.Properties*,java_home=E:\linyun\bug_repo\jdk1.7.0_800,stepLimit=100000,log=printProgress;info;error;debug,working_dir=E:\linyun\bug_repo\Math\30\bug,varLayer=1,launch_class=org.apache.commons.math3.stat.inference.MannWhitneyUTestTest,class_path=E:\linyun\bug_repo\Math\30\bug\target\test-classes;E:\linyun\bug_repo\Math\30\bug\target\classes;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar,dump_file_path=C:\Users\linyun\AppData\Local\Temp\tracePrecheck4467422854665248237.info -cp E:\linyun\bug_repo\Math\30\bug\target\test-classes;E:\linyun\bug_repo\Math\30\bug\target\classes;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar microbat.evaluation.junit.MicroBatTestRunner org.apache.commons.math3.stat.inference.MannWhitneyUTestTest testBigDataSet
```
Thanks!
|
1.0
|
[Instrumentation] Agent Crash - hi @lylytran
The agent crash with the new features. Please kindly check the following cmd:
```
E:\linyun\bug_repo\jdk1.7.0_800\bin\java -Xmx30g -XX:+UseG1GC -ea -noverify -javaagent:E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\instrumentator.jar=precheck=true,excludes=java.util.Properties*,java_home=E:\linyun\bug_repo\jdk1.7.0_800,stepLimit=100000,log=printProgress;info;error;debug,working_dir=E:\linyun\bug_repo\Math\30\bug,varLayer=1,launch_class=org.apache.commons.math3.stat.inference.MannWhitneyUTestTest,class_path=E:\linyun\bug_repo\Math\30\bug\target\test-classes;E:\linyun\bug_repo\Math\30\bug\target\classes;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar,dump_file_path=C:\Users\linyun\AppData\Local\Temp\tracePrecheck4467422854665248237.info -cp E:\linyun\bug_repo\Math\30\bug\target\test-classes;E:\linyun\bug_repo\Math\30\bug\target\classes;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar microbat.evaluation.junit.MicroBatTestRunner org.apache.commons.math3.stat.inference.MannWhitneyUTestTest testBigDataSet
```
Thanks!
|
priority
|
agent crash hi lylytran the agent crash with the new features please kindly check the following cmd e linyun bug repo bin java xx ea noverify javaagent e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib instrumentator jar precheck true excludes java util properties java home e linyun bug repo steplimit log printprogress info error debug working dir e linyun bug repo math bug varlayer launch class org apache commons stat inference mannwhitneyutesttest class path e linyun bug repo math bug target test classes e linyun bug repo math bug target classes e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib junit jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib org hamcrest core jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib testrunner jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib bcel jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib javassist jar dump file path c users linyun appdata local temp info cp e linyun bug repo math bug target test classes e linyun bug repo math bug target classes e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib junit jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib org hamcrest core jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib testrunner jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib bcel jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib javassist jar microbat evaluation junit microbattestrunner org apache commons stat inference mannwhitneyutesttest testbigdataset thanks
| 1
|
482,138
| 13,901,241,142
|
IssuesEvent
|
2020-10-20 02:26:19
|
Seneca-CDOT/telescope
|
https://api.github.com/repos/Seneca-CDOT/telescope
|
closed
|
'Search' returns latest posts instead of results for searched input.
|
Priority: High area: front-end type: bug
|
<!-- Please use this template while reporting a bug and provide as much info as possible. Thanks!
-->
**What happened**:
I entered a string in the search box and it returned the latest posts instead of actual results

**What should have happened**:
I should have seen posts related to the entered string
**How to reproduce it (as precise as possible)**:
Go to https://dev.telescope.cdot.systems/search, enter any string in the search box and compare results to latest posts in `/`
|
1.0
|
'Search' returns latest posts instead of results for searched input. - <!-- Please use this template while reporting a bug and provide as much info as possible. Thanks!
-->
**What happened**:
I entered a string in the search box and it returned the latest posts instead of actual results

**What should have happened**:
I should have seen posts related to the entered string
**How to reproduce it (as precise as possible)**:
Go to https://dev.telescope.cdot.systems/search, enter any string in the search box and compare results to latest posts in `/`
|
priority
|
search returns latest posts instead of results for searched input please use this template while reporting a bug and provide as much info as possible thanks what happened i entered a string in the search box and it returned the latest posts instead of actual results what should have happened i should have seen posts related to the entered string how to reproduce it as precise as possible go to enter any string in the search box and compare results to latest posts in
| 1
|
491,716
| 14,169,806,260
|
IssuesEvent
|
2020-11-12 13:44:30
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Compiler generated names clashes with user defined names with quoted identifiers
|
Area/Desugar Priority/High SwanLakeDump Team/CompilerFE Type/Bug
|
**Description:**
As we are now supporting quoted identifiers in Ballerina, there is a possibility that the user defined identifiers can have similar names as the identifiers generated at the compiler. The generated names need to be fixed to solve this overloading.
Eg. -
The following code checks the generated anon type name `$anonType$...` with a user defined type. This gives a compilation error as follows.
```ballerina
import ballerina/io;
public type Person object {
public int age = 0;
public string name = "";
public function getName() returns string {
return "Person Name";
}
};
type '\$anonType\$0 object {
string name;
function init(string name) {
self.name = name;
}
};
function testAbstractAnonObjectInVarDef() {
abstract object { public function getName() returns string;} p = new Person();
io:println(getName(p));
}
public function main() {
'\$anonType\$0 x = new ("Jack") ;
io:println(x);
testAbstractAnonObjectInVarDef();
}
```
The error received:
```
error: .::initTest.bal:25:23: redeclared symbol '$anonType$0'
error: .::initTest.bal:25:70: incompatible types: expected 'object { string name; }', found 'Person'
error: .::initTest.bal:26:16: undefined function 'getName'
```
**Affected Versions:**
slp2
**Related Issues (optional):**
#18720
**Suggested Labels (optional):**
**Suggested Assignees (optional):**
|
1.0
|
Compiler generated names clashes with user defined names with quoted identifiers - **Description:**
As we are now supporting quoted identifiers in Ballerina, there is a possibility that the user defined identifiers can have similar names as the identifiers generated at the compiler. The generated names need to be fixed to solve this overloading.
Eg. -
The following code checks the generated anon type name `$anonType$...` with a user defined type. This gives a compilation error as follows.
```ballerina
import ballerina/io;
public type Person object {
public int age = 0;
public string name = "";
public function getName() returns string {
return "Person Name";
}
};
type '\$anonType\$0 object {
string name;
function init(string name) {
self.name = name;
}
};
function testAbstractAnonObjectInVarDef() {
abstract object { public function getName() returns string;} p = new Person();
io:println(getName(p));
}
public function main() {
'\$anonType\$0 x = new ("Jack") ;
io:println(x);
testAbstractAnonObjectInVarDef();
}
```
The error received:
```
error: .::initTest.bal:25:23: redeclared symbol '$anonType$0'
error: .::initTest.bal:25:70: incompatible types: expected 'object { string name; }', found 'Person'
error: .::initTest.bal:26:16: undefined function 'getName'
```
**Affected Versions:**
slp2
**Related Issues (optional):**
#18720
**Suggested Labels (optional):**
**Suggested Assignees (optional):**
|
priority
|
compiler generated names clashes with user defined names with quoted identifiers description as we are now supporting quoted identifiers in ballerina there is a possibility that the user defined identifiers can have similar names as the identifiers generated at the compiler the generated names need to be fixed to solve this overloading eg the following code checks the generated anon type name anontype with a user defined type this gives a compilation error as follows ballerina import ballerina io public type person object public int age public string name public function getname returns string return person name type anontype object string name function init string name self name name function testabstractanonobjectinvardef abstract object public function getname returns string p new person io println getname p public function main anontype x new jack io println x testabstractanonobjectinvardef the error received error inittest bal redeclared symbol anontype error inittest bal incompatible types expected object string name found person error inittest bal undefined function getname affected versions related issues optional suggested labels optional suggested assignees optional
| 1
|
526,432
| 15,288,241,258
|
IssuesEvent
|
2021-02-23 16:38:53
|
bennyboer/bbb_app
|
https://api.github.com/repos/bennyboer/bbb_app
|
opened
|
When muting the user still appears as unmuted in the web app
|
bug high priority size: small
|
When muting the user still appears as unmuted in the web app although the user really is muted and just appears not to be.
|
1.0
|
When muting the user still appears as unmuted in the web app - When muting the user still appears as unmuted in the web app although the user really is muted and just appears not to be.
|
priority
|
when muting the user still appears as unmuted in the web app when muting the user still appears as unmuted in the web app although the user really is muted and just appears not to be
| 1
|
675,825
| 23,107,597,333
|
IssuesEvent
|
2022-07-27 10:08:41
|
wfau/aglais
|
https://api.github.com/repos/wfau/aglais
|
closed
|
Add @url handler for public key
|
high priority live-20220725 20220726-zrq-config-settings
|
Add handler to resolve a URL to load the user's public key.
```
publickey: "@http://... "
```
```
publickey: "@file://... "
```
|
1.0
|
Add @url handler for public key - Add handler to resolve a URL to load the user's public key.
```
publickey: "@http://... "
```
```
publickey: "@file://... "
```
|
priority
|
add url handler for public key add handler to resolve a url to load the user s public key publickey publickey file
| 1
|
632,869
| 20,237,450,091
|
IssuesEvent
|
2022-02-14 04:43:43
|
CheongYeeMing/Neontera2
|
https://api.github.com/repos/CheongYeeMing/Neontera2
|
opened
|
Duplication of Items when Save and Load
|
bug must fix priority.High
|
All items are duplicated when the player saves and loads data.

|
1.0
|
Duplication of Items when Save and Load - All items are duplicated when the player saves and loads data.

|
priority
|
duplication of items when save and load all items are duplicated when the player saves and loads data
| 1
|
751,983
| 26,268,684,453
|
IssuesEvent
|
2023-01-06 14:58:42
|
uedar/mongodb-prototype
|
https://api.github.com/repos/uedar/mongodb-prototype
|
closed
|
Embed jsmol in detail page
|
front-end high priority
|
- download and use jsmol source (gitignore)
- embed to script tag
- react-helmet?
Other visualization tools are ok
|
1.0
|
Embed jsmol in detail page - - download and use jsmol source (gitignore)
- embed to script tag
- react-helmet?
Other visualization tools are ok
|
priority
|
embed jsmol in detail page download and use jsmol source gitignore embed to script tag react helmet other visualization tools are ok
| 1
|
692,345
| 23,730,576,799
|
IssuesEvent
|
2022-08-31 01:05:25
|
City-Bureau/city-scrapers-atl
|
https://api.github.com/repos/City-Bureau/city-scrapers-atl
|
closed
|
New Scraper: Cobb County Elections and Registration
|
priority-high
|
Create a new scraper for Cobb County Elections and Registration
Website: https://www.cobbcounty.org/events?field_section_target_id=All&field_event_category_target_id=195&field_event_date_recur_value_2=&field_event_date_recur_end_value=
Jurisdiction: Cobb County
Classification: Elections
Board is empowered with all powers and duties relating to the conduct of elections as election superintendents and the registration of voters and absentee balloting procedures as board of registrars as provided by Ga. Law.
|
1.0
|
New Scraper: Cobb County Elections and Registration - Create a new scraper for Cobb County Elections and Registration
Website: https://www.cobbcounty.org/events?field_section_target_id=All&field_event_category_target_id=195&field_event_date_recur_value_2=&field_event_date_recur_end_value=
Jurisdiction: Cobb County
Classification: Elections
Board is empowered with all powers and duties relating to the conduct of elections as election superintendents and the registration of voters and absentee balloting procedures as board of registrars as provided by Ga. Law.
|
priority
|
new scraper cobb county elections and registration create a new scraper for cobb county elections and registration website jurisdiction cobb county classification elections board is empowered with all powers and duties relating to the conduct of elections as election superintendents and the registration of voters and absentee balloting procedures as board of registrars as provided by ga law
| 1
|
383,289
| 11,353,996,898
|
IssuesEvent
|
2020-01-24 16:38:59
|
shijima666/blood-steel-maid
|
https://api.github.com/repos/shijima666/blood-steel-maid
|
closed
|
褒める文言に「りんごー!」を加えたとき、2つのイベントが走る
|
Priority: High Type: Bug
|
「メイナちゃん、すごいリンゴー!」と言った場合、『褒めたら照れる』イベントと『リンゴでテンションが上がる』イベントが同時に走るため、応答が不自然になる
そのため、『褒めたら照れる』イベントにリンゴを除外する正規表現を追加する必要がある
|
1.0
|
褒める文言に「りんごー!」を加えたとき、2つのイベントが走る - 「メイナちゃん、すごいリンゴー!」と言った場合、『褒めたら照れる』イベントと『リンゴでテンションが上がる』イベントが同時に走るため、応答が不自然になる
そのため、『褒めたら照れる』イベントにリンゴを除外する正規表現を追加する必要がある
|
priority
|
褒める文言に「りんごー!」を加えたとき、 「メイナちゃん、すごいリンゴー!」と言った場合、『褒めたら照れる』イベントと『リンゴでテンションが上がる』イベントが同時に走るため、応答が不自然になる そのため、『褒めたら照れる』イベントにリンゴを除外する正規表現を追加する必要がある
| 1
|
662,761
| 22,152,305,598
|
IssuesEvent
|
2022-06-03 18:16:09
|
RAF-SI-2021/Banka-Back
|
https://api.github.com/repos/RAF-SI-2021/Banka-Back
|
closed
|
Odobravanje porudžbina
|
priority/high area/backend
|
Porudžbinu je potrebno odobriti ukoliko je incirana od strane usera koji je Agent, u sledećim slučajevima:
* User ima setovan flag “Zahtevati odobravanje svake transakcije” na True
* User je iskoristio svoj dodeljeni limit
* Order po svojoj ukupnoj trenutni ceni prelazi agentov dodeljeni limit
Porudžbine odobrava user koji je Supervizor.
|
1.0
|
Odobravanje porudžbina - Porudžbinu je potrebno odobriti ukoliko je incirana od strane usera koji je Agent, u sledećim slučajevima:
* User ima setovan flag “Zahtevati odobravanje svake transakcije” na True
* User je iskoristio svoj dodeljeni limit
* Order po svojoj ukupnoj trenutni ceni prelazi agentov dodeljeni limit
Porudžbine odobrava user koji je Supervizor.
|
priority
|
odobravanje porudžbina porudžbinu je potrebno odobriti ukoliko je incirana od strane usera koji je agent u sledećim slučajevima user ima setovan flag “zahtevati odobravanje svake transakcije” na true user je iskoristio svoj dodeljeni limit order po svojoj ukupnoj trenutni ceni prelazi agentov dodeljeni limit porudžbine odobrava user koji je supervizor
| 1
|
250,424
| 7,976,818,512
|
IssuesEvent
|
2018-07-17 13:47:14
|
Ecam-Eurobot/Eurobot-2019
|
https://api.github.com/repos/Ecam-Eurobot/Eurobot-2019
|
opened
|
Tracking issue: Moving base
|
Priority: High Status: Discussion
|
# Project "Moving base"
This is a tracking issue for the "Moving base" project. This issue can be used to discuss the project, track progress, etc. Specific items should not be discussed here, but on the appropriate sub-issues.
## Proposal
...
## To Do
...
-----
*This issue will be updated periodically to reflect progress*
|
1.0
|
Tracking issue: Moving base - # Project "Moving base"
This is a tracking issue for the "Moving base" project. This issue can be used to discuss the project, track progress, etc. Specific items should not be discussed here, but on the appropriate sub-issues.
## Proposal
...
## To Do
...
-----
*This issue will be updated periodically to reflect progress*
|
priority
|
tracking issue moving base project moving base this is a tracking issue for the moving base project this issue can be used to discuss the project track progress etc specific items should not be discussed here but on the appropriate sub issues proposal to do this issue will be updated periodically to reflect progress
| 1
|
347,916
| 10,436,404,528
|
IssuesEvent
|
2019-09-17 19:29:28
|
nim-lang/Nim
|
https://api.github.com/repos/nim-lang/Nim
|
closed
|
CountTable is not reliable.
|
High Priority Stdlib
|
Today I got a very strange output from my app. After digging, I found the bug is in the CountTable.
### Example
```nim
import tables
var ct = initCountTable[int]()
ct.inc(130, 1)
ct.inc(132, 1)
ct.inc(258, 1)
ct.inc(131, 100)
echo ct[131] # should output 100
ct.inc(132, -1)
echo ct[131] # should output 100, too
```
### Current Output
```
100
0
```
### Expected Output
```
100
100
```
### Additional Information
```
Nim Compiler Version 0.20.99 [Windows: amd64]
```
|
1.0
|
CountTable is not reliable. - Today I got a very strange output from my app. After digging, I found the bug is in the CountTable.
### Example
```nim
import tables
var ct = initCountTable[int]()
ct.inc(130, 1)
ct.inc(132, 1)
ct.inc(258, 1)
ct.inc(131, 100)
echo ct[131] # should output 100
ct.inc(132, -1)
echo ct[131] # should output 100, too
```
### Current Output
```
100
0
```
### Expected Output
```
100
100
```
### Additional Information
```
Nim Compiler Version 0.20.99 [Windows: amd64]
```
|
priority
|
counttable is not reliable today i got a very strange output from my app after digging i found the bug is in the counttable example nim import tables var ct initcounttable ct inc ct inc ct inc ct inc echo ct should output ct inc echo ct should output too current output expected output additional information nim compiler version
| 1
|
389,875
| 11,518,604,749
|
IssuesEvent
|
2020-02-14 10:51:53
|
iteahome/team-project-online-shop
|
https://api.github.com/repos/iteahome/team-project-online-shop
|
closed
|
Fix Login bug
|
High_Priority bug
|
Login doesn't loop. It only accepts credentials from the first user. For all other users, it throws the credentials exception message from LoginUI.java:42
|
1.0
|
Fix Login bug - Login doesn't loop. It only accepts credentials from the first user. For all other users, it throws the credentials exception message from LoginUI.java:42
|
priority
|
fix login bug login doesn t loop it only accepts credentials from the first user for all other users it throws the credentials exception message from loginui java
| 1
|
659,570
| 21,933,582,612
|
IssuesEvent
|
2022-05-23 11:59:08
|
woocommerce/woocommerce-blocks
|
https://api.github.com/repos/woocommerce/woocommerce-blocks
|
closed
|
Cart inner blocks can be removed if unlocking them with the new WP 6.0 UI
|
type: bug ◼️ block: cart priority: high
|
## Describe the bug
WP 6.0 introduces a UI to lock/unlock blocks. That makes it possible to remove Cart inner blocks if the user previously unlocks them.
## To reproduce
1. Install WordPress beta tester plugin and update to WP 6.0 beta 2.
2. Add the Cart block to a post or page.
3. Unlock the Cart Line Items, Cart Items block and Filled Cart blocks.
4. Select the Cart Items block.
5. Press <kbd>Del</kbd> or delete the block using the UI.
6. Notice the Cart Items block could be removed.
https://user-images.githubusercontent.com/3616980/165117738-ac2218da-50aa-4194-9789-0db82e104b2c.mp4
## Expected behavior
It shouldn't be possible to remove the Cart Items block.
## More info
Gutenberg seems to have introduced a `supports.__experimentalLock` flag that might be useful to prevent this (https://github.com/WordPress/gutenberg/pull/39568).
I couldn't reproduce it with the Checkout block, but might be worth testing.
|
1.0
|
Cart inner blocks can be removed if unlocking them with the new WP 6.0 UI - ## Describe the bug
WP 6.0 introduces a UI to lock/unlock blocks. That makes it possible to remove Cart inner blocks if the user previously unlocks them.
## To reproduce
1. Install WordPress beta tester plugin and update to WP 6.0 beta 2.
2. Add the Cart block to a post or page.
3. Unlock the Cart Line Items, Cart Items block and Filled Cart blocks.
4. Select the Cart Items block.
5. Press <kbd>Del</kbd> or delete the block using the UI.
6. Notice the Cart Items block could be removed.
https://user-images.githubusercontent.com/3616980/165117738-ac2218da-50aa-4194-9789-0db82e104b2c.mp4
## Expected behavior
It shouldn't be possible to remove the Cart Items block.
## More info
Gutenberg seems to have introduced a `supports.__experimentalLock` flag that might be useful to prevent this (https://github.com/WordPress/gutenberg/pull/39568).
I couldn't reproduce it with the Checkout block, but might be worth testing.
|
priority
|
cart inner blocks can be removed if unlocking them with the new wp ui describe the bug wp introduces a ui to lock unlock blocks that makes it possible to remove cart inner blocks if the user previously unlocks them to reproduce install wordpress beta tester plugin and update to wp beta add the cart block to a post or page unlock the cart line items cart items block and filled cart blocks select the cart items block press del or delete the block using the ui notice the cart items block could be removed expected behavior it shouldn t be possible to remove the cart items block more info gutenberg seems to have introduced a supports experimentallock flag that might be useful to prevent this i couldn t reproduce it with the checkout block but might be worth testing
| 1
|
643,568
| 20,959,844,366
|
IssuesEvent
|
2022-03-27 16:30:28
|
status-im/status-desktop
|
https://api.github.com/repos/status-im/status-desktop
|
opened
|
after unblocking a user, messages are still not received until app is restarted
|
bug Chat priority 1: high
|
# Bug Report
## Steps to reproduce
1. aadd a mutual contact request
2. after send messages to each other
3. block the user A
4. have user A send messages -> nothing is displayed as expected
5. unblock user A
6. have user A send messages
#### Expected behavior
new messages from user A since the unblocking should be displayed
#### Actual behavior
new messages are not displayed
|
1.0
|
after unblocking a user, messages are still not received until app is restarted - # Bug Report
## Steps to reproduce
1. aadd a mutual contact request
2. after send messages to each other
3. block the user A
4. have user A send messages -> nothing is displayed as expected
5. unblock user A
6. have user A send messages
#### Expected behavior
new messages from user A since the unblocking should be displayed
#### Actual behavior
new messages are not displayed
|
priority
|
after unblocking a user messages are still not received until app is restarted bug report steps to reproduce aadd a mutual contact request after send messages to each other block the user a have user a send messages nothing is displayed as expected unblock user a have user a send messages expected behavior new messages from user a since the unblocking should be displayed actual behavior new messages are not displayed
| 1
|
576,448
| 17,087,146,013
|
IssuesEvent
|
2021-07-08 13:15:37
|
bd-R/bdclean
|
https://api.github.com/repos/bd-R/bdclean
|
closed
|
Verify: bdchecks package has high file size
|
discussion high-priority
|
**Describe the bug**
Verify if bdchecks has unreasonably high package size
**To Reproduce**
Download repo and verify, build executable and verify
|
1.0
|
Verify: bdchecks package has high file size - **Describe the bug**
Verify if bdchecks has unreasonably high package size
**To Reproduce**
Download repo and verify, build executable and verify
|
priority
|
verify bdchecks package has high file size describe the bug verify if bdchecks has unreasonably high package size to reproduce download repo and verify build executable and verify
| 1
|
126,876
| 5,007,020,953
|
IssuesEvent
|
2016-12-12 15:43:47
|
w3c/webpayments-method-identifiers
|
https://api.github.com/repos/w3c/webpayments-method-identifiers
|
closed
|
Should we define nesting/grouping semantics for payment method identifier matching?
|
help wanted Priority: High question
|
Migrated from https://github.com/w3c/browser-payment-api/issues/30 by @ianbjacobs
There is a generally interesting question of how we will define payment method identifier matching. There are many possibilities, ranging from exact string equivalence, to complex URI matching schemes, to regular expressions.
This particular issue is focused on a suggestion from the 24 Feb FTF meeting [1]: should there be a way to define a sort of hierarchy so that one payment method identifier will match more than one other identifiers. (e.g., "match any payment method identifier related to Visa").
(If people want to raise issues for other matching semantics, please create new issues.)
[1] https://www.w3.org/2016/02/24-wpwg-minutes
|
1.0
|
Should we define nesting/grouping semantics for payment method identifier matching? - Migrated from https://github.com/w3c/browser-payment-api/issues/30 by @ianbjacobs
There is a generally interesting question of how we will define payment method identifier matching. There are many possibilities, ranging from exact string equivalence, to complex URI matching schemes, to regular expressions.
This particular issue is focused on a suggestion from the 24 Feb FTF meeting [1]: should there be a way to define a sort of hierarchy so that one payment method identifier will match more than one other identifiers. (e.g., "match any payment method identifier related to Visa").
(If people want to raise issues for other matching semantics, please create new issues.)
[1] https://www.w3.org/2016/02/24-wpwg-minutes
|
priority
|
should we define nesting grouping semantics for payment method identifier matching migrated from by ianbjacobs there is a generally interesting question of how we will define payment method identifier matching there are many possibilities ranging from exact string equivalence to complex uri matching schemes to regular expressions this particular issue is focused on a suggestion from the feb ftf meeting should there be a way to define a sort of hierarchy so that one payment method identifier will match more than one other identifiers e g match any payment method identifier related to visa if people want to raise issues for other matching semantics please create new issues
| 1
|
347,811
| 10,434,497,740
|
IssuesEvent
|
2019-09-17 15:20:36
|
Deepomatic/dmake
|
https://api.github.com/repos/Deepomatic/dmake
|
closed
|
Dmake fail when there is a `#` or `/` in the branch name
|
bug high priority
|
Docker gives a ``` invalid reference format ``` when launching Dmake in a branch with a ```#``` in the name.
|
1.0
|
Dmake fail when there is a `#` or `/` in the branch name - Docker gives a ``` invalid reference format ``` when launching Dmake in a branch with a ```#``` in the name.
|
priority
|
dmake fail when there is a or in the branch name docker gives a invalid reference format when launching dmake in a branch with a in the name
| 1
|
2,953
| 2,534,674,825
|
IssuesEvent
|
2015-01-25 06:45:02
|
bethlakshmi/GBE2
|
https://api.github.com/repos/bethlakshmi/GBE2
|
closed
|
Schedule items should be deletable
|
High Priority
|
A scheduler should be able to delete a schedule item from the lists. It can be used from the list for the given event type (Generic, Show, Class). When it goes, it kills off:
- all resource allocations for the scheduler.event
- ?volunteer opportunities for the event?
If we don't kill the volunteer opportunities, they are floating in limbo, and volunteers will appear to be assigned to them.
The counter concern would be that we eliminate a whole lot of work when we do that... but I don't like having too much dead data around - it just leads to confusion.
|
1.0
|
Schedule items should be deletable - A scheduler should be able to delete a schedule item from the lists. It can be used from the list for the given event type (Generic, Show, Class). When it goes, it kills off:
- all resource allocations for the scheduler.event
- ?volunteer opportunities for the event?
If we don't kill the volunteer opportunities, they are floating in limbo, and volunteers will appear to be assigned to them.
The counter concern would be that we eliminate a whole lot of work when we do that... but I don't like having too much dead data around - it just leads to confusion.
|
priority
|
schedule items should be deletable a scheduler should be able to delete a schedule item from the lists it can be used from the list for the given event type generic show class when it goes it kills off all resource allocations for the scheduler event volunteer opportunities for the event if we don t kill the volunteer opportunities they are floating in limbo and volunteers will appear to be assigned to them the counter concern would be that we eliminate a whole lot of work when we do that but i don t like having too much dead data around it just leads to confusion
| 1
|
527,027
| 15,307,314,599
|
IssuesEvent
|
2021-02-24 20:43:09
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
closed
|
courses hard to find in the doc
|
priority:HIGH
|

--> not intuitive.
we should move the "course" somewhere else
|
1.0
|
courses hard to find in the doc - 
--> not intuitive.
we should move the "course" somewhere else
|
priority
|
courses hard to find in the doc not intuitive we should move the course somewhere else
| 1
|
656,267
| 21,725,175,036
|
IssuesEvent
|
2022-05-11 06:54:40
|
yuukiyamagata/SANKOUSHO-LABO
|
https://api.github.com/repos/yuukiyamagata/SANKOUSHO-LABO
|
closed
|
[機能] firebase Storageで画像を保存する
|
Priority High
|
## 実装内容の説明
- firebase Storageで画像を保存する
## 実装予定のページ・コンポーネントのデザイン画像もしくはURL
## TASK
- [ ] firebase Storageで画像を保存する機能を実装
- [ ] ログイン初期表示の画像の保存をする
- [ ] 投稿画面で画像をダウンロードできるようにする
## その他・備考
-
## その他・備考
-
|
1.0
|
[機能] firebase Storageで画像を保存する - ## 実装内容の説明
- firebase Storageで画像を保存する
## 実装予定のページ・コンポーネントのデザイン画像もしくはURL
## TASK
- [ ] firebase Storageで画像を保存する機能を実装
- [ ] ログイン初期表示の画像の保存をする
- [ ] 投稿画面で画像をダウンロードできるようにする
## その他・備考
-
## その他・備考
-
|
priority
|
firebase storageで画像を保存する 実装内容の説明 firebase storageで画像を保存する 実装予定のページ・コンポーネントのデザイン画像もしくはurl task firebase storageで画像を保存する機能を実装 ログイン初期表示の画像の保存をする 投稿画面で画像をダウンロードできるようにする その他・備考 その他・備考
| 1
|
281,851
| 8,700,403,816
|
IssuesEvent
|
2018-12-05 08:37:59
|
medic/medic-webapp
|
https://api.github.com/repos/medic/medic-webapp
|
opened
|
The released API bundle is missing shared-libs
|
Priority: 1 - High Type: Bug
|
We use npm pack to zip up api and sentinel and all their bundledDependencies so when they get deployed on the server they just work. Currently when deploying the `medic:medic:master` release, API won't start due to a missing dependency on `@shared-libs/server-checks`. The [`medic:medic:master`](https://staging.dev.medicmobile.org/_couch/_utils/document.html?builds/medic%3Amedic%3Amaster) release on the market is missing this node_module, but it exists in the [`medic:medic:3.3.x`](https://staging.dev.medicmobile.org/_couch/_utils/document.html?builds/medic%3Amedic%3A3.3.x) release.
I suspect it has something to do with replacing [yarn with npm](https://github.com/medic/medic-webapp/issues/4814) which was done recently so start there.
|
1.0
|
The released API bundle is missing shared-libs - We use npm pack to zip up api and sentinel and all their bundledDependencies so when they get deployed on the server they just work. Currently when deploying the `medic:medic:master` release, API won't start due to a missing dependency on `@shared-libs/server-checks`. The [`medic:medic:master`](https://staging.dev.medicmobile.org/_couch/_utils/document.html?builds/medic%3Amedic%3Amaster) release on the market is missing this node_module, but it exists in the [`medic:medic:3.3.x`](https://staging.dev.medicmobile.org/_couch/_utils/document.html?builds/medic%3Amedic%3A3.3.x) release.
I suspect it has something to do with replacing [yarn with npm](https://github.com/medic/medic-webapp/issues/4814) which was done recently so start there.
|
priority
|
the released api bundle is missing shared libs we use npm pack to zip up api and sentinel and all their bundleddependencies so when they get deployed on the server they just work currently when deploying the medic medic master release api won t start due to a missing dependency on shared libs server checks the release on the market is missing this node module but it exists in the release i suspect it has something to do with replacing which was done recently so start there
| 1
|
321,740
| 9,808,228,342
|
IssuesEvent
|
2019-06-12 15:12:42
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
opened
|
[Self Registration REST API] Wrong Content-Type in Register User successful response
|
Complexity/Low Component/Identity Governance Priority/High Severity/Major Type/Bug
|
Moved from: https://wso2.org/jira/browse/IDENTITY-7373
Hi,
I'm using the Self Registration REST API (v0.9) of Identity Server 5.3.0.
This is what I added in <IS_HOME>/repository/conf/identity/identity.xml:
```
<SelfRegistration>
<Enable>true</Enable>
<LockOnCreation>true</LockOnCreation>
<Notification>
<InternallyManage>false</InternallyManage>
</Notification>
<ReCaptcha>false</ReCaptcha>
</SelfRegistration>
```
The important thing here is that I have disable the internal management of the notifications because I want to manage it myself.
So when I use the Register User (https://localhost:9443/api/identity/user/v0.9/me) resource of the API it returns the Validation Code so I can handle it. But the Content-Type of the response is application/json and the body is a plain text. Here is an example:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 07:10:42 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
8ff16bfe-64d5-43a4-95b7-d9d2285a4936
```
When there is an error all goes well because in that case the Content-Type is application/json and the body is a JSON:
```
HTTP/1.1 409 Conflict
Date: Thu, 14 Jun 2018 09:46:19 GMT
Content-Type: application/json
Transfer-Encoding: chunked
Server: WSO2 Carbon Server
{"code":"20030","message":"Conflict","description":"User test_user already exists in the system. Please use a different username."}
```
Maybe a simple solution could be include the Validation Code in a JSON:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 07:10:42 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
{"code":"8ff16bfe-64d5-43a4-95b7-d9d2285a4936"}
```
The same happens with the Resend Code (https://localhost:9443/api/identity/user/v0.9/resend-code) resource of the API. Here is an example of response:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 09:49:35 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
81279f88-ae62-454e-9520-838a1ea80364
```
Thanks!
|
1.0
|
[Self Registration REST API] Wrong Content-Type in Register User successful response - Moved from: https://wso2.org/jira/browse/IDENTITY-7373
Hi,
I'm using the Self Registration REST API (v0.9) of Identity Server 5.3.0.
This is what I added in <IS_HOME>/repository/conf/identity/identity.xml:
```
<SelfRegistration>
<Enable>true</Enable>
<LockOnCreation>true</LockOnCreation>
<Notification>
<InternallyManage>false</InternallyManage>
</Notification>
<ReCaptcha>false</ReCaptcha>
</SelfRegistration>
```
The important thing here is that I have disable the internal management of the notifications because I want to manage it myself.
So when I use the Register User (https://localhost:9443/api/identity/user/v0.9/me) resource of the API it returns the Validation Code so I can handle it. But the Content-Type of the response is application/json and the body is a plain text. Here is an example:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 07:10:42 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
8ff16bfe-64d5-43a4-95b7-d9d2285a4936
```
When there is an error all goes well because in that case the Content-Type is application/json and the body is a JSON:
```
HTTP/1.1 409 Conflict
Date: Thu, 14 Jun 2018 09:46:19 GMT
Content-Type: application/json
Transfer-Encoding: chunked
Server: WSO2 Carbon Server
{"code":"20030","message":"Conflict","description":"User test_user already exists in the system. Please use a different username."}
```
Maybe a simple solution could be include the Validation Code in a JSON:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 07:10:42 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
{"code":"8ff16bfe-64d5-43a4-95b7-d9d2285a4936"}
```
The same happens with the Resend Code (https://localhost:9443/api/identity/user/v0.9/resend-code) resource of the API. Here is an example of response:
```
HTTP/1.1 201 Created
Date: Thu, 14 Jun 2018 09:49:35 GMT
Content-Type: application/json
Content-Length: 36
Server: WSO2 Carbon Server
81279f88-ae62-454e-9520-838a1ea80364
```
Thanks!
|
priority
|
wrong content type in register user successful response moved from hi i m using the self registration rest api of identity server this is what i added in repository conf identity identity xml true true false false the important thing here is that i have disable the internal management of the notifications because i want to manage it myself so when i use the register user resource of the api it returns the validation code so i can handle it but the content type of the response is application json and the body is a plain text here is an example http created date thu jun gmt content type application json content length server carbon server when there is an error all goes well because in that case the content type is application json and the body is a json http conflict date thu jun gmt content type application json transfer encoding chunked server carbon server code message conflict description user test user already exists in the system please use a different username maybe a simple solution could be include the validation code in a json http created date thu jun gmt content type application json content length server carbon server code the same happens with the resend code resource of the api here is an example of response http created date thu jun gmt content type application json content length server carbon server thanks
| 1
|
524,702
| 15,219,333,868
|
IssuesEvent
|
2021-02-17 19:03:13
|
xournalpp/xournalpp
|
https://api.github.com/repos/xournalpp/xournalpp
|
closed
|
New instance of Xournal++ gnome icon created when moving window to different monitor
|
bug difficulty::easy hacktoberfest priority::high
|
(Please complete the following information, and then delete this line)
**Affects versions :**
- OS: Ubuntu 19.04
- 1.0.13
**Describe the bug**
The Ubuntu Gnome dock shows two instances of Xournal ++ when you move the Xournal ++ window to a different monitor.
**To Reproduce**
Steps to reproduce the behavior:
1. Have a dual monitor setup.
2. Favorite Xournal++ to the Ubuntu dock.
3. Open an instance of Xournal++
4. Drag the instance of Xournal++ to a different monitor.
**Expected behavior**
There should be only one Xournal++ icon in the Gnome dock with one orange dot. If I open another Xournal++ window, then I should see two dots and so on.
**Screenshots of Problem**
(Optional) If applicable, add screenshots to help explain your problem.

**Additional context**
I don't recall this being a problem before, this just started happening maybe a few updates ago? Not sure.
|
1.0
|
New instance of Xournal++ gnome icon created when moving window to different monitor - (Please complete the following information, and then delete this line)
**Affects versions :**
- OS: Ubuntu 19.04
- 1.0.13
**Describe the bug**
The Ubuntu Gnome dock shows two instances of Xournal ++ when you move the Xournal ++ window to a different monitor.
**To Reproduce**
Steps to reproduce the behavior:
1. Have a dual monitor setup.
2. Favorite Xournal++ to the Ubuntu dock.
3. Open an instance of Xournal++
4. Drag the instance of Xournal++ to a different monitor.
**Expected behavior**
There should be only one Xournal++ icon in the Gnome dock with one orange dot. If I open another Xournal++ window, then I should see two dots and so on.
**Screenshots of Problem**
(Optional) If applicable, add screenshots to help explain your problem.

**Additional context**
I don't recall this being a problem before, this just started happening maybe a few updates ago? Not sure.
|
priority
|
new instance of xournal gnome icon created when moving window to different monitor please complete the following information and then delete this line affects versions os ubuntu describe the bug the ubuntu gnome dock shows two instances of xournal when you move the xournal window to a different monitor to reproduce steps to reproduce the behavior have a dual monitor setup favorite xournal to the ubuntu dock open an instance of xournal drag the instance of xournal to a different monitor expected behavior there should be only one xournal icon in the gnome dock with one orange dot if i open another xournal window then i should see two dots and so on screenshots of problem optional if applicable add screenshots to help explain your problem additional context i don t recall this being a problem before this just started happening maybe a few updates ago not sure
| 1
|
250,243
| 7,973,851,202
|
IssuesEvent
|
2018-07-17 01:43:44
|
ballerina-platform/ballerina-message-broker
|
https://api.github.com/repos/ballerina-platform/ballerina-message-broker
|
opened
|
Implement functionality to retrieve channels for a connection through REST api
|
Module/broker-amqp Priority/High Type/Task
|
**Description:**
Implement functionality to retrieve channels for a connection through REST api
|
1.0
|
Implement functionality to retrieve channels for a connection through REST api - **Description:**
Implement functionality to retrieve channels for a connection through REST api
|
priority
|
implement functionality to retrieve channels for a connection through rest api description implement functionality to retrieve channels for a connection through rest api
| 1
|
198,171
| 6,970,747,011
|
IssuesEvent
|
2017-12-11 11:26:18
|
grmToolbox/grmpy
|
https://api.github.com/repos/grmToolbox/grmpy
|
closed
|
Value of Criterion Function at truth
|
pb-estimation priority-high size-S
|
Please add the information about the value of the criterion function at the truth to the information on simulated datasets in `*.grmpy.info`
|
1.0
|
Value of Criterion Function at truth - Please add the information about the value of the criterion function at the truth to the information on simulated datasets in `*.grmpy.info`
|
priority
|
value of criterion function at truth please add the information about the value of the criterion function at the truth to the information on simulated datasets in grmpy info
| 1
|
722,847
| 24,876,055,780
|
IssuesEvent
|
2022-10-27 19:12:13
|
art-framework-suite/mrb
|
https://api.github.com/repos/art-framework-suite/mrb
|
closed
|
problem with mrb v6_07_04
|
bug high priority
|
**Describe the bug**
Cloning a release defaults to redmine instead of github:
mrb g -r sbncode@release/v09_61_00
ERROR: unable to verify read access to remote http://cdcvs.fnal.gov/projects/sbncode@release/v09_61_00
**To Reproduce**
This can be easily reproduced on scisoftbuild01.
**If this is a high-priority issue**
Please add the "high priority" label to this issue.
YES
|
1.0
|
problem with mrb v6_07_04 - **Describe the bug**
Cloning a release defaults to redmine instead of github:
mrb g -r sbncode@release/v09_61_00
ERROR: unable to verify read access to remote http://cdcvs.fnal.gov/projects/sbncode@release/v09_61_00
**To Reproduce**
This can be easily reproduced on scisoftbuild01.
**If this is a high-priority issue**
Please add the "high priority" label to this issue.
YES
|
priority
|
problem with mrb describe the bug cloning a release defaults to redmine instead of github mrb g r sbncode release error unable to verify read access to remote to reproduce this can be easily reproduced on if this is a high priority issue please add the high priority label to this issue yes
| 1
|
412,341
| 12,041,473,520
|
IssuesEvent
|
2020-04-14 08:54:12
|
ooni/probe-engine
|
https://api.github.com/repos/ooni/probe-engine
|
opened
|
The DASH server does not save the user IP on M-Lab
|
effort/XL priority/high
|
- [ ] The DASH server saves the @m-lab's UUID
- [ ] The DASH client saves the @m-lab's UUID
- [ ] The DASH server does not use the reverse traceroute helper
- [ ] The DASH server measurement includes the probe ASN and CC
- [ ] The DASH server measurement does not include the IP
|
1.0
|
The DASH server does not save the user IP on M-Lab - - [ ] The DASH server saves the @m-lab's UUID
- [ ] The DASH client saves the @m-lab's UUID
- [ ] The DASH server does not use the reverse traceroute helper
- [ ] The DASH server measurement includes the probe ASN and CC
- [ ] The DASH server measurement does not include the IP
|
priority
|
the dash server does not save the user ip on m lab the dash server saves the m lab s uuid the dash client saves the m lab s uuid the dash server does not use the reverse traceroute helper the dash server measurement includes the probe asn and cc the dash server measurement does not include the ip
| 1
|
454,949
| 13,109,537,972
|
IssuesEvent
|
2020-08-04 18:53:20
|
ctm/mb2-doc
|
https://api.github.com/repos/ctm/mb2-doc
|
closed
|
payouts don't balance
|
chore easy high priority
|
> The payouts for the main don't balance. I think there must be a bug in that part of the payout table. Below is what MB2 spit out but my results email will have the right numbers.
> 1 Russell "ABVidale" Fox T$4,050
> 2 John "muscatel" Grout T$3,078
> 3 John "da pickle" Pickels T$2,268
> 4 Gillian "Tegwin" Groves T$1,782
> 5 Christopher "tombayz" Mecklin T$1,458
> 6 Kenny "HoserSimpson" Shei T$1,134
> 7 Bryan "bjuliano" Juliano T$810
> 8 Tanya "MissT74" Peck-Devenport T$486
> 9 Shari "pokerchimp" Silk T$324
> 10 Michael "mjoseph" Brennan T$324
> 11 Barry "MrRaise" Kornspan T$243
> 12 John "JRX" Reed T$243
Here are the correct numbers:
> 1 Russell "ABVidale" Fox T$4,060
> 2 John "muscatel" Grout T$3,086
> 3 John "da pickle" Pickels T$2,274
> 4 Gillian "Tegwin" Groves T$1,786
> 5 Christopher "tombayz" Mecklin T$1,462
> 6 Kenny "HoserSimpson" Shei T$1,137
> 7 Bryan "bjuliano" Juliano T$812
> 8 Tanya "MissT74" Peck-Devenport T$487
> 9 Shari "pokerchimp" Silk T$325
> 10 Michael "mjoseph" Brennan T$325
> 11 Barry "MrRaise" Kornspan T$243
> 12 John "JRX" Reed T$243
Differences (copied and pasted and calculated in my tired head): FM44 difference. WTF?
1. 4060 4050 +10
2. 3086 3078 +8
3. 2274 2268 +6
4. 1786 1782 +4
5. 1462 1458 +4
6. 1137 1134 +3
7. 812 810 +2
8. 487 486 +1
9. 325 324 +1
10. 325 324 +1
11 and 12 are correct
|
1.0
|
payouts don't balance - > The payouts for the main don't balance. I think there must be a bug in that part of the payout table. Below is what MB2 spit out but my results email will have the right numbers.
> 1 Russell "ABVidale" Fox T$4,050
> 2 John "muscatel" Grout T$3,078
> 3 John "da pickle" Pickels T$2,268
> 4 Gillian "Tegwin" Groves T$1,782
> 5 Christopher "tombayz" Mecklin T$1,458
> 6 Kenny "HoserSimpson" Shei T$1,134
> 7 Bryan "bjuliano" Juliano T$810
> 8 Tanya "MissT74" Peck-Devenport T$486
> 9 Shari "pokerchimp" Silk T$324
> 10 Michael "mjoseph" Brennan T$324
> 11 Barry "MrRaise" Kornspan T$243
> 12 John "JRX" Reed T$243
Here are the correct numbers:
> 1 Russell "ABVidale" Fox T$4,060
> 2 John "muscatel" Grout T$3,086
> 3 John "da pickle" Pickels T$2,274
> 4 Gillian "Tegwin" Groves T$1,786
> 5 Christopher "tombayz" Mecklin T$1,462
> 6 Kenny "HoserSimpson" Shei T$1,137
> 7 Bryan "bjuliano" Juliano T$812
> 8 Tanya "MissT74" Peck-Devenport T$487
> 9 Shari "pokerchimp" Silk T$325
> 10 Michael "mjoseph" Brennan T$325
> 11 Barry "MrRaise" Kornspan T$243
> 12 John "JRX" Reed T$243
Differences (copied and pasted and calculated in my tired head): FM44 difference. WTF?
1. 4060 4050 +10
2. 3086 3078 +8
3. 2274 2268 +6
4. 1786 1782 +4
5. 1462 1458 +4
6. 1137 1134 +3
7. 812 810 +2
8. 487 486 +1
9. 325 324 +1
10. 325 324 +1
11 and 12 are correct
|
priority
|
payouts don t balance the payouts for the main don t balance i think there must be a bug in that part of the payout table below is what spit out but my results email will have the right numbers russell abvidale fox t john muscatel grout t john da pickle pickels t gillian tegwin groves t christopher tombayz mecklin t kenny hosersimpson shei t bryan bjuliano juliano t tanya peck devenport t shari pokerchimp silk t michael mjoseph brennan t barry mrraise kornspan t john jrx reed t here are the correct numbers russell abvidale fox t john muscatel grout t john da pickle pickels t gillian tegwin groves t christopher tombayz mecklin t kenny hosersimpson shei t bryan bjuliano juliano t tanya peck devenport t shari pokerchimp silk t michael mjoseph brennan t barry mrraise kornspan t john jrx reed t differences copied and pasted and calculated in my tired head difference wtf and are correct
| 1
|
110,304
| 4,424,866,387
|
IssuesEvent
|
2016-08-16 13:55:29
|
softdevteam/krun
|
https://api.github.com/repos/softdevteam/krun
|
closed
|
Apply patch to PyPy while building to fix instrumentation
|
bug high priority (must have before publication)
|
PyPy rev 6a6545b4a915 needs to be applied, because otherwise categories aren't properly nested.
|
1.0
|
Apply patch to PyPy while building to fix instrumentation - PyPy rev 6a6545b4a915 needs to be applied, because otherwise categories aren't properly nested.
|
priority
|
apply patch to pypy while building to fix instrumentation pypy rev needs to be applied because otherwise categories aren t properly nested
| 1
|
221,644
| 7,393,823,435
|
IssuesEvent
|
2018-03-17 02:15:02
|
VolmitSoftware/React
|
https://api.github.com/repos/VolmitSoftware/React
|
closed
|
React monitoring Freeze
|
R6 blocker bug confirmed high priority
|
Frozen since 6 hours: http://prntscr.com/iro6qj
Monitoring system freeze after few minutes. Need to reload React to fix.
|
1.0
|
React monitoring Freeze - Frozen since 6 hours: http://prntscr.com/iro6qj
Monitoring system freeze after few minutes. Need to reload React to fix.
|
priority
|
react monitoring freeze frozen since hours monitoring system freeze after few minutes need to reload react to fix
| 1
|
594,352
| 18,043,696,702
|
IssuesEvent
|
2021-09-18 14:00:40
|
CaptureCoop/SnipSniper
|
https://api.github.com/repos/CaptureCoop/SnipSniper
|
closed
|
Create a universal IClosable interface to make it easier to close child windows
|
enhancement High Priority
|
For example:
ConfigWindow has an array of IClosables that it closes once it itself closes
|
1.0
|
Create a universal IClosable interface to make it easier to close child windows - For example:
ConfigWindow has an array of IClosables that it closes once it itself closes
|
priority
|
create a universal iclosable interface to make it easier to close child windows for example configwindow has an array of iclosables that it closes once it itself closes
| 1
|
806,683
| 29,868,336,477
|
IssuesEvent
|
2023-06-20 06:38:56
|
ElectricRCAircraftGuy/ElectricRCAircraftGuy.github.io
|
https://api.github.com/repos/ElectricRCAircraftGuy/ElectricRCAircraftGuy.github.io
|
closed
|
swap the "previous" and "next" buttons at the bottom of the main home page; they are backwards
|
highest priority
|
Swap their names. They say the wrong thing:

Swap them in the `_includes/paginator.html` file:

|
1.0
|
swap the "previous" and "next" buttons at the bottom of the main home page; they are backwards - Swap their names. They say the wrong thing:

Swap them in the `_includes/paginator.html` file:

|
priority
|
swap the previous and next buttons at the bottom of the main home page they are backwards swap their names they say the wrong thing swap them in the includes paginator html file
| 1
|
638,977
| 20,743,730,335
|
IssuesEvent
|
2022-03-14 20:22:29
|
infosimples/infosimples
|
https://api.github.com/repos/infosimples/infosimples
|
closed
|
[API de Automação de Consulta] SEFAZ / SP / IPVA
|
status: confirmed product: consultas priority: high
|
### Nome da consulta:
SEFAZ / SP / IPVA
### URL da consulta:
https://www.ipva.fazenda.sp.gov.br/ipvanet/guiater.aspx
COLOQUE_AQUI_A_URL_DA_CONSULTA
### Parâmetros de entrada necessários para fazer a consulta:
- renavam
- placa
- 0km (booleano)
- locadoa (booleano)
### Passo a passo de como realizar manualmente a consulta:



### Qual é uma boa descrição desta consulta?
Pesquisa dados de Imposto sobre Propriedades de Veículos Automotores (IPVA).
### [OPCIONAL] Você gostaria de usar esta API para automatizar que tipo de processo da sua empresa?
- Pode ser usada para prevenir fraudes, confirmando se o veículo informado possui débitos de IPVA.
- Pode ser usada em gestão cadastral e fiscal do veículo, garantindo que os dados informados estão atualizados e corretos.
- Pode ser usada para gerar boletos de pagamentos pendentes.
|
1.0
|
[API de Automação de Consulta] SEFAZ / SP / IPVA - ### Nome da consulta:
SEFAZ / SP / IPVA
### URL da consulta:
https://www.ipva.fazenda.sp.gov.br/ipvanet/guiater.aspx
COLOQUE_AQUI_A_URL_DA_CONSULTA
### Parâmetros de entrada necessários para fazer a consulta:
- renavam
- placa
- 0km (booleano)
- locadoa (booleano)
### Passo a passo de como realizar manualmente a consulta:



### Qual é uma boa descrição desta consulta?
Pesquisa dados de Imposto sobre Propriedades de Veículos Automotores (IPVA).
### [OPCIONAL] Você gostaria de usar esta API para automatizar que tipo de processo da sua empresa?
- Pode ser usada para prevenir fraudes, confirmando se o veículo informado possui débitos de IPVA.
- Pode ser usada em gestão cadastral e fiscal do veículo, garantindo que os dados informados estão atualizados e corretos.
- Pode ser usada para gerar boletos de pagamentos pendentes.
|
priority
|
sefaz sp ipva nome da consulta sefaz sp ipva url da consulta coloque aqui a url da consulta parâmetros de entrada necessários para fazer a consulta renavam placa booleano locadoa booleano passo a passo de como realizar manualmente a consulta qual é uma boa descrição desta consulta pesquisa dados de imposto sobre propriedades de veículos automotores ipva você gostaria de usar esta api para automatizar que tipo de processo da sua empresa pode ser usada para prevenir fraudes confirmando se o veículo informado possui débitos de ipva pode ser usada em gestão cadastral e fiscal do veículo garantindo que os dados informados estão atualizados e corretos pode ser usada para gerar boletos de pagamentos pendentes
| 1
|
143,875
| 5,531,963,259
|
IssuesEvent
|
2017-03-21 09:20:48
|
HeinrichReimer/material-intro
|
https://api.github.com/repos/HeinrichReimer/material-intro
|
closed
|
Changing to MIT license
|
high priority
|
Hi,
It's a nice library and I though of using it for a new project. Unfortunately, you license is complex. I suggest you change it to MIT which is the most common open source license these days. Otherwise, it makes it difficult for others to use it.
Thanks in Advance!
|
1.0
|
Changing to MIT license - Hi,
It's a nice library and I though of using it for a new project. Unfortunately, you license is complex. I suggest you change it to MIT which is the most common open source license these days. Otherwise, it makes it difficult for others to use it.
Thanks in Advance!
|
priority
|
changing to mit license hi it s a nice library and i though of using it for a new project unfortunately you license is complex i suggest you change it to mit which is the most common open source license these days otherwise it makes it difficult for others to use it thanks in advance
| 1
|
712,091
| 24,484,326,547
|
IssuesEvent
|
2022-10-09 08:17:03
|
AY2223S1-CS2103-F13-4/tp
|
https://api.github.com/repos/AY2223S1-CS2103-F13-4/tp
|
closed
|
As a medical administrator I can delete a patient
|
type.Story priority.High
|
who no longer requires home visits, so that I do not need to include them in the scheduling exercise.
|
1.0
|
As a medical administrator I can delete a patient - who no longer requires home visits, so that I do not need to include them in the scheduling exercise.
|
priority
|
as a medical administrator i can delete a patient who no longer requires home visits so that i do not need to include them in the scheduling exercise
| 1
|
809,431
| 30,192,827,561
|
IssuesEvent
|
2023-07-04 17:02:36
|
tcet-opensource/erp-backend
|
https://api.github.com/repos/tcet-opensource/erp-backend
|
closed
|
[feat]: make all models consistent
|
enhancement Issue Size: 3 Priority: High Models
|
# Feature Request
## Description
Bring consistency across all the model files.
## Use Case
This will help in easing out the workflow
## Proposed Solution
making sure all variable case are same. and create will take an object as input
---
Thank you for submitting your feature request! We appreciate your feedback and will carefully review your suggestions. Your input helps us improve our product and better meet the needs of our users.
|
1.0
|
[feat]: make all models consistent - # Feature Request
## Description
Bring consistency across all the model files.
## Use Case
This will help in easing out the workflow
## Proposed Solution
making sure all variable case are same. and create will take an object as input
---
Thank you for submitting your feature request! We appreciate your feedback and will carefully review your suggestions. Your input helps us improve our product and better meet the needs of our users.
|
priority
|
make all models consistent feature request description bring consistency across all the model files use case this will help in easing out the workflow proposed solution making sure all variable case are same and create will take an object as input thank you for submitting your feature request we appreciate your feedback and will carefully review your suggestions your input helps us improve our product and better meet the needs of our users
| 1
|
261,077
| 8,223,753,004
|
IssuesEvent
|
2018-09-06 11:41:03
|
dr-wilkinson/traveller
|
https://api.github.com/repos/dr-wilkinson/traveller
|
opened
|
bug : new save file ui reset
|
priority: high status: accepted type: bug
|
When 'New' menu item is selected and the currently displayed characters have not been saved, an alert is displayed asking if the user wants to save the current characters. If the user selects yes, the characters are saved but the ui doesn't reset itself to blank i.e. new when it should.
|
1.0
|
bug : new save file ui reset - When 'New' menu item is selected and the currently displayed characters have not been saved, an alert is displayed asking if the user wants to save the current characters. If the user selects yes, the characters are saved but the ui doesn't reset itself to blank i.e. new when it should.
|
priority
|
bug new save file ui reset when new menu item is selected and the currently displayed characters have not been saved an alert is displayed asking if the user wants to save the current characters if the user selects yes the characters are saved but the ui doesn t reset itself to blank i e new when it should
| 1
|
694,359
| 23,810,876,371
|
IssuesEvent
|
2022-09-04 18:45:01
|
opendatahub-io/odh-dashboard
|
https://api.github.com/repos/opendatahub-io/odh-dashboard
|
opened
|
[DSG]: Support Notebook & Data Source Lists
|
kind/enhancement priority/high feature/dsg
|
### Feature description
Create tables to hold onto notebooks & storage items (using an existing Notebook from the Jupyter Tile for data should be enough to get test data while you wait on #483)
- Must support
- collapsible rows
- sortable columns
- external links
- kebab items
### Describe alternatives you've considered
_No response_
### Anything else?
Mocks - https://marvelapp.com/prototype/afa716d/screen/88397428
|
1.0
|
[DSG]: Support Notebook & Data Source Lists - ### Feature description
Create tables to hold onto notebooks & storage items (using an existing Notebook from the Jupyter Tile for data should be enough to get test data while you wait on #483)
- Must support
- collapsible rows
- sortable columns
- external links
- kebab items
### Describe alternatives you've considered
_No response_
### Anything else?
Mocks - https://marvelapp.com/prototype/afa716d/screen/88397428
|
priority
|
support notebook data source lists feature description create tables to hold onto notebooks storage items using an existing notebook from the jupyter tile for data should be enough to get test data while you wait on must support collapsible rows sortable columns external links kebab items describe alternatives you ve considered no response anything else mocks
| 1
|
174,604
| 6,541,528,072
|
IssuesEvent
|
2017-09-01 20:25:36
|
jhpoelen/fb-osmose-bridge
|
https://api.github.com/repos/jhpoelen/fb-osmose-bridge
|
reopened
|
Parameters of the background functional groups for which it is not reasonable to define a default value
|
UI High priority
|
@jhpoelen @Dengaloo
The “osm_param-ltl.csv” file provides estimates for three parameters: “plankton.size.min.plkX”, “plankton.size.max.plkX” and “plankton.TL.plkX”.
The default values of these three parameters (which are specified in “fishbase-mapping.csv”) are problematic.
Indeed, it is impossible to specify a default value for these three parameters for all the fish and invertebrates recorded in FishBase and SeaLifeBase.
To solve this issue, we have two options:
(Option 1) Setting the default value of the “plankton.size.min.plkX”, “plankton.size.max.plkX” and “plankton.TL.plkX” parameters to “NA” (not available) in the “fishbase-mapping.csv” file.
Or
(Option 2) Implementing a rule of thumb in the API such that, if the API is unable to define a value for one parameter (for example “plankton.size.min.plkX”) for a given functional group, it uses information for the other species belonging to the genuses, families and orders considered in the functional group. (These other species can be found anywhere across the globe).
For instance, let’s say that:
(i) We are dealing with the Iceland Shelf/Sea ecosystem and that the bridge between FishBase/SeaLifeBase and OSMOSE defines a functional group called “demersalmollusc” for us.
(ii) We choose to define “demersalmollusc” as a “background” functional group.
(iii) The “demersalmollusc” functional group includes three species: Falcidens thorensis (Aplacophora), Prochaetoderma clenchi (Aplacophora) and Micropilina minuta (Monoplacophora).
(iv) Estimates are not available for Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta to define a value for the parameter “plankton.size.min.plkX” for the “demersalmollusc” functional group.
(v) Then, the API: identifies the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong; and uses the information available for all the species that belong to the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong to define a value for the parameter “plankton.size.min.plkX” for the “demersalmollusc” functional group.
Personnally, I prefer Option 2. We do not want to provide a lot of NA’s to the user; this would be frustrating to them.
Plus I am not sure that the API can handle something else than numbers regarding the default values taken from the “Default value” column of the “fishbase-mapping.csv” file. @jhpoelen Is that correct?
But then, if we choose Option 2, how should the estimates be computed?
I was thinking that the best option in the case presented above would be to: (i) consider all the minimum size estimates available for the species that belong to the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong; and (ii) produce a mean estimate from all the available estimates.
Please let me know what you think.
After I have received your input, I will close the present issue and will create new API issues.
Many thanks.
|
1.0
|
Parameters of the background functional groups for which it is not reasonable to define a default value - @jhpoelen @Dengaloo
The “osm_param-ltl.csv” file provides estimates for three parameters: “plankton.size.min.plkX”, “plankton.size.max.plkX” and “plankton.TL.plkX”.
The default values of these three parameters (which are specified in “fishbase-mapping.csv”) are problematic.
Indeed, it is impossible to specify a default value for these three parameters for all the fish and invertebrates recorded in FishBase and SeaLifeBase.
To solve this issue, we have two options:
(Option 1) Setting the default value of the “plankton.size.min.plkX”, “plankton.size.max.plkX” and “plankton.TL.plkX” parameters to “NA” (not available) in the “fishbase-mapping.csv” file.
Or
(Option 2) Implementing a rule of thumb in the API such that, if the API is unable to define a value for one parameter (for example “plankton.size.min.plkX”) for a given functional group, it uses information for the other species belonging to the genuses, families and orders considered in the functional group. (These other species can be found anywhere across the globe).
For instance, let’s say that:
(i) We are dealing with the Iceland Shelf/Sea ecosystem and that the bridge between FishBase/SeaLifeBase and OSMOSE defines a functional group called “demersalmollusc” for us.
(ii) We choose to define “demersalmollusc” as a “background” functional group.
(iii) The “demersalmollusc” functional group includes three species: Falcidens thorensis (Aplacophora), Prochaetoderma clenchi (Aplacophora) and Micropilina minuta (Monoplacophora).
(iv) Estimates are not available for Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta to define a value for the parameter “plankton.size.min.plkX” for the “demersalmollusc” functional group.
(v) Then, the API: identifies the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong; and uses the information available for all the species that belong to the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong to define a value for the parameter “plankton.size.min.plkX” for the “demersalmollusc” functional group.
Personnally, I prefer Option 2. We do not want to provide a lot of NA’s to the user; this would be frustrating to them.
Plus I am not sure that the API can handle something else than numbers regarding the default values taken from the “Default value” column of the “fishbase-mapping.csv” file. @jhpoelen Is that correct?
But then, if we choose Option 2, how should the estimates be computed?
I was thinking that the best option in the case presented above would be to: (i) consider all the minimum size estimates available for the species that belong to the genuses, families and orders to which Falcidens thorensis, Prochaetoderma clenchi and Micropilina minuta belong; and (ii) produce a mean estimate from all the available estimates.
Please let me know what you think.
After I have received your input, I will close the present issue and will create new API issues.
Many thanks.
|
priority
|
parameters of the background functional groups for which it is not reasonable to define a default value jhpoelen dengaloo the “osm param ltl csv” file provides estimates for three parameters “plankton size min plkx” “plankton size max plkx” and “plankton tl plkx” the default values of these three parameters which are specified in “fishbase mapping csv” are problematic indeed it is impossible to specify a default value for these three parameters for all the fish and invertebrates recorded in fishbase and sealifebase to solve this issue we have two options option setting the default value of the “plankton size min plkx” “plankton size max plkx” and “plankton tl plkx” parameters to “na” not available in the “fishbase mapping csv” file or option implementing a rule of thumb in the api such that if the api is unable to define a value for one parameter for example “plankton size min plkx” for a given functional group it uses information for the other species belonging to the genuses families and orders considered in the functional group these other species can be found anywhere across the globe for instance let’s say that i we are dealing with the iceland shelf sea ecosystem and that the bridge between fishbase sealifebase and osmose defines a functional group called “demersalmollusc” for us ii we choose to define “demersalmollusc” as a “background” functional group iii the “demersalmollusc” functional group includes three species falcidens thorensis aplacophora prochaetoderma clenchi aplacophora and micropilina minuta monoplacophora iv estimates are not available for falcidens thorensis prochaetoderma clenchi and micropilina minuta to define a value for the parameter “plankton size min plkx” for the “demersalmollusc” functional group v then the api identifies the genuses families and orders to which falcidens thorensis prochaetoderma clenchi and micropilina minuta belong and uses the information available for all the species that belong to the genuses families and orders to which falcidens thorensis prochaetoderma clenchi and micropilina minuta belong to define a value for the parameter “plankton size min plkx” for the “demersalmollusc” functional group personnally i prefer option we do not want to provide a lot of na’s to the user this would be frustrating to them plus i am not sure that the api can handle something else than numbers regarding the default values taken from the “default value” column of the “fishbase mapping csv” file jhpoelen is that correct but then if we choose option how should the estimates be computed i was thinking that the best option in the case presented above would be to i consider all the minimum size estimates available for the species that belong to the genuses families and orders to which falcidens thorensis prochaetoderma clenchi and micropilina minuta belong and ii produce a mean estimate from all the available estimates please let me know what you think after i have received your input i will close the present issue and will create new api issues many thanks
| 1
|
506,328
| 14,662,938,496
|
IssuesEvent
|
2020-12-29 08:32:43
|
bounswe/bounswe2020group3
|
https://api.github.com/repos/bounswe/bounswe2020group3
|
closed
|
Android - User can not edit private projects
|
Android Priority: High Type: Bug
|
* **Project: ANDROID**
* **This is a: BUG REPORT**
* **Description of the issue**
Currently logged in user can not edit private projects of his/her.
* **For bug reports: Explanation of how to reproduce the bug, and what was the expected behaviour.**
1- Open project details of a current user's private project
2- Click edit button
3- Private fields are not shown therefore cant update some fields of project
* **Deadline for resolution:**
29.12.2020
|
1.0
|
Android - User can not edit private projects - * **Project: ANDROID**
* **This is a: BUG REPORT**
* **Description of the issue**
Currently logged in user can not edit private projects of his/her.
* **For bug reports: Explanation of how to reproduce the bug, and what was the expected behaviour.**
1- Open project details of a current user's private project
2- Click edit button
3- Private fields are not shown therefore cant update some fields of project
* **Deadline for resolution:**
29.12.2020
|
priority
|
android user can not edit private projects project android this is a bug report description of the issue currently logged in user can not edit private projects of his her for bug reports explanation of how to reproduce the bug and what was the expected behaviour open project details of a current user s private project click edit button private fields are not shown therefore cant update some fields of project deadline for resolution
| 1
|
212,222
| 7,229,508,613
|
IssuesEvent
|
2018-02-11 20:28:07
|
builderbook/builderbook
|
https://api.github.com/repos/builderbook/builderbook
|
closed
|
remove sanitize-html
|
high priority
|
@tima101 sanitize-html causes issues with displaying some content in books. Since users will write all code in books (and will not add anything unsafe), we can remove sanitize-html.
|
1.0
|
remove sanitize-html - @tima101 sanitize-html causes issues with displaying some content in books. Since users will write all code in books (and will not add anything unsafe), we can remove sanitize-html.
|
priority
|
remove sanitize html sanitize html causes issues with displaying some content in books since users will write all code in books and will not add anything unsafe we can remove sanitize html
| 1
|
315,729
| 9,631,433,194
|
IssuesEvent
|
2019-05-15 14:13:30
|
jonnhy12333/hmct-web
|
https://api.github.com/repos/jonnhy12333/hmct-web
|
closed
|
Using settings should not force the chat to reload
|
Enhancement High Priority
|
With the new decryption things are a little slow on initial load. Instead of navigating to a new route for the settings page we should just make it a dialog like how the about page is.
|
1.0
|
Using settings should not force the chat to reload - With the new decryption things are a little slow on initial load. Instead of navigating to a new route for the settings page we should just make it a dialog like how the about page is.
|
priority
|
using settings should not force the chat to reload with the new decryption things are a little slow on initial load instead of navigating to a new route for the settings page we should just make it a dialog like how the about page is
| 1
|
508,102
| 14,689,950,849
|
IssuesEvent
|
2021-01-02 12:46:57
|
1ForeverHD/HDAdmin
|
https://api.github.com/repos/1ForeverHD/HDAdmin
|
opened
|
Fully rebrand to V3 to Nanoblox
|
Priority: High Scope: Command Scope: Docs Type: Enhancement
|
- [ ] complete topbarplusv2 and move out of repo
- [ ] complete zoneplusv2 and move out of repo
- [ ] move datastoreplus out of repo
- [ ] consider dependency management for modules such as 'State'k
- [ ] transfer all utility modules directly into the core
- [ ] remove projects and utility from docs
- [ ] rename docs from HDAdmin to Nanoblox
- [ ] change repo name to 1ForeverHD/Nanoblox
- [ ] check all devforum links now refer to 1ForeverHD/Nanoblox, 1ForeverHD/TopbarPlus or 1ForeverHD/ZonePlus
|
1.0
|
Fully rebrand to V3 to Nanoblox - - [ ] complete topbarplusv2 and move out of repo
- [ ] complete zoneplusv2 and move out of repo
- [ ] move datastoreplus out of repo
- [ ] consider dependency management for modules such as 'State'k
- [ ] transfer all utility modules directly into the core
- [ ] remove projects and utility from docs
- [ ] rename docs from HDAdmin to Nanoblox
- [ ] change repo name to 1ForeverHD/Nanoblox
- [ ] check all devforum links now refer to 1ForeverHD/Nanoblox, 1ForeverHD/TopbarPlus or 1ForeverHD/ZonePlus
|
priority
|
fully rebrand to to nanoblox complete and move out of repo complete and move out of repo move datastoreplus out of repo consider dependency management for modules such as state k transfer all utility modules directly into the core remove projects and utility from docs rename docs from hdadmin to nanoblox change repo name to nanoblox check all devforum links now refer to nanoblox topbarplus or zoneplus
| 1
|
75,531
| 3,466,186,570
|
IssuesEvent
|
2015-12-22 01:15:26
|
byaka/flaskJSONRPCServer
|
https://api.github.com/repos/byaka/flaskJSONRPCServer
|
opened
|
В API параллельного бекенда нужно включить сжатие для длинных сообщений
|
enhancement High-priority
|
Но нужно проверить, поддерживается ли сжатие используемая библиотека запросов
|
1.0
|
В API параллельного бекенда нужно включить сжатие для длинных сообщений - Но нужно проверить, поддерживается ли сжатие используемая библиотека запросов
|
priority
|
в api параллельного бекенда нужно включить сжатие для длинных сообщений но нужно проверить поддерживается ли сжатие используемая библиотека запросов
| 1
|
190,688
| 6,821,556,869
|
IssuesEvent
|
2017-11-07 17:05:42
|
vmware/vic-product
|
https://api.github.com/repos/vmware/vic-product
|
closed
|
Identify VIC Appliance VM
|
component/ova priority/high team/lifecycle
|
**User Statement:**
As a UI I need to be able to identify which VMs are VIC Appliances.
**Details:**
Related to #920
http://pubs.vmware.com/vsphere-6-5/index.jsp#com.vmware.vspsdk.apiref.doc/vim.ExtensibleManagedObject.html?resultof=%2522%2573%2565%2574%2543%2575%2573%2574%256f%256d%2556%2561%256c%2575%2565%2522%2520%2522%2573%2565%2574%2563%2575%2573%2574%256f%256d%2576%2561%256c%2575%2522%2520
Get the VirtualMachine object and set a custom value on it
Tag should be on key `identifier` with value `vic-appliance-<BUILD_NUMBER>`
Figure out build number format with @jooskim and @mdharamadas1
**Acceptance Criteria:**
- [ ] Use API to tag the VM during the appliance initialization process
- [ ] Same process needs to apply during all cases of initialization - getting started page, upgrade, registration API
|
1.0
|
Identify VIC Appliance VM - **User Statement:**
As a UI I need to be able to identify which VMs are VIC Appliances.
**Details:**
Related to #920
http://pubs.vmware.com/vsphere-6-5/index.jsp#com.vmware.vspsdk.apiref.doc/vim.ExtensibleManagedObject.html?resultof=%2522%2573%2565%2574%2543%2575%2573%2574%256f%256d%2556%2561%256c%2575%2565%2522%2520%2522%2573%2565%2574%2563%2575%2573%2574%256f%256d%2576%2561%256c%2575%2522%2520
Get the VirtualMachine object and set a custom value on it
Tag should be on key `identifier` with value `vic-appliance-<BUILD_NUMBER>`
Figure out build number format with @jooskim and @mdharamadas1
**Acceptance Criteria:**
- [ ] Use API to tag the VM during the appliance initialization process
- [ ] Same process needs to apply during all cases of initialization - getting started page, upgrade, registration API
|
priority
|
identify vic appliance vm user statement as a ui i need to be able to identify which vms are vic appliances details related to get the virtualmachine object and set a custom value on it tag should be on key identifier with value vic appliance figure out build number format with jooskim and acceptance criteria use api to tag the vm during the appliance initialization process same process needs to apply during all cases of initialization getting started page upgrade registration api
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.