Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
281,988 | 8,701,595,208 | IssuesEvent | 2018-12-05 12:02:53 | antonwilc0x/NewSO | https://api.github.com/repos/antonwilc0x/NewSO | closed | Basic Server Database | complexity: medium hiatus priority: high server | In order to add or extend certain features, a database needs to be established. Entity Framework, Microsoft's cross-platform ORM framework, ~~LiteDB, a noSQL database,~~ will serve as the foundation. For better flexibility, each database will be separate and operate independently of each other. In theory, this _should_ cut down on the number of writes.
## Databases
- [ ] Users (Username, ID, Password Hash, Assigned Avatar IDs)
- [ ] Avatars (Avatar Name, ID, Assigned User ID, City, Money)
- [ ] Cities (Lot, Lot location, Assigned Avatar ID, Category)
The city database is not included in this since it's the most complicated part. | 1.0 | Basic Server Database - In order to add or extend certain features, a database needs to be established. Entity Framework, Microsoft's cross-platform ORM framework, ~~LiteDB, a noSQL database,~~ will serve as the foundation. For better flexibility, each database will be separate and operate independently of each other. In theory, this _should_ cut down on the number of writes.
## Databases
- [ ] Users (Username, ID, Password Hash, Assigned Avatar IDs)
- [ ] Avatars (Avatar Name, ID, Assigned User ID, City, Money)
- [ ] Cities (Lot, Lot location, Assigned Avatar ID, Category)
The city database is not included in this since it's the most complicated part. | priority | basic server database in order to add or extend certain features a database needs to be established entity framework microsoft s cross platform orm framework litedb a nosql database will serve as the foundation for better flexibility each database will be separate and operate independently of each other in theory this should cut down on the number of writes databases users username id password hash assigned avatar ids avatars avatar name id assigned user id city money cities lot lot location assigned avatar id category the city database is not included in this since it s the most complicated part | 1 |
425,456 | 12,340,661,537 | IssuesEvent | 2020-05-14 20:21:13 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Missing forward_key_balanced configuration in radius config | Priority: Medium Type: Bug | **Describe the bug**
In Configuration -> System configuration -> RADIUS it miss the forward_key_balanced configuration parameter.
| 1.0 | Missing forward_key_balanced configuration in radius config - **Describe the bug**
In Configuration -> System configuration -> RADIUS it miss the forward_key_balanced configuration parameter.
| priority | missing forward key balanced configuration in radius config describe the bug in configuration system configuration radius it miss the forward key balanced configuration parameter | 1 |
311,575 | 9,535,613,212 | IssuesEvent | 2019-04-30 07:27:50 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Don't set EventsPerJob bigger than RequestNumEvents | Enhancement Medium Priority ReqMgr2 | Of course, also considering the FilterEfficiency and the usual multiple of EventsPerLumi.
This JIRA ticket reports an workflow that goes to failed:
https://its.cern.ch/jira/browse/CMSCOMPPR-5394
and the reason is that the TimePerEvent is so so small that the job splitting gets automatically set to 42 million events per job (420k lumis per job), while it's requesting only 100k events.
| 1.0 | Don't set EventsPerJob bigger than RequestNumEvents - Of course, also considering the FilterEfficiency and the usual multiple of EventsPerLumi.
This JIRA ticket reports an workflow that goes to failed:
https://its.cern.ch/jira/browse/CMSCOMPPR-5394
and the reason is that the TimePerEvent is so so small that the job splitting gets automatically set to 42 million events per job (420k lumis per job), while it's requesting only 100k events.
| priority | don t set eventsperjob bigger than requestnumevents of course also considering the filterefficiency and the usual multiple of eventsperlumi this jira ticket reports an workflow that goes to failed and the reason is that the timeperevent is so so small that the job splitting gets automatically set to million events per job lumis per job while it s requesting only events | 1 |
782,727 | 27,504,944,200 | IssuesEvent | 2023-03-06 02:07:33 | AY2223S2-CS2103T-W13-3/tp | https://api.github.com/repos/AY2223S2-CS2103T-W13-3/tp | closed | As a fickle user, I want to edit a task | type.Story priority.Medium | So that I can correct mistakes without deleting a task. | 1.0 | As a fickle user, I want to edit a task - So that I can correct mistakes without deleting a task. | priority | as a fickle user i want to edit a task so that i can correct mistakes without deleting a task | 1 |
260,355 | 8,209,099,834 | IssuesEvent | 2018-09-04 06:10:27 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | Division search (website), PROD, #J247 | kind/support priority/medium | Доброго дня, зареєстровано три відділення ("legal_entity_id": "18791c31-3281-4e5a-9ea5-2eef9a6fe18e"):
1) "id": "b4985f5a-a2e4-4a52-b65f-59c8a7a90549","Амбулаторія загальної практики сімейної медицини № 1"
2) "id": "80904c4a-80e9-4680-9101-3c8d87bad1ba", "Амбулаторія загальної практики сімейної медицини № 2"
3) "id": "ac50fb91-36dc-449c-9205-61365e725dc5", "Амбулаторія загальної практики сімейної медицини № 3"
по запиту через API вони є, поточний статус "ACTIVE" і вказані координати, проте при пошуку на сайті https://portal.ehealth.gov.ua/divisions.html#/ їх не видно, як через ЄДРПОУ закладу так і по ід.
такий запит нічого не повертає хоча division guid зареєстрований
https://portal.ehealth.gov.ua/divisions.html#/b4985f5a-a2e4-4a52-b65f-59c8a7a90549
Скриншот здесь | 1.0 | Division search (website), PROD, #J247 - Доброго дня, зареєстровано три відділення ("legal_entity_id": "18791c31-3281-4e5a-9ea5-2eef9a6fe18e"):
1) "id": "b4985f5a-a2e4-4a52-b65f-59c8a7a90549","Амбулаторія загальної практики сімейної медицини № 1"
2) "id": "80904c4a-80e9-4680-9101-3c8d87bad1ba", "Амбулаторія загальної практики сімейної медицини № 2"
3) "id": "ac50fb91-36dc-449c-9205-61365e725dc5", "Амбулаторія загальної практики сімейної медицини № 3"
по запиту через API вони є, поточний статус "ACTIVE" і вказані координати, проте при пошуку на сайті https://portal.ehealth.gov.ua/divisions.html#/ їх не видно, як через ЄДРПОУ закладу так і по ід.
такий запит нічого не повертає хоча division guid зареєстрований
https://portal.ehealth.gov.ua/divisions.html#/b4985f5a-a2e4-4a52-b65f-59c8a7a90549
Скриншот здесь | priority | division search website prod доброго дня зареєстровано три відділення legal entity id id амбулаторія загальної практики сімейної медицини № id амбулаторія загальної практики сімейної медицини № id амбулаторія загальної практики сімейної медицини № по запиту через api вони є поточний статус active і вказані координати проте при пошуку на сайті їх не видно як через єдрпоу закладу так і по ід такий запит нічого не повертає хоча division guid зареєстрований скриншот здесь | 1 |
490,517 | 14,135,343,527 | IssuesEvent | 2020-11-10 01:28:07 | visit-dav/visit | https://api.github.com/repos/visit-dav/visit | opened | Add arbitrary simple shapes to plots | enhancement impact medium likelihood medium priority | Michael Hohensee would like the ability to add arbitrary simple shapes to the plots. For example, add a sphere or a cylinder.
### Describe alternatives you've considered.
Considered the sphere/line/box tool, but those are more for analytics. Also considered annotations, but that's not gonna work.
### Additional context
I don't know if Michael wants this feature just for visualization or if he wants the arbitrary shapes to be treated as additional data.
| 1.0 | Add arbitrary simple shapes to plots - Michael Hohensee would like the ability to add arbitrary simple shapes to the plots. For example, add a sphere or a cylinder.
### Describe alternatives you've considered.
Considered the sphere/line/box tool, but those are more for analytics. Also considered annotations, but that's not gonna work.
### Additional context
I don't know if Michael wants this feature just for visualization or if he wants the arbitrary shapes to be treated as additional data.
| priority | add arbitrary simple shapes to plots michael hohensee would like the ability to add arbitrary simple shapes to the plots for example add a sphere or a cylinder describe alternatives you ve considered considered the sphere line box tool but those are more for analytics also considered annotations but that s not gonna work additional context i don t know if michael wants this feature just for visualization or if he wants the arbitrary shapes to be treated as additional data | 1 |
637,186 | 20,622,963,517 | IssuesEvent | 2022-03-07 19:18:58 | supercrafter333/theSpawn | https://api.github.com/repos/supercrafter333/theSpawn | closed | [BUG:] TPA's cannot be awnsered | bug TO-DO priority: medium Status: Confirmed | # Bug: TPA's cannot be awnsered
### Informations
theSpawn Version: 1.6.1
Server-OS: Linux Ubuntu 20.04
PHP Version: 8.0.16
PocketMine-MP Version: 4.2.3+dev
### Error
None
### readjustment instructions
1. send a tpa to another player
2. the player should try to awnser the tpa
3. RESULT -> "You don't have any pending tpa" | 1.0 | [BUG:] TPA's cannot be awnsered - # Bug: TPA's cannot be awnsered
### Informations
theSpawn Version: 1.6.1
Server-OS: Linux Ubuntu 20.04
PHP Version: 8.0.16
PocketMine-MP Version: 4.2.3+dev
### Error
None
### readjustment instructions
1. send a tpa to another player
2. the player should try to awnser the tpa
3. RESULT -> "You don't have any pending tpa" | priority | tpa s cannot be awnsered bug tpa s cannot be awnsered informations thespawn version server os linux ubuntu php version pocketmine mp version dev error none readjustment instructions send a tpa to another player the player should try to awnser the tpa result you don t have any pending tpa | 1 |
624,120 | 19,687,212,903 | IssuesEvent | 2022-01-12 00:07:58 | PlaceOS/staff-api | https://api.github.com/repos/PlaceOS/staff-api | opened | add support for configurable booking limits on resources | type: enhancement product: placeos priority: medium focus: backend | The bookings controller provides an API for booking items (desks, car parking spaces etc)
We want to be able to put a limit on how many concurrent bookings a single user is allowed to have.
There is already code to check if there is a clash (two different people trying to book the same resource)
```crystal
# check there isn't a clashing booking
clashing_bookings = check_clashing(existing_booking)
render :conflict, json: clashing_bookings.first if clashing_bookings.size > 0
```
So we need a new bit of code at the same point that checks if the current user will go over any limits
* there might not be a limit
* if there is a limit then check the current request doesn't breach any limits
* when updating a booking the booking being updated can be ignored in the limit check (can't clash with itself)
* users can book on behalf of other users, the check needs to be against the user who is being allocated the resource
To implement this I was thinking that we
* modify the [tenant model](https://github.com/PlaceOS/staff-api/blob/master/src/models/tenant.cr#L49)
* add a [JSON column](https://github.com/PlaceOS/staff-api/blob/master/src/models/booking.cr#L53) called something like `booking_limits`
* the value of which will be `"desk" => 2` i.e. a user can book at most two desk resources at the same time
This way different domains can have different limits | 1.0 | add support for configurable booking limits on resources - The bookings controller provides an API for booking items (desks, car parking spaces etc)
We want to be able to put a limit on how many concurrent bookings a single user is allowed to have.
There is already code to check if there is a clash (two different people trying to book the same resource)
```crystal
# check there isn't a clashing booking
clashing_bookings = check_clashing(existing_booking)
render :conflict, json: clashing_bookings.first if clashing_bookings.size > 0
```
So we need a new bit of code at the same point that checks if the current user will go over any limits
* there might not be a limit
* if there is a limit then check the current request doesn't breach any limits
* when updating a booking the booking being updated can be ignored in the limit check (can't clash with itself)
* users can book on behalf of other users, the check needs to be against the user who is being allocated the resource
To implement this I was thinking that we
* modify the [tenant model](https://github.com/PlaceOS/staff-api/blob/master/src/models/tenant.cr#L49)
* add a [JSON column](https://github.com/PlaceOS/staff-api/blob/master/src/models/booking.cr#L53) called something like `booking_limits`
* the value of which will be `"desk" => 2` i.e. a user can book at most two desk resources at the same time
This way different domains can have different limits | priority | add support for configurable booking limits on resources the bookings controller provides an api for booking items desks car parking spaces etc we want to be able to put a limit on how many concurrent bookings a single user is allowed to have there is already code to check if there is a clash two different people trying to book the same resource crystal check there isn t a clashing booking clashing bookings check clashing existing booking render conflict json clashing bookings first if clashing bookings size so we need a new bit of code at the same point that checks if the current user will go over any limits there might not be a limit if there is a limit then check the current request doesn t breach any limits when updating a booking the booking being updated can be ignored in the limit check can t clash with itself users can book on behalf of other users the check needs to be against the user who is being allocated the resource to implement this i was thinking that we modify the add a called something like booking limits the value of which will be desk i e a user can book at most two desk resources at the same time this way different domains can have different limits | 1 |
632,336 | 20,192,452,930 | IssuesEvent | 2022-02-11 07:21:17 | asyml/forte | https://api.github.com/repos/asyml/forte | closed | Handle unknown attributes and unknown types | priority: medium e/5 | This issue is blocked by https://github.com/asyml/forte/issues/405.
**Is your feature request related to a problem? Please describe.**
Forte ontology defines the data types that can be used in the system, sometimes the ontology evolves by adding new classes or new attributes. Some existing pipelines will break (during deserialization) since they encountered unknown types and attributes.
Example:
The old ontology may define an `EntityMention`, with attribute `ner_type`. At some point we updated the definition, adding a new attribute, say `score` to `EntityMention`, now the `EntityMention` class changes from
```
class EntityMention:
ner_type = ""
```
to
```
class EntityMention:
ner_type = ""
score = 0
```
Loading data with the new types or attributes will break the deserialization.
**Describe the solution you'd like**
Forte can be robust to these changes when a boolean flag is set to `on`, in such cases, it should ignore unknown entry types.
On the other direction, when we load old data to a new pipeline, we can assign default `None` values to unknown entry attributes.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Handle unknown attributes and unknown types - This issue is blocked by https://github.com/asyml/forte/issues/405.
**Is your feature request related to a problem? Please describe.**
Forte ontology defines the data types that can be used in the system, sometimes the ontology evolves by adding new classes or new attributes. Some existing pipelines will break (during deserialization) since they encountered unknown types and attributes.
Example:
The old ontology may define an `EntityMention`, with attribute `ner_type`. At some point we updated the definition, adding a new attribute, say `score` to `EntityMention`, now the `EntityMention` class changes from
```
class EntityMention:
ner_type = ""
```
to
```
class EntityMention:
ner_type = ""
score = 0
```
Loading data with the new types or attributes will break the deserialization.
**Describe the solution you'd like**
Forte can be robust to these changes when a boolean flag is set to `on`, in such cases, it should ignore unknown entry types.
On the other direction, when we load old data to a new pipeline, we can assign default `None` values to unknown entry attributes.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | handle unknown attributes and unknown types this issue is blocked by is your feature request related to a problem please describe forte ontology defines the data types that can be used in the system sometimes the ontology evolves by adding new classes or new attributes some existing pipelines will break during deserialization since they encountered unknown types and attributes example the old ontology may define an entitymention with attribute ner type at some point we updated the definition adding a new attribute say score to entitymention now the entitymention class changes from class entitymention ner type to class entitymention ner type score loading data with the new types or attributes will break the deserialization describe the solution you d like forte can be robust to these changes when a boolean flag is set to on in such cases it should ignore unknown entry types on the other direction when we load old data to a new pipeline we can assign default none values to unknown entry attributes describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 1 |
54,493 | 3,068,413,051 | IssuesEvent | 2015-08-18 15:32:15 | TroyManary/EasyPlow | https://api.github.com/repos/TroyManary/EasyPlow | reopened | Smart Phone Home Page | Function: General Priority: Medium State: In Progress Type: Bug | As with all browsers not every browser including smart phone- the look and feel is not at all common. With the Smart Phone we even see a new bar below the header indicating "Snow Storm" a clock and share icon....

| 1.0 | Smart Phone Home Page - As with all browsers not every browser including smart phone- the look and feel is not at all common. With the Smart Phone we even see a new bar below the header indicating "Snow Storm" a clock and share icon....

| priority | smart phone home page as with all browsers not every browser including smart phone the look and feel is not at all common with the smart phone we even see a new bar below the header indicating snow storm a clock and share icon | 1 |
126,729 | 5,003,040,353 | IssuesEvent | 2016-12-11 18:18:33 | dteviot/WebToEpub | https://api.github.com/repos/dteviot/WebToEpub | closed | Parser issue: Skythewood | bug medium priority | I noticed that the skythewood.blogspot.it parser doesn't always grab the higher resolution image.
For example: "Altina the Sword Princess" volume 10.
There is a b/w picture (0002_p017.jpg) which is 412kB if you click on the pic when reading (1st chapter); the addon grabs just the 81kB preview you see on the page.
It happens every now and the on all of the Altina; it does not happen every time, since the addon usually grabs the correct picture.
I don't know if it happens with other LN on the site.
Can you look into it? Thanks. | 1.0 | Parser issue: Skythewood - I noticed that the skythewood.blogspot.it parser doesn't always grab the higher resolution image.
For example: "Altina the Sword Princess" volume 10.
There is a b/w picture (0002_p017.jpg) which is 412kB if you click on the pic when reading (1st chapter); the addon grabs just the 81kB preview you see on the page.
It happens every now and the on all of the Altina; it does not happen every time, since the addon usually grabs the correct picture.
I don't know if it happens with other LN on the site.
Can you look into it? Thanks. | priority | parser issue skythewood i noticed that the skythewood blogspot it parser doesn t always grab the higher resolution image for example altina the sword princess volume there is a b w picture jpg which is if you click on the pic when reading chapter the addon grabs just the preview you see on the page it happens every now and the on all of the altina it does not happen every time since the addon usually grabs the correct picture i don t know if it happens with other ln on the site can you look into it thanks | 1 |
755,944 | 26,448,651,893 | IssuesEvent | 2023-01-16 09:31:53 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | Warnings and other checks with symlinks | type: feature stage: in-progress priority: medium complex: low | - Compressing `tgz` files should warn if the file is not contained in the package.
- Compressing `tgz` files should convert a symlink to relative (as the file copier does) when a symlink is abs but contained in the package.
- Decompressing should warn when discarding a file: #4966
- File copier should warn/info for each thing ignored or discarded or converted.
See this branch containing some tests and incomplete/preliminary implementation: https://github.com/lasote/conan/pull/new/feature/warnings_and_tests | 1.0 | Warnings and other checks with symlinks - - Compressing `tgz` files should warn if the file is not contained in the package.
- Compressing `tgz` files should convert a symlink to relative (as the file copier does) when a symlink is abs but contained in the package.
- Decompressing should warn when discarding a file: #4966
- File copier should warn/info for each thing ignored or discarded or converted.
See this branch containing some tests and incomplete/preliminary implementation: https://github.com/lasote/conan/pull/new/feature/warnings_and_tests | priority | warnings and other checks with symlinks compressing tgz files should warn if the file is not contained in the package compressing tgz files should convert a symlink to relative as the file copier does when a symlink is abs but contained in the package decompressing should warn when discarding a file file copier should warn info for each thing ignored or discarded or converted see this branch containing some tests and incomplete preliminary implementation | 1 |
594,438 | 18,045,628,760 | IssuesEvent | 2021-09-18 21:08:20 | MudBlazor/MudBlazor | https://api.github.com/repos/MudBlazor/MudBlazor | closed | MudSelect dropdown should highlight selected item | enhancement Priority: Medium | ### Feature request type
Enhance component
### Component name
MudSelect
### Is your feature request related to a problem?
When an item has been selected and I open the select dropdown list, I can't see which item I selected from the list.
### Describe the solution you'd like
The selected item should be highlighted in the list similar to how it is highlighted in the Autocomplete component.
### Have you seen this feature anywhere else?
This is a commonly used pattern that can be seen in many apps:
Google fonts:
<img width="227" alt="Screen Shot 2021-09-17 at 10 35 21 AM" src="https://user-images.githubusercontent.com/17790790/133694379-4bcb6423-dc3d-45ac-b180-3ec2c1beca8a.png">
Google flights:
<img width="230" alt="Screen Shot 2021-09-17 at 10 35 44 AM" src="https://user-images.githubusercontent.com/17790790/133694383-df04250f-f105-4a65-8e09-df6da9055eaa.png">
### Describe alternatives you've considered
_No response_
### Pull Request
- [ ] I would like to do a Pull Request
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | 1.0 | MudSelect dropdown should highlight selected item - ### Feature request type
Enhance component
### Component name
MudSelect
### Is your feature request related to a problem?
When an item has been selected and I open the select dropdown list, I can't see which item I selected from the list.
### Describe the solution you'd like
The selected item should be highlighted in the list similar to how it is highlighted in the Autocomplete component.
### Have you seen this feature anywhere else?
This is a commonly used pattern that can be seen in many apps:
Google fonts:
<img width="227" alt="Screen Shot 2021-09-17 at 10 35 21 AM" src="https://user-images.githubusercontent.com/17790790/133694379-4bcb6423-dc3d-45ac-b180-3ec2c1beca8a.png">
Google flights:
<img width="230" alt="Screen Shot 2021-09-17 at 10 35 44 AM" src="https://user-images.githubusercontent.com/17790790/133694383-df04250f-f105-4a65-8e09-df6da9055eaa.png">
### Describe alternatives you've considered
_No response_
### Pull Request
- [ ] I would like to do a Pull Request
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | priority | mudselect dropdown should highlight selected item feature request type enhance component component name mudselect is your feature request related to a problem when an item has been selected and i open the select dropdown list i can t see which item i selected from the list describe the solution you d like the selected item should be highlighted in the list similar to how it is highlighted in the autocomplete component have you seen this feature anywhere else this is a commonly used pattern that can be seen in many apps google fonts img width alt screen shot at am src google flights img width alt screen shot at am src describe alternatives you ve considered no response pull request i would like to do a pull request code of conduct i agree to follow this project s code of conduct | 1 |
716,268 | 24,626,734,044 | IssuesEvent | 2022-10-16 16:10:30 | AY2223S1-CS2103T-W08-3/tp | https://api.github.com/repos/AY2223S1-CS2103T-W08-3/tp | closed | As a lazy user, I want to be able to open the github profile page of addresses in my address book with a command | priority.Medium type.Story | so that I can view my friends/Teaching Assistants/Professors github projects easily. | 1.0 | As a lazy user, I want to be able to open the github profile page of addresses in my address book with a command - so that I can view my friends/Teaching Assistants/Professors github projects easily. | priority | as a lazy user i want to be able to open the github profile page of addresses in my address book with a command so that i can view my friends teaching assistants professors github projects easily | 1 |
312,792 | 9,553,115,029 | IssuesEvent | 2019-05-02 18:24:16 | phetsims/fraction-matcher | https://api.github.com/repos/phetsims/fraction-matcher | closed | Some previously translated strings are no longer translated | priority:3-medium status:blocks-sim-publication status:ready-for-review | The level selection screen for the published Farsi (Persian) version of Fraction Matcher looks like this:

On the current master version, using locale=fa, it looks like this:

The reason that English words are now appearing is that a number of strings were moved from the fraction-matcher repo to fractions-common during the recent work on the fractions suite, and the translated strings weren't moved over. This should probably be fixed, otherwise the next time Fraction Master is published off of master, it may cause existing translations to fall back to English in several places as seen above.
I'm guessing it would be a couple hours of work max to either propagate the strings manually or write a script to do it. Either @jonathanolson or I could do it, or perhaps someone who we want to get a better understanding of how the translation utility works. Assigning to @ariel-phet for prioritization and assignment. | 1.0 | Some previously translated strings are no longer translated - The level selection screen for the published Farsi (Persian) version of Fraction Matcher looks like this:

On the current master version, using locale=fa, it looks like this:

The reason that English words are now appearing is that a number of strings were moved from the fraction-matcher repo to fractions-common during the recent work on the fractions suite, and the translated strings weren't moved over. This should probably be fixed, otherwise the next time Fraction Master is published off of master, it may cause existing translations to fall back to English in several places as seen above.
I'm guessing it would be a couple hours of work max to either propagate the strings manually or write a script to do it. Either @jonathanolson or I could do it, or perhaps someone who we want to get a better understanding of how the translation utility works. Assigning to @ariel-phet for prioritization and assignment. | priority | some previously translated strings are no longer translated the level selection screen for the published farsi persian version of fraction matcher looks like this on the current master version using locale fa it looks like this the reason that english words are now appearing is that a number of strings were moved from the fraction matcher repo to fractions common during the recent work on the fractions suite and the translated strings weren t moved over this should probably be fixed otherwise the next time fraction master is published off of master it may cause existing translations to fall back to english in several places as seen above i m guessing it would be a couple hours of work max to either propagate the strings manually or write a script to do it either jonathanolson or i could do it or perhaps someone who we want to get a better understanding of how the translation utility works assigning to ariel phet for prioritization and assignment | 1 |
401,550 | 11,795,107,200 | IssuesEvent | 2020-03-18 08:16:18 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | The user with deleted account appears as 'Anonymous' in forum discussion. | bug priority: medium | **Describe the bug**
if a user deleted his own account, His data is not deleted from forum and displayed as 'anonymous' user.
https://prnt.sc/rere3y
**To Reproduce**
Steps to reproduce the behavior:
1.Log in as a test user.
2. Post in discussion.
3. Go to Account > Delete Account
4. Mark check ' I understand the consequence' and click on 'Delete Account'.
5. Log in again as another user. Go the same discussion.
6. The user appears as 'Anonymous'.
**Expected behavior**
The content should be deleted.
**Screenshots**
https://prnt.sc/rere3y
**Support ticket links**
https://buddyboss.zendesk.com/agent/tickets/63594
| 1.0 | The user with deleted account appears as 'Anonymous' in forum discussion. - **Describe the bug**
if a user deleted his own account, His data is not deleted from forum and displayed as 'anonymous' user.
https://prnt.sc/rere3y
**To Reproduce**
Steps to reproduce the behavior:
1.Log in as a test user.
2. Post in discussion.
3. Go to Account > Delete Account
4. Mark check ' I understand the consequence' and click on 'Delete Account'.
5. Log in again as another user. Go the same discussion.
6. The user appears as 'Anonymous'.
**Expected behavior**
The content should be deleted.
**Screenshots**
https://prnt.sc/rere3y
**Support ticket links**
https://buddyboss.zendesk.com/agent/tickets/63594
| priority | the user with deleted account appears as anonymous in forum discussion describe the bug if a user deleted his own account his data is not deleted from forum and displayed as anonymous user to reproduce steps to reproduce the behavior log in as a test user post in discussion go to account delete account mark check i understand the consequence and click on delete account log in again as another user go the same discussion the user appears as anonymous expected behavior the content should be deleted screenshots support ticket links | 1 |
683,978 | 23,401,577,879 | IssuesEvent | 2022-08-12 08:32:13 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | sample.drivers.flash.shell: Failed on atmel targets | bug priority: medium platform: Microchip SAM | **Describe the bug**
Seen here: https://github.com/zephyrproject-rtos/zephyr/runs/7769981753?check_suite_focus=true
`samples/drivers/flash_shell/sample.drivers.flash.shell ` is failed on following targets:
- sam4s_xplained
- sam4l_ek
- sam4e_xpro
- arduino_due
**Log Error**
```
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_get_page':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:88:18: error: 'IFLASH_PAGE_SIZE' undeclared (first use in this function); did you mean 'IFLASH0_PAGE_SIZE'?
88 | return offset / IFLASH_PAGE_SIZE;
| ^~~~~~~~~~~~~~~~
| IFLASH0_PAGE_SIZE
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:88:18: note: each undeclared identifier is reported only once for each function it appears in
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_wait_ready':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:108:13: error: 'EEFC_FSR_FLERR' undeclared (first use in this function); did you mean 'EEFC_FSR_FRDY'?
108 | if (fsr & EEFC_FSR_FLERR) {
| ^~~~~~~~~~~~~~
| EEFC_FSR_FRDY
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_write':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:204:27: error: 'IFLASH_PAGE_SIZE' undeclared (first use in this function); did you mean 'IFLASH0_PAGE_SIZE'?
204 | eop_len = -(offset | ~(IFLASH_PAGE_SIZE - 1));
| ^~~~~~~~~~~~~~~~
| IFLASH0_PAGE_SIZE
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_erase_block':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:255:4: error: 'EEFC_FCR_FCMD_EPA' undeclared (first use in this function); did you mean 'EEFC_FCR_FCMD_EA'?
255 | EEFC_FCR_FCMD_EPA;
| ^~~~~~~~~~~~~~~~~
| EEFC_FCR_FCMD_EA
In file included from /local/mcu/zephyrproject/zephyr/include/zephyr/device.h:29,
from /local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:13:
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_erase':
/local/mcu/zephyrproject/zephyr/twister-out/arduino_due/samples/drivers/flash_shell/sample.drivers.flash.shell/zephyr/include/generated/devicetree_unfixed.h:2883:34: error: 'DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000_P_erase_block_size' undeclared (first use in this function); did you mean 'DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000_P_write_block_size'?
2883 | #define DT_N_INST_0_soc_nv_flash DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
```
**To Reproduce**
`twister -b -s samples/drivers/flash_shell/sample.drivers.flash.shell -p sam4s_xplained`
**Impact**
Blocking CI (https://github.com/zephyrproject-rtos/zephyr/pull/45221)
**Environment (please complete the following information):**
zephyr-v3.1.0-3289-gc93361a5bf
| 1.0 | sample.drivers.flash.shell: Failed on atmel targets - **Describe the bug**
Seen here: https://github.com/zephyrproject-rtos/zephyr/runs/7769981753?check_suite_focus=true
`samples/drivers/flash_shell/sample.drivers.flash.shell ` is failed on following targets:
- sam4s_xplained
- sam4l_ek
- sam4e_xpro
- arduino_due
**Log Error**
```
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_get_page':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:88:18: error: 'IFLASH_PAGE_SIZE' undeclared (first use in this function); did you mean 'IFLASH0_PAGE_SIZE'?
88 | return offset / IFLASH_PAGE_SIZE;
| ^~~~~~~~~~~~~~~~
| IFLASH0_PAGE_SIZE
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:88:18: note: each undeclared identifier is reported only once for each function it appears in
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_wait_ready':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:108:13: error: 'EEFC_FSR_FLERR' undeclared (first use in this function); did you mean 'EEFC_FSR_FRDY'?
108 | if (fsr & EEFC_FSR_FLERR) {
| ^~~~~~~~~~~~~~
| EEFC_FSR_FRDY
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_write':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:204:27: error: 'IFLASH_PAGE_SIZE' undeclared (first use in this function); did you mean 'IFLASH0_PAGE_SIZE'?
204 | eop_len = -(offset | ~(IFLASH_PAGE_SIZE - 1));
| ^~~~~~~~~~~~~~~~
| IFLASH0_PAGE_SIZE
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_erase_block':
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:255:4: error: 'EEFC_FCR_FCMD_EPA' undeclared (first use in this function); did you mean 'EEFC_FCR_FCMD_EA'?
255 | EEFC_FCR_FCMD_EPA;
| ^~~~~~~~~~~~~~~~~
| EEFC_FCR_FCMD_EA
In file included from /local/mcu/zephyrproject/zephyr/include/zephyr/device.h:29,
from /local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c:13:
/local/mcu/zephyrproject/zephyr/drivers/flash/flash_sam.c: In function 'flash_sam_erase':
/local/mcu/zephyrproject/zephyr/twister-out/arduino_due/samples/drivers/flash_shell/sample.drivers.flash.shell/zephyr/include/generated/devicetree_unfixed.h:2883:34: error: 'DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000_P_erase_block_size' undeclared (first use in this function); did you mean 'DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000_P_write_block_size'?
2883 | #define DT_N_INST_0_soc_nv_flash DT_N_S_soc_S_flash_controller_400e0a00_S_flash_80000
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
```
**To Reproduce**
`twister -b -s samples/drivers/flash_shell/sample.drivers.flash.shell -p sam4s_xplained`
**Impact**
Blocking CI (https://github.com/zephyrproject-rtos/zephyr/pull/45221)
**Environment (please complete the following information):**
zephyr-v3.1.0-3289-gc93361a5bf
| priority | sample drivers flash shell failed on atmel targets describe the bug seen here samples drivers flash shell sample drivers flash shell is failed on following targets xplained ek xpro arduino due log error local mcu zephyrproject zephyr drivers flash flash sam c in function flash sam get page local mcu zephyrproject zephyr drivers flash flash sam c error iflash page size undeclared first use in this function did you mean page size return offset iflash page size page size local mcu zephyrproject zephyr drivers flash flash sam c note each undeclared identifier is reported only once for each function it appears in local mcu zephyrproject zephyr drivers flash flash sam c in function flash sam wait ready local mcu zephyrproject zephyr drivers flash flash sam c error eefc fsr flerr undeclared first use in this function did you mean eefc fsr frdy if fsr eefc fsr flerr eefc fsr frdy local mcu zephyrproject zephyr drivers flash flash sam c in function flash sam write local mcu zephyrproject zephyr drivers flash flash sam c error iflash page size undeclared first use in this function did you mean page size eop len offset iflash page size page size local mcu zephyrproject zephyr drivers flash flash sam c in function flash sam erase block local mcu zephyrproject zephyr drivers flash flash sam c error eefc fcr fcmd epa undeclared first use in this function did you mean eefc fcr fcmd ea eefc fcr fcmd epa eefc fcr fcmd ea in file included from local mcu zephyrproject zephyr include zephyr device h from local mcu zephyrproject zephyr drivers flash flash sam c local mcu zephyrproject zephyr drivers flash flash sam c in function flash sam erase local mcu zephyrproject zephyr twister out arduino due samples drivers flash shell sample drivers flash shell zephyr include generated devicetree unfixed h error dt n s soc s flash controller s flash p erase block size undeclared first use in this function did you mean dt n s soc s flash controller s flash p write block size define dt n inst soc nv flash dt n s soc s flash controller s flash to reproduce twister b s samples drivers flash shell sample drivers flash shell p xplained impact blocking ci environment please complete the following information zephyr | 1 |
205,200 | 7,094,693,339 | IssuesEvent | 2018-01-13 07:06:12 | facelessuser/backrefs | https://api.github.com/repos/facelessuser/backrefs | closed | Ignore comments | Bug Priority - Medium Severity - Major | I never use them, but this is kind of a big oversight. `(?#comment)` should not have content processed. | 1.0 | Ignore comments - I never use them, but this is kind of a big oversight. `(?#comment)` should not have content processed. | priority | ignore comments i never use them but this is kind of a big oversight comment should not have content processed | 1 |
43,851 | 2,893,439,709 | IssuesEvent | 2015-06-15 17:59:51 | SteamDatabase/steamSummerMinigame | https://api.github.com/repos/SteamDatabase/steamSummerMinigame | closed | Autobuy Cheapest Upgrade | 2 - Medium Priority | If Clicking messes up the autoplayer and the goal is efficiency, then we need some way of the game simply buying the cheapest available upgrade to continue effective progression without crippling the script itself. | 1.0 | Autobuy Cheapest Upgrade - If Clicking messes up the autoplayer and the goal is efficiency, then we need some way of the game simply buying the cheapest available upgrade to continue effective progression without crippling the script itself. | priority | autobuy cheapest upgrade if clicking messes up the autoplayer and the goal is efficiency then we need some way of the game simply buying the cheapest available upgrade to continue effective progression without crippling the script itself | 1 |
548,759 | 16,075,210,722 | IssuesEvent | 2021-04-25 08:04:34 | dodona-edu/dodona | https://api.github.com/repos/dodona-edu/dodona | opened | DNS migration | medium priority | This issue tracks the todo's of the DNS migration
- [ ] register dodona.be
- [ ] point name servers to cloudflare
- [ ] set A and CNAME records (https://github.com/dodona-edu/dodona-ansible/wiki/DNS-records)
- [ ] add domains to ansible (let's encrypt) https://github.com/dodona-edu/dodona-ansible/pull/117
- [ ] rewrite dodona.be and www.dodona.be to dodona.ugent.be on apache
- [ ] rewrite naos.dodona.be to naos.ugent.be on apache
- [ ] rewrite mestra.dodona.be to mestra.ugent.be on apache
- [ ] add sandbox.dodona.be, naos-sandbox.dodona.be and mestra-sandbox.dodona.be to the CSP headers
- [ ] serve the exercises from the new sandbox domains | 1.0 | DNS migration - This issue tracks the todo's of the DNS migration
- [ ] register dodona.be
- [ ] point name servers to cloudflare
- [ ] set A and CNAME records (https://github.com/dodona-edu/dodona-ansible/wiki/DNS-records)
- [ ] add domains to ansible (let's encrypt) https://github.com/dodona-edu/dodona-ansible/pull/117
- [ ] rewrite dodona.be and www.dodona.be to dodona.ugent.be on apache
- [ ] rewrite naos.dodona.be to naos.ugent.be on apache
- [ ] rewrite mestra.dodona.be to mestra.ugent.be on apache
- [ ] add sandbox.dodona.be, naos-sandbox.dodona.be and mestra-sandbox.dodona.be to the CSP headers
- [ ] serve the exercises from the new sandbox domains | priority | dns migration this issue tracks the todo s of the dns migration register dodona be point name servers to cloudflare set a and cname records add domains to ansible let s encrypt rewrite dodona be and to dodona ugent be on apache rewrite naos dodona be to naos ugent be on apache rewrite mestra dodona be to mestra ugent be on apache add sandbox dodona be naos sandbox dodona be and mestra sandbox dodona be to the csp headers serve the exercises from the new sandbox domains | 1 |
317,932 | 9,671,280,169 | IssuesEvent | 2019-05-21 22:15:16 | GingerWalnut/SQBeyondPublic | https://api.github.com/repos/GingerWalnut/SQBeyondPublic | closed | SQ Tech Machines not giving blocks back if block is moving | medium priority | So, If a piston is pushing a block when the SQ Tech Machine tries to break it, then it does not give the block back. This is fairly easy to see by comparing the two example machines in the Screenshot breaking blocks. The one on the right pushes a block in front to the breaker each cycle, making sure that there is never a moving block in front of the breaker. This one returns all the blocks placed. The one on the left just pushes them in as fast as possible, which means that sometimes the block in front of the breaker is moving when it activates. This one looses blocks as it goes.
I think a fix might be making the breaker unable to break block 42 (or whatever the moving block thing has turned into after 1.13).

| 1.0 | SQ Tech Machines not giving blocks back if block is moving - So, If a piston is pushing a block when the SQ Tech Machine tries to break it, then it does not give the block back. This is fairly easy to see by comparing the two example machines in the Screenshot breaking blocks. The one on the right pushes a block in front to the breaker each cycle, making sure that there is never a moving block in front of the breaker. This one returns all the blocks placed. The one on the left just pushes them in as fast as possible, which means that sometimes the block in front of the breaker is moving when it activates. This one looses blocks as it goes.
I think a fix might be making the breaker unable to break block 42 (or whatever the moving block thing has turned into after 1.13).

| priority | sq tech machines not giving blocks back if block is moving so if a piston is pushing a block when the sq tech machine tries to break it then it does not give the block back this is fairly easy to see by comparing the two example machines in the screenshot breaking blocks the one on the right pushes a block in front to the breaker each cycle making sure that there is never a moving block in front of the breaker this one returns all the blocks placed the one on the left just pushes them in as fast as possible which means that sometimes the block in front of the breaker is moving when it activates this one looses blocks as it goes i think a fix might be making the breaker unable to break block or whatever the moving block thing has turned into after | 1 |
287,173 | 8,805,284,325 | IssuesEvent | 2018-12-26 18:39:26 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | В группу ассортимента добавить поле Лимит. | Meat Priority-Medium Type-Enhancement | Originally reported on Google Code with ID 2189
```
Если добавляем сырьё как заменитель в рецепт, и если это сырьё уже входит в рецепт то
дать возможность добавить. Если добавляем не как заменитель, а как позицию рецепта,
то показать сообщение что сырьё уже присутствует.
```
Reported by `stasgm` on 2010-10-20 13:41:14
| 1.0 | В группу ассортимента добавить поле Лимит. - Originally reported on Google Code with ID 2189
```
Если добавляем сырьё как заменитель в рецепт, и если это сырьё уже входит в рецепт то
дать возможность добавить. Если добавляем не как заменитель, а как позицию рецепта,
то показать сообщение что сырьё уже присутствует.
```
Reported by `stasgm` on 2010-10-20 13:41:14
| priority | в группу ассортимента добавить поле лимит originally reported on google code with id если добавляем сырьё как заменитель в рецепт и если это сырьё уже входит в рецепт то дать возможность добавить если добавляем не как заменитель а как позицию рецепта то показать сообщение что сырьё уже присутствует reported by stasgm on | 1 |
317,332 | 9,663,456,805 | IssuesEvent | 2019-05-21 00:43:47 | medic/medic | https://api.github.com/repos/medic/medic | closed | Import people blocked by CSP | Configuration Priority: 2 - Medium Type: Bug | **Describe the bug**
Uploading people using the admin app isn't working.
**To Reproduce**
1. Create a JSON file with some people data in it. A minimal example:
`[{"_id": "a", "name": "something"}]`
1. Go to Admin app > Import & Export > People
2. Click on the Browse button in the Import section
3. Select the file you created in 1.
4. Click Submit
5. An error is shown in the UI. There's an error in the log. The person is not imported.
**Expected behavior**
No errors are shown and a contact is imported.
**Logs**
```
Content Security Policy: The page’s settings blocked the loading of a resource at http://_db/undefined (“connect-src”).
```
**Screenshots**


**Environment**
- Instance: localhost
- Browser: Firefox
- Client platform: Ubuntu
- App: admin
- Version: 3.5.0 (master)
**Additional context**
1. Before starting on this bug consider if we can drop this feature altogether. medic-conf provides similar functionality.
2. If it is worth fixing, consider refactoring it to use the existing APIs for importing people and places.
3. Also consider using bulk APIs instead of one call per contact.
| 1.0 | Import people blocked by CSP - **Describe the bug**
Uploading people using the admin app isn't working.
**To Reproduce**
1. Create a JSON file with some people data in it. A minimal example:
`[{"_id": "a", "name": "something"}]`
1. Go to Admin app > Import & Export > People
2. Click on the Browse button in the Import section
3. Select the file you created in 1.
4. Click Submit
5. An error is shown in the UI. There's an error in the log. The person is not imported.
**Expected behavior**
No errors are shown and a contact is imported.
**Logs**
```
Content Security Policy: The page’s settings blocked the loading of a resource at http://_db/undefined (“connect-src”).
```
**Screenshots**


**Environment**
- Instance: localhost
- Browser: Firefox
- Client platform: Ubuntu
- App: admin
- Version: 3.5.0 (master)
**Additional context**
1. Before starting on this bug consider if we can drop this feature altogether. medic-conf provides similar functionality.
2. If it is worth fixing, consider refactoring it to use the existing APIs for importing people and places.
3. Also consider using bulk APIs instead of one call per contact.
| priority | import people blocked by csp describe the bug uploading people using the admin app isn t working to reproduce create a json file with some people data in it a minimal example go to admin app import export people click on the browse button in the import section select the file you created in click submit an error is shown in the ui there s an error in the log the person is not imported expected behavior no errors are shown and a contact is imported logs content security policy the page’s settings blocked the loading of a resource at “connect src” screenshots environment instance localhost browser firefox client platform ubuntu app admin version master additional context before starting on this bug consider if we can drop this feature altogether medic conf provides similar functionality if it is worth fixing consider refactoring it to use the existing apis for importing people and places also consider using bulk apis instead of one call per contact | 1 |
155,565 | 5,957,006,590 | IssuesEvent | 2017-05-28 22:04:45 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Windows upgrader doesn't always close bitfighter | 019c 020 bug imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on December 01, 2013 04:20:37_
On some Windows machines (but not all), the upgrader does not close the running Bitfighter session. This will cause the upgrade to fail if the user does not close the window themselves.
At a minimum, we should add a message to the upgrade window telling people they need to make sure the window is closed. Also, we should figure out why this happens.
I have one machine where I can reliably reproduce.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=322_
| 1.0 | Windows upgrader doesn't always close bitfighter - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on December 01, 2013 04:20:37_
On some Windows machines (but not all), the upgrader does not close the running Bitfighter session. This will cause the upgrade to fail if the user does not close the window themselves.
At a minimum, we should add a message to the upgrade window telling people they need to make sure the window is closed. Also, we should figure out why this happens.
I have one machine where I can reliably reproduce.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=322_
| priority | windows upgrader doesn t always close bitfighter from on december on some windows machines but not all the upgrader does not close the running bitfighter session this will cause the upgrade to fail if the user does not close the window themselves at a minimum we should add a message to the upgrade window telling people they need to make sure the window is closed also we should figure out why this happens i have one machine where i can reliably reproduce original issue | 1 |
490,682 | 14,138,607,515 | IssuesEvent | 2020-11-10 08:41:12 | vmware/singleton | https://api.github.com/repos/vmware/singleton | closed | [ENHANCEMENT] Optimize VIPService by removing unused code. Automate initialization of VIPService in VIPCfg. | area/java-client kind/enhancement priority/medium | **Is your feature request related to a problem? Please describe.**
VIPService has redundant/unused properties such as productID and version properties.
Singleton pattern is not needed for VIPService because VIPService is a member of VIPCfg.
Initialization of VIPService is unnecessarily a separate API call after VIPCfg.initialize.
**Describe the solution you'd like**
Optimize VIPService by removing unused code such as productId and version properties.
No need to have a singleton instance of VIPService because VIPService is a member of VIPCfg.
Automate instantiation of VIPService in VIPCfg. Do not allow instantiation of VIPCfg with null vipServer or null/faultyHttpRequester | 1.0 | [ENHANCEMENT] Optimize VIPService by removing unused code. Automate initialization of VIPService in VIPCfg. - **Is your feature request related to a problem? Please describe.**
VIPService has redundant/unused properties such as productID and version properties.
Singleton pattern is not needed for VIPService because VIPService is a member of VIPCfg.
Initialization of VIPService is unnecessarily a separate API call after VIPCfg.initialize.
**Describe the solution you'd like**
Optimize VIPService by removing unused code such as productId and version properties.
No need to have a singleton instance of VIPService because VIPService is a member of VIPCfg.
Automate instantiation of VIPService in VIPCfg. Do not allow instantiation of VIPCfg with null vipServer or null/faultyHttpRequester | priority | optimize vipservice by removing unused code automate initialization of vipservice in vipcfg is your feature request related to a problem please describe vipservice has redundant unused properties such as productid and version properties singleton pattern is not needed for vipservice because vipservice is a member of vipcfg initialization of vipservice is unnecessarily a separate api call after vipcfg initialize describe the solution you d like optimize vipservice by removing unused code such as productid and version properties no need to have a singleton instance of vipservice because vipservice is a member of vipcfg automate instantiation of vipservice in vipcfg do not allow instantiation of vipcfg with null vipserver or null faultyhttprequester | 1 |
268,225 | 8,405,040,351 | IssuesEvent | 2018-10-11 14:21:07 | geosolutions-it/smb-app | https://api.github.com/repos/geosolutions-it/smb-app | closed | Landing View | Priority: Medium review | Create Landing view to show call to actions buttons to relevant functionalities (tracking, lost/found notifications, etc.) with short description.
This will substituto the current "Tracks registration" view as the landing view.
In the future it could also host news and messages | 1.0 | Landing View - Create Landing view to show call to actions buttons to relevant functionalities (tracking, lost/found notifications, etc.) with short description.
This will substituto the current "Tracks registration" view as the landing view.
In the future it could also host news and messages | priority | landing view create landing view to show call to actions buttons to relevant functionalities tracking lost found notifications etc with short description this will substituto the current tracks registration view as the landing view in the future it could also host news and messages | 1 |
22,141 | 2,645,690,087 | IssuesEvent | 2015-03-13 01:08:34 | prikhi/evoluspencil | https://api.github.com/repos/prikhi/evoluspencil | closed | Allow reordering of page tabs | 1 star bug imported Priority-Medium | _From [brownsp...@gmail.com](https://code.google.com/u/100468905672215389538/) on August 22, 2008 06:38:15_
What steps will reproduce the problem? 1. Create a new document, one page available.
2. Add a new page.
3. Add another new page, but you want the tab to appear next to the first one. What is the expected output? What do you see instead? I would like to have some sort of function to facilitate reordering of the
page tabs, for organization. Drag-and-drop just like Firefox does would be
nice. What version of the product are you using? On what operating system? 1.0 standalone on Windows XP. Please provide any additional information below.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=34_ | 1.0 | Allow reordering of page tabs - _From [brownsp...@gmail.com](https://code.google.com/u/100468905672215389538/) on August 22, 2008 06:38:15_
What steps will reproduce the problem? 1. Create a new document, one page available.
2. Add a new page.
3. Add another new page, but you want the tab to appear next to the first one. What is the expected output? What do you see instead? I would like to have some sort of function to facilitate reordering of the
page tabs, for organization. Drag-and-drop just like Firefox does would be
nice. What version of the product are you using? On what operating system? 1.0 standalone on Windows XP. Please provide any additional information below.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=34_ | priority | allow reordering of page tabs from on august what steps will reproduce the problem create a new document one page available add a new page add another new page but you want the tab to appear next to the first one what is the expected output what do you see instead i would like to have some sort of function to facilitate reordering of the page tabs for organization drag and drop just like firefox does would be nice what version of the product are you using on what operating system standalone on windows xp please provide any additional information below original issue | 1 |
345,432 | 10,367,956,770 | IssuesEvent | 2019-09-07 13:05:56 | eternialz/moeverdose | https://api.github.com/repos/eternialz/moeverdose | closed | Main menu isn't working correctly in Edge <= 18 | Bug Priority:Medium Theme:Frontend | The main-menu isn't displaying at all using Microsoft Edge | 1.0 | Main menu isn't working correctly in Edge <= 18 - The main-menu isn't displaying at all using Microsoft Edge | priority | main menu isn t working correctly in edge the main menu isn t displaying at all using microsoft edge | 1 |
115,808 | 4,682,506,958 | IssuesEvent | 2016-10-09 09:27:29 | CS2103AUG2016-F10-C2/main | https://api.github.com/repos/CS2103AUG2016-F10-C2/main | closed | Logic parser: Implement a better parser algo | priority.medium status.ongoing type.enchancement type.epic | Currently, if the user messes up the order of an valid command (such as putting the tags before the description of an add command), the parser will deem it as an invalid command when in essence it is a totally valid command with all the required information to add a new task.
Thus this overhaul plans to introduce more flexibility in the parser class in order to make commands less rigid and supports the easy addition of new arguments to new and existing commands.
### Examples
---
**Example 1:** Valid command
```java
ArgumentsParser parser = new ArgumentsParser() ;
parser.addNoFlagsArgs(CommandArgs.NAME).addOptionalArg(CommandArgs.DESC).addOptionalArg(CommandArgs.TAGS) ;
parser.parse ("add hello d/hi e/tag1 e/tag2") ;
```
> parser.getArgValue (CommandArgs.NAME) = "hello" <br>
> parser.getArgValue (CommandArgs.DESC) = "hi" <br>
> parser.getArgValue (CommandArgs.TAGS) = {tag1, tag2} <br>
**Example 2:** Required Arguments not present
```java
ArgumentsParser parser = new ArgumentsParser() ;
parser.addNoFlagsArgs(CommandArgs.NAME).addOptionalArg(CommandArgs.DESC).addOptionalArg(CommandArgs.TAGS) ;
parser.parse ("add") ;
```
> IllegalValueException : Invalid command format | 1.0 | Logic parser: Implement a better parser algo - Currently, if the user messes up the order of an valid command (such as putting the tags before the description of an add command), the parser will deem it as an invalid command when in essence it is a totally valid command with all the required information to add a new task.
Thus this overhaul plans to introduce more flexibility in the parser class in order to make commands less rigid and supports the easy addition of new arguments to new and existing commands.
### Examples
---
**Example 1:** Valid command
```java
ArgumentsParser parser = new ArgumentsParser() ;
parser.addNoFlagsArgs(CommandArgs.NAME).addOptionalArg(CommandArgs.DESC).addOptionalArg(CommandArgs.TAGS) ;
parser.parse ("add hello d/hi e/tag1 e/tag2") ;
```
> parser.getArgValue (CommandArgs.NAME) = "hello" <br>
> parser.getArgValue (CommandArgs.DESC) = "hi" <br>
> parser.getArgValue (CommandArgs.TAGS) = {tag1, tag2} <br>
**Example 2:** Required Arguments not present
```java
ArgumentsParser parser = new ArgumentsParser() ;
parser.addNoFlagsArgs(CommandArgs.NAME).addOptionalArg(CommandArgs.DESC).addOptionalArg(CommandArgs.TAGS) ;
parser.parse ("add") ;
```
> IllegalValueException : Invalid command format | priority | logic parser implement a better parser algo currently if the user messes up the order of an valid command such as putting the tags before the description of an add command the parser will deem it as an invalid command when in essence it is a totally valid command with all the required information to add a new task thus this overhaul plans to introduce more flexibility in the parser class in order to make commands less rigid and supports the easy addition of new arguments to new and existing commands examples example valid command java argumentsparser parser new argumentsparser parser addnoflagsargs commandargs name addoptionalarg commandargs desc addoptionalarg commandargs tags parser parse add hello d hi e e parser getargvalue commandargs name hello parser getargvalue commandargs desc hi parser getargvalue commandargs tags example required arguments not present java argumentsparser parser new argumentsparser parser addnoflagsargs commandargs name addoptionalarg commandargs desc addoptionalarg commandargs tags parser parse add illegalvalueexception invalid command format | 1 |
768,770 | 26,979,695,394 | IssuesEvent | 2023-02-09 12:12:27 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Bluetooth: Host: Periodic scanner does not differentiate between partial and incomplete data | bug priority: medium area: Bluetooth area: Bluetooth Host | **Describe the bug**
When the periodic scanner is reconstructing the periodic advertising data to report to the application, it only checks if the data_status is marked as complete or not, it does not differentiate between incomplete and partial data, see [here](https://github.com/zephyrproject-rtos/zephyr/blob/main/subsys/bluetooth/host/scan.c#L766-L835). In the case where the controller does not receive the data complete packet and the data_status is marked as "incomplete, no more data to come", the data of the next incoming periodic advertising event is appended to the truncated data resulting in incorrect data being passed to the application.
**To Reproduce**
This is quite a difficult bug to reproduce as it relies on a packet being dropped. The test environment where this was discovered had a very high RSSI just above the threshold of the radios sensitivity which increased the likely of the packet not being received.
Steps to reproduce the behavior:
1. Establish a synchronization between a periodic advertiser and scanner. The periodic advertiser should be advertising known data that is long enough to be split across multiple packets.
2. Repeatedly check the data received by the scanner and compare it with the expected advertised data.
3. When one of the packets to construct the data is not received the data should differ between the received data and the expected data
**Expected behavior**
The incomplete data should not be passed up to the application and should be dropped.
**Impact**
The application does not receive the correct data from the host.
**Logs and console output**
Expected data:
```
\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xf60\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,\x03'
```
Received data:
```
\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xf60\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,\x03'
```
In the following logs, the data total data length of the advertised data was 300bytes, which was split into 3 packets: 247, 3, 50. After logging the data length of the incoming periodic advertisement packets before they're reconstructed, it can be seen in the following logs that right before the test failed, a packet of data length 50 (the final data in the sequence) was dropped. This resulted in the data being 250 bytes longer than expected as it appended 247, 3, 247, 3 and 50 bytes to the data buffer for the advertisement report. (I've indented the events of interest)
```
\*\*\* Booting Zephyr OS build v3.2.99-ncs1-1495-gaf6add53aa1f \*\*\*
[00:00:01.836,822] <inf> bt_hci_core: hci_vs_init: HW Platform: Nordic Semiconductor (0x0002)
[00:00:01.836,883] <inf> bt_hci_core: hci_vs_init: HW Variant: nRF53x (0x0003)
[00:00:01.836,914] <inf> bt_hci_core: hci_vs_init: Firmware: Standard Bluetooth controller (0x00) Version 184.62253 Build 1392531289
[00:00:01.839,080] <inf> bt_hci_core: bt_dev_show_info: Identity: D6:30:BF:22:A4:15 (random)
[00:00:01.839,111] <inf> bt_hci_core: bt_dev_show_info: HCI: version 5.3 (0x0c) revision 0x221c, manufacturer 0x0059
[00:00:01.839,141] <inf> bt_hci_core: bt_dev_show_info: LMP: version 5.3 (0x0c) subver 0x221c
[00:00:07.969,604] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:07.969,726] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:07.969,848] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:08.276,550] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:08.276,672] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:08.276,794] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:08.351,409] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:08.351,501] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:08.351,654] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.415,588] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.415,740] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.415,863] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.718,414] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.718,536] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.793,792] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.793,914] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.794,036] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.868,743] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.868,865] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.868,988] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
```
On further investigation I found in repeated failures the data_status on the packet before the dropped packet had data status set to 2: data incomplete no more to come as expected, it just isn't handled by the host.
**Environment:**
- OS: Linux (Ubuntu 20.04)
- Target: nrf5340dk
- Toolchain: NCS
- SHA: [af6add53](https://github.com/zephyrproject-rtos/zephyr/tree/af6add53aa1f357e5216a6207af0572b53830783)
| 1.0 | Bluetooth: Host: Periodic scanner does not differentiate between partial and incomplete data - **Describe the bug**
When the periodic scanner is reconstructing the periodic advertising data to report to the application, it only checks if the data_status is marked as complete or not, it does not differentiate between incomplete and partial data, see [here](https://github.com/zephyrproject-rtos/zephyr/blob/main/subsys/bluetooth/host/scan.c#L766-L835). In the case where the controller does not receive the data complete packet and the data_status is marked as "incomplete, no more data to come", the data of the next incoming periodic advertising event is appended to the truncated data resulting in incorrect data being passed to the application.
**To Reproduce**
This is quite a difficult bug to reproduce as it relies on a packet being dropped. The test environment where this was discovered had a very high RSSI just above the threshold of the radios sensitivity which increased the likely of the packet not being received.
Steps to reproduce the behavior:
1. Establish a synchronization between a periodic advertiser and scanner. The periodic advertiser should be advertising known data that is long enough to be split across multiple packets.
2. Repeatedly check the data received by the scanner and compare it with the expected advertised data.
3. When one of the packets to construct the data is not received the data should differ between the received data and the expected data
**Expected behavior**
The incomplete data should not be passed up to the application and should be dropped.
**Impact**
The application does not receive the correct data from the host.
**Logs and console output**
Expected data:
```
\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xf60\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,\x03'
```
Received data:
```
\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xfa\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\\]^_`abcdefghijklmnopqrstuvwxyz{\|}~\x7f\x80\x81\x82\x83\x84\x85\x86\x87\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f\x90\x91\x92\x93\x94\x95\x96\x97\x98\x99\x9a\x9b\x9c\x9d\x9e\x9f\xa0\xa1\xa2\xa3\xa4\xa5\xa6\xa7\xa8\xa9\xaa\xab\xac\xad\xae\xaf\xb0\xb1\xb2\xb3\xb4\xb5\xb6\xb7\xb8\xb9\xba\xbb\xbc\xbd\xbe\xbf\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf\xd0\xd1\xd2\xd3\xd4\xd5\xd6\xd7\xd8\xd9\xda\xdb\xdc\xdd\xde\xdf\xe0\xe1\xe2\xe3\xe4\xe5\xe6\xe7\xe8\xe9\xea\xeb\xec\xed\xee\xef\xf0\xf1\xf2\xf3\xf4\xf5\xf60\xffY\x00\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f !"#$%&\'()*+,\x03'
```
In the following logs, the data total data length of the advertised data was 300bytes, which was split into 3 packets: 247, 3, 50. After logging the data length of the incoming periodic advertisement packets before they're reconstructed, it can be seen in the following logs that right before the test failed, a packet of data length 50 (the final data in the sequence) was dropped. This resulted in the data being 250 bytes longer than expected as it appended 247, 3, 247, 3 and 50 bytes to the data buffer for the advertisement report. (I've indented the events of interest)
```
\*\*\* Booting Zephyr OS build v3.2.99-ncs1-1495-gaf6add53aa1f \*\*\*
[00:00:01.836,822] <inf> bt_hci_core: hci_vs_init: HW Platform: Nordic Semiconductor (0x0002)
[00:00:01.836,883] <inf> bt_hci_core: hci_vs_init: HW Variant: nRF53x (0x0003)
[00:00:01.836,914] <inf> bt_hci_core: hci_vs_init: Firmware: Standard Bluetooth controller (0x00) Version 184.62253 Build 1392531289
[00:00:01.839,080] <inf> bt_hci_core: bt_dev_show_info: Identity: D6:30:BF:22:A4:15 (random)
[00:00:01.839,111] <inf> bt_hci_core: bt_dev_show_info: HCI: version 5.3 (0x0c) revision 0x221c, manufacturer 0x0059
[00:00:01.839,141] <inf> bt_hci_core: bt_dev_show_info: LMP: version 5.3 (0x0c) subver 0x221c
[00:00:07.969,604] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:07.969,726] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:07.969,848] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:08.276,550] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:08.276,672] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:08.276,794] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:08.351,409] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:08.351,501] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:08.351,654] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.415,588] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.415,740] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.415,863] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.718,414] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.718,536] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.793,792] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.793,914] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.794,036] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
[00:00:09.868,743] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 247
[00:00:09.868,865] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 3
[00:00:09.868,988] <err> bt_scan: bt_hci_le_per_adv_report: evt len: 50
```
On further investigation I found in repeated failures the data_status on the packet before the dropped packet had data status set to 2: data incomplete no more to come as expected, it just isn't handled by the host.
**Environment:**
- OS: Linux (Ubuntu 20.04)
- Target: nrf5340dk
- Toolchain: NCS
- SHA: [af6add53](https://github.com/zephyrproject-rtos/zephyr/tree/af6add53aa1f357e5216a6207af0572b53830783)
| priority | bluetooth host periodic scanner does not differentiate between partial and incomplete data describe the bug when the periodic scanner is reconstructing the periodic advertising data to report to the application it only checks if the data status is marked as complete or not it does not differentiate between incomplete and partial data see in the case where the controller does not receive the data complete packet and the data status is marked as incomplete no more data to come the data of the next incoming periodic advertising event is appended to the truncated data resulting in incorrect data being passed to the application to reproduce this is quite a difficult bug to reproduce as it relies on a packet being dropped the test environment where this was discovered had a very high rssi just above the threshold of the radios sensitivity which increased the likely of the packet not being received steps to reproduce the behavior establish a synchronization between a periodic advertiser and scanner the periodic advertiser should be advertising known data that is long enough to be split across multiple packets repeatedly check the data received by the scanner and compare it with the expected advertised data when one of the packets to construct the data is not received the data should differ between the received data and the expected data expected behavior the incomplete data should not be passed up to the application and should be dropped impact the application does not receive the correct data from the host logs and console output expected data xfa xffy t n r abcdefghijklmnopqrstuvwxyz abcdefghijklmnopqrstuvwxyz xaa xab xac xad xae xaf xba xbb xbc xbd xbe xbf xca xcb xcc xcd xce xcf xda xdb xdc xdd xde xdf xea xeb xec xed xee xef xffy t n r received data xfa xffy t n r abcdefghijklmnopqrstuvwxyz abcdefghijklmnopqrstuvwxyz xaa xab xac xad xae xaf xba xbb xbc xbd xbe xbf xca xcb xcc xcd xce xcf xda xdb xdc xdd xde xdf xea xeb xec xed xee xef xfa xffy t n r abcdefghijklmnopqrstuvwxyz abcdefghijklmnopqrstuvwxyz xaa xab xac xad xae xaf xba xbb xbc xbd xbe xbf xca xcb xcc xcd xce xcf xda xdb xdc xdd xde xdf xea xeb xec xed xee xef xffy t n r in the following logs the data total data length of the advertised data was which was split into packets after logging the data length of the incoming periodic advertisement packets before they re reconstructed it can be seen in the following logs that right before the test failed a packet of data length the final data in the sequence was dropped this resulted in the data being bytes longer than expected as it appended and bytes to the data buffer for the advertisement report i ve indented the events of interest booting zephyr os build bt hci core hci vs init hw platform nordic semiconductor bt hci core hci vs init hw variant bt hci core hci vs init firmware standard bluetooth controller version build bt hci core bt dev show info identity bf random bt hci core bt dev show info hci version revision manufacturer bt hci core bt dev show info lmp version subver bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len bt scan bt hci le per adv report evt len on further investigation i found in repeated failures the data status on the packet before the dropped packet had data status set to data incomplete no more to come as expected it just isn t handled by the host environment os linux ubuntu target toolchain ncs sha | 1 |
11,662 | 2,610,156,299 | IssuesEvent | 2015-02-26 18:49:40 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Med Droid | auto-migrated Priority-Medium Type-Review | ```
Put med droid at tech 1
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:31 | 1.0 | Med Droid - ```
Put med droid at tech 1
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:31 | priority | med droid put med droid at tech original issue reported on code google com by gmail com on jan at | 1 |
77,081 | 3,506,259,200 | IssuesEvent | 2016-01-08 05:02:42 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Consume magic blood elf priest racial (BB #143) | Category: Miscellaneous migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** niyo
**Original Date:** 03.05.2010 21:47:01 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/143
<hr>
* The blood elf priest racial does not remove any magical buff at the moment and you can cast it even if you have no buff. It is really overpowered for PvE priests because you get like 700-800 mana every 2 mins for doing nothing.
* It should remove one magic buff on the caster and if there is no buff you can't cast it
| 1.0 | Consume magic blood elf priest racial (BB #143) - This issue was migrated from bitbucket.
**Original Reporter:** niyo
**Original Date:** 03.05.2010 21:47:01 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/143
<hr>
* The blood elf priest racial does not remove any magical buff at the moment and you can cast it even if you have no buff. It is really overpowered for PvE priests because you get like 700-800 mana every 2 mins for doing nothing.
* It should remove one magic buff on the caster and if there is no buff you can't cast it
| priority | consume magic blood elf priest racial bb this issue was migrated from bitbucket original reporter niyo original date gmt original priority major original type bug original state resolved direct link the blood elf priest racial does not remove any magical buff at the moment and you can cast it even if you have no buff it is really overpowered for pve priests because you get like mana every mins for doing nothing it should remove one magic buff on the caster and if there is no buff you can t cast it | 1 |
133,128 | 5,197,617,244 | IssuesEvent | 2017-01-23 16:04:09 | Cadasta/cadasta-platform | https://api.github.com/repos/Cadasta/cadasta-platform | closed | Permissions issue: Data Collector Role | bug priority: medium ui/ux | ### Steps to reproduce the error
Based off this user's permissions (which I setup locally):

The user should have data collector rights to the Pekapuran Laut Test Project. Generally this user should have the ability to view/add/edit all locations, relationships, parties, resources, as well as import and download project data.
### Actual behavior
I'm getting an alert that I don't have permission to import data. This new import data dropdown option is only available with PR #942. There isn't an option to download data. I think we need to revisit this from a UI perspective too. Maybe they should have a modified view of the ellipsis menu instead of removing it entirely.


### Expected behavior
I should be able to import data if I'm a data collector. | 1.0 | Permissions issue: Data Collector Role - ### Steps to reproduce the error
Based off this user's permissions (which I setup locally):

The user should have data collector rights to the Pekapuran Laut Test Project. Generally this user should have the ability to view/add/edit all locations, relationships, parties, resources, as well as import and download project data.
### Actual behavior
I'm getting an alert that I don't have permission to import data. This new import data dropdown option is only available with PR #942. There isn't an option to download data. I think we need to revisit this from a UI perspective too. Maybe they should have a modified view of the ellipsis menu instead of removing it entirely.


### Expected behavior
I should be able to import data if I'm a data collector. | priority | permissions issue data collector role steps to reproduce the error based off this user s permissions which i setup locally the user should have data collector rights to the pekapuran laut test project generally this user should have the ability to view add edit all locations relationships parties resources as well as import and download project data actual behavior i m getting an alert that i don t have permission to import data this new import data dropdown option is only available with pr there isn t an option to download data i think we need to revisit this from a ui perspective too maybe they should have a modified view of the ellipsis menu instead of removing it entirely expected behavior i should be able to import data if i m a data collector | 1 |
540,792 | 15,817,196,942 | IssuesEvent | 2021-04-05 14:15:54 | AY2021S2-CS2103-T16-1/tp | https://api.github.com/repos/AY2021S2-CS2103-T16-1/tp | closed | [PE-D] Keyword provided for `sort` could be case insensitive | priority.medium response.Accepted type.FeatureFlaw | (Suggestion!) `Ascending` or `descending` keywords could possibly be case insensitive for the convenience and speed for users using the function, otherwise it works well!
<!--session: 1617429967547-821d5e25-def9-4f4f-b7c7-a341c7e0f8e7-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: tsh22/ped#3 | 1.0 | [PE-D] Keyword provided for `sort` could be case insensitive - (Suggestion!) `Ascending` or `descending` keywords could possibly be case insensitive for the convenience and speed for users using the function, otherwise it works well!
<!--session: 1617429967547-821d5e25-def9-4f4f-b7c7-a341c7e0f8e7-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: tsh22/ped#3 | priority | keyword provided for sort could be case insensitive suggestion ascending or descending keywords could possibly be case insensitive for the convenience and speed for users using the function otherwise it works well labels severity low type featureflaw original ped | 1 |
26,501 | 2,684,633,784 | IssuesEvent | 2015-03-29 06:08:12 | gtcasl/gpuocelot | https://api.github.com/repos/gtcasl/gpuocelot | opened | build missing LIB | bug imported Priority-Medium | _From [ebo.2...@gmail.com](https://code.google.com/u/109078985440113502765/) on August 19, 2014 01:02:35_
What steps will reproduce the problem? 1. svn co ...
2. ./build.py -w What is the expected output? What do you see instead? Expected it to build, but got the following error:
cons: building associated VariantDir targets: .release_build/ocelot
config_h_build([".release_build/ocelot/configure.h"], ["ocelot/configure.h.in"])
g++ -o .release_build/ocelot/LoadPtx -L/usr/lib64 -lz -lpthread -lffi -ldl -lm .release_build/ocelot/ocelot/tools/LoadPtx.o -L/usr/lib -L/usr/lib64 -locelot -lboost_system-mt -lboost_filesystem-mt -lboost_thread-mt -lGLEW -ldl -lrt -lpthread -lLLVMAsmParser -lLLVMX86Disassembler -lLLVMX86AsmParser -lLLVMX86CodeGen -lLLVMSelectionDAG -lLLVMAsmPrinter -lLLVMMCParser -lLLVMX86Desc -lLLVMX86Info -lLLVMX86AsmPrinter -lLLVMX86Utils -lLLVMJIT -lLLVMRuntimeDyld -lLLVMExecutionEngine -lLLVMCodeGen -lLLVMObjCARCOpts -lLLVMScalarOpts -lLLVMInstCombine -lLLVMTransformUtils -lLLVMipa -lLLVMAnalysis -lLLVMTarget -lLLVMMC -lLLVMObject -lLLVMCore -lLLVMSupport
/usr/lib/gcc/x86_64-pc-linux-gnu/4.8.3/../../../../x86_64-pc-linux-gnu/bin/ld: cannot find -locelot
collect2: error: ld returned 1 exit status
scons: *** [.release_build/ocelot/LoadPtx] Error 1 What version of the product are you using? On what operating system? Repository Root: http://gpuocelot.googlecode.com/svn Repository UUID: ec522b64-56cd-11de-8d7d-01d876f1962d
Revision: 2256
OS: Gentoo Linux-3.12.21-gentoo- r1 g++ (Gentoo 4.8.3 p1.1, pie-0.5.9) 4.8.3 Please provide any additional information below. I can get ocelot to build by adding -L.release_build to the beginning of the last G++ command. I have seen this before and was able to fix it by defining "LIBPATH=['.release_build']" in the environment. I was not able to figure out how to do this to the provided ./ocelot/SConscript
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=96_ | 1.0 | build missing LIB - _From [ebo.2...@gmail.com](https://code.google.com/u/109078985440113502765/) on August 19, 2014 01:02:35_
What steps will reproduce the problem? 1. svn co ...
2. ./build.py -w What is the expected output? What do you see instead? Expected it to build, but got the following error:
cons: building associated VariantDir targets: .release_build/ocelot
config_h_build([".release_build/ocelot/configure.h"], ["ocelot/configure.h.in"])
g++ -o .release_build/ocelot/LoadPtx -L/usr/lib64 -lz -lpthread -lffi -ldl -lm .release_build/ocelot/ocelot/tools/LoadPtx.o -L/usr/lib -L/usr/lib64 -locelot -lboost_system-mt -lboost_filesystem-mt -lboost_thread-mt -lGLEW -ldl -lrt -lpthread -lLLVMAsmParser -lLLVMX86Disassembler -lLLVMX86AsmParser -lLLVMX86CodeGen -lLLVMSelectionDAG -lLLVMAsmPrinter -lLLVMMCParser -lLLVMX86Desc -lLLVMX86Info -lLLVMX86AsmPrinter -lLLVMX86Utils -lLLVMJIT -lLLVMRuntimeDyld -lLLVMExecutionEngine -lLLVMCodeGen -lLLVMObjCARCOpts -lLLVMScalarOpts -lLLVMInstCombine -lLLVMTransformUtils -lLLVMipa -lLLVMAnalysis -lLLVMTarget -lLLVMMC -lLLVMObject -lLLVMCore -lLLVMSupport
/usr/lib/gcc/x86_64-pc-linux-gnu/4.8.3/../../../../x86_64-pc-linux-gnu/bin/ld: cannot find -locelot
collect2: error: ld returned 1 exit status
scons: *** [.release_build/ocelot/LoadPtx] Error 1 What version of the product are you using? On what operating system? Repository Root: http://gpuocelot.googlecode.com/svn Repository UUID: ec522b64-56cd-11de-8d7d-01d876f1962d
Revision: 2256
OS: Gentoo Linux-3.12.21-gentoo- r1 g++ (Gentoo 4.8.3 p1.1, pie-0.5.9) 4.8.3 Please provide any additional information below. I can get ocelot to build by adding -L.release_build to the beginning of the last G++ command. I have seen this before and was able to fix it by defining "LIBPATH=['.release_build']" in the environment. I was not able to figure out how to do this to the provided ./ocelot/SConscript
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=96_ | priority | build missing lib from on august what steps will reproduce the problem svn co build py w what is the expected output what do you see instead expected it to build but got the following error cons building associated variantdir targets release build ocelot config h build g o release build ocelot loadptx l usr lz lpthread lffi ldl lm release build ocelot ocelot tools loadptx o l usr lib l usr locelot lboost system mt lboost filesystem mt lboost thread mt lglew ldl lrt lpthread lllvmasmparser lllvmselectiondag lllvmasmprinter lllvmmcparser lllvmjit lllvmruntimedyld lllvmexecutionengine lllvmcodegen lllvmobjcarcopts lllvmscalaropts lllvminstcombine lllvmtransformutils lllvmipa lllvmanalysis lllvmtarget lllvmmc lllvmobject lllvmcore lllvmsupport usr lib gcc pc linux gnu pc linux gnu bin ld cannot find locelot error ld returned exit status scons error what version of the product are you using on what operating system repository root repository uuid revision os gentoo linux gentoo g gentoo pie please provide any additional information below i can get ocelot to build by adding l release build to the beginning of the last g command i have seen this before and was able to fix it by defining libpath in the environment i was not able to figure out how to do this to the provided ocelot sconscript original issue | 1 |
356,273 | 10,591,014,591 | IssuesEvent | 2019-10-09 09:56:42 | fac-17/My-Body-Back | https://api.github.com/repos/fac-17/My-Body-Back | opened | Research ChatBot / Socket | Feature Medium Priority One Hour Research Required | - [ ] Spend some time researching implementation of ChatBot
- [ ] Post research notes to Wiki/HackMD. | 1.0 | Research ChatBot / Socket - - [ ] Spend some time researching implementation of ChatBot
- [ ] Post research notes to Wiki/HackMD. | priority | research chatbot socket spend some time researching implementation of chatbot post research notes to wiki hackmd | 1 |
194,088 | 6,891,474,763 | IssuesEvent | 2017-11-22 17:09:56 | JeffersonLab/hcana | https://api.github.com/repos/JeffersonLab/hcana | closed | Add config event handler for "TI setup data" | Medium priority | There is a (semi-)new TI config related blob in Event 125 that stores some trigger/TM related settings that needs a decoder in THcConfigEvtHandler.cxx. Runs after SHMS 1295 and HMS 718 will have blobs with the information below.
The word format is defined by this code in the hcvme0[12].c ROLs:
```
/* Log prescales, misc TI related information from ffile */
*rol->dabufp++ = (0xd0000000); // 'misc' header block
*rol->dabufp++ = (0xabcd0000+1); // last +N is version ID number
*rol->dabufp++ = nped; //
*rol->dabufp++ = scaler_period; //
*rol->dabufp++ = syncCnt; //
for(i=0; i<NUM_PRESCALES; i++) { // Input Trigger prescale settings
*rol->dabufp++ = prescales[i];
}
*rol->dabufp++ = (0xd000000f); // 'misc' trailer block
``` | 1.0 | Add config event handler for "TI setup data" - There is a (semi-)new TI config related blob in Event 125 that stores some trigger/TM related settings that needs a decoder in THcConfigEvtHandler.cxx. Runs after SHMS 1295 and HMS 718 will have blobs with the information below.
The word format is defined by this code in the hcvme0[12].c ROLs:
```
/* Log prescales, misc TI related information from ffile */
*rol->dabufp++ = (0xd0000000); // 'misc' header block
*rol->dabufp++ = (0xabcd0000+1); // last +N is version ID number
*rol->dabufp++ = nped; //
*rol->dabufp++ = scaler_period; //
*rol->dabufp++ = syncCnt; //
for(i=0; i<NUM_PRESCALES; i++) { // Input Trigger prescale settings
*rol->dabufp++ = prescales[i];
}
*rol->dabufp++ = (0xd000000f); // 'misc' trailer block
``` | priority | add config event handler for ti setup data there is a semi new ti config related blob in event that stores some trigger tm related settings that needs a decoder in thcconfigevthandler cxx runs after shms and hms will have blobs with the information below the word format is defined by this code in the c rols log prescales misc ti related information from ffile rol dabufp misc header block rol dabufp last n is version id number rol dabufp nped rol dabufp scaler period rol dabufp synccnt for i i num prescales i input trigger prescale settings rol dabufp prescales rol dabufp misc trailer block | 1 |
118,940 | 4,758,101,313 | IssuesEvent | 2016-10-24 18:31:06 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | Project locations - Google maps | Priority - Medium Type - Bug | Al hacer clic en esos botones, se pierde el foco del mapa, y entonces los puntos ya no aparecen.
Lo probé logueandome como AJARVIS y entando al proyecto 18 en modo read.

| 1.0 | Project locations - Google maps - Al hacer clic en esos botones, se pierde el foco del mapa, y entonces los puntos ya no aparecen.
Lo probé logueandome como AJARVIS y entando al proyecto 18 en modo read.

| priority | project locations google maps al hacer clic en esos botones se pierde el foco del mapa y entonces los puntos ya no aparecen lo probé logueandome como ajarvis y entando al proyecto en modo read | 1 |
285,533 | 8,766,219,672 | IssuesEvent | 2018-12-17 16:17:32 | carbon-design-system/carbon-website | https://api.github.com/repos/carbon-design-system/carbon-website | closed | [Website Components] Shell behavior for narrow screens | priority: medium project: website type: ux :popcorn: type: visual :art: v10 :new: | The website shell is using a fixed panel side nav that does not expand or collapse until screen width goes below a certain threshold.
https://w3.ibm.com/design/product/experimental/layout/#grid-behaviors
This task is to define that behavior and produce redlines for the mockups.


 | 1.0 | [Website Components] Shell behavior for narrow screens - The website shell is using a fixed panel side nav that does not expand or collapse until screen width goes below a certain threshold.
https://w3.ibm.com/design/product/experimental/layout/#grid-behaviors
This task is to define that behavior and produce redlines for the mockups.


 | priority | shell behavior for narrow screens the website shell is using a fixed panel side nav that does not expand or collapse until screen width goes below a certain threshold this task is to define that behavior and produce redlines for the mockups | 1 |
76,693 | 3,491,074,112 | IssuesEvent | 2016-01-04 14:01:10 | ngageoint/hootenanny-ui | https://api.github.com/repos/ngageoint/hootenanny-ui | opened | Clipped Dataset goes to incorrect folder | Category: UI Priority: Medium Status: New/Undefined Type: Bug | After clipping a dataset, it does not go to the selected folder. | 1.0 | Clipped Dataset goes to incorrect folder - After clipping a dataset, it does not go to the selected folder. | priority | clipped dataset goes to incorrect folder after clipping a dataset it does not go to the selected folder | 1 |
111,630 | 4,479,531,675 | IssuesEvent | 2016-08-27 17:25:16 | hook/champions | https://api.github.com/repos/hook/champions | closed | Importing / Exporting | Priority: Medium Status: Completed Type: Question | I absolutely love this Champions tool that has been put together. And, I may be looking at this from way out in left field and missing this one, but each time I try and import or export .csv files, it appears to do nothing for me. It definitely does not bring up any dialog box to choose a .csv file. Are these commands being updated currently? | 1.0 | Importing / Exporting - I absolutely love this Champions tool that has been put together. And, I may be looking at this from way out in left field and missing this one, but each time I try and import or export .csv files, it appears to do nothing for me. It definitely does not bring up any dialog box to choose a .csv file. Are these commands being updated currently? | priority | importing exporting i absolutely love this champions tool that has been put together and i may be looking at this from way out in left field and missing this one but each time i try and import or export csv files it appears to do nothing for me it definitely does not bring up any dialog box to choose a csv file are these commands being updated currently | 1 |
660,779 | 22,012,940,162 | IssuesEvent | 2022-05-27 00:12:20 | GameFreedomGG/Sindragosa-SII | https://api.github.com/repos/GameFreedomGG/Sindragosa-SII | closed | In Need of Felblood. TBC quest not 100% drop | Status: Closed- Resolved Status: Confirmed Bug (Awaiting Development) Priority: Medium Custom mechanic Quest | Decription:
In Need of Felblood. https://wotlkdb.com/?quest=9366
How it works:
Quest item doesnt drop with 100% chance
How it should work:
It should drops 100% chance according to Discord announce for Vanilla and TBC zones
Source (you should point out proofs of your report, please give us some source):
https://discord.com/channels/227036070850723840/967887165163982880/975843740876947486 | 1.0 | In Need of Felblood. TBC quest not 100% drop - Decription:
In Need of Felblood. https://wotlkdb.com/?quest=9366
How it works:
Quest item doesnt drop with 100% chance
How it should work:
It should drops 100% chance according to Discord announce for Vanilla and TBC zones
Source (you should point out proofs of your report, please give us some source):
https://discord.com/channels/227036070850723840/967887165163982880/975843740876947486 | priority | in need of felblood tbc quest not drop decription in need of felblood how it works quest item doesnt drop with chance how it should work it should drops chance according to discord announce for vanilla and tbc zones source you should point out proofs of your report please give us some source | 1 |
755,897 | 26,446,134,695 | IssuesEvent | 2023-01-16 07:28:13 | zhamao-robot/zhamao-framework | https://api.github.com/repos/zhamao-robot/zhamao-framework | closed | 文档补全计划 | enhancement priority/medium | ### 描述
- [x] #65 #110
- [x] 中间件传参
- [ ] TaskWorker 教程
- [ ] 远程终端自定义命令
- [x] #76
- [ ] #96
- [x] #97 #103
- [x] #99 #67
- [x] #111
- [ ] `--audit-mode` 相关文档 #114
- [x] #130 | 1.0 | 文档补全计划 - ### 描述
- [x] #65 #110
- [x] 中间件传参
- [ ] TaskWorker 教程
- [ ] 远程终端自定义命令
- [x] #76
- [ ] #96
- [x] #97 #103
- [x] #99 #67
- [x] #111
- [ ] `--audit-mode` 相关文档 #114
- [x] #130 | priority | 文档补全计划 描述 中间件传参 taskworker 教程 远程终端自定义命令 audit mode 相关文档 | 1 |
131,297 | 5,145,875,675 | IssuesEvent | 2017-01-12 22:54:46 | swri-robotics/bag-database | https://api.github.com/repos/swri-robotics/bag-database | closed | "Copy Link" from Folder View puts bag path instead of id in the URL | Priority: Medium Status: Completed Type: Bug | The endpoint used to download bag files expects to be passed a database ID. The "Copy Link" context menu item in the List View puts the id in there as expected, but the Folder View puts the full path to the bag in there, which won't work at all. | 1.0 | "Copy Link" from Folder View puts bag path instead of id in the URL - The endpoint used to download bag files expects to be passed a database ID. The "Copy Link" context menu item in the List View puts the id in there as expected, but the Folder View puts the full path to the bag in there, which won't work at all. | priority | copy link from folder view puts bag path instead of id in the url the endpoint used to download bag files expects to be passed a database id the copy link context menu item in the list view puts the id in there as expected but the folder view puts the full path to the bag in there which won t work at all | 1 |
491,474 | 14,164,909,500 | IssuesEvent | 2020-11-12 06:10:56 | pypa/pipenv | https://api.github.com/repos/pypa/pipenv | closed | editable dependency causes wrong versions to be installed | Category: Dependency Resolution Category: VCS Priority: Medium Type: Possible Bug | ### Issue description
An error is not reported at the end of a `pipenv sync` command if an installed dependency violates a constraint / does not match the version specified in the lockfile.
An error is not reported at the end of creating a lockfile via `pipenv install` if an installed dependency violates a constraint specified in the Pipenv file / newly created lockfile.
Looks like an editable dependency is key to causing this issue.
May be related to https://github.com/pypa/pipenv/issues/3296
Note that there appears to be a datarace involved as SOMETIMES when I run `pipenv install` the versions that are installed match the generated/established lockfile and SOMETIMES they do not!
The test case I describe here will be able to repeat the issue consistently. It appears to happen whenever there is a dependency of an editable dependency that is loosely locked down ( > or >= or < or <= or * ) this rarely happens if at all when such a dependency is locked down exactly ( == )
Output from test:
```text
STARTING TESTS
testing 'pipenv sync' operation; iteration = 1
Creating a virtualenv for this project…
Pipfile: /workspace/Pipfile
Using /usr/local/bin/python (3.7.5) to create virtualenv…
⠋ Creating virtual environment...Already using interpreter /usr/local/bin/python
Using base prefix '/usr/local'
New python executable in /workspace/.venv/bin/python
Installing setuptools, pip, wheel...
done.
✔ Successfully created virtual environment!
Virtualenv location: /workspace/.venv
Installing dependencies from Pipfile.lock (bfdd12)…
🐍 ▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉ 18/18 — 00:01:08
To activate this project's virtualenv, run pipenv shell.
Alternatively, run a command inside the virtualenv with pipenv run.
All dependencies are now up-to-date!
verifying...FAIL
the following python packages did not install correctly:
amqp: expected 1.4.9 got 2.5.2 instead
celery: expected 3.1.19 got 4.4.0 instead
billiard: expected 3.3.0.23 got 3.6.1.0 instead
```
### Expected result
I expect installed dependency versions to not violate constraints specified in the Pipenv file and I expect dependencies that are installed when running `pipenv sync` or `pipenv install` to match exactly the version specified in the lockfile - whether that lockfile be pre-existing or newly created.
If this does not happen, I expect an error to be reported.
( In addition, if the lockfile exists and I run `pipenv sync` I still expect the lockfile and Pipenv file to remain unchanged. Though there appears to be no issue around that combination of conditions I feel it should be enumerated. )
### Actual result
Versions of dependencies that violate the top-most Pipenv file are installed when running `pipenv install`. Also they do not match the version stated in the newly created lockfile.
Versions of dependencies that do not match the lockfile are installed when running `pipenv sync`.
No error is thrown.
### Steps to replicate
Made a couple of minor tests that express the problem:
On a linux friendly system with make:
1. clone https://github.com/bottom-text-everywhere/test-python-project
2. cd test-python-project
3. install docker
5. `make test`
View the README.md file in the project root for more details.
There are two tests: calling `pipenv sync` and calling `pipenv install` which creates a lockfile.
-------------------------------------------------------------------------------
<details><summary>$ pipenv --support</summary>
Pipenv version: `'2018.11.26'`
Pipenv location: `'/usr/local/lib/python3.7/site-packages/pipenv'`
Python location: `'/usr/local/bin/python'`
Python installations found:
- `3.7.5`: `/usr/local/bin/python`
- `3.7.5`: `/usr/local/bin/python3.7m`
- `3.7.3`: `/usr/bin/python3.7m`
- `3.7.3`: `/usr/bin/python3.7`
- `2.7.16`: `/usr/bin/python2`
PEP 508 Information:
```
{'implementation_name': 'cpython',
'implementation_version': '3.7.5',
'os_name': 'posix',
'platform_machine': 'x86_64',
'platform_python_implementation': 'CPython',
'platform_release': '4.9.184-linuxkit',
'platform_system': 'Linux',
'platform_version': '#1 SMP Tue Jul 2 22:58:16 UTC 2019',
'python_full_version': '3.7.5',
'python_version': '3.7',
'sys_platform': 'linux'}
```
System environment variables:
- `PIPENV_VENV_IN_PROJECT`
- `HOSTNAME`
- `PYTHON_VERSION`
- `DOCKER_SHELL_ACTIVE`
- `PWD`
- `HOME`
- `LANG`
- `GPG_KEY`
- `TERM`
- `SHLVL`
- `PYTHON_PIP_VERSION`
- `PYTHON_GET_PIP_SHA256`
- `PYTHON_GET_PIP_URL`
- `PATH`
- `_`
- `PIP_DISABLE_PIP_VERSION_CHECK`
- `PYTHONDONTWRITEBYTECODE`
- `PIP_SHIMS_BASE_MODULE`
- `PIP_PYTHON_PATH`
- `PYTHONFINDER_IGNORE_UNSUPPORTED`
Pipenv–specific environment variables:
- `PIPENV_VENV_IN_PROJECT`: `1`
Debug–specific environment variables:
- `PATH`: `/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin`
- `LANG`: `C.UTF-8`
- `PWD`: `/workspace`
---------------------------
Contents of `Pipfile` ('/workspace/Pipfile'):
```toml
[[source]]
name = "pypi"
url = "https://pypi.org/simple"
verify_ssl = true
[dev-packages]
[packages]
requests = "==2.22.0"
anyjson = "==0.3.3"
Babel = "==2.2.0"
certifi = "==2019.6.16"
chardet = "==3.0.4"
idna = "==2.8"
pytz = "==2015.7"
tornado = "==4.2"
urllib3 = "==1.25.3"
vine = "==1.3.0"
flower = {editable = true,git = "https://github.com/mher/flower",ref = "f20f43c858df3c0a94e5e10015fcd7d52089e3f0"}
celery = "==3.1.19"
kombu = "==3.0.37"
amqp = "==1.4.9"
billiard = "==3.3.0.23"
importlib-metadata = "==1.1.0"
zipp = "==0.6.0"
more-itertools = "==8.0.2"
[requires]
python_version = "3.7"
```
Contents of `Pipfile.lock` ('/workspace/Pipfile.lock'):
```json
{
"_meta": {
"hash": {
"sha256": "5a05b2be1247f8f5df396dbe9a557d43c34516cb9b9c0ae3a3c2eb4b9fbfdd12"
},
"pipfile-spec": 6,
"requires": {
"python_version": "3.7"
},
"sources": [
{
"name": "pypi",
"url": "https://pypi.org/simple",
"verify_ssl": true
}
]
},
"default": {
"amqp": {
"hashes": [
"sha256:2dea4d16d073c902c3b89d9b96620fb6729ac0f7a923bbc777cb4ad827c0c61a",
"sha256:e0ed0ce6b8ffe5690a2e856c7908dc557e0e605283d6885dd1361d79f2928908"
],
"index": "pypi",
"version": "==1.4.9"
},
"anyjson": {
"hashes": [
"sha256:37812d863c9ad3e35c0734c42e0bf0320ce8c3bed82cd20ad54cb34d158157ba"
],
"index": "pypi",
"version": "==0.3.3"
},
"babel": {
"hashes": [
"sha256:d8cb4c0e78148aee89560f9fe21587aa57739c975bb89ff66b1e842cc697428f",
"sha256:fed07cbcdcb3de79b53a8220eebed21c93f8dbb3dbce1d9c6b1c4b09e8aecf2b"
],
"index": "pypi",
"version": "==2.2.0"
},
"billiard": {
"hashes": [
"sha256:204e75d390ef8f839c30a93b696bd842c3941916e15921745d05edc2a83868ab",
"sha256:23cb71472712e96bff3e0d45763b7b8a99e5040385fffb96816028352c255682",
"sha256:692a2a5a55ee39a42bcb7557930e2541da85df9ea81c6e24827f63b80cd39d0b",
"sha256:82041dbaa62f7fde1464d7ab449978618a38b241b40c0d31dafabb36446635dc",
"sha256:958fc9f8fd5cc9b936b2cb9d96f02aa5ec3613ba13ee7f089c77ff0bcc368fac",
"sha256:c0cbe8d45ba8d8213ad68ef9a1881002a151569c9424d551634195a18c3a4160",
"sha256:ccfe0419eb5e49f27ad35cf06e75360af903df6d576c66cb8073246d4e023e5c",
"sha256:d4d2fed1a251ea58eed47b48db3778ebb92f5ff4407dc91869c6f41c3a9249d0"
],
"index": "pypi",
"version": "==3.3.0.23"
},
"celery": {
"hashes": [
"sha256:434b4e62308428b0e83fbcd1cebb8853beb230237b1b23484063fcf1ff610b6e",
"sha256:bd6a21290c822ecf78947a7dbc8b2270d88a5fa08a7a2bfcd908bb3d5ed66b71"
],
"index": "pypi",
"version": "==3.1.19"
},
"certifi": {
"hashes": [
"sha256:046832c04d4e752f37383b628bc601a7ea7211496b4638f6514d0e5b9acc4939",
"sha256:945e3ba63a0b9f577b1395204e13c3a231f9bc0223888be653286534e5873695"
],
"index": "pypi",
"version": "==2019.6.16"
},
"chardet": {
"hashes": [
"sha256:84ab92ed1c4d4f16916e05906b6b75a6c0fb5db821cc65e70cbd64a3e2a5eaae",
"sha256:fc323ffcaeaed0e0a02bf4d117757b98aed530d9ed4531e3e15460124c106691"
],
"index": "pypi",
"version": "==3.0.4"
},
"flower": {
"editable": true,
"git": "https://github.com/mher/flower",
"ref": "f20f43c858df3c0a94e5e10015fcd7d52089e3f0"
},
"idna": {
"hashes": [
"sha256:c357b3f628cf53ae2c4c05627ecc484553142ca23264e593d327bcde5e9c3407",
"sha256:ea8b7f6188e6fa117537c3df7da9fc686d485087abf6ac197f9c46432f7e4a3c"
],
"index": "pypi",
"version": "==2.8"
},
"importlib-metadata": {
"hashes": [
"sha256:b044f07694ef14a6683b097ba56bd081dbc7cdc7c7fe46011e499dfecc082f21",
"sha256:e6ac600a142cf2db707b1998382cc7fc3b02befb7273876e01b8ad10b9652742"
],
"index": "pypi",
"version": "==1.1.0"
},
"kombu": {
"hashes": [
"sha256:7ceab743e3e974f3e5736082e8cc514c009e254e646d6167342e0e192aee81a6",
"sha256:e064a00c66b4d1058cd2b0523fb8d98c82c18450244177b6c0f7913016642650"
],
"index": "pypi",
"version": "==3.0.37"
},
"more-itertools": {
"hashes": [
"sha256:b84b238cce0d9adad5ed87e745778d20a3f8487d0f0cb8b8a586816c7496458d",
"sha256:c833ef592a0324bcc6a60e48440da07645063c453880c9477ceb22490aec1564"
],
"index": "pypi",
"version": "==8.0.2"
},
"pytz": {
"hashes": [
"sha256:3ede470d3d17ba3c07638dfa0d10452bc1b6e5ad326127a65ba77e6aaeb11bec",
"sha256:8b6ce1c993909783bc96e0b4f34ea223bff7a4df2c90bdb9c4e0f1ac928689e3",
"sha256:99266ef30a37e43932deec2b7ca73e83c8dbc3b9ff703ec73eca6b1dae6befea",
"sha256:fbd26746772c24cb93c8b97cbdad5cb9e46c86bbdb1b9d8a743ee00e2fb1fc5d"
],
"index": "pypi",
"version": "==2015.7"
},
"requests": {
"hashes": [
"sha256:11e007a8a2aa0323f5a921e9e6a2d7e4e67d9877e85773fba9ba6419025cbeb4",
"sha256:9cf5292fcd0f598c671cfc1e0d7d1a7f13bb8085e9a590f48c010551dc6c4b31"
],
"index": "pypi",
"version": "==2.22.0"
},
"tornado": {
"hashes": [
"sha256:e8b1207da67dbdceebfb291292b4ef1b547d6171525bec1b366853f923456a5f"
],
"index": "pypi",
"version": "==4.2"
},
"urllib3": {
"hashes": [
"sha256:b246607a25ac80bedac05c6f282e3cdaf3afb65420fd024ac94435cabe6e18d1",
"sha256:dbe59173209418ae49d485b87d1681aefa36252ee85884c31346debd19463232"
],
"index": "pypi",
"version": "==1.25.3"
},
"vine": {
"hashes": [
"sha256:133ee6d7a9016f177ddeaf191c1f58421a1dcc6ee9a42c58b34bed40e1d2cd87",
"sha256:ea4947cc56d1fd6f2095c8d543ee25dad966f78692528e68b4fada11ba3f98af"
],
"index": "pypi",
"version": "==1.3.0"
},
"zipp": {
"hashes": [
"sha256:3718b1cbcd963c7d4c5511a8240812904164b7f381b647143a89d3b98f9bcd8e",
"sha256:f06903e9f1f43b12d371004b4ac7b06ab39a44adc747266928ae6debfa7b3335"
],
"index": "pypi",
"version": "==0.6.0"
}
},
"develop": {}
}
```
</details> | 1.0 | editable dependency causes wrong versions to be installed - ### Issue description
An error is not reported at the end of a `pipenv sync` command if an installed dependency violates a constraint / does not match the version specified in the lockfile.
An error is not reported at the end of creating a lockfile via `pipenv install` if an installed dependency violates a constraint specified in the Pipenv file / newly created lockfile.
Looks like an editable dependency is key to causing this issue.
May be related to https://github.com/pypa/pipenv/issues/3296
Note that there appears to be a datarace involved as SOMETIMES when I run `pipenv install` the versions that are installed match the generated/established lockfile and SOMETIMES they do not!
The test case I describe here will be able to repeat the issue consistently. It appears to happen whenever there is a dependency of an editable dependency that is loosely locked down ( > or >= or < or <= or * ) this rarely happens if at all when such a dependency is locked down exactly ( == )
Output from test:
```text
STARTING TESTS
testing 'pipenv sync' operation; iteration = 1
Creating a virtualenv for this project…
Pipfile: /workspace/Pipfile
Using /usr/local/bin/python (3.7.5) to create virtualenv…
⠋ Creating virtual environment...Already using interpreter /usr/local/bin/python
Using base prefix '/usr/local'
New python executable in /workspace/.venv/bin/python
Installing setuptools, pip, wheel...
done.
✔ Successfully created virtual environment!
Virtualenv location: /workspace/.venv
Installing dependencies from Pipfile.lock (bfdd12)…
🐍 ▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉ 18/18 — 00:01:08
To activate this project's virtualenv, run pipenv shell.
Alternatively, run a command inside the virtualenv with pipenv run.
All dependencies are now up-to-date!
verifying...FAIL
the following python packages did not install correctly:
amqp: expected 1.4.9 got 2.5.2 instead
celery: expected 3.1.19 got 4.4.0 instead
billiard: expected 3.3.0.23 got 3.6.1.0 instead
```
### Expected result
I expect installed dependency versions to not violate constraints specified in the Pipenv file and I expect dependencies that are installed when running `pipenv sync` or `pipenv install` to match exactly the version specified in the lockfile - whether that lockfile be pre-existing or newly created.
If this does not happen, I expect an error to be reported.
( In addition, if the lockfile exists and I run `pipenv sync` I still expect the lockfile and Pipenv file to remain unchanged. Though there appears to be no issue around that combination of conditions I feel it should be enumerated. )
### Actual result
Versions of dependencies that violate the top-most Pipenv file are installed when running `pipenv install`. Also they do not match the version stated in the newly created lockfile.
Versions of dependencies that do not match the lockfile are installed when running `pipenv sync`.
No error is thrown.
### Steps to replicate
Made a couple of minor tests that express the problem:
On a linux friendly system with make:
1. clone https://github.com/bottom-text-everywhere/test-python-project
2. cd test-python-project
3. install docker
5. `make test`
View the README.md file in the project root for more details.
There are two tests: calling `pipenv sync` and calling `pipenv install` which creates a lockfile.
-------------------------------------------------------------------------------
<details><summary>$ pipenv --support</summary>
Pipenv version: `'2018.11.26'`
Pipenv location: `'/usr/local/lib/python3.7/site-packages/pipenv'`
Python location: `'/usr/local/bin/python'`
Python installations found:
- `3.7.5`: `/usr/local/bin/python`
- `3.7.5`: `/usr/local/bin/python3.7m`
- `3.7.3`: `/usr/bin/python3.7m`
- `3.7.3`: `/usr/bin/python3.7`
- `2.7.16`: `/usr/bin/python2`
PEP 508 Information:
```
{'implementation_name': 'cpython',
'implementation_version': '3.7.5',
'os_name': 'posix',
'platform_machine': 'x86_64',
'platform_python_implementation': 'CPython',
'platform_release': '4.9.184-linuxkit',
'platform_system': 'Linux',
'platform_version': '#1 SMP Tue Jul 2 22:58:16 UTC 2019',
'python_full_version': '3.7.5',
'python_version': '3.7',
'sys_platform': 'linux'}
```
System environment variables:
- `PIPENV_VENV_IN_PROJECT`
- `HOSTNAME`
- `PYTHON_VERSION`
- `DOCKER_SHELL_ACTIVE`
- `PWD`
- `HOME`
- `LANG`
- `GPG_KEY`
- `TERM`
- `SHLVL`
- `PYTHON_PIP_VERSION`
- `PYTHON_GET_PIP_SHA256`
- `PYTHON_GET_PIP_URL`
- `PATH`
- `_`
- `PIP_DISABLE_PIP_VERSION_CHECK`
- `PYTHONDONTWRITEBYTECODE`
- `PIP_SHIMS_BASE_MODULE`
- `PIP_PYTHON_PATH`
- `PYTHONFINDER_IGNORE_UNSUPPORTED`
Pipenv–specific environment variables:
- `PIPENV_VENV_IN_PROJECT`: `1`
Debug–specific environment variables:
- `PATH`: `/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin`
- `LANG`: `C.UTF-8`
- `PWD`: `/workspace`
---------------------------
Contents of `Pipfile` ('/workspace/Pipfile'):
```toml
[[source]]
name = "pypi"
url = "https://pypi.org/simple"
verify_ssl = true
[dev-packages]
[packages]
requests = "==2.22.0"
anyjson = "==0.3.3"
Babel = "==2.2.0"
certifi = "==2019.6.16"
chardet = "==3.0.4"
idna = "==2.8"
pytz = "==2015.7"
tornado = "==4.2"
urllib3 = "==1.25.3"
vine = "==1.3.0"
flower = {editable = true,git = "https://github.com/mher/flower",ref = "f20f43c858df3c0a94e5e10015fcd7d52089e3f0"}
celery = "==3.1.19"
kombu = "==3.0.37"
amqp = "==1.4.9"
billiard = "==3.3.0.23"
importlib-metadata = "==1.1.0"
zipp = "==0.6.0"
more-itertools = "==8.0.2"
[requires]
python_version = "3.7"
```
Contents of `Pipfile.lock` ('/workspace/Pipfile.lock'):
```json
{
"_meta": {
"hash": {
"sha256": "5a05b2be1247f8f5df396dbe9a557d43c34516cb9b9c0ae3a3c2eb4b9fbfdd12"
},
"pipfile-spec": 6,
"requires": {
"python_version": "3.7"
},
"sources": [
{
"name": "pypi",
"url": "https://pypi.org/simple",
"verify_ssl": true
}
]
},
"default": {
"amqp": {
"hashes": [
"sha256:2dea4d16d073c902c3b89d9b96620fb6729ac0f7a923bbc777cb4ad827c0c61a",
"sha256:e0ed0ce6b8ffe5690a2e856c7908dc557e0e605283d6885dd1361d79f2928908"
],
"index": "pypi",
"version": "==1.4.9"
},
"anyjson": {
"hashes": [
"sha256:37812d863c9ad3e35c0734c42e0bf0320ce8c3bed82cd20ad54cb34d158157ba"
],
"index": "pypi",
"version": "==0.3.3"
},
"babel": {
"hashes": [
"sha256:d8cb4c0e78148aee89560f9fe21587aa57739c975bb89ff66b1e842cc697428f",
"sha256:fed07cbcdcb3de79b53a8220eebed21c93f8dbb3dbce1d9c6b1c4b09e8aecf2b"
],
"index": "pypi",
"version": "==2.2.0"
},
"billiard": {
"hashes": [
"sha256:204e75d390ef8f839c30a93b696bd842c3941916e15921745d05edc2a83868ab",
"sha256:23cb71472712e96bff3e0d45763b7b8a99e5040385fffb96816028352c255682",
"sha256:692a2a5a55ee39a42bcb7557930e2541da85df9ea81c6e24827f63b80cd39d0b",
"sha256:82041dbaa62f7fde1464d7ab449978618a38b241b40c0d31dafabb36446635dc",
"sha256:958fc9f8fd5cc9b936b2cb9d96f02aa5ec3613ba13ee7f089c77ff0bcc368fac",
"sha256:c0cbe8d45ba8d8213ad68ef9a1881002a151569c9424d551634195a18c3a4160",
"sha256:ccfe0419eb5e49f27ad35cf06e75360af903df6d576c66cb8073246d4e023e5c",
"sha256:d4d2fed1a251ea58eed47b48db3778ebb92f5ff4407dc91869c6f41c3a9249d0"
],
"index": "pypi",
"version": "==3.3.0.23"
},
"celery": {
"hashes": [
"sha256:434b4e62308428b0e83fbcd1cebb8853beb230237b1b23484063fcf1ff610b6e",
"sha256:bd6a21290c822ecf78947a7dbc8b2270d88a5fa08a7a2bfcd908bb3d5ed66b71"
],
"index": "pypi",
"version": "==3.1.19"
},
"certifi": {
"hashes": [
"sha256:046832c04d4e752f37383b628bc601a7ea7211496b4638f6514d0e5b9acc4939",
"sha256:945e3ba63a0b9f577b1395204e13c3a231f9bc0223888be653286534e5873695"
],
"index": "pypi",
"version": "==2019.6.16"
},
"chardet": {
"hashes": [
"sha256:84ab92ed1c4d4f16916e05906b6b75a6c0fb5db821cc65e70cbd64a3e2a5eaae",
"sha256:fc323ffcaeaed0e0a02bf4d117757b98aed530d9ed4531e3e15460124c106691"
],
"index": "pypi",
"version": "==3.0.4"
},
"flower": {
"editable": true,
"git": "https://github.com/mher/flower",
"ref": "f20f43c858df3c0a94e5e10015fcd7d52089e3f0"
},
"idna": {
"hashes": [
"sha256:c357b3f628cf53ae2c4c05627ecc484553142ca23264e593d327bcde5e9c3407",
"sha256:ea8b7f6188e6fa117537c3df7da9fc686d485087abf6ac197f9c46432f7e4a3c"
],
"index": "pypi",
"version": "==2.8"
},
"importlib-metadata": {
"hashes": [
"sha256:b044f07694ef14a6683b097ba56bd081dbc7cdc7c7fe46011e499dfecc082f21",
"sha256:e6ac600a142cf2db707b1998382cc7fc3b02befb7273876e01b8ad10b9652742"
],
"index": "pypi",
"version": "==1.1.0"
},
"kombu": {
"hashes": [
"sha256:7ceab743e3e974f3e5736082e8cc514c009e254e646d6167342e0e192aee81a6",
"sha256:e064a00c66b4d1058cd2b0523fb8d98c82c18450244177b6c0f7913016642650"
],
"index": "pypi",
"version": "==3.0.37"
},
"more-itertools": {
"hashes": [
"sha256:b84b238cce0d9adad5ed87e745778d20a3f8487d0f0cb8b8a586816c7496458d",
"sha256:c833ef592a0324bcc6a60e48440da07645063c453880c9477ceb22490aec1564"
],
"index": "pypi",
"version": "==8.0.2"
},
"pytz": {
"hashes": [
"sha256:3ede470d3d17ba3c07638dfa0d10452bc1b6e5ad326127a65ba77e6aaeb11bec",
"sha256:8b6ce1c993909783bc96e0b4f34ea223bff7a4df2c90bdb9c4e0f1ac928689e3",
"sha256:99266ef30a37e43932deec2b7ca73e83c8dbc3b9ff703ec73eca6b1dae6befea",
"sha256:fbd26746772c24cb93c8b97cbdad5cb9e46c86bbdb1b9d8a743ee00e2fb1fc5d"
],
"index": "pypi",
"version": "==2015.7"
},
"requests": {
"hashes": [
"sha256:11e007a8a2aa0323f5a921e9e6a2d7e4e67d9877e85773fba9ba6419025cbeb4",
"sha256:9cf5292fcd0f598c671cfc1e0d7d1a7f13bb8085e9a590f48c010551dc6c4b31"
],
"index": "pypi",
"version": "==2.22.0"
},
"tornado": {
"hashes": [
"sha256:e8b1207da67dbdceebfb291292b4ef1b547d6171525bec1b366853f923456a5f"
],
"index": "pypi",
"version": "==4.2"
},
"urllib3": {
"hashes": [
"sha256:b246607a25ac80bedac05c6f282e3cdaf3afb65420fd024ac94435cabe6e18d1",
"sha256:dbe59173209418ae49d485b87d1681aefa36252ee85884c31346debd19463232"
],
"index": "pypi",
"version": "==1.25.3"
},
"vine": {
"hashes": [
"sha256:133ee6d7a9016f177ddeaf191c1f58421a1dcc6ee9a42c58b34bed40e1d2cd87",
"sha256:ea4947cc56d1fd6f2095c8d543ee25dad966f78692528e68b4fada11ba3f98af"
],
"index": "pypi",
"version": "==1.3.0"
},
"zipp": {
"hashes": [
"sha256:3718b1cbcd963c7d4c5511a8240812904164b7f381b647143a89d3b98f9bcd8e",
"sha256:f06903e9f1f43b12d371004b4ac7b06ab39a44adc747266928ae6debfa7b3335"
],
"index": "pypi",
"version": "==0.6.0"
}
},
"develop": {}
}
```
</details> | priority | editable dependency causes wrong versions to be installed issue description an error is not reported at the end of a pipenv sync command if an installed dependency violates a constraint does not match the version specified in the lockfile an error is not reported at the end of creating a lockfile via pipenv install if an installed dependency violates a constraint specified in the pipenv file newly created lockfile looks like an editable dependency is key to causing this issue may be related to note that there appears to be a datarace involved as sometimes when i run pipenv install the versions that are installed match the generated established lockfile and sometimes they do not the test case i describe here will be able to repeat the issue consistently it appears to happen whenever there is a dependency of an editable dependency that is loosely locked down or or or or this rarely happens if at all when such a dependency is locked down exactly output from test text starting tests testing pipenv sync operation iteration creating a virtualenv for this project… pipfile workspace pipfile using usr local bin python to create virtualenv… ⠋ creating virtual environment already using interpreter usr local bin python using base prefix usr local new python executable in workspace venv bin python installing setuptools pip wheel done ✔ successfully created virtual environment virtualenv location workspace venv installing dependencies from pipfile lock … 🐍 ▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉▉ — to activate this project s virtualenv run pipenv shell alternatively run a command inside the virtualenv with pipenv run all dependencies are now up to date verifying fail the following python packages did not install correctly amqp expected got instead celery expected got instead billiard expected got instead expected result i expect installed dependency versions to not violate constraints specified in the pipenv file and i expect dependencies that are installed when running pipenv sync or pipenv install to match exactly the version specified in the lockfile whether that lockfile be pre existing or newly created if this does not happen i expect an error to be reported in addition if the lockfile exists and i run pipenv sync i still expect the lockfile and pipenv file to remain unchanged though there appears to be no issue around that combination of conditions i feel it should be enumerated actual result versions of dependencies that violate the top most pipenv file are installed when running pipenv install also they do not match the version stated in the newly created lockfile versions of dependencies that do not match the lockfile are installed when running pipenv sync no error is thrown steps to replicate made a couple of minor tests that express the problem on a linux friendly system with make clone cd test python project install docker make test view the readme md file in the project root for more details there are two tests calling pipenv sync and calling pipenv install which creates a lockfile pipenv support pipenv version pipenv location usr local lib site packages pipenv python location usr local bin python python installations found usr local bin python usr local bin usr bin usr bin usr bin pep information implementation name cpython implementation version os name posix platform machine platform python implementation cpython platform release linuxkit platform system linux platform version smp tue jul utc python full version python version sys platform linux system environment variables pipenv venv in project hostname python version docker shell active pwd home lang gpg key term shlvl python pip version python get pip python get pip url path pip disable pip version check pythondontwritebytecode pip shims base module pip python path pythonfinder ignore unsupported pipenv–specific environment variables pipenv venv in project debug–specific environment variables path usr local bin usr local sbin usr local bin usr sbin usr bin sbin bin lang c utf pwd workspace contents of pipfile workspace pipfile toml name pypi url verify ssl true requests anyjson babel certifi chardet idna pytz tornado vine flower editable true git celery kombu amqp billiard importlib metadata zipp more itertools python version contents of pipfile lock workspace pipfile lock json meta hash pipfile spec requires python version sources name pypi url verify ssl true default amqp hashes index pypi version anyjson hashes index pypi version babel hashes index pypi version billiard hashes index pypi version celery hashes index pypi version certifi hashes index pypi version chardet hashes index pypi version flower editable true git ref idna hashes index pypi version importlib metadata hashes index pypi version kombu hashes index pypi version more itertools hashes index pypi version pytz hashes index pypi version requests hashes index pypi version tornado hashes index pypi version hashes index pypi version vine hashes index pypi version zipp hashes index pypi version develop | 1 |
187,514 | 6,758,499,995 | IssuesEvent | 2017-10-24 14:23:13 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Using FIx Character Values to adjust stats temporarily makes the website append not add | hacktoberfest priority: medium section: Settings status: issue: in progress | Using FIx Character Values to increase your gold by a certain number temporarily makes the website **append** that number to your gold balance, not **add** it. To replicate the bug:
1. Go to User Icon > Settings > Fix Character Values
2. Adjust your gold to have any non-zero value (e.g. 42).
3. Click "Save & Close".
4. Go to your task list and click a positive Habit.
5. Observe your gold balance: the gold you just earned has been **appended** to your gold balance, not **added** to it. E.g., if you earned 1.5 gold, your balance will change from 42 to 421.5 (it should be 43.5)
6. Click a positive Habit again.
7. Your gold balance drops to zero.
8. Reload the website. Your gold balance goes back to what it should be (e.g., about 45).
Examining my values in the database during that test showed that my actual gold balance was always correct in my account; it's just a display issue in the website.
Something similar happens for Experience and presumably other stats. | 1.0 | Using FIx Character Values to adjust stats temporarily makes the website append not add - Using FIx Character Values to increase your gold by a certain number temporarily makes the website **append** that number to your gold balance, not **add** it. To replicate the bug:
1. Go to User Icon > Settings > Fix Character Values
2. Adjust your gold to have any non-zero value (e.g. 42).
3. Click "Save & Close".
4. Go to your task list and click a positive Habit.
5. Observe your gold balance: the gold you just earned has been **appended** to your gold balance, not **added** to it. E.g., if you earned 1.5 gold, your balance will change from 42 to 421.5 (it should be 43.5)
6. Click a positive Habit again.
7. Your gold balance drops to zero.
8. Reload the website. Your gold balance goes back to what it should be (e.g., about 45).
Examining my values in the database during that test showed that my actual gold balance was always correct in my account; it's just a display issue in the website.
Something similar happens for Experience and presumably other stats. | priority | using fix character values to adjust stats temporarily makes the website append not add using fix character values to increase your gold by a certain number temporarily makes the website append that number to your gold balance not add it to replicate the bug go to user icon settings fix character values adjust your gold to have any non zero value e g click save close go to your task list and click a positive habit observe your gold balance the gold you just earned has been appended to your gold balance not added to it e g if you earned gold your balance will change from to it should be click a positive habit again your gold balance drops to zero reload the website your gold balance goes back to what it should be e g about examining my values in the database during that test showed that my actual gold balance was always correct in my account it s just a display issue in the website something similar happens for experience and presumably other stats | 1 |
528,326 | 15,364,509,838 | IssuesEvent | 2021-03-01 22:06:17 | Baystation12/Baystation12 | https://api.github.com/repos/Baystation12/Baystation12 | closed | Cycling airlocks lose air | Atmos :bug: Could Not Reproduce :bug: Priority: Medium | I thought this was already here, tried searching for it but might have missed it.
1. Enter airlock. I observed this on research outpost airlock and mining outpost airlock (that one in western part of the outpost) or on engineering outpost.
2. Cycle out and in few times.
3. Observe as the airlock loses large amounts of air with every out-in cycle. (Note: I know there would be minor loss due to pumps not venting the air completely, this is different, this minor loss would be few kPa per cycle, this is more than one hundred!)
4. Eventually, after two or three cycles (maybe less) the airlock ceases to function when cycling in. The airlock will be stuck on partial pressure (example: 73kPa), with pumps being unable to fill it as the canisters are completely empty.
I suspect the vents somehow "eat" part of the gas.
| 1.0 | Cycling airlocks lose air - I thought this was already here, tried searching for it but might have missed it.
1. Enter airlock. I observed this on research outpost airlock and mining outpost airlock (that one in western part of the outpost) or on engineering outpost.
2. Cycle out and in few times.
3. Observe as the airlock loses large amounts of air with every out-in cycle. (Note: I know there would be minor loss due to pumps not venting the air completely, this is different, this minor loss would be few kPa per cycle, this is more than one hundred!)
4. Eventually, after two or three cycles (maybe less) the airlock ceases to function when cycling in. The airlock will be stuck on partial pressure (example: 73kPa), with pumps being unable to fill it as the canisters are completely empty.
I suspect the vents somehow "eat" part of the gas.
| priority | cycling airlocks lose air i thought this was already here tried searching for it but might have missed it enter airlock i observed this on research outpost airlock and mining outpost airlock that one in western part of the outpost or on engineering outpost cycle out and in few times observe as the airlock loses large amounts of air with every out in cycle note i know there would be minor loss due to pumps not venting the air completely this is different this minor loss would be few kpa per cycle this is more than one hundred eventually after two or three cycles maybe less the airlock ceases to function when cycling in the airlock will be stuck on partial pressure example with pumps being unable to fill it as the canisters are completely empty i suspect the vents somehow eat part of the gas | 1 |
716,187 | 24,625,063,162 | IssuesEvent | 2022-10-16 12:12:46 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Renovate creates multiple PRs for common versions in gradle version catalog | type:bug priority-3-medium manager:gradle status:ready reproduction:provided | ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
renovatebot/github-action@v29.32.1
### Please select which platform you are using if self-hosting.
github.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Describe the bug
When using gradle version catalog version refs for common versions, a single change in the common version will trigger the creation of multiple PRs, one for each dependency.
The changes and the CI to check them will be the same.
Reproducer: https://github.com/F43nd1r/renovate-bot-multiple-prs-bug-reproducer
### Relevant debug logs
_No response_
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | 1.0 | Renovate creates multiple PRs for common versions in gradle version catalog - ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
renovatebot/github-action@v29.32.1
### Please select which platform you are using if self-hosting.
github.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Describe the bug
When using gradle version catalog version refs for common versions, a single change in the common version will trigger the creation of multiple PRs, one for each dependency.
The changes and the CI to check them will be the same.
Reproducer: https://github.com/F43nd1r/renovate-bot-multiple-prs-bug-reproducer
### Relevant debug logs
_No response_
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | priority | renovate creates multiple prs for common versions in gradle version catalog how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run renovatebot github action please select which platform you are using if self hosting github com if you re self hosting renovate tell us what version of the platform you run no response describe the bug when using gradle version catalog version refs for common versions a single change in the common version will trigger the creation of multiple prs one for each dependency the changes and the ci to check them will be the same reproducer relevant debug logs no response have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description | 1 |
718,760 | 24,731,058,291 | IssuesEvent | 2022-10-20 17:37:41 | wasmerio/wasmer | https://api.github.com/repos/wasmerio/wasmer | closed | Fastest way to transfer data between WASM and Rust? | ❓ question 📦 lib-types 🏚 stale priority-medium | Currently I'm using memory.view() to transfer bytedata between WASM and Rust. This data is generally arbitrarily placed and sized. It may be possible for the size to be a compile-time constant, but it is currently set at runtime.
I've been disappointed that to get or set the data I have to call cell.get() or cell.set() for every single byte. This seems like it will likely needlessly slow things down, unless significant optimizations are made. In addition, it prevents me from using WASM memory directly eg with read as a buffer.
I've noticed there's an atomically() function. This changes the type of each byte in AtomicU8 rather than Cell<u8>.
Which of these is closer to the native representation?
Which is expected to generally be faster?
Why isn't there a way to directly access the bytes of the WASM memory space?
Thanks. | 1.0 | Fastest way to transfer data between WASM and Rust? - Currently I'm using memory.view() to transfer bytedata between WASM and Rust. This data is generally arbitrarily placed and sized. It may be possible for the size to be a compile-time constant, but it is currently set at runtime.
I've been disappointed that to get or set the data I have to call cell.get() or cell.set() for every single byte. This seems like it will likely needlessly slow things down, unless significant optimizations are made. In addition, it prevents me from using WASM memory directly eg with read as a buffer.
I've noticed there's an atomically() function. This changes the type of each byte in AtomicU8 rather than Cell<u8>.
Which of these is closer to the native representation?
Which is expected to generally be faster?
Why isn't there a way to directly access the bytes of the WASM memory space?
Thanks. | priority | fastest way to transfer data between wasm and rust currently i m using memory view to transfer bytedata between wasm and rust this data is generally arbitrarily placed and sized it may be possible for the size to be a compile time constant but it is currently set at runtime i ve been disappointed that to get or set the data i have to call cell get or cell set for every single byte this seems like it will likely needlessly slow things down unless significant optimizations are made in addition it prevents me from using wasm memory directly eg with read as a buffer i ve noticed there s an atomically function this changes the type of each byte in rather than cell which of these is closer to the native representation which is expected to generally be faster why isn t there a way to directly access the bytes of the wasm memory space thanks | 1 |
349,532 | 10,470,777,230 | IssuesEvent | 2019-09-23 05:32:22 | AY1920S1-CS2113T-W12-3/main | https://api.github.com/repos/AY1920S1-CS2113T-W12-3/main | opened | As a hall resident, I want the booking timings to be more flexible | priority.Medium type.Story | so that I do not book the facilities for longer than I need | 1.0 | As a hall resident, I want the booking timings to be more flexible - so that I do not book the facilities for longer than I need | priority | as a hall resident i want the booking timings to be more flexible so that i do not book the facilities for longer than i need | 1 |
96,442 | 3,968,472,601 | IssuesEvent | 2016-05-03 19:47:34 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Index: Index according to data type to facilitate range facets and queries. | Component: Search/Browse Priority: Medium Status: Duplicate Status: Triaged Type: Feature | ---
Author Name: **Kevin Condon** (@kcondon)
Original Redmine Issue: 3785, https://redmine.hmdc.harvard.edu/issues/3785
Original Date: 2014-03-27
Original Assignee: Philip Durbin
---
This ticket is a placeholder for indexing fields that contain something other than strings, eg. dates and floats, to support range queries for either facets or searches, depending on what is finally decided.
Example date: production date
Example float: Coverage Central Wavelength(m)
---
Related issue(s): #70, #485
Redmine related issue(s): [3478](https://redmine.hmdc.harvard.edu/issues/3478), [3900](https://redmine.hmdc.harvard.edu/issues/3900)
---
| 1.0 | Index: Index according to data type to facilitate range facets and queries. - ---
Author Name: **Kevin Condon** (@kcondon)
Original Redmine Issue: 3785, https://redmine.hmdc.harvard.edu/issues/3785
Original Date: 2014-03-27
Original Assignee: Philip Durbin
---
This ticket is a placeholder for indexing fields that contain something other than strings, eg. dates and floats, to support range queries for either facets or searches, depending on what is finally decided.
Example date: production date
Example float: Coverage Central Wavelength(m)
---
Related issue(s): #70, #485
Redmine related issue(s): [3478](https://redmine.hmdc.harvard.edu/issues/3478), [3900](https://redmine.hmdc.harvard.edu/issues/3900)
---
| priority | index index according to data type to facilitate range facets and queries author name kevin condon kcondon original redmine issue original date original assignee philip durbin this ticket is a placeholder for indexing fields that contain something other than strings eg dates and floats to support range queries for either facets or searches depending on what is finally decided example date production date example float coverage central wavelength m related issue s redmine related issue s | 1 |
743,293 | 25,893,795,590 | IssuesEvent | 2022-12-14 20:22:58 | RoleModel/rolemodel-design-system | https://api.github.com/repos/RoleModel/rolemodel-design-system | opened | Revisit Forms | Cleanup Breaking Change Priority: Medium Simplification | **Is your feature request related to a problem? Please describe.**
The form styles are great, but we have had some simplifying ideas to make them easier to use
**Describe the solution you'd like**
```
.form-group {}
.form-label {}
.form-control {}
.form-control--error {}
.form-control--read-only {}
.form-error {}
.form-hint {}
.form-error-summary {}
```
.form-group
label.form-label for='thing' My Label
input.form-control#thing type='text'
.form-error conditional error message
.form-hint optional hint message
.form-group
label.form-label for='thing' My Label
textarea.form-control#thing
.form-error conditional error message
.form-hint optional hint message
```
We may want to have an alternative form group for horizontal alignment instead of always vertical | 1.0 | Revisit Forms - **Is your feature request related to a problem? Please describe.**
The form styles are great, but we have had some simplifying ideas to make them easier to use
**Describe the solution you'd like**
```
.form-group {}
.form-label {}
.form-control {}
.form-control--error {}
.form-control--read-only {}
.form-error {}
.form-hint {}
.form-error-summary {}
```
.form-group
label.form-label for='thing' My Label
input.form-control#thing type='text'
.form-error conditional error message
.form-hint optional hint message
.form-group
label.form-label for='thing' My Label
textarea.form-control#thing
.form-error conditional error message
.form-hint optional hint message
```
We may want to have an alternative form group for horizontal alignment instead of always vertical | priority | revisit forms is your feature request related to a problem please describe the form styles are great but we have had some simplifying ideas to make them easier to use describe the solution you d like form group form label form control form control error form control read only form error form hint form error summary form group label form label for thing my label input form control thing type text form error conditional error message form hint optional hint message form group label form label for thing my label textarea form control thing form error conditional error message form hint optional hint message we may want to have an alternative form group for horizontal alignment instead of always vertical | 1 |
40,779 | 2,868,941,788 | IssuesEvent | 2015-06-05 22:05:47 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | closed | Consider using normalize.css instead of the Compass Reset. | Component-UI enhancement Priority-Medium wontfix | <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8384_
----
**Attachment:**
[css-reset-debugging.png](https://storage.googleapis.com/google-code-attachments/dart/issue-8384/comment-0/css-reset-debugging.png) (49.42 KB) | 1.0 | Consider using normalize.css instead of the Compass Reset. - <a href="https://github.com/amouravski"><img src="https://avatars.githubusercontent.com/u/264967?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [amouravski](https://github.com/amouravski)**
_Originally opened as dart-lang/sdk#8384_
----
**Attachment:**
[css-reset-debugging.png](https://storage.googleapis.com/google-code-attachments/dart/issue-8384/comment-0/css-reset-debugging.png) (49.42 KB) | priority | consider using normalize css instead of the compass reset issue by originally opened as dart lang sdk attachment kb | 1 |
736,669 | 25,482,492,867 | IssuesEvent | 2022-11-26 00:38:45 | bcgov/cas-cif | https://api.github.com/repos/bcgov/cas-cif | closed | As a CIF administrator, I want to be alerted if the calculated province's share is different than the one entered on the "budgets, expenses and payments" form, so that I can ensure data consistency | Backlog Refinement attn: Business Area User Story Needs AC Medium Priority | #### Description:
Should we have an alert / an indicator showing if the calculated province's share is different than the target one (typically 50%) ?
#### Acceptance Criteria:
Given I am on the 'budgets' page
When the calculated province's share is different (by > 1%? TBC Business area) than the funding agreement value
Then I can see an alert near the field
#### Development Checklist:
- [ ] Implement the alert component
- [ ] Display the alert component near the calculated province share field if the value exceeds the threshold
- [ ] Meets the DOD
**Definition of Ready** (Note: If any of these points are not applicable, mark N/A)
- [ ] User story is included
- [ ] User role and type are identified
- [ ] Acceptance criteria are included
- [ ] Wireframes are included (if required)
- [ ] Design / Solution is accepted by Product Owner
- [ ] Dependencies are identified (technical, business, regulatory/policy)
- [ ] Story has been estimated (under 13 pts)
·**Definition of Done** (Note: If any of these points are not applicable, mark N/A)
- [ ] Acceptance criteria are tested by the CI pipeline
- [ ] UI meets accessibility requirements
- [ ] Configuration changes are documented, documentation and designs are updated
- [ ] Passes code peer-review
- [ ] Passes QA of Acceptance Criteria with verification in Dev and Test
- [ ] Ticket is ready to be merged to main branch
- [ ] Can be demoed in Sprint Review
- [ ] Bugs or future work cards are identified and created
- [ ] Reviewed and approved by Product Owner
#### Notes:
-
| 1.0 | As a CIF administrator, I want to be alerted if the calculated province's share is different than the one entered on the "budgets, expenses and payments" form, so that I can ensure data consistency - #### Description:
Should we have an alert / an indicator showing if the calculated province's share is different than the target one (typically 50%) ?
#### Acceptance Criteria:
Given I am on the 'budgets' page
When the calculated province's share is different (by > 1%? TBC Business area) than the funding agreement value
Then I can see an alert near the field
#### Development Checklist:
- [ ] Implement the alert component
- [ ] Display the alert component near the calculated province share field if the value exceeds the threshold
- [ ] Meets the DOD
**Definition of Ready** (Note: If any of these points are not applicable, mark N/A)
- [ ] User story is included
- [ ] User role and type are identified
- [ ] Acceptance criteria are included
- [ ] Wireframes are included (if required)
- [ ] Design / Solution is accepted by Product Owner
- [ ] Dependencies are identified (technical, business, regulatory/policy)
- [ ] Story has been estimated (under 13 pts)
·**Definition of Done** (Note: If any of these points are not applicable, mark N/A)
- [ ] Acceptance criteria are tested by the CI pipeline
- [ ] UI meets accessibility requirements
- [ ] Configuration changes are documented, documentation and designs are updated
- [ ] Passes code peer-review
- [ ] Passes QA of Acceptance Criteria with verification in Dev and Test
- [ ] Ticket is ready to be merged to main branch
- [ ] Can be demoed in Sprint Review
- [ ] Bugs or future work cards are identified and created
- [ ] Reviewed and approved by Product Owner
#### Notes:
-
| priority | as a cif administrator i want to be alerted if the calculated province s share is different than the one entered on the budgets expenses and payments form so that i can ensure data consistency description should we have an alert an indicator showing if the calculated province s share is different than the target one typically acceptance criteria given i am on the budgets page when the calculated province s share is different by tbc business area than the funding agreement value then i can see an alert near the field development checklist implement the alert component display the alert component near the calculated province share field if the value exceeds the threshold meets the dod definition of ready note if any of these points are not applicable mark n a user story is included user role and type are identified acceptance criteria are included wireframes are included if required design solution is accepted by product owner dependencies are identified technical business regulatory policy story has been estimated under pts · definition of done note if any of these points are not applicable mark n a acceptance criteria are tested by the ci pipeline ui meets accessibility requirements configuration changes are documented documentation and designs are updated passes code peer review passes qa of acceptance criteria with verification in dev and test ticket is ready to be merged to main branch can be demoed in sprint review bugs or future work cards are identified and created reviewed and approved by product owner notes | 1 |
587,242 | 17,610,728,321 | IssuesEvent | 2021-08-18 00:36:00 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | opened | Migrate the new "beta" UI from AngularJS to Angular | Priority-Medium Component-UI Maintainability Difficulty-Medium Language-JS | https://blog.angular.io/stable-angularjs-and-long-term-support-7e077635ee9c
"AngularJS is an extremely stable framework for building web applications, and has been used by millions of developers across the web. Angular is its successor and uses the same philosophies such as declarative templates and dependency injection."
"July 1, 2018 — December 31, 2021 AngularJS 1.7 LTS Period"
https://angular.io/guide/upgrade | 1.0 | Migrate the new "beta" UI from AngularJS to Angular - https://blog.angular.io/stable-angularjs-and-long-term-support-7e077635ee9c
"AngularJS is an extremely stable framework for building web applications, and has been used by millions of developers across the web. Angular is its successor and uses the same philosophies such as declarative templates and dependency injection."
"July 1, 2018 — December 31, 2021 AngularJS 1.7 LTS Period"
https://angular.io/guide/upgrade | priority | migrate the new beta ui from angularjs to angular angularjs is an extremely stable framework for building web applications and has been used by millions of developers across the web angular is its successor and uses the same philosophies such as declarative templates and dependency injection july — december angularjs lts period | 1 |
435,397 | 12,535,017,505 | IssuesEvent | 2020-06-04 20:32:47 | MLH-Fellowship/CodeVidLive | https://api.github.com/repos/MLH-Fellowship/CodeVidLive | opened | Simulation Based Prediction Model | priority/medium | Run agent based SIRs with spatial movement simulation then make a prediction model:
(case_location_coordinates, user_location_coordinates, time window)=> probability | 1.0 | Simulation Based Prediction Model - Run agent based SIRs with spatial movement simulation then make a prediction model:
(case_location_coordinates, user_location_coordinates, time window)=> probability | priority | simulation based prediction model run agent based sirs with spatial movement simulation then make a prediction model case location coordinates user location coordinates time window probability | 1 |
374,703 | 11,094,212,329 | IssuesEvent | 2019-12-16 05:23:33 | BCcampus/edehr | https://api.github.com/repos/BCcampus/edehr | closed | pdf files | Effort - Low Epic - Styles Priority - Medium ~Feature | PDF files should appear with the "pdf" icon to it's left. These files do not require "external link" icons.
| 1.0 | pdf files - PDF files should appear with the "pdf" icon to it's left. These files do not require "external link" icons.
| priority | pdf files pdf files should appear with the pdf icon to it s left these files do not require external link icons | 1 |
585,705 | 17,515,283,926 | IssuesEvent | 2021-08-11 05:36:10 | WordPress/learn | https://api.github.com/repos/WordPress/learn | opened | Learner recognition on WordPress.org profiles | [Type] Enhancement [Priority] Medium | As per [this proposal](https://make.wordpress.org/training/2021/08/10/proposal-learner-achievements-on-profiles/), we would like to recognise learner achievements on WordPress profiles. The ways we would like to do this are:
* An activity stream entry whenever a learner completes a full course.
* A persistent line showing the learner’s average grade on Learn WordPress – I envision this in the top right info block above the social links, but happy to be shown otherwise.
* A new tab under ‘Activity’ called ‘Learning’ (or similar) that shows all of the courses they have completed along with their average grade for the course and what date they completed it on. The total average grade must also be shown in here.
* A ‘Learner’ badge for everyone who has completed at least one course. This badge will use the same icon as the Training Team badge, but in a different colour that still needs to be confirmed.
Technically we can already do all of that, but I'm aware that I'm asking for changes to be made to the profiles themselves with the average grade and the new tab. Are there any potential issues there? Do we need some additional approval before making those changes, or should we log that elsewhere to be added? | 1.0 | Learner recognition on WordPress.org profiles - As per [this proposal](https://make.wordpress.org/training/2021/08/10/proposal-learner-achievements-on-profiles/), we would like to recognise learner achievements on WordPress profiles. The ways we would like to do this are:
* An activity stream entry whenever a learner completes a full course.
* A persistent line showing the learner’s average grade on Learn WordPress – I envision this in the top right info block above the social links, but happy to be shown otherwise.
* A new tab under ‘Activity’ called ‘Learning’ (or similar) that shows all of the courses they have completed along with their average grade for the course and what date they completed it on. The total average grade must also be shown in here.
* A ‘Learner’ badge for everyone who has completed at least one course. This badge will use the same icon as the Training Team badge, but in a different colour that still needs to be confirmed.
Technically we can already do all of that, but I'm aware that I'm asking for changes to be made to the profiles themselves with the average grade and the new tab. Are there any potential issues there? Do we need some additional approval before making those changes, or should we log that elsewhere to be added? | priority | learner recognition on wordpress org profiles as per we would like to recognise learner achievements on wordpress profiles the ways we would like to do this are an activity stream entry whenever a learner completes a full course a persistent line showing the learner’s average grade on learn wordpress – i envision this in the top right info block above the social links but happy to be shown otherwise a new tab under ‘activity’ called ‘learning’ or similar that shows all of the courses they have completed along with their average grade for the course and what date they completed it on the total average grade must also be shown in here a ‘learner’ badge for everyone who has completed at least one course this badge will use the same icon as the training team badge but in a different colour that still needs to be confirmed technically we can already do all of that but i m aware that i m asking for changes to be made to the profiles themselves with the average grade and the new tab are there any potential issues there do we need some additional approval before making those changes or should we log that elsewhere to be added | 1 |
107,148 | 4,290,068,419 | IssuesEvent | 2016-07-18 08:03:45 | juju/docs | https://api.github.com/repos/juju/docs | closed | Requested: Juju Quickstart documentation | medium priority | Juju Quickstart is missing documentation in both official and unofficial capacity in the docs. | 1.0 | Requested: Juju Quickstart documentation - Juju Quickstart is missing documentation in both official and unofficial capacity in the docs. | priority | requested juju quickstart documentation juju quickstart is missing documentation in both official and unofficial capacity in the docs | 1 |
325,230 | 9,921,416,438 | IssuesEvent | 2019-06-30 18:02:38 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | opened | Move all LOCALIZE_TOKEN (and other Localization methods) for HUD elements to Reset() function | Priority: Medium Size: Small Type: Enhancement | Since Init() or the constructor is only called once, reloading localization files on the fly will not work unless you restart the game. Moving to Reset() where it gets called every player spawn/map reload is much better for testing localization. | 1.0 | Move all LOCALIZE_TOKEN (and other Localization methods) for HUD elements to Reset() function - Since Init() or the constructor is only called once, reloading localization files on the fly will not work unless you restart the game. Moving to Reset() where it gets called every player spawn/map reload is much better for testing localization. | priority | move all localize token and other localization methods for hud elements to reset function since init or the constructor is only called once reloading localization files on the fly will not work unless you restart the game moving to reset where it gets called every player spawn map reload is much better for testing localization | 1 |
466,980 | 13,438,387,358 | IssuesEvent | 2020-09-07 17:57:55 | flextype/flextype | https://api.github.com/repos/flextype/flextype | closed | Use filp/whoops instead of zeuxisoo/slim-whoops | priority: medium type: feature | We should use native **filp/whoops** instead of legacy **zeuxisoo/slim-whoops** for errors handling | 1.0 | Use filp/whoops instead of zeuxisoo/slim-whoops - We should use native **filp/whoops** instead of legacy **zeuxisoo/slim-whoops** for errors handling | priority | use filp whoops instead of zeuxisoo slim whoops we should use native filp whoops instead of legacy zeuxisoo slim whoops for errors handling | 1 |
321,889 | 9,810,299,703 | IssuesEvent | 2019-06-12 20:08:39 | SETI/pds-opus | https://api.github.com/repos/SETI/pds-opus | closed | Infinite scroll in the table view is spastic | A-Bug B-OPUS JS Effort 2 Medium Priority 2 Essential | - In the table view, when you scroll the slider does not move by the gallery step size, but instead moves by units of 1
- When cache filling is required, the table jumps a long distance (so does the slider)
- For awhile we had the behavior that the slider would *always* step by the gallery row size. I believe this is the desired behavior, even though it looks weird in table view. I don't know what broke this.
| 1.0 | Infinite scroll in the table view is spastic - - In the table view, when you scroll the slider does not move by the gallery step size, but instead moves by units of 1
- When cache filling is required, the table jumps a long distance (so does the slider)
- For awhile we had the behavior that the slider would *always* step by the gallery row size. I believe this is the desired behavior, even though it looks weird in table view. I don't know what broke this.
| priority | infinite scroll in the table view is spastic in the table view when you scroll the slider does not move by the gallery step size but instead moves by units of when cache filling is required the table jumps a long distance so does the slider for awhile we had the behavior that the slider would always step by the gallery row size i believe this is the desired behavior even though it looks weird in table view i don t know what broke this | 1 |
682,474 | 23,345,985,907 | IssuesEvent | 2022-08-09 18:01:13 | projectdiscovery/dnsx | https://api.github.com/repos/projectdiscovery/dnsx | closed | allow/deny host list support | Priority: Medium Status: On Hold Type: Enhancement Type: Question | ### Please describe your feature request:
Porting **allow/deny** support from [httpx](https://github.com/projectdiscovery/httpx/issues/374) to dnsx
### Describe the use case of this feature:
allows us to define which hosts to be passed through or denied for processing. | 1.0 | allow/deny host list support - ### Please describe your feature request:
Porting **allow/deny** support from [httpx](https://github.com/projectdiscovery/httpx/issues/374) to dnsx
### Describe the use case of this feature:
allows us to define which hosts to be passed through or denied for processing. | priority | allow deny host list support please describe your feature request porting allow deny support from to dnsx describe the use case of this feature allows us to define which hosts to be passed through or denied for processing | 1 |
510,193 | 14,786,529,177 | IssuesEvent | 2021-01-12 05:44:38 | OpenSRP/opensrp-client-reveal | https://api.github.com/repos/OpenSRP/opensrp-client-reveal | closed | RVL-755 - Inactive versus ineligible structure colours | Blocked Priority: High Size: Medium (2-3) | **Current state:**
1. Structures marked as ineligible (ie they are not eligible during this plan) are black
2. Inactive structures (ie they are not structures that will ever be eligible, maybe were enumerated accidentally or entered in error) are grey.
**Desired Operation**
1. Ineligible - Black (determined based on current in-field data collection) - retain current functionality (ability to view & edit the status of the black structure)
2. Inactive - Grey (this only applies to NON-residential structures that were determined to be non-residential and ineligible in the previous campaign) - need to be able to update ‘inactive’ to ‘not-visited’ (yellow structure) and then proceed as normal.
See the acceptance Criteria [here](https://smartregister.atlassian.net/browse/RVL-755):
| 1.0 | RVL-755 - Inactive versus ineligible structure colours - **Current state:**
1. Structures marked as ineligible (ie they are not eligible during this plan) are black
2. Inactive structures (ie they are not structures that will ever be eligible, maybe were enumerated accidentally or entered in error) are grey.
**Desired Operation**
1. Ineligible - Black (determined based on current in-field data collection) - retain current functionality (ability to view & edit the status of the black structure)
2. Inactive - Grey (this only applies to NON-residential structures that were determined to be non-residential and ineligible in the previous campaign) - need to be able to update ‘inactive’ to ‘not-visited’ (yellow structure) and then proceed as normal.
See the acceptance Criteria [here](https://smartregister.atlassian.net/browse/RVL-755):
| priority | rvl inactive versus ineligible structure colours current state structures marked as ineligible ie they are not eligible during this plan are black inactive structures ie they are not structures that will ever be eligible maybe were enumerated accidentally or entered in error are grey desired operation ineligible black determined based on current in field data collection retain current functionality ability to view edit the status of the black structure inactive grey this only applies to non residential structures that were determined to be non residential and ineligible in the previous campaign need to be able to update ‘inactive’ to ‘not visited’ yellow structure and then proceed as normal see the acceptance criteria | 1 |
206,662 | 7,114,690,627 | IssuesEvent | 2018-01-18 02:10:18 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Test scalability of #1473 work on EC2 | Category: Core Priority: Medium Status: Defined Type: Task | Need to figure out how to launch a hoot vagrant instance on EC2 as a hadoop cluster. EMR will probably come into play somehow. | 1.0 | Test scalability of #1473 work on EC2 - Need to figure out how to launch a hoot vagrant instance on EC2 as a hadoop cluster. EMR will probably come into play somehow. | priority | test scalability of work on need to figure out how to launch a hoot vagrant instance on as a hadoop cluster emr will probably come into play somehow | 1 |
732,647 | 25,270,763,026 | IssuesEvent | 2022-11-16 09:09:20 | joakimAbdinurIusuf/population-app | https://api.github.com/repos/joakimAbdinurIusuf/population-app | closed | Use TypeScript best practices and change from "any" type to custom types | enhancement medium priority | At the moment the app is written in TypeScript, but it is defining variables and functions using the "any" data type a lot. There is one custom data type in the types folder, but the whole app should have custom data types and not use the any type. | 1.0 | Use TypeScript best practices and change from "any" type to custom types - At the moment the app is written in TypeScript, but it is defining variables and functions using the "any" data type a lot. There is one custom data type in the types folder, but the whole app should have custom data types and not use the any type. | priority | use typescript best practices and change from any type to custom types at the moment the app is written in typescript but it is defining variables and functions using the any data type a lot there is one custom data type in the types folder but the whole app should have custom data types and not use the any type | 1 |
42,562 | 2,871,557,969 | IssuesEvent | 2015-06-08 04:36:05 | lua-carbon/carbon | https://api.github.com/repos/lua-carbon/carbon | closed | Carbide line mapping and error rewriting | difficulty:hard feature graphene-2.0 priority:medium | Transpiling languages usually provide a mechanism to reference lines in the original code. This would be a very helpful feature for Carbide, especially as it grows in complexity. It could integrate with the proposed Graphene features for translating error stacks and would make debugging much easier.
Errors should also be rewritten to give original line numbers. | 1.0 | Carbide line mapping and error rewriting - Transpiling languages usually provide a mechanism to reference lines in the original code. This would be a very helpful feature for Carbide, especially as it grows in complexity. It could integrate with the proposed Graphene features for translating error stacks and would make debugging much easier.
Errors should also be rewritten to give original line numbers. | priority | carbide line mapping and error rewriting transpiling languages usually provide a mechanism to reference lines in the original code this would be a very helpful feature for carbide especially as it grows in complexity it could integrate with the proposed graphene features for translating error stacks and would make debugging much easier errors should also be rewritten to give original line numbers | 1 |
792,386 | 27,958,240,027 | IssuesEvent | 2023-03-24 13:55:52 | MaibornWolff/codecharta | https://api.github.com/repos/MaibornWolff/codecharta | closed | Add notes to a Custom View | pr-visualization priority:medium feature difficulty:low | # Feature request
<Please note by far the quickest way to get a new feature is to file a Pull Request.>
<We will consider your request but it may be closed,>
<if it's something we're not actively planning to work on>
<or it does not match our visison of this software.>
## Description
As an Auditor, I want to add custom notes to a Custom View, so that I e.g. can remember what this Custom View was intended for.
## Acceptance criteria
- I can add custom notes when creating a Custom View
- I can edit custom notes for a existing Custom View
- The notes are included in the config file and therefore can be exported automatically
| 1.0 | Add notes to a Custom View - # Feature request
<Please note by far the quickest way to get a new feature is to file a Pull Request.>
<We will consider your request but it may be closed,>
<if it's something we're not actively planning to work on>
<or it does not match our visison of this software.>
## Description
As an Auditor, I want to add custom notes to a Custom View, so that I e.g. can remember what this Custom View was intended for.
## Acceptance criteria
- I can add custom notes when creating a Custom View
- I can edit custom notes for a existing Custom View
- The notes are included in the config file and therefore can be exported automatically
| priority | add notes to a custom view feature request description as an auditor i want to add custom notes to a custom view so that i e g can remember what this custom view was intended for acceptance criteria i can add custom notes when creating a custom view i can edit custom notes for a existing custom view the notes are included in the config file and therefore can be exported automatically | 1 |
273,658 | 8,551,251,966 | IssuesEvent | 2018-11-07 17:30:42 | blackbaud/skyux2 | https://api.github.com/repos/blackbaud/skyux2 | reopened | Modals sometimes receive a visible 'focused' border around content | From: Consumer Priority: Medium accessibility | ### Expected behavior
No border appears
### Actual behavior
On latest chrome, a small blue border appears surrounding the modal-content element when the modal content does not contain an input element. Doesn't appear to happen on latest firefox.
### Steps to reproduce
Create a modal without an input field in the `sky-modal-content`, open the modal in chrome, see the blue-ish focus border around the modal-content.
### Plunker
https://next.plnkr.co/edit/uv44WosD1IYeLfp6zYl0?p=preview&utm_source=legacy&utm_medium=worker&utm_campaign=next&preview
### Severity
Minimal
### Impact
Only appears to affect modals without an input element - the confirm component shows a similar border around the buttons on the bottom (indicating they are given focus initially) | 1.0 | Modals sometimes receive a visible 'focused' border around content - ### Expected behavior
No border appears
### Actual behavior
On latest chrome, a small blue border appears surrounding the modal-content element when the modal content does not contain an input element. Doesn't appear to happen on latest firefox.
### Steps to reproduce
Create a modal without an input field in the `sky-modal-content`, open the modal in chrome, see the blue-ish focus border around the modal-content.
### Plunker
https://next.plnkr.co/edit/uv44WosD1IYeLfp6zYl0?p=preview&utm_source=legacy&utm_medium=worker&utm_campaign=next&preview
### Severity
Minimal
### Impact
Only appears to affect modals without an input element - the confirm component shows a similar border around the buttons on the bottom (indicating they are given focus initially) | priority | modals sometimes receive a visible focused border around content expected behavior no border appears actual behavior on latest chrome a small blue border appears surrounding the modal content element when the modal content does not contain an input element doesn t appear to happen on latest firefox steps to reproduce create a modal without an input field in the sky modal content open the modal in chrome see the blue ish focus border around the modal content plunker severity minimal impact only appears to affect modals without an input element the confirm component shows a similar border around the buttons on the bottom indicating they are given focus initially | 1 |
430,129 | 12,440,746,331 | IssuesEvent | 2020-05-26 12:33:34 | BgeeDB/bgee_apps | https://api.github.com/repos/BgeeDB/bgee_apps | closed | Facet menus | enhancement priority: medium ui enhancement | In GitLab by @marcrr on Oct 15, 2015, 12:09
Allow selection of download files by facet menus. | 1.0 | Facet menus - In GitLab by @marcrr on Oct 15, 2015, 12:09
Allow selection of download files by facet menus. | priority | facet menus in gitlab by marcrr on oct allow selection of download files by facet menus | 1 |
20,266 | 2,622,337,525 | IssuesEvent | 2015-03-04 01:40:47 | 0xtob/nitrotracker | https://api.github.com/repos/0xtob/nitrotracker | closed | Sample Start/End | auto-migrated Priority-Medium Type-Enhancement | ```
Don't think it's already been said, but maybe a sample start/end function?
This way, you could edit a sample without having to make a bunch of copies
of the same sample with different trimming.
```
Original issue reported on code.google.com by `Marr....@gmail.com` on 8 Apr 2010 at 2:12
* Merged into: #1 | 1.0 | Sample Start/End - ```
Don't think it's already been said, but maybe a sample start/end function?
This way, you could edit a sample without having to make a bunch of copies
of the same sample with different trimming.
```
Original issue reported on code.google.com by `Marr....@gmail.com` on 8 Apr 2010 at 2:12
* Merged into: #1 | priority | sample start end don t think it s already been said but maybe a sample start end function this way you could edit a sample without having to make a bunch of copies of the same sample with different trimming original issue reported on code google com by marr gmail com on apr at merged into | 1 |
425,803 | 12,359,346,989 | IssuesEvent | 2020-05-17 10:28:23 | UltimateCodeMonkeys/CodeMonkeysMVVM | https://api.github.com/repos/UltimateCodeMonkeys/CodeMonkeysMVVM | closed | Visual Studio Project Filter to only load the framework related assemblies | Priority: Medium Status: Review Needed Type: Enhancement | It annoys me to unload projects which are not directly part of the framework. For example samples. We should create a visual studio filter(slnf) to exclude these. | 1.0 | Visual Studio Project Filter to only load the framework related assemblies - It annoys me to unload projects which are not directly part of the framework. For example samples. We should create a visual studio filter(slnf) to exclude these. | priority | visual studio project filter to only load the framework related assemblies it annoys me to unload projects which are not directly part of the framework for example samples we should create a visual studio filter slnf to exclude these | 1 |
367,017 | 10,833,230,381 | IssuesEvent | 2019-11-11 12:27:20 | code4romania/mon-vot-android-kotlin | https://api.github.com/repos/code4romania/mon-vot-android-kotlin | opened | [Improvement] Show the 'has notes' icon for questions also in the question details screen | enhancement help wanted kotlin medium-priority november-2019 | Currently an icon is displayed in the questions list, for the questions that have notes on them. Display the icon also in the question details screen.
Also, investigate if the 'Adauga o note la intrebare' text should be changed if the question already has some notes added. | 1.0 | [Improvement] Show the 'has notes' icon for questions also in the question details screen - Currently an icon is displayed in the questions list, for the questions that have notes on them. Display the icon also in the question details screen.
Also, investigate if the 'Adauga o note la intrebare' text should be changed if the question already has some notes added. | priority | show the has notes icon for questions also in the question details screen currently an icon is displayed in the questions list for the questions that have notes on them display the icon also in the question details screen also investigate if the adauga o note la intrebare text should be changed if the question already has some notes added | 1 |
449,801 | 12,975,178,171 | IssuesEvent | 2020-07-21 16:33:43 | oslc-op/oslc-specs | https://api.github.com/repos/oslc-op/oslc-specs | closed | Core spec has 17 HTML5 validation errors | Area: Core Priority: Medium | With https://chrome.google.com/webstore/detail/validity/bbicmjjbohdfglopkidebfccilipgeif, I get 17 HTML5 errors in the Core spec.
<details>
<summary>HTML5 validity report</summary>
VM94:1 line 191: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 196: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 197: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 254: Consider using the “h1” element as a top-level heading only (all “h1” elements are treated as top-level headings by many screen readers and other tools).
(anonymous) @ VM94:1
VM94:1 line 280: Element “br” not allowed as child of element “dl” in this context. (Suppressing further errors from this subtree.)
(anonymous) @ VM94:1
VM94:1 line 281: Element “p” not allowed as child of element “dl” in this context. (Suppressing further errors from this subtree.)
(anonymous) @ VM94:1
VM94:1 line 337: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 352: No “p” element in scope but a “p” end tag seen.
(anonymous) @ VM94:1
VM94:1 line 420: End tag “div” seen, but there were open elements.
(anonymous) @ VM94:1
VM94:1 line 416: Unclosed element “section”.
(anonymous) @ VM94:1
VM94:1 line 416: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 725: No “p” element in scope but a “p” end tag seen.
(anonymous) @ VM94:1
VM94:1 line 776: The “cellspacing” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 776: The “cellpadding” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 776: The “border” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 779: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 779: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 781: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 781: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 783: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 783: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 785: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 785: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 2: This document appears to be written in English. Consider adding “lang="en"” (or variant) to the “html” start tag.
</details> | 1.0 | Core spec has 17 HTML5 validation errors - With https://chrome.google.com/webstore/detail/validity/bbicmjjbohdfglopkidebfccilipgeif, I get 17 HTML5 errors in the Core spec.
<details>
<summary>HTML5 validity report</summary>
VM94:1 line 191: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 196: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 197: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 254: Consider using the “h1” element as a top-level heading only (all “h1” elements are treated as top-level headings by many screen readers and other tools).
(anonymous) @ VM94:1
VM94:1 line 280: Element “br” not allowed as child of element “dl” in this context. (Suppressing further errors from this subtree.)
(anonymous) @ VM94:1
VM94:1 line 281: Element “p” not allowed as child of element “dl” in this context. (Suppressing further errors from this subtree.)
(anonymous) @ VM94:1
VM94:1 line 337: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 352: No “p” element in scope but a “p” end tag seen.
(anonymous) @ VM94:1
VM94:1 line 420: End tag “div” seen, but there were open elements.
(anonymous) @ VM94:1
VM94:1 line 416: Unclosed element “section”.
(anonymous) @ VM94:1
VM94:1 line 416: Section lacks heading. Consider using “h2”-“h6” elements to add identifying headings to all sections.
(anonymous) @ VM94:1
VM94:1 line 725: No “p” element in scope but a “p” end tag seen.
(anonymous) @ VM94:1
VM94:1 line 776: The “cellspacing” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 776: The “cellpadding” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 776: The “border” attribute on the “table” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 779: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 779: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 781: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 781: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 783: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 783: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 785: The “width” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 785: The “valign” attribute on the “td” element is obsolete. Use CSS instead.
(anonymous) @ VM94:1
VM94:1 line 2: This document appears to be written in English. Consider adding “lang="en"” (or variant) to the “html” start tag.
</details> | priority | core spec has validation errors with i get errors in the core spec validity report line section lacks heading consider using “ ” “ ” elements to add identifying headings to all sections anonymous line section lacks heading consider using “ ” “ ” elements to add identifying headings to all sections anonymous line section lacks heading consider using “ ” “ ” elements to add identifying headings to all sections anonymous line consider using the “ ” element as a top level heading only all “ ” elements are treated as top level headings by many screen readers and other tools anonymous line element “br” not allowed as child of element “dl” in this context suppressing further errors from this subtree anonymous line element “p” not allowed as child of element “dl” in this context suppressing further errors from this subtree anonymous line section lacks heading consider using “ ” “ ” elements to add identifying headings to all sections anonymous line no “p” element in scope but a “p” end tag seen anonymous line end tag “div” seen but there were open elements anonymous line unclosed element “section” anonymous line section lacks heading consider using “ ” “ ” elements to add identifying headings to all sections anonymous line no “p” element in scope but a “p” end tag seen anonymous line the “cellspacing” attribute on the “table” element is obsolete use css instead anonymous line the “cellpadding” attribute on the “table” element is obsolete use css instead anonymous line the “border” attribute on the “table” element is obsolete use css instead anonymous line the “width” attribute on the “td” element is obsolete use css instead anonymous line the “valign” attribute on the “td” element is obsolete use css instead anonymous line the “width” attribute on the “td” element is obsolete use css instead anonymous line the “valign” attribute on the “td” element is obsolete use css instead anonymous line the “width” attribute on the “td” element is obsolete use css instead anonymous line the “valign” attribute on the “td” element is obsolete use css instead anonymous line the “width” attribute on the “td” element is obsolete use css instead anonymous line the “valign” attribute on the “td” element is obsolete use css instead anonymous line this document appears to be written in english consider adding “lang en ” or variant to the “html” start tag | 1 |
663,750 | 22,205,257,469 | IssuesEvent | 2022-06-07 14:25:10 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | Seed phrase words are cut | bug ui onboarding priority 2: medium E:Onboarding | ### Description
<img width="1541" alt="Screenshot 2022-06-02 at 12 18 42" src="https://user-images.githubusercontent.com/82375995/171598697-c836c554-d239-4006-8848-402327edf775.png">
**version**: master `046aa365655ab65e67fbf5beff0d36d713d9df59` | 1.0 | Seed phrase words are cut - ### Description
<img width="1541" alt="Screenshot 2022-06-02 at 12 18 42" src="https://user-images.githubusercontent.com/82375995/171598697-c836c554-d239-4006-8848-402327edf775.png">
**version**: master `046aa365655ab65e67fbf5beff0d36d713d9df59` | priority | seed phrase words are cut description img width alt screenshot at src version master | 1 |
311,806 | 9,539,373,915 | IssuesEvent | 2019-04-30 16:46:59 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | Api/V8 create record does not support unicode and space in attributes | API Fix Proposed Medium Priority Resolved: Next Release bug size: 2 | When create record with Api/V8 it validate attributes with regex pattern **`const REGEX_FIELD_PATTERN = '/[^\w-,]/'`**, so if you use unicode or space on value will occur error **"The option attributes with value array is invalid."** in the response. | 1.0 | Api/V8 create record does not support unicode and space in attributes - When create record with Api/V8 it validate attributes with regex pattern **`const REGEX_FIELD_PATTERN = '/[^\w-,]/'`**, so if you use unicode or space on value will occur error **"The option attributes with value array is invalid."** in the response. | priority | api create record does not support unicode and space in attributes when create record with api it validate attributes with regex pattern const regex field pattern so if you use unicode or space on value will occur error the option attributes with value array is invalid in the response | 1 |
417,680 | 12,168,405,368 | IssuesEvent | 2020-04-27 12:38:19 | canonical-web-and-design/snapcraft.io | https://api.github.com/repos/canonical-web-and-design/snapcraft.io | closed | Unclear how to log in to build.snapcraft.io | Priority: Medium | ### Expected behaviour
I would like a clear indicator of how to login.
I only visit the site occasionally when I need to manually kick off a snap build (normally prompted by a security warning email). When I look at the page, there are two options that present themselves to me "Set up in minutes :octocat:" (which I disregard, because I'm already set up) and "Developer account" (which I click on because for some reason I think that I want to log in to my snapcraft developer account to do snapcraft things as a developer). Then I'm presented with a list of snaps, this is going great, but when I click through to the snap in question, and keep clicking on things, I never find my way to something that lets me build my snaps. This is frustrating!
(Of course, I eventually work out that :octocat: is code for "click this to just login with Github as well" and get on with things, but this was a couple of frustrating minutes that I think would put off plenty of users!)
### Steps to reproduce the problem
1) Navigate to build.snapcraft.io
| 1.0 | Unclear how to log in to build.snapcraft.io - ### Expected behaviour
I would like a clear indicator of how to login.
I only visit the site occasionally when I need to manually kick off a snap build (normally prompted by a security warning email). When I look at the page, there are two options that present themselves to me "Set up in minutes :octocat:" (which I disregard, because I'm already set up) and "Developer account" (which I click on because for some reason I think that I want to log in to my snapcraft developer account to do snapcraft things as a developer). Then I'm presented with a list of snaps, this is going great, but when I click through to the snap in question, and keep clicking on things, I never find my way to something that lets me build my snaps. This is frustrating!
(Of course, I eventually work out that :octocat: is code for "click this to just login with Github as well" and get on with things, but this was a couple of frustrating minutes that I think would put off plenty of users!)
### Steps to reproduce the problem
1) Navigate to build.snapcraft.io
| priority | unclear how to log in to build snapcraft io expected behaviour i would like a clear indicator of how to login i only visit the site occasionally when i need to manually kick off a snap build normally prompted by a security warning email when i look at the page there are two options that present themselves to me set up in minutes octocat which i disregard because i m already set up and developer account which i click on because for some reason i think that i want to log in to my snapcraft developer account to do snapcraft things as a developer then i m presented with a list of snaps this is going great but when i click through to the snap in question and keep clicking on things i never find my way to something that lets me build my snaps this is frustrating of course i eventually work out that octocat is code for click this to just login with github as well and get on with things but this was a couple of frustrating minutes that i think would put off plenty of users steps to reproduce the problem navigate to build snapcraft io | 1 |
170,966 | 6,475,439,364 | IssuesEvent | 2017-08-17 20:24:14 | systers/conference-android | https://api.github.com/repos/systers/conference-android | closed | User can bookmark events to create customized schedule | Priority: MEDIUM Program: GSoC17 | **Description**
As an attendee,
I want to bookmark the events I want to attend so that a customized schedule can be created.
**Mocks**
**Acceptance Criteria**
- User can bookmark an event.
- User can remove event from bookmark.
**Definition of Done**
- Passes all regression tests.
- Passes testing per acceptance criteria items.
- Approved by UI team.
- Able to show feature in demo.
**Estimation**
- 3-4 Hours | 1.0 | User can bookmark events to create customized schedule - **Description**
As an attendee,
I want to bookmark the events I want to attend so that a customized schedule can be created.
**Mocks**
**Acceptance Criteria**
- User can bookmark an event.
- User can remove event from bookmark.
**Definition of Done**
- Passes all regression tests.
- Passes testing per acceptance criteria items.
- Approved by UI team.
- Able to show feature in demo.
**Estimation**
- 3-4 Hours | priority | user can bookmark events to create customized schedule description as an attendee i want to bookmark the events i want to attend so that a customized schedule can be created mocks acceptance criteria user can bookmark an event user can remove event from bookmark definition of done passes all regression tests passes testing per acceptance criteria items approved by ui team able to show feature in demo estimation hours | 1 |
620,489 | 19,563,402,607 | IssuesEvent | 2022-01-03 19:38:24 | bounswe/2021SpringGroup1 | https://api.github.com/repos/bounswe/2021SpringGroup1 | closed | Converting all Backend functions to Json-LD format | Type: Enhancement Priority: Medium Platform: Backend | Previously, necessary changes were made to return it in Json-LD format for Community functions. Now this format should be added to other functions as well.
Here, relatively static fields such as "@context", "@type" should be added first. If there is no problem here, necessary changes should be made on the database and the field names in the Models. | 1.0 | Converting all Backend functions to Json-LD format - Previously, necessary changes were made to return it in Json-LD format for Community functions. Now this format should be added to other functions as well.
Here, relatively static fields such as "@context", "@type" should be added first. If there is no problem here, necessary changes should be made on the database and the field names in the Models. | priority | converting all backend functions to json ld format previously necessary changes were made to return it in json ld format for community functions now this format should be added to other functions as well here relatively static fields such as context type should be added first if there is no problem here necessary changes should be made on the database and the field names in the models | 1 |
42,083 | 2,869,096,672 | IssuesEvent | 2015-06-05 23:18:30 | dart-lang/test | https://api.github.com/repos/dart-lang/test | closed | unittest: html_enhanced_config is using Element.elements -> fails with SDK 0.3.4.0 | bug Fixed Priority-Medium | <a href="https://github.com/kevmoo"><img src="https://avatars.githubusercontent.com/u/17034?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [kevmoo](https://github.com/kevmoo)**
_Originally opened as dart-lang/sdk#8335_
----
**What steps will reproduce the problem?**
1. Get r18115
2. Try to use html_enhanced_config from unittest
**What is the expected output? What do you see instead?**
Expected: It works.
Instead: noSuchMethod 'elements'
**What version of the product are you using? On what operating system?**
0.3.4.0 r18115
Mac 10.8.2
**Please provide any additional information below.**
Perhaps run analyzer over these?
| 1.0 | unittest: html_enhanced_config is using Element.elements -> fails with SDK 0.3.4.0 - <a href="https://github.com/kevmoo"><img src="https://avatars.githubusercontent.com/u/17034?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [kevmoo](https://github.com/kevmoo)**
_Originally opened as dart-lang/sdk#8335_
----
**What steps will reproduce the problem?**
1. Get r18115
2. Try to use html_enhanced_config from unittest
**What is the expected output? What do you see instead?**
Expected: It works.
Instead: noSuchMethod 'elements'
**What version of the product are you using? On what operating system?**
0.3.4.0 r18115
Mac 10.8.2
**Please provide any additional information below.**
Perhaps run analyzer over these?
| priority | unittest html enhanced config is using element elements fails with sdk issue by originally opened as dart lang sdk what steps will reproduce the problem get try to use html enhanced config from unittest what is the expected output what do you see instead expected it works instead nosuchmethod elements what version of the product are you using on what operating system mac please provide any additional information below perhaps run analyzer over these | 1 |
544,181 | 15,890,621,276 | IssuesEvent | 2021-04-10 16:05:33 | davidfstr/Crystal-Web-Archiver | https://api.github.com/repos/davidfstr/Crystal-Web-Archiver | opened | Fix _OrderedTreeCtrl warnings upon ABNORMAL quit (on at least Mac) | priority-medium type-bug | Priority: Medium
* Likely† **prevents reopening Crystal after it is quit**, if did quit while in the middle of downloading.
* Has workaround: Right-click Crystal icon in Dock, hold Option, and select Force Quit.
* † Need to reverify
Repro Steps:
* Create project to download xkcd.
* Define and start downloading the "xkcd Comic" group.
* After ~5 members of the group have downloaded, press the close button on Crystal's main window.
Expected Results:
* The window does close.
* The app icon does disappear, indicating the app has actually quit.
Actual Results:
* The window does close.
* The following error is printed to the console:
* `RuntimeError: wrapped C/C++ object of type _OrderedTreeCtrl has been deleted`
* The app icon remains in the Dock, indicating the app has not actually quit.
* Probably this will prevent attempts to reopen Crystal later. 😬
Although this has only been observed on macOS so far, I **suspect it is a platform-agnostic issue**. (Verification required.)
Related Issues:
* A similar issue happens on Windows on EVERY quit, even normal quits: https://github.com/davidfstr/Crystal-Web-Archiver/issues/14
<img width="1829" alt="Screen Shot 2021-04-10 at 7 27 00 AM" src="https://user-images.githubusercontent.com/764688/114276278-cf040700-99da-11eb-9d8d-13b735872983.png">
| 1.0 | Fix _OrderedTreeCtrl warnings upon ABNORMAL quit (on at least Mac) - Priority: Medium
* Likely† **prevents reopening Crystal after it is quit**, if did quit while in the middle of downloading.
* Has workaround: Right-click Crystal icon in Dock, hold Option, and select Force Quit.
* † Need to reverify
Repro Steps:
* Create project to download xkcd.
* Define and start downloading the "xkcd Comic" group.
* After ~5 members of the group have downloaded, press the close button on Crystal's main window.
Expected Results:
* The window does close.
* The app icon does disappear, indicating the app has actually quit.
Actual Results:
* The window does close.
* The following error is printed to the console:
* `RuntimeError: wrapped C/C++ object of type _OrderedTreeCtrl has been deleted`
* The app icon remains in the Dock, indicating the app has not actually quit.
* Probably this will prevent attempts to reopen Crystal later. 😬
Although this has only been observed on macOS so far, I **suspect it is a platform-agnostic issue**. (Verification required.)
Related Issues:
* A similar issue happens on Windows on EVERY quit, even normal quits: https://github.com/davidfstr/Crystal-Web-Archiver/issues/14
<img width="1829" alt="Screen Shot 2021-04-10 at 7 27 00 AM" src="https://user-images.githubusercontent.com/764688/114276278-cf040700-99da-11eb-9d8d-13b735872983.png">
| priority | fix orderedtreectrl warnings upon abnormal quit on at least mac priority medium likely† prevents reopening crystal after it is quit if did quit while in the middle of downloading has workaround right click crystal icon in dock hold option and select force quit † need to reverify repro steps create project to download xkcd define and start downloading the xkcd comic group after members of the group have downloaded press the close button on crystal s main window expected results the window does close the app icon does disappear indicating the app has actually quit actual results the window does close the following error is printed to the console runtimeerror wrapped c c object of type orderedtreectrl has been deleted the app icon remains in the dock indicating the app has not actually quit probably this will prevent attempts to reopen crystal later 😬 although this has only been observed on macos so far i suspect it is a platform agnostic issue verification required related issues a similar issue happens on windows on every quit even normal quits img width alt screen shot at am src | 1 |
22,209 | 2,645,776,340 | IssuesEvent | 2015-03-13 02:09:15 | prikhi/evoluspencil | https://api.github.com/repos/prikhi/evoluspencil | opened | Position/Dimensions toolbar disabled in 1.0.6 | 2–5 stars bug imported Priority-Medium | _From [benny.mk...@gmail.com](https://code.google.com/u/102854032411103376457/) on September 02, 2009 23:42:37_
What steps will reproduce the problem? 1. Select any rectangular object on a page.
2. The position/dimensions toolbar should show the X Y coordinates and
width and height, but the fields are disabled. What is the expected output? What do you see instead? I should be able to change the position and size of an object by updating
these fields. This used to work. What version of the product are you using? On what operating system? 1.0.6 on Windows XP SP2
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=101_ | 1.0 | Position/Dimensions toolbar disabled in 1.0.6 - _From [benny.mk...@gmail.com](https://code.google.com/u/102854032411103376457/) on September 02, 2009 23:42:37_
What steps will reproduce the problem? 1. Select any rectangular object on a page.
2. The position/dimensions toolbar should show the X Y coordinates and
width and height, but the fields are disabled. What is the expected output? What do you see instead? I should be able to change the position and size of an object by updating
these fields. This used to work. What version of the product are you using? On what operating system? 1.0.6 on Windows XP SP2
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=101_ | priority | position dimensions toolbar disabled in from on september what steps will reproduce the problem select any rectangular object on a page the position dimensions toolbar should show the x y coordinates and width and height but the fields are disabled what is the expected output what do you see instead i should be able to change the position and size of an object by updating these fields this used to work what version of the product are you using on what operating system on windows xp original issue | 1 |
687,992 | 23,544,316,688 | IssuesEvent | 2022-08-20 22:16:46 | awslabs/aws-saas-boost | https://api.github.com/repos/awslabs/aws-saas-boost | closed | FSx ONTAP configurations fail to launch Containers | bug priority-medium tenant-onboarding | The user data boot script is failing with `New-SmbGlobalMapping : The network name cannot be found.` and task definitions with containers expecting a mount point to be available fail to launch.
### Reproduction Steps
Configure an application with Windows, shared file system type of FSx ONTAP, build and push the Windows sample app and onboard a tenant.
### What did you expect to happen?
Windows sample app to be deployed and functional for the onboarded tenant
### What actually happened?
503 error on the tenant subdomain
CodePipeline for the tenant fails or never finishes the deployment step with ECS
ECS task deployment shows Stopped|Cannot Create Container
### Environment
- **AWS Region :** Any
- **AWS SaaS Boost Version :** HEAD
- **Workload OS (Linux or Windows) :** Windows
### Other
In the user data script for Windows we if/else on the New-SmbGlobalMapping command in Powershell to deal with the different syntax between ONTAP and Windows File Server shares. The if/else statement is testing for the string `FSXO` but we're using the string `FSX_ONTAP` now for FSx for NetApp ONTAP file systems.
---
This is :bug: Bug Report
| 1.0 | FSx ONTAP configurations fail to launch Containers - The user data boot script is failing with `New-SmbGlobalMapping : The network name cannot be found.` and task definitions with containers expecting a mount point to be available fail to launch.
### Reproduction Steps
Configure an application with Windows, shared file system type of FSx ONTAP, build and push the Windows sample app and onboard a tenant.
### What did you expect to happen?
Windows sample app to be deployed and functional for the onboarded tenant
### What actually happened?
503 error on the tenant subdomain
CodePipeline for the tenant fails or never finishes the deployment step with ECS
ECS task deployment shows Stopped|Cannot Create Container
### Environment
- **AWS Region :** Any
- **AWS SaaS Boost Version :** HEAD
- **Workload OS (Linux or Windows) :** Windows
### Other
In the user data script for Windows we if/else on the New-SmbGlobalMapping command in Powershell to deal with the different syntax between ONTAP and Windows File Server shares. The if/else statement is testing for the string `FSXO` but we're using the string `FSX_ONTAP` now for FSx for NetApp ONTAP file systems.
---
This is :bug: Bug Report
| priority | fsx ontap configurations fail to launch containers the user data boot script is failing with new smbglobalmapping the network name cannot be found and task definitions with containers expecting a mount point to be available fail to launch reproduction steps configure an application with windows shared file system type of fsx ontap build and push the windows sample app and onboard a tenant what did you expect to happen windows sample app to be deployed and functional for the onboarded tenant what actually happened error on the tenant subdomain codepipeline for the tenant fails or never finishes the deployment step with ecs ecs task deployment shows stopped cannot create container environment aws region any aws saas boost version head workload os linux or windows windows other in the user data script for windows we if else on the new smbglobalmapping command in powershell to deal with the different syntax between ontap and windows file server shares the if else statement is testing for the string fsxo but we re using the string fsx ontap now for fsx for netapp ontap file systems this is bug bug report | 1 |
362,780 | 10,731,916,109 | IssuesEvent | 2019-10-28 20:37:44 | carbon-design-system/ibm-dotcom-library | https://api.github.com/repos/carbon-design-system/ibm-dotcom-library | closed | Cupcake disclaimer messaging | content priority: medium sprint demo website: cupcake | _jamespistilli created the following on Oct 09:_
<!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
As a `web visitor of the forthcoming Cupcake website`:
I need to:
Be alerted at the very beginning of the page that the contents below are part of a controlled adoption and subject to change, and should be treated as so.
so that I can:
Be aware that the contents of the page is not finalized.
#### Additional information
- [ ] Jim to provide a few options for disclaimer messaging, approved by @chsanche
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria – when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1823_ | 1.0 | Cupcake disclaimer messaging - _jamespistilli created the following on Oct 09:_
<!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
As a `web visitor of the forthcoming Cupcake website`:
I need to:
Be alerted at the very beginning of the page that the contents below are part of a controlled adoption and subject to change, and should be treated as so.
so that I can:
Be aware that the contents of the page is not finalized.
#### Additional information
- [ ] Jim to provide a few options for disclaimer messaging, approved by @chsanche
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria – when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1823_ | priority | cupcake disclaimer messaging jamespistilli created the following on oct user story as a web visitor of the forthcoming cupcake website i need to be alerted at the very beginning of the page that the contents below are part of a controlled adoption and subject to change and should be treated as so so that i can be aware that the contents of the page is not finalized additional information jim to provide a few options for disclaimer messaging approved by chsanche original issue | 1 |
550,858 | 16,133,630,468 | IssuesEvent | 2021-04-29 08:56:58 | kubesphere/console | https://api.github.com/repos/kubesphere/console | closed | the Create button needs to be removed in System Project page | kind/bug kind/need-to-verify priority/medium |
**Describe the bug**
Enter cluster management-- >project, in the System Project Tab page, the Create button needs to be removed

/kind bug
/assign @leoendless
/milestone 3.1.0
/priority medium | 1.0 | the Create button needs to be removed in System Project page -
**Describe the bug**
Enter cluster management-- >project, in the System Project Tab page, the Create button needs to be removed

/kind bug
/assign @leoendless
/milestone 3.1.0
/priority medium | priority | the create button needs to be removed in system project page describe the bug enter cluster management project in the system project tab page the create button needs to be removed kind bug assign leoendless milestone priority medium | 1 |
108,834 | 4,351,399,946 | IssuesEvent | 2016-07-31 20:49:55 | fog/fog-google | https://api.github.com/repos/fog/fog-google | closed | Storage json api doesn't support directories.all | enhancement priority/medium ready | Looks like it was left unimplemented.
```ruby
irb(main):008:0* Fog::Storage::Google.new.directories.all
NoMethodError: undefined method `all' for #<Fog::Storage::GoogleJSON::Directories:0x007ffa10a80950>
from (irb):8
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli/console.rb:15:in `run'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli.rb:333:in `console'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/command.rb:27:in `run'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/invocation.rb:126:in `invoke_command'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor.rb:359:in `dispatch'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/base.rb:440:in `start'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli.rb:11:in `start'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/exe/bundle:27:in `block in <top (required)>'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/friendly_errors.rb:98:in `with_friendly_errors'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/exe/bundle:19:in `<top (required)>'
from /home/jsselman/.rbenv/versions/2.2.2/bin/bundle:23:in `load'
from /home/jsselman/.rbenv/versions/2.2.2/bin/bundle:23:in `<main>'
``` | 1.0 | Storage json api doesn't support directories.all - Looks like it was left unimplemented.
```ruby
irb(main):008:0* Fog::Storage::Google.new.directories.all
NoMethodError: undefined method `all' for #<Fog::Storage::GoogleJSON::Directories:0x007ffa10a80950>
from (irb):8
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli/console.rb:15:in `run'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli.rb:333:in `console'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/command.rb:27:in `run'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/invocation.rb:126:in `invoke_command'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor.rb:359:in `dispatch'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/vendor/thor/lib/thor/base.rb:440:in `start'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/cli.rb:11:in `start'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/exe/bundle:27:in `block in <top (required)>'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/lib/bundler/friendly_errors.rb:98:in `with_friendly_errors'
from /home/jsselman/.rbenv/versions/2.2.2/lib/ruby/gems/2.2.0/gems/bundler-1.12.5/exe/bundle:19:in `<top (required)>'
from /home/jsselman/.rbenv/versions/2.2.2/bin/bundle:23:in `load'
from /home/jsselman/.rbenv/versions/2.2.2/bin/bundle:23:in `<main>'
``` | priority | storage json api doesn t support directories all looks like it was left unimplemented ruby irb main fog storage google new directories all nomethoderror undefined method all for from irb from home jsselman rbenv versions lib ruby gems gems bundler lib bundler cli console rb in run from home jsselman rbenv versions lib ruby gems gems bundler lib bundler cli rb in console from home jsselman rbenv versions lib ruby gems gems bundler lib bundler vendor thor lib thor command rb in run from home jsselman rbenv versions lib ruby gems gems bundler lib bundler vendor thor lib thor invocation rb in invoke command from home jsselman rbenv versions lib ruby gems gems bundler lib bundler vendor thor lib thor rb in dispatch from home jsselman rbenv versions lib ruby gems gems bundler lib bundler vendor thor lib thor base rb in start from home jsselman rbenv versions lib ruby gems gems bundler lib bundler cli rb in start from home jsselman rbenv versions lib ruby gems gems bundler exe bundle in block in from home jsselman rbenv versions lib ruby gems gems bundler lib bundler friendly errors rb in with friendly errors from home jsselman rbenv versions lib ruby gems gems bundler exe bundle in from home jsselman rbenv versions bin bundle in load from home jsselman rbenv versions bin bundle in | 1 |
607,062 | 18,772,673,882 | IssuesEvent | 2021-11-07 04:56:51 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Inventory UI bug <3 | Priority: 3-Not Required Changes: UI Difficulty: 2-Medium | Item pile pop-ups are overlayed on top of inventory UI, clicking on an item from a pile will go through to the bottom layer and hit the items in your inventory.
Attempting to pick something up causes you to accidently unequip items. | 1.0 | Inventory UI bug <3 - Item pile pop-ups are overlayed on top of inventory UI, clicking on an item from a pile will go through to the bottom layer and hit the items in your inventory.
Attempting to pick something up causes you to accidently unequip items. | priority | inventory ui bug item pile pop ups are overlayed on top of inventory ui clicking on an item from a pile will go through to the bottom layer and hit the items in your inventory attempting to pick something up causes you to accidently unequip items | 1 |
310,544 | 9,515,518,894 | IssuesEvent | 2019-04-26 05:59:09 | ucb-bar/hammer | https://api.github.com/repos/ucb-bar/hammer | opened | Pin placement API with width and position control | API enhancement medium priority | "Low gear" API for fine-grained control and incremental adoption.
The plan is to add the following API to the pin placement API as follows: `location_mode: str`, `location_by_tracks` (unimplemented), `location_by_coords` | 1.0 | Pin placement API with width and position control - "Low gear" API for fine-grained control and incremental adoption.
The plan is to add the following API to the pin placement API as follows: `location_mode: str`, `location_by_tracks` (unimplemented), `location_by_coords` | priority | pin placement api with width and position control low gear api for fine grained control and incremental adoption the plan is to add the following api to the pin placement api as follows location mode str location by tracks unimplemented location by coords | 1 |
331,373 | 10,065,172,349 | IssuesEvent | 2019-07-23 10:15:08 | garden-io/garden | https://api.github.com/repos/garden-io/garden | opened | .gardenignore doesn't work with unstaged files/dirs | bug priority:medium | ## Bug
### Current Behavior
Adding an entry to `.gardenignore` for a file or directory that hasn't been staged with Git won't work and Garden will continue to watch and build the content.
### Expected behavior
It should work for un-staged content.
### Reproducible example
1. Go to `demo-project`
2. Create `test.txt` and `.gardenignore` files in `frontend` dir
3. Add `test.txt` to `.gardenignore`
4. Run a Garden command
5. Notice that `test.txt` is copied to `.garden/build/frontend`
6. `git add` the file and notice that it now works.
### Workaround
Stage the file you want to ignore. | 1.0 | .gardenignore doesn't work with unstaged files/dirs - ## Bug
### Current Behavior
Adding an entry to `.gardenignore` for a file or directory that hasn't been staged with Git won't work and Garden will continue to watch and build the content.
### Expected behavior
It should work for un-staged content.
### Reproducible example
1. Go to `demo-project`
2. Create `test.txt` and `.gardenignore` files in `frontend` dir
3. Add `test.txt` to `.gardenignore`
4. Run a Garden command
5. Notice that `test.txt` is copied to `.garden/build/frontend`
6. `git add` the file and notice that it now works.
### Workaround
Stage the file you want to ignore. | priority | gardenignore doesn t work with unstaged files dirs bug current behavior adding an entry to gardenignore for a file or directory that hasn t been staged with git won t work and garden will continue to watch and build the content expected behavior it should work for un staged content reproducible example go to demo project create test txt and gardenignore files in frontend dir add test txt to gardenignore run a garden command notice that test txt is copied to garden build frontend git add the file and notice that it now works workaround stage the file you want to ignore | 1 |
109,069 | 4,369,709,541 | IssuesEvent | 2016-08-04 01:33:28 | williewillus/Botania | https://api.github.com/repos/williewillus/Botania | opened | End to end interaction testing | priority-medium | Wands switch mode and act on the block, same for black hole talisman, etc.
Clean all of them up | 1.0 | End to end interaction testing - Wands switch mode and act on the block, same for black hole talisman, etc.
Clean all of them up | priority | end to end interaction testing wands switch mode and act on the block same for black hole talisman etc clean all of them up | 1 |
420,189 | 12,234,436,174 | IssuesEvent | 2020-05-04 13:25:38 | threefoldtech/3bot_wallet | https://api.github.com/repos/threefoldtech/3bot_wallet | opened | Stellar Staging - Dragging to re-order wallets does not always work. | priority_medium type_bug | **Repro steps**
1) Drag to re-order wallets
2) Repeat this several times
**Expected Result**
Works every time
**Actual Result**
The drag often doesn't register, I'd say in about 60% of the cases
**System Info**
| 1.0 | Stellar Staging - Dragging to re-order wallets does not always work. - **Repro steps**
1) Drag to re-order wallets
2) Repeat this several times
**Expected Result**
Works every time
**Actual Result**
The drag often doesn't register, I'd say in about 60% of the cases
**System Info**
| priority | stellar staging dragging to re order wallets does not always work repro steps drag to re order wallets repeat this several times expected result works every time actual result the drag often doesn t register i d say in about of the cases system info | 1 |
49,299 | 3,001,918,144 | IssuesEvent | 2015-07-24 14:26:47 | jayway/powermock | https://api.github.com/repos/jayway/powermock | opened | Better error messages when I forget to PrepareForTest | bug imported Priority-Medium | _From [dan.fabu...@redfin.com](https://code.google.com/u/112087859980982342548/) on May 19, 2011 03:43:26_
Run the tests in the attached Maven project.
The test runs this code:
PowerMockito.doThrow(new RuntimeException()).when(app).finalMethod();
But it "forgets" to prepare the class for test.
Actual: Exception not thrown; the test fails and it's not clear why.
Expected: When I try to mock out a final method, but the class isn't prepared for test, I wish PowerMock would give me an error message reminding me to prepare. (Or, even better, just PrepareForTest on the fly!)
(I lost an hour on this today. :-p)
**Attachment:** [silentfailure.zip](http://code.google.com/p/powermock/issues/detail?id=330)
_Original issue: http://code.google.com/p/powermock/issues/detail?id=330_ | 1.0 | Better error messages when I forget to PrepareForTest - _From [dan.fabu...@redfin.com](https://code.google.com/u/112087859980982342548/) on May 19, 2011 03:43:26_
Run the tests in the attached Maven project.
The test runs this code:
PowerMockito.doThrow(new RuntimeException()).when(app).finalMethod();
But it "forgets" to prepare the class for test.
Actual: Exception not thrown; the test fails and it's not clear why.
Expected: When I try to mock out a final method, but the class isn't prepared for test, I wish PowerMock would give me an error message reminding me to prepare. (Or, even better, just PrepareForTest on the fly!)
(I lost an hour on this today. :-p)
**Attachment:** [silentfailure.zip](http://code.google.com/p/powermock/issues/detail?id=330)
_Original issue: http://code.google.com/p/powermock/issues/detail?id=330_ | priority | better error messages when i forget to preparefortest from on may run the tests in the attached maven project the test runs this code powermockito dothrow new runtimeexception when app finalmethod but it forgets to prepare the class for test actual exception not thrown the test fails and it s not clear why expected when i try to mock out a final method but the class isn t prepared for test i wish powermock would give me an error message reminding me to prepare or even better just preparefortest on the fly i lost an hour on this today p attachment original issue | 1 |
712,592 | 24,500,355,659 | IssuesEvent | 2022-10-10 12:18:46 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [DocDB] Healthcheck errors and Soft memory limit exceeded error at log | kind/bug priority/medium area/cdcsdk status/awaiting-triage | Jira Link: [DB-1862](https://yugabyte.atlassian.net/browse/DB-1862)
### Description
Version: 2.13.2.0-b68
Healtcheck error
`Error executing command ['ssh', 'yugabyte@172.151.31.129', '-p', '54422', '-o', 'StrictHostKeyChecking no', '-o', 'ConnectTimeout=10', '-o', 'UserKnownHostsFile /dev/null', '-o', 'LogLevel ERROR', '-i', '/opt/yugabyte/yugaware/data/keys/08c0ba0e-3558-40fc-94e5-a87a627de8c5/yb-15-aws-portal-1-key.pem', 'set -o pipefail; /home/yugabyte/tserver/bin/cqlsh 172.151.31.129 9042 -e "SHOW HOST"']: b'Connection error: (\'Unable to connect to any servers\', {\'172.151.31.129\': <Error from server: code=1001 [Coordinator node overloaded] message="">})\n'
`
and
```
Rejecting CQL call: Soft memory limit exceeded (at 88.08% of capacity), score: 0.98 [suppressed 135 similar messages]
```
in tserver log file
Steps:
Create 2 universes(source and target)
Run workload at source
wait for some time(5-10 min)
Run bootstrap cdc cmd
take backup of tables and restore at target universe
Then I Observed these errors.
Source Universe:
http://portal.dev.yugabyte.com/universes/2cd656a8-981f-4130-8b04-6d2642a934ef/nodes | 1.0 | [DocDB] Healthcheck errors and Soft memory limit exceeded error at log - Jira Link: [DB-1862](https://yugabyte.atlassian.net/browse/DB-1862)
### Description
Version: 2.13.2.0-b68
Healtcheck error
`Error executing command ['ssh', 'yugabyte@172.151.31.129', '-p', '54422', '-o', 'StrictHostKeyChecking no', '-o', 'ConnectTimeout=10', '-o', 'UserKnownHostsFile /dev/null', '-o', 'LogLevel ERROR', '-i', '/opt/yugabyte/yugaware/data/keys/08c0ba0e-3558-40fc-94e5-a87a627de8c5/yb-15-aws-portal-1-key.pem', 'set -o pipefail; /home/yugabyte/tserver/bin/cqlsh 172.151.31.129 9042 -e "SHOW HOST"']: b'Connection error: (\'Unable to connect to any servers\', {\'172.151.31.129\': <Error from server: code=1001 [Coordinator node overloaded] message="">})\n'
`
and
```
Rejecting CQL call: Soft memory limit exceeded (at 88.08% of capacity), score: 0.98 [suppressed 135 similar messages]
```
in tserver log file
Steps:
Create 2 universes(source and target)
Run workload at source
wait for some time(5-10 min)
Run bootstrap cdc cmd
take backup of tables and restore at target universe
Then I Observed these errors.
Source Universe:
http://portal.dev.yugabyte.com/universes/2cd656a8-981f-4130-8b04-6d2642a934ef/nodes | priority | healthcheck errors and soft memory limit exceeded error at log jira link description version healtcheck error error executing command b connection error unable to connect to any servers n and rejecting cql call soft memory limit exceeded at of capacity score in tserver log file steps create universes source and target run workload at source wait for some time min run bootstrap cdc cmd take backup of tables and restore at target universe then i observed these errors source universe | 1 |
245,095 | 7,881,082,269 | IssuesEvent | 2018-06-26 17:52:34 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: Unowned Cart | Medium Priority | **Version:** 0.7.3.2 beta
**Steps to Reproduce:**
After the most recent patch my server rolled back roughly 20 minutes. When logging back online my cart was listed as "unowned" and the deed says the same. I can still move and use the cart but I cannot pick it up with my hammer.
**Expected behavior:**
When hitting my cart with a hammer while having the deed in my inventory I expect to be able to pick up my cart
**Actual behavior:**
After hitting my cart with my hammer it tells me I must have the deed in order to pick it up | 1.0 | USER ISSUE: Unowned Cart - **Version:** 0.7.3.2 beta
**Steps to Reproduce:**
After the most recent patch my server rolled back roughly 20 minutes. When logging back online my cart was listed as "unowned" and the deed says the same. I can still move and use the cart but I cannot pick it up with my hammer.
**Expected behavior:**
When hitting my cart with a hammer while having the deed in my inventory I expect to be able to pick up my cart
**Actual behavior:**
After hitting my cart with my hammer it tells me I must have the deed in order to pick it up | priority | user issue unowned cart version beta steps to reproduce after the most recent patch my server rolled back roughly minutes when logging back online my cart was listed as unowned and the deed says the same i can still move and use the cart but i cannot pick it up with my hammer expected behavior when hitting my cart with a hammer while having the deed in my inventory i expect to be able to pick up my cart actual behavior after hitting my cart with my hammer it tells me i must have the deed in order to pick it up | 1 |
199,355 | 6,988,678,861 | IssuesEvent | 2017-12-14 13:50:08 | nlbdev/pipeline | https://api.github.com/repos/nlbdev/pipeline | closed | Letters become numers | bug Priority:2 - Medium | @josteinaj @usama49 @KariRudjord ⠀⠀⠀⠀⠀⠀
[job#700.zip](https://github.com/nlbdev/pipeline/files/1222071/job.700.zip)
The second sentence in this novel comes out all wrong; The 'J' in 'Jeg vet ikke...' is translated to number 6 (⠼⠋) !
Also, elements from the DTBook frontmatter, is represented twice @josteinaj @usama49 @josteinaj
⠀⠀
```xml
<level2 id="level2_1" class="chapter">
<p>Jeg stoler ikke på noen.</p>
<p>Jeg vet ikke, det er vel en slags skade.</p>
<p>Nei, ikke på deg heller. (...)</p>
(...)
</level2>
```
```
⠀⠀⠀⠀⠠⠚⠑⠛⠀⠎⠞⠕⠇⠑⠗⠀⠊⠅⠅⠑⠀⠏⠡⠀⠝⠕⠑⠝⠄⠀⠀⠀
⠀⠀⠀⠀⠼⠋⠰⠑⠛⠀⠧⠑⠞⠀⠊⠅⠅⠑⠂⠀⠙⠑⠞⠀⠑⠗⠀⠧⠑⠇⠀⠀
⠀⠀⠑⠝⠀⠎⠇⠁⠛⠎⠀⠎⠅⠁⠙⠑⠄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀
⠀⠀⠀⠀⠠⠝⠑⠊⠂⠀⠊⠅⠅⠑⠀⠏⠡⠀⠙⠑⠛⠀⠓⠑⠇⠇⠑⠗⠄⠀
```
Weird braille also in generated text, again letters turned into numbers.
--
012llskrift
012tall _sider: 0339
05oka skal ikke returneres'
_feil eller mangler kan
meldes til 2punkt^nlb'no'[
| 1.0 | Letters become numers - @josteinaj @usama49 @KariRudjord ⠀⠀⠀⠀⠀⠀
[job#700.zip](https://github.com/nlbdev/pipeline/files/1222071/job.700.zip)
The second sentence in this novel comes out all wrong; The 'J' in 'Jeg vet ikke...' is translated to number 6 (⠼⠋) !
Also, elements from the DTBook frontmatter, is represented twice @josteinaj @usama49 @josteinaj
⠀⠀
```xml
<level2 id="level2_1" class="chapter">
<p>Jeg stoler ikke på noen.</p>
<p>Jeg vet ikke, det er vel en slags skade.</p>
<p>Nei, ikke på deg heller. (...)</p>
(...)
</level2>
```
```
⠀⠀⠀⠀⠠⠚⠑⠛⠀⠎⠞⠕⠇⠑⠗⠀⠊⠅⠅⠑⠀⠏⠡⠀⠝⠕⠑⠝⠄⠀⠀⠀
⠀⠀⠀⠀⠼⠋⠰⠑⠛⠀⠧⠑⠞⠀⠊⠅⠅⠑⠂⠀⠙⠑⠞⠀⠑⠗⠀⠧⠑⠇⠀⠀
⠀⠀⠑⠝⠀⠎⠇⠁⠛⠎⠀⠎⠅⠁⠙⠑⠄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀
⠀⠀⠀⠀⠠⠝⠑⠊⠂⠀⠊⠅⠅⠑⠀⠏⠡⠀⠙⠑⠛⠀⠓⠑⠇⠇⠑⠗⠄⠀
```
Weird braille also in generated text, again letters turned into numbers.
--
012llskrift
012tall _sider: 0339
05oka skal ikke returneres'
_feil eller mangler kan
meldes til 2punkt^nlb'no'[
| priority | letters become numers josteinaj karirudjord ⠀⠀⠀⠀⠀⠀ the second sentence in this novel comes out all wrong the j in jeg vet ikke is translated to number ⠼⠋ also elements from the dtbook frontmatter is represented twice josteinaj josteinaj ⠀⠀ xml jeg stoler ikke p noen jeg vet ikke det er vel en slags skade nei ikke p deg heller ⠀⠀⠀⠀⠠⠚⠑⠛⠀⠎⠞⠕⠇⠑⠗⠀⠊⠅⠅⠑⠀⠏⠡⠀⠝⠕⠑⠝⠄⠀⠀⠀ ⠀⠀⠀⠀⠼⠋⠰⠑⠛⠀⠧⠑⠞⠀⠊⠅⠅⠑⠂⠀⠙⠑⠞⠀⠑⠗⠀⠧⠑⠇⠀⠀ ⠀⠀⠑⠝⠀⠎⠇⠁⠛⠎⠀⠎⠅⠁⠙⠑⠄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀ ⠀⠀⠀⠀⠠⠝⠑⠊⠂⠀⠊⠅⠅⠑⠀⠏⠡⠀⠙⠑⠛⠀⠓⠑⠇⠇⠑⠗⠄⠀ weird braille also in generated text again letters turned into numbers sider skal ikke returneres feil eller mangler kan meldes til nlb no | 1 |
774,815 | 27,212,362,438 | IssuesEvent | 2023-02-20 17:40:11 | calcom/cal.com | https://api.github.com/repos/calcom/cal.com | closed | [CAL-1121] reminder emails being sent even after cancellation | 🐛 bug linear Medium priority | Found a bug? Please fill out the sections below. 👍
### Issue Summary
reminder emails being sent even after cancellation
### Steps to Reproduce
1. set up a workflow with emails
2. appoint a meeting with email workflow
3. cancel the meeting
4. reminder emails still being received
Any other relevant information. For example, why do you consider this a bug and what did you expect to happen instead? expected to not receive any reminder
<sub>[CAL-1121](https://linear.app/calcom/issue/CAL-1121/reminder-emails-being-sent-even-after-cancellation)</sub> | 1.0 | [CAL-1121] reminder emails being sent even after cancellation - Found a bug? Please fill out the sections below. 👍
### Issue Summary
reminder emails being sent even after cancellation
### Steps to Reproduce
1. set up a workflow with emails
2. appoint a meeting with email workflow
3. cancel the meeting
4. reminder emails still being received
Any other relevant information. For example, why do you consider this a bug and what did you expect to happen instead? expected to not receive any reminder
<sub>[CAL-1121](https://linear.app/calcom/issue/CAL-1121/reminder-emails-being-sent-even-after-cancellation)</sub> | priority | reminder emails being sent even after cancellation found a bug please fill out the sections below 👍 issue summary reminder emails being sent even after cancellation steps to reproduce set up a workflow with emails appoint a meeting with email workflow cancel the meeting reminder emails still being received any other relevant information for example why do you consider this a bug and what did you expect to happen instead expected to not receive any reminder | 1 |
399,340 | 11,747,510,940 | IssuesEvent | 2020-03-12 13:45:49 | DFO-Ocean-Navigator/Ocean-Data-Map-Project | https://api.github.com/repos/DFO-Ocean-Navigator/Ocean-Data-Map-Project | closed | Refactored UI: "auto" variable scale button not working in some cases | Bug Javascript Priority: Medium | Similar to #616 there appears to be a race condition where the "auto" scale button sends an erroneous api request that results in an exception thrown on the server side. The main issue is the user is unaware of this exception when it happens.
```sh
[2020-01-10 12:02:12 -0400] [1946] [ERROR] Error handling request /api/v1.0/range/giops_fc_2dps/votemper/gaussian/25/10/EPSG:3857/-12312000.908000194,2634036.759076399,1180051.8286728365,11331959.081703175/0/undefined.json?_=1578672061992
Traceback (most recent call last):
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/gunicorn/workers/sync.py", line 135, in handle
self.handle_request(listener, req, client, addr)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/gunicorn/workers/sync.py", line 176, in handle_request
respiter = self.wsgi(environ, resp.start_response)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1997, in __call__
return self.wsgi_app(environ, start_response)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1985, in wsgi_app
response = self.handle_exception(e)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1540, in handle_exception
reraise(exc_type, exc_value, tb)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/_compat.py", line 33, in reraise
raise value
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1982, in wsgi_app
response = self.full_dispatch_request()
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1614, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1517, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/_compat.py", line 33, in reraise
raise value
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1612, in full_dispatch_request
rv = self.dispatch_request()
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1598, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/home/buildadm/ocean-nav/Ocean-Data-Map-Project/routes/api_v1_0.py", line 273, in range_query_v1_0
string_to_datetime(time), config.time_dim_units)
File "/home/buildadm/ocean-nav/Ocean-Data-Map-Project/data/utils.py", line 67, in string_to_datetime
return dateutil.parser.parse(string).replace(tzinfo=pytz.UTC)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/dateutil/parser/_parser.py", line 1312, in parse
return DEFAULTPARSER.parse(timestr, **kwargs)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/dateutil/parser/_parser.py", line 607, in parse
raise ValueError("Unknown string format:", timestr)
ValueError: ('Unknown string format:', 'undefined')
``` | 1.0 | Refactored UI: "auto" variable scale button not working in some cases - Similar to #616 there appears to be a race condition where the "auto" scale button sends an erroneous api request that results in an exception thrown on the server side. The main issue is the user is unaware of this exception when it happens.
```sh
[2020-01-10 12:02:12 -0400] [1946] [ERROR] Error handling request /api/v1.0/range/giops_fc_2dps/votemper/gaussian/25/10/EPSG:3857/-12312000.908000194,2634036.759076399,1180051.8286728365,11331959.081703175/0/undefined.json?_=1578672061992
Traceback (most recent call last):
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/gunicorn/workers/sync.py", line 135, in handle
self.handle_request(listener, req, client, addr)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/gunicorn/workers/sync.py", line 176, in handle_request
respiter = self.wsgi(environ, resp.start_response)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1997, in __call__
return self.wsgi_app(environ, start_response)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1985, in wsgi_app
response = self.handle_exception(e)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1540, in handle_exception
reraise(exc_type, exc_value, tb)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/_compat.py", line 33, in reraise
raise value
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1982, in wsgi_app
response = self.full_dispatch_request()
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1614, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1517, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/_compat.py", line 33, in reraise
raise value
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1612, in full_dispatch_request
rv = self.dispatch_request()
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/flask/app.py", line 1598, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/home/buildadm/ocean-nav/Ocean-Data-Map-Project/routes/api_v1_0.py", line 273, in range_query_v1_0
string_to_datetime(time), config.time_dim_units)
File "/home/buildadm/ocean-nav/Ocean-Data-Map-Project/data/utils.py", line 67, in string_to_datetime
return dateutil.parser.parse(string).replace(tzinfo=pytz.UTC)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/dateutil/parser/_parser.py", line 1312, in parse
return DEFAULTPARSER.parse(timestr, **kwargs)
File "/home/sdfo000/ocean-nav/miniconda/3/envs/navigator/lib/python3.6/site-packages/dateutil/parser/_parser.py", line 607, in parse
raise ValueError("Unknown string format:", timestr)
ValueError: ('Unknown string format:', 'undefined')
``` | priority | refactored ui auto variable scale button not working in some cases similar to there appears to be a race condition where the auto scale button sends an erroneous api request that results in an exception thrown on the server side the main issue is the user is unaware of this exception when it happens sh error handling request api range giops fc votemper gaussian epsg undefined json traceback most recent call last file home ocean nav miniconda envs navigator lib site packages gunicorn workers sync py line in handle self handle request listener req client addr file home ocean nav miniconda envs navigator lib site packages gunicorn workers sync py line in handle request respiter self wsgi environ resp start response file home ocean nav miniconda envs navigator lib site packages flask app py line in call return self wsgi app environ start response file home ocean nav miniconda envs navigator lib site packages flask app py line in wsgi app response self handle exception e file home ocean nav miniconda envs navigator lib site packages flask app py line in handle exception reraise exc type exc value tb file home ocean nav miniconda envs navigator lib site packages flask compat py line in reraise raise value file home ocean nav miniconda envs navigator lib site packages flask app py line in wsgi app response self full dispatch request file home ocean nav miniconda envs navigator lib site packages flask app py line in full dispatch request rv self handle user exception e file home ocean nav miniconda envs navigator lib site packages flask app py line in handle user exception reraise exc type exc value tb file home ocean nav miniconda envs navigator lib site packages flask compat py line in reraise raise value file home ocean nav miniconda envs navigator lib site packages flask app py line in full dispatch request rv self dispatch request file home ocean nav miniconda envs navigator lib site packages flask app py line in dispatch request return self view functions req view args file home buildadm ocean nav ocean data map project routes api py line in range query string to datetime time config time dim units file home buildadm ocean nav ocean data map project data utils py line in string to datetime return dateutil parser parse string replace tzinfo pytz utc file home ocean nav miniconda envs navigator lib site packages dateutil parser parser py line in parse return defaultparser parse timestr kwargs file home ocean nav miniconda envs navigator lib site packages dateutil parser parser py line in parse raise valueerror unknown string format timestr valueerror unknown string format undefined | 1 |
42,031 | 2,869,094,235 | IssuesEvent | 2015-06-05 23:17:16 | dart-lang/test | https://api.github.com/repos/dart-lang/test | closed | unittest: expect doesn't take advantage of named arguments | enhancement Fixed Priority-Medium | <a href="https://github.com/seaneagan"><img src="https://avatars.githubusercontent.com/u/444270?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [seaneagan](https://github.com/seaneagan)**
_Originally opened as dart-lang/sdk#3684_
----
Current dart:unittest has:
expect(actual, [matcherOrReason = null, String reason = ''])
I think instead it should be:
expect(actual, [Matcher matcher = isTrue, String reason = ''])
and then if one wants to provide a reason, but no matcher, they can just do:
expect(a <= b, reason: '...');
which gives type safety for the "matcher" parameter, and allows it to use a default value of the "isTrue" matcher.
| 1.0 | unittest: expect doesn't take advantage of named arguments - <a href="https://github.com/seaneagan"><img src="https://avatars.githubusercontent.com/u/444270?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [seaneagan](https://github.com/seaneagan)**
_Originally opened as dart-lang/sdk#3684_
----
Current dart:unittest has:
expect(actual, [matcherOrReason = null, String reason = ''])
I think instead it should be:
expect(actual, [Matcher matcher = isTrue, String reason = ''])
and then if one wants to provide a reason, but no matcher, they can just do:
expect(a <= b, reason: '...');
which gives type safety for the "matcher" parameter, and allows it to use a default value of the "isTrue" matcher.
| priority | unittest expect doesn t take advantage of named arguments issue by originally opened as dart lang sdk current dart unittest has expect actual i think instead it should be expect actual and then if one wants to provide a reason but no matcher they can just do expect a lt b reason which gives type safety for the quot matcher quot parameter and allows it to use a default value of the quot istrue quot matcher | 1 |
640,701 | 20,796,817,668 | IssuesEvent | 2022-03-17 10:06:07 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | [BUG] Redirect to cloud from agent email alert doesn't highlight the affected chart | bug priority/medium alerts-team-bugs cloud-frontend | Steps to reproduce:
1. Setup email setting to the agent
2. Claim agent to the cloud
3. Trigger alert on the agent
4. Click the Go To Chart CTA inside the **agent email** for that alert (important - not the email sent from the cloud)
**Expected:**
1. User is re-directed to nodes single node view.
2. The affected chart should be highlighted and scroll to its canvas
**Actual:**
1. User should be re-directed to nodes single node view.
2. The affected chart is highlighted, but focus stays on top of the screen.
3. Error message that chart doesn't exist on the agent any more even though it does
**This issue is ONLY for the described flow, other flows, like following the alert from the alert details modal in the cloud, or the cloud emails work as expected!**

| 1.0 | [BUG] Redirect to cloud from agent email alert doesn't highlight the affected chart - Steps to reproduce:
1. Setup email setting to the agent
2. Claim agent to the cloud
3. Trigger alert on the agent
4. Click the Go To Chart CTA inside the **agent email** for that alert (important - not the email sent from the cloud)
**Expected:**
1. User is re-directed to nodes single node view.
2. The affected chart should be highlighted and scroll to its canvas
**Actual:**
1. User should be re-directed to nodes single node view.
2. The affected chart is highlighted, but focus stays on top of the screen.
3. Error message that chart doesn't exist on the agent any more even though it does
**This issue is ONLY for the described flow, other flows, like following the alert from the alert details modal in the cloud, or the cloud emails work as expected!**

| priority | redirect to cloud from agent email alert doesn t highlight the affected chart steps to reproduce setup email setting to the agent claim agent to the cloud trigger alert on the agent click the go to chart cta inside the agent email for that alert important not the email sent from the cloud expected user is re directed to nodes single node view the affected chart should be highlighted and scroll to its canvas actual user should be re directed to nodes single node view the affected chart is highlighted but focus stays on top of the screen error message that chart doesn t exist on the agent any more even though it does this issue is only for the described flow other flows like following the alert from the alert details modal in the cloud or the cloud emails work as expected | 1 |
55,008 | 3,071,811,618 | IssuesEvent | 2015-08-19 14:08:35 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | Want to open a new browser using robotium | bug imported invalid Priority-Medium | _From [kommalap...@gmail.com](https://code.google.com/u/115815323554714072359/) on February 17, 2011 23:13:01_
can you please help me out how to open a browser window using robotium
_Original issue: http://code.google.com/p/robotium/issues/detail?id=74_ | 1.0 | Want to open a new browser using robotium - _From [kommalap...@gmail.com](https://code.google.com/u/115815323554714072359/) on February 17, 2011 23:13:01_
can you please help me out how to open a browser window using robotium
_Original issue: http://code.google.com/p/robotium/issues/detail?id=74_ | priority | want to open a new browser using robotium from on february can you please help me out how to open a browser window using robotium original issue | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.