Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
698,021 | 23,962,711,693 | IssuesEvent | 2022-09-12 20:43:10 | OpenPrinting/libcups | https://api.github.com/repos/OpenPrinting/libcups | closed | Public DNS-SD API for browsing, resolving, and publishing services | enhancement priority-high | Currently the various bits of CUPS, PAPPL, etc. have a conditional mess of code that use the Avahi or mDNSResponder APIs to do DNS-SD discovery and sharing. This continues to be a major pain, and CUPS should have a common API that developers can use to access whatever the local API/service is for this. Goals:
- Support the [Avahi](https://avahi.org), mDNSResponder, and [Win32](https://docs.microsoft.com/en-us/windows/win32/api/_dns/) APIs, with potential for others (systemd/D-Bus) in the future
- Support asynchronous browsing, resolution, and service publication with support for sub-types, domains, and LOC+TXT records
- Provide callbacks as needed
- Provide helper functions for renaming, getting the mDNS hostname, and getting/setting a list of DNS-SD domains
| 1.0 | Public DNS-SD API for browsing, resolving, and publishing services - Currently the various bits of CUPS, PAPPL, etc. have a conditional mess of code that use the Avahi or mDNSResponder APIs to do DNS-SD discovery and sharing. This continues to be a major pain, and CUPS should have a common API that developers can use to access whatever the local API/service is for this. Goals:
- Support the [Avahi](https://avahi.org), mDNSResponder, and [Win32](https://docs.microsoft.com/en-us/windows/win32/api/_dns/) APIs, with potential for others (systemd/D-Bus) in the future
- Support asynchronous browsing, resolution, and service publication with support for sub-types, domains, and LOC+TXT records
- Provide callbacks as needed
- Provide helper functions for renaming, getting the mDNS hostname, and getting/setting a list of DNS-SD domains
| priority | public dns sd api for browsing resolving and publishing services currently the various bits of cups pappl etc have a conditional mess of code that use the avahi or mdnsresponder apis to do dns sd discovery and sharing this continues to be a major pain and cups should have a common api that developers can use to access whatever the local api service is for this goals support the mdnsresponder and apis with potential for others systemd d bus in the future support asynchronous browsing resolution and service publication with support for sub types domains and loc txt records provide callbacks as needed provide helper functions for renaming getting the mdns hostname and getting setting a list of dns sd domains | 1 |
477,971 | 13,770,884,154 | IssuesEvent | 2020-10-07 20:59:00 | Viktor50/mymarket | https://api.github.com/repos/Viktor50/mymarket | opened | Связь между товаром в каталоге и товаром в магазине | High Priority | Т.к. кроме создания товара в магазине, нам будет необходимо держать его в актуальном состоянии (характеристики, цена, доступность), нам нужно хранить связь между карточкой товара в каталоге и магазине.
Что это значит и как должно выглядеть по итогу? Если у нас в магазине уже есть этот товар и мы нажимаем кнопку "Добавить/обновить товар в магазин", то все данные, которые мы отправляем из каталога должны перезаписать имеющиеся данные в магазине для этого товара (а не создать новый товар).
С базовой информацией по товару (наименование, модель, артикул, цена, наличие, характеристики, метатеги и т.д.) всё в принципе просто - ее просто перезаписываем, если она менялась с момента последнего использования кнопки "добавить/обновить".
Если у товара обновились картинки, то мы делаем следующее (**этот момент разработаем вместе**):
1. В БД удаляем старые ссылки на картинки для этого товара и прописываем новые
2. На стороне магазина удаляем в кэше папки с картинками этого товара. Эти папки расположены в двух местах: /var/www/ezon.by/image/cache/catalog/Good и /var/www/ezon.by/image/cache/webp/catalog/Good
3. Запускаем процесс кэширования картинок для этого товара (используем curl). | 1.0 | Связь между товаром в каталоге и товаром в магазине - Т.к. кроме создания товара в магазине, нам будет необходимо держать его в актуальном состоянии (характеристики, цена, доступность), нам нужно хранить связь между карточкой товара в каталоге и магазине.
Что это значит и как должно выглядеть по итогу? Если у нас в магазине уже есть этот товар и мы нажимаем кнопку "Добавить/обновить товар в магазин", то все данные, которые мы отправляем из каталога должны перезаписать имеющиеся данные в магазине для этого товара (а не создать новый товар).
С базовой информацией по товару (наименование, модель, артикул, цена, наличие, характеристики, метатеги и т.д.) всё в принципе просто - ее просто перезаписываем, если она менялась с момента последнего использования кнопки "добавить/обновить".
Если у товара обновились картинки, то мы делаем следующее (**этот момент разработаем вместе**):
1. В БД удаляем старые ссылки на картинки для этого товара и прописываем новые
2. На стороне магазина удаляем в кэше папки с картинками этого товара. Эти папки расположены в двух местах: /var/www/ezon.by/image/cache/catalog/Good и /var/www/ezon.by/image/cache/webp/catalog/Good
3. Запускаем процесс кэширования картинок для этого товара (используем curl). | priority | связь между товаром в каталоге и товаром в магазине т к кроме создания товара в магазине нам будет необходимо держать его в актуальном состоянии характеристики цена доступность нам нужно хранить связь между карточкой товара в каталоге и магазине что это значит и как должно выглядеть по итогу если у нас в магазине уже есть этот товар и мы нажимаем кнопку добавить обновить товар в магазин то все данные которые мы отправляем из каталога должны перезаписать имеющиеся данные в магазине для этого товара а не создать новый товар с базовой информацией по товару наименование модель артикул цена наличие характеристики метатеги и т д всё в принципе просто ее просто перезаписываем если она менялась с момента последнего использования кнопки добавить обновить если у товара обновились картинки то мы делаем следующее этот момент разработаем вместе в бд удаляем старые ссылки на картинки для этого товара и прописываем новые на стороне магазина удаляем в кэше папки с картинками этого товара эти папки расположены в двух местах var www ezon by image cache catalog good и var www ezon by image cache webp catalog good запускаем процесс кэширования картинок для этого товара используем curl | 1 |
212,538 | 7,238,286,040 | IssuesEvent | 2018-02-13 14:09:43 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | Removing and adding center contact points (PPA Partners) is not working correctly | Priority - High Type - Bug | I tried removing a PPA Partner contact point from FTA (Institution Tropenbos) and add a new contact point and:
1. The old contact point can not be removed.
2. The new contact point does not display in the PPA Partner screen (note that when I go to the user list that the new contact point appears and an email notification was sent out).
I am unable to remove Henk Lijftogt and newly added user Herman Savenije does not appear

In the user screen I can see that Herman has been added and Henk has not been removed.

| 1.0 | Removing and adding center contact points (PPA Partners) is not working correctly - I tried removing a PPA Partner contact point from FTA (Institution Tropenbos) and add a new contact point and:
1. The old contact point can not be removed.
2. The new contact point does not display in the PPA Partner screen (note that when I go to the user list that the new contact point appears and an email notification was sent out).
I am unable to remove Henk Lijftogt and newly added user Herman Savenije does not appear

In the user screen I can see that Herman has been added and Henk has not been removed.

| priority | removing and adding center contact points ppa partners is not working correctly i tried removing a ppa partner contact point from fta institution tropenbos and add a new contact point and the old contact point can not be removed the new contact point does not display in the ppa partner screen note that when i go to the user list that the new contact point appears and an email notification was sent out i am unable to remove henk lijftogt and newly added user herman savenije does not appear in the user screen i can see that herman has been added and henk has not been removed | 1 |
209,660 | 7,178,411,402 | IssuesEvent | 2018-01-31 16:25:12 | metasfresh/metasfresh | https://api.github.com/repos/metasfresh/metasfresh | opened | Dunning Level is not set in invoice after generating dunning doc | priority:high type:bug | ### Is this a bug or feature request?
bug
### What is the current behavior?
if you create a dunning doc for an invoice the field c_dunning_level_id remains null
#### Which are the steps to reproduce?
1. follow http://docs.metasfresh.org/webui_collection/EN/Dunning_Run.html
1. open the invoice and check dunning level
1. NOK: its null
### What is the expected or desired behavior?
the field should be filled so people can see the dunning level right on the invoice
if not filled the sql to find is quite heavy:
http://docs.metasfresh.org/sql_collection/c_dunning.html
| 1.0 | Dunning Level is not set in invoice after generating dunning doc - ### Is this a bug or feature request?
bug
### What is the current behavior?
if you create a dunning doc for an invoice the field c_dunning_level_id remains null
#### Which are the steps to reproduce?
1. follow http://docs.metasfresh.org/webui_collection/EN/Dunning_Run.html
1. open the invoice and check dunning level
1. NOK: its null
### What is the expected or desired behavior?
the field should be filled so people can see the dunning level right on the invoice
if not filled the sql to find is quite heavy:
http://docs.metasfresh.org/sql_collection/c_dunning.html
| priority | dunning level is not set in invoice after generating dunning doc is this a bug or feature request bug what is the current behavior if you create a dunning doc for an invoice the field c dunning level id remains null which are the steps to reproduce follow open the invoice and check dunning level nok its null what is the expected or desired behavior the field should be filled so people can see the dunning level right on the invoice if not filled the sql to find is quite heavy | 1 |
665,512 | 22,320,596,891 | IssuesEvent | 2022-06-14 05:52:13 | younginnovations/iatipublisher | https://api.github.com/repos/younginnovations/iatipublisher | opened | #57 Bug: Any user is allowed to access all of the activities created by any user | type: bug priority: high Backend | ### Context ###
- Desktop
- Chrome 102.0.5005.61
### Preconditon ###
https://stage.iatipublisher.yipl.com.np/
- Username: User_A
- Password: test1234
### Steps ###
- Go to
- https://stage.iatipublisher.yipl.com.np/activities/1
- https://stage.iatipublisher.yipl.com.np/activities/2
- https://stage.iatipublisher.yipl.com.np/activities/3
- https://stage.iatipublisher.yipl.com.np/activities/11
- https://stage.iatipublisher.yipl.com.np/activities/37
### Actual result ###
- User allowed to access to all of the activities which is not its own
### Expected result ###
- Activities which is only of its own is to be allowed to be accessible
| 1.0 | #57 Bug: Any user is allowed to access all of the activities created by any user - ### Context ###
- Desktop
- Chrome 102.0.5005.61
### Preconditon ###
https://stage.iatipublisher.yipl.com.np/
- Username: User_A
- Password: test1234
### Steps ###
- Go to
- https://stage.iatipublisher.yipl.com.np/activities/1
- https://stage.iatipublisher.yipl.com.np/activities/2
- https://stage.iatipublisher.yipl.com.np/activities/3
- https://stage.iatipublisher.yipl.com.np/activities/11
- https://stage.iatipublisher.yipl.com.np/activities/37
### Actual result ###
- User allowed to access to all of the activities which is not its own
### Expected result ###
- Activities which is only of its own is to be allowed to be accessible
| priority | bug any user is allowed to access all of the activities created by any user context desktop chrome preconditon username user a password steps go to actual result user allowed to access to all of the activities which is not its own expected result activities which is only of its own is to be allowed to be accessible | 1 |
372,954 | 11,030,987,586 | IssuesEvent | 2019-12-06 16:47:39 | ukwa/ukwa-ui | https://api.github.com/repos/ukwa/ukwa-ui | opened | Add more content to UKWA | Backend priority: high | Make 2015, 2016 (and 2017?) annual domain crawls accessible and searchable in UKWA.
Make 2019 Frequent crawls searchable in UKWA. | 1.0 | Add more content to UKWA - Make 2015, 2016 (and 2017?) annual domain crawls accessible and searchable in UKWA.
Make 2019 Frequent crawls searchable in UKWA. | priority | add more content to ukwa make and annual domain crawls accessible and searchable in ukwa make frequent crawls searchable in ukwa | 1 |
522,588 | 15,162,533,464 | IssuesEvent | 2021-02-12 10:45:08 | Najaran/NAJARANTIS_Modpack | https://api.github.com/repos/Najaran/NAJARANTIS_Modpack | closed | 初期ヘルスがハート3つ分になる | high priority | <!--
問題点に関するIssueを発行する場合、以下のフォームに沿って発行してください
* 基本情報、問題内容の要約については原則漏れなく記載してください。
* 再現性、スクリーンショットについては必須ではありませんが、説明時に必要があれば併せて報告ください。
* クラッシュログにはMinecraftに限らず、様々な情報が含まれています。取り扱いには十分に注意してください。(必要であればマスキングをしてください。)
-->
## 基本情報
* バージョン : alpha-0.1.4
* 環境 : クライアント
* Modpackへの改変 : なし
* 改変内容 :
## 問題内容の要約
初期ヘルスがハート3つ分になります。
## スクリーンショット

## クラッシュログ
クラッシュを伴わなかったため、無し | 1.0 | 初期ヘルスがハート3つ分になる - <!--
問題点に関するIssueを発行する場合、以下のフォームに沿って発行してください
* 基本情報、問題内容の要約については原則漏れなく記載してください。
* 再現性、スクリーンショットについては必須ではありませんが、説明時に必要があれば併せて報告ください。
* クラッシュログにはMinecraftに限らず、様々な情報が含まれています。取り扱いには十分に注意してください。(必要であればマスキングをしてください。)
-->
## 基本情報
* バージョン : alpha-0.1.4
* 環境 : クライアント
* Modpackへの改変 : なし
* 改変内容 :
## 問題内容の要約
初期ヘルスがハート3つ分になります。
## スクリーンショット

## クラッシュログ
クラッシュを伴わなかったため、無し | priority | 問題点に関するissueを発行する場合、以下のフォームに沿って発行してください 基本情報、問題内容の要約については原則漏れなく記載してください。 再現性、スクリーンショットについては必須ではありませんが、説明時に必要があれば併せて報告ください。 クラッシュログにはminecraftに限らず、様々な情報が含まれています。取り扱いには十分に注意してください。 必要であればマスキングをしてください。 基本情報 バージョン alpha 環境 クライアント modpackへの改変 なし 改変内容 問題内容の要約 。 スクリーンショット クラッシュログ クラッシュを伴わなかったため、無し | 1 |
787,871 | 27,734,142,319 | IssuesEvent | 2023-03-15 10:04:46 | AY2223S2-CS2113-T13-1/tp | https://api.github.com/repos/AY2223S2-CS2113-T13-1/tp | closed | "withdraw" command | type.Story priority.High | As a working student, I can withdraw money from my accounts so that I can spend it as cash.
### Acceptance Criteria
- Withdraw the amount of money of specified currency.
- The balance of the account should decrease by the specified amount
- Handle error if no account for this currency exists
- Handle error if amount is too big
- Handle error if input is invalid, e.g. amount cannot be parsed to a decimal
- Format: `withdraw 100 $/MYR` | 1.0 | "withdraw" command - As a working student, I can withdraw money from my accounts so that I can spend it as cash.
### Acceptance Criteria
- Withdraw the amount of money of specified currency.
- The balance of the account should decrease by the specified amount
- Handle error if no account for this currency exists
- Handle error if amount is too big
- Handle error if input is invalid, e.g. amount cannot be parsed to a decimal
- Format: `withdraw 100 $/MYR` | priority | withdraw command as a working student i can withdraw money from my accounts so that i can spend it as cash acceptance criteria withdraw the amount of money of specified currency the balance of the account should decrease by the specified amount handle error if no account for this currency exists handle error if amount is too big handle error if input is invalid e g amount cannot be parsed to a decimal format withdraw myr | 1 |
344,372 | 10,343,691,958 | IssuesEvent | 2019-09-04 09:29:47 | bbc/simorgh | https://api.github.com/repos/bbc/simorgh | closed | Confirm that the manifest.json is correct for Punjabi | high priority simorgh-core-stream ws-fp-phase2 ws-frontpage-stream | Blocked on https://github.com/bbc/simorgh/issues/3516
**Is your feature request related to a problem? Please describe.**
Confirm that the links in the manifest.json are correct for the assets added in https://github.com/bbc/simorgh/issues/3516
Confirm that the link to the manifest.json is correct - it should not live in /article it should live in the root eg bbc.com/punjabi/manifest.json
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Testing notes**
[Tester to complete]
Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc
- [ ] This feature is expected to need manual testing.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Confirm that the manifest.json is correct for Punjabi - Blocked on https://github.com/bbc/simorgh/issues/3516
**Is your feature request related to a problem? Please describe.**
Confirm that the links in the manifest.json are correct for the assets added in https://github.com/bbc/simorgh/issues/3516
Confirm that the link to the manifest.json is correct - it should not live in /article it should live in the root eg bbc.com/punjabi/manifest.json
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Testing notes**
[Tester to complete]
Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc
- [ ] This feature is expected to need manual testing.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | confirm that the manifest json is correct for punjabi blocked on is your feature request related to a problem please describe confirm that the links in the manifest json are correct for the assets added in confirm that the link to the manifest json is correct it should not live in article it should live in the root eg bbc com punjabi manifest json describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered testing notes dev insight will cypress tests be required or are unit tests sufficient will there be any potential regression etc this feature is expected to need manual testing additional context add any other context or screenshots about the feature request here | 1 |
80,643 | 3,572,677,747 | IssuesEvent | 2016-01-27 00:47:35 | pathwaysmedical/frasernw | https://api.github.com/repos/pathwaysmedical/frasernw | closed | CR: Hide Automatic Specialist / Clinic Updates | ChangeRequest High Priority New Feature | CR 13. November 2015. Priority 2.
Automatic **Latest Specialist / Clinic Updates** that appear on home page should be able to be edited or hidden if necessary.
#231 and this should be handled together.
------------
Comment by me: This could be tricky due to our use of `versions` via the `Papertrail` gem for generating Latest Specialist / Clinic Updates. | 1.0 | CR: Hide Automatic Specialist / Clinic Updates - CR 13. November 2015. Priority 2.
Automatic **Latest Specialist / Clinic Updates** that appear on home page should be able to be edited or hidden if necessary.
#231 and this should be handled together.
------------
Comment by me: This could be tricky due to our use of `versions` via the `Papertrail` gem for generating Latest Specialist / Clinic Updates. | priority | cr hide automatic specialist clinic updates cr november priority automatic latest specialist clinic updates that appear on home page should be able to be edited or hidden if necessary and this should be handled together comment by me this could be tricky due to our use of versions via the papertrail gem for generating latest specialist clinic updates | 1 |
540,289 | 15,805,655,086 | IssuesEvent | 2021-04-04 00:17:48 | LittleImprovementsCustom/LittleImprovementsCustom | https://api.github.com/repos/LittleImprovementsCustom/LittleImprovementsCustom | closed | Variated fungi doesn't show it's image | category: resource packs priority: high scale: simple type: bug |
> This issue was created by an automation. It was authored in Discord by Nigelrex#8452, in Beatserver #lounge. | 1.0 | Variated fungi doesn't show it's image -
> This issue was created by an automation. It was authored in Discord by Nigelrex#8452, in Beatserver #lounge. | priority | variated fungi doesn t show it s image this issue was created by an automation it was authored in discord by nigelrex in beatserver lounge | 1 |
272,101 | 8,499,108,026 | IssuesEvent | 2018-10-29 16:22:09 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | Travis build does not start on stable branch | Environment Priority: High pending review review | ### Description
Travis regex does not match stable branch names.
### In case of Bug (otherwise remove this paragraph)
*Browser Affected*
(use this site: https://www.whatsmybrowser.org/ for non expert users)
- [ ] Internet Explorer
- [ ] Chrome
- [ ] Firefox
- [ ] Safari
*Browser Version Affected*
- Indicate the browser version in which the issue has been found
*Steps to reproduce*
- open a pr on stable branch
*Expected Result*
- travis buil will start
*Current Result*
- it does not start
### Other useful information (optional):
| 1.0 | Travis build does not start on stable branch - ### Description
Travis regex does not match stable branch names.
### In case of Bug (otherwise remove this paragraph)
*Browser Affected*
(use this site: https://www.whatsmybrowser.org/ for non expert users)
- [ ] Internet Explorer
- [ ] Chrome
- [ ] Firefox
- [ ] Safari
*Browser Version Affected*
- Indicate the browser version in which the issue has been found
*Steps to reproduce*
- open a pr on stable branch
*Expected Result*
- travis buil will start
*Current Result*
- it does not start
### Other useful information (optional):
| priority | travis build does not start on stable branch description travis regex does not match stable branch names in case of bug otherwise remove this paragraph browser affected use this site for non expert users internet explorer chrome firefox safari browser version affected indicate the browser version in which the issue has been found steps to reproduce open a pr on stable branch expected result travis buil will start current result it does not start other useful information optional | 1 |
116,613 | 4,704,304,169 | IssuesEvent | 2016-10-13 10:58:59 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Absorption for generic geometries - Improve UI, handling can materials etc | Component: Framework Misc: Roadmap Priority: High | The UI should focus on letting the use select from a list of possible geometries of sample holders etc. This may be central, central and filtered by instrument, or instrument specific (yet to decide). The sample in a cynlindrical sample holder the sample depth would need to be entered, but most other things can be defined in advance.
There are some anvil shapes on this tickety for PEARL #3119 | 1.0 | Absorption for generic geometries - Improve UI, handling can materials etc - The UI should focus on letting the use select from a list of possible geometries of sample holders etc. This may be central, central and filtered by instrument, or instrument specific (yet to decide). The sample in a cynlindrical sample holder the sample depth would need to be entered, but most other things can be defined in advance.
There are some anvil shapes on this tickety for PEARL #3119 | priority | absorption for generic geometries improve ui handling can materials etc the ui should focus on letting the use select from a list of possible geometries of sample holders etc this may be central central and filtered by instrument or instrument specific yet to decide the sample in a cynlindrical sample holder the sample depth would need to be entered but most other things can be defined in advance there are some anvil shapes on this tickety for pearl | 1 |
384,731 | 11,402,433,200 | IssuesEvent | 2020-01-31 03:11:08 | woocommerce/woocommerce-gateway-paypal-express-checkout | https://api.github.com/repos/woocommerce/woocommerce-gateway-paypal-express-checkout | closed | Smart Payment Buttons erroring with subscription products | Feature: Smart Payment Buttons Priority: High | 2293488-zd
## Synopsis
When clicking the credit card buttons, also known as _Smart Payment Buttons_ on the product page results in errors. This is only happening on subscription products (variable and simple). Non-subscription products allow inline checkout functionality for single product page.
## Details
**1. Set up Smart payment buttons**
These are the Smart Button settings that will get these buttons to show up on the single product page.

**2. Click Smart payment buttons from a subscription product page**
On subscription product page, click Visa, MasterCard, or Discover, to get this error message:

If you click Try Again, the modal PayPal payment information window loads:

If you click Amex, the modal payment window loads on the first click.

We were able to replicate this issue with subscription products. Based on this, I think we may be looking at another version of this bug, or a different but related bug: https://github.com/woocommerce/woocommerce-gateway-paypal-express-checkout/issues/596
No WC fatal errors logged, no PayPal errors logged.
## Expected behavior
If inline checkout for subscription products isn't supported, this checkout method should default to the modal with no error message.
| 1.0 | Smart Payment Buttons erroring with subscription products - 2293488-zd
## Synopsis
When clicking the credit card buttons, also known as _Smart Payment Buttons_ on the product page results in errors. This is only happening on subscription products (variable and simple). Non-subscription products allow inline checkout functionality for single product page.
## Details
**1. Set up Smart payment buttons**
These are the Smart Button settings that will get these buttons to show up on the single product page.

**2. Click Smart payment buttons from a subscription product page**
On subscription product page, click Visa, MasterCard, or Discover, to get this error message:

If you click Try Again, the modal PayPal payment information window loads:

If you click Amex, the modal payment window loads on the first click.

We were able to replicate this issue with subscription products. Based on this, I think we may be looking at another version of this bug, or a different but related bug: https://github.com/woocommerce/woocommerce-gateway-paypal-express-checkout/issues/596
No WC fatal errors logged, no PayPal errors logged.
## Expected behavior
If inline checkout for subscription products isn't supported, this checkout method should default to the modal with no error message.
| priority | smart payment buttons erroring with subscription products zd synopsis when clicking the credit card buttons also known as smart payment buttons on the product page results in errors this is only happening on subscription products variable and simple non subscription products allow inline checkout functionality for single product page details set up smart payment buttons these are the smart button settings that will get these buttons to show up on the single product page click smart payment buttons from a subscription product page on subscription product page click visa mastercard or discover to get this error message if you click try again the modal paypal payment information window loads if you click amex the modal payment window loads on the first click we were able to replicate this issue with subscription products based on this i think we may be looking at another version of this bug or a different but related bug no wc fatal errors logged no paypal errors logged expected behavior if inline checkout for subscription products isn t supported this checkout method should default to the modal with no error message | 1 |
103,223 | 4,165,523,729 | IssuesEvent | 2016-06-19 15:11:16 | ArdaCraft/IssueTracker | https://api.github.com/repos/ArdaCraft/IssueTracker | closed | Modblock torches missing | bug high priority mod | Noticed some of the modblock torch/lanterns were missing - may be an issue with the world converter | 1.0 | Modblock torches missing - Noticed some of the modblock torch/lanterns were missing - may be an issue with the world converter | priority | modblock torches missing noticed some of the modblock torch lanterns were missing may be an issue with the world converter | 1 |
300,520 | 9,211,330,352 | IssuesEvent | 2019-03-09 14:29:21 | qgisissuebot/QGIS | https://api.github.com/repos/qgisissuebot/QGIS | closed | Regression with expressions in Composer | Bug Priority: high Regression | ---
Author Name: **Pedro Venâncio** (Pedro Venâncio)
Original Redmine Issue: 21471, https://issues.qgis.org/issues/21471
---
I updated QGIS 3.4.5, 3.6.0 and 3.7.0 today to 3.4.5-3 (code revision 8ba90fc441), 3.6.0-4 (code revision ec30c48de9) and 3.7.0-4 (code revision 34a0650177) with OSGeo4W, and the labels with expressions are not rendered as expected, when are used more than one line.
For instance,
```
[% CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now) %]
```
renders ok
```
March 2019
```
but
```
[%
CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now)
%]
```
renders
```
[%
CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now)
%]
```
I'm sure this is a regression, because it was working fine before today's update. Last week this was working fine.
| 1.0 | Regression with expressions in Composer - ---
Author Name: **Pedro Venâncio** (Pedro Venâncio)
Original Redmine Issue: 21471, https://issues.qgis.org/issues/21471
---
I updated QGIS 3.4.5, 3.6.0 and 3.7.0 today to 3.4.5-3 (code revision 8ba90fc441), 3.6.0-4 (code revision ec30c48de9) and 3.7.0-4 (code revision 34a0650177) with OSGeo4W, and the labels with expressions are not rendered as expected, when are used more than one line.
For instance,
```
[% CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now) %]
```
renders ok
```
March 2019
```
but
```
[%
CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now)
%]
```
renders
```
[%
CASE WHEN month($now) = 3 THEN 'March' END || ' ' || year($now)
%]
```
I'm sure this is a regression, because it was working fine before today's update. Last week this was working fine.
| priority | regression with expressions in composer author name pedro venâncio pedro venâncio original redmine issue i updated qgis and today to code revision code revision and code revision with and the labels with expressions are not rendered as expected when are used more than one line for instance renders ok march but case when month now then march end year now renders case when month now then march end year now i m sure this is a regression because it was working fine before today s update last week this was working fine | 1 |
62,184 | 3,175,710,034 | IssuesEvent | 2015-09-24 02:12:37 | mozilla/mdn-tests | https://api.github.com/repos/mozilla/mdn-tests | closed | Upgrade py.test dependency to 2.7 | Community difficulty beginner priority high | We should update to the latest version of py.test. This can be done by updating the pinned versions of pytest, py, and any pytest plugins from `requirements.txt` to the latest versions. See https://github.com/mozilla/Addon-Tests/commit/bfa2336803debe57aaffaed4e1033e5d88f4c198 as an example.
We should also remove any dependencies that are not required, such as `pytest-xdist`, which is useful for running tests in parallel, but is not a hard requirement.
Before submitting a pull request please test the changes locally. If you need any assistance, either comment here or ask in #mozwebqa on irc.mozilla.org. Details of how to get onto IRC can be found [here](https://wiki.mozilla.org/IRC). | 1.0 | Upgrade py.test dependency to 2.7 - We should update to the latest version of py.test. This can be done by updating the pinned versions of pytest, py, and any pytest plugins from `requirements.txt` to the latest versions. See https://github.com/mozilla/Addon-Tests/commit/bfa2336803debe57aaffaed4e1033e5d88f4c198 as an example.
We should also remove any dependencies that are not required, such as `pytest-xdist`, which is useful for running tests in parallel, but is not a hard requirement.
Before submitting a pull request please test the changes locally. If you need any assistance, either comment here or ask in #mozwebqa on irc.mozilla.org. Details of how to get onto IRC can be found [here](https://wiki.mozilla.org/IRC). | priority | upgrade py test dependency to we should update to the latest version of py test this can be done by updating the pinned versions of pytest py and any pytest plugins from requirements txt to the latest versions see as an example we should also remove any dependencies that are not required such as pytest xdist which is useful for running tests in parallel but is not a hard requirement before submitting a pull request please test the changes locally if you need any assistance either comment here or ask in mozwebqa on irc mozilla org details of how to get onto irc can be found | 1 |
766,299 | 26,877,412,694 | IssuesEvent | 2023-02-05 07:34:18 | nokotan/siv3d-studio | https://api.github.com/repos/nokotan/siv3d-studio | opened | 幅の狭い端末でアクティビティバーが押せなくなることがある | priority: high wontfix | ### 再現手順
テキストエディタで右側にカーソルを置いた状態でエクスプローラを開く
### 再現端末
Android 12 + Google Chrome | 1.0 | 幅の狭い端末でアクティビティバーが押せなくなることがある - ### 再現手順
テキストエディタで右側にカーソルを置いた状態でエクスプローラを開く
### 再現端末
Android 12 + Google Chrome | priority | 幅の狭い端末でアクティビティバーが押せなくなることがある 再現手順 テキストエディタで右側にカーソルを置いた状態でエクスプローラを開く 再現端末 android google chrome | 1 |
55,619 | 3,073,968,333 | IssuesEvent | 2015-08-20 02:20:08 | canadainc/ilmtest | https://api.github.com/repos/canadainc/ilmtest | opened | Implement client-server connection for download sessions | logic Priority-High task | To download the database and uncompress it. | 1.0 | Implement client-server connection for download sessions - To download the database and uncompress it. | priority | implement client server connection for download sessions to download the database and uncompress it | 1 |
684,946 | 23,439,206,702 | IssuesEvent | 2022-08-15 13:21:11 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | closed | Detect if WP Rocket is on one.com environment | type: enhancement priority: high effort: [S] one.com | Hi
Similar to what WP Rocket has for other hosting providers such as seen in:
https://github.com/wp-media/wp-rocket/blob/develop/inc/ThirdParty/Hostings/HostResolver.php
It would be great if this could also be done for domains on one.com environment.
We have a lot of server variables for instance $_SERVER['ONECOM_DOMAIN_NAME'] that would only be present on the one.com environment.
So if you have a domain let's call it domain.tld the ONECOM_DOMAIN_NAME server variable would be "domain.tld", simply detecting if this variable gives any value would be good enough to detect if it is on a one.com environment or not.
The reason for this request is that once this detection is in place, we can also improve some other things such as certain tailor-made pre-selected settings for domains on one.com environment but for those I will create separate tickets as to not make this one too cluttered.
Thank you and please let me know if any further information is required from my side. | 1.0 | Detect if WP Rocket is on one.com environment - Hi
Similar to what WP Rocket has for other hosting providers such as seen in:
https://github.com/wp-media/wp-rocket/blob/develop/inc/ThirdParty/Hostings/HostResolver.php
It would be great if this could also be done for domains on one.com environment.
We have a lot of server variables for instance $_SERVER['ONECOM_DOMAIN_NAME'] that would only be present on the one.com environment.
So if you have a domain let's call it domain.tld the ONECOM_DOMAIN_NAME server variable would be "domain.tld", simply detecting if this variable gives any value would be good enough to detect if it is on a one.com environment or not.
The reason for this request is that once this detection is in place, we can also improve some other things such as certain tailor-made pre-selected settings for domains on one.com environment but for those I will create separate tickets as to not make this one too cluttered.
Thank you and please let me know if any further information is required from my side. | priority | detect if wp rocket is on one com environment hi similar to what wp rocket has for other hosting providers such as seen in it would be great if this could also be done for domains on one com environment we have a lot of server variables for instance server that would only be present on the one com environment so if you have a domain let s call it domain tld the onecom domain name server variable would be domain tld simply detecting if this variable gives any value would be good enough to detect if it is on a one com environment or not the reason for this request is that once this detection is in place we can also improve some other things such as certain tailor made pre selected settings for domains on one com environment but for those i will create separate tickets as to not make this one too cluttered thank you and please let me know if any further information is required from my side | 1 |
323,017 | 9,842,039,595 | IssuesEvent | 2019-06-18 08:24:07 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Ability to configure CORS header "Access-Control-Expose-Headers" in WSO2 API Manager | 2.6.0 Priority/High Type/Improvement commitment | The ability to configure the CORS header "Access-Control-Expose-Headers" is not available in the WSO2 API Manager at the moment.
It would be better if this feature can be added to a future release. | 1.0 | Ability to configure CORS header "Access-Control-Expose-Headers" in WSO2 API Manager - The ability to configure the CORS header "Access-Control-Expose-Headers" is not available in the WSO2 API Manager at the moment.
It would be better if this feature can be added to a future release. | priority | ability to configure cors header access control expose headers in api manager the ability to configure the cors header access control expose headers is not available in the api manager at the moment it would be better if this feature can be added to a future release | 1 |
570,080 | 17,018,452,473 | IssuesEvent | 2021-07-02 15:09:16 | lakeboy93/Money4Mobs | https://api.github.com/repos/lakeboy93/Money4Mobs | closed | Using /mk with no parameters throws an error in the console. | High-Priority bug | There appears to be no handler for /mk on its own, resulting on an ArrayIndexOutOfBoundsException.
```
[05:44:17 ERROR]: null
org.bukkit.command.CommandException: Unhandled exception executing command 'mk' in plugin Money4Mobs v1.6.4
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:47) ~[patched_1.17.jar:git-Paper-69]
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:159) ~[patched_1.17.jar:git-Paper-69]
at org.bukkit.craftbukkit.v1_17_R1.CraftServer.dispatchCommand(CraftServer.java:821) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleCommand(ServerGamePacketListenerImpl.java:2185) ~[?:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleChat(ServerGamePacketListenerImpl.java:1996) ~[?:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleChat(ServerGamePacketListenerImpl.java:1977) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundChatPacket.handle(ServerboundChatPacket.java:46) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundChatPacket.handle(ServerboundChatPacket.java:6) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:36) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1340) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.shouldRun(MinecraftServer.java:193) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1319) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1312) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1273) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1184) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:320) ~[patched_1.17.jar:git-Paper-69] at java.lang.Thread.run(Thread.java:831) [?:?]
Caused by: java.lang.ArrayIndexOutOfBoundsException: Index 0 out of bounds for length 0
at Latch.Money4Mobs.MkCommand.onCommand(MkCommand.java:764) ~[?:?]
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:45) ~[patched_1.17.jar:git-Paper-69]
... 21 more`
```
### Server Information
Paper 1.17 build 69 with the following plugins:
- 1.17---Money4Mobs-v1.6.4
- iEconomy Release 1.1
- LuckPerms Bukkit-5.3.48
- Vault
| 1.0 | Using /mk with no parameters throws an error in the console. - There appears to be no handler for /mk on its own, resulting on an ArrayIndexOutOfBoundsException.
```
[05:44:17 ERROR]: null
org.bukkit.command.CommandException: Unhandled exception executing command 'mk' in plugin Money4Mobs v1.6.4
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:47) ~[patched_1.17.jar:git-Paper-69]
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:159) ~[patched_1.17.jar:git-Paper-69]
at org.bukkit.craftbukkit.v1_17_R1.CraftServer.dispatchCommand(CraftServer.java:821) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleCommand(ServerGamePacketListenerImpl.java:2185) ~[?:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleChat(ServerGamePacketListenerImpl.java:1996) ~[?:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleChat(ServerGamePacketListenerImpl.java:1977) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundChatPacket.handle(ServerboundChatPacket.java:46) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundChatPacket.handle(ServerboundChatPacket.java:6) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:36) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1340) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.shouldRun(MinecraftServer.java:193) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1319) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1312) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1273) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1184) ~[patched_1.17.jar:git-Paper-69]
at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:320) ~[patched_1.17.jar:git-Paper-69] at java.lang.Thread.run(Thread.java:831) [?:?]
Caused by: java.lang.ArrayIndexOutOfBoundsException: Index 0 out of bounds for length 0
at Latch.Money4Mobs.MkCommand.onCommand(MkCommand.java:764) ~[?:?]
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:45) ~[patched_1.17.jar:git-Paper-69]
... 21 more`
```
### Server Information
Paper 1.17 build 69 with the following plugins:
- 1.17---Money4Mobs-v1.6.4
- iEconomy Release 1.1
- LuckPerms Bukkit-5.3.48
- Vault
| priority | using mk with no parameters throws an error in the console there appears to be no handler for mk on its own resulting on an arrayindexoutofboundsexception null org bukkit command commandexception unhandled exception executing command mk in plugin at org bukkit command plugincommand execute plugincommand java at org bukkit command simplecommandmap dispatch simplecommandmap java at org bukkit craftbukkit craftserver dispatchcommand craftserver java at net minecraft server network servergamepacketlistenerimpl handlecommand servergamepacketlistenerimpl java at net minecraft server network servergamepacketlistenerimpl handlechat servergamepacketlistenerimpl java at net minecraft server network servergamepacketlistenerimpl handlechat servergamepacketlistenerimpl java at net minecraft network protocol game serverboundchatpacket handle serverboundchatpacket java at net minecraft network protocol game serverboundchatpacket handle serverboundchatpacket java at net minecraft network protocol packetutils lambda ensurerunningonsamethread packetutils java at net minecraft server ticktask run ticktask java at net minecraft util thread blockableeventloop doruntask blockableeventloop java at net minecraft util thread reentrantblockableeventloop doruntask reentrantblockableeventloop java at net minecraft server minecraftserver doruntask minecraftserver java at net minecraft server minecraftserver shouldrun minecraftserver java at net minecraft util thread blockableeventloop polltask blockableeventloop java at net minecraft server minecraftserver polltaskinternal minecraftserver java at net minecraft server minecraftserver polltask minecraftserver java at net minecraft util thread blockableeventloop managedblock blockableeventloop java at net minecraft server minecraftserver waituntilnexttick minecraftserver java at net minecraft server minecraftserver runserver minecraftserver java at net minecraft server minecraftserver lambda spin minecraftserver java at java lang thread run thread java caused by java lang arrayindexoutofboundsexception index out of bounds for length at latch mkcommand oncommand mkcommand java at org bukkit command plugincommand execute plugincommand java more server information paper build with the following plugins ieconomy release luckperms bukkit vault | 1 |
83,449 | 3,634,984,495 | IssuesEvent | 2016-02-11 20:02:35 | ELVIS-Project/elvis-database | https://api.github.com/repos/ELVIS-Project/elvis-database | opened | 1-Click Dev Deployments | Priority: HIGH Status: IN PROGRESS Type: ENHANCEMENT | I'm writing a script to automate `dev` deployments. This technique will make it much easier to deploy builds.
The script clones the dev branch, sets up the virtual environment with requirements. The current build is moved to a back-up location.
Later on, we will adapt this script for 1-Click Prod deployments. | 1.0 | 1-Click Dev Deployments - I'm writing a script to automate `dev` deployments. This technique will make it much easier to deploy builds.
The script clones the dev branch, sets up the virtual environment with requirements. The current build is moved to a back-up location.
Later on, we will adapt this script for 1-Click Prod deployments. | priority | click dev deployments i m writing a script to automate dev deployments this technique will make it much easier to deploy builds the script clones the dev branch sets up the virtual environment with requirements the current build is moved to a back up location later on we will adapt this script for click prod deployments | 1 |
258,084 | 8,154,424,794 | IssuesEvent | 2018-08-23 03:12:03 | WordImpress/wp-business-reviews | https://api.github.com/repos/WordImpress/wp-business-reviews | opened | compat(collection): ensure collections render in Safari 10.X.X | high-priority | ## User Story
As a Safari user, I want collections to work in Safari 10.X.X so I can see my reviews.
## Current Behavior
Reviews do not appear in Safari 10.1.2 due to a console error related to JS uglification:
> SyntaxError: Cannot declare a let variable twice: 'e'.
The issue does not exist in Safari 11+.
## Expected Behavior
Collections render without console errors.
## Acceptance Criteria
- [ ] A collection renders in Safari 10.1.2 (confirm using BrowserStack).
- [ ] The [new Demos page](https://wpbusinessreviews.com/demos/) works in Safari 10.1.2.
## Possible Solution
The error and fix are documented in this [GitHub comment](https://github.com/mishoo/UglifyJS2/issues/1753#issuecomment-324814782). | 1.0 | compat(collection): ensure collections render in Safari 10.X.X - ## User Story
As a Safari user, I want collections to work in Safari 10.X.X so I can see my reviews.
## Current Behavior
Reviews do not appear in Safari 10.1.2 due to a console error related to JS uglification:
> SyntaxError: Cannot declare a let variable twice: 'e'.
The issue does not exist in Safari 11+.
## Expected Behavior
Collections render without console errors.
## Acceptance Criteria
- [ ] A collection renders in Safari 10.1.2 (confirm using BrowserStack).
- [ ] The [new Demos page](https://wpbusinessreviews.com/demos/) works in Safari 10.1.2.
## Possible Solution
The error and fix are documented in this [GitHub comment](https://github.com/mishoo/UglifyJS2/issues/1753#issuecomment-324814782). | priority | compat collection ensure collections render in safari x x user story as a safari user i want collections to work in safari x x so i can see my reviews current behavior reviews do not appear in safari due to a console error related to js uglification syntaxerror cannot declare a let variable twice e the issue does not exist in safari expected behavior collections render without console errors acceptance criteria a collection renders in safari confirm using browserstack the works in safari possible solution the error and fix are documented in this | 1 |
406,781 | 11,902,956,084 | IssuesEvent | 2020-03-30 14:40:42 | sous-chefs/java | https://api.github.com/repos/sous-chefs/java | closed | Add new attr for Homebrew Cask name | Feature Request Priority: High | Because Oracle likes to hide things behind login wall, naturally push is to go towards `adoptopenjdk` , this should extend to macOS + Homebrew.
At present Homebrew is hardcode to `java` in https://github.com/sous-chefs/java/blob/master/recipes/homebrew.rb#L6
Propose a new attribute to allow for passing cask name like `adoptopenjdk`.
Will do up a PR, but looking for idea/input. | 1.0 | Add new attr for Homebrew Cask name - Because Oracle likes to hide things behind login wall, naturally push is to go towards `adoptopenjdk` , this should extend to macOS + Homebrew.
At present Homebrew is hardcode to `java` in https://github.com/sous-chefs/java/blob/master/recipes/homebrew.rb#L6
Propose a new attribute to allow for passing cask name like `adoptopenjdk`.
Will do up a PR, but looking for idea/input. | priority | add new attr for homebrew cask name because oracle likes to hide things behind login wall naturally push is to go towards adoptopenjdk this should extend to macos homebrew at present homebrew is hardcode to java in propose a new attribute to allow for passing cask name like adoptopenjdk will do up a pr but looking for idea input | 1 |
658,193 | 21,880,274,778 | IssuesEvent | 2022-05-19 13:48:50 | bounswe/bounswe2022group8 | https://api.github.com/repos/bounswe/bounswe2022group8 | closed | Practice App: Feature-Comment | Effort: High Priority: High Status: in progress practice app | ### What's up?
Endpoints get,delete,post for comment on an art item will be implemented after the implementations of serializers and database are completed.
| API | API Method | API Description |
| :----: | :----: | :----: |
| ```api/v1/comments/artitem/<int:id>``` | ```GET``` | Get all of the comments of the specific art item (by id) |
| ```api/v1/comments/artitem/<int:id>``` | ```POST``` | Post comment for the specific art item (by id) |
| ```api/v1/user/<int:id>/comment/<int:commentid>``` | ```DELETE``` | Delete the comment by commentid posted by the user having id |
### To Do
- [x] Create new branch 'feature/comment' based on practice_app
- [x] Create endpoints to create, view, delete comments
- [x] Write unit tests
- [x] Create html page in templates
- [x] Create pull request and assign @serdarakol as reviewer as stated in [meeting notes #14](https://github.com/bounswe/bounswe2022group8/wiki/Week-10-Meeting-%2314)
### Deadline
15.05.2022 @20.00
### Additional Information
_No response_
### Reviewers
_No response_ | 1.0 | Practice App: Feature-Comment - ### What's up?
Endpoints get,delete,post for comment on an art item will be implemented after the implementations of serializers and database are completed.
| API | API Method | API Description |
| :----: | :----: | :----: |
| ```api/v1/comments/artitem/<int:id>``` | ```GET``` | Get all of the comments of the specific art item (by id) |
| ```api/v1/comments/artitem/<int:id>``` | ```POST``` | Post comment for the specific art item (by id) |
| ```api/v1/user/<int:id>/comment/<int:commentid>``` | ```DELETE``` | Delete the comment by commentid posted by the user having id |
### To Do
- [x] Create new branch 'feature/comment' based on practice_app
- [x] Create endpoints to create, view, delete comments
- [x] Write unit tests
- [x] Create html page in templates
- [x] Create pull request and assign @serdarakol as reviewer as stated in [meeting notes #14](https://github.com/bounswe/bounswe2022group8/wiki/Week-10-Meeting-%2314)
### Deadline
15.05.2022 @20.00
### Additional Information
_No response_
### Reviewers
_No response_ | priority | practice app feature comment what s up endpoints get delete post for comment on an art item will be implemented after the implementations of serializers and database are completed api api method api description api comments artitem get get all of the comments of the specific art item by id api comments artitem post post comment for the specific art item by id api user comment delete delete the comment by commentid posted by the user having id to do create new branch feature comment based on practice app create endpoints to create view delete comments write unit tests create html page in templates create pull request and assign serdarakol as reviewer as stated in deadline additional information no response reviewers no response | 1 |
365,785 | 10,797,729,419 | IssuesEvent | 2019-11-06 08:34:01 | oceanprotocol/commons | https://api.github.com/repos/oceanprotocol/commons | opened | Can't parse description field in asset registration | bug help wanted priority:high | ## Current Behavior
The following description will cause errors and break the registration flow;
```
Line 1
Line 2
```
Error trace (but just see for yourself)
```
Response message:
[{"message":"'Line 1\\n\\nLine 2' does not match '^(.*)$'","path":"base/description"}]
Error fetching querying metadata:
Response {type: "cors", url: "https://aquarius.commons.oceanprotocol.com/api/v1/aquarius/assets/ddo", redirected: false, status: 400, ok: false, …}
Logger.js:101
DDO stored
Logger.js:101 error: Cannot read property 'id' of null
```
This occurs in BMW, Commons, Nile, ...!
### Possible Solution
Plecos? Try the /validate endpoint in Aquarius service?
| 1.0 | Can't parse description field in asset registration - ## Current Behavior
The following description will cause errors and break the registration flow;
```
Line 1
Line 2
```
Error trace (but just see for yourself)
```
Response message:
[{"message":"'Line 1\\n\\nLine 2' does not match '^(.*)$'","path":"base/description"}]
Error fetching querying metadata:
Response {type: "cors", url: "https://aquarius.commons.oceanprotocol.com/api/v1/aquarius/assets/ddo", redirected: false, status: 400, ok: false, …}
Logger.js:101
DDO stored
Logger.js:101 error: Cannot read property 'id' of null
```
This occurs in BMW, Commons, Nile, ...!
### Possible Solution
Plecos? Try the /validate endpoint in Aquarius service?
| priority | can t parse description field in asset registration current behavior the following description will cause errors and break the registration flow line line error trace but just see for yourself response message error fetching querying metadata response type cors url redirected false status ok false … logger js ddo stored logger js error cannot read property id of null this occurs in bmw commons nile possible solution plecos try the validate endpoint in aquarius service | 1 |
695,923 | 23,876,610,866 | IssuesEvent | 2022-09-07 19:44:05 | Igalia/wpe-android | https://api.github.com/repos/Igalia/wpe-android | closed | Architecture redesign | enhancement priority-high | Current architecture is hard to keep robust and scalable because of many not so flexible approaches in current code.
For example:
All java -> c++ calls are doing via one Browser singleton and glue entity. This entry points keeps growing in size when we implement new features.
This task is to investigate better approaches and do design overhaul.
- Fix native access model (relates to #104 )
- Ensure thread safety and clean up threading model
- More straightforward 1:1 relationship between java and c++
- Reformat whole codebase using common code style and convetions | 1.0 | Architecture redesign - Current architecture is hard to keep robust and scalable because of many not so flexible approaches in current code.
For example:
All java -> c++ calls are doing via one Browser singleton and glue entity. This entry points keeps growing in size when we implement new features.
This task is to investigate better approaches and do design overhaul.
- Fix native access model (relates to #104 )
- Ensure thread safety and clean up threading model
- More straightforward 1:1 relationship between java and c++
- Reformat whole codebase using common code style and convetions | priority | architecture redesign current architecture is hard to keep robust and scalable because of many not so flexible approaches in current code for example all java c calls are doing via one browser singleton and glue entity this entry points keeps growing in size when we implement new features this task is to investigate better approaches and do design overhaul fix native access model relates to ensure thread safety and clean up threading model more straightforward relationship between java and c reformat whole codebase using common code style and convetions | 1 |
313,165 | 9,557,808,027 | IssuesEvent | 2019-05-03 12:42:03 | wso2/product-ei | https://api.github.com/repos/wso2/product-ei | opened | Update MicroIntegrator documentation for EI 6.5.0 | 6.5.0 Priority/Highest Type/Docs | The MicroIntegrator profile is moved out of the EI 6.5.0 distribution. This profile is going to be introduced as an independent distribution.
Documentation needs to be updated on all the related changes for the Micro Integrator profile.
[1] https://docs.wso2.com/display/EI6xx/Working+with+the+Micro+Integrator | 1.0 | Update MicroIntegrator documentation for EI 6.5.0 - The MicroIntegrator profile is moved out of the EI 6.5.0 distribution. This profile is going to be introduced as an independent distribution.
Documentation needs to be updated on all the related changes for the Micro Integrator profile.
[1] https://docs.wso2.com/display/EI6xx/Working+with+the+Micro+Integrator | priority | update microintegrator documentation for ei the microintegrator profile is moved out of the ei distribution this profile is going to be introduced as an independent distribution documentation needs to be updated on all the related changes for the micro integrator profile | 1 |
589,451 | 17,702,806,207 | IssuesEvent | 2021-08-25 01:37:50 | axinc-ai/ailia-models | https://api.github.com/repos/axinc-ai/ailia-models | closed | ADD mediapipe objectron | high priority | We want to add shoe objectron
https://ai.googleblog.com/2020/03/real-time-3d-object-detection-on-mobile.html
https://github.com/google-research-datasets/Objectron
Model conversion example (PintoModelZoo)
https://github.com/PINTO0309/PINTO_model_zoo/tree/main/036_Objectron
Runtime example (OpenVINO)
https://github.com/yas-sim/objectron-3d-object-detection-openvino | 1.0 | ADD mediapipe objectron - We want to add shoe objectron
https://ai.googleblog.com/2020/03/real-time-3d-object-detection-on-mobile.html
https://github.com/google-research-datasets/Objectron
Model conversion example (PintoModelZoo)
https://github.com/PINTO0309/PINTO_model_zoo/tree/main/036_Objectron
Runtime example (OpenVINO)
https://github.com/yas-sim/objectron-3d-object-detection-openvino | priority | add mediapipe objectron we want to add shoe objectron model conversion example pintomodelzoo runtime example openvino | 1 |
89,954 | 3,807,221,586 | IssuesEvent | 2016-03-25 06:23:42 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [FVT]: Hard disk boot failed on NextScale SLES11 SP4 after diskful installation | component:os_provision priority:high status:pending type:bug | Test OS: SLES11 SP4
Test xCAT version: xCAT Version 2.10 (git commit 79c7f6949c09e0b49875f10ede4b75b18ad4a0e7, built Fri Jul 24 05:27:07 EDT 2015)
How to reproduce:
1. Define NextScale node as a compute node
2. Provision SLES11 SP4 on NextScale node
3. reboot NextScale node and find the boot failed.
Here is the output from the console
```
System initializing
System initializing memory
System initializing...
System initializing
System initializing memory
System initializing...
Loading value-add drivers
Scanning system, connecting boot device(s)
<F1> Setup<F2> Diagnostics<F12> Select Boot Device
No key pressed. Preparing to boot normally...Optimized Boot ... 11
Boot Failed - SUSE Linux Enterprise Server 11 SP4: HD(1,GPT,E2A6F59C-6C93-4794-9B46-9FB1623565C0,0x800,0x42000)/\efi\SuSE\elilo.efi
Boot Failed - SUSE Linux Enterprise Server 11 SP4: HD(1,GPT,F245DF20-35F3-4B81-960C-9FFA6E2A6838,0x800,0x42000)/\efi\SuSE\elilo.efi
Boot Failed - PXE Network: PciRoot(0x0)/Pci(0x2,0x0)/Pci(0x0,0x0)/Ctrl(0x1)/MAC(E41D2D796471,0x0)/IPv4(0.0.0.0,0x0,DHCP,0.0.0.0,0.0.0.0,0.0.0.0)
Boot Failed - PXE Network: PciRoot(0x0)/Pci(0x2,0x0)/Pci(0x0,0x0)/Ctrl(0x2)/MAC(E41D2D796472,0x0)/IPv4(0.0.0.0,0x0,DHCP,0.0.0.0,0.0.0.0,0.0.0.0)
``` | 1.0 | [FVT]: Hard disk boot failed on NextScale SLES11 SP4 after diskful installation - Test OS: SLES11 SP4
Test xCAT version: xCAT Version 2.10 (git commit 79c7f6949c09e0b49875f10ede4b75b18ad4a0e7, built Fri Jul 24 05:27:07 EDT 2015)
How to reproduce:
1. Define NextScale node as a compute node
2. Provision SLES11 SP4 on NextScale node
3. reboot NextScale node and find the boot failed.
Here is the output from the console
```
System initializing
System initializing memory
System initializing...
System initializing
System initializing memory
System initializing...
Loading value-add drivers
Scanning system, connecting boot device(s)
<F1> Setup<F2> Diagnostics<F12> Select Boot Device
No key pressed. Preparing to boot normally...Optimized Boot ... 11
Boot Failed - SUSE Linux Enterprise Server 11 SP4: HD(1,GPT,E2A6F59C-6C93-4794-9B46-9FB1623565C0,0x800,0x42000)/\efi\SuSE\elilo.efi
Boot Failed - SUSE Linux Enterprise Server 11 SP4: HD(1,GPT,F245DF20-35F3-4B81-960C-9FFA6E2A6838,0x800,0x42000)/\efi\SuSE\elilo.efi
Boot Failed - PXE Network: PciRoot(0x0)/Pci(0x2,0x0)/Pci(0x0,0x0)/Ctrl(0x1)/MAC(E41D2D796471,0x0)/IPv4(0.0.0.0,0x0,DHCP,0.0.0.0,0.0.0.0,0.0.0.0)
Boot Failed - PXE Network: PciRoot(0x0)/Pci(0x2,0x0)/Pci(0x0,0x0)/Ctrl(0x2)/MAC(E41D2D796472,0x0)/IPv4(0.0.0.0,0x0,DHCP,0.0.0.0,0.0.0.0,0.0.0.0)
``` | priority | hard disk boot failed on nextscale after diskful installation test os test xcat version xcat version git commit built fri jul edt how to reproduce define nextscale node as a compute node provision on nextscale node reboot nextscale node and find the boot failed here is the output from the console system initializing system initializing memory system initializing system initializing system initializing memory system initializing loading value add drivers scanning system connecting boot device s setup diagnostics select boot device no key pressed preparing to boot normally optimized boot boot failed suse linux enterprise server hd gpt efi suse elilo efi boot failed suse linux enterprise server hd gpt efi suse elilo efi boot failed pxe network pciroot pci pci ctrl mac dhcp boot failed pxe network pciroot pci pci ctrl mac dhcp | 1 |
787,650 | 27,725,810,382 | IssuesEvent | 2023-03-15 02:03:02 | AY2223S2-CS2103T-W15-1/tp | https://api.github.com/repos/AY2223S2-CS2103T-W15-1/tp | closed | As a student, I want to be able to add a new contact without tags | type.Story priority.High | ...so I can add people who don't have a common CCA or module with me | 1.0 | As a student, I want to be able to add a new contact without tags - ...so I can add people who don't have a common CCA or module with me | priority | as a student i want to be able to add a new contact without tags so i can add people who don t have a common cca or module with me | 1 |
558,305 | 16,528,980,204 | IssuesEvent | 2021-05-27 01:36:31 | kobotoolbox/formpack | https://api.github.com/repos/kobotoolbox/formpack | closed | Handle pyxform v1.5.0 features | enhancement high priority | - Step 1: ensure exports succeed with new pxform features
- Step 2: integrate features properly into exports
| 1.0 | Handle pyxform v1.5.0 features - - Step 1: ensure exports succeed with new pxform features
- Step 2: integrate features properly into exports
| priority | handle pyxform features step ensure exports succeed with new pxform features step integrate features properly into exports | 1 |
524,115 | 15,196,594,022 | IssuesEvent | 2021-02-16 08:31:38 | wso2/streaming-integrator-tooling | https://api.github.com/repos/wso2/streaming-integrator-tooling | closed | Export to Kubernetes - Outdated Docker images are referred | Priority/High Severity/Critical Type/Bug | **Description:**
In the Export to Kubernetes feature, outdated Docker images are referred in the artifacts. This needs to be fixed
**Affected Product Version:**
SI Tooling 1.1.0 | 1.0 | Export to Kubernetes - Outdated Docker images are referred - **Description:**
In the Export to Kubernetes feature, outdated Docker images are referred in the artifacts. This needs to be fixed
**Affected Product Version:**
SI Tooling 1.1.0 | priority | export to kubernetes outdated docker images are referred description in the export to kubernetes feature outdated docker images are referred in the artifacts this needs to be fixed affected product version si tooling | 1 |
812,634 | 30,345,467,097 | IssuesEvent | 2023-07-11 15:09:30 | calcom/cal.com | https://api.github.com/repos/calcom/cal.com | opened | Events and emails are being sent to an old, disconnected Google account | 🐛 bug High priority calendar-apps | Found a bug? Please fill out the sections below. 👍
### Issue Summary
I recently was asked if I was joining a call that someone had scheduled with me and I noticed the event was not put onto the calendar I have linked (keith@cal.com) but instead it was sent to an old, disconnected Google account that I had already completely removed from my Cal.com account.
### Steps to Reproduce
1. Link 1 Google account and set that as the default calendar for where events are added
2. Link another Google account
3. Delete the first Google account added and make sure the 2nd one is now the default.
4. Submit bookings to your account.
### Actual Results
- I'm getting Google calendar events and emails to an old, disconnected Google account instead of the one that is set in app.cal.com (it's the only one even connected).
### Expected Results
- To get events on the calendar that is linked and set as default/event types
### Evidence
<img width="791" alt="Screen Shot 2023-07-11 at 4 29 51 PM" src="https://github.com/calcom/cal.com/assets/2538462/9b757e62-33f4-4af8-b281-d2f936214eab">
<img width="1415" alt="Screen Shot 2023-07-11 at 5 04 38 PM" src="https://github.com/calcom/cal.com/assets/2538462/6b10928b-e411-4b62-be99-9b9c225e4d59">
| 1.0 | Events and emails are being sent to an old, disconnected Google account - Found a bug? Please fill out the sections below. 👍
### Issue Summary
I recently was asked if I was joining a call that someone had scheduled with me and I noticed the event was not put onto the calendar I have linked (keith@cal.com) but instead it was sent to an old, disconnected Google account that I had already completely removed from my Cal.com account.
### Steps to Reproduce
1. Link 1 Google account and set that as the default calendar for where events are added
2. Link another Google account
3. Delete the first Google account added and make sure the 2nd one is now the default.
4. Submit bookings to your account.
### Actual Results
- I'm getting Google calendar events and emails to an old, disconnected Google account instead of the one that is set in app.cal.com (it's the only one even connected).
### Expected Results
- To get events on the calendar that is linked and set as default/event types
### Evidence
<img width="791" alt="Screen Shot 2023-07-11 at 4 29 51 PM" src="https://github.com/calcom/cal.com/assets/2538462/9b757e62-33f4-4af8-b281-d2f936214eab">
<img width="1415" alt="Screen Shot 2023-07-11 at 5 04 38 PM" src="https://github.com/calcom/cal.com/assets/2538462/6b10928b-e411-4b62-be99-9b9c225e4d59">
| priority | events and emails are being sent to an old disconnected google account found a bug please fill out the sections below 👍 issue summary i recently was asked if i was joining a call that someone had scheduled with me and i noticed the event was not put onto the calendar i have linked keith cal com but instead it was sent to an old disconnected google account that i had already completely removed from my cal com account steps to reproduce link google account and set that as the default calendar for where events are added link another google account delete the first google account added and make sure the one is now the default submit bookings to your account actual results i m getting google calendar events and emails to an old disconnected google account instead of the one that is set in app cal com it s the only one even connected expected results to get events on the calendar that is linked and set as default event types evidence img width alt screen shot at pm src img width alt screen shot at pm src | 1 |
671,099 | 22,743,156,628 | IssuesEvent | 2022-07-07 06:41:11 | Elice-SW-2-Team14/Animal-Hospital | https://api.github.com/repos/Elice-SW-2-Team14/Animal-Hospital | closed | [FE] 기초 셋팅 ( 모듈 , 전역 스타일 ) | 🔩 setup ❗️high-priority 🖥 Frontend | ## 🔨 기능 설명
기초 셋팅 ( 모듈 , 전역 스타일 )
## 📑 완료 조건
에러 없이 실행
## 💭 관련 백로그
[[FE] 초기 셋팅]-[프론트엔드 초기 셋팅]-[전역 스타일 설정]
## 💭 예상 작업 시간
(작업 시간)0.5h
| 1.0 | [FE] 기초 셋팅 ( 모듈 , 전역 스타일 ) - ## 🔨 기능 설명
기초 셋팅 ( 모듈 , 전역 스타일 )
## 📑 완료 조건
에러 없이 실행
## 💭 관련 백로그
[[FE] 초기 셋팅]-[프론트엔드 초기 셋팅]-[전역 스타일 설정]
## 💭 예상 작업 시간
(작업 시간)0.5h
| priority | 기초 셋팅 모듈 전역 스타일 🔨 기능 설명 기초 셋팅 모듈 전역 스타일 📑 완료 조건 에러 없이 실행 💭 관련 백로그 초기 셋팅 💭 예상 작업 시간 작업 시간 | 1 |
240,241 | 7,800,708,545 | IssuesEvent | 2018-06-09 12:46:36 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0011034:
Uncaught TypeError when saving items | Bug Calendar Mantis high priority | **Reported by shochdoerfer on 13 May 2015 20:53**
**Version:** Koriander (2014.09.10)
When saving an item (e.g. add an event in the calendar, add a new timeaccout) I get the following JS error:
Uncaught TypeError: Cannot read property 'ext-comp-1225' of undefined
ext-all.js:7 Ext.Element
ext-all.js:7 h.get
ext-all.js:7 Ext.extend.onRender
ext-all.js:7 Ext.Panel.Ext.extend.onRender
ext-all.js:14 Ext.Window.Ext.extend.onRender
ext-all.js:7 Ext.extend.render
ext-all.js:14 Ext.MessageBox.getDialog
ext-all.js:14 Ext.MessageBox.show
ext-all.js:14 Ext.MessageBox.alert
index.php?method=Tinebase.getJsFiles:37 (anonymous function)
**Additional information:** I made sure that I cleared the browser cache so this should not be an issue.
| 1.0 | 0011034:
Uncaught TypeError when saving items - **Reported by shochdoerfer on 13 May 2015 20:53**
**Version:** Koriander (2014.09.10)
When saving an item (e.g. add an event in the calendar, add a new timeaccout) I get the following JS error:
Uncaught TypeError: Cannot read property 'ext-comp-1225' of undefined
ext-all.js:7 Ext.Element
ext-all.js:7 h.get
ext-all.js:7 Ext.extend.onRender
ext-all.js:7 Ext.Panel.Ext.extend.onRender
ext-all.js:14 Ext.Window.Ext.extend.onRender
ext-all.js:7 Ext.extend.render
ext-all.js:14 Ext.MessageBox.getDialog
ext-all.js:14 Ext.MessageBox.show
ext-all.js:14 Ext.MessageBox.alert
index.php?method=Tinebase.getJsFiles:37 (anonymous function)
**Additional information:** I made sure that I cleared the browser cache so this should not be an issue.
| priority | uncaught typeerror when saving items reported by shochdoerfer on may version koriander when saving an item e g add an event in the calendar add a new timeaccout i get the following js error uncaught typeerror cannot read property ext comp of undefined ext all js ext element ext all js h get ext all js ext extend onrender ext all js ext panel ext extend onrender ext all js ext window ext extend onrender ext all js ext extend render ext all js ext messagebox getdialog ext all js ext messagebox show ext all js ext messagebox alert index php method tinebase getjsfiles anonymous function additional information i made sure that i cleared the browser cache so this should not be an issue | 1 |
623,482 | 19,669,514,164 | IssuesEvent | 2022-01-11 04:47:00 | ballerina-platform/openapi-tools | https://api.github.com/repos/ballerina-platform/openapi-tools | opened | Add support to Typed headers for HTTP response | Type/Improvement Points/2 Priority/High BallerinaToOpenAPI | **Description:**
Related issue : https://github.com/ballerina-platform/ballerina-standard-library/issues/2562
```ballerina
public type RateLimitHeaders record {|
string x\-rate\-limit\-id;
int x\-rate\-limit\-remaining;
string[] x\-rate\-limit\-types;
|};
public type OkWithRateLmits record {|
*Ok;
RateLimitHeaders headers;
string body;
|};
service / on new http:Listener(9090) {
resource function get status() returns OkWithRateLmits {
OkWithRateLmits okWithRateLmits = {
headers: {
x\-rate\-limit\-id: "1xed",
x\-rate\-limit\-remaining: 3,
x\-rate\-limit\-types: ["sliver", "gold"]
},
body: "full"
};
return okWithRateLmits;
}
}
```
OAS mapping
```openapi
paths:
/ping:
get:
summary: Checks if the server is alive.
responses:
'200':
description: OK
headers:
x-rate-limit-id:
schema:
type: string
x-rate-limit-remaining:
schema:
type: integer
x-rate-limit-types:
schema:
type: array
items:
type: string
content:
test/plain:
schema:
type: string
```
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Add support to Typed headers for HTTP response - **Description:**
Related issue : https://github.com/ballerina-platform/ballerina-standard-library/issues/2562
```ballerina
public type RateLimitHeaders record {|
string x\-rate\-limit\-id;
int x\-rate\-limit\-remaining;
string[] x\-rate\-limit\-types;
|};
public type OkWithRateLmits record {|
*Ok;
RateLimitHeaders headers;
string body;
|};
service / on new http:Listener(9090) {
resource function get status() returns OkWithRateLmits {
OkWithRateLmits okWithRateLmits = {
headers: {
x\-rate\-limit\-id: "1xed",
x\-rate\-limit\-remaining: 3,
x\-rate\-limit\-types: ["sliver", "gold"]
},
body: "full"
};
return okWithRateLmits;
}
}
```
OAS mapping
```openapi
paths:
/ping:
get:
summary: Checks if the server is alive.
responses:
'200':
description: OK
headers:
x-rate-limit-id:
schema:
type: string
x-rate-limit-remaining:
schema:
type: integer
x-rate-limit-types:
schema:
type: array
items:
type: string
content:
test/plain:
schema:
type: string
```
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | add support to typed headers for http response description related issue ballerina public type ratelimitheaders record string x rate limit id int x rate limit remaining string x rate limit types public type okwithratelmits record ok ratelimitheaders headers string body service on new http listener resource function get status returns okwithratelmits okwithratelmits okwithratelmits headers x rate limit id x rate limit remaining x rate limit types body full return okwithratelmits oas mapping openapi paths ping get summary checks if the server is alive responses description ok headers x rate limit id schema type string x rate limit remaining schema type integer x rate limit types schema type array items type string content test plain schema type string describe your solution s related issues optional suggested labels optional suggested assignees optional | 1 |
421,930 | 12,263,568,264 | IssuesEvent | 2020-05-07 01:26:24 | findthemasks/findthemasks | https://api.github.com/repos/findthemasks/findthemasks | closed | [v2-design] Carry v2 design flourish to embed page - Designs DEV READY on Figma | help wanted high-priority | give.html should look similar to index.html with the caveat that since this page is being embedded on other sites, it should deal with smaller window sizes. Getusppe restricts height to 750px.
Some features we need to continue to support
- [x] any of the embed query params (turn filters on/off, center map to location, center map to state, etc
Test on the following sites:
- [x] https://getusppe.org/give/ (embed is in middle of page)
- [x] https://theppedrive.com/ppe-drive-listing
- [x] https://resource19.org/pages/search.html
- [x] https://lafightingcovid.com/ | 1.0 | [v2-design] Carry v2 design flourish to embed page - Designs DEV READY on Figma - give.html should look similar to index.html with the caveat that since this page is being embedded on other sites, it should deal with smaller window sizes. Getusppe restricts height to 750px.
Some features we need to continue to support
- [x] any of the embed query params (turn filters on/off, center map to location, center map to state, etc
Test on the following sites:
- [x] https://getusppe.org/give/ (embed is in middle of page)
- [x] https://theppedrive.com/ppe-drive-listing
- [x] https://resource19.org/pages/search.html
- [x] https://lafightingcovid.com/ | priority | carry design flourish to embed page designs dev ready on figma give html should look similar to index html with the caveat that since this page is being embedded on other sites it should deal with smaller window sizes getusppe restricts height to some features we need to continue to support any of the embed query params turn filters on off center map to location center map to state etc test on the following sites embed is in middle of page | 1 |
280,913 | 8,688,214,506 | IssuesEvent | 2018-12-03 15:37:10 | jaedan/Outlands | https://api.github.com/repos/jaedan/Outlands | opened | Treasure maps are far too lucrative & buggy | Priority: High bug | Tmaps are broken and are flooding the market with gold and scrolls with little to no risk .. (this is coming from an avid tmaper.. speaking up for the greater good of the community)
imho recommendation for fixing: lower the gold by 50%, lower the drop chance of scrolls by 50%, add in some new uber rare deco items from boxes to make them worth doing.. right now level 1s and 2s are not worth the time to find/do, but 3s,4s,5s, are $$
or some unique cloth hues or something idk, just offering an idea
also, theres one other issue not sure if anyone is exploiting because i just found it last night.. if you're doing a high level tmap, you can lure all the spawn away and if decent enough distance, the mobs will spawn again, not sure if this is exploitable but might be a source of high level creature farm with high flow rate | 1.0 | Treasure maps are far too lucrative & buggy - Tmaps are broken and are flooding the market with gold and scrolls with little to no risk .. (this is coming from an avid tmaper.. speaking up for the greater good of the community)
imho recommendation for fixing: lower the gold by 50%, lower the drop chance of scrolls by 50%, add in some new uber rare deco items from boxes to make them worth doing.. right now level 1s and 2s are not worth the time to find/do, but 3s,4s,5s, are $$
or some unique cloth hues or something idk, just offering an idea
also, theres one other issue not sure if anyone is exploiting because i just found it last night.. if you're doing a high level tmap, you can lure all the spawn away and if decent enough distance, the mobs will spawn again, not sure if this is exploitable but might be a source of high level creature farm with high flow rate | priority | treasure maps are far too lucrative buggy tmaps are broken and are flooding the market with gold and scrolls with little to no risk this is coming from an avid tmaper speaking up for the greater good of the community imho recommendation for fixing lower the gold by lower the drop chance of scrolls by add in some new uber rare deco items from boxes to make them worth doing right now level and are not worth the time to find do but are or some unique cloth hues or something idk just offering an idea also theres one other issue not sure if anyone is exploiting because i just found it last night if you re doing a high level tmap you can lure all the spawn away and if decent enough distance the mobs will spawn again not sure if this is exploitable but might be a source of high level creature farm with high flow rate | 1 |
766,529 | 26,887,070,190 | IssuesEvent | 2023-02-06 04:51:56 | virtualcell/vcell | https://api.github.com/repos/virtualcell/vcell | closed | renamed reaction parameters revert to predefined names on save | bug High Priority VCell-7.5.0 | Any built-in kinetic parameter that is being renamed looses the new name upon save. | 1.0 | renamed reaction parameters revert to predefined names on save - Any built-in kinetic parameter that is being renamed looses the new name upon save. | priority | renamed reaction parameters revert to predefined names on save any built in kinetic parameter that is being renamed looses the new name upon save | 1 |
246,853 | 7,895,809,514 | IssuesEvent | 2018-06-29 05:53:12 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | subset plot wireframe shows domains, instead of blocks | Likelihood: 3 - Occasional OS: All Priority: High Severity: 2 - Minor Irritation Support Group: Any bug version: 2.7.1 | John Hayes was creating a subset plot of blocks, (not domains) and wanted to see the outlines of the blocks using the wireframe option.
When he enabled this, instead of seeing the block outlines, he saw all of the domains outlined.
Cyrus has access to an example data set, but it is quite large. Cyrus will get a smaller reproducer & we can decide the path forward from there.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 03/31/2014 03:50 pm
Original update: 11/19/2014 03:33 pm
Ticket number: 1791 | 1.0 | subset plot wireframe shows domains, instead of blocks - John Hayes was creating a subset plot of blocks, (not domains) and wanted to see the outlines of the blocks using the wireframe option.
When he enabled this, instead of seeing the block outlines, he saw all of the domains outlined.
Cyrus has access to an example data set, but it is quite large. Cyrus will get a smaller reproducer & we can decide the path forward from there.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 03/31/2014 03:50 pm
Original update: 11/19/2014 03:33 pm
Ticket number: 1791 | priority | subset plot wireframe shows domains instead of blocks john hayes was creating a subset plot of blocks not domains and wanted to see the outlines of the blocks using the wireframe option when he enabled this instead of seeing the block outlines he saw all of the domains outlined cyrus has access to an example data set but it is quite large cyrus will get a smaller reproducer we can decide the path forward from there redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number | 1 |
452,471 | 13,050,993,058 | IssuesEvent | 2020-07-29 16:20:49 | localstack/localstack | https://api.github.com/repos/localstack/localstack | closed | API Gateway Error: HTTP_PROXY GET not yet implemented | feature-missing priority-high | When setting up a API Gateway with HTTP_PROXY I get the following error message when accessing the route `http://localhost:4567/restapis/xaper8hype/test/_user_request_/test123`
```
{"message": "API Gateway integration type \"HTTP_PROXY\" for method \"GET\" not yet implemented"}
```
tf file
```
provider "aws" {
access_key = "mock_access_key"
region = "us-east-1"
s3_force_path_style = true
secret_key = "mock_secret_key"
skip_credentials_validation = true
skip_metadata_api_check = true
skip_requesting_account_id = true
endpoints {
apigateway = "http://localhost:4567"
cloudformation = "http://localhost:4581"
cloudwatch = "http://localhost:4582"
dynamodb = "http://localhost:4569"
es = "http://localhost:4578"
firehose = "http://localhost:4573"
iam = "http://localhost:4593"
kinesis = "http://localhost:4568"
lambda = "http://localhost:4574"
route53 = "http://localhost:4580"
redshift = "http://localhost:4577"
s3 = "http://localhost:4572"
secretsmanager = "http://localhost:4584"
ses = "http://localhost:4579"
sns = "http://localhost:4575"
sqs = "http://localhost:4576"
ssm = "http://localhost:4583"
stepfunctions = "http://localhost:4585"
sts = "http://localhost:4592"
}
}
resource "aws_api_gateway_rest_api" "MyDemoAPI" {
name = "MyDemoAPI"
description = "This is my API for demonstration purposes"
}
resource "aws_api_gateway_resource" "MyDemoResource" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
parent_id = aws_api_gateway_rest_api.MyDemoAPI.root_resource_id
path_part = "test123"
}
resource "aws_api_gateway_method" "MyDemoMethod" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
resource_id = aws_api_gateway_resource.MyDemoResource.id
http_method = "GET"
authorization = "NONE"
request_parameters = {
"method.request.path.proxy" = true
}
}
resource "aws_api_gateway_integration" "MyDemoIntegration" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
resource_id = aws_api_gateway_resource.MyDemoResource.id
http_method = aws_api_gateway_method.MyDemoMethod.http_method
type = "HTTP_PROXY"
uri = "https://httpbin.org/anything/test123"
integration_http_method = "GET"
cache_key_parameters = ["method.request.path.proxy"]
timeout_milliseconds = 29000
request_parameters = {
"integration.request.path.proxy" = "method.request.path.proxy"
}
}
resource "aws_api_gateway_deployment" "teststage" {
depends_on = [
aws_api_gateway_integration.MyDemoIntegration
]
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
stage_name = "test"
}
output "base_url" {
value = aws_api_gateway_deployment.teststage.invoke_url
}
```
docker compose
```
version: '3'
services:
localstack:
container_name: "${LOCALSTACK_DOCKER_NAME-localstack_main}"
image: localstack/localstack:latest
ports:
- "4567-4597:4567-4597"
- "${PORT_WEB_UI-8080}:${PORT_WEB_UI-8080}"
environment:
- SERVICES=apigateway
- DEBUG=1
- DATA_DIR=${DATA_DIR- }
- PORT_WEB_UI=${PORT_WEB_UI- }
- LAMBDA_EXECUTOR=${LAMBDA_EXECUTOR- }
- KINESIS_ERROR_PROBABILITY=${KINESIS_ERROR_PROBABILITY- }
- DOCKER_HOST=unix:///var/run/docker.sock
volumes:
- "${TMPDIR:-/tmp/localstack}:/tmp/localstack"
- "/var/run/docker.sock:/var/run/docker.sock"
```
aws output
```
$ aws --endpoint=http://localhost:4567 --region us-east-1 apigateway get-resources --rest-api-id xaper8hype
{
"items": [
{
"id": "5t3u62jjyg",
"path": "/",
"resourceMethods": {
"GET": {}
}
},
{
"id": "fu3fph78jj",
"parentId": "5t3u62jjyg",
"pathPart": "test123",
"path": "/test123",
"resourceMethods": {
"GET": {
"httpMethod": "GET",
"authorizationType": "NONE",
"methodIntegration": {
"type": "HTTP_PROXY",
"httpMethod": "GET",
"uri": "https://httpbin.org/anything/test123",
"requestTemplates": {},
"integrationResponses": {
"200": {
"statusCode": 200,
"responseTemplates": {
"application/json": null
}
}
}
}
}
}
}
]
}
```
I'm not sure if I do something wrong but the terraform code does work on AWS. | 1.0 | API Gateway Error: HTTP_PROXY GET not yet implemented - When setting up a API Gateway with HTTP_PROXY I get the following error message when accessing the route `http://localhost:4567/restapis/xaper8hype/test/_user_request_/test123`
```
{"message": "API Gateway integration type \"HTTP_PROXY\" for method \"GET\" not yet implemented"}
```
tf file
```
provider "aws" {
access_key = "mock_access_key"
region = "us-east-1"
s3_force_path_style = true
secret_key = "mock_secret_key"
skip_credentials_validation = true
skip_metadata_api_check = true
skip_requesting_account_id = true
endpoints {
apigateway = "http://localhost:4567"
cloudformation = "http://localhost:4581"
cloudwatch = "http://localhost:4582"
dynamodb = "http://localhost:4569"
es = "http://localhost:4578"
firehose = "http://localhost:4573"
iam = "http://localhost:4593"
kinesis = "http://localhost:4568"
lambda = "http://localhost:4574"
route53 = "http://localhost:4580"
redshift = "http://localhost:4577"
s3 = "http://localhost:4572"
secretsmanager = "http://localhost:4584"
ses = "http://localhost:4579"
sns = "http://localhost:4575"
sqs = "http://localhost:4576"
ssm = "http://localhost:4583"
stepfunctions = "http://localhost:4585"
sts = "http://localhost:4592"
}
}
resource "aws_api_gateway_rest_api" "MyDemoAPI" {
name = "MyDemoAPI"
description = "This is my API for demonstration purposes"
}
resource "aws_api_gateway_resource" "MyDemoResource" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
parent_id = aws_api_gateway_rest_api.MyDemoAPI.root_resource_id
path_part = "test123"
}
resource "aws_api_gateway_method" "MyDemoMethod" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
resource_id = aws_api_gateway_resource.MyDemoResource.id
http_method = "GET"
authorization = "NONE"
request_parameters = {
"method.request.path.proxy" = true
}
}
resource "aws_api_gateway_integration" "MyDemoIntegration" {
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
resource_id = aws_api_gateway_resource.MyDemoResource.id
http_method = aws_api_gateway_method.MyDemoMethod.http_method
type = "HTTP_PROXY"
uri = "https://httpbin.org/anything/test123"
integration_http_method = "GET"
cache_key_parameters = ["method.request.path.proxy"]
timeout_milliseconds = 29000
request_parameters = {
"integration.request.path.proxy" = "method.request.path.proxy"
}
}
resource "aws_api_gateway_deployment" "teststage" {
depends_on = [
aws_api_gateway_integration.MyDemoIntegration
]
rest_api_id = aws_api_gateway_rest_api.MyDemoAPI.id
stage_name = "test"
}
output "base_url" {
value = aws_api_gateway_deployment.teststage.invoke_url
}
```
docker compose
```
version: '3'
services:
localstack:
container_name: "${LOCALSTACK_DOCKER_NAME-localstack_main}"
image: localstack/localstack:latest
ports:
- "4567-4597:4567-4597"
- "${PORT_WEB_UI-8080}:${PORT_WEB_UI-8080}"
environment:
- SERVICES=apigateway
- DEBUG=1
- DATA_DIR=${DATA_DIR- }
- PORT_WEB_UI=${PORT_WEB_UI- }
- LAMBDA_EXECUTOR=${LAMBDA_EXECUTOR- }
- KINESIS_ERROR_PROBABILITY=${KINESIS_ERROR_PROBABILITY- }
- DOCKER_HOST=unix:///var/run/docker.sock
volumes:
- "${TMPDIR:-/tmp/localstack}:/tmp/localstack"
- "/var/run/docker.sock:/var/run/docker.sock"
```
aws output
```
$ aws --endpoint=http://localhost:4567 --region us-east-1 apigateway get-resources --rest-api-id xaper8hype
{
"items": [
{
"id": "5t3u62jjyg",
"path": "/",
"resourceMethods": {
"GET": {}
}
},
{
"id": "fu3fph78jj",
"parentId": "5t3u62jjyg",
"pathPart": "test123",
"path": "/test123",
"resourceMethods": {
"GET": {
"httpMethod": "GET",
"authorizationType": "NONE",
"methodIntegration": {
"type": "HTTP_PROXY",
"httpMethod": "GET",
"uri": "https://httpbin.org/anything/test123",
"requestTemplates": {},
"integrationResponses": {
"200": {
"statusCode": 200,
"responseTemplates": {
"application/json": null
}
}
}
}
}
}
}
]
}
```
I'm not sure if I do something wrong but the terraform code does work on AWS. | priority | api gateway error http proxy get not yet implemented when setting up a api gateway with http proxy i get the following error message when accessing the route message api gateway integration type http proxy for method get not yet implemented tf file provider aws access key mock access key region us east force path style true secret key mock secret key skip credentials validation true skip metadata api check true skip requesting account id true endpoints apigateway cloudformation cloudwatch dynamodb es firehose iam kinesis lambda redshift secretsmanager ses sns sqs ssm stepfunctions sts resource aws api gateway rest api mydemoapi name mydemoapi description this is my api for demonstration purposes resource aws api gateway resource mydemoresource rest api id aws api gateway rest api mydemoapi id parent id aws api gateway rest api mydemoapi root resource id path part resource aws api gateway method mydemomethod rest api id aws api gateway rest api mydemoapi id resource id aws api gateway resource mydemoresource id http method get authorization none request parameters method request path proxy true resource aws api gateway integration mydemointegration rest api id aws api gateway rest api mydemoapi id resource id aws api gateway resource mydemoresource id http method aws api gateway method mydemomethod http method type http proxy uri integration http method get cache key parameters timeout milliseconds request parameters integration request path proxy method request path proxy resource aws api gateway deployment teststage depends on aws api gateway integration mydemointegration rest api id aws api gateway rest api mydemoapi id stage name test output base url value aws api gateway deployment teststage invoke url docker compose version services localstack container name localstack docker name localstack main image localstack localstack latest ports port web ui port web ui environment services apigateway debug data dir data dir port web ui port web ui lambda executor lambda executor kinesis error probability kinesis error probability docker host unix var run docker sock volumes tmpdir tmp localstack tmp localstack var run docker sock var run docker sock aws output aws endpoint region us east apigateway get resources rest api id items id path resourcemethods get id parentid pathpart path resourcemethods get httpmethod get authorizationtype none methodintegration type http proxy httpmethod get uri requesttemplates integrationresponses statuscode responsetemplates application json null i m not sure if i do something wrong but the terraform code does work on aws | 1 |
453,255 | 13,067,283,054 | IssuesEvent | 2020-07-30 23:58:45 | Energy-Innovation/eps-us | https://api.github.com/repos/Energy-Innovation/eps-us | closed | Bug in calculating distributed solar PV costs and solution | high priority | I found a bug in the calculation of distributed solar subsidy costs.
For the calculation of _Subsidy Paid on Distributed Solar Capacity_, the variable _Perc Subsidy for Distributed Solar PV Capacity_ is used, but instead it should be _Perc Subsidy for Distributed Solar PV Capacity this Year_ | 1.0 | Bug in calculating distributed solar PV costs and solution - I found a bug in the calculation of distributed solar subsidy costs.
For the calculation of _Subsidy Paid on Distributed Solar Capacity_, the variable _Perc Subsidy for Distributed Solar PV Capacity_ is used, but instead it should be _Perc Subsidy for Distributed Solar PV Capacity this Year_ | priority | bug in calculating distributed solar pv costs and solution i found a bug in the calculation of distributed solar subsidy costs for the calculation of subsidy paid on distributed solar capacity the variable perc subsidy for distributed solar pv capacity is used but instead it should be perc subsidy for distributed solar pv capacity this year | 1 |
158,555 | 6,029,823,451 | IssuesEvent | 2017-06-08 18:55:55 | renatobenks/treasy-challenge | https://api.github.com/repos/renatobenks/treasy-challenge | closed | TREASY-02 - Integrar webpack middleware ao server | Priority: Highest Type: enhancement | Integrar webpack middleware ao server
---
Integrar o middleware do `webpack` ao `express` *server* para expedir o bundle da aplicação com o express.
- [webpack-dev-middleware](https://github.com/webpack/webpack-dev-middleware) | 1.0 | TREASY-02 - Integrar webpack middleware ao server - Integrar webpack middleware ao server
---
Integrar o middleware do `webpack` ao `express` *server* para expedir o bundle da aplicação com o express.
- [webpack-dev-middleware](https://github.com/webpack/webpack-dev-middleware) | priority | treasy integrar webpack middleware ao server integrar webpack middleware ao server integrar o middleware do webpack ao express server para expedir o bundle da aplicação com o express | 1 |
821,601 | 30,828,067,633 | IssuesEvent | 2023-08-01 21:57:45 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | Design System: Implement `button` component | Team : Lunik Type : New Functionality Triage OKR : Core Features Priority : 2 High | ### Parent Issue
#25355
### User Story
As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets.
### Acceptance Criteria
1. Match the design provided with all the types and severities in primeng
2. Add Storybook stories for the buttons with the same structure of primeng
3. Implement the buttons in dojo
4. User the CSS variables accordingly
### Proposed Objective
Core Features
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
[Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev)
### Assumptions & Initiation Needs
[_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e)
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | Design System: Implement `button` component - ### Parent Issue
#25355
### User Story
As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets.
### Acceptance Criteria
1. Match the design provided with all the types and severities in primeng
2. Add Storybook stories for the buttons with the same structure of primeng
3. Implement the buttons in dojo
4. User the CSS variables accordingly
### Proposed Objective
Core Features
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
[Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev)
### Assumptions & Initiation Needs
[_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e)
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | priority | design system implement button component parent issue user story as a stakeholder i want you to implement the button component new design for primeng and dojo widgets acceptance criteria match the design provided with all the types and severities in primeng add storybook stories for the buttons with the same structure of primeng implement the buttons in dojo user the css variables accordingly proposed objective core features proposed priority priority average external links slack conversations support tickets figma designs etc assumptions initiation needs quality assurance notes workarounds no response sub tasks estimates no response | 1 |
678,805 | 23,211,352,367 | IssuesEvent | 2022-08-02 10:24:21 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Getting Landed to a Authentication Error page when you select No for Logout Consent Prompt with React SDK | Priority/Highest Severity/Major bug developer-sdk 6.0.0-bug-fixing Affected-6.0.0 QA-Reported | **How to reproduce:**
1. Configure React SDK with IS 5.12.0 alpaha 17 pack by following the guide section **Try out a sample React SDK on Identity Server** in https://docs.google.com/document/d/1n-7DswTz4YJJl4T2wbpylBSa4pPL0UTQ_p3eDPh0VMU/edit#heading=h.y5d6rc170m0s
2. From management console enable config allow authenticate without client secret or public client config from console ( otherwise authentication will fail)
3. Access https://localhost:5000/
4. Login as admin:admin for the sample
5. Click Logout
6. From the prompt for are you sure to Logout select No
7. You will be redirected to Login page
8. Click again on Login will end up in a authentication error page
https://user-images.githubusercontent.com/31848014/160809207-401dab0f-7045-4f23-9222-52db3d44eeff.mp4

**Expected behavior:**
When you give No as Are you sure to Logout prompt you should not get navigated back to login page
Also once you click again on Login Button should not end up in a authentication error page
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
IS 5.12.0 alpha 17
postgres 14 | 1.0 | Getting Landed to a Authentication Error page when you select No for Logout Consent Prompt with React SDK - **How to reproduce:**
1. Configure React SDK with IS 5.12.0 alpaha 17 pack by following the guide section **Try out a sample React SDK on Identity Server** in https://docs.google.com/document/d/1n-7DswTz4YJJl4T2wbpylBSa4pPL0UTQ_p3eDPh0VMU/edit#heading=h.y5d6rc170m0s
2. From management console enable config allow authenticate without client secret or public client config from console ( otherwise authentication will fail)
3. Access https://localhost:5000/
4. Login as admin:admin for the sample
5. Click Logout
6. From the prompt for are you sure to Logout select No
7. You will be redirected to Login page
8. Click again on Login will end up in a authentication error page
https://user-images.githubusercontent.com/31848014/160809207-401dab0f-7045-4f23-9222-52db3d44eeff.mp4

**Expected behavior:**
When you give No as Are you sure to Logout prompt you should not get navigated back to login page
Also once you click again on Login Button should not end up in a authentication error page
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
IS 5.12.0 alpha 17
postgres 14 | priority | getting landed to a authentication error page when you select no for logout consent prompt with react sdk how to reproduce configure react sdk with is alpaha pack by following the guide section try out a sample react sdk on identity server in from management console enable config allow authenticate without client secret or public client config from console otherwise authentication will fail access login as admin admin for the sample click logout from the prompt for are you sure to logout select no you will be redirected to login page click again on login will end up in a authentication error page expected behavior when you give no as are you sure to logout prompt you should not get navigated back to login page also once you click again on login button should not end up in a authentication error page environment information please complete the following information remove any unnecessary fields is alpha postgres | 1 |
418,367 | 12,196,986,606 | IssuesEvent | 2020-04-29 19:58:21 | altest-com/vue-schemas | https://api.github.com/repos/altest-com/vue-schemas | closed | Bug in SchemaFieldsView, ItemSchemaView and ItemEditor components | bug high priority small size | When assigning fields or values to sections in these components, if the section of a field or a value does not exist (because a previously assigned section was deleted or it has never been assigned), the field is not rendered. | 1.0 | Bug in SchemaFieldsView, ItemSchemaView and ItemEditor components - When assigning fields or values to sections in these components, if the section of a field or a value does not exist (because a previously assigned section was deleted or it has never been assigned), the field is not rendered. | priority | bug in schemafieldsview itemschemaview and itemeditor components when assigning fields or values to sections in these components if the section of a field or a value does not exist because a previously assigned section was deleted or it has never been assigned the field is not rendered | 1 |
166,122 | 6,291,203,636 | IssuesEvent | 2017-07-19 23:49:53 | influxdata/docs.influxdata.com | https://api.github.com/repos/influxdata/docs.influxdata.com | closed | [InfluxDB] Document top() and bottom() divergent behavior when writing | Fix underway Priority: high Time commitment: project Type: update | ###### URL for relevant page?
https://github.com/influxdata/influxdb/pull/8398
###### What products and version are you using?
InfluxDB 1.3.0
These functions will have strange behavior when writing that will need to be documented. Whenever a tag is included in the function, it will be passed to be written as a tag instead of a field. This behavior is weird and these are the only functions that will have this behavior. | 1.0 | [InfluxDB] Document top() and bottom() divergent behavior when writing - ###### URL for relevant page?
https://github.com/influxdata/influxdb/pull/8398
###### What products and version are you using?
InfluxDB 1.3.0
These functions will have strange behavior when writing that will need to be documented. Whenever a tag is included in the function, it will be passed to be written as a tag instead of a field. This behavior is weird and these are the only functions that will have this behavior. | priority | document top and bottom divergent behavior when writing url for relevant page what products and version are you using influxdb these functions will have strange behavior when writing that will need to be documented whenever a tag is included in the function it will be passed to be written as a tag instead of a field this behavior is weird and these are the only functions that will have this behavior | 1 |
830,623 | 32,019,038,653 | IssuesEvent | 2023-09-22 01:48:08 | oceanbase/odc | https://api.github.com/repos/oceanbase/odc | closed | [Feature]: Data masking supports columns in database view | type-feature priority-high module-Data security | ### Describe your use case
ODC provide data masking for database tables since v4.2.0, but the database views are not taken into consideration.
In the current situation, sensitive data can be access illegally through creating a view and select data from this view.
1. So, I propose to incorporate views into data desensitization control.
### Describe the solution you'd like
1. Allow setting columns in views as sensitive columns;
2. When extracting columns from SQL, just make views into consideration.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_ | 1.0 | [Feature]: Data masking supports columns in database view - ### Describe your use case
ODC provide data masking for database tables since v4.2.0, but the database views are not taken into consideration.
In the current situation, sensitive data can be access illegally through creating a view and select data from this view.
1. So, I propose to incorporate views into data desensitization control.
### Describe the solution you'd like
1. Allow setting columns in views as sensitive columns;
2. When extracting columns from SQL, just make views into consideration.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_ | priority | data masking supports columns in database view describe your use case odc provide data masking for database tables since but the database views are not taken into consideration in the current situation sensitive data can be access illegally through creating a view and select data from this view so i propose to incorporate views into data desensitization control describe the solution you d like allow setting columns in views as sensitive columns; when extracting columns from sql just make views into consideration describe alternatives you ve considered no response additional context no response | 1 |
513,482 | 14,922,402,862 | IssuesEvent | 2021-01-23 14:41:08 | bounswe/bounswe2020group2 | https://api.github.com/repos/bounswe/bounswe2020group2 | closed | Implement Add/Edit Product Modal | effort: high priority: high type: front-end | We need to create a component that lets us add or edit a product. It needs to contain necessary product fields and also an image upload section for product images. | 1.0 | Implement Add/Edit Product Modal - We need to create a component that lets us add or edit a product. It needs to contain necessary product fields and also an image upload section for product images. | priority | implement add edit product modal we need to create a component that lets us add or edit a product it needs to contain necessary product fields and also an image upload section for product images | 1 |
359,799 | 10,681,001,444 | IssuesEvent | 2019-10-21 23:03:47 | bounswe/bounswe2019group7 | https://api.github.com/repos/bounswe/bounswe2019group7 | closed | API requests will be added on login and register page. | Effort: Few Hours Frontend Priority: High | It should be able to get token from API and correctly send register request.
Deadline: 20 October 2019, 23:59 | 1.0 | API requests will be added on login and register page. - It should be able to get token from API and correctly send register request.
Deadline: 20 October 2019, 23:59 | priority | api requests will be added on login and register page it should be able to get token from api and correctly send register request deadline october | 1 |
699,557 | 24,021,148,671 | IssuesEvent | 2022-09-15 07:46:27 | MathMarEcol/WSMPA2 | https://api.github.com/repos/MathMarEcol/WSMPA2 | opened | Change kernel density plot | enhancement NPI High Priority | - remove the colors (change it to just grey and blue)
- add the dashed lines for the means
- remove the numbers and replace with less and more
- maybe remove the legend with the color + numbers on the right-hand side? | 1.0 | Change kernel density plot - - remove the colors (change it to just grey and blue)
- add the dashed lines for the means
- remove the numbers and replace with less and more
- maybe remove the legend with the color + numbers on the right-hand side? | priority | change kernel density plot remove the colors change it to just grey and blue add the dashed lines for the means remove the numbers and replace with less and more maybe remove the legend with the color numbers on the right hand side | 1 |
426,429 | 12,372,362,171 | IssuesEvent | 2020-05-18 20:16:12 | jetrails/magento-cloudflare | https://api.github.com/repos/jetrails/magento-cloudflare | closed | Bugs When Read Only Permissions Are Present | bug priority: high | - [x] caching.caching_level: On fail, revert to value
- [x] page_rules.page_rules: Switch order for description and limits.
- [x] remove cloudflare import with common import
- [x] On add/update for modal, make modal inactive and grayed out.
- [x] Check $result->success in models
- [x] firewall.user_agent_blocking Default value for usage. | 1.0 | Bugs When Read Only Permissions Are Present - - [x] caching.caching_level: On fail, revert to value
- [x] page_rules.page_rules: Switch order for description and limits.
- [x] remove cloudflare import with common import
- [x] On add/update for modal, make modal inactive and grayed out.
- [x] Check $result->success in models
- [x] firewall.user_agent_blocking Default value for usage. | priority | bugs when read only permissions are present caching caching level on fail revert to value page rules page rules switch order for description and limits remove cloudflare import with common import on add update for modal make modal inactive and grayed out check result success in models firewall user agent blocking default value for usage | 1 |
555,134 | 16,447,771,126 | IssuesEvent | 2021-05-20 22:04:49 | capitalone/DataProfiler | https://api.github.com/repos/capitalone/DataProfiler | closed | Top Level UnstructuredProfiler (User Interaction) | High Priority New Feature | **Is your feature request related to a problem? Please describe.**
Need unstructured profiling
**Describe the outcome you'd like:**
Should follow same format as original profiler to mimic structured, however do the same for unstructured.
**Additional context:**
Starting code (can change):
```python
class UnstructuredProfiler(BaseProfiler):
def __init__(self, data, samples_per_update=None, min_true_samples=None,
profiler_options=None):
super().__init__(data,
samples_per_update,
min_true_samples,
profiler_options)
raise NotImplementedError()
def __add__(self, other):
"""
Merges two Unstructured profilers together overriding the `+` operator.
:param other: unstructured profile being add to this one.
:type other: UnstructuredProfiler
:return: merger of the two unstructured profilers
"""
raise NotImplementedError()
def _get_sample_size(self, data):
"""
Determines the minimum sampling size for detecting column type.
:param data: data to be profiled
:type data: pandas.core.series.Series
:return: integer sampling size
:rtype: int
"""
len_df = len(data)
if len_df < self._min_sample_size:
return int(len_df)
return max(int(self._sampling_ratio * len_df), self._min_sample_size)
def report(self, report_options=None):
"""Empty lines"""
raise NotImplementedError()
def _clean_and_sample_lines(self, data, min_true_samples=None, sample_ids=None):
"""
Identify null lines and return them in a dictionary as well as
remove any nulls data.
:param df_series: a given column
:type df_series: pandas.core.series.Series
:param sample_size: Number of samples to use in generating the profile
:type sample_size: int
:param min_true_samples: Minimum number of samples required for the
profiler
:type min_true_samples: int
:param sample_ids: Randomized list of sample indices
:type sample_ids: list(list)
:return: updated column with null removed and dictionary of null
parameters
:rtype: pd.Series, dict
"""
raise NotImplementedError()
def _update_base_stats(self, base_stats):
raise NotImplementedError()
def _update_line_statistics(self, data):
"""
Iterate over the provided dataset line by line and calculate
the line statistics. Specifically, number of unique lines,
lines containing null values, and total lines reviewed. This
function is safe to use in batches.
:param data: a dataset
:type data: pandas.DataFrame
"""
raise NotImplementedError()
def _update_profiler_helper(self, data, sample_size, min_true_samples, pool=None):
"""
Send the data to the profile compilers to get proifles on the data.
:param data: a dataset
:type data: pandas.DataFrame
:param sample_size: number of samples for profiling the data
:type sample_size: int
:param min_true_samples: minimum number samples which are not empty
or white space characters
:type min_true_samples: int
"""
if not sample_size:
sample_size = len(df_series)
if not sample_size:
sample_size = self._get_sample_size(df_series)
# Shuffle indices ones and share with columns
sample_ids = [*utils.shuffle_in_chunks(len(df), len(df))]
clean_sampled_df, base_stats = self._clean_and_sample_lines(data)
self._update_base_stats(base_stats)
self._update_line_statistics(data)
pool = None
if options.multiprocess.is_enabled:
cpu_count = 1
try:
cpu_count = mp.cpu_count()
except NotImplementedError as e:
cpu_count = 1
# No additional advantage beyond 8 processes
# Always leave 1 cores free
if cpu_count > 2:
cpu_count = min(cpu_count-1, 8)
pool = mp.Pool(cpu_count)
print("Utilizing", cpu_count, "processes for profiling")
# Profile compilers being updated
for profile in self.profiles.values():
profile.update_profile(clean_sampled_df, pool)
if pool is not None:
pool.close() # Close pool for new tasks
pool.join() # Wait for all workers to complete
return profile
def update_profile(self, data, sample_size=None):
"""
Update the profile for data provided. User can specify the sample
size to profile the data with. Additionally, the user can specify the
minimum number of non-null samples to profile.
:param data: data to be profiled
:type data: Union[data_readers.base_data.BaseData, pandas.DataFrame]
:param sample_size: number of samples to profile from the data
:type sample_size: int
:param min_true_samples: minimum number of non-null samples to profile
:type min_true_samples
:return: None
"""
if not sample_size:
sample_size = self._samples_per_update
if not min_true_samples:
min_true_samples = self._min_true_samples
if isinstance(data, data_readers.base_data.BaseData):
self._update_profile_helper(
data.data, sample_size, min_true_samples, self.options)
self.encoding = data.file_encoding
self.file_type = data.data_type
elif isinstance(data, pd.DataFrame):
self._update_profiler_helper(
data, sample_size, min_true_samples, self.options)
self.file_type = str(data.__class__)
else:
raise ValueError(
"Data must either be imported using the data_readers or "
"pd.DataFrame."
)
``` | 1.0 | Top Level UnstructuredProfiler (User Interaction) - **Is your feature request related to a problem? Please describe.**
Need unstructured profiling
**Describe the outcome you'd like:**
Should follow same format as original profiler to mimic structured, however do the same for unstructured.
**Additional context:**
Starting code (can change):
```python
class UnstructuredProfiler(BaseProfiler):
def __init__(self, data, samples_per_update=None, min_true_samples=None,
profiler_options=None):
super().__init__(data,
samples_per_update,
min_true_samples,
profiler_options)
raise NotImplementedError()
def __add__(self, other):
"""
Merges two Unstructured profilers together overriding the `+` operator.
:param other: unstructured profile being add to this one.
:type other: UnstructuredProfiler
:return: merger of the two unstructured profilers
"""
raise NotImplementedError()
def _get_sample_size(self, data):
"""
Determines the minimum sampling size for detecting column type.
:param data: data to be profiled
:type data: pandas.core.series.Series
:return: integer sampling size
:rtype: int
"""
len_df = len(data)
if len_df < self._min_sample_size:
return int(len_df)
return max(int(self._sampling_ratio * len_df), self._min_sample_size)
def report(self, report_options=None):
"""Empty lines"""
raise NotImplementedError()
def _clean_and_sample_lines(self, data, min_true_samples=None, sample_ids=None):
"""
Identify null lines and return them in a dictionary as well as
remove any nulls data.
:param df_series: a given column
:type df_series: pandas.core.series.Series
:param sample_size: Number of samples to use in generating the profile
:type sample_size: int
:param min_true_samples: Minimum number of samples required for the
profiler
:type min_true_samples: int
:param sample_ids: Randomized list of sample indices
:type sample_ids: list(list)
:return: updated column with null removed and dictionary of null
parameters
:rtype: pd.Series, dict
"""
raise NotImplementedError()
def _update_base_stats(self, base_stats):
raise NotImplementedError()
def _update_line_statistics(self, data):
"""
Iterate over the provided dataset line by line and calculate
the line statistics. Specifically, number of unique lines,
lines containing null values, and total lines reviewed. This
function is safe to use in batches.
:param data: a dataset
:type data: pandas.DataFrame
"""
raise NotImplementedError()
def _update_profiler_helper(self, data, sample_size, min_true_samples, pool=None):
"""
Send the data to the profile compilers to get proifles on the data.
:param data: a dataset
:type data: pandas.DataFrame
:param sample_size: number of samples for profiling the data
:type sample_size: int
:param min_true_samples: minimum number samples which are not empty
or white space characters
:type min_true_samples: int
"""
if not sample_size:
sample_size = len(df_series)
if not sample_size:
sample_size = self._get_sample_size(df_series)
# Shuffle indices ones and share with columns
sample_ids = [*utils.shuffle_in_chunks(len(df), len(df))]
clean_sampled_df, base_stats = self._clean_and_sample_lines(data)
self._update_base_stats(base_stats)
self._update_line_statistics(data)
pool = None
if options.multiprocess.is_enabled:
cpu_count = 1
try:
cpu_count = mp.cpu_count()
except NotImplementedError as e:
cpu_count = 1
# No additional advantage beyond 8 processes
# Always leave 1 cores free
if cpu_count > 2:
cpu_count = min(cpu_count-1, 8)
pool = mp.Pool(cpu_count)
print("Utilizing", cpu_count, "processes for profiling")
# Profile compilers being updated
for profile in self.profiles.values():
profile.update_profile(clean_sampled_df, pool)
if pool is not None:
pool.close() # Close pool for new tasks
pool.join() # Wait for all workers to complete
return profile
def update_profile(self, data, sample_size=None):
"""
Update the profile for data provided. User can specify the sample
size to profile the data with. Additionally, the user can specify the
minimum number of non-null samples to profile.
:param data: data to be profiled
:type data: Union[data_readers.base_data.BaseData, pandas.DataFrame]
:param sample_size: number of samples to profile from the data
:type sample_size: int
:param min_true_samples: minimum number of non-null samples to profile
:type min_true_samples
:return: None
"""
if not sample_size:
sample_size = self._samples_per_update
if not min_true_samples:
min_true_samples = self._min_true_samples
if isinstance(data, data_readers.base_data.BaseData):
self._update_profile_helper(
data.data, sample_size, min_true_samples, self.options)
self.encoding = data.file_encoding
self.file_type = data.data_type
elif isinstance(data, pd.DataFrame):
self._update_profiler_helper(
data, sample_size, min_true_samples, self.options)
self.file_type = str(data.__class__)
else:
raise ValueError(
"Data must either be imported using the data_readers or "
"pd.DataFrame."
)
``` | priority | top level unstructuredprofiler user interaction is your feature request related to a problem please describe need unstructured profiling describe the outcome you d like should follow same format as original profiler to mimic structured however do the same for unstructured additional context starting code can change python class unstructuredprofiler baseprofiler def init self data samples per update none min true samples none profiler options none super init data samples per update min true samples profiler options raise notimplementederror def add self other merges two unstructured profilers together overriding the operator param other unstructured profile being add to this one type other unstructuredprofiler return merger of the two unstructured profilers raise notimplementederror def get sample size self data determines the minimum sampling size for detecting column type param data data to be profiled type data pandas core series series return integer sampling size rtype int len df len data if len df self min sample size return int len df return max int self sampling ratio len df self min sample size def report self report options none empty lines raise notimplementederror def clean and sample lines self data min true samples none sample ids none identify null lines and return them in a dictionary as well as remove any nulls data param df series a given column type df series pandas core series series param sample size number of samples to use in generating the profile type sample size int param min true samples minimum number of samples required for the profiler type min true samples int param sample ids randomized list of sample indices type sample ids list list return updated column with null removed and dictionary of null parameters rtype pd series dict raise notimplementederror def update base stats self base stats raise notimplementederror def update line statistics self data iterate over the provided dataset line by line and calculate the line statistics specifically number of unique lines lines containing null values and total lines reviewed this function is safe to use in batches param data a dataset type data pandas dataframe raise notimplementederror def update profiler helper self data sample size min true samples pool none send the data to the profile compilers to get proifles on the data param data a dataset type data pandas dataframe param sample size number of samples for profiling the data type sample size int param min true samples minimum number samples which are not empty or white space characters type min true samples int if not sample size sample size len df series if not sample size sample size self get sample size df series shuffle indices ones and share with columns sample ids clean sampled df base stats self clean and sample lines data self update base stats base stats self update line statistics data pool none if options multiprocess is enabled cpu count try cpu count mp cpu count except notimplementederror as e cpu count no additional advantage beyond processes always leave cores free if cpu count cpu count min cpu count pool mp pool cpu count print utilizing cpu count processes for profiling profile compilers being updated for profile in self profiles values profile update profile clean sampled df pool if pool is not none pool close close pool for new tasks pool join wait for all workers to complete return profile def update profile self data sample size none update the profile for data provided user can specify the sample size to profile the data with additionally the user can specify the minimum number of non null samples to profile param data data to be profiled type data union param sample size number of samples to profile from the data type sample size int param min true samples minimum number of non null samples to profile type min true samples return none if not sample size sample size self samples per update if not min true samples min true samples self min true samples if isinstance data data readers base data basedata self update profile helper data data sample size min true samples self options self encoding data file encoding self file type data data type elif isinstance data pd dataframe self update profiler helper data sample size min true samples self options self file type str data class else raise valueerror data must either be imported using the data readers or pd dataframe | 1 |
743,772 | 25,914,410,732 | IssuesEvent | 2022-12-15 16:14:35 | daisy/pipeline-ui | https://api.github.com/repos/daisy/pipeline-ui | closed | `null` is passed as option value in certain cases | priority:high | An example is the "stylesheet-parameters" option in the "DTBook to braille" script (when left empty). This option is defined in XProc as follows:
```xml
<p:option name="stylesheet-parameters"
required="false"
px:type="transform-query"
select="''">
```
Another example is the "template" option in the "DTBook to ODT" script (when left empty). This option is defined in XProc as follows:
```xml
<p:option name="template"
required="false"
px:type="anyFileURI"
select="''"
px:media-type="application/vnd.oasis.opendocument.text-template">
```
Perhaps it's the `required="false"` in combination with the `select=""`? | 1.0 | `null` is passed as option value in certain cases - An example is the "stylesheet-parameters" option in the "DTBook to braille" script (when left empty). This option is defined in XProc as follows:
```xml
<p:option name="stylesheet-parameters"
required="false"
px:type="transform-query"
select="''">
```
Another example is the "template" option in the "DTBook to ODT" script (when left empty). This option is defined in XProc as follows:
```xml
<p:option name="template"
required="false"
px:type="anyFileURI"
select="''"
px:media-type="application/vnd.oasis.opendocument.text-template">
```
Perhaps it's the `required="false"` in combination with the `select=""`? | priority | null is passed as option value in certain cases an example is the stylesheet parameters option in the dtbook to braille script when left empty this option is defined in xproc as follows xml p option name stylesheet parameters required false px type transform query select another example is the template option in the dtbook to odt script when left empty this option is defined in xproc as follows xml p option name template required false px type anyfileuri select px media type application vnd oasis opendocument text template perhaps it s the required false in combination with the select | 1 |
330,399 | 10,039,745,667 | IssuesEvent | 2019-07-18 18:11:48 | onaio/reveal-frontend | https://api.github.com/repos/onaio/reveal-frontend | opened | Change the web Login screen to show the Reveal Logo instead of the OpenSRP logo on reveal-stage | Priority: High | URL: https://reveal-stage.smartregister.org/opensrp/login.jsp

- [ ] Change the logo to display the OpenSRP logo
- [ ] Change the word "OpenSRP" to "Powered by OpenSRP" | 1.0 | Change the web Login screen to show the Reveal Logo instead of the OpenSRP logo on reveal-stage - URL: https://reveal-stage.smartregister.org/opensrp/login.jsp

- [ ] Change the logo to display the OpenSRP logo
- [ ] Change the word "OpenSRP" to "Powered by OpenSRP" | priority | change the web login screen to show the reveal logo instead of the opensrp logo on reveal stage url change the logo to display the opensrp logo change the word opensrp to powered by opensrp | 1 |
240,010 | 7,800,272,168 | IssuesEvent | 2018-06-09 07:19:08 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0008366:
js error when selecting a row in contact search grid | Bug Felamimail Mantis high priority | **Reported by pschuele on 13 May 2013 10:36**
**Version:** Kristina (2013.03.3)
js error when selecting a row in contact search grid
**Steps to reproduce:** - open "search recipients" in message edit dialog
- select a row in the contact grid
**Additional information:** Uncaught TypeError: Cannot call method 'push' of undefined ContactGrid.js:217
Tine.Felamimail.ContactGridPanel.Ext.extend.updateRecipients ContactGrid.js:217
Tine.Felamimail.ContactGridPanel.Ext.extend.onCellClick ContactGrid.js:192
EXTUTIL.Event.fire ext-all-debug.js:2415
EXTUTIL.Observable.fireEvent ext-all-debug.js:2061
Ext.grid.GridPanel.Ext.extend.processEvent ext-all-debug.js:64599
Ext.grid.GridPanel.Ext.extend.onClick ext-all-debug.js:64613
h
| 1.0 | 0008366:
js error when selecting a row in contact search grid - **Reported by pschuele on 13 May 2013 10:36**
**Version:** Kristina (2013.03.3)
js error when selecting a row in contact search grid
**Steps to reproduce:** - open "search recipients" in message edit dialog
- select a row in the contact grid
**Additional information:** Uncaught TypeError: Cannot call method 'push' of undefined ContactGrid.js:217
Tine.Felamimail.ContactGridPanel.Ext.extend.updateRecipients ContactGrid.js:217
Tine.Felamimail.ContactGridPanel.Ext.extend.onCellClick ContactGrid.js:192
EXTUTIL.Event.fire ext-all-debug.js:2415
EXTUTIL.Observable.fireEvent ext-all-debug.js:2061
Ext.grid.GridPanel.Ext.extend.processEvent ext-all-debug.js:64599
Ext.grid.GridPanel.Ext.extend.onClick ext-all-debug.js:64613
h
| priority | js error when selecting a row in contact search grid reported by pschuele on may version kristina js error when selecting a row in contact search grid steps to reproduce open quot search recipients quot in message edit dialog select a row in the contact grid additional information uncaught typeerror cannot call method push of undefined contactgrid js tine felamimail contactgridpanel ext extend updaterecipients contactgrid js tine felamimail contactgridpanel ext extend oncellclick contactgrid js extutil event fire ext all debug js extutil observable fireevent ext all debug js ext grid gridpanel ext extend processevent ext all debug js ext grid gridpanel ext extend onclick ext all debug js h | 1 |
721,549 | 24,830,840,375 | IssuesEvent | 2022-10-26 03:13:06 | kubesphere/console | https://api.github.com/repos/kubesphere/console | closed | Can not select cluster in tab named 'Devops projects' of workspace member | kind/bug kind/need-to-verify priority/high | **Describe the bug**
Only request member cluster, and can not select other clusters.

**Versions used(KubeSphere/Kubernetes)**
KubeSphere: `v3.3.1-rc.5`
| 1.0 | Can not select cluster in tab named 'Devops projects' of workspace member - **Describe the bug**
Only request member cluster, and can not select other clusters.

**Versions used(KubeSphere/Kubernetes)**
KubeSphere: `v3.3.1-rc.5`
| priority | can not select cluster in tab named devops projects of workspace member describe the bug only request member cluster and can not select other clusters versions used kubesphere kubernetes kubesphere rc | 1 |
90,751 | 3,830,100,032 | IssuesEvent | 2016-03-31 13:30:01 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | opened | Snapshot unit test sometimes fail on travis | bug Priority: High | Sometimes the snapshot fails on travis.
We should also avoid to save snapshot test files.
| 1.0 | Snapshot unit test sometimes fail on travis - Sometimes the snapshot fails on travis.
We should also avoid to save snapshot test files.
| priority | snapshot unit test sometimes fail on travis sometimes the snapshot fails on travis we should also avoid to save snapshot test files | 1 |
160,971 | 6,106,276,223 | IssuesEvent | 2017-06-21 03:24:59 | jdmack/tlb | https://api.github.com/repos/jdmack/tlb | closed | LightManager class | Coding priority:high | Create initial light manager class that acts as a factory for lights and keep tracks of them. | 1.0 | LightManager class - Create initial light manager class that acts as a factory for lights and keep tracks of them. | priority | lightmanager class create initial light manager class that acts as a factory for lights and keep tracks of them | 1 |
566,790 | 16,830,388,281 | IssuesEvent | 2021-06-18 03:12:12 | ansible/awx-operator | https://api.github.com/repos/ansible/awx-operator | reopened | Restore fails if a value in the deployment has colon | priority:high state:needs_devel type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
In awx object spec I have `ingress_annotations: 'cert-manager.io/cluster-issuer: letsencrypt'`, the value that has a colon. In such a case the awx object does not serialize correctly and restore fails with "unknown playbook failure".
##### ENVIRONMENT
* AWX version: 19.2.0
* Operator version: 0.10.0
* Kubernetes version: v1.21.0+k3s1
* AWX install method: kubernetes
##### STEPS TO REPRODUCE
Backup:
```
---
apiVersion: awx.ansible.com/v1beta1
kind: AWXBackup
metadata:
name: awxbackup-manual-2021-06-04-01
spec:
deployment_name: awx
postgres_label_selector: app.kubernetes.io/instance=postgres-awx
backup_pvc: manual-claim
```
Restore:
```
---
apiVersion: awx.ansible.com/v1beta1
kind: AWXRestore
metadata:
name: awxrestore-init
spec:
deployment_name: awx
backup_pvc: manual-claim
backup_dir: '/backups/tower-openshift-backup-2021-06-04-11:42:15'
backup_pvc_namespace: default
```
##### EXPECTED RESULTS
Restore should succeed
##### ACTUAL RESULTS
Restore fails with "unknown playbook failure".
##### ADDITIONAL INFORMATION
This is what persisted `awx_object` looks like:
```
{admin_user: admin, api_version: awx.ansible.com/v1beta1, create_preload_data: True, deployment_type: awx, garbage_collect_secrets: False, hostname: nevermind.domain.tld, image_pull_policy: IfNotPresent, ingress_annotations: cert-manager.io/cluster-issuer: letsencrypt, ingress_tls_secret: awx-le-certificate, ingress_type: Ingress, kind: AWX, loadbalancer_port: 80, loadbalancer_protocol: http, projects_persistence: False, projects_storage_access_mode: ReadWriteMany, projects_storage_size: 8Gi, replicas: 1, route_tls_termination_mechanism: Edge, task_privileged: False}
```
Note how there is no quotes around `cert-manager.io/cluster-issuer: letsencrypt`. The reason for this is that [this line](https://github.com/ansible/awx-operator/blob/af5983cb6879cbb2b653af142d9a65d7f7d96117/roles/backup/tasks/awx-cro.yml#L24) does not quite work as the author intended.
Suggested fix:
```
bash -c 'echo "$0" > {{ backup_dir }}/awx_object' {{ awx_spec|quote }}
```
##### AWX-OPERATOR LOGS
Can be found here: <https://gist.github.com/AndrewSav/e9e73d9b4ab19341bf4926707ec52540> | 1.0 | Restore fails if a value in the deployment has colon - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
In awx object spec I have `ingress_annotations: 'cert-manager.io/cluster-issuer: letsencrypt'`, the value that has a colon. In such a case the awx object does not serialize correctly and restore fails with "unknown playbook failure".
##### ENVIRONMENT
* AWX version: 19.2.0
* Operator version: 0.10.0
* Kubernetes version: v1.21.0+k3s1
* AWX install method: kubernetes
##### STEPS TO REPRODUCE
Backup:
```
---
apiVersion: awx.ansible.com/v1beta1
kind: AWXBackup
metadata:
name: awxbackup-manual-2021-06-04-01
spec:
deployment_name: awx
postgres_label_selector: app.kubernetes.io/instance=postgres-awx
backup_pvc: manual-claim
```
Restore:
```
---
apiVersion: awx.ansible.com/v1beta1
kind: AWXRestore
metadata:
name: awxrestore-init
spec:
deployment_name: awx
backup_pvc: manual-claim
backup_dir: '/backups/tower-openshift-backup-2021-06-04-11:42:15'
backup_pvc_namespace: default
```
##### EXPECTED RESULTS
Restore should succeed
##### ACTUAL RESULTS
Restore fails with "unknown playbook failure".
##### ADDITIONAL INFORMATION
This is what persisted `awx_object` looks like:
```
{admin_user: admin, api_version: awx.ansible.com/v1beta1, create_preload_data: True, deployment_type: awx, garbage_collect_secrets: False, hostname: nevermind.domain.tld, image_pull_policy: IfNotPresent, ingress_annotations: cert-manager.io/cluster-issuer: letsencrypt, ingress_tls_secret: awx-le-certificate, ingress_type: Ingress, kind: AWX, loadbalancer_port: 80, loadbalancer_protocol: http, projects_persistence: False, projects_storage_access_mode: ReadWriteMany, projects_storage_size: 8Gi, replicas: 1, route_tls_termination_mechanism: Edge, task_privileged: False}
```
Note how there is no quotes around `cert-manager.io/cluster-issuer: letsencrypt`. The reason for this is that [this line](https://github.com/ansible/awx-operator/blob/af5983cb6879cbb2b653af142d9a65d7f7d96117/roles/backup/tasks/awx-cro.yml#L24) does not quite work as the author intended.
Suggested fix:
```
bash -c 'echo "$0" > {{ backup_dir }}/awx_object' {{ awx_spec|quote }}
```
##### AWX-OPERATOR LOGS
Can be found here: <https://gist.github.com/AndrewSav/e9e73d9b4ab19341bf4926707ec52540> | priority | restore fails if a value in the deployment has colon issue type bug report summary in awx object spec i have ingress annotations cert manager io cluster issuer letsencrypt the value that has a colon in such a case the awx object does not serialize correctly and restore fails with unknown playbook failure environment awx version operator version kubernetes version awx install method kubernetes steps to reproduce backup apiversion awx ansible com kind awxbackup metadata name awxbackup manual spec deployment name awx postgres label selector app kubernetes io instance postgres awx backup pvc manual claim restore apiversion awx ansible com kind awxrestore metadata name awxrestore init spec deployment name awx backup pvc manual claim backup dir backups tower openshift backup backup pvc namespace default expected results restore should succeed actual results restore fails with unknown playbook failure additional information this is what persisted awx object looks like admin user admin api version awx ansible com create preload data true deployment type awx garbage collect secrets false hostname nevermind domain tld image pull policy ifnotpresent ingress annotations cert manager io cluster issuer letsencrypt ingress tls secret awx le certificate ingress type ingress kind awx loadbalancer port loadbalancer protocol http projects persistence false projects storage access mode readwritemany projects storage size replicas route tls termination mechanism edge task privileged false note how there is no quotes around cert manager io cluster issuer letsencrypt the reason for this is that does not quite work as the author intended suggested fix bash c echo backup dir awx object awx spec quote awx operator logs can be found here | 1 |
661,294 | 22,046,635,513 | IssuesEvent | 2022-05-30 03:02:00 | EvolutionX-10/Radon | https://api.github.com/repos/EvolutionX-10/Radon | closed | bug: Bot locks itself out of channels prematurely | Good first issue Bug: Confirmed SEM: Patch Priority: High | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description of the bug
When locking a channel to the \@everyone role when Radon doesn't have send messages specifically granted to it on that channel, the bot will be locked out by its own actions and unable to send the message.
### Steps To Reproduce
1. Create a channel with all permissions set to gray, and ensure that the bot and the bot's roles don't have send messages specifically granted on that channel
2. Lock the channel, and try to send a message with the lock
### Expected behavior
The message should be sent before locking the channel, to avoid locking the bot out.
### Screenshots
_No response_
### Additional context
The bot remains in a pending state for 15 minutes until it times out because of this bug, but the channel is still locked successfully | 1.0 | bug: Bot locks itself out of channels prematurely - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description of the bug
When locking a channel to the \@everyone role when Radon doesn't have send messages specifically granted to it on that channel, the bot will be locked out by its own actions and unable to send the message.
### Steps To Reproduce
1. Create a channel with all permissions set to gray, and ensure that the bot and the bot's roles don't have send messages specifically granted on that channel
2. Lock the channel, and try to send a message with the lock
### Expected behavior
The message should be sent before locking the channel, to avoid locking the bot out.
### Screenshots
_No response_
### Additional context
The bot remains in a pending state for 15 minutes until it times out because of this bug, but the channel is still locked successfully | priority | bug bot locks itself out of channels prematurely is there an existing issue for this i have searched the existing issues description of the bug when locking a channel to the everyone role when radon doesn t have send messages specifically granted to it on that channel the bot will be locked out by its own actions and unable to send the message steps to reproduce create a channel with all permissions set to gray and ensure that the bot and the bot s roles don t have send messages specifically granted on that channel lock the channel and try to send a message with the lock expected behavior the message should be sent before locking the channel to avoid locking the bot out screenshots no response additional context the bot remains in a pending state for minutes until it times out because of this bug but the channel is still locked successfully | 1 |
581,314 | 17,290,959,388 | IssuesEvent | 2021-07-24 18:41:46 | TestCentric/testcentric-gui | https://api.github.com/repos/TestCentric/testcentric-gui | closed | Hide result box in TestPropertiesDialog when there is no result | Enhancement High Priority | As is done in the TestPropertiesView. | 1.0 | Hide result box in TestPropertiesDialog when there is no result - As is done in the TestPropertiesView. | priority | hide result box in testpropertiesdialog when there is no result as is done in the testpropertiesview | 1 |
816,072 | 30,587,210,825 | IssuesEvent | 2023-07-21 14:14:36 | go-shiori/shiori | https://api.github.com/repos/go-shiori/shiori | closed | Title is never retrieved when adding bookmark | type:bug good first issue component:backend priority:high | Heads up: title is never retrieved
https://github.com/go-shiori/shiori/blob/v1.5.5-rc.1/internal/webserver/handler-api.go#L320-L322
```go
// Make sure bookmark's title not empty
if book.Title == "" {
book.Title = book.URL
}
// ^ because if this
// …
bookmark, err := downloadBookmarkContent(book, h.DataDir, r, book.Title != "", book.Excerpt != "")
// ^ it's always non-empty
// …
book, err = downloadBookmarkContent(book, h.DataDir, r, book.Title != "", book.Excerpt != "")
// ^ it's always non-empty
```
_Originally posted by @eitau in https://github.com/go-shiori/shiori/discussions/637#discussioncomment-6476870_ | 1.0 | Title is never retrieved when adding bookmark - Heads up: title is never retrieved
https://github.com/go-shiori/shiori/blob/v1.5.5-rc.1/internal/webserver/handler-api.go#L320-L322
```go
// Make sure bookmark's title not empty
if book.Title == "" {
book.Title = book.URL
}
// ^ because if this
// …
bookmark, err := downloadBookmarkContent(book, h.DataDir, r, book.Title != "", book.Excerpt != "")
// ^ it's always non-empty
// …
book, err = downloadBookmarkContent(book, h.DataDir, r, book.Title != "", book.Excerpt != "")
// ^ it's always non-empty
```
_Originally posted by @eitau in https://github.com/go-shiori/shiori/discussions/637#discussioncomment-6476870_ | priority | title is never retrieved when adding bookmark heads up title is never retrieved go make sure bookmark s title not empty if book title book title book url because if this … bookmark err downloadbookmarkcontent book h datadir r book title book excerpt it s always non empty … book err downloadbookmarkcontent book h datadir r book title book excerpt it s always non empty originally posted by eitau in | 1 |
435,562 | 12,536,533,328 | IssuesEvent | 2020-06-05 00:23:23 | IngenioUN/front_end | https://api.github.com/repos/IngenioUN/front_end | closed | Crear UI. Registro | Cap: Front-End Priority: High Type: New | Debe ser responsive.
Tener en cuenta la sintaxis en el ingreso del correo y la contraseña. | 1.0 | Crear UI. Registro - Debe ser responsive.
Tener en cuenta la sintaxis en el ingreso del correo y la contraseña. | priority | crear ui registro debe ser responsive tener en cuenta la sintaxis en el ingreso del correo y la contraseña | 1 |
715,307 | 24,593,750,940 | IssuesEvent | 2022-10-14 06:17:59 | AY2223S1-CS2103T-T15-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-T15-1/tp | closed | Update `find` function and `edit` function | priority.High type.Task | Both functions are already implemented (from AB3), so just take a look and make appropriate changes.
| 1.0 | Update `find` function and `edit` function - Both functions are already implemented (from AB3), so just take a look and make appropriate changes.
| priority | update find function and edit function both functions are already implemented from so just take a look and make appropriate changes | 1 |
83,543 | 3,637,047,322 | IssuesEvent | 2016-02-12 08:49:00 | Valhalla-Gaming/Tracker | https://api.github.com/repos/Valhalla-Gaming/Tracker | closed | [spell] paladin - guarded by the light | Class-Paladin Priority-High | how it should work: Your spell power is now equal to 100% of your attack power, and you no longer benefit from other sources of spell power.
How it works: Spell Power is always "0" , means no convert from attack power. That also means for the paladin ALL dmg- and heal-spells do much too less dmg/healing!
http://www.wowhead.com/spell=53592/guarded-by-the-light | 1.0 | [spell] paladin - guarded by the light - how it should work: Your spell power is now equal to 100% of your attack power, and you no longer benefit from other sources of spell power.
How it works: Spell Power is always "0" , means no convert from attack power. That also means for the paladin ALL dmg- and heal-spells do much too less dmg/healing!
http://www.wowhead.com/spell=53592/guarded-by-the-light | priority | paladin guarded by the light how it should work your spell power is now equal to of your attack power and you no longer benefit from other sources of spell power how it works spell power is always means no convert from attack power that also means for the paladin all dmg and heal spells do much too less dmg healing | 1 |
65,958 | 3,248,292,970 | IssuesEvent | 2015-10-17 05:51:02 | dhowe/RiTa | https://api.github.com/repos/dhowe/RiTa | closed | Improvements to Tutorials | enhancement PRIORITY: High website | - [x] A code example from RiTa for EVERY sub-section/section (including what the output would be)
- [x] Markov-page is great, but only about halfway done; continue describing
- [x] how to create a model from an existing text, and
- [x] how to use that model to generate a new text
- [x] code examples
- [x] link to Kafgenstein example: /RiTaJS/examples/p5js/Kafgenstein
- [x] Grammar page
- [x] include a larger grammar (with standard YAML format, no '|'s)
- [x] show an example with generateFrom() -- see RiGrammar.php
- [x] link to Haiku example: /RiTaJS/examples/p5js/HaikuGrammar
- [x] Concordance page needs a lot more work
- [x] better explanation, see [this](http://shiffman.net/teaching/a2z/analysis/#concordance)
- [x] code examples
- [x] info on [KWIC models](https://en.wikipedia.org/wiki/Key_Word_in_Context)
- [x] link to KWICModel example: /RiTaJS/examples/p5js/KWICmodel
- [x] Syntax highlighting (see #209)
Note: this [page](http://shiffman.net/teaching/a2z/generate/) has some good info for deepening our generation pages. Feel free to borrow from it (with citation of course) | 1.0 | Improvements to Tutorials - - [x] A code example from RiTa for EVERY sub-section/section (including what the output would be)
- [x] Markov-page is great, but only about halfway done; continue describing
- [x] how to create a model from an existing text, and
- [x] how to use that model to generate a new text
- [x] code examples
- [x] link to Kafgenstein example: /RiTaJS/examples/p5js/Kafgenstein
- [x] Grammar page
- [x] include a larger grammar (with standard YAML format, no '|'s)
- [x] show an example with generateFrom() -- see RiGrammar.php
- [x] link to Haiku example: /RiTaJS/examples/p5js/HaikuGrammar
- [x] Concordance page needs a lot more work
- [x] better explanation, see [this](http://shiffman.net/teaching/a2z/analysis/#concordance)
- [x] code examples
- [x] info on [KWIC models](https://en.wikipedia.org/wiki/Key_Word_in_Context)
- [x] link to KWICModel example: /RiTaJS/examples/p5js/KWICmodel
- [x] Syntax highlighting (see #209)
Note: this [page](http://shiffman.net/teaching/a2z/generate/) has some good info for deepening our generation pages. Feel free to borrow from it (with citation of course) | priority | improvements to tutorials a code example from rita for every sub section section including what the output would be markov page is great but only about halfway done continue describing how to create a model from an existing text and how to use that model to generate a new text code examples link to kafgenstein example ritajs examples kafgenstein grammar page include a larger grammar with standard yaml format no s show an example with generatefrom see rigrammar php link to haiku example ritajs examples haikugrammar concordance page needs a lot more work better explanation see code examples info on link to kwicmodel example ritajs examples kwicmodel syntax highlighting see note this has some good info for deepening our generation pages feel free to borrow from it with citation of course | 1 |
349,944 | 10,475,983,508 | IssuesEvent | 2019-09-23 17:33:18 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | PyTorch "previous versions" install instructions are confusing | high priority module: binaries triaged | ## ❓ Questions and Help
I'm trying to download pytorch to run on my university machines that have CUDA 9. The link at:
https://download.pytorch.org/whl/cu90/torch_stable.html however, doesn't seem to work.
cc @ezyang @gchanan @zou3519 | 1.0 | PyTorch "previous versions" install instructions are confusing - ## ❓ Questions and Help
I'm trying to download pytorch to run on my university machines that have CUDA 9. The link at:
https://download.pytorch.org/whl/cu90/torch_stable.html however, doesn't seem to work.
cc @ezyang @gchanan @zou3519 | priority | pytorch previous versions install instructions are confusing ❓ questions and help i m trying to download pytorch to run on my university machines that have cuda the link at however doesn t seem to work cc ezyang gchanan | 1 |
168,327 | 6,369,542,193 | IssuesEvent | 2017-08-01 12:11:07 | fossasia/susper.com | https://api.github.com/repos/fossasia/susper.com | closed | Code Maintainability: Arrange the services and components in a separate folder | enhancement Priority: High | **Actual Behaviour**
There are a lot of services inside `src/app/`
**Expected Behaviour**
Create a separate folder for the services and a separate folder for the components to easily distinguish between them.
**Screenshots of the issue**
(Not applicable here) | 1.0 | Code Maintainability: Arrange the services and components in a separate folder - **Actual Behaviour**
There are a lot of services inside `src/app/`
**Expected Behaviour**
Create a separate folder for the services and a separate folder for the components to easily distinguish between them.
**Screenshots of the issue**
(Not applicable here) | priority | code maintainability arrange the services and components in a separate folder actual behaviour there are a lot of services inside src app expected behaviour create a separate folder for the services and a separate folder for the components to easily distinguish between them screenshots of the issue not applicable here | 1 |
359,426 | 10,676,009,408 | IssuesEvent | 2019-10-21 12:57:25 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Port over stat optimizations | Feature High Priority | I've rewritten the stat system to reduce boilerplate, we'll need to port over your optimizations you made and bullet proof the system. Lets meet on this when you join the party on 9.0 later. | 1.0 | Port over stat optimizations - I've rewritten the stat system to reduce boilerplate, we'll need to port over your optimizations you made and bullet proof the system. Lets meet on this when you join the party on 9.0 later. | priority | port over stat optimizations i ve rewritten the stat system to reduce boilerplate we ll need to port over your optimizations you made and bullet proof the system lets meet on this when you join the party on later | 1 |
683,494 | 23,384,474,224 | IssuesEvent | 2022-08-11 12:39:04 | pvs-hd-tea/PyTypes | https://api.github.com/repos/pvs-hd-tea/PyTypes | closed | Fix Tracer not inferring correct module paths | Bug High Priority | It must account for builtin types, stdlib types, project types and venv imports | 1.0 | Fix Tracer not inferring correct module paths - It must account for builtin types, stdlib types, project types and venv imports | priority | fix tracer not inferring correct module paths it must account for builtin types stdlib types project types and venv imports | 1 |
317,870 | 9,670,222,181 | IssuesEvent | 2019-05-21 19:21:08 | Masuzu/SarasaBot | https://api.github.com/repos/Masuzu/SarasaBot | closed | Sarasa bot crashing when raid boss die | bug high priority | Sarasa bot run perfectly for short time and sometime when raid boss die it crashing without any warn
my .lua
character_1:UseSkill(3)
character_1:UseSkill(1)
Summon(6)
if turn == 2 then
character_2:UseSkill(1)
character_2:UseSkill(3)
character_3:UseSkill(4)
character_3:UseSkill(1)
character_4:UseSkill(3)
character_4:UseSkill(2)
character_4:UseSkill(1)
end
my .ini
RaidsToTrack=Lv120 グリームニル,Lvl 120 Grimnir
MaxNumPlayersInRaid=10
MinimumRemainingBossHpPercentage=50
MaximumRemainingBossHpPercentage=100
RaidRequestTimeToLiveInMs=180000
MinDelayInMsBeforeTryingToRejoinRaid=20000
DisplayRaids=true
[General]
Language=en
MaxNumSoulBerries=9999
AdditionalNumSoulBerriesToUsePerRefill=0
MaxNumSoulBalms=9999
UseSoulBalmsWhenNoRemainingSoulBerries=true
UseSoulBalmsFirst=false
MaxPageLoadDelayInMsBeforeRetry=3500
MaxTriggerDelayInMsBeforeFallback=5000
MaxResponseDelayInMs=2000
MaxNumActionRetries=1
MinWaitTimeInMsAfterRefresh=2000
TimeLimitInSeconds=21000
UseViramate=false
OngoingRaidsCheckFrequencyInSeconds=10
WaitForRaidEndBeforeJoiningNewRaid=false
CountdownTimerHorizontalPosition=0
CountdownTimerVerticalPosition=0
FastRaidRewardCollection=false
MaximumDelayInMinutesBeforeAttemptingToRetreatFromARaid=10
QuickRaidJoining=false
[Inputs]
DelayInMsBetweenMouseDownAndUp=40
RandomDelayInMsBetweenMouseDownAndUp=20
MouseSpeed=700
MouseScrollSpeed=700
ExitKeyCode=112
WaitTimeInMsBeforeClickInput=50
[Summons]
PreferredSummons=shiva
DefaultSummonAttributeTab=Misc
RerollSummonWhenNoPreferredSummonWasFound=false
[Combat]
MinWaitTimeInMsAfterAttack=200
MinWaitTimeInMsAfterSummon=3000
MinWaitTimeInMsAfterAbility=1750
LuaScript=Scripts/HL2.lua
ReloadPageOnLastBattle=true
MaxInactivityPeriodBeforeLeavingTheRaidInMs=6000
EnableEarlyTurnStartWithViramate=false
MinWaitTimeBeforeActionAfterEarlyTurnStartInMs=1500
AdditionalDelayInMsAfter2ChainBurst=1500
AdditionalDelayInMsAfter3ChainBurst=5000
AdditionalDelayInMsAfter4ChainBurst=7600
AdditionalDelayInMsAfterOverchainBurst=7600
[Debug]
Clicks=false
ForceRaidRewardCollection=false
ForceRaidJoining=true
[Alerting]
CaptchaNotificationSoundPath=captcha_notification.mp3
NumNotifications=9999
MaxNumSummonSelectionFailuresBeforePlayingSoundNotification=3
Check4SymbolsPendantsCount=false
Max4SymbolsPendantsCount=30000 | 1.0 | Sarasa bot crashing when raid boss die - Sarasa bot run perfectly for short time and sometime when raid boss die it crashing without any warn
my .lua
character_1:UseSkill(3)
character_1:UseSkill(1)
Summon(6)
if turn == 2 then
character_2:UseSkill(1)
character_2:UseSkill(3)
character_3:UseSkill(4)
character_3:UseSkill(1)
character_4:UseSkill(3)
character_4:UseSkill(2)
character_4:UseSkill(1)
end
my .ini
RaidsToTrack=Lv120 グリームニル,Lvl 120 Grimnir
MaxNumPlayersInRaid=10
MinimumRemainingBossHpPercentage=50
MaximumRemainingBossHpPercentage=100
RaidRequestTimeToLiveInMs=180000
MinDelayInMsBeforeTryingToRejoinRaid=20000
DisplayRaids=true
[General]
Language=en
MaxNumSoulBerries=9999
AdditionalNumSoulBerriesToUsePerRefill=0
MaxNumSoulBalms=9999
UseSoulBalmsWhenNoRemainingSoulBerries=true
UseSoulBalmsFirst=false
MaxPageLoadDelayInMsBeforeRetry=3500
MaxTriggerDelayInMsBeforeFallback=5000
MaxResponseDelayInMs=2000
MaxNumActionRetries=1
MinWaitTimeInMsAfterRefresh=2000
TimeLimitInSeconds=21000
UseViramate=false
OngoingRaidsCheckFrequencyInSeconds=10
WaitForRaidEndBeforeJoiningNewRaid=false
CountdownTimerHorizontalPosition=0
CountdownTimerVerticalPosition=0
FastRaidRewardCollection=false
MaximumDelayInMinutesBeforeAttemptingToRetreatFromARaid=10
QuickRaidJoining=false
[Inputs]
DelayInMsBetweenMouseDownAndUp=40
RandomDelayInMsBetweenMouseDownAndUp=20
MouseSpeed=700
MouseScrollSpeed=700
ExitKeyCode=112
WaitTimeInMsBeforeClickInput=50
[Summons]
PreferredSummons=shiva
DefaultSummonAttributeTab=Misc
RerollSummonWhenNoPreferredSummonWasFound=false
[Combat]
MinWaitTimeInMsAfterAttack=200
MinWaitTimeInMsAfterSummon=3000
MinWaitTimeInMsAfterAbility=1750
LuaScript=Scripts/HL2.lua
ReloadPageOnLastBattle=true
MaxInactivityPeriodBeforeLeavingTheRaidInMs=6000
EnableEarlyTurnStartWithViramate=false
MinWaitTimeBeforeActionAfterEarlyTurnStartInMs=1500
AdditionalDelayInMsAfter2ChainBurst=1500
AdditionalDelayInMsAfter3ChainBurst=5000
AdditionalDelayInMsAfter4ChainBurst=7600
AdditionalDelayInMsAfterOverchainBurst=7600
[Debug]
Clicks=false
ForceRaidRewardCollection=false
ForceRaidJoining=true
[Alerting]
CaptchaNotificationSoundPath=captcha_notification.mp3
NumNotifications=9999
MaxNumSummonSelectionFailuresBeforePlayingSoundNotification=3
Check4SymbolsPendantsCount=false
Max4SymbolsPendantsCount=30000 | priority | sarasa bot crashing when raid boss die sarasa bot run perfectly for short time and sometime when raid boss die it crashing without any warn my lua character useskill character useskill summon if turn then character useskill character useskill character useskill character useskill character useskill character useskill character useskill end my ini raidstotrack グリームニル lvl grimnir maxnumplayersinraid minimumremainingbosshppercentage maximumremainingbosshppercentage raidrequesttimetoliveinms mindelayinmsbeforetryingtorejoinraid displayraids true language en maxnumsoulberries additionalnumsoulberriestouseperrefill maxnumsoulbalms usesoulbalmswhennoremainingsoulberries true usesoulbalmsfirst false maxpageloaddelayinmsbeforeretry maxtriggerdelayinmsbeforefallback maxresponsedelayinms maxnumactionretries minwaittimeinmsafterrefresh timelimitinseconds useviramate false ongoingraidscheckfrequencyinseconds waitforraidendbeforejoiningnewraid false countdowntimerhorizontalposition countdowntimerverticalposition fastraidrewardcollection false maximumdelayinminutesbeforeattemptingtoretreatfromaraid quickraidjoining false delayinmsbetweenmousedownandup randomdelayinmsbetweenmousedownandup mousespeed mousescrollspeed exitkeycode waittimeinmsbeforeclickinput preferredsummons shiva defaultsummonattributetab misc rerollsummonwhennopreferredsummonwasfound false minwaittimeinmsafterattack minwaittimeinmsaftersummon minwaittimeinmsafterability luascript scripts lua reloadpageonlastbattle true maxinactivityperiodbeforeleavingtheraidinms enableearlyturnstartwithviramate false minwaittimebeforeactionafterearlyturnstartinms additionaldelayinmsafteroverchainburst clicks false forceraidrewardcollection false forceraidjoining true captchanotificationsoundpath captcha notification numnotifications maxnumsummonselectionfailuresbeforeplayingsoundnotification false | 1 |
33,607 | 2,770,422,121 | IssuesEvent | 2015-05-01 14:30:45 | nickdhondt/dungeons-and-dragons | https://api.github.com/repos/nickdhondt/dungeons-and-dragons | closed | make skill array | enhancement High priority javascript minor new feature php SQL | Make the main array 'skill' and clean get_user_skill_data function to the normalised array structure.
MEMO: get rid of the ';' delimeters and replace then in a loopable array. | 1.0 | make skill array - Make the main array 'skill' and clean get_user_skill_data function to the normalised array structure.
MEMO: get rid of the ';' delimeters and replace then in a loopable array. | priority | make skill array make the main array skill and clean get user skill data function to the normalised array structure memo get rid of the delimeters and replace then in a loopable array | 1 |
116,695 | 4,705,326,696 | IssuesEvent | 2016-10-13 14:14:31 | centre-for-educational-technology/edidaktikum | https://api.github.com/repos/centre-for-educational-technology/edidaktikum | closed | Grupi määramine ülesande vastuse juures | enhancement high priority | Ülesande vastuse juures arvestada gruppi, mille on määranud ülesande looja - vastuse sisestaja ei tohi saada eemaldada või lisada gruppe. Kasutaja määramine toimib endiselt.
Probleemiks on vastused, millele on määratud grupp, kuhu hindav õppejõud ei kuulu. Sellisel juhul hetkel õppejõud vastuse staatust muuta ei saa.
Kattub piletiga #357 ja sulgeb selle. | 1.0 | Grupi määramine ülesande vastuse juures - Ülesande vastuse juures arvestada gruppi, mille on määranud ülesande looja - vastuse sisestaja ei tohi saada eemaldada või lisada gruppe. Kasutaja määramine toimib endiselt.
Probleemiks on vastused, millele on määratud grupp, kuhu hindav õppejõud ei kuulu. Sellisel juhul hetkel õppejõud vastuse staatust muuta ei saa.
Kattub piletiga #357 ja sulgeb selle. | priority | grupi määramine ülesande vastuse juures ülesande vastuse juures arvestada gruppi mille on määranud ülesande looja vastuse sisestaja ei tohi saada eemaldada või lisada gruppe kasutaja määramine toimib endiselt probleemiks on vastused millele on määratud grupp kuhu hindav õppejõud ei kuulu sellisel juhul hetkel õppejõud vastuse staatust muuta ei saa kattub piletiga ja sulgeb selle | 1 |
359,896 | 10,682,258,843 | IssuesEvent | 2019-10-22 04:32:32 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Publisher UI should show warning when editing API resurces that are used by API Products | 3.0.0 Priority/High | Editing API resources that are used by API Products could affect the respective API Product. Therefore it is better to display a warning in the UI when someone tries to edit such an API resource | 1.0 | Publisher UI should show warning when editing API resurces that are used by API Products - Editing API resources that are used by API Products could affect the respective API Product. Therefore it is better to display a warning in the UI when someone tries to edit such an API resource | priority | publisher ui should show warning when editing api resurces that are used by api products editing api resources that are used by api products could affect the respective api product therefore it is better to display a warning in the ui when someone tries to edit such an api resource | 1 |
468,379 | 13,466,524,457 | IssuesEvent | 2020-09-09 23:09:27 | OpenApoc/OpenApoc | https://api.github.com/repos/OpenApoc/OpenApoc | closed | No Vehicle object matching ID "VEHICLE_19" | !BUG! HIGH PRIORITY Duplicate Verified / Replicated | Hi, I have been speaking to FilmBoy and he has ask me to log this and send save and log
[save_lukewhale.zip](https://github.com/OpenApoc/OpenApoc/files/5198244/save_lukewhale.zip)
[log.txt](https://github.com/OpenApoc/OpenApoc/files/5198253/log.txt)
All i done in the game was buy/sell and build base.
(Note) Sold ground units and bought hover bikes x4 , hover cars x4 and equip them with lasers and better engines. Change the name of soldiers and recruit new researchers/engineers. Dont this is this may have caused a issues but i moved all the researchers and engineers out of the base and then picked the highest skill level
Anything else just let me know.. Thank you
OpenApoc error
class std::shared_ptr<class OpenApoc::vehicle> _cdecl OpenApoc::stateObject<class OpenApoc::Vehicle>::get(const class OpenApoc::GameState &,const class std::basic_string<char_traits<char>,class std::allocator<char> > &): No vehicle matching ID "VEHICLE_19"
void _cdecl OpenApoc::StateRef<class OpenApoc::vehicle>::resolve(void) const: No Vehicle object matching ID "VEHICLE_19" found | 1.0 | No Vehicle object matching ID "VEHICLE_19" - Hi, I have been speaking to FilmBoy and he has ask me to log this and send save and log
[save_lukewhale.zip](https://github.com/OpenApoc/OpenApoc/files/5198244/save_lukewhale.zip)
[log.txt](https://github.com/OpenApoc/OpenApoc/files/5198253/log.txt)
All i done in the game was buy/sell and build base.
(Note) Sold ground units and bought hover bikes x4 , hover cars x4 and equip them with lasers and better engines. Change the name of soldiers and recruit new researchers/engineers. Dont this is this may have caused a issues but i moved all the researchers and engineers out of the base and then picked the highest skill level
Anything else just let me know.. Thank you
OpenApoc error
class std::shared_ptr<class OpenApoc::vehicle> _cdecl OpenApoc::stateObject<class OpenApoc::Vehicle>::get(const class OpenApoc::GameState &,const class std::basic_string<char_traits<char>,class std::allocator<char> > &): No vehicle matching ID "VEHICLE_19"
void _cdecl OpenApoc::StateRef<class OpenApoc::vehicle>::resolve(void) const: No Vehicle object matching ID "VEHICLE_19" found | priority | no vehicle object matching id vehicle hi i have been speaking to filmboy and he has ask me to log this and send save and log all i done in the game was buy sell and build base note sold ground units and bought hover bikes hover cars and equip them with lasers and better engines change the name of soldiers and recruit new researchers engineers dont this is this may have caused a issues but i moved all the researchers and engineers out of the base and then picked the highest skill level anything else just let me know thank you openapoc error class std shared ptr cdecl openapoc stateobject get const class openapoc gamestate const class std basic string class std allocator no vehicle matching id vehicle void cdecl openapoc stateref resolve void const no vehicle object matching id vehicle found | 1 |
652,834 | 21,563,210,210 | IssuesEvent | 2022-05-01 13:28:49 | Polymer/tools | https://api.github.com/repos/Polymer/tools | closed | linter warns on item property of vaadin-grid templates | Package: linter Status: Available Priority: High Type: Bug wontfix | `<vaadin-grid>` uses the keyword `item` to represent iterated data points in it's grid templates, like so
```html
<vaadin-grid aria-label="Two-way Binding Example" items="[[users.result]]">
<vaadin-grid-column>
<template class="header">First Name</template>
<template>
<input value="{{item.firstName::input}}" readonly$="[[!editing]]">
</template>
</vaadin-grid-column>
<vaadin-grid-column>
<template class="header">Last Name</template>
<template>
<input value="{{item.lastName::input}}" readonly$="[[!editing]]">
</template>
</vaadin-grid-column>
<vaadin-grid-column width="200px">
<template class="header">Email</template>
<template>
<div>[[item.firstName]].[[item.lastName]]@example.com</div>
</template>
</vaadin-grid-column>
</vaadin-grid>
```
The above example does not trigger linting errors, however, if `[[item.prop]]` is bound in a child template, polymer linter warns.
```html
<vaadin-grid id="grid" name="grid" items="[[videos]]">
<vaadin-grid-column flex-grow="2">
<template class="header">Title</template>
<template>
<span class="title">[[item.videoData.title]]</span>
<template is="dom-if" if="[[!same(item.videoData.title, item.videoData.filename)]]" restamp="true">
<span class="filename">[[item.videoData.filename]]</span>
</template>
</template>
</vaadin-grid-column>
```
in which case, the binding expression `[[item.videoData.filename]]` triggers a warning. ` `item is not declared or used more than once.` | 1.0 | linter warns on item property of vaadin-grid templates - `<vaadin-grid>` uses the keyword `item` to represent iterated data points in it's grid templates, like so
```html
<vaadin-grid aria-label="Two-way Binding Example" items="[[users.result]]">
<vaadin-grid-column>
<template class="header">First Name</template>
<template>
<input value="{{item.firstName::input}}" readonly$="[[!editing]]">
</template>
</vaadin-grid-column>
<vaadin-grid-column>
<template class="header">Last Name</template>
<template>
<input value="{{item.lastName::input}}" readonly$="[[!editing]]">
</template>
</vaadin-grid-column>
<vaadin-grid-column width="200px">
<template class="header">Email</template>
<template>
<div>[[item.firstName]].[[item.lastName]]@example.com</div>
</template>
</vaadin-grid-column>
</vaadin-grid>
```
The above example does not trigger linting errors, however, if `[[item.prop]]` is bound in a child template, polymer linter warns.
```html
<vaadin-grid id="grid" name="grid" items="[[videos]]">
<vaadin-grid-column flex-grow="2">
<template class="header">Title</template>
<template>
<span class="title">[[item.videoData.title]]</span>
<template is="dom-if" if="[[!same(item.videoData.title, item.videoData.filename)]]" restamp="true">
<span class="filename">[[item.videoData.filename]]</span>
</template>
</template>
</vaadin-grid-column>
```
in which case, the binding expression `[[item.videoData.filename]]` triggers a warning. ` `item is not declared or used more than once.` | priority | linter warns on item property of vaadin grid templates uses the keyword item to represent iterated data points in it s grid templates like so html first name last name email example com the above example does not trigger linting errors however if is bound in a child template polymer linter warns html title in which case the binding expression triggers a warning item is not declared or used more than once | 1 |
830,664 | 32,020,460,471 | IssuesEvent | 2023-09-22 03:47:34 | dimensionhq/infralink | https://api.github.com/repos/dimensionhq/infralink | reopened | `build` Command | high-priority in-progress | This will be responsible for
* Building the OCI-compliant image (using nixpacks)
* Generating a terraform configuration based on the app & configuration (will then be deployed by the `deploy` command.
* Start off with support for AWS. | 1.0 | `build` Command - This will be responsible for
* Building the OCI-compliant image (using nixpacks)
* Generating a terraform configuration based on the app & configuration (will then be deployed by the `deploy` command.
* Start off with support for AWS. | priority | build command this will be responsible for building the oci compliant image using nixpacks generating a terraform configuration based on the app configuration will then be deployed by the deploy command start off with support for aws | 1 |
38,441 | 2,847,300,120 | IssuesEvent | 2015-05-29 16:12:30 | phetsims/tasks | https://api.github.com/repos/phetsims/tasks | closed | Dev test Curve Fitting 1.0.0-dev.4 | High Priority QA | @orejolaphet, @phet-steele, @bryo5363
Please dev test the latest version of Curve Fitting (http://www.colorado.edu/physics/phet/dev/html/curve-fitting/1.0.0-dev.4/curve-fitting_en.html)
- [x] Windows + Chrome
- [x] OS X + Safari 7.1+
- [x] iPad iOS 8 + Safari
Consider this testing the next priority after any final website testing | 1.0 | Dev test Curve Fitting 1.0.0-dev.4 - @orejolaphet, @phet-steele, @bryo5363
Please dev test the latest version of Curve Fitting (http://www.colorado.edu/physics/phet/dev/html/curve-fitting/1.0.0-dev.4/curve-fitting_en.html)
- [x] Windows + Chrome
- [x] OS X + Safari 7.1+
- [x] iPad iOS 8 + Safari
Consider this testing the next priority after any final website testing | priority | dev test curve fitting dev orejolaphet phet steele please dev test the latest version of curve fitting windows chrome os x safari ipad ios safari consider this testing the next priority after any final website testing | 1 |
351,206 | 10,514,140,863 | IssuesEvent | 2019-09-27 23:01:24 | xournalpp/xournalpp | https://api.github.com/repos/xournalpp/xournalpp | closed | Crash on export | Crash bug priority: high | I can reliably cause X++ to crash, by doing the following steps:
1. Open an XOPP file (or create a new one)
2. Click "file->export as pdf".
3. Enter a name, click "OK"
4. On "Export" dialog (where we can select page ranges), click "Cancel"
5. Draw a single stroke in X++
6. The program is now frozen.
I'm running latest master on linux.
| 1.0 | Crash on export - I can reliably cause X++ to crash, by doing the following steps:
1. Open an XOPP file (or create a new one)
2. Click "file->export as pdf".
3. Enter a name, click "OK"
4. On "Export" dialog (where we can select page ranges), click "Cancel"
5. Draw a single stroke in X++
6. The program is now frozen.
I'm running latest master on linux.
| priority | crash on export i can reliably cause x to crash by doing the following steps open an xopp file or create a new one click file export as pdf enter a name click ok on export dialog where we can select page ranges click cancel draw a single stroke in x the program is now frozen i m running latest master on linux | 1 |
366,856 | 10,831,562,897 | IssuesEvent | 2019-11-11 08:37:21 | crossminer/scava | https://api.github.com/repos/crossminer/scava | closed | Cannot interrupt task in worker | bug high priority | We need to be able to stop a running task in a worker. There is a button on the worker page for that, but nothing happens when clicked. It seems @mhow2 confirmed the issue on his instance. | 1.0 | Cannot interrupt task in worker - We need to be able to stop a running task in a worker. There is a button on the worker page for that, but nothing happens when clicked. It seems @mhow2 confirmed the issue on his instance. | priority | cannot interrupt task in worker we need to be able to stop a running task in a worker there is a button on the worker page for that but nothing happens when clicked it seems confirmed the issue on his instance | 1 |
396,778 | 11,713,929,314 | IssuesEvent | 2020-03-09 11:17:03 | CyberCRI/learn-ext | https://api.github.com/repos/CyberCRI/learn-ext | closed | [AI and Map] Personnalize map: Use tags as meta-categories | priority: high | User story: The user adds resources and wants to group them on the map by not only concepts but also a specific meta category. They create a tag corresponding to the category, annotate resources with this tag while adding. Then, they go to the map and chose to use tag a1, ...an as meta category (example: methodology).
They click on "regenerate map" and the map appears where resources are first grouped by meta category, then within each meta category by concepts. The regions for a specific meta category would be ideally in the background colored with a specific color and their name would be indicated if we zoom out of the concepts. The user then should keep this personalization while adding new resources.
- [ ] Dropdown menu called Meta categories and choice between None(default), and all tags used.
- [ ] Button to regenerate map
- [ ] An algorithm that takes into account this meta grouping first
- [ ] The annotation of the map with these meta-categories (and eventually the background color)
- [ ] Capacity to keep such a personalized visualization when the user logs out and logs in to account, or when user shares map.
| 1.0 | [AI and Map] Personnalize map: Use tags as meta-categories - User story: The user adds resources and wants to group them on the map by not only concepts but also a specific meta category. They create a tag corresponding to the category, annotate resources with this tag while adding. Then, they go to the map and chose to use tag a1, ...an as meta category (example: methodology).
They click on "regenerate map" and the map appears where resources are first grouped by meta category, then within each meta category by concepts. The regions for a specific meta category would be ideally in the background colored with a specific color and their name would be indicated if we zoom out of the concepts. The user then should keep this personalization while adding new resources.
- [ ] Dropdown menu called Meta categories and choice between None(default), and all tags used.
- [ ] Button to regenerate map
- [ ] An algorithm that takes into account this meta grouping first
- [ ] The annotation of the map with these meta-categories (and eventually the background color)
- [ ] Capacity to keep such a personalized visualization when the user logs out and logs in to account, or when user shares map.
| priority | personnalize map use tags as meta categories user story the user adds resources and wants to group them on the map by not only concepts but also a specific meta category they create a tag corresponding to the category annotate resources with this tag while adding then they go to the map and chose to use tag an as meta category example methodology they click on regenerate map and the map appears where resources are first grouped by meta category then within each meta category by concepts the regions for a specific meta category would be ideally in the background colored with a specific color and their name would be indicated if we zoom out of the concepts the user then should keep this personalization while adding new resources dropdown menu called meta categories and choice between none default and all tags used button to regenerate map an algorithm that takes into account this meta grouping first the annotation of the map with these meta categories and eventually the background color capacity to keep such a personalized visualization when the user logs out and logs in to account or when user shares map | 1 |
320,312 | 9,779,348,281 | IssuesEvent | 2019-06-07 14:18:52 | poanetwork/blockscout | https://api.github.com/repos/poanetwork/blockscout | closed | Store all blocks for the 1st page in ETS | in progress performance priority: high | Relates to https://github.com/poanetwork/blockscout/issues/1880
Store all blocks for the 1st page in ETS (There are 50 blocks by-default there). When new blocks are coming, add them to cache and remove the same amount of blocks from the tail of the cache. | 1.0 | Store all blocks for the 1st page in ETS - Relates to https://github.com/poanetwork/blockscout/issues/1880
Store all blocks for the 1st page in ETS (There are 50 blocks by-default there). When new blocks are coming, add them to cache and remove the same amount of blocks from the tail of the cache. | priority | store all blocks for the page in ets relates to store all blocks for the page in ets there are blocks by default there when new blocks are coming add them to cache and remove the same amount of blocks from the tail of the cache | 1 |
719,221 | 24,751,653,223 | IssuesEvent | 2022-10-21 14:11:33 | SuperCoopBerlin/tapir | https://api.github.com/repos/SuperCoopBerlin/tapir | closed | Improve accessible accounts on the test instance | high priority product | Currently on the test instance, there is only one account (admin/admin).
- Since the test instance is available publicly, we should probably use a more secure password, even if there is no sensitive data
- It would be nice to have member office access (currently it's almost the same as admin access but still) and normal member access
- While we're at it, let's add a Vorstand account (also almost the same as member office for now) | 1.0 | Improve accessible accounts on the test instance - Currently on the test instance, there is only one account (admin/admin).
- Since the test instance is available publicly, we should probably use a more secure password, even if there is no sensitive data
- It would be nice to have member office access (currently it's almost the same as admin access but still) and normal member access
- While we're at it, let's add a Vorstand account (also almost the same as member office for now) | priority | improve accessible accounts on the test instance currently on the test instance there is only one account admin admin since the test instance is available publicly we should probably use a more secure password even if there is no sensitive data it would be nice to have member office access currently it s almost the same as admin access but still and normal member access while we re at it let s add a vorstand account also almost the same as member office for now | 1 |
590,517 | 17,779,356,004 | IssuesEvent | 2021-08-31 00:51:40 | fuchicorp/main | https://api.github.com/repos/fuchicorp/main | closed | Deployment of ELK-Stack to bastion-host | Kubernetes Priority High Bastion elastic-search kibana basic | Deployment of ELK-Stack to the bastion-host.
Here's the Github repo with documentation to follow!
https://github.com/fuchicorp/elk-stack | 1.0 | Deployment of ELK-Stack to bastion-host - Deployment of ELK-Stack to the bastion-host.
Here's the Github repo with documentation to follow!
https://github.com/fuchicorp/elk-stack | priority | deployment of elk stack to bastion host deployment of elk stack to the bastion host here s the github repo with documentation to follow | 1 |
357,391 | 10,606,028,653 | IssuesEvent | 2019-10-10 21:57:06 | hdashnow/str-dev | https://api.github.com/repos/hdashnow/str-dev | closed | Open-ended and negative bounds | HIGH PRIORITY fixed pending merge | When only one of the bounds is known, set the others to the same value. For example this locus is reported as a "large bounds" but it's really just that the right bound is unknown.
Examples:
```
large bounds:(tid: 9, left: 5573387, right: 0, center_mass: 0, n_left: 3, n_right: 0, n_total: 3, repeat: "GG") skipping
large bounds:(tid: 9, left: 9515681, right: 0, center_mass: 0, n_left: 7, n_right: 0, n_total: 7, repeat: "GG") skipping
```
Also seeing some where right bound is before left bound.
Examples:
```
large bounds:(tid: 3, left: 6126536, right: 6125848, center_mass: 6126395, n_left: 2, n_right: 3, n_total: 8, repeat: "CA") skipping
large bounds:(tid: 3, left: 6726087, right: 6725725, center_mass: 6725953, n_left: 1, n_right: 1, n_total: 29, repeat: "CA") skipping
``` | 1.0 | Open-ended and negative bounds - When only one of the bounds is known, set the others to the same value. For example this locus is reported as a "large bounds" but it's really just that the right bound is unknown.
Examples:
```
large bounds:(tid: 9, left: 5573387, right: 0, center_mass: 0, n_left: 3, n_right: 0, n_total: 3, repeat: "GG") skipping
large bounds:(tid: 9, left: 9515681, right: 0, center_mass: 0, n_left: 7, n_right: 0, n_total: 7, repeat: "GG") skipping
```
Also seeing some where right bound is before left bound.
Examples:
```
large bounds:(tid: 3, left: 6126536, right: 6125848, center_mass: 6126395, n_left: 2, n_right: 3, n_total: 8, repeat: "CA") skipping
large bounds:(tid: 3, left: 6726087, right: 6725725, center_mass: 6725953, n_left: 1, n_right: 1, n_total: 29, repeat: "CA") skipping
``` | priority | open ended and negative bounds when only one of the bounds is known set the others to the same value for example this locus is reported as a large bounds but it s really just that the right bound is unknown examples large bounds tid left right center mass n left n right n total repeat gg skipping large bounds tid left right center mass n left n right n total repeat gg skipping also seeing some where right bound is before left bound examples large bounds tid left right center mass n left n right n total repeat ca skipping large bounds tid left right center mass n left n right n total repeat ca skipping | 1 |
508,288 | 14,697,891,624 | IssuesEvent | 2021-01-04 04:58:05 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Getting unreachable pattern for nested match-stmt | Priority/High Team/CompilerFE Type/Bug | **Description:**
$subject
**Steps to reproduce:**
```
public function main() {
string a = "foo";
string b = "bar";
match a {
"foo" => {
match b {
"bar" => {
}
_ => {
}
}
}
_ => {
}
}
}
```
**Affected Versions:**
slp5
| 1.0 | Getting unreachable pattern for nested match-stmt - **Description:**
$subject
**Steps to reproduce:**
```
public function main() {
string a = "foo";
string b = "bar";
match a {
"foo" => {
match b {
"bar" => {
}
_ => {
}
}
}
_ => {
}
}
}
```
**Affected Versions:**
slp5
| priority | getting unreachable pattern for nested match stmt description subject steps to reproduce public function main string a foo string b bar match a foo match b bar affected versions | 1 |
739,149 | 25,582,508,622 | IssuesEvent | 2022-12-01 06:16:10 | SzFMV2022-Osz/AutomatedCar-A | https://api.github.com/repos/SzFMV2022-Osz/AutomatedCar-A | closed | ACC funkcionális működése | effort: high priority: normal | Definition of Done
- [x] Ha nincs saját sávban előttünk autó, akkor a vezérelt autó tartja a kiválasztott célsebességet
Ha a saját sávban található autó:
- [x] Ha az előttünk levő autó lassabb, akkor fel kell venni a sebességét
- [x] Ha gyorsabb, akkor tartja a kiválasztott sebességet | 1.0 | ACC funkcionális működése - Definition of Done
- [x] Ha nincs saját sávban előttünk autó, akkor a vezérelt autó tartja a kiválasztott célsebességet
Ha a saját sávban található autó:
- [x] Ha az előttünk levő autó lassabb, akkor fel kell venni a sebességét
- [x] Ha gyorsabb, akkor tartja a kiválasztott sebességet | priority | acc funkcionális működése definition of done ha nincs saját sávban előttünk autó akkor a vezérelt autó tartja a kiválasztott célsebességet ha a saját sávban található autó ha az előttünk levő autó lassabb akkor fel kell venni a sebességét ha gyorsabb akkor tartja a kiválasztott sebességet | 1 |
552,755 | 16,325,716,863 | IssuesEvent | 2021-05-12 00:49:12 | openmsupply/msupply-cold-chain | https://api.github.com/repos/openmsupply/msupply-cold-chain | closed | remove a sensor | Priority: High | be able to disconnect/unpair/remove a sensor from a tablet (sensor is malfunctioning, sensor is needed in another cold storage...)
| 1.0 | remove a sensor - be able to disconnect/unpair/remove a sensor from a tablet (sensor is malfunctioning, sensor is needed in another cold storage...)
| priority | remove a sensor be able to disconnect unpair remove a sensor from a tablet sensor is malfunctioning sensor is needed in another cold storage | 1 |
355,032 | 10,575,947,922 | IssuesEvent | 2019-10-07 16:46:27 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | opened | v1.8.0 regression: Numpad keys clear hint keystring | bug: behavior component: keyinput priority: 0 - high | From IRC:
```
18:24 <dive> I've noticed that when using :hint -m number the digits vanish after the first number is pressed. Is there a way to keep them visible?
18:24 <dive> The same happens when using -m letter but characters are set to numbers.
18:26 <enot> dive: it seems to stay for me. highlighting the number pressed already
18:26 <dive> Strange. I tested with -T too.
18:27 <dive> brb
18:30 <dive> All I see in the log when I press numbers is something about forward_unbound_keys, but changing it to all the settings makes no difference.
18:33 <dive> WIth 1.8.1 the numbers all disappear after pressing the first one :/
18:33 <jgkamat> I can't reproduce, -T, hints.mode->number
18:34 <dive> At least the numbpad ones do. Didn't check the top row.
18:34 <jgkamat> oho I can reproduce that
18:34 <dive> Yeah the top row stay, but the numpad vanish :/
18:35 <dive> I thought it would be usefuil to use the numpad. I got fed up with home row because on slow loading pages I press 'f' then 'd' and the hints have vanished so the tab closes.
18:35 <dive> Switched to the top row, which is better, but numpad would be very nice.
18:37 <jgkamat> The-Compiler: 9557885ace0bc93257a36b1db420ccd722fa0b69
18:41 <The-Compiler> jgkamat: what about it?
18:42 <jgkamat> that's the first commit with the hints vanishing on the first press of a numpad key (what dive was talking about before)
```
commit: 9557885ace0bc93257a36b1db420ccd722fa0b69 | 1.0 | v1.8.0 regression: Numpad keys clear hint keystring - From IRC:
```
18:24 <dive> I've noticed that when using :hint -m number the digits vanish after the first number is pressed. Is there a way to keep them visible?
18:24 <dive> The same happens when using -m letter but characters are set to numbers.
18:26 <enot> dive: it seems to stay for me. highlighting the number pressed already
18:26 <dive> Strange. I tested with -T too.
18:27 <dive> brb
18:30 <dive> All I see in the log when I press numbers is something about forward_unbound_keys, but changing it to all the settings makes no difference.
18:33 <dive> WIth 1.8.1 the numbers all disappear after pressing the first one :/
18:33 <jgkamat> I can't reproduce, -T, hints.mode->number
18:34 <dive> At least the numbpad ones do. Didn't check the top row.
18:34 <jgkamat> oho I can reproduce that
18:34 <dive> Yeah the top row stay, but the numpad vanish :/
18:35 <dive> I thought it would be usefuil to use the numpad. I got fed up with home row because on slow loading pages I press 'f' then 'd' and the hints have vanished so the tab closes.
18:35 <dive> Switched to the top row, which is better, but numpad would be very nice.
18:37 <jgkamat> The-Compiler: 9557885ace0bc93257a36b1db420ccd722fa0b69
18:41 <The-Compiler> jgkamat: what about it?
18:42 <jgkamat> that's the first commit with the hints vanishing on the first press of a numpad key (what dive was talking about before)
```
commit: 9557885ace0bc93257a36b1db420ccd722fa0b69 | priority | regression numpad keys clear hint keystring from irc i ve noticed that when using hint m number the digits vanish after the first number is pressed is there a way to keep them visible the same happens when using m letter but characters are set to numbers dive it seems to stay for me highlighting the number pressed already strange i tested with t too brb all i see in the log when i press numbers is something about forward unbound keys but changing it to all the settings makes no difference with the numbers all disappear after pressing the first one i can t reproduce t hints mode number at least the numbpad ones do didn t check the top row oho i can reproduce that yeah the top row stay but the numpad vanish i thought it would be usefuil to use the numpad i got fed up with home row because on slow loading pages i press f then d and the hints have vanished so the tab closes switched to the top row which is better but numpad would be very nice the compiler jgkamat what about it that s the first commit with the hints vanishing on the first press of a numpad key what dive was talking about before commit | 1 |
475,432 | 13,710,281,544 | IssuesEvent | 2020-10-02 00:34:17 | RoboJackets/apiary | https://api.github.com/repos/RoboJackets/apiary | closed | Put back observers for model changes | area / jedi priority / high type / feature | - [ ] `DuesPackageObserver` scheduling expiration job when access expires
- [ ] `UserObserver` scheduling JEDI job when access override expires | 1.0 | Put back observers for model changes - - [ ] `DuesPackageObserver` scheduling expiration job when access expires
- [ ] `UserObserver` scheduling JEDI job when access override expires | priority | put back observers for model changes duespackageobserver scheduling expiration job when access expires userobserver scheduling jedi job when access override expires | 1 |
621,913 | 19,599,381,467 | IssuesEvent | 2022-01-05 22:21:13 | oilshell/oil | https://api.github.com/repos/oilshell/oil | closed | Better error message for == | high-priority pending-release errors user-feedback | lobste.rs user hit this. Also hit in #999
```
if (arg == '--hard') {
^
'/Users/timkellogg/bin/git-rebranch':19: Syntax error in expression (near Id.Arith_Equal)
``` | 1.0 | Better error message for == - lobste.rs user hit this. Also hit in #999
```
if (arg == '--hard') {
^
'/Users/timkellogg/bin/git-rebranch':19: Syntax error in expression (near Id.Arith_Equal)
``` | priority | better error message for lobste rs user hit this also hit in if arg hard users timkellogg bin git rebranch syntax error in expression near id arith equal | 1 |
744,300 | 25,937,595,170 | IssuesEvent | 2022-12-16 15:26:03 | trimble-oss/website-modus.trimble.com | https://api.github.com/repos/trimble-oss/website-modus.trimble.com | closed | Dark Mode General Description to Modus Foundations | 5 story priority:high dark-mode | As a user of Modus, I need to understand how to use the dark mode components in my design and development. | 1.0 | Dark Mode General Description to Modus Foundations - As a user of Modus, I need to understand how to use the dark mode components in my design and development. | priority | dark mode general description to modus foundations as a user of modus i need to understand how to use the dark mode components in my design and development | 1 |
224,531 | 7,471,633,528 | IssuesEvent | 2018-04-03 09:53:01 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Thread requirements in RFC4291 | area: Networking enhancement priority: high | **_Reported by Gajinder Vij:_**
The Thread Protocol imposes the following requirements on RFC4291:
RFC4291.2: IPv6 Addressing
RFC4291.2.1: Addressing Model
RFC4291.2.2: Text Representation of Addresses
RFC4291.2.3: Text Representation of Address Prefixes
RFC4291.2.4: Address Type Identification
RFC4291.2.5: Unicast Addresses
RFC4291.2.5.1: Interface Identifiers
RFC4291.2.5.2: The Unspecified Address
RFC4291.2.5.3: The Loopback Address
RFC4291.2.5.4: Global Unicast Addresses
RFC4291.2.5.6: Link-Local IPv6 Unicast Addresses
RFC4291.2.6: Anycast Addresses
RFC4291.2.8: A Node�s Required Addresses
� Note: A Node MAY recognize the loopback address
RFC4291.2.6.1: Required Anycast Address
RFC4291.2.7: Multicast Addresses
RFC4291.2.7.1: Pre-Defined Multicast Addresses
(Imported from Jira ZEP-836) | 1.0 | Thread requirements in RFC4291 - **_Reported by Gajinder Vij:_**
The Thread Protocol imposes the following requirements on RFC4291:
RFC4291.2: IPv6 Addressing
RFC4291.2.1: Addressing Model
RFC4291.2.2: Text Representation of Addresses
RFC4291.2.3: Text Representation of Address Prefixes
RFC4291.2.4: Address Type Identification
RFC4291.2.5: Unicast Addresses
RFC4291.2.5.1: Interface Identifiers
RFC4291.2.5.2: The Unspecified Address
RFC4291.2.5.3: The Loopback Address
RFC4291.2.5.4: Global Unicast Addresses
RFC4291.2.5.6: Link-Local IPv6 Unicast Addresses
RFC4291.2.6: Anycast Addresses
RFC4291.2.8: A Node�s Required Addresses
� Note: A Node MAY recognize the loopback address
RFC4291.2.6.1: Required Anycast Address
RFC4291.2.7: Multicast Addresses
RFC4291.2.7.1: Pre-Defined Multicast Addresses
(Imported from Jira ZEP-836) | priority | thread requirements in reported by gajinder vij the thread protocol imposes the following requirements on addressing addressing model text representation of addresses text representation of address prefixes address type identification unicast addresses interface identifiers the unspecified address the loopback address global unicast addresses link local unicast addresses anycast addresses a node�s required addresses � note a node may recognize the loopback address required anycast address multicast addresses pre defined multicast addresses imported from jira zep | 1 |
759,682 | 26,605,884,303 | IssuesEvent | 2023-01-23 19:17:59 | KlimaDAO/klimadao | https://api.github.com/repos/KlimaDAO/klimadao | closed | [connect modal] Don't assume window.ethereum is always metamask | bug HIGH PRIORITY QA needed | If you use the coinbase mobile app and load the site in the browser, it still injects `window.ethereum`
https://www.coinbase.com/cloud/discover/dev-foundations/three-ways-to-integrate-coinbase-wallet
So this line of code is wrong, and falsely assumes its metamask, when it's coinbase browser.
https://github.com/KlimaDAO/klimadao/blob/9d32d9678e53dea2566fd45ca3b61b4283fc6803/lib/components/ConnectModal/ConnectContent.tsx#L46
* [x] 1. look for `isCoinbaseWallet` in `window.ethereum` and handle accordingly. In this case we don't use the sdk, just use the window.ethereum provider directly.
* [x] 2. if not metamask, not brave, not coinbase, but window.ethereum still exists, show a generic "Connect browser wallet" button. Improvise the style/icon for it I trust ya | 1.0 | [connect modal] Don't assume window.ethereum is always metamask - If you use the coinbase mobile app and load the site in the browser, it still injects `window.ethereum`
https://www.coinbase.com/cloud/discover/dev-foundations/three-ways-to-integrate-coinbase-wallet
So this line of code is wrong, and falsely assumes its metamask, when it's coinbase browser.
https://github.com/KlimaDAO/klimadao/blob/9d32d9678e53dea2566fd45ca3b61b4283fc6803/lib/components/ConnectModal/ConnectContent.tsx#L46
* [x] 1. look for `isCoinbaseWallet` in `window.ethereum` and handle accordingly. In this case we don't use the sdk, just use the window.ethereum provider directly.
* [x] 2. if not metamask, not brave, not coinbase, but window.ethereum still exists, show a generic "Connect browser wallet" button. Improvise the style/icon for it I trust ya | priority | don t assume window ethereum is always metamask if you use the coinbase mobile app and load the site in the browser it still injects window ethereum so this line of code is wrong and falsely assumes its metamask when it s coinbase browser look for iscoinbasewallet in window ethereum and handle accordingly in this case we don t use the sdk just use the window ethereum provider directly if not metamask not brave not coinbase but window ethereum still exists show a generic connect browser wallet button improvise the style icon for it i trust ya | 1 |
754,149 | 26,373,788,342 | IssuesEvent | 2023-01-11 23:26:03 | bcgov/cas-cif | https://api.github.com/repos/bcgov/cas-cif | closed | Identify and confirm the scope of the proponent-facing portal for 2023 application intake | Task UX Research High Priority | #### Describe the task
Shelby has indicated the need to have a proponent-facing portal/tool **before March 2023** for the CIF team to handle the 2023 application intake (and for the proponents to submit their application). Due to the urgency of it, the developers suggested to have a clearly defined scope of features for this MVP, so that we can plan the most effective solution to address the need.
#### Acceptance Criteria
- [x] Draft a list of required features based on previous discussions with Shelby, Jesse, and Amanda
- [ ] Review, refine, and finalize the feature list engaging the CIF team and us
#### Additional context
- Initial feedback from CIF team in relation to required fields/features for an MVP include: funding stream, funding year, legal name, and allowing for attachments
- Nice to have: auto generation of RFP number
- Question for devs: can we automate funding year for proponents submitting an application? | 1.0 | Identify and confirm the scope of the proponent-facing portal for 2023 application intake - #### Describe the task
Shelby has indicated the need to have a proponent-facing portal/tool **before March 2023** for the CIF team to handle the 2023 application intake (and for the proponents to submit their application). Due to the urgency of it, the developers suggested to have a clearly defined scope of features for this MVP, so that we can plan the most effective solution to address the need.
#### Acceptance Criteria
- [x] Draft a list of required features based on previous discussions with Shelby, Jesse, and Amanda
- [ ] Review, refine, and finalize the feature list engaging the CIF team and us
#### Additional context
- Initial feedback from CIF team in relation to required fields/features for an MVP include: funding stream, funding year, legal name, and allowing for attachments
- Nice to have: auto generation of RFP number
- Question for devs: can we automate funding year for proponents submitting an application? | priority | identify and confirm the scope of the proponent facing portal for application intake describe the task shelby has indicated the need to have a proponent facing portal tool before march for the cif team to handle the application intake and for the proponents to submit their application due to the urgency of it the developers suggested to have a clearly defined scope of features for this mvp so that we can plan the most effective solution to address the need acceptance criteria draft a list of required features based on previous discussions with shelby jesse and amanda review refine and finalize the feature list engaging the cif team and us additional context initial feedback from cif team in relation to required fields features for an mvp include funding stream funding year legal name and allowing for attachments nice to have auto generation of rfp number question for devs can we automate funding year for proponents submitting an application | 1 |
440,173 | 12,694,627,608 | IssuesEvent | 2020-06-22 06:56:06 | radareorg/radare2 | https://api.github.com/repos/radareorg/radare2 | closed | drtj all broken | RDebug high-priority | # Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits (mandatory) | Ubuntu x86 64
| File format of the file you reverse (mandatory) | ELF
| Architecture/bits of the file (mandatory) | x86/64
| r2 -v full output, **not truncated** (mandatory) | radare2 4.5.0-git 1 @ linux-x86-64 git.4.5.0-git commit: dae5bb43ee8e3c6ba9964caf6569ce28918dd000 build: 2020-04-16__23:30:26
### Expected behavior
So the command `drtj all` used to return all registers r2 knows about, as a json object. At some point, this stopped working and it has broken my angr symbion integration.
### Actual behavior
```
drt?
Usage: drt [type] [size] # debug register types
| drt List all available register types
| drt [size] Show all regs in the profile of size
| drt 16 Show 16 bit registers
| drt [type] Show all regs in the profile of this type
| drt all Show all registers
| drt fpu Show fpu registers
| drt [type] [size] Same as above for type and size
| drt [type] [size] Same as above for type and size
| drt* List flags in r commands
[0x7f1634629090]> drt all
[0x7f1634629090]> drtj all
```
### Steps to reproduce the behavior
- Just try to use `drtj all` on any binary. In my case, the example is the vanilla /bin/ls.
### Additional Logs, screenshots, source-code, configuration dump, ...
| 1.0 | drtj all broken - # Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits (mandatory) | Ubuntu x86 64
| File format of the file you reverse (mandatory) | ELF
| Architecture/bits of the file (mandatory) | x86/64
| r2 -v full output, **not truncated** (mandatory) | radare2 4.5.0-git 1 @ linux-x86-64 git.4.5.0-git commit: dae5bb43ee8e3c6ba9964caf6569ce28918dd000 build: 2020-04-16__23:30:26
### Expected behavior
So the command `drtj all` used to return all registers r2 knows about, as a json object. At some point, this stopped working and it has broken my angr symbion integration.
### Actual behavior
```
drt?
Usage: drt [type] [size] # debug register types
| drt List all available register types
| drt [size] Show all regs in the profile of size
| drt 16 Show 16 bit registers
| drt [type] Show all regs in the profile of this type
| drt all Show all registers
| drt fpu Show fpu registers
| drt [type] [size] Same as above for type and size
| drt [type] [size] Same as above for type and size
| drt* List flags in r commands
[0x7f1634629090]> drt all
[0x7f1634629090]> drtj all
```
### Steps to reproduce the behavior
- Just try to use `drtj all` on any binary. In my case, the example is the vanilla /bin/ls.
### Additional Logs, screenshots, source-code, configuration dump, ...
| priority | drtj all broken work environment questions answers os arch bits mandatory ubuntu file format of the file you reverse mandatory elf architecture bits of the file mandatory v full output not truncated mandatory git linux git git commit build expected behavior so the command drtj all used to return all registers knows about as a json object at some point this stopped working and it has broken my angr symbion integration actual behavior drt usage drt debug register types drt list all available register types drt show all regs in the profile of size drt show bit registers drt show all regs in the profile of this type drt all show all registers drt fpu show fpu registers drt same as above for type and size drt same as above for type and size drt list flags in r commands drt all drtj all steps to reproduce the behavior just try to use drtj all on any binary in my case the example is the vanilla bin ls additional logs screenshots source code configuration dump | 1 |
378,403 | 11,201,579,561 | IssuesEvent | 2020-01-04 04:47:02 | sethballantyne/Game-Demos | https://api.github.com/repos/sethballantyne/Game-Demos | closed | Replace powerup sprites | Plexis-Game enhancement priority-high | All the power up sprites need to be replaced; keep the colour scheme and dimensions but make them less shit. :P | 1.0 | Replace powerup sprites - All the power up sprites need to be replaced; keep the colour scheme and dimensions but make them less shit. :P | priority | replace powerup sprites all the power up sprites need to be replaced keep the colour scheme and dimensions but make them less shit p | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.