Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
622,324 | 19,621,403,909 | IssuesEvent | 2022-01-07 07:14:26 | tgpethan/EUS | https://api.github.com/repos/tgpethan/EUS | closed | Migrate to using database. | enhancement High priority | Currently as it stands EUS stores every single image in a json file in it's module folder. This is pretty bad ***json is not a database*** and it shouldn't be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided. | 1.0 | Migrate to using database. - Currently as it stands EUS stores every single image in a json file in it's module folder. This is pretty bad ***json is not a database*** and it shouldn't be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided. | priority | migrate to using database currently as it stands eus stores every single image in a json file in it s module folder this is pretty bad json is not a database and it shouldn t be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided | 1 |
652,196 | 21,525,055,478 | IssuesEvent | 2022-04-28 17:32:56 | ooni/explorer | https://api.github.com/repos/ooni/explorer | closed | Help text for MAT | priority/high MAT | To enable use of the MAT, we can include the following copy under the "Help" section of the MAT:
```
# What is the MAT?
OONI's Measurement Aggregation Toolkit (MAT) is a tool that enables you to generate your own custom charts based on **aggregate views of real-time OONI data** collected from around the world.
OONI data consists of network measurements collected by [OONI Probe](https://ooni.org/install/) users around the world.
These measurements contain information about various types of **internet censorship**, such as the [blocking of websites and apps](https://ooni.org/nettest/) around the world.
# Who is the MAT for?
The MAT was built for researchers, journalists, and human rights defenders interested in examining internet censorship around the world.
# Why use the MAT?
When examining cases of internet censorship, it's important to **look at many measurements at once** ("in aggregate") in order to answer key questions like the following:
* Does the testing of a service (e.g. Facebook) present **signs of blocking every time that it is tested** in a country? This can be helpful for ruling out [false positives](https://ooni.org/support/faq/#what-are-false-positives).
* What types of websites (e.g. human rights websites) are blocked in each country?
* In which countries is a specific website (e.g. `bbc.com`) blocked?
* How does the blocking of different apps (e.g. WhatsApp or Telegram) vary across countries?
* How does the blocking of a service vary across countries and [ASNs](https://ooni.org/support/glossary/#asn)?
* How does the blocking of a service change over time?
When trying to answer questions like the above, we normally perform relevant data analysis (instead of inspecting measurements one by one).
The MAT incorporates our data analysis techniques, enabling you to answer such questions without any data analysis skills, and with the click of a button!
# How to use the MAT?
Through the filters at the start of the page, select the parameters you care about in order to plot charts based on aggregate views of OONI data.
The MAT includes the following filters:
* **Countries:** Select a country through the drop-down menu (the "All Countries" option will show global coverage)
* **Test Name:** Select an [OONI Probe test](https://ooni.org/nettest/) based on which you would like to get measurements (for example, select `Web Connectivity` to view the testing of websites)
* **Domain:** Type the domain for the website you would like to get measurements (e.g. `twitter.com`)
* **Website categories:** Select the [website category](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv) for which you would like to get measurements (e.g. `News Media` for news media websites)
* **ASN:** Type the [ASN](https://ooni.org/support/glossary/#asn) of the network for which you would like to get measurements (e.g. `AS30722` for Vodafone Italia)
* **Date range:** Select the date range of the measurements by adjusting the `Since` and `Until` filters
* **X axis:** Select the values that you would like to appear on the horizontal axis of your chart
* **Y axis:** Select the values that you would like to appear on the vertical axis of your chart
Depending on what you would like to explore, adjust the MAT filters accordingly and click `Submit`.
For example, if you would like to check the testing of BBC in all countries around the world:
* Type `www.bbc.com` under `Domain`
* Select `Countries` under the `Y axis`
* Click `Submit`
This will plot numerous charts based on the OONI Probe testing of `www.bbc.com` worldwide.
# Interpreting MAT charts
The MAT charts (and associated tables) include the following values:
* **OK count:** Successful measurements (i.e. NO sign of internet censorship)
* **Confirmed count:** Measurements from automatically **confirmed blocked websites** (e.g. a [block page](https://ooni.org/support/glossary/#block-page) was served)
* **Anomaly count:** Measurements that provided **signs of potential blocking** (however, [false positives](https://ooni.org/support/faq/#what-are-false-positives) can occur)
* **Failure count:** Failed experiments that should be discarded
* **Measurement count:** Total volume of OONI measurements (pertaining to the selected country, resource, etc.)
When trying to identify the blocking of a service (e.g. `twitter.com`), it's useful to check whether:
* Measurements are annotated as `confirmed`, automatically confirming the blocking of websites
* A large volume of measurements (in comparison to the overall measurement count) present `anomalies` (i.e. signs of potential censorship)
You can access the raw data by clicking on the bars of charts, and subsequently clicking on the relevant measurement links.
# Website categories
[OONI Probe](https://ooni.org/install/) users test a wide range of [websites](https://ooni.org/support/faq/#which-websites-will-i-test-for-censorship-with-ooni-probe) that fall under the following [30 standardized categories](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv).
``` | 1.0 | Help text for MAT - To enable use of the MAT, we can include the following copy under the "Help" section of the MAT:
```
# What is the MAT?
OONI's Measurement Aggregation Toolkit (MAT) is a tool that enables you to generate your own custom charts based on **aggregate views of real-time OONI data** collected from around the world.
OONI data consists of network measurements collected by [OONI Probe](https://ooni.org/install/) users around the world.
These measurements contain information about various types of **internet censorship**, such as the [blocking of websites and apps](https://ooni.org/nettest/) around the world.
# Who is the MAT for?
The MAT was built for researchers, journalists, and human rights defenders interested in examining internet censorship around the world.
# Why use the MAT?
When examining cases of internet censorship, it's important to **look at many measurements at once** ("in aggregate") in order to answer key questions like the following:
* Does the testing of a service (e.g. Facebook) present **signs of blocking every time that it is tested** in a country? This can be helpful for ruling out [false positives](https://ooni.org/support/faq/#what-are-false-positives).
* What types of websites (e.g. human rights websites) are blocked in each country?
* In which countries is a specific website (e.g. `bbc.com`) blocked?
* How does the blocking of different apps (e.g. WhatsApp or Telegram) vary across countries?
* How does the blocking of a service vary across countries and [ASNs](https://ooni.org/support/glossary/#asn)?
* How does the blocking of a service change over time?
When trying to answer questions like the above, we normally perform relevant data analysis (instead of inspecting measurements one by one).
The MAT incorporates our data analysis techniques, enabling you to answer such questions without any data analysis skills, and with the click of a button!
# How to use the MAT?
Through the filters at the start of the page, select the parameters you care about in order to plot charts based on aggregate views of OONI data.
The MAT includes the following filters:
* **Countries:** Select a country through the drop-down menu (the "All Countries" option will show global coverage)
* **Test Name:** Select an [OONI Probe test](https://ooni.org/nettest/) based on which you would like to get measurements (for example, select `Web Connectivity` to view the testing of websites)
* **Domain:** Type the domain for the website you would like to get measurements (e.g. `twitter.com`)
* **Website categories:** Select the [website category](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv) for which you would like to get measurements (e.g. `News Media` for news media websites)
* **ASN:** Type the [ASN](https://ooni.org/support/glossary/#asn) of the network for which you would like to get measurements (e.g. `AS30722` for Vodafone Italia)
* **Date range:** Select the date range of the measurements by adjusting the `Since` and `Until` filters
* **X axis:** Select the values that you would like to appear on the horizontal axis of your chart
* **Y axis:** Select the values that you would like to appear on the vertical axis of your chart
Depending on what you would like to explore, adjust the MAT filters accordingly and click `Submit`.
For example, if you would like to check the testing of BBC in all countries around the world:
* Type `www.bbc.com` under `Domain`
* Select `Countries` under the `Y axis`
* Click `Submit`
This will plot numerous charts based on the OONI Probe testing of `www.bbc.com` worldwide.
# Interpreting MAT charts
The MAT charts (and associated tables) include the following values:
* **OK count:** Successful measurements (i.e. NO sign of internet censorship)
* **Confirmed count:** Measurements from automatically **confirmed blocked websites** (e.g. a [block page](https://ooni.org/support/glossary/#block-page) was served)
* **Anomaly count:** Measurements that provided **signs of potential blocking** (however, [false positives](https://ooni.org/support/faq/#what-are-false-positives) can occur)
* **Failure count:** Failed experiments that should be discarded
* **Measurement count:** Total volume of OONI measurements (pertaining to the selected country, resource, etc.)
When trying to identify the blocking of a service (e.g. `twitter.com`), it's useful to check whether:
* Measurements are annotated as `confirmed`, automatically confirming the blocking of websites
* A large volume of measurements (in comparison to the overall measurement count) present `anomalies` (i.e. signs of potential censorship)
You can access the raw data by clicking on the bars of charts, and subsequently clicking on the relevant measurement links.
# Website categories
[OONI Probe](https://ooni.org/install/) users test a wide range of [websites](https://ooni.org/support/faq/#which-websites-will-i-test-for-censorship-with-ooni-probe) that fall under the following [30 standardized categories](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv).
``` | priority | help text for mat to enable use of the mat we can include the following copy under the help section of the mat what is the mat ooni s measurement aggregation toolkit mat is a tool that enables you to generate your own custom charts based on aggregate views of real time ooni data collected from around the world ooni data consists of network measurements collected by users around the world these measurements contain information about various types of internet censorship such as the around the world who is the mat for the mat was built for researchers journalists and human rights defenders interested in examining internet censorship around the world why use the mat when examining cases of internet censorship it s important to look at many measurements at once in aggregate in order to answer key questions like the following does the testing of a service e g facebook present signs of blocking every time that it is tested in a country this can be helpful for ruling out what types of websites e g human rights websites are blocked in each country in which countries is a specific website e g bbc com blocked how does the blocking of different apps e g whatsapp or telegram vary across countries how does the blocking of a service vary across countries and how does the blocking of a service change over time when trying to answer questions like the above we normally perform relevant data analysis instead of inspecting measurements one by one the mat incorporates our data analysis techniques enabling you to answer such questions without any data analysis skills and with the click of a button how to use the mat through the filters at the start of the page select the parameters you care about in order to plot charts based on aggregate views of ooni data the mat includes the following filters countries select a country through the drop down menu the all countries option will show global coverage test name select an based on which you would like to get measurements for example select web connectivity to view the testing of websites domain type the domain for the website you would like to get measurements e g twitter com website categories select the for which you would like to get measurements e g news media for news media websites asn type the of the network for which you would like to get measurements e g for vodafone italia date range select the date range of the measurements by adjusting the since and until filters x axis select the values that you would like to appear on the horizontal axis of your chart y axis select the values that you would like to appear on the vertical axis of your chart depending on what you would like to explore adjust the mat filters accordingly and click submit for example if you would like to check the testing of bbc in all countries around the world type under domain select countries under the y axis click submit this will plot numerous charts based on the ooni probe testing of worldwide interpreting mat charts the mat charts and associated tables include the following values ok count successful measurements i e no sign of internet censorship confirmed count measurements from automatically confirmed blocked websites e g a was served anomaly count measurements that provided signs of potential blocking however can occur failure count failed experiments that should be discarded measurement count total volume of ooni measurements pertaining to the selected country resource etc when trying to identify the blocking of a service e g twitter com it s useful to check whether measurements are annotated as confirmed automatically confirming the blocking of websites a large volume of measurements in comparison to the overall measurement count present anomalies i e signs of potential censorship you can access the raw data by clicking on the bars of charts and subsequently clicking on the relevant measurement links website categories users test a wide range of that fall under the following | 1 |
518,736 | 15,033,774,767 | IssuesEvent | 2021-02-02 11:58:10 | opentargets/platform | https://api.github.com/repos/opentargets/platform | closed | I would like to have an example for Reactome for the new JSON schema | Kind: Data Priority: High | Provide an example evidence from Reactome of how it looks with the current JSON schema and how it should look like when they use the new schema | 1.0 | I would like to have an example for Reactome for the new JSON schema - Provide an example evidence from Reactome of how it looks with the current JSON schema and how it should look like when they use the new schema | priority | i would like to have an example for reactome for the new json schema provide an example evidence from reactome of how it looks with the current json schema and how it should look like when they use the new schema | 1 |
146,910 | 5,630,412,514 | IssuesEvent | 2017-04-05 12:12:48 | CS2103JAN2017-T11-B2/main | https://api.github.com/repos/CS2103JAN2017-T11-B2/main | closed | V0.5rc Documentation | priority.high type.task | All .md files need to be updated, including UserGuide, DeveloperGuide, AboutUs, and README | 1.0 | V0.5rc Documentation - All .md files need to be updated, including UserGuide, DeveloperGuide, AboutUs, and README | priority | documentation all md files need to be updated including userguide developerguide aboutus and readme | 1 |
470,009 | 13,529,607,776 | IssuesEvent | 2020-09-15 18:34:12 | Kedyn/fusliez-notes | https://api.github.com/repos/Kedyn/fusliez-notes | closed | Add title attribute to h1 input so the user knows its editable | Priority: High Status: Pending Type: Maintenance | In my fork I had a title attribute on the h1 input that read "Click to edit". it's not obvious that it can be edited. | 1.0 | Add title attribute to h1 input so the user knows its editable - In my fork I had a title attribute on the h1 input that read "Click to edit". it's not obvious that it can be edited. | priority | add title attribute to input so the user knows its editable in my fork i had a title attribute on the input that read click to edit it s not obvious that it can be edited | 1 |
803,779 | 29,189,021,261 | IssuesEvent | 2023-05-19 18:03:43 | minio/docs | https://api.github.com/repos/minio/docs | opened | [RELEASE] MinIO RELEASE.2023-05-18T00-05-36Z doc changes | priority: high | **Summary**
See https://github.com/minio/minio/releases/tag/RELEASE.2023-05-18T00-05-36Z for full changelog
** ToDo
- [ ] Persistent Queue Store for system/audit logs - [PR 17121](https://github.com/minio/minio/pull/17121)
- [ ] Max policy size of 2KiB for Service Account / STS policies (clarify w/ engineer) - [PR 17161](https://github.com/minio/minio/pull/17167)
- [ ] Webhook usage metrics [PR 17179)(https://github.com/minio/minio/pull/17179)
- [ ] healing updates parity based on current Storage Class - clarify w/ engineer [PR 17187](https://github.com/minio/minio/pull/17187)
- [ ] Improved support for topology changes during decomm - clarify w/ engineer [PR 17221](https://github.com/minio/minio/pull/17221)
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | [RELEASE] MinIO RELEASE.2023-05-18T00-05-36Z doc changes - **Summary**
See https://github.com/minio/minio/releases/tag/RELEASE.2023-05-18T00-05-36Z for full changelog
** ToDo
- [ ] Persistent Queue Store for system/audit logs - [PR 17121](https://github.com/minio/minio/pull/17121)
- [ ] Max policy size of 2KiB for Service Account / STS policies (clarify w/ engineer) - [PR 17161](https://github.com/minio/minio/pull/17167)
- [ ] Webhook usage metrics [PR 17179)(https://github.com/minio/minio/pull/17179)
- [ ] healing updates parity based on current Storage Class - clarify w/ engineer [PR 17187](https://github.com/minio/minio/pull/17187)
- [ ] Improved support for topology changes during decomm - clarify w/ engineer [PR 17221](https://github.com/minio/minio/pull/17221)
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | minio release doc changes summary see for full changelog todo persistent queue store for system audit logs max policy size of for service account sts policies clarify w engineer webhook usage metrics pr healing updates parity based on current storage class clarify w engineer improved support for topology changes during decomm clarify w engineer additional context add any other context or screenshots about the feature request here | 1 |
751,708 | 26,254,527,449 | IssuesEvent | 2023-01-05 22:47:30 | lambdaclass/cairo-rs | https://api.github.com/repos/lambdaclass/cairo-rs | closed | Abstract the representation of field elements | high-priority | We're currently using `BigInt` explicitly, which forces us to compute the `.mod_floor` of results after every operation, use costly divisions, etc.
The first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way. | 1.0 | Abstract the representation of field elements - We're currently using `BigInt` explicitly, which forces us to compute the `.mod_floor` of results after every operation, use costly divisions, etc.
The first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way. | priority | abstract the representation of field elements we re currently using bigint explicitly which forces us to compute the mod floor of results after every operation use costly divisions etc the first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way | 1 |
479,386 | 13,795,650,187 | IssuesEvent | 2020-10-09 18:23:48 | vanjarosoftware/Vanjaro.Platform | https://api.github.com/repos/vanjarosoftware/Vanjaro.Platform | closed | Add resource file in Authentication package | Area: Backend Priority: High Release: Minor | extract in file in below path
DesktopModules\AuthenticationServices\Vanjaro\App_LocalResources | 1.0 | Add resource file in Authentication package - extract in file in below path
DesktopModules\AuthenticationServices\Vanjaro\App_LocalResources | priority | add resource file in authentication package extract in file in below path desktopmodules authenticationservices vanjaro app localresources | 1 |
565,991 | 16,777,734,309 | IssuesEvent | 2021-06-15 00:52:57 | myConsciousness/twitter-bot-j | https://api.github.com/repos/myConsciousness/twitter-bot-j | opened | 不要な差分シンボルの追加処理がある | Priority: high Problem: bug | # High Priotity Bug Report
## 1. Bug Details
数値が負数であった場合にその値を文字列に変換すると既に負数を表す「-」が付いている状態だが、
現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「-」が2つ付いている状態で値が出力される。
```java
private String toReportCount(@NonNull final Difference difference) {
return switch (difference.getDifferenceType()) {
case NONE -> DifferenceSymbolUtils.toNoneString(difference.getValue());
case INCREASE -> DifferenceSymbolUtils.toIncreaseString(difference.getValue());
case DECREASE -> DifferenceSymbolUtils.toDecreaseString(difference.getValue());
};
}
```
## 2. What you did caused that bug
運用時の出力結果で確認。
## 3. How it should be
上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。
## 4. References
| 1.0 | 不要な差分シンボルの追加処理がある - # High Priotity Bug Report
## 1. Bug Details
数値が負数であった場合にその値を文字列に変換すると既に負数を表す「-」が付いている状態だが、
現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「-」が2つ付いている状態で値が出力される。
```java
private String toReportCount(@NonNull final Difference difference) {
return switch (difference.getDifferenceType()) {
case NONE -> DifferenceSymbolUtils.toNoneString(difference.getValue());
case INCREASE -> DifferenceSymbolUtils.toIncreaseString(difference.getValue());
case DECREASE -> DifferenceSymbolUtils.toDecreaseString(difference.getValue());
};
}
```
## 2. What you did caused that bug
運用時の出力結果で確認。
## 3. How it should be
上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。
## 4. References
| priority | 不要な差分シンボルの追加処理がある high priotity bug report bug details 数値が負数であった場合にその値を文字列に変換すると既に負数を表す「 」が付いている状態だが、 現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「 」 。 java private string toreportcount nonnull final difference difference return switch difference getdifferencetype case none differencesymbolutils tononestring difference getvalue case increase differencesymbolutils toincreasestring difference getvalue case decrease differencesymbolutils todecreasestring difference getvalue what you did caused that bug 運用時の出力結果で確認。 how it should be 上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。 references | 1 |
363,605 | 10,745,079,350 | IssuesEvent | 2019-10-30 08:11:59 | execom-eu/hawaii | https://api.github.com/repos/execom-eu/hawaii | closed | Manually changing leave profile for employee crashes the application | fix frontend high priority | Steps to reproduce (must have HR_MANAGER role):
- Go to administration and select Employees tab
- Find your profile information and change leave profile
- Save the change, and go to the dashboard
Expected result:
- New leave profile is set and the corresponding allowance values are set
Actual result:
- Application crashes with an error message (image below)

| 1.0 | Manually changing leave profile for employee crashes the application - Steps to reproduce (must have HR_MANAGER role):
- Go to administration and select Employees tab
- Find your profile information and change leave profile
- Save the change, and go to the dashboard
Expected result:
- New leave profile is set and the corresponding allowance values are set
Actual result:
- Application crashes with an error message (image below)

| priority | manually changing leave profile for employee crashes the application steps to reproduce must have hr manager role go to administration and select employees tab find your profile information and change leave profile save the change and go to the dashboard expected result new leave profile is set and the corresponding allowance values are set actual result application crashes with an error message image below | 1 |
811,732 | 30,297,940,309 | IssuesEvent | 2023-07-10 01:48:30 | steedos/steedos-platform | https://api.github.com/repos/steedos/steedos-platform | closed | [Bug]: 导入数据,lookup字段的“关联失败时保存key”功能无效 | bug done priority: High | ### Description
底层代码问题,当关联失败时,判断数据条数的地方已经通过 .length 转换成数值了,后面的if里面又来了一次 .length 造成判断无效,全部进入最后的else

### Steps To Reproduce 重现步骤
1. 配置数据导入
2. 配置关联表字段的导入
3. 配置对应的 “关联失败时保存key” 为打勾状态
4. 保存
5. 到对应的对象上导入数据
### Version 版本
所有版本 | 1.0 | [Bug]: 导入数据,lookup字段的“关联失败时保存key”功能无效 - ### Description
底层代码问题,当关联失败时,判断数据条数的地方已经通过 .length 转换成数值了,后面的if里面又来了一次 .length 造成判断无效,全部进入最后的else

### Steps To Reproduce 重现步骤
1. 配置数据导入
2. 配置关联表字段的导入
3. 配置对应的 “关联失败时保存key” 为打勾状态
4. 保存
5. 到对应的对象上导入数据
### Version 版本
所有版本 | priority | 导入数据,lookup字段的“关联失败时保存key”功能无效 description 底层代码问题,当关联失败时,判断数据条数的地方已经通过 length 转换成数值了,后面的if里面又来了一次 length 造成判断无效,全部进入最后的else steps to reproduce 重现步骤 配置数据导入 配置关联表字段的导入 配置对应的 “关联失败时保存key” 为打勾状态 保存 到对应的对象上导入数据 version 版本 所有版本 | 1 |
340,390 | 10,271,858,700 | IssuesEvent | 2019-08-23 15:02:23 | storybookjs/storybook | https://api.github.com/repos/storybookjs/storybook | closed | Addon-docs: User #root styles breaks Docs tab | addon: docs bug high priority ui | ### Problem
User applied the following global style to the preview iframe:
```css
#root {
height: 100vh;
display: flex;
flex-direction: column;
}
```
This breaks docs in the following way. When the user click on the `Docs` tab, Storybook applies the `hidden` attribute to `#root`, which triggers the following browser CSS:
```css
[hidden] {
display: none;
}
```
However, the `#root` CSS is more specific, so the Story renders on top of the docs.
### Solution
When we added the following CSS to the user's code, it fixed it.
```css
#root[hidden] {
display: none;
}
```
We can add this to Storybook itself to avoid this issue for users that style `#root` (which is a completely reasonable thing to do).
However, I'm not sure this is enough...
| 1.0 | Addon-docs: User #root styles breaks Docs tab - ### Problem
User applied the following global style to the preview iframe:
```css
#root {
height: 100vh;
display: flex;
flex-direction: column;
}
```
This breaks docs in the following way. When the user click on the `Docs` tab, Storybook applies the `hidden` attribute to `#root`, which triggers the following browser CSS:
```css
[hidden] {
display: none;
}
```
However, the `#root` CSS is more specific, so the Story renders on top of the docs.
### Solution
When we added the following CSS to the user's code, it fixed it.
```css
#root[hidden] {
display: none;
}
```
We can add this to Storybook itself to avoid this issue for users that style `#root` (which is a completely reasonable thing to do).
However, I'm not sure this is enough...
| priority | addon docs user root styles breaks docs tab problem user applied the following global style to the preview iframe css root height display flex flex direction column this breaks docs in the following way when the user click on the docs tab storybook applies the hidden attribute to root which triggers the following browser css css display none however the root css is more specific so the story renders on top of the docs solution when we added the following css to the user s code it fixed it css root display none we can add this to storybook itself to avoid this issue for users that style root which is a completely reasonable thing to do however i m not sure this is enough | 1 |
430,587 | 12,463,497,772 | IssuesEvent | 2020-05-28 10:43:25 | UTRS2/utrs | https://api.github.com/repos/UTRS2/utrs | closed | Unable to view appeals | Priority: High bug | Users w/ the global flag in users.wikis or maybe with multiple wikis are not seeing appeals at all.
Current workaround: Manual override of DB field to "enwiki" until fixed as there are no global appeals yet.
Done for these users so far, replace them when done.
Xoasflux
JJMC89
ST47
TonyBallioni
TheSandDoctor | 1.0 | Unable to view appeals - Users w/ the global flag in users.wikis or maybe with multiple wikis are not seeing appeals at all.
Current workaround: Manual override of DB field to "enwiki" until fixed as there are no global appeals yet.
Done for these users so far, replace them when done.
Xoasflux
JJMC89
ST47
TonyBallioni
TheSandDoctor | priority | unable to view appeals users w the global flag in users wikis or maybe with multiple wikis are not seeing appeals at all current workaround manual override of db field to enwiki until fixed as there are no global appeals yet done for these users so far replace them when done xoasflux tonyballioni thesanddoctor | 1 |
745,862 | 26,004,398,832 | IssuesEvent | 2022-12-20 17:55:57 | vyper-protocol/vyper-otc-ui | https://api.github.com/repos/vyper-protocol/vyper-otc-ui | closed | add support for featured | high priority | add a menu in the topbar with a dropdown with featured products. These can be different depending on the cluster. #364 also relevant for this | 1.0 | add support for featured - add a menu in the topbar with a dropdown with featured products. These can be different depending on the cluster. #364 also relevant for this | priority | add support for featured add a menu in the topbar with a dropdown with featured products these can be different depending on the cluster also relevant for this | 1 |
517,500 | 15,014,950,055 | IssuesEvent | 2021-02-01 07:32:55 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | connect.garmin.com - site is not usable | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal | <!-- @browser: Firefox 86.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:86.0) Gecko/20100101 Firefox/86.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66526 -->
**URL**: https://connect.garmin.com/modern/
**Browser / Version**: Firefox 86.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Tab loads with the page but page remains blank
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/9f3fc4ee-7863-46dd-98e4-136ef8be4ac3.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210128185743</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/4457943e-bc80-48c5-a650-bcb7d7736133)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | connect.garmin.com - site is not usable - <!-- @browser: Firefox 86.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:86.0) Gecko/20100101 Firefox/86.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66526 -->
**URL**: https://connect.garmin.com/modern/
**Browser / Version**: Firefox 86.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Tab loads with the page but page remains blank
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/9f3fc4ee-7863-46dd-98e4-136ef8be4ac3.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210128185743</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/4457943e-bc80-48c5-a650-bcb7d7736133)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | connect garmin com site is not usable url browser version firefox operating system windows tested another browser no problem type site is not usable description page not loading correctly steps to reproduce tab loads with the page but page remains blank view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
805,896 | 29,736,130,298 | IssuesEvent | 2023-06-14 01:10:08 | GSM-MSG/SMS-BackEnd | https://api.github.com/repos/GSM-MSG/SMS-BackEnd | closed | Spring Security에서 예외 핸들링 반대로 되어져 있음 | 1️⃣ Priority: High | ### Describe
인증에 403이 뜨고, 인가에 401이 뜨게 잘못 핸들링 해서 수정하겠습니다.
### Additional
_No response_ | 1.0 | Spring Security에서 예외 핸들링 반대로 되어져 있음 - ### Describe
인증에 403이 뜨고, 인가에 401이 뜨게 잘못 핸들링 해서 수정하겠습니다.
### Additional
_No response_ | priority | spring security에서 예외 핸들링 반대로 되어져 있음 describe 인증에 뜨고 인가에 뜨게 잘못 핸들링 해서 수정하겠습니다 additional no response | 1 |
387,720 | 11,467,130,334 | IssuesEvent | 2020-02-08 02:20:38 | allenai/scholar-reader | https://api.github.com/repos/allenai/scholar-reader | closed | Fault-tolerant processing of one entity at a time | high-priority pipeline | One challenge with colorizing multiple entities at a time is cascading failures---an error in colorizing one entity may change the location of all after it, or cause an entire batch of entities not to be colorized.
This task includes:
- [x] Add option to colorizing commands to process entities one at a time
- [x] Add option to full pipeline for processing entities one at a time
- [x] Add visual validation to check for any black pixels in image diffs. If there are black pixels, then do not attempt to detect bounding boxes
- [x] Documentation that conveys that running one entity at a time will result in much greater usage of storage
Follow-up analysis includes (put this in a separate issue later):
- [ ] Characterize the time it takes to process the 'typical' paper one entity at a time
- [ ] Compare to the relative costs and benefits of reworking the TeX engine to handle colorizing without affecting paper layout
- [ ] Characterize the number of entities that will be left out during one-at-a-time entity colorizing | 1.0 | Fault-tolerant processing of one entity at a time - One challenge with colorizing multiple entities at a time is cascading failures---an error in colorizing one entity may change the location of all after it, or cause an entire batch of entities not to be colorized.
This task includes:
- [x] Add option to colorizing commands to process entities one at a time
- [x] Add option to full pipeline for processing entities one at a time
- [x] Add visual validation to check for any black pixels in image diffs. If there are black pixels, then do not attempt to detect bounding boxes
- [x] Documentation that conveys that running one entity at a time will result in much greater usage of storage
Follow-up analysis includes (put this in a separate issue later):
- [ ] Characterize the time it takes to process the 'typical' paper one entity at a time
- [ ] Compare to the relative costs and benefits of reworking the TeX engine to handle colorizing without affecting paper layout
- [ ] Characterize the number of entities that will be left out during one-at-a-time entity colorizing | priority | fault tolerant processing of one entity at a time one challenge with colorizing multiple entities at a time is cascading failures an error in colorizing one entity may change the location of all after it or cause an entire batch of entities not to be colorized this task includes add option to colorizing commands to process entities one at a time add option to full pipeline for processing entities one at a time add visual validation to check for any black pixels in image diffs if there are black pixels then do not attempt to detect bounding boxes documentation that conveys that running one entity at a time will result in much greater usage of storage follow up analysis includes put this in a separate issue later characterize the time it takes to process the typical paper one entity at a time compare to the relative costs and benefits of reworking the tex engine to handle colorizing without affecting paper layout characterize the number of entities that will be left out during one at a time entity colorizing | 1 |
355,374 | 10,579,943,329 | IssuesEvent | 2019-10-08 04:50:44 | CalNourish/ucbfpa-webapp | https://api.github.com/repos/CalNourish/ucbfpa-webapp | opened | Update Add/Edit UI | high priority 🔥 | - [ ] Remove points
- [ ] Remove images
- [ ] Introduce some UI for easily adding and subtracting counts without having to math (and be able to edit the total amount directly somehow) | 1.0 | Update Add/Edit UI - - [ ] Remove points
- [ ] Remove images
- [ ] Introduce some UI for easily adding and subtracting counts without having to math (and be able to edit the total amount directly somehow) | priority | update add edit ui remove points remove images introduce some ui for easily adding and subtracting counts without having to math and be able to edit the total amount directly somehow | 1 |
166,221 | 6,300,163,869 | IssuesEvent | 2017-07-21 02:21:05 | minio/minio | https://api.github.com/repos/minio/minio | closed | [mint] Tests fail in Azure/GCS gateway mode | priority: high working as intended | ## Expected Behavior
Mint test cases should pass.
## Current Behavior
When Mint is run against the `minio gateway azure` or `minio gateway gcs`, requests fail with server log as below
```
$ ./minio gateway azure
Endpoint: http://192.168.86.129:9000 http://172.17.0.1:9000 http://172.18.0.1:9000 http://127.0.0.1:9000
ERRO[0007] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-10234","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=ac91ba1d601e7b2d5136e531a3111770db0603705d9f388aee05c36f53fc6f65"],"Aws-Sdk-Invocation-Id":["8ccba5715ed1842e5862c4ec36a94c8c"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
ERRO[0007] {"method":"DELETE","reqURI":"/aws-sdk-php-bucket-10234/obj1","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=7f448c75c4f825a397624b1fc1836841dffd8ecbf402a474c1a28486c026b9ae"],"Aws-Sdk-Invocation-Id":["908b9e424ad67ce48c2d5cde6c2659ec"],"Aws-Sdk-Retry":["0/0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
```
```
$ ./minio gateway gcs peak-essence-171622
*** Warning: Not Ready for Production ***
ERRO[0025] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-43021","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=2297da513a5dabc3230d17adc6713c54a5af97f4f0573ee86c3486add17d36c9"],"Aws-Sdk-Invocation-Id":["3ac3bd8d67bb14aedf64ee263ae42d6d"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T102637Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
```
Note that region is set using `export MINIO_REGION="us-east-1"`
## Steps to Reproduce (for bugs)
1. Start Minio server in azure or gcs gateway mode.
2. Run Mint against the Minio server instance. | 1.0 | [mint] Tests fail in Azure/GCS gateway mode - ## Expected Behavior
Mint test cases should pass.
## Current Behavior
When Mint is run against the `minio gateway azure` or `minio gateway gcs`, requests fail with server log as below
```
$ ./minio gateway azure
Endpoint: http://192.168.86.129:9000 http://172.17.0.1:9000 http://172.18.0.1:9000 http://127.0.0.1:9000
ERRO[0007] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-10234","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=ac91ba1d601e7b2d5136e531a3111770db0603705d9f388aee05c36f53fc6f65"],"Aws-Sdk-Invocation-Id":["8ccba5715ed1842e5862c4ec36a94c8c"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
ERRO[0007] {"method":"DELETE","reqURI":"/aws-sdk-php-bucket-10234/obj1","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=7f448c75c4f825a397624b1fc1836841dffd8ecbf402a474c1a28486c026b9ae"],"Aws-Sdk-Invocation-Id":["908b9e424ad67ce48c2d5cde6c2659ec"],"Aws-Sdk-Retry":["0/0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
```
```
$ ./minio gateway gcs peak-essence-171622
*** Warning: Not Ready for Production ***
ERRO[0025] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-43021","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=2297da513a5dabc3230d17adc6713c54a5af97f4f0573ee86c3486add17d36c9"],"Aws-Sdk-Invocation-Id":["3ac3bd8d67bb14aedf64ee263ae42d6d"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T102637Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()]
```
Note that region is set using `export MINIO_REGION="us-east-1"`
## Steps to Reproduce (for bugs)
1. Start Minio server in azure or gcs gateway mode.
2. Run Mint against the Minio server instance. | priority | tests fail in azure gcs gateway mode expected behavior mint test cases should pass current behavior when mint is run against the minio gateway azure or minio gateway gcs requests fail with server log as below minio gateway azure endpoint erro method put requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry content length host user agent x amz content x amz date cause signature does not match source erro method delete requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry host user agent x amz content x amz date cause signature does not match source minio gateway gcs peak essence warning not ready for production erro method put requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry content length host user agent x amz content x amz date cause signature does not match source note that region is set using export minio region us east steps to reproduce for bugs start minio server in azure or gcs gateway mode run mint against the minio server instance | 1 |
520,042 | 15,077,759,202 | IssuesEvent | 2021-02-05 07:34:09 | wso2/cellery | https://api.github.com/repos/wso2/cellery | closed | Dry run mode should be implemented for running cellery component tests | Priority/High Resolution/Won’t Fix Type/Improvement | **Description:**
After the CLI refactoring, the kubectl commands are run in a ballerina native function. Therefore the component tests are required to be run on dry run mode to `cellery run` command perform a kubectl dry run to stop communication to the api server. As a workaround for the moment, the exceptions thrown by kubectl apply command are skipped after checking. | 1.0 | Dry run mode should be implemented for running cellery component tests - **Description:**
After the CLI refactoring, the kubectl commands are run in a ballerina native function. Therefore the component tests are required to be run on dry run mode to `cellery run` command perform a kubectl dry run to stop communication to the api server. As a workaround for the moment, the exceptions thrown by kubectl apply command are skipped after checking. | priority | dry run mode should be implemented for running cellery component tests description after the cli refactoring the kubectl commands are run in a ballerina native function therefore the component tests are required to be run on dry run mode to cellery run command perform a kubectl dry run to stop communication to the api server as a workaround for the moment the exceptions thrown by kubectl apply command are skipped after checking | 1 |
362,260 | 10,724,513,707 | IssuesEvent | 2019-10-28 02:02:55 | LuanKovacs/LittleMatchGirlGame | https://api.github.com/repos/LuanKovacs/LittleMatchGirlGame | closed | [UPDATE, NOW WITH CAMERA PROBLEMS] I've fallen over and can get back up | Priority: High bug | **Describe the bug**
you can fall over if you die near the end of the game
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the final area
2. die
3. if the spawn in falling stutters a little then the player can now fall over when walking
**Expected behavior**
you shouldn't be able to fall over like that, the player while running should stay upright
**Additional context**
Add any other context about the problem here.
| 1.0 | [UPDATE, NOW WITH CAMERA PROBLEMS] I've fallen over and can get back up - **Describe the bug**
you can fall over if you die near the end of the game
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the final area
2. die
3. if the spawn in falling stutters a little then the player can now fall over when walking
**Expected behavior**
you shouldn't be able to fall over like that, the player while running should stay upright
**Additional context**
Add any other context about the problem here.
| priority | i ve fallen over and can get back up describe the bug you can fall over if you die near the end of the game to reproduce steps to reproduce the behavior go to the final area die if the spawn in falling stutters a little then the player can now fall over when walking expected behavior you shouldn t be able to fall over like that the player while running should stay upright additional context add any other context about the problem here | 1 |
192,217 | 6,847,705,995 | IssuesEvent | 2017-11-13 16:11:06 | cceh/capitularia | https://api.github.com/repos/cceh/capitularia | closed | Implementierung des Ortsregisters in den Handschriftenfilter | High Priority | Das Ortsregister sollte zeitnah in den Handschriftenfilter implementiert werden. | 1.0 | Implementierung des Ortsregisters in den Handschriftenfilter - Das Ortsregister sollte zeitnah in den Handschriftenfilter implementiert werden. | priority | implementierung des ortsregisters in den handschriftenfilter das ortsregister sollte zeitnah in den handschriftenfilter implementiert werden | 1 |
656,227 | 21,724,004,263 | IssuesEvent | 2022-05-11 05:21:29 | ever-co/ever-gauzy | https://api.github.com/repos/ever-co/ever-gauzy | closed | Fix: Theme Card Scrollbars (Firefox) | priority: highest Firefox Chorme Safari | The theme "scrollbars" in Firefox so they will be "thin" like in Chrome.

| 1.0 | Fix: Theme Card Scrollbars (Firefox) - The theme "scrollbars" in Firefox so they will be "thin" like in Chrome.

| priority | fix theme card scrollbars firefox the theme scrollbars in firefox so they will be thin like in chrome | 1 |
151,026 | 5,795,538,621 | IssuesEvent | 2017-05-02 17:20:06 | JiscRDSS/rdss-canonical-data-model | https://api.github.com/repos/JiscRDSS/rdss-canonical-data-model | closed | UC87 Metadata fields | alpha priority:High use case | UC no.: 87
Theme: Metadata fields
As a Data creator
I want Widespread use of unique IDs for people and organisations
So that I don't have to keep up with correct/canonical names changing all the time, particularly for organisations and individuals' email addresses
Comments
Identifiers are a key component of the MVP (and data model) | 1.0 | UC87 Metadata fields - UC no.: 87
Theme: Metadata fields
As a Data creator
I want Widespread use of unique IDs for people and organisations
So that I don't have to keep up with correct/canonical names changing all the time, particularly for organisations and individuals' email addresses
Comments
Identifiers are a key component of the MVP (and data model) | priority | metadata fields uc no theme metadata fields as a data creator i want widespread use of unique ids for people and organisations so that i don t have to keep up with correct canonical names changing all the time particularly for organisations and individuals email addresses comments identifiers are a key component of the mvp and data model | 1 |
432,023 | 12,488,173,186 | IssuesEvent | 2020-05-31 13:01:57 | STAMACODING/RSA-App | https://api.github.com/repos/STAMACODING/RSA-App | opened | Switch to OpenJDK | high priority meeting relevant organization | Auf meiner Raspberry Pi 4B gibt es leider nur schweren Support für das kommerzielle Oracle JDK. Für die neuste Version (14) sogar gar keine. Der Standard ist [OpenJDK](https://openjdk.java.net/), eine Open-Source-Variante des JDKs. Fürs Programmieren würde das quasi keine Unterschiede machen. Es müsste nur jeder bei sich einrichten. Dafür könnte ich auch ein Tutorial schreiben. Am besten besprechen wir das ganze in einem **Meeting**. | 1.0 | Switch to OpenJDK - Auf meiner Raspberry Pi 4B gibt es leider nur schweren Support für das kommerzielle Oracle JDK. Für die neuste Version (14) sogar gar keine. Der Standard ist [OpenJDK](https://openjdk.java.net/), eine Open-Source-Variante des JDKs. Fürs Programmieren würde das quasi keine Unterschiede machen. Es müsste nur jeder bei sich einrichten. Dafür könnte ich auch ein Tutorial schreiben. Am besten besprechen wir das ganze in einem **Meeting**. | priority | switch to openjdk auf meiner raspberry pi gibt es leider nur schweren support für das kommerzielle oracle jdk für die neuste version sogar gar keine der standard ist eine open source variante des jdks fürs programmieren würde das quasi keine unterschiede machen es müsste nur jeder bei sich einrichten dafür könnte ich auch ein tutorial schreiben am besten besprechen wir das ganze in einem meeting | 1 |
413,005 | 12,059,178,805 | IssuesEvent | 2020-04-15 18:48:16 | tern-tools/tern | https://api.github.com/repos/tern-tools/tern | closed | Update SPDX format to include file level analysis | high-priority | **Description**
Update the SPDX report format to include situations where there is file level data. Use the http://13.57.134.254/app/validate/ online tool to validate the generated SPDX document for various container images.
**Background**
This depends on https://github.com/vmware/tern/pull/582 to be merged.
**Super Issues**
#583
| 1.0 | Update SPDX format to include file level analysis - **Description**
Update the SPDX report format to include situations where there is file level data. Use the http://13.57.134.254/app/validate/ online tool to validate the generated SPDX document for various container images.
**Background**
This depends on https://github.com/vmware/tern/pull/582 to be merged.
**Super Issues**
#583
| priority | update spdx format to include file level analysis description update the spdx report format to include situations where there is file level data use the online tool to validate the generated spdx document for various container images background this depends on to be merged super issues | 1 |
296,020 | 9,103,469,990 | IssuesEvent | 2019-02-20 15:57:57 | infor-design/website | https://api.github.com/repos/infor-design/website | closed | Source Sans fonts aren't included in ng7 build | for: dev priority: high | **Describe the bug**
Oops.
**To Reproduce**
Navigate to site on a device that doesn't have Source Sans installed.
**Expected behavior**
Include using Google fonts as previously or you could include and serve the fonts from `ids-identity`. | 1.0 | Source Sans fonts aren't included in ng7 build - **Describe the bug**
Oops.
**To Reproduce**
Navigate to site on a device that doesn't have Source Sans installed.
**Expected behavior**
Include using Google fonts as previously or you could include and serve the fonts from `ids-identity`. | priority | source sans fonts aren t included in build describe the bug oops to reproduce navigate to site on a device that doesn t have source sans installed expected behavior include using google fonts as previously or you could include and serve the fonts from ids identity | 1 |
727,707 | 25,044,734,986 | IssuesEvent | 2022-11-05 04:50:12 | WFP-VAM/prism-app | https://api.github.com/repos/WFP-VAM/prism-app | closed | PRISM reports module | enhancement help wanted idea priority:high | Most users of information accessible through PRISM will not frequently visit the application. In addition, more advanced users such as analysts and GIS officers produce outputs that are then shared to a wider audience. Recognizing this, PRISM requires the ability to 1) condense various data inputs into a single output for a single snapshot view of key risk and impact factors; 2) reduce the steps involved for a user to generate a shareable output to a wide audience. A reports feature in PRISM will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart. The report could be triggered through the ‘Export’ button, and a direct link will be accessible via a URL such as https://prism.org?view=report&date=2021-05-17&hazard=tropical-storm
Google doc to comment on here: https://docs.google.com/document/d/1FaV2jp526Xa134j32iXPZ_U4lhxFanlr/edit?usp=sharing&ouid=105953411765103382631&rtpof=true&sd=true
Mockups:


| 1.0 | PRISM reports module - Most users of information accessible through PRISM will not frequently visit the application. In addition, more advanced users such as analysts and GIS officers produce outputs that are then shared to a wider audience. Recognizing this, PRISM requires the ability to 1) condense various data inputs into a single output for a single snapshot view of key risk and impact factors; 2) reduce the steps involved for a user to generate a shareable output to a wide audience. A reports feature in PRISM will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart. The report could be triggered through the ‘Export’ button, and a direct link will be accessible via a URL such as https://prism.org?view=report&date=2021-05-17&hazard=tropical-storm
Google doc to comment on here: https://docs.google.com/document/d/1FaV2jp526Xa134j32iXPZ_U4lhxFanlr/edit?usp=sharing&ouid=105953411765103382631&rtpof=true&sd=true
Mockups:


| priority | prism reports module most users of information accessible through prism will not frequently visit the application in addition more advanced users such as analysts and gis officers produce outputs that are then shared to a wider audience recognizing this prism requires the ability to condense various data inputs into a single output for a single snapshot view of key risk and impact factors reduce the steps involved for a user to generate a shareable output to a wide audience a reports feature in prism will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart the report could be triggered through the ‘export’ button and a direct link will be accessible via a url such as google doc to comment on here mockups | 1 |
527,434 | 15,342,640,821 | IssuesEvent | 2021-02-27 17:02:21 | getting-things-gnome/gtg | https://api.github.com/repos/getting-things-gnome/gtg | opened | Tags picker popover is empty after the v2 file format branch merge | bug priority:high reproducible-in-git | With current master, with the "screenshots" dataset, if you create a new task, in the task editor, clicking the tags button will give you an empty popover rather than showing existing tags. | 1.0 | Tags picker popover is empty after the v2 file format branch merge - With current master, with the "screenshots" dataset, if you create a new task, in the task editor, clicking the tags button will give you an empty popover rather than showing existing tags. | priority | tags picker popover is empty after the file format branch merge with current master with the screenshots dataset if you create a new task in the task editor clicking the tags button will give you an empty popover rather than showing existing tags | 1 |
135,793 | 5,258,857,791 | IssuesEvent | 2017-02-03 00:57:57 | ucdavis/ipa-client-angular | https://api.github.com/repos/ucdavis/ipa-client-angular | closed | TeachingCallResponse report: suggested courses not displayed correctly | bug high priority | From email:
If you look at faculty member "Hanti Kao Lin" in Philosophy - his interested courses are all listed as sabbaticals in the report.
| 1.0 | TeachingCallResponse report: suggested courses not displayed correctly - From email:
If you look at faculty member "Hanti Kao Lin" in Philosophy - his interested courses are all listed as sabbaticals in the report.
| priority | teachingcallresponse report suggested courses not displayed correctly from email if you look at faculty member hanti kao lin in philosophy his interested courses are all listed as sabbaticals in the report | 1 |
274,845 | 8,568,542,085 | IssuesEvent | 2018-11-10 22:37:32 | giftdibs/giftdibs-browser | https://api.github.com/repos/giftdibs/giftdibs-browser | closed | Gift Detail > Gift delivered message | priority: high | Show message on gift detail (and `gift_delivered` notification) when someone delivers a gift, to let the owner of the gift mark it as received.
Clean up the "Delivered by" section on the gift detail. | 1.0 | Gift Detail > Gift delivered message - Show message on gift detail (and `gift_delivered` notification) when someone delivers a gift, to let the owner of the gift mark it as received.
Clean up the "Delivered by" section on the gift detail. | priority | gift detail gift delivered message show message on gift detail and gift delivered notification when someone delivers a gift to let the owner of the gift mark it as received clean up the delivered by section on the gift detail | 1 |
597,806 | 18,172,502,063 | IssuesEvent | 2021-09-27 21:45:22 | StatisticsNZ/simplevis | https://api.github.com/repos/StatisticsNZ/simplevis | closed | bar: x_var date labels are not working correctly | high priority | ```
library(tidyverse)
library(er.helpers)
library(simplevis)
setup_datalake_access()
no2_nzta <- er.helpers::read_from_datalake( "air/2021/tidy/no2_nzta.RDS")
sitecheck_data <- no2_nzta %>%
select(site, "value" = concentration, month, year) %>%
mutate(len = str_length(site)) %>%
mutate(temp_id = as.character(substring(site, 1,6))) %>%
group_by(temp_id) %>%
filter(any(str_length(site) > 6)) %>%
mutate(measurement_date = lubridate::my(paste0(month, year)) %>% lubridate::as_date()) %>%
mutate(site = as.character(site))
p <- sitecheck_data %>%
filter(temp_id == "AUC004") %>%
simplevis::gg_bar_col(x_var = measurement_date,
y_var = value,
col_var = site,
x_pretty_n = 10,
x_labels = scales::date_format("%y"))
p
sitecheck_data %>%
filter(temp_id == "AUC004") %>%
ggplot(aes(x = measurement_date, y = value, fill = site)) +
geom_col()
plotly::ggplotly(p)
```
| 1.0 | bar: x_var date labels are not working correctly - ```
library(tidyverse)
library(er.helpers)
library(simplevis)
setup_datalake_access()
no2_nzta <- er.helpers::read_from_datalake( "air/2021/tidy/no2_nzta.RDS")
sitecheck_data <- no2_nzta %>%
select(site, "value" = concentration, month, year) %>%
mutate(len = str_length(site)) %>%
mutate(temp_id = as.character(substring(site, 1,6))) %>%
group_by(temp_id) %>%
filter(any(str_length(site) > 6)) %>%
mutate(measurement_date = lubridate::my(paste0(month, year)) %>% lubridate::as_date()) %>%
mutate(site = as.character(site))
p <- sitecheck_data %>%
filter(temp_id == "AUC004") %>%
simplevis::gg_bar_col(x_var = measurement_date,
y_var = value,
col_var = site,
x_pretty_n = 10,
x_labels = scales::date_format("%y"))
p
sitecheck_data %>%
filter(temp_id == "AUC004") %>%
ggplot(aes(x = measurement_date, y = value, fill = site)) +
geom_col()
plotly::ggplotly(p)
```
| priority | bar x var date labels are not working correctly library tidyverse library er helpers library simplevis setup datalake access nzta er helpers read from datalake air tidy nzta rds sitecheck data select site value concentration month year mutate len str length site mutate temp id as character substring site group by temp id filter any str length site mutate measurement date lubridate my month year lubridate as date mutate site as character site p filter temp id simplevis gg bar col x var measurement date y var value col var site x pretty n x labels scales date format y p sitecheck data filter temp id ggplot aes x measurement date y value fill site geom col plotly ggplotly p | 1 |
478,897 | 13,787,839,858 | IssuesEvent | 2020-10-09 05:58:08 | wso2/streaming-integrator | https://api.github.com/repos/wso2/streaming-integrator | opened | Improvement for Siddhi Aggregation process | Priority/High Severity/Major Type/Improvement | **Description:**
We need to have a check before purging tables.
Ex. when purging the "days" table, we need to check whether the "months" table aggregations have happened with the relevant data which is going to be purged. And if not, we need to log a warning or an error.
So that if any error happens, there will be data in the tables without purging so after providing a fix, it will resume aggregate data from where it left off. This will result in no data loss.
**Affected Product Version:**
SI-1.1.0 | 1.0 | Improvement for Siddhi Aggregation process - **Description:**
We need to have a check before purging tables.
Ex. when purging the "days" table, we need to check whether the "months" table aggregations have happened with the relevant data which is going to be purged. And if not, we need to log a warning or an error.
So that if any error happens, there will be data in the tables without purging so after providing a fix, it will resume aggregate data from where it left off. This will result in no data loss.
**Affected Product Version:**
SI-1.1.0 | priority | improvement for siddhi aggregation process description we need to have a check before purging tables ex when purging the days table we need to check whether the months table aggregations have happened with the relevant data which is going to be purged and if not we need to log a warning or an error so that if any error happens there will be data in the tables without purging so after providing a fix it will resume aggregate data from where it left off this will result in no data loss affected product version si | 1 |
45,343 | 2,928,232,999 | IssuesEvent | 2015-06-27 00:55:54 | EFForg/privacybadgerchrome | https://api.github.com/repos/EFForg/privacybadgerchrome | closed | One click whitelist is broken for youtube.com | bug High priority | It should pop up when a user tries to comment on youtube but it doesn't. We should check disqus as well. | 1.0 | One click whitelist is broken for youtube.com - It should pop up when a user tries to comment on youtube but it doesn't. We should check disqus as well. | priority | one click whitelist is broken for youtube com it should pop up when a user tries to comment on youtube but it doesn t we should check disqus as well | 1 |
106,641 | 4,281,570,843 | IssuesEvent | 2016-07-15 03:57:55 | fflewddur/archivo | https://api.github.com/repos/fflewddur/archivo | closed | Pixelization/macroblocking in archived videos | bug high priority | Using Windows 10 and PrivateInternetAccess VPN, PC is unable to find TiVo device. They are on the same network - disabling firewall for private connections didn't help. If I disable the VPN it works fine. Any ideas on how to use without disabling the VPN? | 1.0 | Pixelization/macroblocking in archived videos - Using Windows 10 and PrivateInternetAccess VPN, PC is unable to find TiVo device. They are on the same network - disabling firewall for private connections didn't help. If I disable the VPN it works fine. Any ideas on how to use without disabling the VPN? | priority | pixelization macroblocking in archived videos using windows and privateinternetaccess vpn pc is unable to find tivo device they are on the same network disabling firewall for private connections didn t help if i disable the vpn it works fine any ideas on how to use without disabling the vpn | 1 |
624,964 | 19,714,774,121 | IssuesEvent | 2022-01-13 09:55:10 | hermeznetwork/wallet-ui | https://api.github.com/repos/hermeznetwork/wallet-ui | closed | Token Swap/Implement Design Token selector | type: enhancement priority: high | - Needs API to query tokens
- Needs to check valid swaps for second token
- Needs way to get images for token
 | 1.0 | Token Swap/Implement Design Token selector - - Needs API to query tokens
- Needs to check valid swaps for second token
- Needs way to get images for token
 | priority | token swap implement design token selector needs api to query tokens needs to check valid swaps for second token needs way to get images for token | 1 |
454,140 | 13,095,491,648 | IssuesEvent | 2020-08-03 14:12:24 | firecracker-microvm/firecracker | https://api.github.com/repos/firecracker-microvm/firecracker | closed | InvalidOffset (virtio-block) error when resuming after loading a snapshot (with rootfs from firecracker-containerd) | Feature: Snapshotting Priority: High Quality: Bug | Hi, we started developing support for snapshot `pause/resume/create/load` inside our fork of firecracker-containerd (we can make the code public). While we had no problem with supporting `pause/resume/create-snapshot` methods we ran into an error inside Firecracker's virtio-block module with `load-snapshot->resume`.
The workflow is startVM(boot) -> Pause -> Create-Snap -> Offload (kill the VM with SIGTERM) -> SnapshotLoad -> Resume
The problem we face is at the Resume point where we get the following error from firecracker's log:
```
2020-06-25T07:21:10.652265445 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Put request on "/snapshot/load" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\"}".
2020-06-25T07:21:10.664695532 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:21:10.665787599 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Patch request on "/vm" with body "{\"state\":\"Resumed\"}".
2020-06-25T07:21:10.665908820 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:21:14.799994834 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset)
2020-06-25T07:21:14.802628426 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset)
```
We tend to think that this can be related to the way firecracker-containerd manages block devices. The VM mounts block devices in 2 phases. First, the VM boots from a generic rootfs (with the containerd agent):
```
The API server received a Put request on "/drives/root_drive" with body "The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n".
```
Then the agent needs to mount a second block device that contains the container-specific data. To do so, containerd attaches another drive and patches the path to the drive twice:
```
The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n".
<<..>>
The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n".
The request was executed successfully. Status code: 204 No Content.
The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n".
```
Can the issue be connected to the way [PATCH drive works](https://github.com/firecracker-microvm/firecracker/blob/master/docs/api_requests/patch-block.md)? This drive is supposed to remain mounted into the restored guest.
We would greatly appreciate comments and ideas on what could be the root cause from Firecracker and firecracker-containerd maintainers: for example, @acatangiu @kzys . Once we fix the issue, we would be happy to contribute our changes to firecracker-containerd upstream.
Full workflow log (Firecracker's log):
```
Running Firecracker v0.21.0
2020-06-25T07:20:17.287742839 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.288144594 [anonymous-instance:INFO] The API server received a Put request on "/machine-config" with body "{\"cpu_template\":\"T2\",\"ht_enabled\":false,\"mem_size_mib\":512,\"vcpu_count\":1}\n".
2020-06-25T07:20:17.288291418 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.288521865 [anonymous-instance:INFO] The API server received a Get request on "/machine-config".
2020-06-25T07:20:17.288625819 [anonymous-instance:INFO] The request was executed successfully. Status code: 200 OK.
2020-06-25T07:20:17.288971127 [anonymous-instance:INFO] The API server received a Put request on "/boot-source" with body "{\"boot_args\":\"8250.nr_uarts=0 ip=190.128.0.2::190.128.0.1:255.192.0.0:::off::: systemd.log_color=false init=/sbin/overlay-init systemd.unit=firecracker.target quiet noapic nomodules ipv6.disable=1 ro panic=1 tsc=reliable reboot=k pci=off\",\"kernel_image_path\":\"/var/lib/firecracker-containerd/runtime/hello-vmlinux.bin\"}\n".
2020-06-25T07:20:17.289115444 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.289571124 [anonymous-instance:INFO] The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n".
2020-06-25T07:20:17.289732138 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.290046499 [anonymous-instance:INFO] The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n".
2020-06-25T07:20:17.290151860 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.290522553 [anonymous-instance:INFO] The API server received a Put request on "/network-interfaces/1" with body "{\"guest_mac\":\"02:FC:00:00:00:00\",\"host_dev_name\":\"fc-0-tap0\",\"iface_id\":\"1\"}\n".
2020-06-25T07:20:17.292898333 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.293236487 [anonymous-instance:INFO] The API server received a Put request on "/vsock" with body "{\"guest_cid\":0,\"uds_path\":\"firecracker.vsock\",\"vsock_id\":\"agent_api\"}\n".
2020-06-25T07:20:17.293472939 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.293754191 [anonymous-instance:INFO] The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n".
2020-06-25T07:20:17.305625745 [anonymous-instance:WARN] Could not add serial input event to epoll: Error during epoll call: Operation not permitted (os error 1)
2020-06-25T07:20:17.306309907 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:19.177877690 [anonymous-instance:INFO] The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n".
2020-06-25T07:20:19.178152684 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:20.104482947 [anonymous-instance:INFO] The API server received a Patch request on "/vm" with body "{\"state\":\"Paused\"}".
2020-06-25T07:20:20.104702379 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:20.106119166 [anonymous-instance:INFO] The API server received a Put request on "/snapshot/create" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\",\"snapshot_type\":\"Full\"}".
2020-06-25T07:20:20.649402863 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
```
And the corresponding log from firecracker-containerd (containerd's log)
```
INFO[2020-06-25T07:20:17.275031352-04:00] creating new VM runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:17.275464729-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" shim_stream=stderr vmID=505
setupLogging
setupMetrics
ERRO[2020-06-25T07:20:17.288041835-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="VMM metrics disabled." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.288909406-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate:T2 HtEnabled:0xc000582f43 MemSizeMib:0xc000582f38 VcpuCount:0xc000582f30}" shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289306935-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289343891-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289933498-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289952823-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0, slot MN2HE43UOVRDA, root false." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.290347821-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.290377745-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching NIC fc-0-tap0 (hwaddr 02:FC:00:00:00:00) at index 1" shim_stream=stderr vmID=505
INFO[2020-06-25T07:20:17.306497334-04:00] calling agent runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:17.306529875-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " shim_stream=stderr vmID=505
DEBU[2020-06-25T07:20:18.004394761-04:00] garbage collected d=10.79516ms
INFO[2020-06-25T07:20:18.410674791-04:00] successfully started the VM runtime=aws.firecracker vmID=505
DEBU[2020-06-25T07:20:18.411052023-04:00] event forwarded ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
DEBU[2020-06-25T07:20:18.440217583-04:00] prepare snapshot key=505 parent="sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc"
DEBU[2020-06-25T07:20:18.440347865-04:00] prepare key=firecracker-containerd/17/505 parent="firecracker-containerd/16/sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc"
DEBU[2020-06-25T07:20:18.440504004-04:00] creating snapshot device 'fc-dev-thinpool-snap-9' from 'fc-dev-thinpool-snap-8'
DEBU[2020-06-25T07:20:18.795683399-04:00] event published ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
DEBU[2020-06-25T07:20:18.800390309-04:00] get snapshot mounts key=505
DEBU[2020-06-25T07:20:18.800469759-04:00] mounts key=firecracker-containerd/17/505
DEBU[2020-06-25T07:20:19.120516743-04:00] event published ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
DEBU[2020-06-25T07:20:19.138167895-04:00] get snapshot mounts key=505
DEBU[2020-06-25T07:20:19.138206395-04:00] mounts key=firecracker-containerd/17/505
time="2020-06-25T07:20:19.174115148-04:00" level=debug msg=StartShim runtime=aws.firecracker task_id=505
DEBU[2020-06-25T07:20:19.175656280-04:00] create VM request: VMID:"505"
DEBU[2020-06-25T07:20:19.175707632-04:00] using namespace: firecracker-containerd
time="2020-06-25T07:20:19.176188684-04:00" level=info msg="successfully started shim (git commit: 22b7cfceb4901b91c998a380df74d9b825614967)." runtime=aws.firecracker task_id=505 vmID=505
ERRO[2020-06-25T07:20:19.178386061-04:00] time="2020-06-25T07:20:19-04:00" level=info msg="PatchGuestDrive successful" shim_stream=stderr vmID=505
INFO[2020-06-25T07:20:19.330752483-04:00] successfully created task ExecID= TaskID=505 pid_in_vm=727 runtime=aws.firecracker vmID=505
DEBU[2020-06-25T07:20:19.331251986-04:00] event forwarded ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
DEBU[2020-06-25T07:20:19.337046160-04:00] garbage collected d=12.841932ms
DEBU[2020-06-25T07:20:19.339260338-04:00] event forwarded ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
DEBU[2020-06-25T07:20:20.103637583-04:00] pause VM request: VMID:"505"
DEBU[2020-06-25T07:20:20.105566980-04:00] make snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file"
DEBU[2020-06-25T07:20:20.650066779-04:00] offload original request: VMID:"505"
ERRO[2020-06-25T07:20:25.892253044-04:00] error waiting for exit ExecID= TaskID=505 error="ttrpc: closed" runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:25.892257191-04:00] error while forwarding events from VM agent error="ttrpc: closed" runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:25.892484333-04:00] error copying io ExecID= TaskID=505 error="reading from a closed fifo" runtime=aws.firecracker stream=stdin vmID=505
ERRO[2020-06-25T07:20:25.892580185-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=505
ERRO[2020-06-25T07:20:25.923729272-04:00] time="2020-06-25T07:20:25-04:00" level=warning msg="firecracker exited: signal: killed" shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:30.892486598-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=505
ERRO[2020-06-25T07:20:30.892525517-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=505
DEBU[2020-06-25T07:21:10.651353555-04:00] load snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file"
DEBU[2020-06-25T07:21:10.665366928-04:00] resume VM request: VMID:"505"
```
Platform: Host: Ubuntu 18, v4.15. Guest: containerd's default kernel and rootfs | 1.0 | InvalidOffset (virtio-block) error when resuming after loading a snapshot (with rootfs from firecracker-containerd) - Hi, we started developing support for snapshot `pause/resume/create/load` inside our fork of firecracker-containerd (we can make the code public). While we had no problem with supporting `pause/resume/create-snapshot` methods we ran into an error inside Firecracker's virtio-block module with `load-snapshot->resume`.
The workflow is startVM(boot) -> Pause -> Create-Snap -> Offload (kill the VM with SIGTERM) -> SnapshotLoad -> Resume
The problem we face is at the Resume point where we get the following error from firecracker's log:
```
2020-06-25T07:21:10.652265445 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Put request on "/snapshot/load" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\"}".
2020-06-25T07:21:10.664695532 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:21:10.665787599 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Patch request on "/vm" with body "{\"state\":\"Resumed\"}".
2020-06-25T07:21:10.665908820 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:21:14.799994834 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset)
2020-06-25T07:21:14.802628426 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset)
```
We tend to think that this can be related to the way firecracker-containerd manages block devices. The VM mounts block devices in 2 phases. First, the VM boots from a generic rootfs (with the containerd agent):
```
The API server received a Put request on "/drives/root_drive" with body "The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n".
```
Then the agent needs to mount a second block device that contains the container-specific data. To do so, containerd attaches another drive and patches the path to the drive twice:
```
The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n".
<<..>>
The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n".
The request was executed successfully. Status code: 204 No Content.
The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n".
```
Can the issue be connected to the way [PATCH drive works](https://github.com/firecracker-microvm/firecracker/blob/master/docs/api_requests/patch-block.md)? This drive is supposed to remain mounted into the restored guest.
We would greatly appreciate comments and ideas on what could be the root cause from Firecracker and firecracker-containerd maintainers: for example, @acatangiu @kzys . Once we fix the issue, we would be happy to contribute our changes to firecracker-containerd upstream.
Full workflow log (Firecracker's log):
```
Running Firecracker v0.21.0
2020-06-25T07:20:17.287742839 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.288144594 [anonymous-instance:INFO] The API server received a Put request on "/machine-config" with body "{\"cpu_template\":\"T2\",\"ht_enabled\":false,\"mem_size_mib\":512,\"vcpu_count\":1}\n".
2020-06-25T07:20:17.288291418 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.288521865 [anonymous-instance:INFO] The API server received a Get request on "/machine-config".
2020-06-25T07:20:17.288625819 [anonymous-instance:INFO] The request was executed successfully. Status code: 200 OK.
2020-06-25T07:20:17.288971127 [anonymous-instance:INFO] The API server received a Put request on "/boot-source" with body "{\"boot_args\":\"8250.nr_uarts=0 ip=190.128.0.2::190.128.0.1:255.192.0.0:::off::: systemd.log_color=false init=/sbin/overlay-init systemd.unit=firecracker.target quiet noapic nomodules ipv6.disable=1 ro panic=1 tsc=reliable reboot=k pci=off\",\"kernel_image_path\":\"/var/lib/firecracker-containerd/runtime/hello-vmlinux.bin\"}\n".
2020-06-25T07:20:17.289115444 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.289571124 [anonymous-instance:INFO] The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n".
2020-06-25T07:20:17.289732138 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.290046499 [anonymous-instance:INFO] The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n".
2020-06-25T07:20:17.290151860 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.290522553 [anonymous-instance:INFO] The API server received a Put request on "/network-interfaces/1" with body "{\"guest_mac\":\"02:FC:00:00:00:00\",\"host_dev_name\":\"fc-0-tap0\",\"iface_id\":\"1\"}\n".
2020-06-25T07:20:17.292898333 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.293236487 [anonymous-instance:INFO] The API server received a Put request on "/vsock" with body "{\"guest_cid\":0,\"uds_path\":\"firecracker.vsock\",\"vsock_id\":\"agent_api\"}\n".
2020-06-25T07:20:17.293472939 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:17.293754191 [anonymous-instance:INFO] The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n".
2020-06-25T07:20:17.305625745 [anonymous-instance:WARN] Could not add serial input event to epoll: Error during epoll call: Operation not permitted (os error 1)
2020-06-25T07:20:17.306309907 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:19.177877690 [anonymous-instance:INFO] The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n".
2020-06-25T07:20:19.178152684 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:20.104482947 [anonymous-instance:INFO] The API server received a Patch request on "/vm" with body "{\"state\":\"Paused\"}".
2020-06-25T07:20:20.104702379 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
2020-06-25T07:20:20.106119166 [anonymous-instance:INFO] The API server received a Put request on "/snapshot/create" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\",\"snapshot_type\":\"Full\"}".
2020-06-25T07:20:20.649402863 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content.
```
And the corresponding log from firecracker-containerd (containerd's log)
```
INFO[2020-06-25T07:20:17.275031352-04:00] creating new VM runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:17.275464729-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" shim_stream=stderr vmID=505
setupLogging
setupMetrics
ERRO[2020-06-25T07:20:17.288041835-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="VMM metrics disabled." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.288909406-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate:T2 HtEnabled:0xc000582f43 MemSizeMib:0xc000582f38 VcpuCount:0xc000582f30}" shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289306935-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289343891-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289933498-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.289952823-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0, slot MN2HE43UOVRDA, root false." shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.290347821-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:17.290377745-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching NIC fc-0-tap0 (hwaddr 02:FC:00:00:00:00) at index 1" shim_stream=stderr vmID=505
INFO[2020-06-25T07:20:17.306497334-04:00] calling agent runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:17.306529875-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " shim_stream=stderr vmID=505
DEBU[2020-06-25T07:20:18.004394761-04:00] garbage collected d=10.79516ms
INFO[2020-06-25T07:20:18.410674791-04:00] successfully started the VM runtime=aws.firecracker vmID=505
DEBU[2020-06-25T07:20:18.411052023-04:00] event forwarded ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
DEBU[2020-06-25T07:20:18.440217583-04:00] prepare snapshot key=505 parent="sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc"
DEBU[2020-06-25T07:20:18.440347865-04:00] prepare key=firecracker-containerd/17/505 parent="firecracker-containerd/16/sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc"
DEBU[2020-06-25T07:20:18.440504004-04:00] creating snapshot device 'fc-dev-thinpool-snap-9' from 'fc-dev-thinpool-snap-8'
DEBU[2020-06-25T07:20:18.795683399-04:00] event published ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
DEBU[2020-06-25T07:20:18.800390309-04:00] get snapshot mounts key=505
DEBU[2020-06-25T07:20:18.800469759-04:00] mounts key=firecracker-containerd/17/505
DEBU[2020-06-25T07:20:19.120516743-04:00] event published ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
DEBU[2020-06-25T07:20:19.138167895-04:00] get snapshot mounts key=505
DEBU[2020-06-25T07:20:19.138206395-04:00] mounts key=firecracker-containerd/17/505
time="2020-06-25T07:20:19.174115148-04:00" level=debug msg=StartShim runtime=aws.firecracker task_id=505
DEBU[2020-06-25T07:20:19.175656280-04:00] create VM request: VMID:"505"
DEBU[2020-06-25T07:20:19.175707632-04:00] using namespace: firecracker-containerd
time="2020-06-25T07:20:19.176188684-04:00" level=info msg="successfully started shim (git commit: 22b7cfceb4901b91c998a380df74d9b825614967)." runtime=aws.firecracker task_id=505 vmID=505
ERRO[2020-06-25T07:20:19.178386061-04:00] time="2020-06-25T07:20:19-04:00" level=info msg="PatchGuestDrive successful" shim_stream=stderr vmID=505
INFO[2020-06-25T07:20:19.330752483-04:00] successfully created task ExecID= TaskID=505 pid_in_vm=727 runtime=aws.firecracker vmID=505
DEBU[2020-06-25T07:20:19.331251986-04:00] event forwarded ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
DEBU[2020-06-25T07:20:19.337046160-04:00] garbage collected d=12.841932ms
DEBU[2020-06-25T07:20:19.339260338-04:00] event forwarded ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
DEBU[2020-06-25T07:20:20.103637583-04:00] pause VM request: VMID:"505"
DEBU[2020-06-25T07:20:20.105566980-04:00] make snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file"
DEBU[2020-06-25T07:20:20.650066779-04:00] offload original request: VMID:"505"
ERRO[2020-06-25T07:20:25.892253044-04:00] error waiting for exit ExecID= TaskID=505 error="ttrpc: closed" runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:25.892257191-04:00] error while forwarding events from VM agent error="ttrpc: closed" runtime=aws.firecracker vmID=505
ERRO[2020-06-25T07:20:25.892484333-04:00] error copying io ExecID= TaskID=505 error="reading from a closed fifo" runtime=aws.firecracker stream=stdin vmID=505
ERRO[2020-06-25T07:20:25.892580185-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=505
ERRO[2020-06-25T07:20:25.923729272-04:00] time="2020-06-25T07:20:25-04:00" level=warning msg="firecracker exited: signal: killed" shim_stream=stderr vmID=505
ERRO[2020-06-25T07:20:30.892486598-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=505
ERRO[2020-06-25T07:20:30.892525517-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=505
DEBU[2020-06-25T07:21:10.651353555-04:00] load snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file"
DEBU[2020-06-25T07:21:10.665366928-04:00] resume VM request: VMID:"505"
```
Platform: Host: Ubuntu 18, v4.15. Guest: containerd's default kernel and rootfs | priority | invalidoffset virtio block error when resuming after loading a snapshot with rootfs from firecracker containerd hi we started developing support for snapshot pause resume create load inside our fork of firecracker containerd we can make the code public while we had no problem with supporting pause resume create snapshot methods we ran into an error inside firecracker s virtio block module with load snapshot resume the workflow is startvm boot pause create snap offload kill the vm with sigterm snapshotload resume the problem we face is at the resume point where we get the following error from firecracker s log the api server received a put request on snapshot load with body mem file path tmp mem file snapshot path tmp snapshot file the request was executed successfully status code no content the api server received a patch request on vm with body state resumed the request was executed successfully status code no content failed to execute request badrequest invalidoffset failed to execute request badrequest invalidoffset we tend to think that this can be related to the way firecracker containerd manages block devices the vm mounts block devices in phases first the vm boots from a generic rootfs with the containerd agent the api server received a put request on drives root drive with body the api server received a put request on drives root drive with body drive id root drive is read only true is root device true path on host var lib firecracker containerd runtime default rootfs img n then the agent needs to mount a second block device that contains the container specific data to do so containerd attaches another drive and patches the path to the drive twice the api server received a put request on drives with body drive id is read only false is root device false path on host var lib firecracker containerd shim base firecracker containerd n the api server received a put request on actions with body action type instancestart n the request was executed successfully status code no content the api server received a patch request on drives with body drive id path on host dev mapper fc dev thinpool snap n can the issue be connected to the way this drive is supposed to remain mounted into the restored guest we would greatly appreciate comments and ideas on what could be the root cause from firecracker and firecracker containerd maintainers for example acatangiu kzys once we fix the issue we would be happy to contribute our changes to firecracker containerd upstream full workflow log firecracker s log running firecracker the request was executed successfully status code no content the api server received a put request on machine config with body cpu template ht enabled false mem size mib vcpu count n the request was executed successfully status code no content the api server received a get request on machine config the request was executed successfully status code ok the api server received a put request on boot source with body boot args nr uarts ip off systemd log color false init sbin overlay init systemd unit firecracker target quiet noapic nomodules disable ro panic tsc reliable reboot k pci off kernel image path var lib firecracker containerd runtime hello vmlinux bin n the request was executed successfully status code no content the api server received a put request on drives root drive with body drive id root drive is read only true is root device true path on host var lib firecracker containerd runtime default rootfs img n the request was executed successfully status code no content the api server received a put request on drives with body drive id is read only false is root device false path on host var lib firecracker containerd shim base firecracker containerd n the request was executed successfully status code no content the api server received a put request on network interfaces with body guest mac fc host dev name fc iface id n the request was executed successfully status code no content the api server received a put request on vsock with body guest cid uds path firecracker vsock vsock id agent api n the request was executed successfully status code no content the api server received a put request on actions with body action type instancestart n could not add serial input event to epoll error during epoll call operation not permitted os error the request was executed successfully status code no content the api server received a patch request on drives with body drive id path on host dev mapper fc dev thinpool snap n the request was executed successfully status code no content the api server received a patch request on vm with body state paused the request was executed successfully status code no content the api server received a put request on snapshot create with body mem file path tmp mem file snapshot path tmp snapshot file snapshot type full the request was executed successfully status code no content and the corresponding log from firecracker containerd containerd s log info creating new vm runtime aws firecracker vmid erro time level info msg called startvmm setting up a vmm on firecracker sock shim stream stderr vmid setuplogging setupmetrics erro time level info msg vmm metrics disabled shim stream stderr vmid erro time level info msg refreshmachineconfiguration getmachineconfigurationok cputemplate htenabled memsizemib vcpucount shim stream stderr vmid erro time level info msg putguestbootsource putguestbootsourcenocontent shim stream stderr vmid erro time level info msg attaching drive var lib firecracker containerd runtime default rootfs img slot root drive root true shim stream stderr vmid erro time level info msg attached drive var lib firecracker containerd runtime default rootfs img putguestdrivebyidnocontent shim stream stderr vmid erro time level info msg attaching drive var lib firecracker containerd shim base firecracker containerd slot root false shim stream stderr vmid erro time level info msg attached drive var lib firecracker containerd shim base firecracker containerd putguestdrivebyidnocontent shim stream stderr vmid erro time level info msg attaching nic fc hwaddr fc at index shim stream stderr vmid info calling agent runtime aws firecracker vmid erro time level info msg startinstance successful createsyncactionnocontent shim stream stderr vmid debu garbage collected d info successfully started the vm runtime aws firecracker vmid debu event forwarded ns firecracker containerd topic firecracker vm start type vmstart debu prepare snapshot key parent debu prepare key firecracker containerd parent firecracker containerd debu creating snapshot device fc dev thinpool snap from fc dev thinpool snap debu event published ns firecracker containerd topic snapshot prepare type containerd events snapshotprepare debu get snapshot mounts key debu mounts key firecracker containerd debu event published ns firecracker containerd topic containers create type containerd events containercreate debu get snapshot mounts key debu mounts key firecracker containerd time level debug msg startshim runtime aws firecracker task id debu create vm request vmid debu using namespace firecracker containerd time level info msg successfully started shim git commit runtime aws firecracker task id vmid erro time level info msg patchguestdrive successful shim stream stderr vmid info successfully created task execid taskid pid in vm runtime aws firecracker vmid debu event forwarded ns firecracker containerd topic tasks create type containerd events taskcreate debu garbage collected d debu event forwarded ns firecracker containerd topic tasks start type containerd events taskstart debu pause vm request vmid debu make snapshot request vmid snapshotfilepath tmp snapshot file memfilepath tmp mem file debu offload original request vmid erro error waiting for exit execid taskid error ttrpc closed runtime aws firecracker vmid erro error while forwarding events from vm agent error ttrpc closed runtime aws firecracker vmid erro error copying io execid taskid error reading from a closed fifo runtime aws firecracker stream stdin vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stdin vmid erro time level warning msg firecracker exited signal killed shim stream stderr vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stderr vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stdout vmid debu load snapshot request vmid snapshotfilepath tmp snapshot file memfilepath tmp mem file debu resume vm request vmid platform host ubuntu guest containerd s default kernel and rootfs | 1 |
159,064 | 6,040,075,512 | IssuesEvent | 2017-06-10 10:34:14 | zulip/zulip-electron | https://api.github.com/repos/zulip/zulip-electron | closed | Add desktop notifications support on Windows 7 | help wanted Priority: High Type: Enhancement windows | According to electron documentation Notification api does not work on Windows7, it has to use Tray baloon api.
Is it possible to implement baloon notifications ? It would most propably require work also on standard website.
| 1.0 | Add desktop notifications support on Windows 7 - According to electron documentation Notification api does not work on Windows7, it has to use Tray baloon api.
Is it possible to implement baloon notifications ? It would most propably require work also on standard website.
| priority | add desktop notifications support on windows according to electron documentation notification api does not work on it has to use tray baloon api is it possible to implement baloon notifications it would most propably require work also on standard website | 1 |
343,913 | 10,338,268,977 | IssuesEvent | 2019-09-03 16:31:15 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | [feature request] `select_index` for sparse tensors | feature high priority module: sparse todo triaged | ## Issue description
Sparse tensors do not currently support indexing (neither `gather`, nor `select_index`). This could be handy for e.g. embedding categorical features with non-contiguous vocabulary.
Sidenote: a similar result can be achieved via bucketization (see #7284), without using sparse tensors.
## Code example
```
>>> import numpy as np
>>> import torch
>>> vocab = [102, 104, 2, 103, 0, 101, 3]
>>> lookup_table = torch.sparse.IntTensor(
... torch.tensor(np.atleast_2d(vocab)),
... torch.arange(len(vocab), dtype=torch.int),
... torch.Size((max(vocab) + 1, )))
...
>>> lookup_table[102]
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
RuntimeError: Sparse tensors do not have strides.
```
## System Info
```
PyTorch version: 0.5.0a0+3e785d5
Is debug build: No
CUDA used to build PyTorch: None
OS: Mac OSX 10.13.3
GCC version: Could not collect
CMake version: version 3.9.4
Python version: 3.6
Is CUDA available: No
CUDA runtime version: No CUDA
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
Versions of relevant libraries:
[pip3] numpy (1.14.2)
[pip3] torch (0.5.0a0+3e785d5)
[pip3] torchvision (0.2.1)
[conda] torch 0.4.0 <pip>
[conda] torch 0.5.0a0+3e785d5 <pip>
[conda] torchvision 0.2.1 <pip>
```
| 1.0 | [feature request] `select_index` for sparse tensors - ## Issue description
Sparse tensors do not currently support indexing (neither `gather`, nor `select_index`). This could be handy for e.g. embedding categorical features with non-contiguous vocabulary.
Sidenote: a similar result can be achieved via bucketization (see #7284), without using sparse tensors.
## Code example
```
>>> import numpy as np
>>> import torch
>>> vocab = [102, 104, 2, 103, 0, 101, 3]
>>> lookup_table = torch.sparse.IntTensor(
... torch.tensor(np.atleast_2d(vocab)),
... torch.arange(len(vocab), dtype=torch.int),
... torch.Size((max(vocab) + 1, )))
...
>>> lookup_table[102]
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
RuntimeError: Sparse tensors do not have strides.
```
## System Info
```
PyTorch version: 0.5.0a0+3e785d5
Is debug build: No
CUDA used to build PyTorch: None
OS: Mac OSX 10.13.3
GCC version: Could not collect
CMake version: version 3.9.4
Python version: 3.6
Is CUDA available: No
CUDA runtime version: No CUDA
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
Versions of relevant libraries:
[pip3] numpy (1.14.2)
[pip3] torch (0.5.0a0+3e785d5)
[pip3] torchvision (0.2.1)
[conda] torch 0.4.0 <pip>
[conda] torch 0.5.0a0+3e785d5 <pip>
[conda] torchvision 0.2.1 <pip>
```
| priority | select index for sparse tensors issue description sparse tensors do not currently support indexing neither gather nor select index this could be handy for e g embedding categorical features with non contiguous vocabulary sidenote a similar result can be achieved via bucketization see without using sparse tensors code example import numpy as np import torch vocab lookup table torch sparse inttensor torch tensor np atleast vocab torch arange len vocab dtype torch int torch size max vocab lookup table traceback most recent call last file line in runtimeerror sparse tensors do not have strides system info pytorch version is debug build no cuda used to build pytorch none os mac osx gcc version could not collect cmake version version python version is cuda available no cuda runtime version no cuda gpu models and configuration no cuda nvidia driver version no cuda cudnn version no cuda versions of relevant libraries numpy torch torchvision torch torch torchvision | 1 |
343,825 | 10,336,956,183 | IssuesEvent | 2019-09-03 13:59:29 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Cannot load multivariable netcdf files | Bug Feedback High Priority Raster Regression | Author Name: **saverio vicario** (saverio vicario)
Original Redmine Issue: [21822](https://issues.qgis.org/issues/21822)
Affected QGIS version: 3.4.6
Redmine category:rasters
---
I upgraded QGis and I cannot load some netcdf anymore. All the file multivariable that I have cannot be read.
The netcdf that I can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files.
When I try to load a multivariable netcdf Qgis report on the GUI that source is not valid and on terminal I am getting only warnings relative to the naming of dimensions
"
Warning 1: dimension #2 (x) is not a Longitude/X dimension.
Warning 1: dimension #1 (y) is not a Latitude/Y dimension.
Warning 1: dimension #0 (Year) is not a Time or Vertical dimension.
"
so no critical error.
When I am using gdalinfo is all fine.
when I am using gdal_translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself.
---
- [Testy.nc](https://issues.qgis.org/attachments/download/14779/Testy.nc) (saverio vicario)
---
Related issue(s): #29598 (relates)
Redmine related issue(s): [21783](https://issues.qgis.org/issues/21783)
---
| 1.0 | Cannot load multivariable netcdf files - Author Name: **saverio vicario** (saverio vicario)
Original Redmine Issue: [21822](https://issues.qgis.org/issues/21822)
Affected QGIS version: 3.4.6
Redmine category:rasters
---
I upgraded QGis and I cannot load some netcdf anymore. All the file multivariable that I have cannot be read.
The netcdf that I can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files.
When I try to load a multivariable netcdf Qgis report on the GUI that source is not valid and on terminal I am getting only warnings relative to the naming of dimensions
"
Warning 1: dimension #2 (x) is not a Longitude/X dimension.
Warning 1: dimension #1 (y) is not a Latitude/Y dimension.
Warning 1: dimension #0 (Year) is not a Time or Vertical dimension.
"
so no critical error.
When I am using gdalinfo is all fine.
when I am using gdal_translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself.
---
- [Testy.nc](https://issues.qgis.org/attachments/download/14779/Testy.nc) (saverio vicario)
---
Related issue(s): #29598 (relates)
Redmine related issue(s): [21783](https://issues.qgis.org/issues/21783)
---
| priority | cannot load multivariable netcdf files author name saverio vicario saverio vicario original redmine issue affected qgis version redmine category rasters i upgraded qgis and i cannot load some netcdf anymore all the file multivariable that i have cannot be read the netcdf that i can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files when i try to load a multivariable netcdf qgis report on the gui that source is not valid and on terminal i am getting only warnings relative to the naming of dimensions warning dimension x is not a longitude x dimension warning dimension y is not a latitude y dimension warning dimension year is not a time or vertical dimension so no critical error when i am using gdalinfo is all fine when i am using gdal translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself saverio vicario related issue s relates redmine related issue s | 1 |
263,120 | 8,273,715,368 | IssuesEvent | 2018-09-17 07:16:56 | theblacksea/issues | https://api.github.com/repos/theblacksea/issues | closed | social media | High Priority | Social media sharing link needs to be fixed.
With Facebook and Twitter, sharing doesn't work properly, it doesn't show the story, only shows the theblacksea.eu address. Please see the picture explaining

| 1.0 | social media - Social media sharing link needs to be fixed.
With Facebook and Twitter, sharing doesn't work properly, it doesn't show the story, only shows the theblacksea.eu address. Please see the picture explaining

| priority | social media social media sharing link needs to be fixed with facebook and twitter sharing doesn t work properly it doesn t show the story only shows the theblacksea eu address please see the picture explaining | 1 |
200,281 | 7,005,201,993 | IssuesEvent | 2017-12-19 00:37:50 | python/mypy | https://api.github.com/repos/python/mypy | closed | Calling `callable` in an if-condition makes mypy think branch in unreachable | bug priority-0-high | In the following code, mypy thinking `callable(o)` always returns `False`. This might be related to #3603.
```py
def f(o: object) -> None:
if callable(o):
# this branch is not typechecked
o()
1 + 'boom' # no error from mypy
o()
f(lambda: print('hello'))
```
Here is the output from `mypy` and `python`.
```
$ mypy n.py
$ python3 n.py
hello
Traceback (most recent call last):
File "n.py", line 8, in <module>
f(lambda: print('hello'))
File "n.py", line 5, in f
1 + 'boom'
TypeError: unsupported operand type(s) for +: 'int' and 'str'
``` | 1.0 | Calling `callable` in an if-condition makes mypy think branch in unreachable - In the following code, mypy thinking `callable(o)` always returns `False`. This might be related to #3603.
```py
def f(o: object) -> None:
if callable(o):
# this branch is not typechecked
o()
1 + 'boom' # no error from mypy
o()
f(lambda: print('hello'))
```
Here is the output from `mypy` and `python`.
```
$ mypy n.py
$ python3 n.py
hello
Traceback (most recent call last):
File "n.py", line 8, in <module>
f(lambda: print('hello'))
File "n.py", line 5, in f
1 + 'boom'
TypeError: unsupported operand type(s) for +: 'int' and 'str'
``` | priority | calling callable in an if condition makes mypy think branch in unreachable in the following code mypy thinking callable o always returns false this might be related to py def f o object none if callable o this branch is not typechecked o boom no error from mypy o f lambda print hello here is the output from mypy and python mypy n py n py hello traceback most recent call last file n py line in f lambda print hello file n py line in f boom typeerror unsupported operand type s for int and str | 1 |
601,855 | 18,437,684,507 | IssuesEvent | 2021-10-14 14:37:05 | DXgovernance/dxvote | https://api.github.com/repos/DXgovernance/dxvote | closed | New proposal metadata page | Enhancement High Priority | **Blocked by #187**
In the simplified UX flow for proposal creation we will have a first view after selecting proposal type be a metadata editor.
For the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later.
This first implementation should be accessible upon clicking the "Contributor proposal" option. (Not CUSTOM)
This page should be paramterised and reusable, multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposalType must be passed through to know where to navigate the user afterwards. (For now the page should not lead anywhere)
The screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look, feel free to ask for feedback from everyone.
Basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform MD editor with preview on the right.
<img width="905" alt="スクリーンショット 2021-09-28 9 38 04" src="https://user-images.githubusercontent.com/39137239/135053467-f745b6c8-f14d-456f-8a40-ebf905629c7b.png">
https://www.figma.com/file/N4e2YuSiv1SycIhmqhIXGB/DXgov?node-id=95%3A2 | 1.0 | New proposal metadata page - **Blocked by #187**
In the simplified UX flow for proposal creation we will have a first view after selecting proposal type be a metadata editor.
For the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later.
This first implementation should be accessible upon clicking the "Contributor proposal" option. (Not CUSTOM)
This page should be paramterised and reusable, multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposalType must be passed through to know where to navigate the user afterwards. (For now the page should not lead anywhere)
The screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look, feel free to ask for feedback from everyone.
Basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform MD editor with preview on the right.
<img width="905" alt="スクリーンショット 2021-09-28 9 38 04" src="https://user-images.githubusercontent.com/39137239/135053467-f745b6c8-f14d-456f-8a40-ebf905629c7b.png">
https://www.figma.com/file/N4e2YuSiv1SycIhmqhIXGB/DXgov?node-id=95%3A2 | priority | new proposal metadata page blocked by in the simplified ux flow for proposal creation we will have a first view after selecting proposal type be a metadata editor for the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later this first implementation should be accessible upon clicking the contributor proposal option not custom this page should be paramterised and reusable multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposaltype must be passed through to know where to navigate the user afterwards for now the page should not lead anywhere the screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look feel free to ask for feedback from everyone basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform md editor with preview on the right img width alt スクリーンショット src | 1 |
544,253 | 15,891,627,858 | IssuesEvent | 2021-04-10 20:02:28 | wso2/streaming-integrator | https://api.github.com/repos/wso2/streaming-integrator | closed | SI Extension installer overwrites the jars in the <SI_HOME>/lib directory | Priority/Highest Severity/Major T7 Type/Bug | **Description:**
SI Extension installer overwrites the jars in the <SI_HOME>/lib directory
**Affected Product Version:**
SI 1.1.0
**OS, DB, other environment details and versions:**
N/A | 1.0 | SI Extension installer overwrites the jars in the <SI_HOME>/lib directory - **Description:**
SI Extension installer overwrites the jars in the <SI_HOME>/lib directory
**Affected Product Version:**
SI 1.1.0
**OS, DB, other environment details and versions:**
N/A | priority | si extension installer overwrites the jars in the lib directory description si extension installer overwrites the jars in the lib directory affected product version si os db other environment details and versions n a | 1 |
269,403 | 8,435,428,556 | IssuesEvent | 2018-10-17 13:08:34 | vasyoid/cremebrulee | https://api.github.com/repos/vasyoid/cremebrulee | opened | Ошибка в сортировке при отображении списка экскурсий | bug high importance high priority | При указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом, чем следующие варианты. | 1.0 | Ошибка в сортировке при отображении списка экскурсий - При указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом, чем следующие варианты. | priority | ошибка в сортировке при отображении списка экскурсий при указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом чем следующие варианты | 1 |
660,824 | 22,032,586,746 | IssuesEvent | 2022-05-28 04:21:53 | sqlalchemy/sqlalchemy | https://api.github.com/repos/sqlalchemy/sqlalchemy | closed | no control of DDL sequences for indexes, FK constraints that are mutually dependent | bug duplicate high priority schema | **Migrated issue, originally created by Michael Bayer ([@zzzeek](https://github.com/zzzeek))**
the whole system described at http://docs.sqlalchemy.org/en/rel_1_0/core/ddl.html#controlling-ddl-sequences only takes place for constraints that are rendered inline within the table definition via the _create_rule() callable checked in compiler.py. This callable is not consulted anywhere in ddl.py when it goes to create indexes, foreign keys, or for that matter sequences.
e.g. note below they are all execute_if(postgresql) but the engine is MySQL, and they all get built
```
from sqlalchemy import *
from sqlalchemy.orm import *
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
class A(Base):
__tablename__ = 'a'
id = Column(Integer, primary_key=True)
x = Column(Integer)
favorite_b = Column(Integer)
class B(Base):
__tablename__ = 'b'
id = Column(Integer, primary_key=True)
a_id = Column(Integer)
from sqlalchemy.schema import CreateIndex, AddConstraint
idx = Index('aidx', A.x)
fk1 = ForeignKeyConstraint([B.a_id], [A.id], name='afk')
fk2 = ForeignKeyConstraint([A.favorite_b], [B.id], name='bfk')
from sqlalchemy import event
event.listen(
Base.metadata,
'after_create',
CreateIndex(idx).execute_if(dialect='postgresql')
)
event.listen(
Base.metadata,
'after_create',
AddConstraint(fk1).execute_if(dialect='postgresql')
)
event.listen(
Base.metadata,
'after_create',
AddConstraint(fk2).execute_if(dialect='postgresql')
)
e = create_engine("mysql://scott:tiger@localhost/test", echo=True)
Base.metadata.drop_all(e)
Base.metadata.create_all(e)
```
----------------------------------------
Attachments: [3442.patch](../wiki/imported_issue_attachments/3442/3442.patch)
| 1.0 | no control of DDL sequences for indexes, FK constraints that are mutually dependent - **Migrated issue, originally created by Michael Bayer ([@zzzeek](https://github.com/zzzeek))**
the whole system described at http://docs.sqlalchemy.org/en/rel_1_0/core/ddl.html#controlling-ddl-sequences only takes place for constraints that are rendered inline within the table definition via the _create_rule() callable checked in compiler.py. This callable is not consulted anywhere in ddl.py when it goes to create indexes, foreign keys, or for that matter sequences.
e.g. note below they are all execute_if(postgresql) but the engine is MySQL, and they all get built
```
from sqlalchemy import *
from sqlalchemy.orm import *
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
class A(Base):
__tablename__ = 'a'
id = Column(Integer, primary_key=True)
x = Column(Integer)
favorite_b = Column(Integer)
class B(Base):
__tablename__ = 'b'
id = Column(Integer, primary_key=True)
a_id = Column(Integer)
from sqlalchemy.schema import CreateIndex, AddConstraint
idx = Index('aidx', A.x)
fk1 = ForeignKeyConstraint([B.a_id], [A.id], name='afk')
fk2 = ForeignKeyConstraint([A.favorite_b], [B.id], name='bfk')
from sqlalchemy import event
event.listen(
Base.metadata,
'after_create',
CreateIndex(idx).execute_if(dialect='postgresql')
)
event.listen(
Base.metadata,
'after_create',
AddConstraint(fk1).execute_if(dialect='postgresql')
)
event.listen(
Base.metadata,
'after_create',
AddConstraint(fk2).execute_if(dialect='postgresql')
)
e = create_engine("mysql://scott:tiger@localhost/test", echo=True)
Base.metadata.drop_all(e)
Base.metadata.create_all(e)
```
----------------------------------------
Attachments: [3442.patch](../wiki/imported_issue_attachments/3442/3442.patch)
| priority | no control of ddl sequences for indexes fk constraints that are mutually dependent migrated issue originally created by michael bayer the whole system described at only takes place for constraints that are rendered inline within the table definition via the create rule callable checked in compiler py this callable is not consulted anywhere in ddl py when it goes to create indexes foreign keys or for that matter sequences e g note below they are all execute if postgresql but the engine is mysql and they all get built from sqlalchemy import from sqlalchemy orm import from sqlalchemy ext declarative import declarative base base declarative base class a base tablename a id column integer primary key true x column integer favorite b column integer class b base tablename b id column integer primary key true a id column integer from sqlalchemy schema import createindex addconstraint idx index aidx a x foreignkeyconstraint name afk foreignkeyconstraint name bfk from sqlalchemy import event event listen base metadata after create createindex idx execute if dialect postgresql event listen base metadata after create addconstraint execute if dialect postgresql event listen base metadata after create addconstraint execute if dialect postgresql e create engine mysql scott tiger localhost test echo true base metadata drop all e base metadata create all e attachments wiki imported issue attachments patch | 1 |
606,048 | 18,753,941,060 | IssuesEvent | 2021-11-05 08:12:43 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Golang API | Type: Backlog Category: API Status: Accepted Priority: High | # Enhancement Request
## Description
Implement official Go binding for OpenNebula. It should be based on https://github.com/OpenNebula/goca.
## Use case
OpenNebula should have a new API for Golang. It should be used in a similar way as the Java API.
## Interface Changes
New API.
# Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Insert into release process
- [ ] Documentation (including code examples)
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
- [ ] Close https://github.com/OpenNebula/goca and reference the new docs | 1.0 | Golang API - # Enhancement Request
## Description
Implement official Go binding for OpenNebula. It should be based on https://github.com/OpenNebula/goca.
## Use case
OpenNebula should have a new API for Golang. It should be used in a similar way as the Java API.
## Interface Changes
New API.
# Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Insert into release process
- [ ] Documentation (including code examples)
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
- [ ] Close https://github.com/OpenNebula/goca and reference the new docs | priority | golang api enhancement request description implement official go binding for opennebula it should be based on use case opennebula should have a new api for golang it should be used in a similar way as the java api interface changes new api progress status branch created code committed to development branch testing qa insert into release process documentation including code examples release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches close and reference the new docs | 1 |
442,863 | 12,751,964,282 | IssuesEvent | 2020-06-27 14:00:55 | LewisGaul/minegauler | https://api.github.com/repos/LewisGaul/minegauler | closed | Changing Max mines per cell mid game | bug high-priority resolved | Starting a game with max mines per cell as 3 and then changing it to 1 makes you unable to flag multiple mines on a single cell | 1.0 | Changing Max mines per cell mid game - Starting a game with max mines per cell as 3 and then changing it to 1 makes you unable to flag multiple mines on a single cell | priority | changing max mines per cell mid game starting a game with max mines per cell as and then changing it to makes you unable to flag multiple mines on a single cell | 1 |
595,452 | 18,067,330,395 | IssuesEvent | 2021-09-20 20:51:34 | zapzombies/zap-community | https://api.github.com/repos/zapzombies/zap-community | opened | Mobs riding other mobs do not pathfind properly | bug feature-ai high-priority | **Describe the bug**
Mobs, such as the r25 boss(pig), cannot walk up blocks.
**Where did the bug occur**
Hades' Dominion
**To Reproduce**
Spawn bosspig, see how it does not move up blocks
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
This happens because the pathfinder is using the topmost pig's bounding box for pathfinding. It should be using the lowest bounding box.
| 1.0 | Mobs riding other mobs do not pathfind properly - **Describe the bug**
Mobs, such as the r25 boss(pig), cannot walk up blocks.
**Where did the bug occur**
Hades' Dominion
**To Reproduce**
Spawn bosspig, see how it does not move up blocks
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
This happens because the pathfinder is using the topmost pig's bounding box for pathfinding. It should be using the lowest bounding box.
| priority | mobs riding other mobs do not pathfind properly describe the bug mobs such as the boss pig cannot walk up blocks where did the bug occur hades dominion to reproduce spawn bosspig see how it does not move up blocks expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem additional context this happens because the pathfinder is using the topmost pig s bounding box for pathfinding it should be using the lowest bounding box | 1 |
389,627 | 11,504,732,060 | IssuesEvent | 2020-02-13 00:09:20 | fecgov/fec-cms | https://api.github.com/repos/fecgov/fec-cms | opened | Fix IE version filter to return current and unknown version when applied | Bug High priority | **What we're after:**
When the current version filter is applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true and most_recent=NULL, or current and unknown versions. At this time we are only seeing most_recent=true.
When the current version filter is not applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true, most_recent=false and most_recent=NULL.
**Key:**
- most_recent=true is current version
- most_recent=false is past version
- most_recent=NULL is unknown version
### Related issues
- #3545
### Completion criteria
- [ ] Make sure that the version filter works according to the above description.
| 1.0 | Fix IE version filter to return current and unknown version when applied - **What we're after:**
When the current version filter is applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true and most_recent=NULL, or current and unknown versions. At this time we are only seeing most_recent=true.
When the current version filter is not applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true, most_recent=false and most_recent=NULL.
**Key:**
- most_recent=true is current version
- most_recent=false is past version
- most_recent=NULL is unknown version
### Related issues
- #3545
### Completion criteria
- [ ] Make sure that the version filter works according to the above description.
| priority | fix ie version filter to return current and unknown version when applied what we re after when the current version filter is applied for raw all and processed and hour reports we should see most recent true and most recent null or current and unknown versions at this time we are only seeing most recent true when the current version filter is not applied for raw all and processed and hour reports we should see most recent true most recent false and most recent null key most recent true is current version most recent false is past version most recent null is unknown version related issues completion criteria make sure that the version filter works according to the above description | 1 |
78,272 | 3,509,547,465 | IssuesEvent | 2016-01-08 23:23:12 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Help Tavara Quest (BB #1043) | Category: Quests migrated Priority: High Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** Ezopko
**Original Date:** 10.07.2015 10:38:11 GMT+0000
**Original Priority:** critical
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1043
<hr>
Quest Tavara not Working..
http://www.wowhead.com/quest=9586/help-tavara
https://github.com/TrinityCore/TrinityCore/issues/2744 | 1.0 | Help Tavara Quest (BB #1043) - This issue was migrated from bitbucket.
**Original Reporter:** Ezopko
**Original Date:** 10.07.2015 10:38:11 GMT+0000
**Original Priority:** critical
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1043
<hr>
Quest Tavara not Working..
http://www.wowhead.com/quest=9586/help-tavara
https://github.com/TrinityCore/TrinityCore/issues/2744 | priority | help tavara quest bb this issue was migrated from bitbucket original reporter ezopko original date gmt original priority critical original type bug original state resolved direct link quest tavara not working | 1 |
736,718 | 25,484,426,393 | IssuesEvent | 2022-11-26 07:09:32 | xsuite/xsuite | https://api.github.com/repos/xsuite/xsuite | opened | When regenerating json for hllhc15 example test_match_tune_chroma fails | bug High priority | Old json was in the format with elements as list instead of elements as dict. Maybe a bug in Line.to_dict/Line.from_dict? | 1.0 | When regenerating json for hllhc15 example test_match_tune_chroma fails - Old json was in the format with elements as list instead of elements as dict. Maybe a bug in Line.to_dict/Line.from_dict? | priority | when regenerating json for example test match tune chroma fails old json was in the format with elements as list instead of elements as dict maybe a bug in line to dict line from dict | 1 |
330,550 | 10,050,207,110 | IssuesEvent | 2019-07-20 08:59:25 | WoWManiaUK/Blackwing-Lair | https://api.github.com/repos/WoWManiaUK/Blackwing-Lair | closed | [Zone] Stranglethorn Vale - Death | Confirmed Fixed Confirmed Fixed in Dev Priority-High zone 30 - 40 | **Links:**
https://www.wowhead.com/the-cape-of-stranglethorn
from WoWHead or our Armory
**What is happening:**
died in stranglethorn vale and was portaled to the spirit healer in westfall
**What should happen:**
be portaled to stranglethorn vale spirit healer
Coords : 44.01, 91.52
| 1.0 | [Zone] Stranglethorn Vale - Death - **Links:**
https://www.wowhead.com/the-cape-of-stranglethorn
from WoWHead or our Armory
**What is happening:**
died in stranglethorn vale and was portaled to the spirit healer in westfall
**What should happen:**
be portaled to stranglethorn vale spirit healer
Coords : 44.01, 91.52
| priority | stranglethorn vale death links from wowhead or our armory what is happening died in stranglethorn vale and was portaled to the spirit healer in westfall what should happen be portaled to stranglethorn vale spirit healer coords | 1 |
163,949 | 6,217,494,309 | IssuesEvent | 2017-07-08 15:30:01 | cytoscape/cytoscape.js | https://api.github.com/repos/cytoscape/cytoscape.js | closed | Benchmark and improve collection creation performance | performance priority-1-high | New collections are created by many APIs that devs use. They are also used internally in places.
It would benefit overall performance to make collection creation less expensive.
| 1.0 | Benchmark and improve collection creation performance - New collections are created by many APIs that devs use. They are also used internally in places.
It would benefit overall performance to make collection creation less expensive.
| priority | benchmark and improve collection creation performance new collections are created by many apis that devs use they are also used internally in places it would benefit overall performance to make collection creation less expensive | 1 |
502,746 | 14,566,183,977 | IssuesEvent | 2020-12-17 08:31:36 | CatalogueOfLife/portal | https://api.github.com/repos/CatalogueOfLife/portal | closed | References are missing with synonyms on species page | bug data high priority | Feedback from Paul Schoolmeesters:
>I regret that the original reference for a species is not used when there has been any modification, in which case the last reference is shown.
The example Ammoecioides catulus (Balthasar, 1946)
In 2019 reference is correct (in the book picture)
Best wishes, Paul Schoolmeesters
| 1.0 | References are missing with synonyms on species page - Feedback from Paul Schoolmeesters:
>I regret that the original reference for a species is not used when there has been any modification, in which case the last reference is shown.
The example Ammoecioides catulus (Balthasar, 1946)
In 2019 reference is correct (in the book picture)
Best wishes, Paul Schoolmeesters
| priority | references are missing with synonyms on species page feedback from paul schoolmeesters i regret that the original reference for a species is not used when there has been any modification in which case the last reference is shown the example ammoecioides catulus balthasar in reference is correct in the book picture best wishes paul schoolmeesters | 1 |
392,381 | 11,590,691,310 | IssuesEvent | 2020-02-24 07:34:28 | StanfordHCI/bang | https://api.github.com/repos/StanfordHCI/bang | closed | Dynamic Team size (1 v n) Breaks the Manipulation Check | bug high priority | When the team size is dynamic, and you have team size = 1, you cannot choose any partners during the final 'manipulation check' because there are no other partners.
<img width="1071" alt="Screen Shot 2020-02-11 at 11 02 21 PM" src="https://user-images.githubusercontent.com/28793641/74313299-492df280-4d28-11ea-8439-4ddb863c9091.png">
As a result, the partner-select is null. When the survey is submitted, this causes a crash.
**Below is the dump of the error I got.**
[reconvene-logic-issue-dev-break.txt](https://github.com/StanfordHCI/bang/files/4190452/reconvene-logic-issue-dev-break.txt)
**Suggestion:** When asking the question about the same partners, check to make sure you are comparing between rounds that actually have partners. | 1.0 | Dynamic Team size (1 v n) Breaks the Manipulation Check - When the team size is dynamic, and you have team size = 1, you cannot choose any partners during the final 'manipulation check' because there are no other partners.
<img width="1071" alt="Screen Shot 2020-02-11 at 11 02 21 PM" src="https://user-images.githubusercontent.com/28793641/74313299-492df280-4d28-11ea-8439-4ddb863c9091.png">
As a result, the partner-select is null. When the survey is submitted, this causes a crash.
**Below is the dump of the error I got.**
[reconvene-logic-issue-dev-break.txt](https://github.com/StanfordHCI/bang/files/4190452/reconvene-logic-issue-dev-break.txt)
**Suggestion:** When asking the question about the same partners, check to make sure you are comparing between rounds that actually have partners. | priority | dynamic team size v n breaks the manipulation check when the team size is dynamic and you have team size you cannot choose any partners during the final manipulation check because there are no other partners img width alt screen shot at pm src as a result the partner select is null when the survey is submitted this causes a crash below is the dump of the error i got suggestion when asking the question about the same partners check to make sure you are comparing between rounds that actually have partners | 1 |
538,492 | 15,769,961,700 | IssuesEvent | 2021-03-31 18:55:00 | Sage-Bionetworks/dccvalidator | https://api.github.com/repos/Sage-Bionetworks/dccvalidator | closed | Update AD instructions to point to docs submission form | consortia-related high priority | The vignette for the AD instructions is [here](https://github.com/Sage-Bionetworks/dccvalidator/blob/master/inst/using-the-dccvalidator-app-amp-ad.Rmd). This needs to be updated to remove information about uploading via the documentation tab and, instead, pointing the user to the Synapse submission form [here](https://www.synapse.org/#!Synapse:syn25051271).
@avanlinden and @pitviper6, due to this being an instruction set, I want to make sure it's correct for how you would like people to submit/validate data. I am a little out of the loop on the AD process other than knowing that docs are going to the submission form. Would you be able to give me the information you want in these two sections below? Alternatively, @avanlinden, this is an Rmarkdown file. If you would feel comfortable enough, feel free to start a branch with PR to update with changes.
https://github.com/Sage-Bionetworks/dccvalidator/blob/b122c93a796d613cb6846fa908a1625d72f5fd9c/inst/using-the-dccvalidator-app-amp-ad.Rmd#L47-L67 | 1.0 | Update AD instructions to point to docs submission form - The vignette for the AD instructions is [here](https://github.com/Sage-Bionetworks/dccvalidator/blob/master/inst/using-the-dccvalidator-app-amp-ad.Rmd). This needs to be updated to remove information about uploading via the documentation tab and, instead, pointing the user to the Synapse submission form [here](https://www.synapse.org/#!Synapse:syn25051271).
@avanlinden and @pitviper6, due to this being an instruction set, I want to make sure it's correct for how you would like people to submit/validate data. I am a little out of the loop on the AD process other than knowing that docs are going to the submission form. Would you be able to give me the information you want in these two sections below? Alternatively, @avanlinden, this is an Rmarkdown file. If you would feel comfortable enough, feel free to start a branch with PR to update with changes.
https://github.com/Sage-Bionetworks/dccvalidator/blob/b122c93a796d613cb6846fa908a1625d72f5fd9c/inst/using-the-dccvalidator-app-amp-ad.Rmd#L47-L67 | priority | update ad instructions to point to docs submission form the vignette for the ad instructions is this needs to be updated to remove information about uploading via the documentation tab and instead pointing the user to the synapse submission form avanlinden and due to this being an instruction set i want to make sure it s correct for how you would like people to submit validate data i am a little out of the loop on the ad process other than knowing that docs are going to the submission form would you be able to give me the information you want in these two sections below alternatively avanlinden this is an rmarkdown file if you would feel comfortable enough feel free to start a branch with pr to update with changes | 1 |
480,992 | 13,878,980,682 | IssuesEvent | 2020-10-17 12:22:03 | sButtons/sbuttons | https://api.github.com/repos/sButtons/sbuttons | closed | Place "Snake Border" button in the correct alphabetical order under Animated | Hacktoberfest Priority: High bug good first issue help wanted up-for-grabs website | Place "Snake border" in the correct alphabetical order under Animated. Changes should be made only in `assets/js/buttons-examples.js` | 1.0 | Place "Snake Border" button in the correct alphabetical order under Animated - Place "Snake border" in the correct alphabetical order under Animated. Changes should be made only in `assets/js/buttons-examples.js` | priority | place snake border button in the correct alphabetical order under animated place snake border in the correct alphabetical order under animated changes should be made only in assets js buttons examples js | 1 |
265,222 | 8,345,304,020 | IssuesEvent | 2018-10-01 01:36:07 | lgou2w/ldk | https://api.github.com/repos/lgou2w/ldk | closed | 0.1-rc | Enhancement Priority: Highest work in progress | ### 0.0.1-beta1 ✔️
- [x] Common
- [x] #2 ChatComponent
- [x] #3 NBT
- [x] #4 Document
- [x] #5 TAG
- [x] #6 I18N
- [x] Bukkit
- [x] #8
- [x] #9
- [x] ItemBuilder
- [x] ldk-bukkit-common Optimization
### 0.0.1-beta2 ✔️
- [x] Cmd API
- [x] Gui API
- [x] Region API
- [x] Depdend API
- [x] #15 Compatibility
- [x] ItemBuilder Advanced
- [x] Optimization
### 0.0.1-beta3 ✔️
- [x] Hikari module
- [x] Unility
- [x] Optimization
### 0.1-rc1 ✔️
- [x] Particle
- [x] Problem solved
- [x] Optimization
| 1.0 | 0.1-rc - ### 0.0.1-beta1 ✔️
- [x] Common
- [x] #2 ChatComponent
- [x] #3 NBT
- [x] #4 Document
- [x] #5 TAG
- [x] #6 I18N
- [x] Bukkit
- [x] #8
- [x] #9
- [x] ItemBuilder
- [x] ldk-bukkit-common Optimization
### 0.0.1-beta2 ✔️
- [x] Cmd API
- [x] Gui API
- [x] Region API
- [x] Depdend API
- [x] #15 Compatibility
- [x] ItemBuilder Advanced
- [x] Optimization
### 0.0.1-beta3 ✔️
- [x] Hikari module
- [x] Unility
- [x] Optimization
### 0.1-rc1 ✔️
- [x] Particle
- [x] Problem solved
- [x] Optimization
| priority | rc ✔️ common chatcomponent nbt document tag bukkit itembuilder ldk bukkit common optimization ✔️ cmd api gui api region api depdend api compatibility itembuilder advanced optimization ✔️ hikari module unility optimization ✔️ particle problem solved optimization | 1 |
117,213 | 4,712,451,699 | IssuesEvent | 2016-10-14 16:48:21 | genome/civic-client | https://api.github.com/repos/genome/civic-client | closed | Add funding acknowledgement section to about page | high priority reviewer feedback | We should add a section on the about page to gratefully acknowledge funding sources for CIViC. Maybe with their logos displayed as well (for NCI, NHGRI, NIH).
Example acknowledgement text as a starting point:
Acknowledgements and funding:
First and foremost we are grateful to the community of curators, editors, domain experts, and users who make CIViC possible. Malachi Griffith was supported by the National Human Genome Research Institute (NHGRI) of the National Institutes of Health (NIH) under Award Number K99HG007940. Obi Lee Griffith was supported by the National Cancer Institute of the NIH under Award Number K22CA188163. This work was also supported by a grant to Richard K. Wilson from the NHGRI under Award Number U54HG003079. The CIViC project is also supported by the NCI under Award Number U01CA209936 to Obi Lee Griffith (with Malachi Griffith and Elaine R Mardis as co-PIs). The content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the National Institutes of Health.
| 1.0 | Add funding acknowledgement section to about page - We should add a section on the about page to gratefully acknowledge funding sources for CIViC. Maybe with their logos displayed as well (for NCI, NHGRI, NIH).
Example acknowledgement text as a starting point:
Acknowledgements and funding:
First and foremost we are grateful to the community of curators, editors, domain experts, and users who make CIViC possible. Malachi Griffith was supported by the National Human Genome Research Institute (NHGRI) of the National Institutes of Health (NIH) under Award Number K99HG007940. Obi Lee Griffith was supported by the National Cancer Institute of the NIH under Award Number K22CA188163. This work was also supported by a grant to Richard K. Wilson from the NHGRI under Award Number U54HG003079. The CIViC project is also supported by the NCI under Award Number U01CA209936 to Obi Lee Griffith (with Malachi Griffith and Elaine R Mardis as co-PIs). The content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the National Institutes of Health.
| priority | add funding acknowledgement section to about page we should add a section on the about page to gratefully acknowledge funding sources for civic maybe with their logos displayed as well for nci nhgri nih example acknowledgement text as a starting point acknowledgements and funding first and foremost we are grateful to the community of curators editors domain experts and users who make civic possible malachi griffith was supported by the national human genome research institute nhgri of the national institutes of health nih under award number obi lee griffith was supported by the national cancer institute of the nih under award number this work was also supported by a grant to richard k wilson from the nhgri under award number the civic project is also supported by the nci under award number to obi lee griffith with malachi griffith and elaine r mardis as co pis the content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the national institutes of health | 1 |
261,223 | 8,228,136,381 | IssuesEvent | 2018-09-07 03:16:40 | phetsims/perennial | https://api.github.com/repos/phetsims/perennial | closed | Refining the latest links for phet-io | dev:phet-io meeting:phet-io priority:2-high | @jonathanolson, @zepumph, and I discussed the possibility of changing the `/latest` links for phet-io to instead be `/latest/{{major.minor}}`. The reasoning is this will allow clients to automatically pick up non-breaking maintenance releases, but avoid potentially api-breaking minor/major releases.
Here is an example of how the redirect structure would look in the .htaccess file:
```
RewriteEngine on
RewriteBase /sims/chains/
RewriteRule latest/1.17(.*) 1.17.0$1
RewriteRule latest/1.16(.*) 1.16.4$1
```
Here, the link `/sims/chains/latest/1.17` redirects to `/sims/chains/1.17.0` and `/sims/chains/latest/1.16` redirects to `/sims/chains/1.16.4`.
This will be difficult to manage using our current process, where the .htaccess file is recreated on every production deploy based on the current version being deployed. We discussed three possible solutions.
1. The build-server scans the directory structure on every production deploy, determines all of the latest maintenance releases based on a parsing of the file system, and writes a new .htaccess file.
2. The build server edits the .htaccess file in place, using string parsing to determine if a line needs to be changed or a new line added.
3. We add a database backend to the build server which maintains a record of all phet-io deploys. On a production deploy, the build-server queries the database of all latest minor versions and rewrites the .htaccess from scratch.
(3) is probably the most robust and scalable solution, but would require some time to figure out how to integrate a db. | 1.0 | Refining the latest links for phet-io - @jonathanolson, @zepumph, and I discussed the possibility of changing the `/latest` links for phet-io to instead be `/latest/{{major.minor}}`. The reasoning is this will allow clients to automatically pick up non-breaking maintenance releases, but avoid potentially api-breaking minor/major releases.
Here is an example of how the redirect structure would look in the .htaccess file:
```
RewriteEngine on
RewriteBase /sims/chains/
RewriteRule latest/1.17(.*) 1.17.0$1
RewriteRule latest/1.16(.*) 1.16.4$1
```
Here, the link `/sims/chains/latest/1.17` redirects to `/sims/chains/1.17.0` and `/sims/chains/latest/1.16` redirects to `/sims/chains/1.16.4`.
This will be difficult to manage using our current process, where the .htaccess file is recreated on every production deploy based on the current version being deployed. We discussed three possible solutions.
1. The build-server scans the directory structure on every production deploy, determines all of the latest maintenance releases based on a parsing of the file system, and writes a new .htaccess file.
2. The build server edits the .htaccess file in place, using string parsing to determine if a line needs to be changed or a new line added.
3. We add a database backend to the build server which maintains a record of all phet-io deploys. On a production deploy, the build-server queries the database of all latest minor versions and rewrites the .htaccess from scratch.
(3) is probably the most robust and scalable solution, but would require some time to figure out how to integrate a db. | priority | refining the latest links for phet io jonathanolson zepumph and i discussed the possibility of changing the latest links for phet io to instead be latest major minor the reasoning is this will allow clients to automatically pick up non breaking maintenance releases but avoid potentially api breaking minor major releases here is an example of how the redirect structure would look in the htaccess file rewriteengine on rewritebase sims chains rewriterule latest rewriterule latest here the link sims chains latest redirects to sims chains and sims chains latest redirects to sims chains this will be difficult to manage using our current process where the htaccess file is recreated on every production deploy based on the current version being deployed we discussed three possible solutions the build server scans the directory structure on every production deploy determines all of the latest maintenance releases based on a parsing of the file system and writes a new htaccess file the build server edits the htaccess file in place using string parsing to determine if a line needs to be changed or a new line added we add a database backend to the build server which maintains a record of all phet io deploys on a production deploy the build server queries the database of all latest minor versions and rewrites the htaccess from scratch is probably the most robust and scalable solution but would require some time to figure out how to integrate a db | 1 |
364,297 | 10,761,848,769 | IssuesEvent | 2019-10-31 21:46:54 | seakers/ABSP-EOSS | https://api.github.com/repos/seakers/ABSP-EOSS | closed | planner: Add post-mortum list updates | High Priority bug | After an agent has died, it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation | 1.0 | planner: Add post-mortum list updates - After an agent has died, it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation | priority | planner add post mortum list updates after an agent has died it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation | 1 |
445,623 | 12,833,895,185 | IssuesEvent | 2020-07-07 10:04:23 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.0 staging-1644] Civis: broken UI with set conditions deleting | Category: UI Priority: High | 1. Create a law
2. In the IF section choose "Set of conditions" and then delete it

3. If section will be broken, PLUS won't ass any positions

4. After reopening law ui

| 1.0 | [0.9.0 staging-1644] Civis: broken UI with set conditions deleting - 1. Create a law
2. In the IF section choose "Set of conditions" and then delete it

3. If section will be broken, PLUS won't ass any positions

4. After reopening law ui

| priority | civis broken ui with set conditions deleting create a law in the if section choose set of conditions and then delete it if section will be broken plus won t ass any positions after reopening law ui | 1 |
621,850 | 19,597,802,602 | IssuesEvent | 2022-01-05 20:10:12 | DSpace/dspace-angular | https://api.github.com/repos/DSpace/dspace-angular | closed | Edit EPerson | Reset Password disabled | bug high priority authentication: general authentication: password component: administrative tools e/2 | **Describe the bug**
Logged on as a member of Administrator, on the Edit EPerson page, Reset Password is disabled.
**To Reproduce**
Steps to reproduce the behavior:
1. Log on as a member of Administrator.
2. Try to reset someone's password.
**Expected behavior**
A member of Administrator should be able to reset someone's password.
**Related work**
Link to any related tickets or PRs here.
| 1.0 | Edit EPerson | Reset Password disabled - **Describe the bug**
Logged on as a member of Administrator, on the Edit EPerson page, Reset Password is disabled.
**To Reproduce**
Steps to reproduce the behavior:
1. Log on as a member of Administrator.
2. Try to reset someone's password.
**Expected behavior**
A member of Administrator should be able to reset someone's password.
**Related work**
Link to any related tickets or PRs here.
| priority | edit eperson reset password disabled describe the bug logged on as a member of administrator on the edit eperson page reset password is disabled to reproduce steps to reproduce the behavior log on as a member of administrator try to reset someone s password expected behavior a member of administrator should be able to reset someone s password related work link to any related tickets or prs here | 1 |
381,912 | 11,297,700,876 | IssuesEvent | 2020-01-17 06:56:55 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Issue in editing an oauth application which is created using Application management Rest API | Affected/5.10.0-Alpha2 Priority/Highest Severity/Critical Type/Bug | I am using Application management Rest API to create oauth application in IS-5.10.0-alpha pack. Once I create an oauth app using Rest API and when I'm trying to edit this app it gives a blank page.

| 1.0 | Issue in editing an oauth application which is created using Application management Rest API - I am using Application management Rest API to create oauth application in IS-5.10.0-alpha pack. Once I create an oauth app using Rest API and when I'm trying to edit this app it gives a blank page.

| priority | issue in editing an oauth application which is created using application management rest api i am using application management rest api to create oauth application in is alpha pack once i create an oauth app using rest api and when i m trying to edit this app it gives a blank page | 1 |
248,180 | 7,928,251,050 | IssuesEvent | 2018-07-06 10:55:28 | checkorg1/test | https://api.github.com/repos/checkorg1/test | closed | Greek characters in Genes collection | Highest Priority | _From egarciarivera on 13-06-2018 00:32 IST_
deepcompute-agent the Genes collection source file contains several tokens with greek characters that are not appearing on the staging collection. Can we check what the error is here? Source file is here: https://www.dropbox.com/s/s8vuebqcpa8eo7h/Nferx_Genes.csv?dl=0
**Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/322] | 1.0 | Greek characters in Genes collection - _From egarciarivera on 13-06-2018 00:32 IST_
deepcompute-agent the Genes collection source file contains several tokens with greek characters that are not appearing on the staging collection. Can we check what the error is here? Source file is here: https://www.dropbox.com/s/s8vuebqcpa8eo7h/Nferx_Genes.csv?dl=0
**Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/322] | priority | greek characters in genes collection from egarciarivera on ist deepcompute agent the genes collection source file contains several tokens with greek characters that are not appearing on the staging collection can we check what the error is here source file is here copied from original issue | 1 |
377,157 | 11,165,295,367 | IssuesEvent | 2019-12-27 08:52:33 | VirtoCommerce/vc-platform | https://api.github.com/repos/VirtoCommerce/vc-platform | closed | Improved frontend search UI | Epic Feature Priority: High | **Purpose**
Improve frontend search so that it will consider autocomplete, recommendation, empty results, spell checking, search pages and blogs
**Requirements**
- [ ] ?
**Related issues**
- [ ] #950 Storefront: Add relevant suggestion for search phrase
**User stories**
- [ ] ?
**Acceptance criteria**
- [ ] ?
**Definition of Done**
- [ ] ? | 1.0 | Improved frontend search UI - **Purpose**
Improve frontend search so that it will consider autocomplete, recommendation, empty results, spell checking, search pages and blogs
**Requirements**
- [ ] ?
**Related issues**
- [ ] #950 Storefront: Add relevant suggestion for search phrase
**User stories**
- [ ] ?
**Acceptance criteria**
- [ ] ?
**Definition of Done**
- [ ] ? | priority | improved frontend search ui purpose improve frontend search so that it will consider autocomplete recommendation empty results spell checking search pages and blogs requirements related issues storefront add relevant suggestion for search phrase user stories acceptance criteria definition of done | 1 |
325,109 | 9,917,071,214 | IssuesEvent | 2019-06-28 22:16:52 | amgrubb/BloomingLeaf | https://api.github.com/repos/amgrubb/BloomingLeaf | closed | Deleted Elements remain in the List of Assignments | HIGH PRIORITY bug | I created a big model and ran some simulations.
Then I deleted half the model elements and the deleted elements remained List of Assignments Table (see screenshot of model and View List of Assignments Window).
Branch: Develop

| 1.0 | Deleted Elements remain in the List of Assignments - I created a big model and ran some simulations.
Then I deleted half the model elements and the deleted elements remained List of Assignments Table (see screenshot of model and View List of Assignments Window).
Branch: Develop

| priority | deleted elements remain in the list of assignments i created a big model and ran some simulations then i deleted half the model elements and the deleted elements remained list of assignments table see screenshot of model and view list of assignments window branch develop | 1 |
816,551 | 30,602,469,033 | IssuesEvent | 2023-07-22 15:11:23 | alex4401/mediawiki-extensions-DataMaps | https://api.github.com/repos/alex4401/mediawiki-extensions-DataMaps | closed | Change zoom range via data | 1. feature 3. priority: high affects wiki.gg 1. refactoring | Replace zoom Leaflet parameters with a `zoomRange` or `zoom` property. Replace `disableZoom` with `zoom: false`.
This will require a data migration bot on wiki.gg. | 1.0 | Change zoom range via data - Replace zoom Leaflet parameters with a `zoomRange` or `zoom` property. Replace `disableZoom` with `zoom: false`.
This will require a data migration bot on wiki.gg. | priority | change zoom range via data replace zoom leaflet parameters with a zoomrange or zoom property replace disablezoom with zoom false this will require a data migration bot on wiki gg | 1 |
220,487 | 7,360,332,313 | IssuesEvent | 2018-03-10 17:30:14 | bounswe/bounswe2018group5 | https://api.github.com/repos/bounswe/bounswe2018group5 | opened | Revise Requirements | Effort: Medium Priority: High Status: Available Type: Wiki | Per Cihat's comment:
> * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities.
> * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides. | 1.0 | Revise Requirements - Per Cihat's comment:
> * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities.
> * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides. | priority | revise requirements per cihat s comment i am just lazy to check the previous version of your requirements and my feedback however one thing just got my attention for user requirements it seems that you have grouped the requirements according to the user types it could be better if you cluster them according to the functionalities one more comment instead of the user shall be able to see you can write them as system requirements since seeing is not something like a user requirement it just sees it if the system provides | 1 |
145,069 | 5,558,139,994 | IssuesEvent | 2017-03-24 14:03:34 | Valhalla-Gaming/Tracker | https://api.github.com/repos/Valhalla-Gaming/Tracker | closed | [Death Knight][Artifact] Gravitational Pull | Class-Death Knight Priority-High Type-Spell | **Describe the issue you're having**: Doesn't work at all
**Explain how you expect it work**: When using death grip, the enemy should be slowed by 50% for 6 seconds.
**Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=191721/gravitational-pull
| 1.0 | [Death Knight][Artifact] Gravitational Pull - **Describe the issue you're having**: Doesn't work at all
**Explain how you expect it work**: When using death grip, the enemy should be slowed by 50% for 6 seconds.
**Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=191721/gravitational-pull
| priority | gravitational pull describe the issue you re having doesn t work at all explain how you expect it work when using death grip the enemy should be slowed by for seconds links to wowhead youtube etc | 1 |
91,983 | 3,864,031,990 | IssuesEvent | 2016-04-08 12:14:59 | eXistSolutions/hsg-shell | https://api.github.com/repos/eXistSolutions/hsg-shell | closed | Propose & implement approach for secrets | help wanted high priority | where needed:
- s3: used on http://localhost:8080/exist/apps/hsg-shell/historicaldocuments/ebooks and FRUS volume landing pages (the latter not yet implemented in new site)
- twitter: used currently on homepage
- tumblr: used currently on homepage
- eXist user account info: used in our ant scripts for populating an empty database and uploading/deleting content via oXygen | 1.0 | Propose & implement approach for secrets - where needed:
- s3: used on http://localhost:8080/exist/apps/hsg-shell/historicaldocuments/ebooks and FRUS volume landing pages (the latter not yet implemented in new site)
- twitter: used currently on homepage
- tumblr: used currently on homepage
- eXist user account info: used in our ant scripts for populating an empty database and uploading/deleting content via oXygen | priority | propose implement approach for secrets where needed used on and frus volume landing pages the latter not yet implemented in new site twitter used currently on homepage tumblr used currently on homepage exist user account info used in our ant scripts for populating an empty database and uploading deleting content via oxygen | 1 |
304,082 | 9,320,822,208 | IssuesEvent | 2019-03-27 01:00:55 | cuappdev/ithaca-transit-backend | https://api.github.com/repos/cuappdev/ithaca-transit-backend | closed | Show delayed routes on route display page | Priority: High | - Right now, if it is past 11:15pm, we don't see the first bus route even though we should since its delayed

| 1.0 | Show delayed routes on route display page - - Right now, if it is past 11:15pm, we don't see the first bus route even though we should since its delayed

| priority | show delayed routes on route display page right now if it is past we don t see the first bus route even though we should since its delayed | 1 |
504,979 | 14,625,257,019 | IssuesEvent | 2020-12-23 08:10:23 | bounswe/bounswe2020group7 | https://api.github.com/repos/bounswe/bounswe2020group7 | closed | Android - Upload Profile Photo Implementation | Priority: High Status: In Review Subteam: Android Type: New Feature | A ui option must be provided to upload profile photo to user. | 1.0 | Android - Upload Profile Photo Implementation - A ui option must be provided to upload profile photo to user. | priority | android upload profile photo implementation a ui option must be provided to upload profile photo to user | 1 |
461,976 | 13,239,182,225 | IssuesEvent | 2020-08-19 02:40:09 | bazzinotti/snestracker | https://api.github.com/repos/bazzinotti/snestracker | closed | [PatternEditor] cursor bug when switching to pattern of less rows | bug closed-in-earlypaid high priority | To Reproduce:
- Create one pattern (00) with default length $40.
- Create another pattern (01), but this time set the length to much lower.
- Now, while in pattern (01), click the cursor on a high numbered row near the bottom of the screen. The highlight bar should move there as normal.
- Finally, switch back to pattern (00). The highlight bar stays floating in black space because pattern (00) doesn't have as many rows.
Solution: Add some logic when switching rows.
(Pardon the horrid scaling in the photos)
 -> 
| 1.0 | [PatternEditor] cursor bug when switching to pattern of less rows - To Reproduce:
- Create one pattern (00) with default length $40.
- Create another pattern (01), but this time set the length to much lower.
- Now, while in pattern (01), click the cursor on a high numbered row near the bottom of the screen. The highlight bar should move there as normal.
- Finally, switch back to pattern (00). The highlight bar stays floating in black space because pattern (00) doesn't have as many rows.
Solution: Add some logic when switching rows.
(Pardon the horrid scaling in the photos)
 -> 
| priority | cursor bug when switching to pattern of less rows to reproduce create one pattern with default length create another pattern but this time set the length to much lower now while in pattern click the cursor on a high numbered row near the bottom of the screen the highlight bar should move there as normal finally switch back to pattern the highlight bar stays floating in black space because pattern doesn t have as many rows solution add some logic when switching rows pardon the horrid scaling in the photos | 1 |
511,305 | 14,857,988,571 | IssuesEvent | 2021-01-18 16:09:34 | ooni/run | https://api.github.com/repos/ooni/run | closed | Extremely high page size of links shared | bug priority/high | Simple typical links like [this](https://run.ooni.io/nettest\?tn\=web_connectivity\&ta\=%7B%22urls%22%3A%5B%22https%3A%2F%2Ftwitter.com%22%5D%7D\&mv\=1.2.0) when opened are too heavy. Close to 15MB.

Link to [webpage-size-checker-tool](https://www.seoptimer.com/web-page-size-check/run.ooni.io/nettest?tn=web_connectivity&ta=%7b%22urls%22%3a%5b%22https%3a%2f%2ftwitter.com%22%5d%7d&mv=1.2.0) | 1.0 | Extremely high page size of links shared - Simple typical links like [this](https://run.ooni.io/nettest\?tn\=web_connectivity\&ta\=%7B%22urls%22%3A%5B%22https%3A%2F%2Ftwitter.com%22%5D%7D\&mv\=1.2.0) when opened are too heavy. Close to 15MB.

Link to [webpage-size-checker-tool](https://www.seoptimer.com/web-page-size-check/run.ooni.io/nettest?tn=web_connectivity&ta=%7b%22urls%22%3a%5b%22https%3a%2f%2ftwitter.com%22%5d%7d&mv=1.2.0) | priority | extremely high page size of links shared simple typical links like when opened are too heavy close to link to | 1 |
618,396 | 19,440,711,616 | IssuesEvent | 2021-12-22 00:13:07 | CityOfDetroit/bloom | https://api.github.com/repos/CityOfDetroit/bloom | closed | Add ADA Features Filter | size: 5 P0 epic: filtering M17 HighestPriority | Add filtering based on ADA accessibility features, on the building or on the listing. See #609 for the fields added to the database.
Some notes:
- We only have data on the building, not the specific unit, so that will need to be reflected in the frontend filter (see #591)
- Should the filter be a checkbox, or dropdown that lets you select "some units in building have features" or "all units in building have features"? | 1.0 | Add ADA Features Filter - Add filtering based on ADA accessibility features, on the building or on the listing. See #609 for the fields added to the database.
Some notes:
- We only have data on the building, not the specific unit, so that will need to be reflected in the frontend filter (see #591)
- Should the filter be a checkbox, or dropdown that lets you select "some units in building have features" or "all units in building have features"? | priority | add ada features filter add filtering based on ada accessibility features on the building or on the listing see for the fields added to the database some notes we only have data on the building not the specific unit so that will need to be reflected in the frontend filter see should the filter be a checkbox or dropdown that lets you select some units in building have features or all units in building have features | 1 |
618,254 | 19,430,813,345 | IssuesEvent | 2021-12-21 11:43:48 | turbot/steampipe-plugin-aws | https://api.github.com/repos/turbot/steampipe-plugin-aws | closed | Queries against aws_ebs_snapshot are not returning results for aws ap-southeast-2 region | bug priority:high | **Describe the bug**
Queries against aws_ebs_snapshot are not returning results in the ap-southeast-2 region
**Steampipe version (`steampipe -v`)**
v0.10.0
**Plugin version (`steampipe plugin list`)**
v0.41.0
**To reproduce**
1) Create a snapshot in ap-southeast-2
2) query `select title, region from aws_ebs_snapshot`
**Expected behavior**
Returns results from existing snapshots.
| 1.0 | Queries against aws_ebs_snapshot are not returning results for aws ap-southeast-2 region - **Describe the bug**
Queries against aws_ebs_snapshot are not returning results in the ap-southeast-2 region
**Steampipe version (`steampipe -v`)**
v0.10.0
**Plugin version (`steampipe plugin list`)**
v0.41.0
**To reproduce**
1) Create a snapshot in ap-southeast-2
2) query `select title, region from aws_ebs_snapshot`
**Expected behavior**
Returns results from existing snapshots.
| priority | queries against aws ebs snapshot are not returning results for aws ap southeast region describe the bug queries against aws ebs snapshot are not returning results in the ap southeast region steampipe version steampipe v plugin version steampipe plugin list to reproduce create a snapshot in ap southeast query select title region from aws ebs snapshot expected behavior returns results from existing snapshots | 1 |
48,909 | 3,000,832,732 | IssuesEvent | 2015-07-24 06:33:55 | jayway/powermock | https://api.github.com/repos/jayway/powermock | closed | Test methods are not executed if PowerMockRule declaration is present. | bug imported Milestone-Release1.4 Priority-High | _From [Yaroslav...@gmail.com](https://code.google.com/u/107516775877747812247/) on July 16, 2010 23:42:18_
If PowerMockRule declaration is present in a test class, test methods annotated with @Test are not executed.
Example test below passes without any warnings, while one of test methods throws NPE:
@PrepareForTest(Static.class)
public class TestCase {
@Rule
public PowerMockRule rule = new PowerMockRule ();
@Test
public void testMethodThatThrowsNPE() {
throw new NullPointerException();
}
}
If rule declaration is commented out, test fails as expected.
Quick debug shown that Runnable created using "evaluate()" method of PowerMockStatement is never executed. It seems, like this Runnable is intended to be executed in "Object execute(Object instance, Method method, Object... arguments)" method of ClassloaderExecutor, but it does not, because there is a check based on return type, and return type is expected to be non-void (line 61).
PowerMock Ver.: 1.3.9 (Mockito API, with dependencies)
Java: 1.6.0_21-b06
Eclipse (Helios, Build id: 20100617-1415) plugin was used to run the tests.
_Original issue: http://code.google.com/p/powermock/issues/detail?id=268_ | 1.0 | Test methods are not executed if PowerMockRule declaration is present. - _From [Yaroslav...@gmail.com](https://code.google.com/u/107516775877747812247/) on July 16, 2010 23:42:18_
If PowerMockRule declaration is present in a test class, test methods annotated with @Test are not executed.
Example test below passes without any warnings, while one of test methods throws NPE:
@PrepareForTest(Static.class)
public class TestCase {
@Rule
public PowerMockRule rule = new PowerMockRule ();
@Test
public void testMethodThatThrowsNPE() {
throw new NullPointerException();
}
}
If rule declaration is commented out, test fails as expected.
Quick debug shown that Runnable created using "evaluate()" method of PowerMockStatement is never executed. It seems, like this Runnable is intended to be executed in "Object execute(Object instance, Method method, Object... arguments)" method of ClassloaderExecutor, but it does not, because there is a check based on return type, and return type is expected to be non-void (line 61).
PowerMock Ver.: 1.3.9 (Mockito API, with dependencies)
Java: 1.6.0_21-b06
Eclipse (Helios, Build id: 20100617-1415) plugin was used to run the tests.
_Original issue: http://code.google.com/p/powermock/issues/detail?id=268_ | priority | test methods are not executed if powermockrule declaration is present from on july if powermockrule declaration is present in a test class test methods annotated with test are not executed example test below passes without any warnings while one of test methods throws npe preparefortest static class public class testcase rule public powermockrule rule new powermockrule test public void testmethodthatthrowsnpe throw new nullpointerexception if rule declaration is commented out test fails as expected quick debug shown that runnable created using evaluate method of powermockstatement is never executed it seems like this runnable is intended to be executed in object execute object instance method method object arguments method of classloaderexecutor but it does not because there is a check based on return type and return type is expected to be non void line powermock ver mockito api with dependencies java eclipse helios build id plugin was used to run the tests original issue | 1 |
86,739 | 3,728,766,271 | IssuesEvent | 2016-03-07 02:34:09 | Benrnz/BudgetAnalyser | https://api.github.com/repos/Benrnz/BudgetAnalyser | closed | As a reconciliation user I need to see opening balances on ledgers | enhancement Priority-high | So I can double check consistency of data and see if the ledger is increasing or decreasing. | 1.0 | As a reconciliation user I need to see opening balances on ledgers - So I can double check consistency of data and see if the ledger is increasing or decreasing. | priority | as a reconciliation user i need to see opening balances on ledgers so i can double check consistency of data and see if the ledger is increasing or decreasing | 1 |
354,999 | 10,575,561,037 | IssuesEvent | 2019-10-07 15:58:02 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] @example sections not formatted correctly | Context : JSDoc Priority: High Status: Accepted Time: ~1 hour Type: Bug wontfix | ##### **Overview of the issue**
@example code is incorrectly formatted. Html tags are displaying, `"` and other characters have been translated and output.
```
/**
* This method will display a dialog to the user to capture some arbitary text, you
* supply the buttons that should be displayed with the callback that should be invoked
* for each (if required).
*
* @param {string} title the title for the dialog
* @param {string} message the messsage to be presented to the user
* @param {*} buttonConfigs an array, containing a buttonName and callback, i.e.
* @param {boolean} [enableBackdropDismiss=true] optional parameter - allow user to dismiss the prompt by clicking
* on the backdrop (default to true)?
*
* @example
* ```typescript
* this.alertCtrl.showTextInputWithButtons(
* "Pin Verification",
* "You have been sent an email. Please check your email and then type the pin code that you have received below.",
* [
* { buttonName: "Cancel", callback: null },
* { buttonName: "Verify", callback: (value: string) => { this.sendPin(value); } }
* ]);
* ```
*
* @memberOf IonicAlerterService
*/
public abstract showTextInputWithButtons(
title: string,
message: string,
buttonConfigs: any,
enableBackdropDismiss?: boolean,
): void;
```
version: 1.0.1

Has changed post version 1.0.0-beta.13 example below is as expected

##### **Operating System, Node.js, npm, compodoc version(s)**
Node : v8.2.1
npm : 5.3.0
OS : macOS Sierra
"@compodoc/compodoc": "1.0.1"
##### **Angular configuration, a `package.json` file in the root folder**
"@angular/animations": "4.1.3",
"@angular/common": "4.1.3",
"@angular/compiler": "4.1.3",
"@angular/compiler-cli": "4.1.3",
"@angular/core": "4.1.3",
"@angular/forms": "4.1.3",
"@angular/http": "4.1.3",
"@angular/platform-browser": "4.1.3",
"@angular/platform-browser-dynamic": "4.1.3",
"@angular/platform-server": "4.1.3",
"@compodoc/compodoc": "1.0.1",
"@types/jasmine": "2.5.41",
"@types/node": "7.0.5",
"codelyzer": "^3.1.2",
"ionic-angular": "3.6.0",
"ncp": "^2.0.0",
"rexreplace": "^2.5.2",
"rimraf": "^2.6.1",
"rxjs": "5.4.0",
"ts-node": "2.0.0",
"tslint": "^5.6.0",
"tslint-consistent-codestyle": "^1.6.0",
"tslint-eslint-rules": "^4.1.1",
"tslint-ionic-rules": "^0.0.11",
"typescript": "2.3.4",
"zone.js": "0.8.12"
##### **Compodoc installed globally or locally ?**
locally
##### **Motivation for or Use Case**
See diagrams
##### **Reproduce the error**
Example above
##### **Related issues**
There was some changes to @example between 1.0.0-beta.13 and 1.0.0-beta.14
##### **Suggest a Fix**
| 1.0 | [BUG] @example sections not formatted correctly - ##### **Overview of the issue**
@example code is incorrectly formatted. Html tags are displaying, `"` and other characters have been translated and output.
```
/**
* This method will display a dialog to the user to capture some arbitary text, you
* supply the buttons that should be displayed with the callback that should be invoked
* for each (if required).
*
* @param {string} title the title for the dialog
* @param {string} message the messsage to be presented to the user
* @param {*} buttonConfigs an array, containing a buttonName and callback, i.e.
* @param {boolean} [enableBackdropDismiss=true] optional parameter - allow user to dismiss the prompt by clicking
* on the backdrop (default to true)?
*
* @example
* ```typescript
* this.alertCtrl.showTextInputWithButtons(
* "Pin Verification",
* "You have been sent an email. Please check your email and then type the pin code that you have received below.",
* [
* { buttonName: "Cancel", callback: null },
* { buttonName: "Verify", callback: (value: string) => { this.sendPin(value); } }
* ]);
* ```
*
* @memberOf IonicAlerterService
*/
public abstract showTextInputWithButtons(
title: string,
message: string,
buttonConfigs: any,
enableBackdropDismiss?: boolean,
): void;
```
version: 1.0.1

Has changed post version 1.0.0-beta.13 example below is as expected

##### **Operating System, Node.js, npm, compodoc version(s)**
Node : v8.2.1
npm : 5.3.0
OS : macOS Sierra
"@compodoc/compodoc": "1.0.1"
##### **Angular configuration, a `package.json` file in the root folder**
"@angular/animations": "4.1.3",
"@angular/common": "4.1.3",
"@angular/compiler": "4.1.3",
"@angular/compiler-cli": "4.1.3",
"@angular/core": "4.1.3",
"@angular/forms": "4.1.3",
"@angular/http": "4.1.3",
"@angular/platform-browser": "4.1.3",
"@angular/platform-browser-dynamic": "4.1.3",
"@angular/platform-server": "4.1.3",
"@compodoc/compodoc": "1.0.1",
"@types/jasmine": "2.5.41",
"@types/node": "7.0.5",
"codelyzer": "^3.1.2",
"ionic-angular": "3.6.0",
"ncp": "^2.0.0",
"rexreplace": "^2.5.2",
"rimraf": "^2.6.1",
"rxjs": "5.4.0",
"ts-node": "2.0.0",
"tslint": "^5.6.0",
"tslint-consistent-codestyle": "^1.6.0",
"tslint-eslint-rules": "^4.1.1",
"tslint-ionic-rules": "^0.0.11",
"typescript": "2.3.4",
"zone.js": "0.8.12"
##### **Compodoc installed globally or locally ?**
locally
##### **Motivation for or Use Case**
See diagrams
##### **Reproduce the error**
Example above
##### **Related issues**
There was some changes to @example between 1.0.0-beta.13 and 1.0.0-beta.14
##### **Suggest a Fix**
| priority | example sections not formatted correctly overview of the issue example code is incorrectly formatted html tags are displaying quot and other characters have been translated and output this method will display a dialog to the user to capture some arbitary text you supply the buttons that should be displayed with the callback that should be invoked for each if required param string title the title for the dialog param string message the messsage to be presented to the user param buttonconfigs an array containing a buttonname and callback i e param boolean optional parameter allow user to dismiss the prompt by clicking on the backdrop default to true example typescript this alertctrl showtextinputwithbuttons pin verification you have been sent an email please check your email and then type the pin code that you have received below buttonname cancel callback null buttonname verify callback value string this sendpin value memberof ionicalerterservice public abstract showtextinputwithbuttons title string message string buttonconfigs any enablebackdropdismiss boolean void version has changed post version beta example below is as expected operating system node js npm compodoc version s node npm os macos sierra compodoc compodoc angular configuration a package json file in the root folder angular animations angular common angular compiler angular compiler cli angular core angular forms angular http angular platform browser angular platform browser dynamic angular platform server compodoc compodoc types jasmine types node codelyzer ionic angular ncp rexreplace rimraf rxjs ts node tslint tslint consistent codestyle tslint eslint rules tslint ionic rules typescript zone js compodoc installed globally or locally locally motivation for or use case see diagrams reproduce the error example above related issues there was some changes to example between beta and beta suggest a fix | 1 |
423,891 | 12,303,412,525 | IssuesEvent | 2020-05-11 18:39:04 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | connect.secure.wellsfargo.com - see bug description | browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52765 -->
**URL**: https://connect.secure.wellsfargo.com/auth/login/present?origin=mobilebrowser
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0
**Tested Another Browser**: Yes Opera
**Problem type**: Something else
**Description**: unable to sign in
**Steps to Reproduce**:
yried to sin in, said didnt reconize my username and/or password even tho i used the correct ones over and over again.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190406020101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/5/77b9bd01-4de6-4a40-b0d9-e0f9aff989b9)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | connect.secure.wellsfargo.com - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52765 -->
**URL**: https://connect.secure.wellsfargo.com/auth/login/present?origin=mobilebrowser
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0
**Tested Another Browser**: Yes Opera
**Problem type**: Something else
**Description**: unable to sign in
**Steps to Reproduce**:
yried to sin in, said didnt reconize my username and/or password even tho i used the correct ones over and over again.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190406020101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/5/77b9bd01-4de6-4a40-b0d9-e0f9aff989b9)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | connect secure wellsfargo com see bug description url browser version firefox mobile operating system android tested another browser yes opera problem type something else description unable to sign in steps to reproduce yried to sin in said didnt reconize my username and or password even tho i used the correct ones over and over again browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel alpha hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
274,224 | 8,558,677,568 | IssuesEvent | 2018-11-08 18:58:22 | spacejam/sled | https://api.github.com/repos/spacejam/sled | closed | multithread write seems to lost data | bug high priority | Bug reports must include:
1. sled version
0.16.2
Hi,I wrote a simple multi-threaded write / read test for sled in [https://github.com/WenjunHuang/SledMultiThreadTest](https://github.com/WenjunHuang/SledMultiThreadTest)
I found that when i first used multi-threaded to write some data to sled and then __restart__ the program to start as multi-thread read, some data will lost.
But if i first write then read(without restart the program),everything was ok.
And If i used single thread write and read, everything was ok whether i restart the program after write or not.
| 1.0 | multithread write seems to lost data - Bug reports must include:
1. sled version
0.16.2
Hi,I wrote a simple multi-threaded write / read test for sled in [https://github.com/WenjunHuang/SledMultiThreadTest](https://github.com/WenjunHuang/SledMultiThreadTest)
I found that when i first used multi-threaded to write some data to sled and then __restart__ the program to start as multi-thread read, some data will lost.
But if i first write then read(without restart the program),everything was ok.
And If i used single thread write and read, everything was ok whether i restart the program after write or not.
| priority | multithread write seems to lost data bug reports must include sled version hi i wrote a simple multi threaded write read test for sled in i found that when i first used multi threaded to write some data to sled and then restart the program to start as multi thread read some data will lost but if i first write then read without restart the program everything was ok and if i used single thread write and read everything was ok whether i restart the program after write or not | 1 |
509,846 | 14,750,213,939 | IssuesEvent | 2021-01-08 01:32:08 | alibaba/nacos | https://api.github.com/repos/alibaba/nacos | closed | JRaft RPC :Error in object information registration returned | area/Nacos Core kind/bug priority/high | <!-- Here is for bug reports and feature requests ONLY!
If you're looking for help, please check our mail list、WeChat group and the Gitter room.
Please try to use English to describe your issue, or at least provide a snippet of English translation.
我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。
-->
**Describe the bug**

**Expected behavior**
A clear and concise description of what you expected to happen.
**Acutally behavior**
A clear and concise description of what you actually to happen.
**How to Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Desktop (please complete the following information):**
- OS: [e.g. Centos]
- Version [e.g. nacos-server 1.3.1, nacos-client 1.3.1]
- Module [e.g. naming/config]
- SDK [e.g. original, spring-cloud-alibaba-nacos, dubbo]
**Additional context**
Add any other context about the problem here.
| 1.0 | JRaft RPC :Error in object information registration returned - <!-- Here is for bug reports and feature requests ONLY!
If you're looking for help, please check our mail list、WeChat group and the Gitter room.
Please try to use English to describe your issue, or at least provide a snippet of English translation.
我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。
-->
**Describe the bug**

**Expected behavior**
A clear and concise description of what you expected to happen.
**Acutally behavior**
A clear and concise description of what you actually to happen.
**How to Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Desktop (please complete the following information):**
- OS: [e.g. Centos]
- Version [e.g. nacos-server 1.3.1, nacos-client 1.3.1]
- Module [e.g. naming/config]
- SDK [e.g. original, spring-cloud-alibaba-nacos, dubbo]
**Additional context**
Add any other context about the problem here.
| priority | jraft rpc :error in object information registration returned here is for bug reports and feature requests only if you re looking for help please check our mail list、wechat group and the gitter room please try to use english to describe your issue or at least provide a snippet of english translation 我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。 describe the bug expected behavior a clear and concise description of what you expected to happen acutally behavior a clear and concise description of what you actually to happen how to reproduce steps to reproduce the behavior go to click on scroll down to see error desktop please complete the following information os version module sdk additional context add any other context about the problem here | 1 |
396,291 | 11,707,061,953 | IssuesEvent | 2020-03-08 03:06:43 | sjsumlclub/Website | https://api.github.com/repos/sjsumlclub/Website | opened | Add Committees Section | high priority | Here is the text for committees below
# Committees
Our meetings alternate between _Committee Days_ and _Club Days_. Committee Days let members become experts in their topic of interest, while Club Days promote symbiotic learning as a community.
- **Committee Days**: members work within their committee to learn, share resources, and complete projects as a team.
- **Club Days**: Committees rotate to host workshops, paper readings, presentations, or fun games (e.g. trivia) for symbiotic club learning. Learners are exposed to important concepts from other parts of AI, and presenters refine their knowledge by teaching others.
## Computer Vision Committee
The Computer Vision committee is dedicated to retrieve, process, and analyze digital images with deep learning. Our main focus is on object recognition and classification using a variety of convolutional neural networks. We aim to gain expertise in vision models through hands-on experience.
## Reinforcement Learning Committee
The Reinforcement Learning committee is focused on teaching its members Reinforcement Learning's fundamentals, its applications, and its possibilities. We find exciting projects and case studies to analyze and learn from. Through leading workshops and tutorials, the RL committee teaches actionable skills for members to utilize in their own projects.
## Graphics Committee
The Graphics committee explores the use of deep learning for generative graphics, i.e. with Generative Adversarial Networks. These topics include interactive rendering, visualization, art, as well as virtual and augmented reality applications. We aim to create beneficial applications with AI visual imagination. Members will gain practical implementation and research skills.
## Natural Language Committee
COMING SOON
## Traditional ML Committee
The Traditional Machine Learning committee encompasses both traditional machine learning models such as Random Forests, K-Nearest Neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization, dimensionality reduction, cross-validation. We seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks.
| 1.0 | Add Committees Section - Here is the text for committees below
# Committees
Our meetings alternate between _Committee Days_ and _Club Days_. Committee Days let members become experts in their topic of interest, while Club Days promote symbiotic learning as a community.
- **Committee Days**: members work within their committee to learn, share resources, and complete projects as a team.
- **Club Days**: Committees rotate to host workshops, paper readings, presentations, or fun games (e.g. trivia) for symbiotic club learning. Learners are exposed to important concepts from other parts of AI, and presenters refine their knowledge by teaching others.
## Computer Vision Committee
The Computer Vision committee is dedicated to retrieve, process, and analyze digital images with deep learning. Our main focus is on object recognition and classification using a variety of convolutional neural networks. We aim to gain expertise in vision models through hands-on experience.
## Reinforcement Learning Committee
The Reinforcement Learning committee is focused on teaching its members Reinforcement Learning's fundamentals, its applications, and its possibilities. We find exciting projects and case studies to analyze and learn from. Through leading workshops and tutorials, the RL committee teaches actionable skills for members to utilize in their own projects.
## Graphics Committee
The Graphics committee explores the use of deep learning for generative graphics, i.e. with Generative Adversarial Networks. These topics include interactive rendering, visualization, art, as well as virtual and augmented reality applications. We aim to create beneficial applications with AI visual imagination. Members will gain practical implementation and research skills.
## Natural Language Committee
COMING SOON
## Traditional ML Committee
The Traditional Machine Learning committee encompasses both traditional machine learning models such as Random Forests, K-Nearest Neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization, dimensionality reduction, cross-validation. We seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks.
| priority | add committees section here is the text for committees below committees our meetings alternate between committee days and club days committee days let members become experts in their topic of interest while club days promote symbiotic learning as a community committee days members work within their committee to learn share resources and complete projects as a team club days committees rotate to host workshops paper readings presentations or fun games e g trivia for symbiotic club learning learners are exposed to important concepts from other parts of ai and presenters refine their knowledge by teaching others computer vision committee the computer vision committee is dedicated to retrieve process and analyze digital images with deep learning our main focus is on object recognition and classification using a variety of convolutional neural networks we aim to gain expertise in vision models through hands on experience reinforcement learning committee the reinforcement learning committee is focused on teaching its members reinforcement learning s fundamentals its applications and its possibilities we find exciting projects and case studies to analyze and learn from through leading workshops and tutorials the rl committee teaches actionable skills for members to utilize in their own projects graphics committee the graphics committee explores the use of deep learning for generative graphics i e with generative adversarial networks these topics include interactive rendering visualization art as well as virtual and augmented reality applications we aim to create beneficial applications with ai visual imagination members will gain practical implementation and research skills natural language committee coming soon traditional ml committee the traditional machine learning committee encompasses both traditional machine learning models such as random forests k nearest neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization dimensionality reduction cross validation we seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks | 1 |
422,404 | 12,277,877,796 | IssuesEvent | 2020-05-08 08:49:38 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | github.com - see bug description | browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52612 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://github.com/gorhill/uBlock/releases
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: cannot install the beta version
**Steps to Reproduce**:
Should install instead of download
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | github.com - see bug description - <!-- @browser: Firefox Mobile 77.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52612 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://github.com/gorhill/uBlock/releases
**Browser / Version**: Firefox Mobile 77.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: cannot install the beta version
**Steps to Reproduce**:
Should install instead of download
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | github com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description cannot install the beta version steps to reproduce should install instead of download browser configuration none from with ❤️ | 1 |
83,235 | 3,632,268,035 | IssuesEvent | 2016-02-11 08:57:52 | codebuddiesdotorg/cb-v2-scratch | https://api.github.com/repos/codebuddiesdotorg/cb-v2-scratch | closed | "Currently working on" status | feature high-priority ready | A spot for every person to fill in a “current status” to describe what they’re “currently working on”. This way, you’ll be able to see at a glance what everyone is working on. | 1.0 | "Currently working on" status - A spot for every person to fill in a “current status” to describe what they’re “currently working on”. This way, you’ll be able to see at a glance what everyone is working on. | priority | currently working on status a spot for every person to fill in a “current status” to describe what they’re “currently working on” this way you’ll be able to see at a glance what everyone is working on | 1 |
671,186 | 22,747,201,915 | IssuesEvent | 2022-07-07 10:11:18 | codersforcauses/elucidate | https://api.github.com/repos/codersforcauses/elucidate | closed | Fix frontend CI pipeline | bug enhancement frontend priority::high | ## Basic Information
Rewrite pipeline so that it runs the now dockerized project
## Other Information
N/A | 1.0 | Fix frontend CI pipeline - ## Basic Information
Rewrite pipeline so that it runs the now dockerized project
## Other Information
N/A | priority | fix frontend ci pipeline basic information rewrite pipeline so that it runs the now dockerized project other information n a | 1 |
208,031 | 7,135,222,013 | IssuesEvent | 2018-01-22 23:55:57 | richrd/suplemon | https://api.github.com/repos/richrd/suplemon | opened | Rewrite: Core config handling | priority-high rewrite | The rewrite needs a config handler and now would be a good time to implement it. The idea is to make it more or less compatible with Sublime Text. The new format wont be backwards compatible with the current version of suplemon (0.1.X)
A common implementation should be used for both core config and keymap config.
- [ ] Config file format should be similar to sublime text
- [ ] Variable names should be the same as in sublime text
- [ ] Single line comments should be supported
- [ ] User configs need to be merged on top of defaults
- [ ] Defaults can be different from sublime text defaults
- [ ] Not all sublime text options will be implemented
| 1.0 | Rewrite: Core config handling - The rewrite needs a config handler and now would be a good time to implement it. The idea is to make it more or less compatible with Sublime Text. The new format wont be backwards compatible with the current version of suplemon (0.1.X)
A common implementation should be used for both core config and keymap config.
- [ ] Config file format should be similar to sublime text
- [ ] Variable names should be the same as in sublime text
- [ ] Single line comments should be supported
- [ ] User configs need to be merged on top of defaults
- [ ] Defaults can be different from sublime text defaults
- [ ] Not all sublime text options will be implemented
| priority | rewrite core config handling the rewrite needs a config handler and now would be a good time to implement it the idea is to make it more or less compatible with sublime text the new format wont be backwards compatible with the current version of suplemon x a common implementation should be used for both core config and keymap config config file format should be similar to sublime text variable names should be the same as in sublime text single line comments should be supported user configs need to be merged on top of defaults defaults can be different from sublime text defaults not all sublime text options will be implemented | 1 |
143,803 | 5,530,373,928 | IssuesEvent | 2017-03-21 02:17:38 | kamal1978/LTFHC | https://api.github.com/repos/kamal1978/LTFHC | closed | Back arrow causing crashes | app bug high priority | If I am on the following screens, pressing the back arrow will cause a crash:
- Prenatal Fansidar
- Prenatal Iron Folate
- Menbendazole
...and it seems all subsequent screens in the flow. | 1.0 | Back arrow causing crashes - If I am on the following screens, pressing the back arrow will cause a crash:
- Prenatal Fansidar
- Prenatal Iron Folate
- Menbendazole
...and it seems all subsequent screens in the flow. | priority | back arrow causing crashes if i am on the following screens pressing the back arrow will cause a crash prenatal fansidar prenatal iron folate menbendazole and it seems all subsequent screens in the flow | 1 |
404,712 | 11,861,920,798 | IssuesEvent | 2020-03-25 17:05:12 | ME-ICA/tedana | https://api.github.com/repos/ME-ICA/tedana | closed | importing tedana in python seems to lock my multiprocessing to one core | bug high-priority | <!--
This is a suggested issue template for tedana.
If there is other information that would be helpful to include, please do not hesitate to add it!
Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #.
If this is a usage question, please check out NeuroStars here:
https://neurostars.org/
and tag your topic with "multi-echo"
-->
<!--
Summarize the issue in 1-2 sentences, linking other issues if they are relevant
Note: simply typing # will prompt you for open issues to select from
-->
### Summary
I am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands. Normally afni will either automatically use all available cpus or take a manually given amount to process certain commands, however when I import tendana, nipype will only use one. Could this have to do with #188 ?
<!--
If needed, add additional detail for:
1. Recreating a bug/problem
2. Any additional context necessary to understand the issue
-->
### Additional Detail
Here is a sample code for a wrapper in my script:
```
import nipype
from multiprocessing import cpu_count
from nipype.interfaces import afni as afni
from tedana.workflows import t2smap_workflow
def warp(self,fileobj1,fileobj2,out_file,transformation=None,args=None,saved_mat_file=None):
ThreeDWarp = afni.Warp(in_file=fileobj1,out_file=out_file)
if args is not None:
ThreeDWarp.inputs.args=args
if transformation == 'card2oblique':
ThreeDWarp.inputs.oblique_parent = fileobj2
elif transformation == 'deoblique':
ThreeDWarp.inputs.deoblique = True
elif transformation == 'mni2tta':
ThreeDWarp.inputs.mni2tta = True
elif transformation == 'tta2mni':
ThreeDWarp.inputs.tta2mni = True
elif transformation == 'matrix':
ThreeDWarp.inputs.matparent = fileobj2
elif transformation == None:
print("Warning: no transformation input given")
else:
print("Warning: none of the transformation options given match the possible arguments. Matching arguments are card2oblique,"+
" deoblique, mni2tta, tta2mni, and matrix")
ThreeDWarp.inputs.num_threads = cpu_count()
if saved_mat_file: #this is for if the pipline requires saving the 1D matrix tranformation information
print('saving matrix')
ThreeDWarp.inputs.verbose = True
ThreeDWarp.inputs.save_warp = True
ThreeDWarp.run()
```
<!--
If desired, add suggested next steps.
If you foresee them in a particular order or priority, please use numbering
-->
### Next Steps
printing os.environ gives me ```({ 'OMP_NUM_THREADS': '1', 'LESSOPEN': '| /usr/bin/lesspipe %s', '_': '/usr/bin/python3', 'MKL_NUM_THREADS': '1', 'NUMEXPR_NUM_THREADS': '1', 'VECLIB_MAXIMUM_THREADS': '1', 'KMP_DUPLICATE_LIB_OK': 'True', 'KMP_INIT_AT_FORK': 'FALSE'})```
I changed OMP_NUM_THREADS and fixed my issue, but wanted to post this so that the devs here would know how this can affect nipype.
<!--
Thank you for submitting your issue!
If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention.
Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience!
-->
| 1.0 | importing tedana in python seems to lock my multiprocessing to one core - <!--
This is a suggested issue template for tedana.
If there is other information that would be helpful to include, please do not hesitate to add it!
Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #.
If this is a usage question, please check out NeuroStars here:
https://neurostars.org/
and tag your topic with "multi-echo"
-->
<!--
Summarize the issue in 1-2 sentences, linking other issues if they are relevant
Note: simply typing # will prompt you for open issues to select from
-->
### Summary
I am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands. Normally afni will either automatically use all available cpus or take a manually given amount to process certain commands, however when I import tendana, nipype will only use one. Could this have to do with #188 ?
<!--
If needed, add additional detail for:
1. Recreating a bug/problem
2. Any additional context necessary to understand the issue
-->
### Additional Detail
Here is a sample code for a wrapper in my script:
```
import nipype
from multiprocessing import cpu_count
from nipype.interfaces import afni as afni
from tedana.workflows import t2smap_workflow
def warp(self,fileobj1,fileobj2,out_file,transformation=None,args=None,saved_mat_file=None):
ThreeDWarp = afni.Warp(in_file=fileobj1,out_file=out_file)
if args is not None:
ThreeDWarp.inputs.args=args
if transformation == 'card2oblique':
ThreeDWarp.inputs.oblique_parent = fileobj2
elif transformation == 'deoblique':
ThreeDWarp.inputs.deoblique = True
elif transformation == 'mni2tta':
ThreeDWarp.inputs.mni2tta = True
elif transformation == 'tta2mni':
ThreeDWarp.inputs.tta2mni = True
elif transformation == 'matrix':
ThreeDWarp.inputs.matparent = fileobj2
elif transformation == None:
print("Warning: no transformation input given")
else:
print("Warning: none of the transformation options given match the possible arguments. Matching arguments are card2oblique,"+
" deoblique, mni2tta, tta2mni, and matrix")
ThreeDWarp.inputs.num_threads = cpu_count()
if saved_mat_file: #this is for if the pipline requires saving the 1D matrix tranformation information
print('saving matrix')
ThreeDWarp.inputs.verbose = True
ThreeDWarp.inputs.save_warp = True
ThreeDWarp.run()
```
<!--
If desired, add suggested next steps.
If you foresee them in a particular order or priority, please use numbering
-->
### Next Steps
printing os.environ gives me ```({ 'OMP_NUM_THREADS': '1', 'LESSOPEN': '| /usr/bin/lesspipe %s', '_': '/usr/bin/python3', 'MKL_NUM_THREADS': '1', 'NUMEXPR_NUM_THREADS': '1', 'VECLIB_MAXIMUM_THREADS': '1', 'KMP_DUPLICATE_LIB_OK': 'True', 'KMP_INIT_AT_FORK': 'FALSE'})```
I changed OMP_NUM_THREADS and fixed my issue, but wanted to post this so that the devs here would know how this can affect nipype.
<!--
Thank you for submitting your issue!
If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention.
Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience!
-->
| priority | importing tedana in python seems to lock my multiprocessing to one core this is a suggested issue template for tedana if there is other information that would be helpful to include please do not hesitate to add it before submitting please check to make sure that the issue is not already addressed if there is a related issue then please cross reference it by if this is a usage question please check out neurostars here and tag your topic with multi echo summarize the issue in sentences linking other issues if they are relevant note simply typing will prompt you for open issues to select from summary i am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands normally afni will either automatically use all available cpus or take a manually given amount to process certain commands however when i import tendana nipype will only use one could this have to do with if needed add additional detail for recreating a bug problem any additional context necessary to understand the issue additional detail here is a sample code for a wrapper in my script import nipype from multiprocessing import cpu count from nipype interfaces import afni as afni from tedana workflows import workflow def warp self out file transformation none args none saved mat file none threedwarp afni warp in file out file out file if args is not none threedwarp inputs args args if transformation threedwarp inputs oblique parent elif transformation deoblique threedwarp inputs deoblique true elif transformation threedwarp inputs true elif transformation threedwarp inputs true elif transformation matrix threedwarp inputs matparent elif transformation none print warning no transformation input given else print warning none of the transformation options given match the possible arguments matching arguments are deoblique and matrix threedwarp inputs num threads cpu count if saved mat file this is for if the pipline requires saving the matrix tranformation information print saving matrix threedwarp inputs verbose true threedwarp inputs save warp true threedwarp run if desired add suggested next steps if you foresee them in a particular order or priority please use numbering next steps printing os environ gives me omp num threads lessopen usr bin lesspipe s usr bin mkl num threads numexpr num threads veclib maximum threads kmp duplicate lib ok true kmp init at fork false i changed omp num threads and fixed my issue but wanted to post this so that the devs here would know how this can affect nipype thank you for submitting your issue if you do not receive a response within a calendar week please post a comment on this issue to catch our attention some issues may not be resolved right away due to the volunteer nature of the project thank you for your patience | 1 |
563,181 | 16,677,036,003 | IssuesEvent | 2021-06-07 17:32:04 | ucb-rit/coldfront | https://api.github.com/repos/ucb-rit/coldfront | opened | Track that a request is being processed | enhancement high priority | Currently, there is no way for an admin to know whether a given request (e.g., cluster account request, new project request, etc.) is currently being processed, so multiple admins may try to process the same request at the same time.
One solution may be to add an `assignee` field to each request. Admins would see that it is already assigned, and avoid processing it themselves. | 1.0 | Track that a request is being processed - Currently, there is no way for an admin to know whether a given request (e.g., cluster account request, new project request, etc.) is currently being processed, so multiple admins may try to process the same request at the same time.
One solution may be to add an `assignee` field to each request. Admins would see that it is already assigned, and avoid processing it themselves. | priority | track that a request is being processed currently there is no way for an admin to know whether a given request e g cluster account request new project request etc is currently being processed so multiple admins may try to process the same request at the same time one solution may be to add an assignee field to each request admins would see that it is already assigned and avoid processing it themselves | 1 |
680,103 | 23,258,243,502 | IssuesEvent | 2022-08-04 11:14:53 | chloebrett/mlvet | https://api.github.com/repos/chloebrett/mlvet | closed | Trying to open a project (with ctrl+O) when a project is already open should prompt a save first | high priority | We prompt a save when the user tries to quit without saving, so we should do the same when they try to open a project without saving the one they already have open | 1.0 | Trying to open a project (with ctrl+O) when a project is already open should prompt a save first - We prompt a save when the user tries to quit without saving, so we should do the same when they try to open a project without saving the one they already have open | priority | trying to open a project with ctrl o when a project is already open should prompt a save first we prompt a save when the user tries to quit without saving so we should do the same when they try to open a project without saving the one they already have open | 1 |
710,799 | 24,434,959,578 | IssuesEvent | 2022-10-06 10:48:24 | IAmTamal/Milan | https://api.github.com/repos/IAmTamal/Milan | opened | Minor UI fixes | 🟧 priority: high 🛠 goal: fix 🛠 status : under development hacktoberfest | ### Description
Hey, I was going through the project and I found some places where there's a need to work on the UI
<img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194293332-88e2bbb5-1ebe-487e-83db-7bb0ea8e586f.png">
in the above screenshot, I can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality.
<img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194293737-a724832c-5515-499f-9e90-e2121e0eec43.png">
in this screenshot above, whenever I visit this route it autofocuses to the confirm password field which is a bad UX.
in addition to all of this the footer design is not suitable we need to work on it too.
I can help you in doing all these little things fix, please assign me this task.
Thanks.
### Screenshots
<img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194294318-419b8451-58e9-4db5-b4a6-06ffa0f6aa2c.png">
<img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194294337-7e7848ce-a5b3-44d8-8cee-44edf50ddb3d.png">
### Additional information
_No response_
### 🥦 Browser
Brave
### 👀 Have you checked if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Contributing Guidelines?
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
### Are you willing to work on this issue ?
Yes I am willing to submit a PR! | 1.0 | Minor UI fixes - ### Description
Hey, I was going through the project and I found some places where there's a need to work on the UI
<img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194293332-88e2bbb5-1ebe-487e-83db-7bb0ea8e586f.png">
in the above screenshot, I can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality.
<img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194293737-a724832c-5515-499f-9e90-e2121e0eec43.png">
in this screenshot above, whenever I visit this route it autofocuses to the confirm password field which is a bad UX.
in addition to all of this the footer design is not suitable we need to work on it too.
I can help you in doing all these little things fix, please assign me this task.
Thanks.
### Screenshots
<img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194294318-419b8451-58e9-4db5-b4a6-06ffa0f6aa2c.png">
<img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194294337-7e7848ce-a5b3-44d8-8cee-44edf50ddb3d.png">
### Additional information
_No response_
### 🥦 Browser
Brave
### 👀 Have you checked if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Contributing Guidelines?
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
### Are you willing to work on this issue ?
Yes I am willing to submit a PR! | priority | minor ui fixes description hey i was going through the project and i found some places where there s a need to work on the ui img width alt screenshot at pm src in the above screenshot i can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality img width alt screenshot at pm src in this screenshot above whenever i visit this route it autofocuses to the confirm password field which is a bad ux in addition to all of this the footer design is not suitable we need to work on it too i can help you in doing all these little things fix please assign me this task thanks screenshots img width alt screenshot at pm src img width alt screenshot at pm src additional information no response 🥦 browser brave 👀 have you checked if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the contributing guidelines i have read the are you willing to work on this issue yes i am willing to submit a pr | 1 |
435,289 | 12,533,871,524 | IssuesEvent | 2020-06-04 18:22:58 | brunomargenats/mate | https://api.github.com/repos/brunomargenats/mate | closed | PHP WARNING HTTP_USER_AGENT | Priority: High Status: Completed Tech: PHP Type: Bug | if (strpos($_SERVER['HTTP_USER_AGENT'], 'MSIE') !== FALSE ||
strpos($_SERVER['HTTP_USER_AGENT'], 'Trident') !== FALSE) {
/* LOAD SUPPORT ONLY FOR INTERNET EXPLORER */
if (!empty($mate_ie_support)){
get_template_part('functions/ie-support');
}
}
The fix: https://stackoverflow.com/questions/14130830/php-undefined-index-http-user-agent | 1.0 | PHP WARNING HTTP_USER_AGENT - if (strpos($_SERVER['HTTP_USER_AGENT'], 'MSIE') !== FALSE ||
strpos($_SERVER['HTTP_USER_AGENT'], 'Trident') !== FALSE) {
/* LOAD SUPPORT ONLY FOR INTERNET EXPLORER */
if (!empty($mate_ie_support)){
get_template_part('functions/ie-support');
}
}
The fix: https://stackoverflow.com/questions/14130830/php-undefined-index-http-user-agent | priority | php warning http user agent if strpos server msie false strpos server trident false load support only for internet explorer if empty mate ie support get template part functions ie support the fix | 1 |
190,055 | 6,808,315,577 | IssuesEvent | 2017-11-04 01:04:13 | NewsDiffs/newsdiffs | https://api.github.com/repos/NewsDiffs/newsdiffs | opened | Email scraper errors to admins | enhancement high priority | Right now when the scraper has errors, it only logs them to file. At the end of scraping, if this file is non-empty, email to the Django project admins to communicate errors. | 1.0 | Email scraper errors to admins - Right now when the scraper has errors, it only logs them to file. At the end of scraping, if this file is non-empty, email to the Django project admins to communicate errors. | priority | email scraper errors to admins right now when the scraper has errors it only logs them to file at the end of scraping if this file is non empty email to the django project admins to communicate errors | 1 |
276,629 | 8,603,329,448 | IssuesEvent | 2018-11-16 16:31:48 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Migrate images to/from KVM / vCenter DS | Category: vCenter Priority: High Status: Accepted Type: Feature | **Description**
Automatic conversion to/from .vmdk to .raw/.qcow2/etc upon "oneimage clone" operations can be implemented
**Use case**
To facilitate smooth transitions between KVM and VMware based infrastructures, a
**Interface Changes**
None
## Progress Status
- [x] Branch created
- [x] Code committed to development branch
- [x] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [x] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Migrate images to/from KVM / vCenter DS - **Description**
Automatic conversion to/from .vmdk to .raw/.qcow2/etc upon "oneimage clone" operations can be implemented
**Use case**
To facilitate smooth transitions between KVM and VMware based infrastructures, a
**Interface Changes**
None
## Progress Status
- [x] Branch created
- [x] Code committed to development branch
- [x] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [x] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| priority | migrate images to from kvm vcenter ds description automatic conversion to from vmdk to raw etc upon oneimage clone operations can be implemented use case to facilitate smooth transitions between kvm and vmware based infrastructures a interface changes none progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 1 |
536,696 | 15,712,448,288 | IssuesEvent | 2021-03-27 12:10:09 | sopra-fs21-group-06/remys-best-client | https://api.github.com/repos/sopra-fs21-group-06/remys-best-client | opened | Player throws cards away and sits round out | high priority task | If no card can be played, all remaining cards of the player are put back on the deck and the player sits this round out.
Time: 7h
This task is part of user story #65 | 1.0 | Player throws cards away and sits round out - If no card can be played, all remaining cards of the player are put back on the deck and the player sits this round out.
Time: 7h
This task is part of user story #65 | priority | player throws cards away and sits round out if no card can be played all remaining cards of the player are put back on the deck and the player sits this round out time this task is part of user story | 1 |
642,408 | 20,887,247,986 | IssuesEvent | 2022-03-23 07:12:04 | AY2122S2-CS2103-F09-2/tp | https://api.github.com/repos/AY2122S2-CS2103-F09-2/tp | closed | Clear all command | type.Task priority.High | Edit the clear command to clear all friends, events and logs in the addressbook | 1.0 | Clear all command - Edit the clear command to clear all friends, events and logs in the addressbook | priority | clear all command edit the clear command to clear all friends events and logs in the addressbook | 1 |
575,790 | 17,049,935,013 | IssuesEvent | 2021-07-06 07:43:37 | openshift/odo | https://api.github.com/repos/openshift/odo | closed | Link via ServiceBindingRequest, part 2 | kind/feature kind/user-story points/2 priority/High | ## Description
As a developer, I would like to be able to easily inject my source component with the config/secrets/envvars needed to invoke a backing application, so I can focus on may application code instead of configuring application connections.
## Acceptance criteria
* User can issue link command and then push changes, when pushed it will be determined that the components are compatible for `ServiceRequestBinding` and perform the function.
* User can list linked components
* User can unlink, or remove link.
## User Stories
- [x] odo link should store link information in devfile #4208
- [x] Suport linking devfile component with another component (https://github.com/openshift/odo/issues/3423)
## Good to have
- [x] Disallow self linking of a component (https://github.com/openshift/odo/issues/1052)
- [x] Add linked component/service information in odo describe for devfile component and operator services (https://github.com/openshift/odo/issues/3734)
## JIRA Epic Tracking
[ODO-47]( https://issues.redhat.com/browse/ODO-47) | 1.0 | Link via ServiceBindingRequest, part 2 - ## Description
As a developer, I would like to be able to easily inject my source component with the config/secrets/envvars needed to invoke a backing application, so I can focus on may application code instead of configuring application connections.
## Acceptance criteria
* User can issue link command and then push changes, when pushed it will be determined that the components are compatible for `ServiceRequestBinding` and perform the function.
* User can list linked components
* User can unlink, or remove link.
## User Stories
- [x] odo link should store link information in devfile #4208
- [x] Suport linking devfile component with another component (https://github.com/openshift/odo/issues/3423)
## Good to have
- [x] Disallow self linking of a component (https://github.com/openshift/odo/issues/1052)
- [x] Add linked component/service information in odo describe for devfile component and operator services (https://github.com/openshift/odo/issues/3734)
## JIRA Epic Tracking
[ODO-47]( https://issues.redhat.com/browse/ODO-47) | priority | link via servicebindingrequest part description as a developer i would like to be able to easily inject my source component with the config secrets envvars needed to invoke a backing application so i can focus on may application code instead of configuring application connections acceptance criteria user can issue link command and then push changes when pushed it will be determined that the components are compatible for servicerequestbinding and perform the function user can list linked components user can unlink or remove link user stories odo link should store link information in devfile suport linking devfile component with another component good to have disallow self linking of a component add linked component service information in odo describe for devfile component and operator services jira epic tracking | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.